Jan 30 15:23:28.898315 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Jan 30 15:23:28.898341 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Wed Jan 29 10:12:48 -00 2025 Jan 30 15:23:28.898352 kernel: KASLR enabled Jan 30 15:23:28.898358 kernel: efi: EFI v2.7 by EDK II Jan 30 15:23:28.898364 kernel: efi: SMBIOS 3.0=0x135ed0000 MEMATTR=0x133d4d698 ACPI 2.0=0x132430018 RNG=0x13243e918 MEMRESERVE=0x13232ed18 Jan 30 15:23:28.898370 kernel: random: crng init done Jan 30 15:23:28.898377 kernel: ACPI: Early table checksum verification disabled Jan 30 15:23:28.898384 kernel: ACPI: RSDP 0x0000000132430018 000024 (v02 BOCHS ) Jan 30 15:23:28.898390 kernel: ACPI: XSDT 0x000000013243FE98 00006C (v01 BOCHS BXPC 00000001 01000013) Jan 30 15:23:28.898397 kernel: ACPI: FACP 0x000000013243FA98 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 15:23:28.898405 kernel: ACPI: DSDT 0x0000000132437518 001468 (v02 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 15:23:28.898412 kernel: ACPI: APIC 0x000000013243FC18 000108 (v04 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 15:23:28.898418 kernel: ACPI: PPTT 0x000000013243FD98 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 15:23:28.898425 kernel: ACPI: GTDT 0x000000013243D898 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 15:23:28.898433 kernel: ACPI: MCFG 0x000000013243FF98 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 15:23:28.898442 kernel: ACPI: SPCR 0x000000013243E818 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 15:23:28.898449 kernel: ACPI: DBG2 0x000000013243E898 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 15:23:28.898456 kernel: ACPI: IORT 0x000000013243E418 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Jan 30 15:23:28.898463 kernel: ACPI: BGRT 0x000000013243E798 000038 (v01 INTEL EDK2 00000002 01000013) Jan 30 15:23:28.898469 kernel: ACPI: SPCR: console: pl011,mmio32,0x9000000,9600 Jan 30 15:23:28.898476 kernel: NUMA: Failed to initialise from firmware Jan 30 15:23:28.898483 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x0000000139ffffff] Jan 30 15:23:28.898490 kernel: NUMA: NODE_DATA [mem 0x13981f800-0x139824fff] Jan 30 15:23:28.898496 kernel: Zone ranges: Jan 30 15:23:28.898506 kernel: DMA [mem 0x0000000040000000-0x00000000ffffffff] Jan 30 15:23:28.898513 kernel: DMA32 empty Jan 30 15:23:28.898521 kernel: Normal [mem 0x0000000100000000-0x0000000139ffffff] Jan 30 15:23:28.898528 kernel: Movable zone start for each node Jan 30 15:23:28.898534 kernel: Early memory node ranges Jan 30 15:23:28.898541 kernel: node 0: [mem 0x0000000040000000-0x000000013243ffff] Jan 30 15:23:28.898548 kernel: node 0: [mem 0x0000000132440000-0x000000013272ffff] Jan 30 15:23:28.898555 kernel: node 0: [mem 0x0000000132730000-0x0000000135bfffff] Jan 30 15:23:28.898561 kernel: node 0: [mem 0x0000000135c00000-0x0000000135fdffff] Jan 30 15:23:28.898568 kernel: node 0: [mem 0x0000000135fe0000-0x0000000139ffffff] Jan 30 15:23:28.898575 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x0000000139ffffff] Jan 30 15:23:28.898582 kernel: On node 0, zone Normal: 24576 pages in unavailable ranges Jan 30 15:23:28.898589 kernel: psci: probing for conduit method from ACPI. Jan 30 15:23:28.898597 kernel: psci: PSCIv1.1 detected in firmware. Jan 30 15:23:28.898603 kernel: psci: Using standard PSCI v0.2 function IDs Jan 30 15:23:28.898610 kernel: psci: Trusted OS migration not required Jan 30 15:23:28.898620 kernel: psci: SMC Calling Convention v1.1 Jan 30 15:23:28.898628 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Jan 30 15:23:28.898646 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Jan 30 15:23:28.898658 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Jan 30 15:23:28.898666 kernel: pcpu-alloc: [0] 0 [0] 1 Jan 30 15:23:28.898673 kernel: Detected PIPT I-cache on CPU0 Jan 30 15:23:28.898680 kernel: CPU features: detected: GIC system register CPU interface Jan 30 15:23:28.898688 kernel: CPU features: detected: Hardware dirty bit management Jan 30 15:23:28.898695 kernel: CPU features: detected: Spectre-v4 Jan 30 15:23:28.898702 kernel: CPU features: detected: Spectre-BHB Jan 30 15:23:28.898709 kernel: CPU features: kernel page table isolation forced ON by KASLR Jan 30 15:23:28.898716 kernel: CPU features: detected: Kernel page table isolation (KPTI) Jan 30 15:23:28.898724 kernel: CPU features: detected: ARM erratum 1418040 Jan 30 15:23:28.898731 kernel: CPU features: detected: SSBS not fully self-synchronizing Jan 30 15:23:28.898740 kernel: alternatives: applying boot alternatives Jan 30 15:23:28.898748 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=05d22c8845dec898f2b35f78b7d946edccf803dd23b974a9db2c3070ca1d8f8c Jan 30 15:23:28.898756 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jan 30 15:23:28.898763 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 30 15:23:28.898770 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 30 15:23:28.898778 kernel: Fallback order for Node 0: 0 Jan 30 15:23:28.898785 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1008000 Jan 30 15:23:28.898792 kernel: Policy zone: Normal Jan 30 15:23:28.898799 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 30 15:23:28.898806 kernel: software IO TLB: area num 2. Jan 30 15:23:28.898814 kernel: software IO TLB: mapped [mem 0x00000000fbfff000-0x00000000fffff000] (64MB) Jan 30 15:23:28.898823 kernel: Memory: 3881592K/4096000K available (10240K kernel code, 2186K rwdata, 8096K rodata, 39360K init, 897K bss, 214408K reserved, 0K cma-reserved) Jan 30 15:23:28.898830 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jan 30 15:23:28.898837 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 30 15:23:28.898845 kernel: rcu: RCU event tracing is enabled. Jan 30 15:23:28.898852 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jan 30 15:23:28.898860 kernel: Trampoline variant of Tasks RCU enabled. Jan 30 15:23:28.898867 kernel: Tracing variant of Tasks RCU enabled. Jan 30 15:23:28.898875 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 30 15:23:28.898882 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jan 30 15:23:28.898889 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Jan 30 15:23:28.898896 kernel: GICv3: 256 SPIs implemented Jan 30 15:23:28.898996 kernel: GICv3: 0 Extended SPIs implemented Jan 30 15:23:28.899005 kernel: Root IRQ handler: gic_handle_irq Jan 30 15:23:28.899012 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Jan 30 15:23:28.899020 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Jan 30 15:23:28.899027 kernel: ITS [mem 0x08080000-0x0809ffff] Jan 30 15:23:28.899035 kernel: ITS@0x0000000008080000: allocated 8192 Devices @1000c0000 (indirect, esz 8, psz 64K, shr 1) Jan 30 15:23:28.899042 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @1000d0000 (flat, esz 8, psz 64K, shr 1) Jan 30 15:23:28.899049 kernel: GICv3: using LPI property table @0x00000001000e0000 Jan 30 15:23:28.899057 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000001000f0000 Jan 30 15:23:28.899064 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 30 15:23:28.899072 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jan 30 15:23:28.899083 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Jan 30 15:23:28.899090 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Jan 30 15:23:28.899098 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Jan 30 15:23:28.899105 kernel: Console: colour dummy device 80x25 Jan 30 15:23:28.899113 kernel: ACPI: Core revision 20230628 Jan 30 15:23:28.899121 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Jan 30 15:23:28.899129 kernel: pid_max: default: 32768 minimum: 301 Jan 30 15:23:28.899136 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 30 15:23:28.899144 kernel: landlock: Up and running. Jan 30 15:23:28.899152 kernel: SELinux: Initializing. Jan 30 15:23:28.899161 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 30 15:23:28.899168 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 30 15:23:28.899176 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 30 15:23:28.899184 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 30 15:23:28.899191 kernel: rcu: Hierarchical SRCU implementation. Jan 30 15:23:28.899199 kernel: rcu: Max phase no-delay instances is 400. Jan 30 15:23:28.899206 kernel: Platform MSI: ITS@0x8080000 domain created Jan 30 15:23:28.899214 kernel: PCI/MSI: ITS@0x8080000 domain created Jan 30 15:23:28.899221 kernel: Remapping and enabling EFI services. Jan 30 15:23:28.899230 kernel: smp: Bringing up secondary CPUs ... Jan 30 15:23:28.899238 kernel: Detected PIPT I-cache on CPU1 Jan 30 15:23:28.899246 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Jan 30 15:23:28.899253 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000100100000 Jan 30 15:23:28.899261 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jan 30 15:23:28.899268 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Jan 30 15:23:28.899276 kernel: smp: Brought up 1 node, 2 CPUs Jan 30 15:23:28.899283 kernel: SMP: Total of 2 processors activated. Jan 30 15:23:28.899291 kernel: CPU features: detected: 32-bit EL0 Support Jan 30 15:23:28.899300 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Jan 30 15:23:28.899307 kernel: CPU features: detected: Common not Private translations Jan 30 15:23:28.899315 kernel: CPU features: detected: CRC32 instructions Jan 30 15:23:28.899328 kernel: CPU features: detected: Enhanced Virtualization Traps Jan 30 15:23:28.899338 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Jan 30 15:23:28.899346 kernel: CPU features: detected: LSE atomic instructions Jan 30 15:23:28.899354 kernel: CPU features: detected: Privileged Access Never Jan 30 15:23:28.899362 kernel: CPU features: detected: RAS Extension Support Jan 30 15:23:28.899373 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Jan 30 15:23:28.899382 kernel: CPU: All CPU(s) started at EL1 Jan 30 15:23:28.899391 kernel: alternatives: applying system-wide alternatives Jan 30 15:23:28.899399 kernel: devtmpfs: initialized Jan 30 15:23:28.899407 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 30 15:23:28.899415 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jan 30 15:23:28.899423 kernel: pinctrl core: initialized pinctrl subsystem Jan 30 15:23:28.899431 kernel: SMBIOS 3.0.0 present. Jan 30 15:23:28.899439 kernel: DMI: Hetzner vServer/KVM Virtual Machine, BIOS 20171111 11/11/2017 Jan 30 15:23:28.899449 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 30 15:23:28.899457 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Jan 30 15:23:28.899465 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Jan 30 15:23:28.899473 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Jan 30 15:23:28.899481 kernel: audit: initializing netlink subsys (disabled) Jan 30 15:23:28.899489 kernel: audit: type=2000 audit(0.013:1): state=initialized audit_enabled=0 res=1 Jan 30 15:23:28.899498 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 30 15:23:28.899506 kernel: cpuidle: using governor menu Jan 30 15:23:28.899514 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Jan 30 15:23:28.899523 kernel: ASID allocator initialised with 32768 entries Jan 30 15:23:28.899531 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 30 15:23:28.899539 kernel: Serial: AMBA PL011 UART driver Jan 30 15:23:28.899547 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Jan 30 15:23:28.899555 kernel: Modules: 0 pages in range for non-PLT usage Jan 30 15:23:28.899563 kernel: Modules: 509040 pages in range for PLT usage Jan 30 15:23:28.899571 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 30 15:23:28.899580 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Jan 30 15:23:28.899587 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Jan 30 15:23:28.899597 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Jan 30 15:23:28.899605 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 30 15:23:28.899613 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Jan 30 15:23:28.899621 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Jan 30 15:23:28.899629 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Jan 30 15:23:28.899671 kernel: ACPI: Added _OSI(Module Device) Jan 30 15:23:28.899680 kernel: ACPI: Added _OSI(Processor Device) Jan 30 15:23:28.899688 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jan 30 15:23:28.899695 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 30 15:23:28.899707 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 30 15:23:28.899715 kernel: ACPI: Interpreter enabled Jan 30 15:23:28.899723 kernel: ACPI: Using GIC for interrupt routing Jan 30 15:23:28.899731 kernel: ACPI: MCFG table detected, 1 entries Jan 30 15:23:28.899739 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Jan 30 15:23:28.899747 kernel: printk: console [ttyAMA0] enabled Jan 30 15:23:28.899754 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jan 30 15:23:28.899935 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jan 30 15:23:28.900028 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Jan 30 15:23:28.900098 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Jan 30 15:23:28.900176 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Jan 30 15:23:28.900244 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Jan 30 15:23:28.900255 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Jan 30 15:23:28.900263 kernel: PCI host bridge to bus 0000:00 Jan 30 15:23:28.900340 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Jan 30 15:23:28.900409 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Jan 30 15:23:28.900472 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Jan 30 15:23:28.900532 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jan 30 15:23:28.900617 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Jan 30 15:23:28.900728 kernel: pci 0000:00:01.0: [1af4:1050] type 00 class 0x038000 Jan 30 15:23:28.900805 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x11289000-0x11289fff] Jan 30 15:23:28.900877 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000600000-0x8000603fff 64bit pref] Jan 30 15:23:28.900984 kernel: pci 0000:00:02.0: [1b36:000c] type 01 class 0x060400 Jan 30 15:23:28.901057 kernel: pci 0000:00:02.0: reg 0x10: [mem 0x11288000-0x11288fff] Jan 30 15:23:28.901136 kernel: pci 0000:00:02.1: [1b36:000c] type 01 class 0x060400 Jan 30 15:23:28.901209 kernel: pci 0000:00:02.1: reg 0x10: [mem 0x11287000-0x11287fff] Jan 30 15:23:28.901287 kernel: pci 0000:00:02.2: [1b36:000c] type 01 class 0x060400 Jan 30 15:23:28.901358 kernel: pci 0000:00:02.2: reg 0x10: [mem 0x11286000-0x11286fff] Jan 30 15:23:28.901439 kernel: pci 0000:00:02.3: [1b36:000c] type 01 class 0x060400 Jan 30 15:23:28.901510 kernel: pci 0000:00:02.3: reg 0x10: [mem 0x11285000-0x11285fff] Jan 30 15:23:28.901587 kernel: pci 0000:00:02.4: [1b36:000c] type 01 class 0x060400 Jan 30 15:23:28.901701 kernel: pci 0000:00:02.4: reg 0x10: [mem 0x11284000-0x11284fff] Jan 30 15:23:28.901787 kernel: pci 0000:00:02.5: [1b36:000c] type 01 class 0x060400 Jan 30 15:23:28.901861 kernel: pci 0000:00:02.5: reg 0x10: [mem 0x11283000-0x11283fff] Jan 30 15:23:28.901976 kernel: pci 0000:00:02.6: [1b36:000c] type 01 class 0x060400 Jan 30 15:23:28.902055 kernel: pci 0000:00:02.6: reg 0x10: [mem 0x11282000-0x11282fff] Jan 30 15:23:28.902132 kernel: pci 0000:00:02.7: [1b36:000c] type 01 class 0x060400 Jan 30 15:23:28.902202 kernel: pci 0000:00:02.7: reg 0x10: [mem 0x11281000-0x11281fff] Jan 30 15:23:28.902286 kernel: pci 0000:00:03.0: [1b36:000c] type 01 class 0x060400 Jan 30 15:23:28.902356 kernel: pci 0000:00:03.0: reg 0x10: [mem 0x11280000-0x11280fff] Jan 30 15:23:28.902437 kernel: pci 0000:00:04.0: [1b36:0002] type 00 class 0x070002 Jan 30 15:23:28.902507 kernel: pci 0000:00:04.0: reg 0x10: [io 0x8200-0x8207] Jan 30 15:23:28.902587 kernel: pci 0000:01:00.0: [1af4:1041] type 00 class 0x020000 Jan 30 15:23:28.902682 kernel: pci 0000:01:00.0: reg 0x14: [mem 0x11000000-0x11000fff] Jan 30 15:23:28.902761 kernel: pci 0000:01:00.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Jan 30 15:23:28.902834 kernel: pci 0000:01:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Jan 30 15:23:28.902930 kernel: pci 0000:02:00.0: [1b36:000d] type 00 class 0x0c0330 Jan 30 15:23:28.903013 kernel: pci 0000:02:00.0: reg 0x10: [mem 0x10e00000-0x10e03fff 64bit] Jan 30 15:23:28.903096 kernel: pci 0000:03:00.0: [1af4:1043] type 00 class 0x078000 Jan 30 15:23:28.903169 kernel: pci 0000:03:00.0: reg 0x14: [mem 0x10c00000-0x10c00fff] Jan 30 15:23:28.903241 kernel: pci 0000:03:00.0: reg 0x20: [mem 0x8000100000-0x8000103fff 64bit pref] Jan 30 15:23:28.903322 kernel: pci 0000:04:00.0: [1af4:1045] type 00 class 0x00ff00 Jan 30 15:23:28.903395 kernel: pci 0000:04:00.0: reg 0x20: [mem 0x8000200000-0x8000203fff 64bit pref] Jan 30 15:23:28.903480 kernel: pci 0000:05:00.0: [1af4:1044] type 00 class 0x00ff00 Jan 30 15:23:28.903559 kernel: pci 0000:05:00.0: reg 0x14: [mem 0x10800000-0x10800fff] Jan 30 15:23:28.903633 kernel: pci 0000:05:00.0: reg 0x20: [mem 0x8000300000-0x8000303fff 64bit pref] Jan 30 15:23:28.903758 kernel: pci 0000:06:00.0: [1af4:1048] type 00 class 0x010000 Jan 30 15:23:28.903835 kernel: pci 0000:06:00.0: reg 0x14: [mem 0x10600000-0x10600fff] Jan 30 15:23:28.903919 kernel: pci 0000:06:00.0: reg 0x20: [mem 0x8000400000-0x8000403fff 64bit pref] Jan 30 15:23:28.904021 kernel: pci 0000:07:00.0: [1af4:1041] type 00 class 0x020000 Jan 30 15:23:28.904096 kernel: pci 0000:07:00.0: reg 0x14: [mem 0x10400000-0x10400fff] Jan 30 15:23:28.904169 kernel: pci 0000:07:00.0: reg 0x20: [mem 0x8000500000-0x8000503fff 64bit pref] Jan 30 15:23:28.904242 kernel: pci 0000:07:00.0: reg 0x30: [mem 0xfff80000-0xffffffff pref] Jan 30 15:23:28.904315 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x0fff] to [bus 01] add_size 1000 Jan 30 15:23:28.904386 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 01] add_size 100000 add_align 100000 Jan 30 15:23:28.904455 kernel: pci 0000:00:02.0: bridge window [mem 0x00100000-0x001fffff] to [bus 01] add_size 100000 add_align 100000 Jan 30 15:23:28.904532 kernel: pci 0000:00:02.1: bridge window [io 0x1000-0x0fff] to [bus 02] add_size 1000 Jan 30 15:23:28.904604 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 02] add_size 200000 add_align 100000 Jan 30 15:23:28.904693 kernel: pci 0000:00:02.1: bridge window [mem 0x00100000-0x001fffff] to [bus 02] add_size 100000 add_align 100000 Jan 30 15:23:28.904771 kernel: pci 0000:00:02.2: bridge window [io 0x1000-0x0fff] to [bus 03] add_size 1000 Jan 30 15:23:28.904843 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 03] add_size 100000 add_align 100000 Jan 30 15:23:28.904933 kernel: pci 0000:00:02.2: bridge window [mem 0x00100000-0x001fffff] to [bus 03] add_size 100000 add_align 100000 Jan 30 15:23:28.905015 kernel: pci 0000:00:02.3: bridge window [io 0x1000-0x0fff] to [bus 04] add_size 1000 Jan 30 15:23:28.905089 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 04] add_size 100000 add_align 100000 Jan 30 15:23:28.905166 kernel: pci 0000:00:02.3: bridge window [mem 0x00100000-0x000fffff] to [bus 04] add_size 200000 add_align 100000 Jan 30 15:23:28.905241 kernel: pci 0000:00:02.4: bridge window [io 0x1000-0x0fff] to [bus 05] add_size 1000 Jan 30 15:23:28.905311 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 05] add_size 100000 add_align 100000 Jan 30 15:23:28.905381 kernel: pci 0000:00:02.4: bridge window [mem 0x00100000-0x001fffff] to [bus 05] add_size 100000 add_align 100000 Jan 30 15:23:28.905454 kernel: pci 0000:00:02.5: bridge window [io 0x1000-0x0fff] to [bus 06] add_size 1000 Jan 30 15:23:28.905522 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 06] add_size 100000 add_align 100000 Jan 30 15:23:28.905591 kernel: pci 0000:00:02.5: bridge window [mem 0x00100000-0x001fffff] to [bus 06] add_size 100000 add_align 100000 Jan 30 15:23:28.905684 kernel: pci 0000:00:02.6: bridge window [io 0x1000-0x0fff] to [bus 07] add_size 1000 Jan 30 15:23:28.905760 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff 64bit pref] to [bus 07] add_size 100000 add_align 100000 Jan 30 15:23:28.905829 kernel: pci 0000:00:02.6: bridge window [mem 0x00100000-0x001fffff] to [bus 07] add_size 100000 add_align 100000 Jan 30 15:23:28.905903 kernel: pci 0000:00:02.7: bridge window [io 0x1000-0x0fff] to [bus 08] add_size 1000 Jan 30 15:23:28.905989 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 08] add_size 200000 add_align 100000 Jan 30 15:23:28.906059 kernel: pci 0000:00:02.7: bridge window [mem 0x00100000-0x000fffff] to [bus 08] add_size 200000 add_align 100000 Jan 30 15:23:28.906131 kernel: pci 0000:00:03.0: bridge window [io 0x1000-0x0fff] to [bus 09] add_size 1000 Jan 30 15:23:28.906201 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff 64bit pref] to [bus 09] add_size 200000 add_align 100000 Jan 30 15:23:28.906275 kernel: pci 0000:00:03.0: bridge window [mem 0x00100000-0x000fffff] to [bus 09] add_size 200000 add_align 100000 Jan 30 15:23:28.906345 kernel: pci 0000:00:02.0: BAR 14: assigned [mem 0x10000000-0x101fffff] Jan 30 15:23:28.906413 kernel: pci 0000:00:02.0: BAR 15: assigned [mem 0x8000000000-0x80001fffff 64bit pref] Jan 30 15:23:28.906483 kernel: pci 0000:00:02.1: BAR 14: assigned [mem 0x10200000-0x103fffff] Jan 30 15:23:28.906552 kernel: pci 0000:00:02.1: BAR 15: assigned [mem 0x8000200000-0x80003fffff 64bit pref] Jan 30 15:23:28.906623 kernel: pci 0000:00:02.2: BAR 14: assigned [mem 0x10400000-0x105fffff] Jan 30 15:23:28.906747 kernel: pci 0000:00:02.2: BAR 15: assigned [mem 0x8000400000-0x80005fffff 64bit pref] Jan 30 15:23:28.906824 kernel: pci 0000:00:02.3: BAR 14: assigned [mem 0x10600000-0x107fffff] Jan 30 15:23:28.906893 kernel: pci 0000:00:02.3: BAR 15: assigned [mem 0x8000600000-0x80007fffff 64bit pref] Jan 30 15:23:28.906982 kernel: pci 0000:00:02.4: BAR 14: assigned [mem 0x10800000-0x109fffff] Jan 30 15:23:28.907052 kernel: pci 0000:00:02.4: BAR 15: assigned [mem 0x8000800000-0x80009fffff 64bit pref] Jan 30 15:23:28.907122 kernel: pci 0000:00:02.5: BAR 14: assigned [mem 0x10a00000-0x10bfffff] Jan 30 15:23:28.907191 kernel: pci 0000:00:02.5: BAR 15: assigned [mem 0x8000a00000-0x8000bfffff 64bit pref] Jan 30 15:23:28.907269 kernel: pci 0000:00:02.6: BAR 14: assigned [mem 0x10c00000-0x10dfffff] Jan 30 15:23:28.907339 kernel: pci 0000:00:02.6: BAR 15: assigned [mem 0x8000c00000-0x8000dfffff 64bit pref] Jan 30 15:23:28.907409 kernel: pci 0000:00:02.7: BAR 14: assigned [mem 0x10e00000-0x10ffffff] Jan 30 15:23:28.907478 kernel: pci 0000:00:02.7: BAR 15: assigned [mem 0x8000e00000-0x8000ffffff 64bit pref] Jan 30 15:23:28.907549 kernel: pci 0000:00:03.0: BAR 14: assigned [mem 0x11000000-0x111fffff] Jan 30 15:23:28.907618 kernel: pci 0000:00:03.0: BAR 15: assigned [mem 0x8001000000-0x80011fffff 64bit pref] Jan 30 15:23:28.907711 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8001200000-0x8001203fff 64bit pref] Jan 30 15:23:28.907790 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x11200000-0x11200fff] Jan 30 15:23:28.907862 kernel: pci 0000:00:02.0: BAR 0: assigned [mem 0x11201000-0x11201fff] Jan 30 15:23:28.908334 kernel: pci 0000:00:02.0: BAR 13: assigned [io 0x1000-0x1fff] Jan 30 15:23:28.908421 kernel: pci 0000:00:02.1: BAR 0: assigned [mem 0x11202000-0x11202fff] Jan 30 15:23:28.908492 kernel: pci 0000:00:02.1: BAR 13: assigned [io 0x2000-0x2fff] Jan 30 15:23:28.908562 kernel: pci 0000:00:02.2: BAR 0: assigned [mem 0x11203000-0x11203fff] Jan 30 15:23:28.908682 kernel: pci 0000:00:02.2: BAR 13: assigned [io 0x3000-0x3fff] Jan 30 15:23:28.908764 kernel: pci 0000:00:02.3: BAR 0: assigned [mem 0x11204000-0x11204fff] Jan 30 15:23:28.908855 kernel: pci 0000:00:02.3: BAR 13: assigned [io 0x4000-0x4fff] Jan 30 15:23:28.908956 kernel: pci 0000:00:02.4: BAR 0: assigned [mem 0x11205000-0x11205fff] Jan 30 15:23:28.909029 kernel: pci 0000:00:02.4: BAR 13: assigned [io 0x5000-0x5fff] Jan 30 15:23:28.909100 kernel: pci 0000:00:02.5: BAR 0: assigned [mem 0x11206000-0x11206fff] Jan 30 15:23:28.909174 kernel: pci 0000:00:02.5: BAR 13: assigned [io 0x6000-0x6fff] Jan 30 15:23:28.909247 kernel: pci 0000:00:02.6: BAR 0: assigned [mem 0x11207000-0x11207fff] Jan 30 15:23:28.909316 kernel: pci 0000:00:02.6: BAR 13: assigned [io 0x7000-0x7fff] Jan 30 15:23:28.909387 kernel: pci 0000:00:02.7: BAR 0: assigned [mem 0x11208000-0x11208fff] Jan 30 15:23:28.909460 kernel: pci 0000:00:02.7: BAR 13: assigned [io 0x8000-0x8fff] Jan 30 15:23:28.909531 kernel: pci 0000:00:03.0: BAR 0: assigned [mem 0x11209000-0x11209fff] Jan 30 15:23:28.909599 kernel: pci 0000:00:03.0: BAR 13: assigned [io 0x9000-0x9fff] Jan 30 15:23:28.909689 kernel: pci 0000:00:04.0: BAR 0: assigned [io 0xa000-0xa007] Jan 30 15:23:28.909770 kernel: pci 0000:01:00.0: BAR 6: assigned [mem 0x10000000-0x1007ffff pref] Jan 30 15:23:28.909842 kernel: pci 0000:01:00.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Jan 30 15:23:28.910170 kernel: pci 0000:01:00.0: BAR 1: assigned [mem 0x10080000-0x10080fff] Jan 30 15:23:28.910267 kernel: pci 0000:00:02.0: PCI bridge to [bus 01] Jan 30 15:23:28.910345 kernel: pci 0000:00:02.0: bridge window [io 0x1000-0x1fff] Jan 30 15:23:28.910413 kernel: pci 0000:00:02.0: bridge window [mem 0x10000000-0x101fffff] Jan 30 15:23:28.910480 kernel: pci 0000:00:02.0: bridge window [mem 0x8000000000-0x80001fffff 64bit pref] Jan 30 15:23:28.910555 kernel: pci 0000:02:00.0: BAR 0: assigned [mem 0x10200000-0x10203fff 64bit] Jan 30 15:23:28.911001 kernel: pci 0000:00:02.1: PCI bridge to [bus 02] Jan 30 15:23:28.911087 kernel: pci 0000:00:02.1: bridge window [io 0x2000-0x2fff] Jan 30 15:23:28.911155 kernel: pci 0000:00:02.1: bridge window [mem 0x10200000-0x103fffff] Jan 30 15:23:28.911223 kernel: pci 0000:00:02.1: bridge window [mem 0x8000200000-0x80003fffff 64bit pref] Jan 30 15:23:28.911303 kernel: pci 0000:03:00.0: BAR 4: assigned [mem 0x8000400000-0x8000403fff 64bit pref] Jan 30 15:23:28.911376 kernel: pci 0000:03:00.0: BAR 1: assigned [mem 0x10400000-0x10400fff] Jan 30 15:23:28.911447 kernel: pci 0000:00:02.2: PCI bridge to [bus 03] Jan 30 15:23:28.911516 kernel: pci 0000:00:02.2: bridge window [io 0x3000-0x3fff] Jan 30 15:23:28.911591 kernel: pci 0000:00:02.2: bridge window [mem 0x10400000-0x105fffff] Jan 30 15:23:28.911687 kernel: pci 0000:00:02.2: bridge window [mem 0x8000400000-0x80005fffff 64bit pref] Jan 30 15:23:28.911770 kernel: pci 0000:04:00.0: BAR 4: assigned [mem 0x8000600000-0x8000603fff 64bit pref] Jan 30 15:23:28.911846 kernel: pci 0000:00:02.3: PCI bridge to [bus 04] Jan 30 15:23:28.911934 kernel: pci 0000:00:02.3: bridge window [io 0x4000-0x4fff] Jan 30 15:23:28.912007 kernel: pci 0000:00:02.3: bridge window [mem 0x10600000-0x107fffff] Jan 30 15:23:28.912076 kernel: pci 0000:00:02.3: bridge window [mem 0x8000600000-0x80007fffff 64bit pref] Jan 30 15:23:28.912155 kernel: pci 0000:05:00.0: BAR 4: assigned [mem 0x8000800000-0x8000803fff 64bit pref] Jan 30 15:23:28.912235 kernel: pci 0000:05:00.0: BAR 1: assigned [mem 0x10800000-0x10800fff] Jan 30 15:23:28.912305 kernel: pci 0000:00:02.4: PCI bridge to [bus 05] Jan 30 15:23:28.912376 kernel: pci 0000:00:02.4: bridge window [io 0x5000-0x5fff] Jan 30 15:23:28.912447 kernel: pci 0000:00:02.4: bridge window [mem 0x10800000-0x109fffff] Jan 30 15:23:28.912519 kernel: pci 0000:00:02.4: bridge window [mem 0x8000800000-0x80009fffff 64bit pref] Jan 30 15:23:28.912597 kernel: pci 0000:06:00.0: BAR 4: assigned [mem 0x8000a00000-0x8000a03fff 64bit pref] Jan 30 15:23:28.912686 kernel: pci 0000:06:00.0: BAR 1: assigned [mem 0x10a00000-0x10a00fff] Jan 30 15:23:28.912762 kernel: pci 0000:00:02.5: PCI bridge to [bus 06] Jan 30 15:23:28.912837 kernel: pci 0000:00:02.5: bridge window [io 0x6000-0x6fff] Jan 30 15:23:28.912941 kernel: pci 0000:00:02.5: bridge window [mem 0x10a00000-0x10bfffff] Jan 30 15:23:28.913019 kernel: pci 0000:00:02.5: bridge window [mem 0x8000a00000-0x8000bfffff 64bit pref] Jan 30 15:23:28.913100 kernel: pci 0000:07:00.0: BAR 6: assigned [mem 0x10c00000-0x10c7ffff pref] Jan 30 15:23:28.914963 kernel: pci 0000:07:00.0: BAR 4: assigned [mem 0x8000c00000-0x8000c03fff 64bit pref] Jan 30 15:23:28.915118 kernel: pci 0000:07:00.0: BAR 1: assigned [mem 0x10c80000-0x10c80fff] Jan 30 15:23:28.915194 kernel: pci 0000:00:02.6: PCI bridge to [bus 07] Jan 30 15:23:28.915265 kernel: pci 0000:00:02.6: bridge window [io 0x7000-0x7fff] Jan 30 15:23:28.915341 kernel: pci 0000:00:02.6: bridge window [mem 0x10c00000-0x10dfffff] Jan 30 15:23:28.915408 kernel: pci 0000:00:02.6: bridge window [mem 0x8000c00000-0x8000dfffff 64bit pref] Jan 30 15:23:28.915482 kernel: pci 0000:00:02.7: PCI bridge to [bus 08] Jan 30 15:23:28.915555 kernel: pci 0000:00:02.7: bridge window [io 0x8000-0x8fff] Jan 30 15:23:28.915626 kernel: pci 0000:00:02.7: bridge window [mem 0x10e00000-0x10ffffff] Jan 30 15:23:28.915769 kernel: pci 0000:00:02.7: bridge window [mem 0x8000e00000-0x8000ffffff 64bit pref] Jan 30 15:23:28.915851 kernel: pci 0000:00:03.0: PCI bridge to [bus 09] Jan 30 15:23:28.915958 kernel: pci 0000:00:03.0: bridge window [io 0x9000-0x9fff] Jan 30 15:23:28.916037 kernel: pci 0000:00:03.0: bridge window [mem 0x11000000-0x111fffff] Jan 30 15:23:28.916105 kernel: pci 0000:00:03.0: bridge window [mem 0x8001000000-0x80011fffff 64bit pref] Jan 30 15:23:28.916178 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Jan 30 15:23:28.916239 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Jan 30 15:23:28.916300 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Jan 30 15:23:28.916379 kernel: pci_bus 0000:01: resource 0 [io 0x1000-0x1fff] Jan 30 15:23:28.916446 kernel: pci_bus 0000:01: resource 1 [mem 0x10000000-0x101fffff] Jan 30 15:23:28.916515 kernel: pci_bus 0000:01: resource 2 [mem 0x8000000000-0x80001fffff 64bit pref] Jan 30 15:23:28.916595 kernel: pci_bus 0000:02: resource 0 [io 0x2000-0x2fff] Jan 30 15:23:28.916685 kernel: pci_bus 0000:02: resource 1 [mem 0x10200000-0x103fffff] Jan 30 15:23:28.916754 kernel: pci_bus 0000:02: resource 2 [mem 0x8000200000-0x80003fffff 64bit pref] Jan 30 15:23:28.916838 kernel: pci_bus 0000:03: resource 0 [io 0x3000-0x3fff] Jan 30 15:23:28.920934 kernel: pci_bus 0000:03: resource 1 [mem 0x10400000-0x105fffff] Jan 30 15:23:28.921110 kernel: pci_bus 0000:03: resource 2 [mem 0x8000400000-0x80005fffff 64bit pref] Jan 30 15:23:28.921189 kernel: pci_bus 0000:04: resource 0 [io 0x4000-0x4fff] Jan 30 15:23:28.921274 kernel: pci_bus 0000:04: resource 1 [mem 0x10600000-0x107fffff] Jan 30 15:23:28.921368 kernel: pci_bus 0000:04: resource 2 [mem 0x8000600000-0x80007fffff 64bit pref] Jan 30 15:23:28.921442 kernel: pci_bus 0000:05: resource 0 [io 0x5000-0x5fff] Jan 30 15:23:28.921508 kernel: pci_bus 0000:05: resource 1 [mem 0x10800000-0x109fffff] Jan 30 15:23:28.921571 kernel: pci_bus 0000:05: resource 2 [mem 0x8000800000-0x80009fffff 64bit pref] Jan 30 15:23:28.921691 kernel: pci_bus 0000:06: resource 0 [io 0x6000-0x6fff] Jan 30 15:23:28.921766 kernel: pci_bus 0000:06: resource 1 [mem 0x10a00000-0x10bfffff] Jan 30 15:23:28.921831 kernel: pci_bus 0000:06: resource 2 [mem 0x8000a00000-0x8000bfffff 64bit pref] Jan 30 15:23:28.921938 kernel: pci_bus 0000:07: resource 0 [io 0x7000-0x7fff] Jan 30 15:23:28.922020 kernel: pci_bus 0000:07: resource 1 [mem 0x10c00000-0x10dfffff] Jan 30 15:23:28.922086 kernel: pci_bus 0000:07: resource 2 [mem 0x8000c00000-0x8000dfffff 64bit pref] Jan 30 15:23:28.923121 kernel: pci_bus 0000:08: resource 0 [io 0x8000-0x8fff] Jan 30 15:23:28.923234 kernel: pci_bus 0000:08: resource 1 [mem 0x10e00000-0x10ffffff] Jan 30 15:23:28.923302 kernel: pci_bus 0000:08: resource 2 [mem 0x8000e00000-0x8000ffffff 64bit pref] Jan 30 15:23:28.923375 kernel: pci_bus 0000:09: resource 0 [io 0x9000-0x9fff] Jan 30 15:23:28.923440 kernel: pci_bus 0000:09: resource 1 [mem 0x11000000-0x111fffff] Jan 30 15:23:28.923513 kernel: pci_bus 0000:09: resource 2 [mem 0x8001000000-0x80011fffff 64bit pref] Jan 30 15:23:28.923525 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Jan 30 15:23:28.923534 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Jan 30 15:23:28.923543 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Jan 30 15:23:28.923551 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Jan 30 15:23:28.923559 kernel: iommu: Default domain type: Translated Jan 30 15:23:28.923568 kernel: iommu: DMA domain TLB invalidation policy: strict mode Jan 30 15:23:28.923576 kernel: efivars: Registered efivars operations Jan 30 15:23:28.923585 kernel: vgaarb: loaded Jan 30 15:23:28.923596 kernel: clocksource: Switched to clocksource arch_sys_counter Jan 30 15:23:28.923604 kernel: VFS: Disk quotas dquot_6.6.0 Jan 30 15:23:28.923613 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 30 15:23:28.923621 kernel: pnp: PnP ACPI init Jan 30 15:23:28.923720 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Jan 30 15:23:28.923734 kernel: pnp: PnP ACPI: found 1 devices Jan 30 15:23:28.923743 kernel: NET: Registered PF_INET protocol family Jan 30 15:23:28.923752 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 30 15:23:28.923764 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 30 15:23:28.923772 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 30 15:23:28.923780 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 30 15:23:28.923789 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 30 15:23:28.923797 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 30 15:23:28.923806 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 30 15:23:28.923814 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 30 15:23:28.923823 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 30 15:23:28.923922 kernel: pci 0000:02:00.0: enabling device (0000 -> 0002) Jan 30 15:23:28.923943 kernel: PCI: CLS 0 bytes, default 64 Jan 30 15:23:28.923953 kernel: kvm [1]: HYP mode not available Jan 30 15:23:28.923961 kernel: Initialise system trusted keyrings Jan 30 15:23:28.923970 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 30 15:23:28.923978 kernel: Key type asymmetric registered Jan 30 15:23:28.923986 kernel: Asymmetric key parser 'x509' registered Jan 30 15:23:28.923994 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jan 30 15:23:28.924003 kernel: io scheduler mq-deadline registered Jan 30 15:23:28.924011 kernel: io scheduler kyber registered Jan 30 15:23:28.924022 kernel: io scheduler bfq registered Jan 30 15:23:28.924031 kernel: ACPI: \_SB_.PCI0.GSI2: Enabled at IRQ 37 Jan 30 15:23:28.924115 kernel: pcieport 0000:00:02.0: PME: Signaling with IRQ 50 Jan 30 15:23:28.924187 kernel: pcieport 0000:00:02.0: AER: enabled with IRQ 50 Jan 30 15:23:28.924257 kernel: pcieport 0000:00:02.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Jan 30 15:23:28.924332 kernel: pcieport 0000:00:02.1: PME: Signaling with IRQ 51 Jan 30 15:23:28.924402 kernel: pcieport 0000:00:02.1: AER: enabled with IRQ 51 Jan 30 15:23:28.924476 kernel: pcieport 0000:00:02.1: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Jan 30 15:23:28.924549 kernel: pcieport 0000:00:02.2: PME: Signaling with IRQ 52 Jan 30 15:23:28.924618 kernel: pcieport 0000:00:02.2: AER: enabled with IRQ 52 Jan 30 15:23:28.924739 kernel: pcieport 0000:00:02.2: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Jan 30 15:23:28.924821 kernel: pcieport 0000:00:02.3: PME: Signaling with IRQ 53 Jan 30 15:23:28.924894 kernel: pcieport 0000:00:02.3: AER: enabled with IRQ 53 Jan 30 15:23:28.928113 kernel: pcieport 0000:00:02.3: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Jan 30 15:23:28.928204 kernel: pcieport 0000:00:02.4: PME: Signaling with IRQ 54 Jan 30 15:23:28.928277 kernel: pcieport 0000:00:02.4: AER: enabled with IRQ 54 Jan 30 15:23:28.928348 kernel: pcieport 0000:00:02.4: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Jan 30 15:23:28.928424 kernel: pcieport 0000:00:02.5: PME: Signaling with IRQ 55 Jan 30 15:23:28.928497 kernel: pcieport 0000:00:02.5: AER: enabled with IRQ 55 Jan 30 15:23:28.928576 kernel: pcieport 0000:00:02.5: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Jan 30 15:23:28.928698 kernel: pcieport 0000:00:02.6: PME: Signaling with IRQ 56 Jan 30 15:23:28.928784 kernel: pcieport 0000:00:02.6: AER: enabled with IRQ 56 Jan 30 15:23:28.928855 kernel: pcieport 0000:00:02.6: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Jan 30 15:23:28.928946 kernel: pcieport 0000:00:02.7: PME: Signaling with IRQ 57 Jan 30 15:23:28.929031 kernel: pcieport 0000:00:02.7: AER: enabled with IRQ 57 Jan 30 15:23:28.929113 kernel: pcieport 0000:00:02.7: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Jan 30 15:23:28.929125 kernel: ACPI: \_SB_.PCI0.GSI3: Enabled at IRQ 38 Jan 30 15:23:28.929198 kernel: pcieport 0000:00:03.0: PME: Signaling with IRQ 58 Jan 30 15:23:28.929268 kernel: pcieport 0000:00:03.0: AER: enabled with IRQ 58 Jan 30 15:23:28.929340 kernel: pcieport 0000:00:03.0: pciehp: Slot #0 AttnBtn+ PwrCtrl+ MRL- AttnInd+ PwrInd+ HotPlug+ Surprise+ Interlock+ NoCompl- IbPresDis- LLActRep+ Jan 30 15:23:28.929352 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Jan 30 15:23:28.929364 kernel: ACPI: button: Power Button [PWRB] Jan 30 15:23:28.929375 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Jan 30 15:23:28.929452 kernel: virtio-pci 0000:03:00.0: enabling device (0000 -> 0002) Jan 30 15:23:28.929533 kernel: virtio-pci 0000:04:00.0: enabling device (0000 -> 0002) Jan 30 15:23:28.929613 kernel: virtio-pci 0000:07:00.0: enabling device (0000 -> 0002) Jan 30 15:23:28.929626 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 30 15:23:28.929645 kernel: ACPI: \_SB_.PCI0.GSI0: Enabled at IRQ 35 Jan 30 15:23:28.929730 kernel: serial 0000:00:04.0: enabling device (0000 -> 0001) Jan 30 15:23:28.929746 kernel: 0000:00:04.0: ttyS0 at I/O 0xa000 (irq = 45, base_baud = 115200) is a 16550A Jan 30 15:23:28.929754 kernel: thunder_xcv, ver 1.0 Jan 30 15:23:28.929762 kernel: thunder_bgx, ver 1.0 Jan 30 15:23:28.929771 kernel: nicpf, ver 1.0 Jan 30 15:23:28.929779 kernel: nicvf, ver 1.0 Jan 30 15:23:28.929868 kernel: rtc-efi rtc-efi.0: registered as rtc0 Jan 30 15:23:28.930451 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-01-30T15:23:28 UTC (1738250608) Jan 30 15:23:28.930473 kernel: hid: raw HID events driver (C) Jiri Kosina Jan 30 15:23:28.930487 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Jan 30 15:23:28.930496 kernel: watchdog: Delayed init of the lockup detector failed: -19 Jan 30 15:23:28.930504 kernel: watchdog: Hard watchdog permanently disabled Jan 30 15:23:28.930514 kernel: NET: Registered PF_INET6 protocol family Jan 30 15:23:28.930522 kernel: Segment Routing with IPv6 Jan 30 15:23:28.930531 kernel: In-situ OAM (IOAM) with IPv6 Jan 30 15:23:28.930539 kernel: NET: Registered PF_PACKET protocol family Jan 30 15:23:28.930547 kernel: Key type dns_resolver registered Jan 30 15:23:28.930556 kernel: registered taskstats version 1 Jan 30 15:23:28.930564 kernel: Loading compiled-in X.509 certificates Jan 30 15:23:28.930574 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: f200c60883a4a38d496d9250faf693faee9d7415' Jan 30 15:23:28.930583 kernel: Key type .fscrypt registered Jan 30 15:23:28.930592 kernel: Key type fscrypt-provisioning registered Jan 30 15:23:28.930600 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 30 15:23:28.930608 kernel: ima: Allocated hash algorithm: sha1 Jan 30 15:23:28.930617 kernel: ima: No architecture policies found Jan 30 15:23:28.930625 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Jan 30 15:23:28.930633 kernel: clk: Disabling unused clocks Jan 30 15:23:28.930692 kernel: Freeing unused kernel memory: 39360K Jan 30 15:23:28.930701 kernel: Run /init as init process Jan 30 15:23:28.930709 kernel: with arguments: Jan 30 15:23:28.930718 kernel: /init Jan 30 15:23:28.930726 kernel: with environment: Jan 30 15:23:28.930734 kernel: HOME=/ Jan 30 15:23:28.930742 kernel: TERM=linux Jan 30 15:23:28.930750 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jan 30 15:23:28.930761 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 30 15:23:28.930774 systemd[1]: Detected virtualization kvm. Jan 30 15:23:28.930783 systemd[1]: Detected architecture arm64. Jan 30 15:23:28.930791 systemd[1]: Running in initrd. Jan 30 15:23:28.930800 systemd[1]: No hostname configured, using default hostname. Jan 30 15:23:28.930808 systemd[1]: Hostname set to . Jan 30 15:23:28.930817 systemd[1]: Initializing machine ID from VM UUID. Jan 30 15:23:28.930826 systemd[1]: Queued start job for default target initrd.target. Jan 30 15:23:28.930837 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 15:23:28.930846 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 15:23:28.930856 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 30 15:23:28.930865 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 30 15:23:28.930875 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 30 15:23:28.930884 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 30 15:23:28.930894 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 30 15:23:28.931011 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 30 15:23:28.931024 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 15:23:28.931033 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 30 15:23:28.931046 systemd[1]: Reached target paths.target - Path Units. Jan 30 15:23:28.931055 systemd[1]: Reached target slices.target - Slice Units. Jan 30 15:23:28.931064 systemd[1]: Reached target swap.target - Swaps. Jan 30 15:23:28.931073 systemd[1]: Reached target timers.target - Timer Units. Jan 30 15:23:28.931083 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 30 15:23:28.931093 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 30 15:23:28.931103 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 30 15:23:28.931112 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 30 15:23:28.931121 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 30 15:23:28.931130 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 30 15:23:28.931139 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 15:23:28.931148 systemd[1]: Reached target sockets.target - Socket Units. Jan 30 15:23:28.931157 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 30 15:23:28.931166 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 30 15:23:28.931177 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 30 15:23:28.931186 systemd[1]: Starting systemd-fsck-usr.service... Jan 30 15:23:28.931195 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 30 15:23:28.931204 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 30 15:23:28.931213 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 15:23:28.931250 systemd-journald[237]: Collecting audit messages is disabled. Jan 30 15:23:28.931274 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 30 15:23:28.931284 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 15:23:28.931292 systemd[1]: Finished systemd-fsck-usr.service. Jan 30 15:23:28.931302 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 30 15:23:28.931314 systemd-journald[237]: Journal started Jan 30 15:23:28.931334 systemd-journald[237]: Runtime Journal (/run/log/journal/9be7bbb2090d46e5b9a67a98f7c71fe3) is 8.0M, max 76.5M, 68.5M free. Jan 30 15:23:28.919770 systemd-modules-load[238]: Inserted module 'overlay' Jan 30 15:23:28.938712 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 30 15:23:28.938744 systemd[1]: Started systemd-journald.service - Journal Service. Jan 30 15:23:28.937288 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 15:23:28.943942 kernel: Bridge firewalling registered Jan 30 15:23:28.942002 systemd-modules-load[238]: Inserted module 'br_netfilter' Jan 30 15:23:28.947195 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 15:23:28.950451 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 30 15:23:28.951577 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 30 15:23:28.952400 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 15:23:28.961477 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 30 15:23:28.966148 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 30 15:23:28.980022 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 15:23:28.984046 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 30 15:23:28.985688 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 15:23:28.986722 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 15:23:28.991202 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 30 15:23:28.995117 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 30 15:23:29.009516 dracut-cmdline[273]: dracut-dracut-053 Jan 30 15:23:29.013088 dracut-cmdline[273]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyAMA0,115200n8 flatcar.first_boot=detected acpi=force flatcar.oem.id=hetzner verity.usrhash=05d22c8845dec898f2b35f78b7d946edccf803dd23b974a9db2c3070ca1d8f8c Jan 30 15:23:29.032884 systemd-resolved[274]: Positive Trust Anchors: Jan 30 15:23:29.032917 systemd-resolved[274]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 30 15:23:29.034515 systemd-resolved[274]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 30 15:23:29.044228 systemd-resolved[274]: Defaulting to hostname 'linux'. Jan 30 15:23:29.045402 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 30 15:23:29.047205 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 30 15:23:29.099966 kernel: SCSI subsystem initialized Jan 30 15:23:29.104977 kernel: Loading iSCSI transport class v2.0-870. Jan 30 15:23:29.112944 kernel: iscsi: registered transport (tcp) Jan 30 15:23:29.126183 kernel: iscsi: registered transport (qla4xxx) Jan 30 15:23:29.126291 kernel: QLogic iSCSI HBA Driver Jan 30 15:23:29.174414 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 30 15:23:29.180184 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 30 15:23:29.201192 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 30 15:23:29.201290 kernel: device-mapper: uevent: version 1.0.3 Jan 30 15:23:29.201963 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 30 15:23:29.252971 kernel: raid6: neonx8 gen() 15654 MB/s Jan 30 15:23:29.269962 kernel: raid6: neonx4 gen() 15575 MB/s Jan 30 15:23:29.286956 kernel: raid6: neonx2 gen() 13202 MB/s Jan 30 15:23:29.303981 kernel: raid6: neonx1 gen() 10439 MB/s Jan 30 15:23:29.320965 kernel: raid6: int64x8 gen() 6921 MB/s Jan 30 15:23:29.337982 kernel: raid6: int64x4 gen() 7312 MB/s Jan 30 15:23:29.354971 kernel: raid6: int64x2 gen() 6090 MB/s Jan 30 15:23:29.372019 kernel: raid6: int64x1 gen() 5027 MB/s Jan 30 15:23:29.372132 kernel: raid6: using algorithm neonx8 gen() 15654 MB/s Jan 30 15:23:29.388990 kernel: raid6: .... xor() 11822 MB/s, rmw enabled Jan 30 15:23:29.389085 kernel: raid6: using neon recovery algorithm Jan 30 15:23:29.394063 kernel: xor: measuring software checksum speed Jan 30 15:23:29.394148 kernel: 8regs : 16394 MB/sec Jan 30 15:23:29.394981 kernel: 32regs : 19650 MB/sec Jan 30 15:23:29.395010 kernel: arm64_neon : 23137 MB/sec Jan 30 15:23:29.395034 kernel: xor: using function: arm64_neon (23137 MB/sec) Jan 30 15:23:29.446993 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 30 15:23:29.462849 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 30 15:23:29.470297 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 15:23:29.491151 systemd-udevd[456]: Using default interface naming scheme 'v255'. Jan 30 15:23:29.496064 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 15:23:29.507335 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 30 15:23:29.527965 dracut-pre-trigger[458]: rd.md=0: removing MD RAID activation Jan 30 15:23:29.577100 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 30 15:23:29.584122 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 30 15:23:29.649056 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 15:23:29.658562 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 30 15:23:29.699427 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 30 15:23:29.702207 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 30 15:23:29.704500 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 15:23:29.706102 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 30 15:23:29.713154 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 30 15:23:29.730702 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 30 15:23:29.794310 kernel: scsi host0: Virtio SCSI HBA Jan 30 15:23:29.796019 kernel: scsi 0:0:0:0: CD-ROM QEMU QEMU CD-ROM 2.5+ PQ: 0 ANSI: 5 Jan 30 15:23:29.797932 kernel: scsi 0:0:0:1: Direct-Access QEMU QEMU HARDDISK 2.5+ PQ: 0 ANSI: 5 Jan 30 15:23:29.804806 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 30 15:23:29.805877 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 15:23:29.808327 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 15:23:29.809422 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 15:23:29.814383 kernel: ACPI: bus type USB registered Jan 30 15:23:29.814412 kernel: usbcore: registered new interface driver usbfs Jan 30 15:23:29.814480 kernel: usbcore: registered new interface driver hub Jan 30 15:23:29.814493 kernel: usbcore: registered new device driver usb Jan 30 15:23:29.809513 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 15:23:29.812601 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 15:23:29.823135 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 15:23:29.855181 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 15:23:29.862127 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 15:23:29.866265 kernel: sr 0:0:0:0: Power-on or device reset occurred Jan 30 15:23:29.868803 kernel: sr 0:0:0:0: [sr0] scsi3-mmc drive: 16x/50x cd/rw xa/form2 cdda tray Jan 30 15:23:29.869186 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 30 15:23:29.869201 kernel: sr 0:0:0:0: Attached scsi CD-ROM sr0 Jan 30 15:23:29.877449 kernel: sd 0:0:0:1: Power-on or device reset occurred Jan 30 15:23:29.895136 kernel: sd 0:0:0:1: [sda] 80003072 512-byte logical blocks: (41.0 GB/38.1 GiB) Jan 30 15:23:29.895269 kernel: sd 0:0:0:1: [sda] Write Protect is off Jan 30 15:23:29.895356 kernel: sd 0:0:0:1: [sda] Mode Sense: 63 00 00 08 Jan 30 15:23:29.895449 kernel: sd 0:0:0:1: [sda] Write cache: enabled, read cache: enabled, doesn't support DPO or FUA Jan 30 15:23:29.895566 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Jan 30 15:23:29.897715 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 1 Jan 30 15:23:29.897853 kernel: xhci_hcd 0000:02:00.0: hcc params 0x00087001 hci version 0x100 quirks 0x0000000000000010 Jan 30 15:23:29.899754 kernel: xhci_hcd 0000:02:00.0: xHCI Host Controller Jan 30 15:23:29.899985 kernel: xhci_hcd 0000:02:00.0: new USB bus registered, assigned bus number 2 Jan 30 15:23:29.900091 kernel: xhci_hcd 0000:02:00.0: Host supports USB 3.0 SuperSpeed Jan 30 15:23:29.900224 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jan 30 15:23:29.900238 kernel: GPT:17805311 != 80003071 Jan 30 15:23:29.900257 kernel: GPT:Alternate GPT header not at the end of the disk. Jan 30 15:23:29.900268 kernel: GPT:17805311 != 80003071 Jan 30 15:23:29.900277 kernel: GPT: Use GNU Parted to correct GPT errors. Jan 30 15:23:29.900287 kernel: hub 1-0:1.0: USB hub found Jan 30 15:23:29.900467 kernel: hub 1-0:1.0: 4 ports detected Jan 30 15:23:29.900582 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 15:23:29.900594 kernel: usb usb2: We don't know the algorithms for LPM for this host, disabling LPM. Jan 30 15:23:29.900728 kernel: hub 2-0:1.0: USB hub found Jan 30 15:23:29.900836 kernel: sd 0:0:0:1: [sda] Attached SCSI disk Jan 30 15:23:29.901040 kernel: hub 2-0:1.0: 4 ports detected Jan 30 15:23:29.896209 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 15:23:29.949939 kernel: BTRFS: device fsid f02ec3fd-6702-4c1a-b68e-9001713a3a08 devid 1 transid 38 /dev/sda3 scanned by (udev-worker) (503) Jan 30 15:23:29.951936 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by (udev-worker) (509) Jan 30 15:23:29.957079 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - QEMU_HARDDISK EFI-SYSTEM. Jan 30 15:23:29.969863 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - QEMU_HARDDISK ROOT. Jan 30 15:23:29.975833 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Jan 30 15:23:29.985092 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - QEMU_HARDDISK USR-A. Jan 30 15:23:29.985778 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - QEMU_HARDDISK USR-A. Jan 30 15:23:29.996148 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 30 15:23:30.004110 disk-uuid[575]: Primary Header is updated. Jan 30 15:23:30.004110 disk-uuid[575]: Secondary Entries is updated. Jan 30 15:23:30.004110 disk-uuid[575]: Secondary Header is updated. Jan 30 15:23:30.015942 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 15:23:30.020958 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 15:23:30.137017 kernel: usb 1-1: new high-speed USB device number 2 using xhci_hcd Jan 30 15:23:30.380944 kernel: usb 1-2: new high-speed USB device number 3 using xhci_hcd Jan 30 15:23:30.515937 kernel: input: QEMU QEMU USB Tablet as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-1/1-1:1.0/0003:0627:0001.0001/input/input1 Jan 30 15:23:30.516996 kernel: hid-generic 0003:0627:0001.0001: input,hidraw0: USB HID v0.01 Mouse [QEMU QEMU USB Tablet] on usb-0000:02:00.0-1/input0 Jan 30 15:23:30.518940 kernel: input: QEMU QEMU USB Keyboard as /devices/pci0000:00/0000:00:02.1/0000:02:00.0/usb1/1-2/1-2:1.0/0003:0627:0001.0002/input/input2 Jan 30 15:23:30.573820 kernel: hid-generic 0003:0627:0001.0002: input,hidraw1: USB HID v1.11 Keyboard [QEMU QEMU USB Keyboard] on usb-0000:02:00.0-2/input0 Jan 30 15:23:30.574475 kernel: usbcore: registered new interface driver usbhid Jan 30 15:23:30.574507 kernel: usbhid: USB HID core driver Jan 30 15:23:31.026940 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 15:23:31.029038 disk-uuid[576]: The operation has completed successfully. Jan 30 15:23:31.077965 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 30 15:23:31.078077 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 30 15:23:31.093205 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 30 15:23:31.097108 sh[590]: Success Jan 30 15:23:31.111967 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Jan 30 15:23:31.186784 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 30 15:23:31.189968 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 30 15:23:31.194798 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 30 15:23:31.226085 kernel: BTRFS info (device dm-0): first mount of filesystem f02ec3fd-6702-4c1a-b68e-9001713a3a08 Jan 30 15:23:31.226194 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Jan 30 15:23:31.226223 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 30 15:23:31.227170 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 30 15:23:31.227241 kernel: BTRFS info (device dm-0): using free space tree Jan 30 15:23:31.237015 kernel: BTRFS info (device dm-0): enabling ssd optimizations Jan 30 15:23:31.239962 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 30 15:23:31.240722 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 30 15:23:31.248253 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 30 15:23:31.253032 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 30 15:23:31.275159 kernel: BTRFS info (device sda6): first mount of filesystem db40e17a-cddf-4890-8d80-4d8cda0a956a Jan 30 15:23:31.275229 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Jan 30 15:23:31.275242 kernel: BTRFS info (device sda6): using free space tree Jan 30 15:23:31.281274 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 30 15:23:31.281396 kernel: BTRFS info (device sda6): auto enabling async discard Jan 30 15:23:31.298978 kernel: BTRFS info (device sda6): last unmount of filesystem db40e17a-cddf-4890-8d80-4d8cda0a956a Jan 30 15:23:31.299335 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 30 15:23:31.306935 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 30 15:23:31.311142 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 30 15:23:31.400487 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 30 15:23:31.411209 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 30 15:23:31.418041 ignition[686]: Ignition 2.19.0 Jan 30 15:23:31.418053 ignition[686]: Stage: fetch-offline Jan 30 15:23:31.418097 ignition[686]: no configs at "/usr/lib/ignition/base.d" Jan 30 15:23:31.418107 ignition[686]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Jan 30 15:23:31.418950 ignition[686]: parsed url from cmdline: "" Jan 30 15:23:31.418954 ignition[686]: no config URL provided Jan 30 15:23:31.418962 ignition[686]: reading system config file "/usr/lib/ignition/user.ign" Jan 30 15:23:31.418977 ignition[686]: no config at "/usr/lib/ignition/user.ign" Jan 30 15:23:31.423327 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 30 15:23:31.418983 ignition[686]: failed to fetch config: resource requires networking Jan 30 15:23:31.419197 ignition[686]: Ignition finished successfully Jan 30 15:23:31.438319 systemd-networkd[777]: lo: Link UP Jan 30 15:23:31.438333 systemd-networkd[777]: lo: Gained carrier Jan 30 15:23:31.441344 systemd-networkd[777]: Enumeration completed Jan 30 15:23:31.441968 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 30 15:23:31.443697 systemd-networkd[777]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 15:23:31.443700 systemd-networkd[777]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 15:23:31.444346 systemd[1]: Reached target network.target - Network. Jan 30 15:23:31.446790 systemd-networkd[777]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 15:23:31.446793 systemd-networkd[777]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 15:23:31.447346 systemd-networkd[777]: eth0: Link UP Jan 30 15:23:31.447350 systemd-networkd[777]: eth0: Gained carrier Jan 30 15:23:31.447358 systemd-networkd[777]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 15:23:31.450162 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jan 30 15:23:31.451823 systemd-networkd[777]: eth1: Link UP Jan 30 15:23:31.451826 systemd-networkd[777]: eth1: Gained carrier Jan 30 15:23:31.451837 systemd-networkd[777]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 15:23:31.469816 ignition[780]: Ignition 2.19.0 Jan 30 15:23:31.470546 ignition[780]: Stage: fetch Jan 30 15:23:31.471232 ignition[780]: no configs at "/usr/lib/ignition/base.d" Jan 30 15:23:31.471247 ignition[780]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Jan 30 15:23:31.471430 ignition[780]: parsed url from cmdline: "" Jan 30 15:23:31.471434 ignition[780]: no config URL provided Jan 30 15:23:31.471439 ignition[780]: reading system config file "/usr/lib/ignition/user.ign" Jan 30 15:23:31.471448 ignition[780]: no config at "/usr/lib/ignition/user.ign" Jan 30 15:23:31.471471 ignition[780]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #1 Jan 30 15:23:31.472172 ignition[780]: GET error: Get "http://169.254.169.254/hetzner/v1/userdata": dial tcp 169.254.169.254:80: connect: network is unreachable Jan 30 15:23:31.491022 systemd-networkd[777]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 30 15:23:31.517038 systemd-networkd[777]: eth0: DHCPv4 address 168.119.58.127/32, gateway 172.31.1.1 acquired from 172.31.1.1 Jan 30 15:23:31.672381 ignition[780]: GET http://169.254.169.254/hetzner/v1/userdata: attempt #2 Jan 30 15:23:31.677225 ignition[780]: GET result: OK Jan 30 15:23:31.677329 ignition[780]: parsing config with SHA512: 4a1c3e78c8e320c11b59b9908d6c4f8bae6e14cbde9430b9b4acd6cfbd8310b76bd093d6513f9702aaa7654bbc2706133b716b4cde41ff2e29c1c5ea954e05e2 Jan 30 15:23:31.682888 unknown[780]: fetched base config from "system" Jan 30 15:23:31.682897 unknown[780]: fetched base config from "system" Jan 30 15:23:31.683387 ignition[780]: fetch: fetch complete Jan 30 15:23:31.682902 unknown[780]: fetched user config from "hetzner" Jan 30 15:23:31.683395 ignition[780]: fetch: fetch passed Jan 30 15:23:31.686722 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jan 30 15:23:31.683440 ignition[780]: Ignition finished successfully Jan 30 15:23:31.693180 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 30 15:23:31.712475 ignition[788]: Ignition 2.19.0 Jan 30 15:23:31.712489 ignition[788]: Stage: kargs Jan 30 15:23:31.712712 ignition[788]: no configs at "/usr/lib/ignition/base.d" Jan 30 15:23:31.712722 ignition[788]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Jan 30 15:23:31.713752 ignition[788]: kargs: kargs passed Jan 30 15:23:31.713877 ignition[788]: Ignition finished successfully Jan 30 15:23:31.717901 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 30 15:23:31.724186 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 30 15:23:31.739213 ignition[794]: Ignition 2.19.0 Jan 30 15:23:31.739224 ignition[794]: Stage: disks Jan 30 15:23:31.739415 ignition[794]: no configs at "/usr/lib/ignition/base.d" Jan 30 15:23:31.739425 ignition[794]: no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Jan 30 15:23:31.740604 ignition[794]: disks: disks passed Jan 30 15:23:31.740723 ignition[794]: Ignition finished successfully Jan 30 15:23:31.742955 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 30 15:23:31.744072 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 30 15:23:31.744709 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 30 15:23:31.745859 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 30 15:23:31.747117 systemd[1]: Reached target sysinit.target - System Initialization. Jan 30 15:23:31.748035 systemd[1]: Reached target basic.target - Basic System. Jan 30 15:23:31.764280 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 30 15:23:31.789978 systemd-fsck[802]: ROOT: clean, 14/1628000 files, 120691/1617920 blocks Jan 30 15:23:31.793847 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 30 15:23:31.801314 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 30 15:23:31.856267 kernel: EXT4-fs (sda9): mounted filesystem 8499bb43-f860-448d-b3b8-5a1fc2b80abf r/w with ordered data mode. Quota mode: none. Jan 30 15:23:31.857780 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 30 15:23:31.859884 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 30 15:23:31.868212 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 30 15:23:31.872543 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 30 15:23:31.877178 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Jan 30 15:23:31.877959 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 30 15:23:31.877995 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 30 15:23:31.886111 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/sda6 scanned by mount (810) Jan 30 15:23:31.886982 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 30 15:23:31.889164 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 30 15:23:31.893017 kernel: BTRFS info (device sda6): first mount of filesystem db40e17a-cddf-4890-8d80-4d8cda0a956a Jan 30 15:23:31.893047 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Jan 30 15:23:31.893058 kernel: BTRFS info (device sda6): using free space tree Jan 30 15:23:31.903438 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 30 15:23:31.903506 kernel: BTRFS info (device sda6): auto enabling async discard Jan 30 15:23:31.906364 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 30 15:23:31.947938 coreos-metadata[812]: Jan 30 15:23:31.947 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/hostname: Attempt #1 Jan 30 15:23:31.950027 coreos-metadata[812]: Jan 30 15:23:31.949 INFO Fetch successful Jan 30 15:23:31.952061 coreos-metadata[812]: Jan 30 15:23:31.950 INFO wrote hostname ci-4081-3-0-a-5514912dcd to /sysroot/etc/hostname Jan 30 15:23:31.954284 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 30 15:23:31.956695 initrd-setup-root[838]: cut: /sysroot/etc/passwd: No such file or directory Jan 30 15:23:31.962068 initrd-setup-root[845]: cut: /sysroot/etc/group: No such file or directory Jan 30 15:23:31.966954 initrd-setup-root[852]: cut: /sysroot/etc/shadow: No such file or directory Jan 30 15:23:31.972301 initrd-setup-root[859]: cut: /sysroot/etc/gshadow: No such file or directory Jan 30 15:23:32.085514 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 30 15:23:32.092902 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 30 15:23:32.096194 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 30 15:23:32.104955 kernel: BTRFS info (device sda6): last unmount of filesystem db40e17a-cddf-4890-8d80-4d8cda0a956a Jan 30 15:23:32.130964 ignition[927]: INFO : Ignition 2.19.0 Jan 30 15:23:32.130964 ignition[927]: INFO : Stage: mount Jan 30 15:23:32.130964 ignition[927]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 15:23:32.133968 ignition[927]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Jan 30 15:23:32.133968 ignition[927]: INFO : mount: mount passed Jan 30 15:23:32.133968 ignition[927]: INFO : Ignition finished successfully Jan 30 15:23:32.137270 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 30 15:23:32.138645 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 30 15:23:32.146023 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 30 15:23:32.226567 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 30 15:23:32.232247 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 30 15:23:32.243996 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (939) Jan 30 15:23:32.245604 kernel: BTRFS info (device sda6): first mount of filesystem db40e17a-cddf-4890-8d80-4d8cda0a956a Jan 30 15:23:32.245660 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Jan 30 15:23:32.245672 kernel: BTRFS info (device sda6): using free space tree Jan 30 15:23:32.249023 kernel: BTRFS info (device sda6): enabling ssd optimizations Jan 30 15:23:32.249114 kernel: BTRFS info (device sda6): auto enabling async discard Jan 30 15:23:32.252050 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 30 15:23:32.274865 ignition[956]: INFO : Ignition 2.19.0 Jan 30 15:23:32.274865 ignition[956]: INFO : Stage: files Jan 30 15:23:32.276318 ignition[956]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 15:23:32.276318 ignition[956]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Jan 30 15:23:32.276318 ignition[956]: DEBUG : files: compiled without relabeling support, skipping Jan 30 15:23:32.279397 ignition[956]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 30 15:23:32.279397 ignition[956]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 30 15:23:32.284658 ignition[956]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 30 15:23:32.286799 ignition[956]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 30 15:23:32.288987 unknown[956]: wrote ssh authorized keys file for user: core Jan 30 15:23:32.290047 ignition[956]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 30 15:23:32.293340 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Jan 30 15:23:32.293340 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.17.0-linux-arm64.tar.gz: attempt #1 Jan 30 15:23:32.449251 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 30 15:23:32.609198 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.17.0-linux-arm64.tar.gz" Jan 30 15:23:32.609198 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Jan 30 15:23:32.611501 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Jan 30 15:23:32.628033 systemd-networkd[777]: eth0: Gained IPv6LL Jan 30 15:23:33.173782 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Jan 30 15:23:33.265292 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Jan 30 15:23:33.266640 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Jan 30 15:23:33.267705 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Jan 30 15:23:33.267705 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Jan 30 15:23:33.270692 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.32.0-arm64.raw: attempt #1 Jan 30 15:23:33.269884 systemd-networkd[777]: eth1: Gained IPv6LL Jan 30 15:23:33.532133 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Jan 30 15:23:33.933039 ignition[956]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.32.0-arm64.raw" Jan 30 15:23:33.933039 ignition[956]: INFO : files: op(c): [started] processing unit "prepare-helm.service" Jan 30 15:23:33.937378 ignition[956]: INFO : files: op(c): op(d): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 30 15:23:33.937378 ignition[956]: INFO : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 30 15:23:33.937378 ignition[956]: INFO : files: op(c): [finished] processing unit "prepare-helm.service" Jan 30 15:23:33.937378 ignition[956]: INFO : files: op(e): [started] processing unit "coreos-metadata.service" Jan 30 15:23:33.937378 ignition[956]: INFO : files: op(e): op(f): [started] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Jan 30 15:23:33.937378 ignition[956]: INFO : files: op(e): op(f): [finished] writing systemd drop-in "00-custom-metadata.conf" at "/sysroot/etc/systemd/system/coreos-metadata.service.d/00-custom-metadata.conf" Jan 30 15:23:33.937378 ignition[956]: INFO : files: op(e): [finished] processing unit "coreos-metadata.service" Jan 30 15:23:33.937378 ignition[956]: INFO : files: op(10): [started] setting preset to enabled for "prepare-helm.service" Jan 30 15:23:33.937378 ignition[956]: INFO : files: op(10): [finished] setting preset to enabled for "prepare-helm.service" Jan 30 15:23:33.937378 ignition[956]: INFO : files: createResultFile: createFiles: op(11): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 30 15:23:33.937378 ignition[956]: INFO : files: createResultFile: createFiles: op(11): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 30 15:23:33.937378 ignition[956]: INFO : files: files passed Jan 30 15:23:33.937378 ignition[956]: INFO : Ignition finished successfully Jan 30 15:23:33.941172 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 30 15:23:33.946191 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 30 15:23:33.954194 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 30 15:23:33.962853 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 30 15:23:33.964965 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 30 15:23:33.967528 initrd-setup-root-after-ignition[984]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 30 15:23:33.967528 initrd-setup-root-after-ignition[984]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 30 15:23:33.970891 initrd-setup-root-after-ignition[988]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 30 15:23:33.974953 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 30 15:23:33.977276 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 30 15:23:33.983197 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 30 15:23:34.037353 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 30 15:23:34.037541 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 30 15:23:34.040541 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 30 15:23:34.041940 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 30 15:23:34.043035 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 30 15:23:34.049231 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 30 15:23:34.062840 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 30 15:23:34.070200 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 30 15:23:34.083406 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 30 15:23:34.085448 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 15:23:34.086511 systemd[1]: Stopped target timers.target - Timer Units. Jan 30 15:23:34.088358 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 30 15:23:34.088485 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 30 15:23:34.089847 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 30 15:23:34.090498 systemd[1]: Stopped target basic.target - Basic System. Jan 30 15:23:34.091498 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 30 15:23:34.092479 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 30 15:23:34.093614 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 30 15:23:34.094680 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 30 15:23:34.095696 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 30 15:23:34.096856 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 30 15:23:34.097813 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 30 15:23:34.098919 systemd[1]: Stopped target swap.target - Swaps. Jan 30 15:23:34.099722 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 30 15:23:34.099889 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 30 15:23:34.101184 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 30 15:23:34.102220 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 15:23:34.103225 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 30 15:23:34.103596 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 15:23:34.104401 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 30 15:23:34.104568 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 30 15:23:34.105888 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 30 15:23:34.106079 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 30 15:23:34.107274 systemd[1]: ignition-files.service: Deactivated successfully. Jan 30 15:23:34.107425 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 30 15:23:34.108236 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Jan 30 15:23:34.108383 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 30 15:23:34.114266 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 30 15:23:34.118969 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 30 15:23:34.119458 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 30 15:23:34.119670 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 15:23:34.120551 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 30 15:23:34.120663 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 30 15:23:34.131305 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 30 15:23:34.131399 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 30 15:23:34.137959 ignition[1008]: INFO : Ignition 2.19.0 Jan 30 15:23:34.137959 ignition[1008]: INFO : Stage: umount Jan 30 15:23:34.137959 ignition[1008]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 15:23:34.137959 ignition[1008]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/hetzner" Jan 30 15:23:34.145117 ignition[1008]: INFO : umount: umount passed Jan 30 15:23:34.145117 ignition[1008]: INFO : Ignition finished successfully Jan 30 15:23:34.142395 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 30 15:23:34.142505 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 30 15:23:34.146353 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 30 15:23:34.147298 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 30 15:23:34.147393 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 30 15:23:34.148216 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 30 15:23:34.148263 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 30 15:23:34.149251 systemd[1]: ignition-fetch.service: Deactivated successfully. Jan 30 15:23:34.149292 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jan 30 15:23:34.150309 systemd[1]: Stopped target network.target - Network. Jan 30 15:23:34.151161 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 30 15:23:34.151212 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 30 15:23:34.152180 systemd[1]: Stopped target paths.target - Path Units. Jan 30 15:23:34.152988 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 30 15:23:34.156976 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 15:23:34.158440 systemd[1]: Stopped target slices.target - Slice Units. Jan 30 15:23:34.159645 systemd[1]: Stopped target sockets.target - Socket Units. Jan 30 15:23:34.161094 systemd[1]: iscsid.socket: Deactivated successfully. Jan 30 15:23:34.161158 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 30 15:23:34.162248 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 30 15:23:34.162297 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 30 15:23:34.163443 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 30 15:23:34.163511 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 30 15:23:34.164371 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 30 15:23:34.164409 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 30 15:23:34.165502 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 30 15:23:34.166188 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 30 15:23:34.167342 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 30 15:23:34.167425 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 30 15:23:34.168446 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 30 15:23:34.168523 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 30 15:23:34.173049 systemd-networkd[777]: eth0: DHCPv6 lease lost Jan 30 15:23:34.181029 systemd-networkd[777]: eth1: DHCPv6 lease lost Jan 30 15:23:34.183499 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 30 15:23:34.183827 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 30 15:23:34.188057 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 30 15:23:34.188197 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 30 15:23:34.191160 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 30 15:23:34.191231 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 30 15:23:34.198093 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 30 15:23:34.199042 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 30 15:23:34.199155 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 30 15:23:34.203294 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 30 15:23:34.203375 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 30 15:23:34.205150 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 30 15:23:34.205233 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 30 15:23:34.206455 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 30 15:23:34.206519 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 15:23:34.207863 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 15:23:34.223863 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 30 15:23:34.224109 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 15:23:34.225886 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 30 15:23:34.226056 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 30 15:23:34.228195 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 30 15:23:34.228282 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 30 15:23:34.229215 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 30 15:23:34.229259 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 15:23:34.230160 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 30 15:23:34.230213 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 30 15:23:34.231696 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 30 15:23:34.231743 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 30 15:23:34.233300 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 30 15:23:34.233356 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 15:23:34.240191 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 30 15:23:34.240956 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 30 15:23:34.241054 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 15:23:34.244972 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jan 30 15:23:34.245927 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 15:23:34.247715 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 30 15:23:34.247785 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 15:23:34.249172 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 15:23:34.249230 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 15:23:34.250615 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 30 15:23:34.250726 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 30 15:23:34.251980 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 30 15:23:34.257113 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 30 15:23:34.267862 systemd[1]: Switching root. Jan 30 15:23:34.295945 systemd-journald[237]: Journal stopped Jan 30 15:23:35.339474 systemd-journald[237]: Received SIGTERM from PID 1 (systemd). Jan 30 15:23:35.339551 kernel: SELinux: policy capability network_peer_controls=1 Jan 30 15:23:35.339563 kernel: SELinux: policy capability open_perms=1 Jan 30 15:23:35.339578 kernel: SELinux: policy capability extended_socket_class=1 Jan 30 15:23:35.339597 kernel: SELinux: policy capability always_check_network=0 Jan 30 15:23:35.339611 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 30 15:23:35.339621 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 30 15:23:35.339633 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 30 15:23:35.339643 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 30 15:23:35.339652 kernel: audit: type=1403 audit(1738250614.486:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 30 15:23:35.339663 systemd[1]: Successfully loaded SELinux policy in 38.901ms. Jan 30 15:23:35.339689 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.167ms. Jan 30 15:23:35.339700 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 30 15:23:35.339711 systemd[1]: Detected virtualization kvm. Jan 30 15:23:35.339722 systemd[1]: Detected architecture arm64. Jan 30 15:23:35.339739 systemd[1]: Detected first boot. Jan 30 15:23:35.339750 systemd[1]: Hostname set to . Jan 30 15:23:35.339763 systemd[1]: Initializing machine ID from VM UUID. Jan 30 15:23:35.339776 zram_generator::config[1052]: No configuration found. Jan 30 15:23:35.339793 systemd[1]: Populated /etc with preset unit settings. Jan 30 15:23:35.339807 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 30 15:23:35.339819 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 30 15:23:35.339829 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 30 15:23:35.339842 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 30 15:23:35.339861 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 30 15:23:35.339873 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 30 15:23:35.339885 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 30 15:23:35.339899 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 30 15:23:35.339929 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 30 15:23:35.339941 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 30 15:23:35.339953 systemd[1]: Created slice user.slice - User and Session Slice. Jan 30 15:23:35.339967 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 15:23:35.339981 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 15:23:35.339993 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 30 15:23:35.340055 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 30 15:23:35.340078 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 30 15:23:35.340093 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 30 15:23:35.340107 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Jan 30 15:23:35.340119 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 15:23:35.340130 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 30 15:23:35.340144 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 30 15:23:35.340156 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 30 15:23:35.340168 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 30 15:23:35.340181 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 15:23:35.340193 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 30 15:23:35.340206 systemd[1]: Reached target slices.target - Slice Units. Jan 30 15:23:35.340217 systemd[1]: Reached target swap.target - Swaps. Jan 30 15:23:35.340231 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 30 15:23:35.340244 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 30 15:23:35.340255 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 30 15:23:35.340266 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 30 15:23:35.340277 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 15:23:35.340290 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 30 15:23:35.340300 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 30 15:23:35.340310 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 30 15:23:35.340321 systemd[1]: Mounting media.mount - External Media Directory... Jan 30 15:23:35.340333 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 30 15:23:35.340343 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 30 15:23:35.340357 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 30 15:23:35.340370 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 30 15:23:35.340381 systemd[1]: Reached target machines.target - Containers. Jan 30 15:23:35.340392 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 30 15:23:35.340402 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 15:23:35.340414 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 30 15:23:35.340425 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 30 15:23:35.340438 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 30 15:23:35.340450 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 30 15:23:35.340460 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 30 15:23:35.340475 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 30 15:23:35.340487 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 30 15:23:35.340502 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 30 15:23:35.340515 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 30 15:23:35.340531 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 30 15:23:35.340542 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 30 15:23:35.340555 systemd[1]: Stopped systemd-fsck-usr.service. Jan 30 15:23:35.340566 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 30 15:23:35.340576 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 30 15:23:35.340617 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 30 15:23:35.340635 kernel: loop: module loaded Jan 30 15:23:35.340647 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 30 15:23:35.340658 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 30 15:23:35.340669 systemd[1]: verity-setup.service: Deactivated successfully. Jan 30 15:23:35.340681 systemd[1]: Stopped verity-setup.service. Jan 30 15:23:35.340692 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 30 15:23:35.340706 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 30 15:23:35.340716 systemd[1]: Mounted media.mount - External Media Directory. Jan 30 15:23:35.340726 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 30 15:23:35.340737 kernel: fuse: init (API version 7.39) Jan 30 15:23:35.340747 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 30 15:23:35.340758 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 30 15:23:35.340770 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 15:23:35.340781 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 30 15:23:35.340792 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 30 15:23:35.340803 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 30 15:23:35.340815 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 30 15:23:35.340826 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 30 15:23:35.340837 kernel: ACPI: bus type drm_connector registered Jan 30 15:23:35.340851 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 30 15:23:35.340862 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 30 15:23:35.340873 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 30 15:23:35.340884 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 30 15:23:35.340895 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 30 15:23:35.342970 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 30 15:23:35.343018 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 30 15:23:35.343029 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 30 15:23:35.343041 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 30 15:23:35.343052 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 30 15:23:35.343063 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 30 15:23:35.343103 systemd-journald[1129]: Collecting audit messages is disabled. Jan 30 15:23:35.343133 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 30 15:23:35.343147 systemd-journald[1129]: Journal started Jan 30 15:23:35.343172 systemd-journald[1129]: Runtime Journal (/run/log/journal/9be7bbb2090d46e5b9a67a98f7c71fe3) is 8.0M, max 76.5M, 68.5M free. Jan 30 15:23:35.005023 systemd[1]: Queued start job for default target multi-user.target. Jan 30 15:23:35.026522 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Jan 30 15:23:35.027365 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 30 15:23:35.347116 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 30 15:23:35.350353 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 30 15:23:35.357559 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 30 15:23:35.357682 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 30 15:23:35.364405 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jan 30 15:23:35.375926 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 30 15:23:35.384212 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 30 15:23:35.386042 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 15:23:35.396180 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 30 15:23:35.399944 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 30 15:23:35.402847 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 30 15:23:35.407578 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 30 15:23:35.413543 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 30 15:23:35.422800 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 30 15:23:35.426307 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 30 15:23:35.432022 systemd[1]: Started systemd-journald.service - Journal Service. Jan 30 15:23:35.433056 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 15:23:35.434143 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 30 15:23:35.437394 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 30 15:23:35.441525 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 30 15:23:35.486432 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 30 15:23:35.490941 kernel: loop0: detected capacity change from 0 to 114328 Jan 30 15:23:35.491690 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 30 15:23:35.500173 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 30 15:23:35.505097 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jan 30 15:23:35.511104 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jan 30 15:23:35.522518 systemd-tmpfiles[1149]: ACLs are not supported, ignoring. Jan 30 15:23:35.523074 systemd-tmpfiles[1149]: ACLs are not supported, ignoring. Jan 30 15:23:35.530965 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 30 15:23:35.530735 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 30 15:23:35.540035 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 15:23:35.549153 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 30 15:23:35.555161 systemd-journald[1129]: Time spent on flushing to /var/log/journal/9be7bbb2090d46e5b9a67a98f7c71fe3 is 40.167ms for 1143 entries. Jan 30 15:23:35.555161 systemd-journald[1129]: System Journal (/var/log/journal/9be7bbb2090d46e5b9a67a98f7c71fe3) is 8.0M, max 584.8M, 576.8M free. Jan 30 15:23:35.614805 systemd-journald[1129]: Received client request to flush runtime journal. Jan 30 15:23:35.614857 kernel: loop1: detected capacity change from 0 to 8 Jan 30 15:23:35.614872 kernel: loop2: detected capacity change from 0 to 201592 Jan 30 15:23:35.568960 udevadm[1176]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jan 30 15:23:35.581671 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 30 15:23:35.584979 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jan 30 15:23:35.625625 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 30 15:23:35.637815 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 30 15:23:35.646934 kernel: loop3: detected capacity change from 0 to 114432 Jan 30 15:23:35.651182 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 30 15:23:35.685203 systemd-tmpfiles[1190]: ACLs are not supported, ignoring. Jan 30 15:23:35.685221 systemd-tmpfiles[1190]: ACLs are not supported, ignoring. Jan 30 15:23:35.691950 kernel: loop4: detected capacity change from 0 to 114328 Jan 30 15:23:35.699792 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 15:23:35.717937 kernel: loop5: detected capacity change from 0 to 8 Jan 30 15:23:35.722962 kernel: loop6: detected capacity change from 0 to 201592 Jan 30 15:23:35.752219 kernel: loop7: detected capacity change from 0 to 114432 Jan 30 15:23:35.774733 (sd-merge)[1193]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-hetzner'. Jan 30 15:23:35.775918 (sd-merge)[1193]: Merged extensions into '/usr'. Jan 30 15:23:35.783695 systemd[1]: Reloading requested from client PID 1148 ('systemd-sysext') (unit systemd-sysext.service)... Jan 30 15:23:35.783718 systemd[1]: Reloading... Jan 30 15:23:35.919951 zram_generator::config[1224]: No configuration found. Jan 30 15:23:35.985937 ldconfig[1144]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 30 15:23:36.055262 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 15:23:36.103053 systemd[1]: Reloading finished in 318 ms. Jan 30 15:23:36.156434 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 30 15:23:36.158987 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 30 15:23:36.171128 systemd[1]: Starting ensure-sysext.service... Jan 30 15:23:36.174569 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 30 15:23:36.202307 systemd[1]: Reloading requested from client PID 1258 ('systemctl') (unit ensure-sysext.service)... Jan 30 15:23:36.202341 systemd[1]: Reloading... Jan 30 15:23:36.244811 systemd-tmpfiles[1259]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 30 15:23:36.245672 systemd-tmpfiles[1259]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jan 30 15:23:36.246733 systemd-tmpfiles[1259]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jan 30 15:23:36.247207 systemd-tmpfiles[1259]: ACLs are not supported, ignoring. Jan 30 15:23:36.247335 systemd-tmpfiles[1259]: ACLs are not supported, ignoring. Jan 30 15:23:36.252008 systemd-tmpfiles[1259]: Detected autofs mount point /boot during canonicalization of boot. Jan 30 15:23:36.252176 systemd-tmpfiles[1259]: Skipping /boot Jan 30 15:23:36.262328 systemd-tmpfiles[1259]: Detected autofs mount point /boot during canonicalization of boot. Jan 30 15:23:36.262463 systemd-tmpfiles[1259]: Skipping /boot Jan 30 15:23:36.305939 zram_generator::config[1291]: No configuration found. Jan 30 15:23:36.408951 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 15:23:36.457047 systemd[1]: Reloading finished in 254 ms. Jan 30 15:23:36.477026 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 30 15:23:36.482622 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 15:23:36.500362 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jan 30 15:23:36.507618 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 30 15:23:36.512506 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 30 15:23:36.520898 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 30 15:23:36.525326 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 15:23:36.530380 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 30 15:23:36.535659 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 15:23:36.545274 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 30 15:23:36.556022 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 30 15:23:36.559996 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 30 15:23:36.560826 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 15:23:36.573833 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 30 15:23:36.578536 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 30 15:23:36.592605 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 15:23:36.592830 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 15:23:36.596018 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 30 15:23:36.601698 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 30 15:23:36.603012 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 30 15:23:36.611853 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 30 15:23:36.612128 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 30 15:23:36.619136 systemd[1]: Finished ensure-sysext.service. Jan 30 15:23:36.630030 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 15:23:36.634784 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 30 15:23:36.635944 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 15:23:36.636160 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 30 15:23:36.652568 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jan 30 15:23:36.655608 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 30 15:23:36.658571 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 30 15:23:36.658788 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 30 15:23:36.661387 systemd-udevd[1332]: Using default interface naming scheme 'v255'. Jan 30 15:23:36.662837 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 30 15:23:36.665136 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 30 15:23:36.666545 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 30 15:23:36.670805 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 30 15:23:36.674512 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 30 15:23:36.675202 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 30 15:23:36.705218 augenrules[1361]: No rules Jan 30 15:23:36.707331 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jan 30 15:23:36.713655 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 15:23:36.725325 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 30 15:23:36.726453 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 30 15:23:36.862312 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Jan 30 15:23:36.865380 systemd-networkd[1371]: lo: Link UP Jan 30 15:23:36.865899 systemd-networkd[1371]: lo: Gained carrier Jan 30 15:23:36.869285 systemd-networkd[1371]: Enumeration completed Jan 30 15:23:36.869453 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 30 15:23:36.881408 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 30 15:23:36.895435 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jan 30 15:23:36.896428 systemd[1]: Reached target time-set.target - System Time Set. Jan 30 15:23:36.911676 systemd-resolved[1328]: Positive Trust Anchors: Jan 30 15:23:36.912049 systemd-resolved[1328]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 30 15:23:36.912084 systemd-resolved[1328]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 30 15:23:36.919601 systemd-resolved[1328]: Using system hostname 'ci-4081-3-0-a-5514912dcd'. Jan 30 15:23:36.922094 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 30 15:23:36.923507 systemd[1]: Reached target network.target - Network. Jan 30 15:23:36.924849 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 30 15:23:37.013997 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1394) Jan 30 15:23:37.020360 systemd-networkd[1371]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 15:23:37.020375 systemd-networkd[1371]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 15:23:37.022277 systemd-networkd[1371]: eth0: Link UP Jan 30 15:23:37.022293 systemd-networkd[1371]: eth0: Gained carrier Jan 30 15:23:37.022317 systemd-networkd[1371]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 15:23:37.028935 kernel: mousedev: PS/2 mouse device common for all mice Jan 30 15:23:37.078168 systemd-networkd[1371]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 15:23:37.078179 systemd-networkd[1371]: eth1: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 15:23:37.078766 systemd-networkd[1371]: eth1: Link UP Jan 30 15:23:37.078777 systemd-networkd[1371]: eth1: Gained carrier Jan 30 15:23:37.078794 systemd-networkd[1371]: eth1: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 15:23:37.088138 systemd-networkd[1371]: eth0: DHCPv4 address 168.119.58.127/32, gateway 172.31.1.1 acquired from 172.31.1.1 Jan 30 15:23:37.089349 systemd-timesyncd[1353]: Network configuration changed, trying to establish connection. Jan 30 15:23:37.101483 systemd[1]: Condition check resulted in dev-virtio\x2dports-org.qemu.guest_agent.0.device - /dev/virtio-ports/org.qemu.guest_agent.0 being skipped. Jan 30 15:23:37.101684 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 15:23:37.121287 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 30 15:23:37.123281 systemd-networkd[1371]: eth1: DHCPv4 address 10.0.0.3/32, gateway 10.0.0.1 acquired from 10.0.0.1 Jan 30 15:23:37.131399 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 30 15:23:37.138349 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 30 15:23:37.139078 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 15:23:37.139119 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 30 15:23:37.139512 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 30 15:23:37.140317 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 30 15:23:37.142402 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 30 15:23:37.142592 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 30 15:23:37.158548 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - QEMU_HARDDISK OEM. Jan 30 15:23:37.161553 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 30 15:23:37.164191 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 30 15:23:37.184646 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 30 15:23:37.186496 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 30 15:23:37.186647 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 30 15:23:37.190027 kernel: [drm] pci: virtio-gpu-pci detected at 0000:00:01.0 Jan 30 15:23:37.190112 kernel: [drm] features: -virgl +edid -resource_blob -host_visible Jan 30 15:23:37.190128 kernel: [drm] features: -context_init Jan 30 15:23:37.192928 kernel: [drm] number of scanouts: 1 Jan 30 15:23:37.193009 kernel: [drm] number of cap sets: 0 Jan 30 15:23:37.195939 kernel: [drm] Initialized virtio_gpu 0.1.0 0 for 0000:00:01.0 on minor 0 Jan 30 15:23:37.203081 kernel: Console: switching to colour frame buffer device 160x50 Jan 30 15:23:37.204413 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 15:23:37.209940 kernel: virtio-pci 0000:00:01.0: [drm] fb0: virtio_gpudrmfb frame buffer device Jan 30 15:23:37.227522 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 30 15:23:37.229859 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 15:23:37.230141 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 15:23:37.239280 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 15:23:36.887582 systemd-resolved[1328]: Clock change detected. Flushing caches. Jan 30 15:23:36.898503 systemd-journald[1129]: Time jumped backwards, rotating. Jan 30 15:23:36.888414 systemd-timesyncd[1353]: Contacted time server 207.180.217.145:123 (0.flatcar.pool.ntp.org). Jan 30 15:23:36.888492 systemd-timesyncd[1353]: Initial clock synchronization to Thu 2025-01-30 15:23:36.887500 UTC. Jan 30 15:23:36.912818 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 15:23:36.932397 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jan 30 15:23:36.939131 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jan 30 15:23:36.965803 lvm[1439]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 30 15:23:36.997624 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jan 30 15:23:37.000374 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 30 15:23:37.001298 systemd[1]: Reached target sysinit.target - System Initialization. Jan 30 15:23:37.002607 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 30 15:23:37.003524 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 30 15:23:37.004726 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 30 15:23:37.005681 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 30 15:23:37.006634 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 30 15:23:37.007480 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 30 15:23:37.007640 systemd[1]: Reached target paths.target - Path Units. Jan 30 15:23:37.008231 systemd[1]: Reached target timers.target - Timer Units. Jan 30 15:23:37.010933 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 30 15:23:37.014666 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 30 15:23:37.028143 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 30 15:23:37.040875 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jan 30 15:23:37.042921 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 30 15:23:37.044868 systemd[1]: Reached target sockets.target - Socket Units. Jan 30 15:23:37.046052 systemd[1]: Reached target basic.target - Basic System. Jan 30 15:23:37.046618 lvm[1443]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 30 15:23:37.047312 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 30 15:23:37.047390 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 30 15:23:37.055772 systemd[1]: Starting containerd.service - containerd container runtime... Jan 30 15:23:37.059602 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Jan 30 15:23:37.064022 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 30 15:23:37.075792 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 30 15:23:37.081932 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 30 15:23:37.085169 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 30 15:23:37.087823 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 30 15:23:37.093764 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 30 15:23:37.099837 systemd[1]: Started qemu-guest-agent.service - QEMU Guest Agent. Jan 30 15:23:37.106718 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 30 15:23:37.110953 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 30 15:23:37.117828 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 30 15:23:37.120514 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 30 15:23:37.123165 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 30 15:23:37.124790 systemd[1]: Starting update-engine.service - Update Engine... Jan 30 15:23:37.129274 jq[1447]: false Jan 30 15:23:37.133304 coreos-metadata[1445]: Jan 30 15:23:37.128 INFO Fetching http://169.254.169.254/hetzner/v1/metadata: Attempt #1 Jan 30 15:23:37.133304 coreos-metadata[1445]: Jan 30 15:23:37.132 INFO Fetch successful Jan 30 15:23:37.133304 coreos-metadata[1445]: Jan 30 15:23:37.132 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/private-networks: Attempt #1 Jan 30 15:23:37.133146 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 30 15:23:37.135028 coreos-metadata[1445]: Jan 30 15:23:37.133 INFO Fetch successful Jan 30 15:23:37.135930 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jan 30 15:23:37.146700 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 30 15:23:37.147637 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 30 15:23:37.174419 jq[1460]: true Jan 30 15:23:37.212287 systemd[1]: motdgen.service: Deactivated successfully. Jan 30 15:23:37.213853 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 30 15:23:37.214903 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 30 15:23:37.215569 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 30 15:23:37.219805 update_engine[1458]: I20250130 15:23:37.217754 1458 main.cc:92] Flatcar Update Engine starting Jan 30 15:23:37.224309 extend-filesystems[1448]: Found loop4 Jan 30 15:23:37.224309 extend-filesystems[1448]: Found loop5 Jan 30 15:23:37.224309 extend-filesystems[1448]: Found loop6 Jan 30 15:23:37.224309 extend-filesystems[1448]: Found loop7 Jan 30 15:23:37.224309 extend-filesystems[1448]: Found sda Jan 30 15:23:37.224309 extend-filesystems[1448]: Found sda1 Jan 30 15:23:37.224309 extend-filesystems[1448]: Found sda2 Jan 30 15:23:37.224309 extend-filesystems[1448]: Found sda3 Jan 30 15:23:37.239999 extend-filesystems[1448]: Found usr Jan 30 15:23:37.239999 extend-filesystems[1448]: Found sda4 Jan 30 15:23:37.239999 extend-filesystems[1448]: Found sda6 Jan 30 15:23:37.239999 extend-filesystems[1448]: Found sda7 Jan 30 15:23:37.239999 extend-filesystems[1448]: Found sda9 Jan 30 15:23:37.239999 extend-filesystems[1448]: Checking size of /dev/sda9 Jan 30 15:23:37.286141 kernel: EXT4-fs (sda9): resizing filesystem from 1617920 to 9393147 blocks Jan 30 15:23:37.239912 dbus-daemon[1446]: [system] SELinux support is enabled Jan 30 15:23:37.288436 extend-filesystems[1448]: Resized partition /dev/sda9 Jan 30 15:23:37.240471 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 30 15:23:37.293441 jq[1479]: true Jan 30 15:23:37.294248 update_engine[1458]: I20250130 15:23:37.244133 1458 update_check_scheduler.cc:74] Next update check in 7m39s Jan 30 15:23:37.298279 extend-filesystems[1491]: resize2fs 1.47.1 (20-May-2024) Jan 30 15:23:37.240660 (ntainerd)[1476]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jan 30 15:23:37.307997 tar[1475]: linux-arm64/LICENSE Jan 30 15:23:37.307997 tar[1475]: linux-arm64/helm Jan 30 15:23:37.269964 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 30 15:23:37.270028 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 30 15:23:37.282224 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 30 15:23:37.282257 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 30 15:23:37.295254 systemd[1]: Started update-engine.service - Update Engine. Jan 30 15:23:37.305872 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 30 15:23:37.326502 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Jan 30 15:23:37.328332 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 30 15:23:37.332988 systemd-logind[1456]: New seat seat0. Jan 30 15:23:37.334656 systemd-logind[1456]: Watching system buttons on /dev/input/event0 (Power Button) Jan 30 15:23:37.334675 systemd-logind[1456]: Watching system buttons on /dev/input/event2 (QEMU QEMU USB Keyboard) Jan 30 15:23:37.343877 systemd[1]: Started systemd-logind.service - User Login Management. Jan 30 15:23:37.407976 bash[1515]: Updated "/home/core/.ssh/authorized_keys" Jan 30 15:23:37.407609 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 30 15:23:37.427624 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1384) Jan 30 15:23:37.439333 systemd[1]: Starting sshkeys.service... Jan 30 15:23:37.497006 kernel: EXT4-fs (sda9): resized filesystem to 9393147 Jan 30 15:23:37.498876 systemd[1]: Created slice system-coreos\x2dmetadata\x2dsshkeys.slice - Slice /system/coreos-metadata-sshkeys. Jan 30 15:23:37.513030 systemd[1]: Starting coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys)... Jan 30 15:23:37.533506 extend-filesystems[1491]: Filesystem at /dev/sda9 is mounted on /; on-line resizing required Jan 30 15:23:37.533506 extend-filesystems[1491]: old_desc_blocks = 1, new_desc_blocks = 5 Jan 30 15:23:37.533506 extend-filesystems[1491]: The filesystem on /dev/sda9 is now 9393147 (4k) blocks long. Jan 30 15:23:37.538222 extend-filesystems[1448]: Resized filesystem in /dev/sda9 Jan 30 15:23:37.538222 extend-filesystems[1448]: Found sr0 Jan 30 15:23:37.534922 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 30 15:23:37.535125 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 30 15:23:37.546610 coreos-metadata[1526]: Jan 30 15:23:37.545 INFO Fetching http://169.254.169.254/hetzner/v1/metadata/public-keys: Attempt #1 Jan 30 15:23:37.546974 coreos-metadata[1526]: Jan 30 15:23:37.546 INFO Fetch successful Jan 30 15:23:37.550869 unknown[1526]: wrote ssh authorized keys file for user: core Jan 30 15:23:37.588759 update-ssh-keys[1532]: Updated "/home/core/.ssh/authorized_keys" Jan 30 15:23:37.590511 systemd[1]: Finished coreos-metadata-sshkeys@core.service - Flatcar Metadata Agent (SSH Keys). Jan 30 15:23:37.597132 systemd[1]: Finished sshkeys.service. Jan 30 15:23:37.619117 locksmithd[1501]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 30 15:23:37.706614 containerd[1476]: time="2025-01-30T15:23:37.704870811Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21 Jan 30 15:23:37.777554 containerd[1476]: time="2025-01-30T15:23:37.775452171Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jan 30 15:23:37.783359 containerd[1476]: time="2025-01-30T15:23:37.783296931Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jan 30 15:23:37.783359 containerd[1476]: time="2025-01-30T15:23:37.783349091Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jan 30 15:23:37.783359 containerd[1476]: time="2025-01-30T15:23:37.783371291Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jan 30 15:23:37.784492 containerd[1476]: time="2025-01-30T15:23:37.783559171Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jan 30 15:23:37.784551 containerd[1476]: time="2025-01-30T15:23:37.784504091Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jan 30 15:23:37.784964 containerd[1476]: time="2025-01-30T15:23:37.784928091Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 15:23:37.785004 containerd[1476]: time="2025-01-30T15:23:37.784963291Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jan 30 15:23:37.785217 containerd[1476]: time="2025-01-30T15:23:37.785175731Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 15:23:37.785217 containerd[1476]: time="2025-01-30T15:23:37.785209691Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jan 30 15:23:37.785272 containerd[1476]: time="2025-01-30T15:23:37.785228571Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 15:23:37.785272 containerd[1476]: time="2025-01-30T15:23:37.785240931Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jan 30 15:23:37.785352 containerd[1476]: time="2025-01-30T15:23:37.785333171Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jan 30 15:23:37.787233 containerd[1476]: time="2025-01-30T15:23:37.787180531Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jan 30 15:23:37.787408 containerd[1476]: time="2025-01-30T15:23:37.787384891Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 15:23:37.787435 containerd[1476]: time="2025-01-30T15:23:37.787406931Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jan 30 15:23:37.788074 containerd[1476]: time="2025-01-30T15:23:37.787528691Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jan 30 15:23:37.788166 containerd[1476]: time="2025-01-30T15:23:37.788145291Z" level=info msg="metadata content store policy set" policy=shared Jan 30 15:23:37.796239 containerd[1476]: time="2025-01-30T15:23:37.796159451Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jan 30 15:23:37.796400 containerd[1476]: time="2025-01-30T15:23:37.796273491Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jan 30 15:23:37.796400 containerd[1476]: time="2025-01-30T15:23:37.796294211Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jan 30 15:23:37.796400 containerd[1476]: time="2025-01-30T15:23:37.796321291Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jan 30 15:23:37.796400 containerd[1476]: time="2025-01-30T15:23:37.796335891Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jan 30 15:23:37.797671 containerd[1476]: time="2025-01-30T15:23:37.796528491Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jan 30 15:23:37.797882 containerd[1476]: time="2025-01-30T15:23:37.797858571Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jan 30 15:23:37.798038 containerd[1476]: time="2025-01-30T15:23:37.798016651Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jan 30 15:23:37.798064 containerd[1476]: time="2025-01-30T15:23:37.798040851Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jan 30 15:23:37.798064 containerd[1476]: time="2025-01-30T15:23:37.798055731Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jan 30 15:23:37.798097 containerd[1476]: time="2025-01-30T15:23:37.798081491Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jan 30 15:23:37.798129 containerd[1476]: time="2025-01-30T15:23:37.798096091Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jan 30 15:23:37.798129 containerd[1476]: time="2025-01-30T15:23:37.798112251Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jan 30 15:23:37.798162 containerd[1476]: time="2025-01-30T15:23:37.798127211Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jan 30 15:23:37.798162 containerd[1476]: time="2025-01-30T15:23:37.798143011Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jan 30 15:23:37.798162 containerd[1476]: time="2025-01-30T15:23:37.798155331Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jan 30 15:23:37.798222 containerd[1476]: time="2025-01-30T15:23:37.798168411Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jan 30 15:23:37.798222 containerd[1476]: time="2025-01-30T15:23:37.798180731Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jan 30 15:23:37.798262 containerd[1476]: time="2025-01-30T15:23:37.798242771Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798279 containerd[1476]: time="2025-01-30T15:23:37.798263931Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798300 containerd[1476]: time="2025-01-30T15:23:37.798277891Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798300 containerd[1476]: time="2025-01-30T15:23:37.798292251Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798336 containerd[1476]: time="2025-01-30T15:23:37.798304371Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798336 containerd[1476]: time="2025-01-30T15:23:37.798318171Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798336 containerd[1476]: time="2025-01-30T15:23:37.798330451Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798383 containerd[1476]: time="2025-01-30T15:23:37.798342891Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798383 containerd[1476]: time="2025-01-30T15:23:37.798356531Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798383 containerd[1476]: time="2025-01-30T15:23:37.798371731Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798432 containerd[1476]: time="2025-01-30T15:23:37.798383611Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798432 containerd[1476]: time="2025-01-30T15:23:37.798395811Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798432 containerd[1476]: time="2025-01-30T15:23:37.798408251Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798479 containerd[1476]: time="2025-01-30T15:23:37.798432091Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jan 30 15:23:37.798479 containerd[1476]: time="2025-01-30T15:23:37.798456131Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798479 containerd[1476]: time="2025-01-30T15:23:37.798469451Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.798592 containerd[1476]: time="2025-01-30T15:23:37.798487651Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jan 30 15:23:37.802134 containerd[1476]: time="2025-01-30T15:23:37.800106811Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jan 30 15:23:37.802134 containerd[1476]: time="2025-01-30T15:23:37.800143451Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jan 30 15:23:37.802134 containerd[1476]: time="2025-01-30T15:23:37.800159891Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jan 30 15:23:37.802134 containerd[1476]: time="2025-01-30T15:23:37.800173571Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jan 30 15:23:37.802134 containerd[1476]: time="2025-01-30T15:23:37.800187051Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.802134 containerd[1476]: time="2025-01-30T15:23:37.800211611Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jan 30 15:23:37.802134 containerd[1476]: time="2025-01-30T15:23:37.800225371Z" level=info msg="NRI interface is disabled by configuration." Jan 30 15:23:37.802134 containerd[1476]: time="2025-01-30T15:23:37.800235931Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jan 30 15:23:37.802362 containerd[1476]: time="2025-01-30T15:23:37.801496691Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jan 30 15:23:37.802362 containerd[1476]: time="2025-01-30T15:23:37.801582651Z" level=info msg="Connect containerd service" Jan 30 15:23:37.802362 containerd[1476]: time="2025-01-30T15:23:37.801629131Z" level=info msg="using legacy CRI server" Jan 30 15:23:37.802362 containerd[1476]: time="2025-01-30T15:23:37.801637411Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 30 15:23:37.802362 containerd[1476]: time="2025-01-30T15:23:37.801733971Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jan 30 15:23:37.804023 containerd[1476]: time="2025-01-30T15:23:37.803979531Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 30 15:23:37.804578 containerd[1476]: time="2025-01-30T15:23:37.804241731Z" level=info msg="Start subscribing containerd event" Jan 30 15:23:37.804578 containerd[1476]: time="2025-01-30T15:23:37.804302891Z" level=info msg="Start recovering state" Jan 30 15:23:37.804578 containerd[1476]: time="2025-01-30T15:23:37.804377531Z" level=info msg="Start event monitor" Jan 30 15:23:37.804578 containerd[1476]: time="2025-01-30T15:23:37.804398851Z" level=info msg="Start snapshots syncer" Jan 30 15:23:37.804578 containerd[1476]: time="2025-01-30T15:23:37.804408771Z" level=info msg="Start cni network conf syncer for default" Jan 30 15:23:37.804578 containerd[1476]: time="2025-01-30T15:23:37.804417851Z" level=info msg="Start streaming server" Jan 30 15:23:37.804578 containerd[1476]: time="2025-01-30T15:23:37.804501371Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 30 15:23:37.806153 containerd[1476]: time="2025-01-30T15:23:37.806052611Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 30 15:23:37.806265 systemd[1]: Started containerd.service - containerd container runtime. Jan 30 15:23:37.813118 containerd[1476]: time="2025-01-30T15:23:37.812591611Z" level=info msg="containerd successfully booted in 0.113635s" Jan 30 15:23:37.851758 systemd-networkd[1371]: eth1: Gained IPv6LL Jan 30 15:23:37.858433 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 30 15:23:37.862461 systemd[1]: Reached target network-online.target - Network is Online. Jan 30 15:23:37.871804 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:23:37.880283 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 30 15:23:37.943658 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 30 15:23:38.069405 tar[1475]: linux-arm64/README.md Jan 30 15:23:38.083434 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 30 15:23:38.370183 sshd_keygen[1480]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 30 15:23:38.392764 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 30 15:23:38.401810 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 30 15:23:38.409899 systemd[1]: issuegen.service: Deactivated successfully. Jan 30 15:23:38.410810 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 30 15:23:38.421157 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 30 15:23:38.427758 systemd-networkd[1371]: eth0: Gained IPv6LL Jan 30 15:23:38.433746 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 30 15:23:38.443245 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 30 15:23:38.446435 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Jan 30 15:23:38.447358 systemd[1]: Reached target getty.target - Login Prompts. Jan 30 15:23:38.739794 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:23:38.742599 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 30 15:23:38.744392 (kubelet)[1575]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:23:38.746317 systemd[1]: Startup finished in 800ms (kernel) + 5.788s (initrd) + 4.705s (userspace) = 11.294s. Jan 30 15:23:39.275879 kubelet[1575]: E0130 15:23:39.275828 1575 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:23:39.278623 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:23:39.278823 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:23:49.482849 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 30 15:23:49.492327 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:23:49.625647 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:23:49.641294 (kubelet)[1594]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:23:49.699316 kubelet[1594]: E0130 15:23:49.699253 1594 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:23:49.702226 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:23:49.702489 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:23:59.733051 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 30 15:23:59.749947 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:23:59.866748 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:23:59.872434 (kubelet)[1609]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:23:59.921670 kubelet[1609]: E0130 15:23:59.921591 1609 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:23:59.924481 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:23:59.924631 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:24:09.982674 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Jan 30 15:24:09.994946 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:24:10.107067 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:24:10.118521 (kubelet)[1624]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:24:10.166433 kubelet[1624]: E0130 15:24:10.166364 1624 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:24:10.168592 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:24:10.168755 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:24:20.233159 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Jan 30 15:24:20.242924 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:24:20.362948 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:24:20.364010 (kubelet)[1639]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:24:20.414301 kubelet[1639]: E0130 15:24:20.414207 1639 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:24:20.417823 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:24:20.418039 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:24:22.497084 update_engine[1458]: I20250130 15:24:22.496932 1458 update_attempter.cc:509] Updating boot flags... Jan 30 15:24:22.552593 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1655) Jan 30 15:24:22.604853 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 38 scanned by (udev-worker) (1659) Jan 30 15:24:30.483144 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Jan 30 15:24:30.491822 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:24:30.615759 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:24:30.622522 (kubelet)[1672]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:24:30.669051 kubelet[1672]: E0130 15:24:30.668981 1672 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:24:30.671656 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:24:30.671820 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:24:40.733183 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Jan 30 15:24:40.743981 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:24:40.867105 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:24:40.883394 (kubelet)[1688]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:24:40.927856 kubelet[1688]: E0130 15:24:40.927805 1688 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:24:40.929844 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:24:40.929989 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:24:50.983008 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 7. Jan 30 15:24:50.990870 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:24:51.118589 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:24:51.132205 (kubelet)[1703]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:24:51.176794 kubelet[1703]: E0130 15:24:51.176745 1703 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:24:51.179930 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:24:51.180256 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:25:01.232837 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 8. Jan 30 15:25:01.242893 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:25:01.380922 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:25:01.381100 (kubelet)[1717]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:25:01.432352 kubelet[1717]: E0130 15:25:01.432243 1717 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:25:01.435894 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:25:01.436217 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:25:11.483103 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 9. Jan 30 15:25:11.489851 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:25:11.655251 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:25:11.655396 (kubelet)[1733]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:25:11.699216 kubelet[1733]: E0130 15:25:11.699167 1733 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:25:11.702294 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:25:11.702625 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:25:21.732841 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 10. Jan 30 15:25:21.740963 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:25:21.887814 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:25:21.889915 (kubelet)[1748]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:25:21.933449 kubelet[1748]: E0130 15:25:21.933331 1748 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:25:21.936107 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:25:21.936273 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:25:27.943094 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 30 15:25:27.956985 systemd[1]: Started sshd@0-168.119.58.127:22-139.178.68.195:60694.service - OpenSSH per-connection server daemon (139.178.68.195:60694). Jan 30 15:25:28.954397 sshd[1756]: Accepted publickey for core from 139.178.68.195 port 60694 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:25:28.959114 sshd[1756]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:25:28.970706 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 30 15:25:28.977141 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 30 15:25:28.980735 systemd-logind[1456]: New session 1 of user core. Jan 30 15:25:28.991402 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 30 15:25:29.001412 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 30 15:25:29.006837 (systemd)[1760]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jan 30 15:25:29.110641 systemd[1760]: Queued start job for default target default.target. Jan 30 15:25:29.121619 systemd[1760]: Created slice app.slice - User Application Slice. Jan 30 15:25:29.121661 systemd[1760]: Reached target paths.target - Paths. Jan 30 15:25:29.121679 systemd[1760]: Reached target timers.target - Timers. Jan 30 15:25:29.123470 systemd[1760]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 30 15:25:29.139467 systemd[1760]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 30 15:25:29.139668 systemd[1760]: Reached target sockets.target - Sockets. Jan 30 15:25:29.139688 systemd[1760]: Reached target basic.target - Basic System. Jan 30 15:25:29.139740 systemd[1760]: Reached target default.target - Main User Target. Jan 30 15:25:29.139828 systemd[1760]: Startup finished in 124ms. Jan 30 15:25:29.139976 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 30 15:25:29.147822 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 30 15:25:29.851992 systemd[1]: Started sshd@1-168.119.58.127:22-139.178.68.195:60700.service - OpenSSH per-connection server daemon (139.178.68.195:60700). Jan 30 15:25:30.834154 sshd[1771]: Accepted publickey for core from 139.178.68.195 port 60700 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:25:30.836479 sshd[1771]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:25:30.845449 systemd-logind[1456]: New session 2 of user core. Jan 30 15:25:30.856895 systemd[1]: Started session-2.scope - Session 2 of User core. Jan 30 15:25:31.519833 sshd[1771]: pam_unix(sshd:session): session closed for user core Jan 30 15:25:31.524304 systemd[1]: sshd@1-168.119.58.127:22-139.178.68.195:60700.service: Deactivated successfully. Jan 30 15:25:31.527975 systemd[1]: session-2.scope: Deactivated successfully. Jan 30 15:25:31.529534 systemd-logind[1456]: Session 2 logged out. Waiting for processes to exit. Jan 30 15:25:31.531816 systemd-logind[1456]: Removed session 2. Jan 30 15:25:31.701159 systemd[1]: Started sshd@2-168.119.58.127:22-139.178.68.195:60712.service - OpenSSH per-connection server daemon (139.178.68.195:60712). Jan 30 15:25:31.982626 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 11. Jan 30 15:25:31.993032 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:25:32.123086 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:25:32.136366 (kubelet)[1788]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:25:32.189161 kubelet[1788]: E0130 15:25:32.189110 1788 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:25:32.191892 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:25:32.192065 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:25:32.682118 sshd[1778]: Accepted publickey for core from 139.178.68.195 port 60712 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:25:32.684274 sshd[1778]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:25:32.689658 systemd-logind[1456]: New session 3 of user core. Jan 30 15:25:32.700013 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 30 15:25:33.359975 sshd[1778]: pam_unix(sshd:session): session closed for user core Jan 30 15:25:33.366529 systemd[1]: sshd@2-168.119.58.127:22-139.178.68.195:60712.service: Deactivated successfully. Jan 30 15:25:33.369900 systemd[1]: session-3.scope: Deactivated successfully. Jan 30 15:25:33.371188 systemd-logind[1456]: Session 3 logged out. Waiting for processes to exit. Jan 30 15:25:33.372984 systemd-logind[1456]: Removed session 3. Jan 30 15:25:33.539230 systemd[1]: Started sshd@3-168.119.58.127:22-139.178.68.195:60726.service - OpenSSH per-connection server daemon (139.178.68.195:60726). Jan 30 15:25:34.518245 sshd[1799]: Accepted publickey for core from 139.178.68.195 port 60726 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:25:34.520446 sshd[1799]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:25:34.525724 systemd-logind[1456]: New session 4 of user core. Jan 30 15:25:34.535871 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 30 15:25:35.202900 sshd[1799]: pam_unix(sshd:session): session closed for user core Jan 30 15:25:35.207646 systemd[1]: sshd@3-168.119.58.127:22-139.178.68.195:60726.service: Deactivated successfully. Jan 30 15:25:35.209386 systemd[1]: session-4.scope: Deactivated successfully. Jan 30 15:25:35.212738 systemd-logind[1456]: Session 4 logged out. Waiting for processes to exit. Jan 30 15:25:35.213794 systemd-logind[1456]: Removed session 4. Jan 30 15:25:35.385031 systemd[1]: Started sshd@4-168.119.58.127:22-139.178.68.195:37630.service - OpenSSH per-connection server daemon (139.178.68.195:37630). Jan 30 15:25:36.360595 sshd[1806]: Accepted publickey for core from 139.178.68.195 port 37630 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:25:36.363146 sshd[1806]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:25:36.370941 systemd-logind[1456]: New session 5 of user core. Jan 30 15:25:36.378246 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 30 15:25:36.890203 sudo[1809]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 30 15:25:36.890973 sudo[1809]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 15:25:36.909694 sudo[1809]: pam_unix(sudo:session): session closed for user root Jan 30 15:25:37.071087 sshd[1806]: pam_unix(sshd:session): session closed for user core Jan 30 15:25:37.077333 systemd[1]: sshd@4-168.119.58.127:22-139.178.68.195:37630.service: Deactivated successfully. Jan 30 15:25:37.080060 systemd[1]: session-5.scope: Deactivated successfully. Jan 30 15:25:37.082082 systemd-logind[1456]: Session 5 logged out. Waiting for processes to exit. Jan 30 15:25:37.084438 systemd-logind[1456]: Removed session 5. Jan 30 15:25:37.244940 systemd[1]: Started sshd@5-168.119.58.127:22-139.178.68.195:37638.service - OpenSSH per-connection server daemon (139.178.68.195:37638). Jan 30 15:25:38.212439 sshd[1814]: Accepted publickey for core from 139.178.68.195 port 37638 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:25:38.214825 sshd[1814]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:25:38.221467 systemd-logind[1456]: New session 6 of user core. Jan 30 15:25:38.227948 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 30 15:25:38.729741 sudo[1819]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 30 15:25:38.730056 sudo[1819]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 15:25:38.733830 sudo[1819]: pam_unix(sudo:session): session closed for user root Jan 30 15:25:38.739500 sudo[1818]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Jan 30 15:25:38.739864 sudo[1818]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 15:25:38.757039 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Jan 30 15:25:38.759061 auditctl[1822]: No rules Jan 30 15:25:38.759907 systemd[1]: audit-rules.service: Deactivated successfully. Jan 30 15:25:38.760197 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Jan 30 15:25:38.764722 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jan 30 15:25:38.808233 augenrules[1840]: No rules Jan 30 15:25:38.809952 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jan 30 15:25:38.811837 sudo[1818]: pam_unix(sudo:session): session closed for user root Jan 30 15:25:38.973025 sshd[1814]: pam_unix(sshd:session): session closed for user core Jan 30 15:25:38.977437 systemd[1]: sshd@5-168.119.58.127:22-139.178.68.195:37638.service: Deactivated successfully. Jan 30 15:25:38.979405 systemd[1]: session-6.scope: Deactivated successfully. Jan 30 15:25:38.981898 systemd-logind[1456]: Session 6 logged out. Waiting for processes to exit. Jan 30 15:25:38.984429 systemd-logind[1456]: Removed session 6. Jan 30 15:25:39.154126 systemd[1]: Started sshd@6-168.119.58.127:22-139.178.68.195:37640.service - OpenSSH per-connection server daemon (139.178.68.195:37640). Jan 30 15:25:40.136470 sshd[1848]: Accepted publickey for core from 139.178.68.195 port 37640 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:25:40.138512 sshd[1848]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:25:40.144582 systemd-logind[1456]: New session 7 of user core. Jan 30 15:25:40.154116 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 30 15:25:40.660556 sudo[1851]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 30 15:25:40.660851 sudo[1851]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 15:25:40.986080 (dockerd)[1867]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 30 15:25:40.986699 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 30 15:25:41.253158 dockerd[1867]: time="2025-01-30T15:25:41.252959309Z" level=info msg="Starting up" Jan 30 15:25:41.343419 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2643439460-merged.mount: Deactivated successfully. Jan 30 15:25:41.362030 systemd[1]: var-lib-docker-metacopy\x2dcheck3670137670-merged.mount: Deactivated successfully. Jan 30 15:25:41.375888 dockerd[1867]: time="2025-01-30T15:25:41.375773697Z" level=info msg="Loading containers: start." Jan 30 15:25:41.492676 kernel: Initializing XFRM netlink socket Jan 30 15:25:41.582207 systemd-networkd[1371]: docker0: Link UP Jan 30 15:25:41.612716 dockerd[1867]: time="2025-01-30T15:25:41.612422898Z" level=info msg="Loading containers: done." Jan 30 15:25:41.630530 dockerd[1867]: time="2025-01-30T15:25:41.630442615Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 30 15:25:41.630736 dockerd[1867]: time="2025-01-30T15:25:41.630660147Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0 Jan 30 15:25:41.630872 dockerd[1867]: time="2025-01-30T15:25:41.630831276Z" level=info msg="Daemon has completed initialization" Jan 30 15:25:41.683207 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 30 15:25:41.684109 dockerd[1867]: time="2025-01-30T15:25:41.682812349Z" level=info msg="API listen on /run/docker.sock" Jan 30 15:25:42.233025 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 12. Jan 30 15:25:42.241445 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:25:42.341423 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck1743686419-merged.mount: Deactivated successfully. Jan 30 15:25:42.373794 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:25:42.377433 (kubelet)[2013]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:25:42.426266 kubelet[2013]: E0130 15:25:42.426153 2013 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:25:42.428796 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:25:42.428948 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:25:42.509822 containerd[1476]: time="2025-01-30T15:25:42.509686656Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.1\"" Jan 30 15:25:43.219752 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount836856082.mount: Deactivated successfully. Jan 30 15:25:45.327814 containerd[1476]: time="2025-01-30T15:25:45.327729833Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.32.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:45.330760 containerd[1476]: time="2025-01-30T15:25:45.330213917Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.32.1: active requests=0, bytes read=26221040" Jan 30 15:25:45.331842 containerd[1476]: time="2025-01-30T15:25:45.331763515Z" level=info msg="ImageCreate event name:\"sha256:265c2dedf28ab9b88c7910c1643e210ad62483867f2bab88f56919a6e49a0d19\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:45.337272 containerd[1476]: time="2025-01-30T15:25:45.337157146Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:b88ede8e7c3ce354ca0c45c448c48c094781ce692883ee56f181fa569338c0ac\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:45.339852 containerd[1476]: time="2025-01-30T15:25:45.339625990Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.32.1\" with image id \"sha256:265c2dedf28ab9b88c7910c1643e210ad62483867f2bab88f56919a6e49a0d19\", repo tag \"registry.k8s.io/kube-apiserver:v1.32.1\", repo digest \"registry.k8s.io/kube-apiserver@sha256:b88ede8e7c3ce354ca0c45c448c48c094781ce692883ee56f181fa569338c0ac\", size \"26217748\" in 2.829375663s" Jan 30 15:25:45.340446 containerd[1476]: time="2025-01-30T15:25:45.340122655Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.32.1\" returns image reference \"sha256:265c2dedf28ab9b88c7910c1643e210ad62483867f2bab88f56919a6e49a0d19\"" Jan 30 15:25:45.341236 containerd[1476]: time="2025-01-30T15:25:45.341124705Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.1\"" Jan 30 15:25:47.571007 containerd[1476]: time="2025-01-30T15:25:47.570827710Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.32.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:47.573024 containerd[1476]: time="2025-01-30T15:25:47.572969733Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.32.1: active requests=0, bytes read=22527127" Jan 30 15:25:47.573795 containerd[1476]: time="2025-01-30T15:25:47.573420434Z" level=info msg="ImageCreate event name:\"sha256:2933761aa7adae93679cdde1c0bf457bd4dc4b53f95fc066a4c50aa9c375ea13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:47.580649 containerd[1476]: time="2025-01-30T15:25:47.580169397Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:7e86b2b274365bbc5f5d1e08f0d32d8bb04b8484ac6a92484c298dc695025954\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:47.582505 containerd[1476]: time="2025-01-30T15:25:47.582458787Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.32.1\" with image id \"sha256:2933761aa7adae93679cdde1c0bf457bd4dc4b53f95fc066a4c50aa9c375ea13\", repo tag \"registry.k8s.io/kube-controller-manager:v1.32.1\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:7e86b2b274365bbc5f5d1e08f0d32d8bb04b8484ac6a92484c298dc695025954\", size \"23968433\" in 2.240972384s" Jan 30 15:25:47.582505 containerd[1476]: time="2025-01-30T15:25:47.582506469Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.32.1\" returns image reference \"sha256:2933761aa7adae93679cdde1c0bf457bd4dc4b53f95fc066a4c50aa9c375ea13\"" Jan 30 15:25:47.583149 containerd[1476]: time="2025-01-30T15:25:47.583126539Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.1\"" Jan 30 15:25:49.472940 containerd[1476]: time="2025-01-30T15:25:49.472858496Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.32.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:49.474733 containerd[1476]: time="2025-01-30T15:25:49.474673899Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.32.1: active requests=0, bytes read=17481133" Jan 30 15:25:49.475593 containerd[1476]: time="2025-01-30T15:25:49.475247045Z" level=info msg="ImageCreate event name:\"sha256:ddb38cac617cb18802e09e448db4b3aa70e9e469b02defa76e6de7192847a71c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:49.479975 containerd[1476]: time="2025-01-30T15:25:49.479151583Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:b8fcbcd2afe44acf368b24b61813686f64be4d7fff224d305d78a05bac38f72e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:49.480775 containerd[1476]: time="2025-01-30T15:25:49.480731815Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.32.1\" with image id \"sha256:ddb38cac617cb18802e09e448db4b3aa70e9e469b02defa76e6de7192847a71c\", repo tag \"registry.k8s.io/kube-scheduler:v1.32.1\", repo digest \"registry.k8s.io/kube-scheduler@sha256:b8fcbcd2afe44acf368b24b61813686f64be4d7fff224d305d78a05bac38f72e\", size \"18922457\" in 1.897571715s" Jan 30 15:25:49.480775 containerd[1476]: time="2025-01-30T15:25:49.480771337Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.32.1\" returns image reference \"sha256:ddb38cac617cb18802e09e448db4b3aa70e9e469b02defa76e6de7192847a71c\"" Jan 30 15:25:49.481684 containerd[1476]: time="2025-01-30T15:25:49.481611736Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.1\"" Jan 30 15:25:50.921938 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3428679512.mount: Deactivated successfully. Jan 30 15:25:51.232863 containerd[1476]: time="2025-01-30T15:25:51.232809140Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.32.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:51.233781 containerd[1476]: time="2025-01-30T15:25:51.233737220Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.32.1: active requests=0, bytes read=27364423" Jan 30 15:25:51.235242 containerd[1476]: time="2025-01-30T15:25:51.235164403Z" level=info msg="ImageCreate event name:\"sha256:e124fbed851d756107a6153db4dc52269a2fd34af3cc46f00a2ef113f868aab0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:51.238985 containerd[1476]: time="2025-01-30T15:25:51.238845363Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:0244651801747edf2368222f93a7d17cba6e668a890db72532d6b67a7e06dca5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:51.240310 containerd[1476]: time="2025-01-30T15:25:51.239825166Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.32.1\" with image id \"sha256:e124fbed851d756107a6153db4dc52269a2fd34af3cc46f00a2ef113f868aab0\", repo tag \"registry.k8s.io/kube-proxy:v1.32.1\", repo digest \"registry.k8s.io/kube-proxy@sha256:0244651801747edf2368222f93a7d17cba6e668a890db72532d6b67a7e06dca5\", size \"27363416\" in 1.758034262s" Jan 30 15:25:51.240310 containerd[1476]: time="2025-01-30T15:25:51.239871168Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.32.1\" returns image reference \"sha256:e124fbed851d756107a6153db4dc52269a2fd34af3cc46f00a2ef113f868aab0\"" Jan 30 15:25:51.240731 containerd[1476]: time="2025-01-30T15:25:51.240661563Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\"" Jan 30 15:25:51.862595 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3611638276.mount: Deactivated successfully. Jan 30 15:25:52.482930 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 13. Jan 30 15:25:52.493929 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:25:52.674802 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:25:52.676420 (kubelet)[2146]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:25:52.725234 kubelet[2146]: E0130 15:25:52.725130 2146 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:25:52.728238 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:25:52.728376 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:25:53.093615 containerd[1476]: time="2025-01-30T15:25:53.093367306Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:53.096212 containerd[1476]: time="2025-01-30T15:25:53.095878811Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.3: active requests=0, bytes read=16951714" Jan 30 15:25:53.097744 containerd[1476]: time="2025-01-30T15:25:53.097692606Z" level=info msg="ImageCreate event name:\"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:53.102407 containerd[1476]: time="2025-01-30T15:25:53.102315359Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:53.104460 containerd[1476]: time="2025-01-30T15:25:53.104257920Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.3\" with image id \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.3\", repo digest \"registry.k8s.io/coredns/coredns@sha256:9caabbf6238b189a65d0d6e6ac138de60d6a1c419e5a341fbbb7c78382559c6e\", size \"16948420\" in 1.863342387s" Jan 30 15:25:53.104460 containerd[1476]: time="2025-01-30T15:25:53.104316243Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.3\" returns image reference \"sha256:2f6c962e7b8311337352d9fdea917da2184d9919f4da7695bc2a6517cf392fe4\"" Jan 30 15:25:53.105373 containerd[1476]: time="2025-01-30T15:25:53.105086075Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\"" Jan 30 15:25:53.655509 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1997579862.mount: Deactivated successfully. Jan 30 15:25:53.669858 containerd[1476]: time="2025-01-30T15:25:53.669664246Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:53.671388 containerd[1476]: time="2025-01-30T15:25:53.671096865Z" level=info msg="stop pulling image registry.k8s.io/pause:3.10: active requests=0, bytes read=268723" Jan 30 15:25:53.672571 containerd[1476]: time="2025-01-30T15:25:53.672495884Z" level=info msg="ImageCreate event name:\"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:53.676409 containerd[1476]: time="2025-01-30T15:25:53.676320924Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:53.678703 containerd[1476]: time="2025-01-30T15:25:53.677133997Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.10\" with image id \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\", repo tag \"registry.k8s.io/pause:3.10\", repo digest \"registry.k8s.io/pause@sha256:ee6521f290b2168b6e0935a181d4cff9be1ac3f505666ef0e3c98fae8199917a\", size \"267933\" in 572.006801ms" Jan 30 15:25:53.678703 containerd[1476]: time="2025-01-30T15:25:53.677179439Z" level=info msg="PullImage \"registry.k8s.io/pause:3.10\" returns image reference \"sha256:afb61768ce381961ca0beff95337601f29dc70ff3ed14e5e4b3e5699057e6aa8\"" Jan 30 15:25:53.678903 containerd[1476]: time="2025-01-30T15:25:53.678840749Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\"" Jan 30 15:25:54.317238 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1826033898.mount: Deactivated successfully. Jan 30 15:25:57.429147 containerd[1476]: time="2025-01-30T15:25:57.429086032Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.16-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:57.430813 containerd[1476]: time="2025-01-30T15:25:57.430753056Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.16-0: active requests=0, bytes read=67812491" Jan 30 15:25:57.431342 containerd[1476]: time="2025-01-30T15:25:57.431292997Z" level=info msg="ImageCreate event name:\"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:57.435617 containerd[1476]: time="2025-01-30T15:25:57.435549960Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:25:57.437038 containerd[1476]: time="2025-01-30T15:25:57.436999255Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.16-0\" with image id \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\", repo tag \"registry.k8s.io/etcd:3.5.16-0\", repo digest \"registry.k8s.io/etcd@sha256:c6a9d11cc5c04b114ccdef39a9265eeef818e3d02f5359be035ae784097fdec5\", size \"67941650\" in 3.758119225s" Jan 30 15:25:57.437256 containerd[1476]: time="2025-01-30T15:25:57.437149981Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.16-0\" returns image reference \"sha256:7fc9d4aa817aa6a3e549f3cd49d1f7b496407be979fc36dd5f356d59ce8c3a82\"" Jan 30 15:26:02.732459 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 14. Jan 30 15:26:02.741816 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:26:02.866684 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:26:02.878217 (kubelet)[2242]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 15:26:02.922570 kubelet[2242]: E0130 15:26:02.921905 2242 run.go:72] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 15:26:02.924819 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 15:26:02.925362 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 15:26:04.149227 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:26:04.156900 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:26:04.191779 systemd[1]: Reloading requested from client PID 2257 ('systemctl') (unit session-7.scope)... Jan 30 15:26:04.191948 systemd[1]: Reloading... Jan 30 15:26:04.307159 zram_generator::config[2300]: No configuration found. Jan 30 15:26:04.412265 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 15:26:04.483338 systemd[1]: Reloading finished in 290 ms. Jan 30 15:26:04.538670 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 30 15:26:04.538770 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 30 15:26:04.539200 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:26:04.546080 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:26:04.680920 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:26:04.681004 (kubelet)[2345]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 30 15:26:04.726589 kubelet[2345]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 15:26:04.726589 kubelet[2345]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 30 15:26:04.726589 kubelet[2345]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 15:26:04.726589 kubelet[2345]: I0130 15:26:04.725864 2345 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 30 15:26:05.425582 kubelet[2345]: I0130 15:26:05.424926 2345 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Jan 30 15:26:05.425582 kubelet[2345]: I0130 15:26:05.424964 2345 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 30 15:26:05.425582 kubelet[2345]: I0130 15:26:05.425291 2345 server.go:954] "Client rotation is on, will bootstrap in background" Jan 30 15:26:05.454399 kubelet[2345]: E0130 15:26:05.454314 2345 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://168.119.58.127:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 168.119.58.127:6443: connect: connection refused" logger="UnhandledError" Jan 30 15:26:05.460292 kubelet[2345]: I0130 15:26:05.459939 2345 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 30 15:26:05.466256 kubelet[2345]: E0130 15:26:05.466214 2345 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jan 30 15:26:05.466913 kubelet[2345]: I0130 15:26:05.466423 2345 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jan 30 15:26:05.468975 kubelet[2345]: I0130 15:26:05.468949 2345 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 30 15:26:05.470863 kubelet[2345]: I0130 15:26:05.470792 2345 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 30 15:26:05.471055 kubelet[2345]: I0130 15:26:05.470856 2345 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-0-a-5514912dcd","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 30 15:26:05.471157 kubelet[2345]: I0130 15:26:05.471126 2345 topology_manager.go:138] "Creating topology manager with none policy" Jan 30 15:26:05.471157 kubelet[2345]: I0130 15:26:05.471138 2345 container_manager_linux.go:304] "Creating device plugin manager" Jan 30 15:26:05.471378 kubelet[2345]: I0130 15:26:05.471346 2345 state_mem.go:36] "Initialized new in-memory state store" Jan 30 15:26:05.474989 kubelet[2345]: I0130 15:26:05.474769 2345 kubelet.go:446] "Attempting to sync node with API server" Jan 30 15:26:05.474989 kubelet[2345]: I0130 15:26:05.474804 2345 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 30 15:26:05.474989 kubelet[2345]: I0130 15:26:05.474832 2345 kubelet.go:352] "Adding apiserver pod source" Jan 30 15:26:05.474989 kubelet[2345]: I0130 15:26:05.474846 2345 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 30 15:26:05.480587 kubelet[2345]: W0130 15:26:05.479977 2345 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://168.119.58.127:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-0-a-5514912dcd&limit=500&resourceVersion=0": dial tcp 168.119.58.127:6443: connect: connection refused Jan 30 15:26:05.480587 kubelet[2345]: E0130 15:26:05.480065 2345 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://168.119.58.127:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-0-a-5514912dcd&limit=500&resourceVersion=0\": dial tcp 168.119.58.127:6443: connect: connection refused" logger="UnhandledError" Jan 30 15:26:05.480587 kubelet[2345]: I0130 15:26:05.480246 2345 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jan 30 15:26:05.481263 kubelet[2345]: I0130 15:26:05.481229 2345 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 30 15:26:05.481454 kubelet[2345]: W0130 15:26:05.481442 2345 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 30 15:26:05.483841 kubelet[2345]: I0130 15:26:05.483811 2345 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 30 15:26:05.484069 kubelet[2345]: I0130 15:26:05.484018 2345 server.go:1287] "Started kubelet" Jan 30 15:26:05.490444 kubelet[2345]: W0130 15:26:05.490353 2345 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://168.119.58.127:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 168.119.58.127:6443: connect: connection refused Jan 30 15:26:05.490444 kubelet[2345]: E0130 15:26:05.490412 2345 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://168.119.58.127:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 168.119.58.127:6443: connect: connection refused" logger="UnhandledError" Jan 30 15:26:05.492620 kubelet[2345]: E0130 15:26:05.490469 2345 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://168.119.58.127:6443/api/v1/namespaces/default/events\": dial tcp 168.119.58.127:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4081-3-0-a-5514912dcd.181f81dc8f3967e6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4081-3-0-a-5514912dcd,UID:ci-4081-3-0-a-5514912dcd,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4081-3-0-a-5514912dcd,},FirstTimestamp:2025-01-30 15:26:05.483993062 +0000 UTC m=+0.797190238,LastTimestamp:2025-01-30 15:26:05.483993062 +0000 UTC m=+0.797190238,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-0-a-5514912dcd,}" Jan 30 15:26:05.492620 kubelet[2345]: I0130 15:26:05.490761 2345 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Jan 30 15:26:05.493826 kubelet[2345]: I0130 15:26:05.493753 2345 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 30 15:26:05.494047 kubelet[2345]: I0130 15:26:05.494016 2345 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 30 15:26:05.494386 kubelet[2345]: I0130 15:26:05.494365 2345 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 30 15:26:05.496876 kubelet[2345]: I0130 15:26:05.496847 2345 server.go:490] "Adding debug handlers to kubelet server" Jan 30 15:26:05.497585 kubelet[2345]: I0130 15:26:05.497290 2345 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 30 15:26:05.497585 kubelet[2345]: I0130 15:26:05.497410 2345 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 30 15:26:05.497742 kubelet[2345]: E0130 15:26:05.497723 2345 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4081-3-0-a-5514912dcd\" not found" Jan 30 15:26:05.502008 kubelet[2345]: I0130 15:26:05.501963 2345 factory.go:221] Registration of the systemd container factory successfully Jan 30 15:26:05.502135 kubelet[2345]: I0130 15:26:05.502098 2345 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 30 15:26:05.502712 kubelet[2345]: E0130 15:26:05.502668 2345 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://168.119.58.127:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-0-a-5514912dcd?timeout=10s\": dial tcp 168.119.58.127:6443: connect: connection refused" interval="200ms" Jan 30 15:26:05.503443 kubelet[2345]: I0130 15:26:05.503181 2345 reconciler.go:26] "Reconciler: start to sync state" Jan 30 15:26:05.503443 kubelet[2345]: I0130 15:26:05.503240 2345 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 30 15:26:05.504305 kubelet[2345]: W0130 15:26:05.504261 2345 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://168.119.58.127:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 168.119.58.127:6443: connect: connection refused Jan 30 15:26:05.504412 kubelet[2345]: E0130 15:26:05.504393 2345 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://168.119.58.127:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 168.119.58.127:6443: connect: connection refused" logger="UnhandledError" Jan 30 15:26:05.506865 kubelet[2345]: I0130 15:26:05.506322 2345 factory.go:221] Registration of the containerd container factory successfully Jan 30 15:26:05.508458 kubelet[2345]: E0130 15:26:05.508254 2345 kubelet.go:1561] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 30 15:26:05.519503 kubelet[2345]: I0130 15:26:05.519463 2345 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 30 15:26:05.522746 kubelet[2345]: I0130 15:26:05.522696 2345 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 30 15:26:05.523501 kubelet[2345]: I0130 15:26:05.522885 2345 status_manager.go:227] "Starting to sync pod status with apiserver" Jan 30 15:26:05.523501 kubelet[2345]: I0130 15:26:05.522964 2345 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 30 15:26:05.523501 kubelet[2345]: I0130 15:26:05.522981 2345 kubelet.go:2388] "Starting kubelet main sync loop" Jan 30 15:26:05.523501 kubelet[2345]: E0130 15:26:05.523043 2345 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 30 15:26:05.531590 kubelet[2345]: W0130 15:26:05.531488 2345 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://168.119.58.127:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 168.119.58.127:6443: connect: connection refused Jan 30 15:26:05.531730 kubelet[2345]: E0130 15:26:05.531610 2345 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://168.119.58.127:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 168.119.58.127:6443: connect: connection refused" logger="UnhandledError" Jan 30 15:26:05.537429 kubelet[2345]: I0130 15:26:05.537401 2345 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 30 15:26:05.537429 kubelet[2345]: I0130 15:26:05.537429 2345 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 30 15:26:05.537723 kubelet[2345]: I0130 15:26:05.537459 2345 state_mem.go:36] "Initialized new in-memory state store" Jan 30 15:26:05.539927 kubelet[2345]: I0130 15:26:05.539896 2345 policy_none.go:49] "None policy: Start" Jan 30 15:26:05.539927 kubelet[2345]: I0130 15:26:05.539927 2345 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 30 15:26:05.540078 kubelet[2345]: I0130 15:26:05.539939 2345 state_mem.go:35] "Initializing new in-memory state store" Jan 30 15:26:05.547434 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 30 15:26:05.558191 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 30 15:26:05.572174 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 30 15:26:05.573909 kubelet[2345]: I0130 15:26:05.573865 2345 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 30 15:26:05.574178 kubelet[2345]: I0130 15:26:05.574095 2345 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 30 15:26:05.574178 kubelet[2345]: I0130 15:26:05.574116 2345 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 30 15:26:05.574891 kubelet[2345]: I0130 15:26:05.574789 2345 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 30 15:26:05.577922 kubelet[2345]: E0130 15:26:05.577829 2345 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 30 15:26:05.577922 kubelet[2345]: E0130 15:26:05.577888 2345 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4081-3-0-a-5514912dcd\" not found" Jan 30 15:26:05.636459 systemd[1]: Created slice kubepods-burstable-pod971326e260333f76ace84ca47537837d.slice - libcontainer container kubepods-burstable-pod971326e260333f76ace84ca47537837d.slice. Jan 30 15:26:05.658763 kubelet[2345]: E0130 15:26:05.658462 2345 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-0-a-5514912dcd\" not found" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.664840 systemd[1]: Created slice kubepods-burstable-podd6083b14b485d95141f40cab4b058be7.slice - libcontainer container kubepods-burstable-podd6083b14b485d95141f40cab4b058be7.slice. Jan 30 15:26:05.674377 kubelet[2345]: E0130 15:26:05.674312 2345 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-0-a-5514912dcd\" not found" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.678280 kubelet[2345]: I0130 15:26:05.677776 2345 kubelet_node_status.go:76] "Attempting to register node" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.679589 kubelet[2345]: E0130 15:26:05.678703 2345 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://168.119.58.127:6443/api/v1/nodes\": dial tcp 168.119.58.127:6443: connect: connection refused" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.684059 systemd[1]: Created slice kubepods-burstable-poda6a059240d7e803477388b9e0b38df61.slice - libcontainer container kubepods-burstable-poda6a059240d7e803477388b9e0b38df61.slice. Jan 30 15:26:05.686420 kubelet[2345]: E0130 15:26:05.686361 2345 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-0-a-5514912dcd\" not found" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.703469 kubelet[2345]: E0130 15:26:05.703387 2345 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://168.119.58.127:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-0-a-5514912dcd?timeout=10s\": dial tcp 168.119.58.127:6443: connect: connection refused" interval="400ms" Jan 30 15:26:05.705081 kubelet[2345]: I0130 15:26:05.704758 2345 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/971326e260333f76ace84ca47537837d-ca-certs\") pod \"kube-apiserver-ci-4081-3-0-a-5514912dcd\" (UID: \"971326e260333f76ace84ca47537837d\") " pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.705081 kubelet[2345]: I0130 15:26:05.704817 2345 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/971326e260333f76ace84ca47537837d-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-0-a-5514912dcd\" (UID: \"971326e260333f76ace84ca47537837d\") " pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.705081 kubelet[2345]: I0130 15:26:05.704919 2345 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.705081 kubelet[2345]: I0130 15:26:05.704999 2345 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.705081 kubelet[2345]: I0130 15:26:05.705039 2345 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a6a059240d7e803477388b9e0b38df61-kubeconfig\") pod \"kube-scheduler-ci-4081-3-0-a-5514912dcd\" (UID: \"a6a059240d7e803477388b9e0b38df61\") " pod="kube-system/kube-scheduler-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.705496 kubelet[2345]: I0130 15:26:05.705081 2345 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/971326e260333f76ace84ca47537837d-k8s-certs\") pod \"kube-apiserver-ci-4081-3-0-a-5514912dcd\" (UID: \"971326e260333f76ace84ca47537837d\") " pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.705496 kubelet[2345]: I0130 15:26:05.705141 2345 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-ca-certs\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.705496 kubelet[2345]: I0130 15:26:05.705178 2345 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.705496 kubelet[2345]: I0130 15:26:05.705210 2345 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.882051 kubelet[2345]: I0130 15:26:05.881986 2345 kubelet_node_status.go:76] "Attempting to register node" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.882452 kubelet[2345]: E0130 15:26:05.882394 2345 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://168.119.58.127:6443/api/v1/nodes\": dial tcp 168.119.58.127:6443: connect: connection refused" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:05.959945 containerd[1476]: time="2025-01-30T15:26:05.959786369Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-0-a-5514912dcd,Uid:971326e260333f76ace84ca47537837d,Namespace:kube-system,Attempt:0,}" Jan 30 15:26:05.975936 containerd[1476]: time="2025-01-30T15:26:05.975859451Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-0-a-5514912dcd,Uid:d6083b14b485d95141f40cab4b058be7,Namespace:kube-system,Attempt:0,}" Jan 30 15:26:05.988890 containerd[1476]: time="2025-01-30T15:26:05.988332777Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-0-a-5514912dcd,Uid:a6a059240d7e803477388b9e0b38df61,Namespace:kube-system,Attempt:0,}" Jan 30 15:26:06.104732 kubelet[2345]: E0130 15:26:06.104687 2345 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://168.119.58.127:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-0-a-5514912dcd?timeout=10s\": dial tcp 168.119.58.127:6443: connect: connection refused" interval="800ms" Jan 30 15:26:06.285789 kubelet[2345]: I0130 15:26:06.285731 2345 kubelet_node_status.go:76] "Attempting to register node" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:06.286320 kubelet[2345]: E0130 15:26:06.286281 2345 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://168.119.58.127:6443/api/v1/nodes\": dial tcp 168.119.58.127:6443: connect: connection refused" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:06.416383 kubelet[2345]: W0130 15:26:06.416337 2345 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://168.119.58.127:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 168.119.58.127:6443: connect: connection refused Jan 30 15:26:06.416607 kubelet[2345]: E0130 15:26:06.416417 2345 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://168.119.58.127:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 168.119.58.127:6443: connect: connection refused" logger="UnhandledError" Jan 30 15:26:06.507981 kubelet[2345]: W0130 15:26:06.507896 2345 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://168.119.58.127:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 168.119.58.127:6443: connect: connection refused Jan 30 15:26:06.507981 kubelet[2345]: E0130 15:26:06.507977 2345 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://168.119.58.127:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 168.119.58.127:6443: connect: connection refused" logger="UnhandledError" Jan 30 15:26:06.539096 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2849586117.mount: Deactivated successfully. Jan 30 15:26:06.550617 containerd[1476]: time="2025-01-30T15:26:06.549655483Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 15:26:06.553359 containerd[1476]: time="2025-01-30T15:26:06.553268798Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269193" Jan 30 15:26:06.554083 containerd[1476]: time="2025-01-30T15:26:06.554036102Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 15:26:06.555646 containerd[1476]: time="2025-01-30T15:26:06.555595472Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 30 15:26:06.557248 containerd[1476]: time="2025-01-30T15:26:06.557058999Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 15:26:06.559672 containerd[1476]: time="2025-01-30T15:26:06.559306710Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 15:26:06.559672 containerd[1476]: time="2025-01-30T15:26:06.559606360Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 30 15:26:06.561818 containerd[1476]: time="2025-01-30T15:26:06.561746548Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 15:26:06.566315 containerd[1476]: time="2025-01-30T15:26:06.565933082Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 589.938946ms" Jan 30 15:26:06.567884 containerd[1476]: time="2025-01-30T15:26:06.567805901Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 579.3404ms" Jan 30 15:26:06.568092 containerd[1476]: time="2025-01-30T15:26:06.567979947Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 608.074335ms" Jan 30 15:26:06.691633 containerd[1476]: time="2025-01-30T15:26:06.690705060Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 15:26:06.691633 containerd[1476]: time="2025-01-30T15:26:06.690790223Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 15:26:06.691633 containerd[1476]: time="2025-01-30T15:26:06.690801783Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:06.691633 containerd[1476]: time="2025-01-30T15:26:06.690906786Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:06.694664 containerd[1476]: time="2025-01-30T15:26:06.694174131Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 15:26:06.694664 containerd[1476]: time="2025-01-30T15:26:06.694239453Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 15:26:06.694664 containerd[1476]: time="2025-01-30T15:26:06.694276614Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:06.694664 containerd[1476]: time="2025-01-30T15:26:06.694584184Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:06.700019 containerd[1476]: time="2025-01-30T15:26:06.699899953Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 15:26:06.702352 containerd[1476]: time="2025-01-30T15:26:06.700609056Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 15:26:06.702352 containerd[1476]: time="2025-01-30T15:26:06.700636697Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:06.702352 containerd[1476]: time="2025-01-30T15:26:06.700767701Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:06.722784 systemd[1]: Started cri-containerd-26e88f4f2edce8f3fb3212c689825fa1f0c2f9b6b3bb52869fee6e6210b06656.scope - libcontainer container 26e88f4f2edce8f3fb3212c689825fa1f0c2f9b6b3bb52869fee6e6210b06656. Jan 30 15:26:06.737846 systemd[1]: Started cri-containerd-443706cb51f09c76a59b3d00bae25e9f629293fb13424ab2cdbf8798fb8fc8e8.scope - libcontainer container 443706cb51f09c76a59b3d00bae25e9f629293fb13424ab2cdbf8798fb8fc8e8. Jan 30 15:26:06.743707 systemd[1]: Started cri-containerd-5e530141da586405be65973f8518cd3c9225836d39bcd3d7f238c4b2b2a4b654.scope - libcontainer container 5e530141da586405be65973f8518cd3c9225836d39bcd3d7f238c4b2b2a4b654. Jan 30 15:26:06.807591 containerd[1476]: time="2025-01-30T15:26:06.807444582Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4081-3-0-a-5514912dcd,Uid:971326e260333f76ace84ca47537837d,Namespace:kube-system,Attempt:0,} returns sandbox id \"26e88f4f2edce8f3fb3212c689825fa1f0c2f9b6b3bb52869fee6e6210b06656\"" Jan 30 15:26:06.811297 containerd[1476]: time="2025-01-30T15:26:06.811248223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4081-3-0-a-5514912dcd,Uid:d6083b14b485d95141f40cab4b058be7,Namespace:kube-system,Attempt:0,} returns sandbox id \"443706cb51f09c76a59b3d00bae25e9f629293fb13424ab2cdbf8798fb8fc8e8\"" Jan 30 15:26:06.816447 containerd[1476]: time="2025-01-30T15:26:06.815715966Z" level=info msg="CreateContainer within sandbox \"26e88f4f2edce8f3fb3212c689825fa1f0c2f9b6b3bb52869fee6e6210b06656\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 30 15:26:06.816893 containerd[1476]: time="2025-01-30T15:26:06.816781960Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4081-3-0-a-5514912dcd,Uid:a6a059240d7e803477388b9e0b38df61,Namespace:kube-system,Attempt:0,} returns sandbox id \"5e530141da586405be65973f8518cd3c9225836d39bcd3d7f238c4b2b2a4b654\"" Jan 30 15:26:06.820661 containerd[1476]: time="2025-01-30T15:26:06.820602202Z" level=info msg="CreateContainer within sandbox \"443706cb51f09c76a59b3d00bae25e9f629293fb13424ab2cdbf8798fb8fc8e8\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 30 15:26:06.824024 containerd[1476]: time="2025-01-30T15:26:06.823973709Z" level=info msg="CreateContainer within sandbox \"5e530141da586405be65973f8518cd3c9225836d39bcd3d7f238c4b2b2a4b654\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 30 15:26:06.833415 containerd[1476]: time="2025-01-30T15:26:06.833179043Z" level=info msg="CreateContainer within sandbox \"26e88f4f2edce8f3fb3212c689825fa1f0c2f9b6b3bb52869fee6e6210b06656\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"3e826fb81fa7137b7e27eff46eac1b4c10d81de0d6143f9347ee4aa2a2185eb3\"" Jan 30 15:26:06.836589 containerd[1476]: time="2025-01-30T15:26:06.834762453Z" level=info msg="StartContainer for \"3e826fb81fa7137b7e27eff46eac1b4c10d81de0d6143f9347ee4aa2a2185eb3\"" Jan 30 15:26:06.848089 containerd[1476]: time="2025-01-30T15:26:06.848012036Z" level=info msg="CreateContainer within sandbox \"443706cb51f09c76a59b3d00bae25e9f629293fb13424ab2cdbf8798fb8fc8e8\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4\"" Jan 30 15:26:06.848873 containerd[1476]: time="2025-01-30T15:26:06.848840302Z" level=info msg="StartContainer for \"9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4\"" Jan 30 15:26:06.855775 containerd[1476]: time="2025-01-30T15:26:06.855586477Z" level=info msg="CreateContainer within sandbox \"5e530141da586405be65973f8518cd3c9225836d39bcd3d7f238c4b2b2a4b654\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838\"" Jan 30 15:26:06.856617 containerd[1476]: time="2025-01-30T15:26:06.856587949Z" level=info msg="StartContainer for \"165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838\"" Jan 30 15:26:06.882790 systemd[1]: Started cri-containerd-3e826fb81fa7137b7e27eff46eac1b4c10d81de0d6143f9347ee4aa2a2185eb3.scope - libcontainer container 3e826fb81fa7137b7e27eff46eac1b4c10d81de0d6143f9347ee4aa2a2185eb3. Jan 30 15:26:06.898826 systemd[1]: Started cri-containerd-165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838.scope - libcontainer container 165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838. Jan 30 15:26:06.905949 kubelet[2345]: E0130 15:26:06.905789 2345 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://168.119.58.127:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4081-3-0-a-5514912dcd?timeout=10s\": dial tcp 168.119.58.127:6443: connect: connection refused" interval="1.6s" Jan 30 15:26:06.913313 systemd[1]: Started cri-containerd-9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4.scope - libcontainer container 9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4. Jan 30 15:26:06.957116 containerd[1476]: time="2025-01-30T15:26:06.956087242Z" level=info msg="StartContainer for \"3e826fb81fa7137b7e27eff46eac1b4c10d81de0d6143f9347ee4aa2a2185eb3\" returns successfully" Jan 30 15:26:06.964257 kubelet[2345]: W0130 15:26:06.964051 2345 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://168.119.58.127:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 168.119.58.127:6443: connect: connection refused Jan 30 15:26:06.964257 kubelet[2345]: E0130 15:26:06.964145 2345 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://168.119.58.127:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 168.119.58.127:6443: connect: connection refused" logger="UnhandledError" Jan 30 15:26:06.988952 containerd[1476]: time="2025-01-30T15:26:06.988898608Z" level=info msg="StartContainer for \"9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4\" returns successfully" Jan 30 15:26:07.000786 kubelet[2345]: W0130 15:26:07.000644 2345 reflector.go:569] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://168.119.58.127:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-0-a-5514912dcd&limit=500&resourceVersion=0": dial tcp 168.119.58.127:6443: connect: connection refused Jan 30 15:26:07.000786 kubelet[2345]: E0130 15:26:07.000726 2345 reflector.go:166] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://168.119.58.127:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4081-3-0-a-5514912dcd&limit=500&resourceVersion=0\": dial tcp 168.119.58.127:6443: connect: connection refused" logger="UnhandledError" Jan 30 15:26:07.002093 containerd[1476]: time="2025-01-30T15:26:07.001973464Z" level=info msg="StartContainer for \"165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838\" returns successfully" Jan 30 15:26:07.089490 kubelet[2345]: I0130 15:26:07.088921 2345 kubelet_node_status.go:76] "Attempting to register node" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:07.089490 kubelet[2345]: E0130 15:26:07.089344 2345 kubelet_node_status.go:108] "Unable to register node with API server" err="Post \"https://168.119.58.127:6443/api/v1/nodes\": dial tcp 168.119.58.127:6443: connect: connection refused" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:07.549568 kubelet[2345]: E0130 15:26:07.547196 2345 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-0-a-5514912dcd\" not found" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:07.553113 kubelet[2345]: E0130 15:26:07.553082 2345 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-0-a-5514912dcd\" not found" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:07.554122 kubelet[2345]: E0130 15:26:07.553790 2345 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-0-a-5514912dcd\" not found" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:08.557566 kubelet[2345]: E0130 15:26:08.557251 2345 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-0-a-5514912dcd\" not found" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:08.559829 kubelet[2345]: E0130 15:26:08.559672 2345 kubelet.go:3196] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"ci-4081-3-0-a-5514912dcd\" not found" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:08.693432 kubelet[2345]: I0130 15:26:08.693267 2345 kubelet_node_status.go:76] "Attempting to register node" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:09.863679 kubelet[2345]: E0130 15:26:09.863632 2345 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-4081-3-0-a-5514912dcd\" not found" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:09.948320 kubelet[2345]: I0130 15:26:09.948028 2345 kubelet_node_status.go:79] "Successfully registered node" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:09.948320 kubelet[2345]: E0130 15:26:09.948080 2345 kubelet_node_status.go:549] "Error updating node status, will retry" err="error getting node \"ci-4081-3-0-a-5514912dcd\": node \"ci-4081-3-0-a-5514912dcd\" not found" Jan 30 15:26:09.959793 kubelet[2345]: E0130 15:26:09.959753 2345 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4081-3-0-a-5514912dcd\" not found" Jan 30 15:26:10.099353 kubelet[2345]: I0130 15:26:10.099020 2345 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:10.118550 kubelet[2345]: E0130 15:26:10.117904 2345 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-0-a-5514912dcd\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:10.118550 kubelet[2345]: I0130 15:26:10.117940 2345 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:10.123593 kubelet[2345]: E0130 15:26:10.121666 2345 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:10.123593 kubelet[2345]: I0130 15:26:10.121700 2345 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:10.126720 kubelet[2345]: E0130 15:26:10.126676 2345 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-0-a-5514912dcd\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:10.305211 kubelet[2345]: I0130 15:26:10.303360 2345 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:10.305928 kubelet[2345]: E0130 15:26:10.305740 2345 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-scheduler-ci-4081-3-0-a-5514912dcd\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:10.493338 kubelet[2345]: I0130 15:26:10.492946 2345 apiserver.go:52] "Watching apiserver" Jan 30 15:26:10.504295 kubelet[2345]: I0130 15:26:10.504255 2345 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Jan 30 15:26:12.145968 systemd[1]: Reloading requested from client PID 2619 ('systemctl') (unit session-7.scope)... Jan 30 15:26:12.145992 systemd[1]: Reloading... Jan 30 15:26:12.276767 zram_generator::config[2665]: No configuration found. Jan 30 15:26:12.396669 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 15:26:12.483839 systemd[1]: Reloading finished in 337 ms. Jan 30 15:26:12.532413 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:26:12.546976 systemd[1]: kubelet.service: Deactivated successfully. Jan 30 15:26:12.547344 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:26:12.547432 systemd[1]: kubelet.service: Consumed 1.239s CPU time, 126.8M memory peak, 0B memory swap peak. Jan 30 15:26:12.551959 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 15:26:12.687763 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 15:26:12.702085 (kubelet)[2704]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 30 15:26:12.763335 kubelet[2704]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 15:26:12.763335 kubelet[2704]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Jan 30 15:26:12.763335 kubelet[2704]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 15:26:12.763779 kubelet[2704]: I0130 15:26:12.763453 2704 server.go:215] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 30 15:26:12.774931 kubelet[2704]: I0130 15:26:12.774797 2704 server.go:520] "Kubelet version" kubeletVersion="v1.32.0" Jan 30 15:26:12.774931 kubelet[2704]: I0130 15:26:12.774830 2704 server.go:522] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 30 15:26:12.776184 kubelet[2704]: I0130 15:26:12.775583 2704 server.go:954] "Client rotation is on, will bootstrap in background" Jan 30 15:26:12.777407 kubelet[2704]: I0130 15:26:12.777375 2704 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 30 15:26:12.782893 kubelet[2704]: I0130 15:26:12.782344 2704 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 30 15:26:12.787851 kubelet[2704]: E0130 15:26:12.787744 2704 log.go:32] "RuntimeConfig from runtime service failed" err="rpc error: code = Unimplemented desc = unknown method RuntimeConfig for service runtime.v1.RuntimeService" Jan 30 15:26:12.788172 kubelet[2704]: I0130 15:26:12.788140 2704 server.go:1421] "CRI implementation should be updated to support RuntimeConfig when KubeletCgroupDriverFromCRI feature gate has been enabled. Falling back to using cgroupDriver from kubelet config." Jan 30 15:26:12.791572 kubelet[2704]: I0130 15:26:12.791509 2704 server.go:772] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 30 15:26:12.791839 kubelet[2704]: I0130 15:26:12.791772 2704 container_manager_linux.go:268] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 30 15:26:12.792190 kubelet[2704]: I0130 15:26:12.791814 2704 container_manager_linux.go:273] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4081-3-0-a-5514912dcd","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Jan 30 15:26:12.792190 kubelet[2704]: I0130 15:26:12.792177 2704 topology_manager.go:138] "Creating topology manager with none policy" Jan 30 15:26:12.792190 kubelet[2704]: I0130 15:26:12.792190 2704 container_manager_linux.go:304] "Creating device plugin manager" Jan 30 15:26:12.792347 kubelet[2704]: I0130 15:26:12.792250 2704 state_mem.go:36] "Initialized new in-memory state store" Jan 30 15:26:12.792448 kubelet[2704]: I0130 15:26:12.792426 2704 kubelet.go:446] "Attempting to sync node with API server" Jan 30 15:26:12.792448 kubelet[2704]: I0130 15:26:12.792448 2704 kubelet.go:341] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 30 15:26:12.792508 kubelet[2704]: I0130 15:26:12.792468 2704 kubelet.go:352] "Adding apiserver pod source" Jan 30 15:26:12.792508 kubelet[2704]: I0130 15:26:12.792485 2704 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 30 15:26:12.793966 kubelet[2704]: I0130 15:26:12.793704 2704 kuberuntime_manager.go:269] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1" Jan 30 15:26:12.794666 kubelet[2704]: I0130 15:26:12.794649 2704 kubelet.go:890] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 30 15:26:12.796166 kubelet[2704]: I0130 15:26:12.796091 2704 watchdog_linux.go:99] "Systemd watchdog is not enabled" Jan 30 15:26:12.796166 kubelet[2704]: I0130 15:26:12.796131 2704 server.go:1287] "Started kubelet" Jan 30 15:26:12.800433 kubelet[2704]: I0130 15:26:12.800401 2704 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 30 15:26:12.803634 kubelet[2704]: I0130 15:26:12.803591 2704 server.go:169] "Starting to listen" address="0.0.0.0" port=10250 Jan 30 15:26:12.815347 kubelet[2704]: I0130 15:26:12.814922 2704 server.go:490] "Adding debug handlers to kubelet server" Jan 30 15:26:12.816040 kubelet[2704]: I0130 15:26:12.804008 2704 dynamic_serving_content.go:135] "Starting controller" name="kubelet-server-cert-files::/var/lib/kubelet/pki/kubelet.crt::/var/lib/kubelet/pki/kubelet.key" Jan 30 15:26:12.819345 kubelet[2704]: I0130 15:26:12.819107 2704 factory.go:221] Registration of the systemd container factory successfully Jan 30 15:26:12.821388 kubelet[2704]: I0130 15:26:12.820530 2704 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 30 15:26:12.821524 kubelet[2704]: I0130 15:26:12.821473 2704 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 30 15:26:12.822983 kubelet[2704]: I0130 15:26:12.822926 2704 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 30 15:26:12.822983 kubelet[2704]: I0130 15:26:12.822973 2704 status_manager.go:227] "Starting to sync pod status with apiserver" Jan 30 15:26:12.823130 kubelet[2704]: I0130 15:26:12.822995 2704 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Jan 30 15:26:12.823130 kubelet[2704]: I0130 15:26:12.823002 2704 kubelet.go:2388] "Starting kubelet main sync loop" Jan 30 15:26:12.823130 kubelet[2704]: E0130 15:26:12.823075 2704 kubelet.go:2412] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 30 15:26:12.823208 kubelet[2704]: I0130 15:26:12.803670 2704 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 30 15:26:12.823557 kubelet[2704]: I0130 15:26:12.823445 2704 server.go:243] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 30 15:26:12.824582 kubelet[2704]: I0130 15:26:12.805892 2704 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 30 15:26:12.824746 kubelet[2704]: E0130 15:26:12.806057 2704 kubelet_node_status.go:467] "Error getting the current node from lister" err="node \"ci-4081-3-0-a-5514912dcd\" not found" Jan 30 15:26:12.829225 kubelet[2704]: I0130 15:26:12.805871 2704 volume_manager.go:297] "Starting Kubelet Volume Manager" Jan 30 15:26:12.832740 kubelet[2704]: I0130 15:26:12.831505 2704 reconciler.go:26] "Reconciler: start to sync state" Jan 30 15:26:12.834544 kubelet[2704]: I0130 15:26:12.832967 2704 factory.go:221] Registration of the containerd container factory successfully Jan 30 15:26:12.914956 kubelet[2704]: I0130 15:26:12.914923 2704 cpu_manager.go:221] "Starting CPU manager" policy="none" Jan 30 15:26:12.914956 kubelet[2704]: I0130 15:26:12.914942 2704 cpu_manager.go:222] "Reconciling" reconcilePeriod="10s" Jan 30 15:26:12.914956 kubelet[2704]: I0130 15:26:12.914966 2704 state_mem.go:36] "Initialized new in-memory state store" Jan 30 15:26:12.915248 kubelet[2704]: I0130 15:26:12.915147 2704 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 30 15:26:12.915248 kubelet[2704]: I0130 15:26:12.915165 2704 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 30 15:26:12.915248 kubelet[2704]: I0130 15:26:12.915184 2704 policy_none.go:49] "None policy: Start" Jan 30 15:26:12.915248 kubelet[2704]: I0130 15:26:12.915192 2704 memory_manager.go:186] "Starting memorymanager" policy="None" Jan 30 15:26:12.915248 kubelet[2704]: I0130 15:26:12.915200 2704 state_mem.go:35] "Initializing new in-memory state store" Jan 30 15:26:12.915506 kubelet[2704]: I0130 15:26:12.915335 2704 state_mem.go:75] "Updated machine memory state" Jan 30 15:26:12.922035 kubelet[2704]: I0130 15:26:12.921999 2704 manager.go:519] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 30 15:26:12.922248 kubelet[2704]: I0130 15:26:12.922227 2704 eviction_manager.go:189] "Eviction manager: starting control loop" Jan 30 15:26:12.922248 kubelet[2704]: I0130 15:26:12.922246 2704 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 30 15:26:12.922987 kubelet[2704]: I0130 15:26:12.922965 2704 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 30 15:26:12.925605 kubelet[2704]: I0130 15:26:12.923857 2704 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-scheduler-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.925605 kubelet[2704]: I0130 15:26:12.924448 2704 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.926573 kubelet[2704]: I0130 15:26:12.926105 2704 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.929041 kubelet[2704]: E0130 15:26:12.928976 2704 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="no imagefs label for configured runtime" Jan 30 15:26:12.932771 kubelet[2704]: I0130 15:26:12.931988 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/971326e260333f76ace84ca47537837d-ca-certs\") pod \"kube-apiserver-ci-4081-3-0-a-5514912dcd\" (UID: \"971326e260333f76ace84ca47537837d\") " pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.932771 kubelet[2704]: I0130 15:26:12.932026 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-flexvolume-dir\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.932771 kubelet[2704]: I0130 15:26:12.932078 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-k8s-certs\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.932771 kubelet[2704]: I0130 15:26:12.932097 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-kubeconfig\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.932771 kubelet[2704]: I0130 15:26:12.932118 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.933057 kubelet[2704]: I0130 15:26:12.932160 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/a6a059240d7e803477388b9e0b38df61-kubeconfig\") pod \"kube-scheduler-ci-4081-3-0-a-5514912dcd\" (UID: \"a6a059240d7e803477388b9e0b38df61\") " pod="kube-system/kube-scheduler-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.933057 kubelet[2704]: I0130 15:26:12.932177 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/971326e260333f76ace84ca47537837d-k8s-certs\") pod \"kube-apiserver-ci-4081-3-0-a-5514912dcd\" (UID: \"971326e260333f76ace84ca47537837d\") " pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.933057 kubelet[2704]: I0130 15:26:12.932193 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/971326e260333f76ace84ca47537837d-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4081-3-0-a-5514912dcd\" (UID: \"971326e260333f76ace84ca47537837d\") " pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:12.933057 kubelet[2704]: I0130 15:26:12.932230 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d6083b14b485d95141f40cab4b058be7-ca-certs\") pod \"kube-controller-manager-ci-4081-3-0-a-5514912dcd\" (UID: \"d6083b14b485d95141f40cab4b058be7\") " pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:13.030233 kubelet[2704]: I0130 15:26:13.030194 2704 kubelet_node_status.go:76] "Attempting to register node" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:13.041707 kubelet[2704]: I0130 15:26:13.041244 2704 kubelet_node_status.go:125] "Node was previously registered" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:13.041707 kubelet[2704]: I0130 15:26:13.041395 2704 kubelet_node_status.go:79] "Successfully registered node" node="ci-4081-3-0-a-5514912dcd" Jan 30 15:26:13.148757 sudo[2738]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Jan 30 15:26:13.149089 sudo[2738]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=0) Jan 30 15:26:13.657693 sudo[2738]: pam_unix(sudo:session): session closed for user root Jan 30 15:26:13.804264 kubelet[2704]: I0130 15:26:13.803942 2704 apiserver.go:52] "Watching apiserver" Jan 30 15:26:13.826011 kubelet[2704]: I0130 15:26:13.825958 2704 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Jan 30 15:26:13.884041 kubelet[2704]: I0130 15:26:13.883767 2704 kubelet.go:3200] "Creating a mirror pod for static pod" pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:13.897618 kubelet[2704]: E0130 15:26:13.895808 2704 kubelet.go:3202] "Failed creating a mirror pod" err="pods \"kube-apiserver-ci-4081-3-0-a-5514912dcd\" already exists" pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" Jan 30 15:26:13.916382 kubelet[2704]: I0130 15:26:13.915696 2704 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4081-3-0-a-5514912dcd" podStartSLOduration=1.915664362 podStartE2EDuration="1.915664362s" podCreationTimestamp="2025-01-30 15:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 15:26:13.899653033 +0000 UTC m=+1.191066863" watchObservedRunningTime="2025-01-30 15:26:13.915664362 +0000 UTC m=+1.207078352" Jan 30 15:26:13.935929 kubelet[2704]: I0130 15:26:13.935854 2704 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4081-3-0-a-5514912dcd" podStartSLOduration=1.935834847 podStartE2EDuration="1.935834847s" podCreationTimestamp="2025-01-30 15:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 15:26:13.918186352 +0000 UTC m=+1.209600222" watchObservedRunningTime="2025-01-30 15:26:13.935834847 +0000 UTC m=+1.227248677" Jan 30 15:26:13.950603 kubelet[2704]: I0130 15:26:13.950510 2704 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4081-3-0-a-5514912dcd" podStartSLOduration=1.950490098 podStartE2EDuration="1.950490098s" podCreationTimestamp="2025-01-30 15:26:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 15:26:13.935832687 +0000 UTC m=+1.227246517" watchObservedRunningTime="2025-01-30 15:26:13.950490098 +0000 UTC m=+1.241903928" Jan 30 15:26:16.030086 sudo[1851]: pam_unix(sudo:session): session closed for user root Jan 30 15:26:16.191053 sshd[1848]: pam_unix(sshd:session): session closed for user core Jan 30 15:26:16.196321 systemd[1]: sshd@6-168.119.58.127:22-139.178.68.195:37640.service: Deactivated successfully. Jan 30 15:26:16.200999 systemd[1]: session-7.scope: Deactivated successfully. Jan 30 15:26:16.201711 systemd[1]: session-7.scope: Consumed 9.501s CPU time, 153.6M memory peak, 0B memory swap peak. Jan 30 15:26:16.204194 systemd-logind[1456]: Session 7 logged out. Waiting for processes to exit. Jan 30 15:26:16.205993 systemd-logind[1456]: Removed session 7. Jan 30 15:26:16.836488 kubelet[2704]: I0130 15:26:16.836443 2704 kuberuntime_manager.go:1702] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 30 15:26:16.837790 containerd[1476]: time="2025-01-30T15:26:16.837699200Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 30 15:26:16.839142 kubelet[2704]: I0130 15:26:16.838532 2704 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 30 15:26:17.475273 systemd[1]: Created slice kubepods-besteffort-podd947134f_2e6a_46f6_b3d6_c243bfd56209.slice - libcontainer container kubepods-besteffort-podd947134f_2e6a_46f6_b3d6_c243bfd56209.slice. Jan 30 15:26:17.495648 systemd[1]: Created slice kubepods-burstable-podbd67b7e3_7bf4_4174_8e95_118b465075f9.slice - libcontainer container kubepods-burstable-podbd67b7e3_7bf4_4174_8e95_118b465075f9.slice. Jan 30 15:26:17.565087 kubelet[2704]: I0130 15:26:17.564647 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/bd67b7e3-7bf4-4174-8e95-118b465075f9-clustermesh-secrets\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565087 kubelet[2704]: I0130 15:26:17.564696 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-hostproc\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565087 kubelet[2704]: I0130 15:26:17.564714 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-run\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565087 kubelet[2704]: I0130 15:26:17.564730 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-cgroup\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565087 kubelet[2704]: I0130 15:26:17.564747 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-host-proc-sys-net\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565087 kubelet[2704]: I0130 15:26:17.564762 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d947134f-2e6a-46f6-b3d6-c243bfd56209-lib-modules\") pod \"kube-proxy-59qlb\" (UID: \"d947134f-2e6a-46f6-b3d6-c243bfd56209\") " pod="kube-system/kube-proxy-59qlb" Jan 30 15:26:17.565326 kubelet[2704]: I0130 15:26:17.564778 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqfdm\" (UniqueName: \"kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-kube-api-access-wqfdm\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565326 kubelet[2704]: I0130 15:26:17.564797 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-config-path\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565326 kubelet[2704]: I0130 15:26:17.564811 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-host-proc-sys-kernel\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565326 kubelet[2704]: I0130 15:26:17.564827 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d947134f-2e6a-46f6-b3d6-c243bfd56209-xtables-lock\") pod \"kube-proxy-59qlb\" (UID: \"d947134f-2e6a-46f6-b3d6-c243bfd56209\") " pod="kube-system/kube-proxy-59qlb" Jan 30 15:26:17.565326 kubelet[2704]: I0130 15:26:17.564842 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-bpf-maps\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565448 kubelet[2704]: I0130 15:26:17.564857 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-lib-modules\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565448 kubelet[2704]: I0130 15:26:17.564871 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/d947134f-2e6a-46f6-b3d6-c243bfd56209-kube-proxy\") pod \"kube-proxy-59qlb\" (UID: \"d947134f-2e6a-46f6-b3d6-c243bfd56209\") " pod="kube-system/kube-proxy-59qlb" Jan 30 15:26:17.565448 kubelet[2704]: I0130 15:26:17.564899 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-xtables-lock\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565448 kubelet[2704]: I0130 15:26:17.564914 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-hubble-tls\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565448 kubelet[2704]: I0130 15:26:17.564929 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sfgv4\" (UniqueName: \"kubernetes.io/projected/d947134f-2e6a-46f6-b3d6-c243bfd56209-kube-api-access-sfgv4\") pod \"kube-proxy-59qlb\" (UID: \"d947134f-2e6a-46f6-b3d6-c243bfd56209\") " pod="kube-system/kube-proxy-59qlb" Jan 30 15:26:17.565448 kubelet[2704]: I0130 15:26:17.564944 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cni-path\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.565593 kubelet[2704]: I0130 15:26:17.564964 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-etc-cni-netd\") pod \"cilium-76dm8\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " pod="kube-system/cilium-76dm8" Jan 30 15:26:17.694021 kubelet[2704]: E0130 15:26:17.693935 2704 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Jan 30 15:26:17.694021 kubelet[2704]: E0130 15:26:17.693986 2704 projected.go:194] Error preparing data for projected volume kube-api-access-wqfdm for pod kube-system/cilium-76dm8: configmap "kube-root-ca.crt" not found Jan 30 15:26:17.694192 kubelet[2704]: E0130 15:26:17.694112 2704 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-kube-api-access-wqfdm podName:bd67b7e3-7bf4-4174-8e95-118b465075f9 nodeName:}" failed. No retries permitted until 2025-01-30 15:26:18.194081145 +0000 UTC m=+5.485494975 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-wqfdm" (UniqueName: "kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-kube-api-access-wqfdm") pod "cilium-76dm8" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9") : configmap "kube-root-ca.crt" not found Jan 30 15:26:17.702244 kubelet[2704]: E0130 15:26:17.702201 2704 projected.go:288] Couldn't get configMap kube-system/kube-root-ca.crt: configmap "kube-root-ca.crt" not found Jan 30 15:26:17.702445 kubelet[2704]: E0130 15:26:17.702431 2704 projected.go:194] Error preparing data for projected volume kube-api-access-sfgv4 for pod kube-system/kube-proxy-59qlb: configmap "kube-root-ca.crt" not found Jan 30 15:26:17.702675 kubelet[2704]: E0130 15:26:17.702659 2704 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/d947134f-2e6a-46f6-b3d6-c243bfd56209-kube-api-access-sfgv4 podName:d947134f-2e6a-46f6-b3d6-c243bfd56209 nodeName:}" failed. No retries permitted until 2025-01-30 15:26:18.202583208 +0000 UTC m=+5.493997038 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-sfgv4" (UniqueName: "kubernetes.io/projected/d947134f-2e6a-46f6-b3d6-c243bfd56209-kube-api-access-sfgv4") pod "kube-proxy-59qlb" (UID: "d947134f-2e6a-46f6-b3d6-c243bfd56209") : configmap "kube-root-ca.crt" not found Jan 30 15:26:17.958937 systemd[1]: Created slice kubepods-besteffort-pod5263f0fa_2944_49ce_8475_80af8b5aa5e2.slice - libcontainer container kubepods-besteffort-pod5263f0fa_2944_49ce_8475_80af8b5aa5e2.slice. Jan 30 15:26:17.968692 kubelet[2704]: I0130 15:26:17.968639 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/5263f0fa-2944-49ce-8475-80af8b5aa5e2-cilium-config-path\") pod \"cilium-operator-6c4d7847fc-fjcsn\" (UID: \"5263f0fa-2944-49ce-8475-80af8b5aa5e2\") " pod="kube-system/cilium-operator-6c4d7847fc-fjcsn" Jan 30 15:26:17.969107 kubelet[2704]: I0130 15:26:17.968744 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54b9t\" (UniqueName: \"kubernetes.io/projected/5263f0fa-2944-49ce-8475-80af8b5aa5e2-kube-api-access-54b9t\") pod \"cilium-operator-6c4d7847fc-fjcsn\" (UID: \"5263f0fa-2944-49ce-8475-80af8b5aa5e2\") " pod="kube-system/cilium-operator-6c4d7847fc-fjcsn" Jan 30 15:26:18.265386 containerd[1476]: time="2025-01-30T15:26:18.265283795Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-fjcsn,Uid:5263f0fa-2944-49ce-8475-80af8b5aa5e2,Namespace:kube-system,Attempt:0,}" Jan 30 15:26:18.300223 containerd[1476]: time="2025-01-30T15:26:18.300097012Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 15:26:18.300458 containerd[1476]: time="2025-01-30T15:26:18.300236735Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 15:26:18.300458 containerd[1476]: time="2025-01-30T15:26:18.300267776Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:18.300458 containerd[1476]: time="2025-01-30T15:26:18.300415620Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:18.320130 systemd[1]: Started cri-containerd-fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075.scope - libcontainer container fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075. Jan 30 15:26:18.366166 containerd[1476]: time="2025-01-30T15:26:18.366049031Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6c4d7847fc-fjcsn,Uid:5263f0fa-2944-49ce-8475-80af8b5aa5e2,Namespace:kube-system,Attempt:0,} returns sandbox id \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\"" Jan 30 15:26:18.371438 containerd[1476]: time="2025-01-30T15:26:18.371151243Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Jan 30 15:26:18.386896 containerd[1476]: time="2025-01-30T15:26:18.386723764Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-59qlb,Uid:d947134f-2e6a-46f6-b3d6-c243bfd56209,Namespace:kube-system,Attempt:0,}" Jan 30 15:26:18.403037 containerd[1476]: time="2025-01-30T15:26:18.402864900Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-76dm8,Uid:bd67b7e3-7bf4-4174-8e95-118b465075f9,Namespace:kube-system,Attempt:0,}" Jan 30 15:26:18.424508 containerd[1476]: time="2025-01-30T15:26:18.424199810Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 15:26:18.424508 containerd[1476]: time="2025-01-30T15:26:18.424265851Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 15:26:18.424508 containerd[1476]: time="2025-01-30T15:26:18.424293572Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:18.424508 containerd[1476]: time="2025-01-30T15:26:18.424409335Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:18.436818 containerd[1476]: time="2025-01-30T15:26:18.436713932Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 15:26:18.436983 containerd[1476]: time="2025-01-30T15:26:18.436850016Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 15:26:18.436983 containerd[1476]: time="2025-01-30T15:26:18.436883537Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:18.437246 containerd[1476]: time="2025-01-30T15:26:18.437044221Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:18.445768 systemd[1]: Started cri-containerd-d05e336360482490962b45b295a17fd61ba76cbe6fc2fc53ea4e66bf7d7f3b80.scope - libcontainer container d05e336360482490962b45b295a17fd61ba76cbe6fc2fc53ea4e66bf7d7f3b80. Jan 30 15:26:18.465912 systemd[1]: Started cri-containerd-474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac.scope - libcontainer container 474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac. Jan 30 15:26:18.494249 containerd[1476]: time="2025-01-30T15:26:18.493932807Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-59qlb,Uid:d947134f-2e6a-46f6-b3d6-c243bfd56209,Namespace:kube-system,Attempt:0,} returns sandbox id \"d05e336360482490962b45b295a17fd61ba76cbe6fc2fc53ea4e66bf7d7f3b80\"" Jan 30 15:26:18.500837 containerd[1476]: time="2025-01-30T15:26:18.500470815Z" level=info msg="CreateContainer within sandbox \"d05e336360482490962b45b295a17fd61ba76cbe6fc2fc53ea4e66bf7d7f3b80\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 30 15:26:18.507055 containerd[1476]: time="2025-01-30T15:26:18.506970543Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-76dm8,Uid:bd67b7e3-7bf4-4174-8e95-118b465075f9,Namespace:kube-system,Attempt:0,} returns sandbox id \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\"" Jan 30 15:26:18.521317 containerd[1476]: time="2025-01-30T15:26:18.520567733Z" level=info msg="CreateContainer within sandbox \"d05e336360482490962b45b295a17fd61ba76cbe6fc2fc53ea4e66bf7d7f3b80\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"3ad9ce09536647efac379826040f354ec8c007594dff17b44ed37b6cb60fca30\"" Jan 30 15:26:18.522944 containerd[1476]: time="2025-01-30T15:26:18.522771070Z" level=info msg="StartContainer for \"3ad9ce09536647efac379826040f354ec8c007594dff17b44ed37b6cb60fca30\"" Jan 30 15:26:18.562829 systemd[1]: Started cri-containerd-3ad9ce09536647efac379826040f354ec8c007594dff17b44ed37b6cb60fca30.scope - libcontainer container 3ad9ce09536647efac379826040f354ec8c007594dff17b44ed37b6cb60fca30. Jan 30 15:26:18.599474 containerd[1476]: time="2025-01-30T15:26:18.599423365Z" level=info msg="StartContainer for \"3ad9ce09536647efac379826040f354ec8c007594dff17b44ed37b6cb60fca30\" returns successfully" Jan 30 15:26:19.597415 kubelet[2704]: I0130 15:26:19.597278 2704 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-59qlb" podStartSLOduration=2.597258432 podStartE2EDuration="2.597258432s" podCreationTimestamp="2025-01-30 15:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 15:26:18.91512758 +0000 UTC m=+6.206541450" watchObservedRunningTime="2025-01-30 15:26:19.597258432 +0000 UTC m=+6.888672262" Jan 30 15:26:20.106887 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3529010694.mount: Deactivated successfully. Jan 30 15:26:20.527805 containerd[1476]: time="2025-01-30T15:26:20.527743896Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:26:20.529046 containerd[1476]: time="2025-01-30T15:26:20.528877724Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17135306" Jan 30 15:26:20.530618 containerd[1476]: time="2025-01-30T15:26:20.529711705Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:26:20.531575 containerd[1476]: time="2025-01-30T15:26:20.531512430Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 2.160313425s" Jan 30 15:26:20.531724 containerd[1476]: time="2025-01-30T15:26:20.531579071Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Jan 30 15:26:20.534813 containerd[1476]: time="2025-01-30T15:26:20.534757191Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Jan 30 15:26:20.536054 containerd[1476]: time="2025-01-30T15:26:20.536013182Z" level=info msg="CreateContainer within sandbox \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Jan 30 15:26:20.554197 containerd[1476]: time="2025-01-30T15:26:20.554136114Z" level=info msg="CreateContainer within sandbox \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\"" Jan 30 15:26:20.555780 containerd[1476]: time="2025-01-30T15:26:20.555726954Z" level=info msg="StartContainer for \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\"" Jan 30 15:26:20.592912 systemd[1]: Started cri-containerd-0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c.scope - libcontainer container 0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c. Jan 30 15:26:20.622866 containerd[1476]: time="2025-01-30T15:26:20.622745667Z" level=info msg="StartContainer for \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\" returns successfully" Jan 30 15:26:24.462698 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2185538078.mount: Deactivated successfully. Jan 30 15:26:24.820447 kubelet[2704]: I0130 15:26:24.820372 2704 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-6c4d7847fc-fjcsn" podStartSLOduration=5.655884093 podStartE2EDuration="7.820352544s" podCreationTimestamp="2025-01-30 15:26:17 +0000 UTC" firstStartedPulling="2025-01-30 15:26:18.368912465 +0000 UTC m=+5.660326335" lastFinishedPulling="2025-01-30 15:26:20.533380956 +0000 UTC m=+7.824794786" observedRunningTime="2025-01-30 15:26:20.952376773 +0000 UTC m=+8.243790603" watchObservedRunningTime="2025-01-30 15:26:24.820352544 +0000 UTC m=+12.111766374" Jan 30 15:26:25.946683 containerd[1476]: time="2025-01-30T15:26:25.946501984Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:26:25.949074 containerd[1476]: time="2025-01-30T15:26:25.948843718Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157646710" Jan 30 15:26:25.950167 containerd[1476]: time="2025-01-30T15:26:25.950134068Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 15:26:25.952260 containerd[1476]: time="2025-01-30T15:26:25.952121554Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 5.417311442s" Jan 30 15:26:25.952260 containerd[1476]: time="2025-01-30T15:26:25.952162395Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Jan 30 15:26:25.956607 containerd[1476]: time="2025-01-30T15:26:25.955770078Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jan 30 15:26:25.970327 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3102889940.mount: Deactivated successfully. Jan 30 15:26:25.973005 containerd[1476]: time="2025-01-30T15:26:25.972949076Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\"" Jan 30 15:26:25.973740 containerd[1476]: time="2025-01-30T15:26:25.973699333Z" level=info msg="StartContainer for \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\"" Jan 30 15:26:26.007892 systemd[1]: Started cri-containerd-9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b.scope - libcontainer container 9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b. Jan 30 15:26:26.043946 containerd[1476]: time="2025-01-30T15:26:26.043892863Z" level=info msg="StartContainer for \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\" returns successfully" Jan 30 15:26:26.059851 systemd[1]: cri-containerd-9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b.scope: Deactivated successfully. Jan 30 15:26:26.239655 containerd[1476]: time="2025-01-30T15:26:26.239346802Z" level=info msg="shim disconnected" id=9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b namespace=k8s.io Jan 30 15:26:26.239655 containerd[1476]: time="2025-01-30T15:26:26.239412483Z" level=warning msg="cleaning up after shim disconnected" id=9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b namespace=k8s.io Jan 30 15:26:26.239655 containerd[1476]: time="2025-01-30T15:26:26.239423963Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:26:26.938997 containerd[1476]: time="2025-01-30T15:26:26.938822757Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jan 30 15:26:26.956110 containerd[1476]: time="2025-01-30T15:26:26.956036109Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\"" Jan 30 15:26:26.956897 containerd[1476]: time="2025-01-30T15:26:26.956646083Z" level=info msg="StartContainer for \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\"" Jan 30 15:26:26.970563 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b-rootfs.mount: Deactivated successfully. Jan 30 15:26:26.996838 systemd[1]: Started cri-containerd-6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3.scope - libcontainer container 6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3. Jan 30 15:26:27.036046 containerd[1476]: time="2025-01-30T15:26:27.035439669Z" level=info msg="StartContainer for \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\" returns successfully" Jan 30 15:26:27.048416 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 30 15:26:27.049046 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 30 15:26:27.049374 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables... Jan 30 15:26:27.056120 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 30 15:26:27.058965 systemd[1]: cri-containerd-6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3.scope: Deactivated successfully. Jan 30 15:26:27.074604 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 30 15:26:27.088152 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3-rootfs.mount: Deactivated successfully. Jan 30 15:26:27.097002 containerd[1476]: time="2025-01-30T15:26:27.096903212Z" level=info msg="shim disconnected" id=6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3 namespace=k8s.io Jan 30 15:26:27.097213 containerd[1476]: time="2025-01-30T15:26:27.097003854Z" level=warning msg="cleaning up after shim disconnected" id=6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3 namespace=k8s.io Jan 30 15:26:27.097213 containerd[1476]: time="2025-01-30T15:26:27.097136577Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:26:27.944487 containerd[1476]: time="2025-01-30T15:26:27.944418592Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jan 30 15:26:27.967533 containerd[1476]: time="2025-01-30T15:26:27.967176184Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\"" Jan 30 15:26:27.968724 containerd[1476]: time="2025-01-30T15:26:27.968620137Z" level=info msg="StartContainer for \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\"" Jan 30 15:26:28.009826 systemd[1]: Started cri-containerd-257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc.scope - libcontainer container 257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc. Jan 30 15:26:28.050879 systemd[1]: cri-containerd-257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc.scope: Deactivated successfully. Jan 30 15:26:28.052894 containerd[1476]: time="2025-01-30T15:26:28.052843935Z" level=info msg="StartContainer for \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\" returns successfully" Jan 30 15:26:28.076708 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc-rootfs.mount: Deactivated successfully. Jan 30 15:26:28.082369 containerd[1476]: time="2025-01-30T15:26:28.082185345Z" level=info msg="shim disconnected" id=257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc namespace=k8s.io Jan 30 15:26:28.082369 containerd[1476]: time="2025-01-30T15:26:28.082241347Z" level=warning msg="cleaning up after shim disconnected" id=257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc namespace=k8s.io Jan 30 15:26:28.082369 containerd[1476]: time="2025-01-30T15:26:28.082249667Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:26:28.951835 containerd[1476]: time="2025-01-30T15:26:28.951423306Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jan 30 15:26:28.972666 containerd[1476]: time="2025-01-30T15:26:28.972224927Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\"" Jan 30 15:26:28.974266 containerd[1476]: time="2025-01-30T15:26:28.973037225Z" level=info msg="StartContainer for \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\"" Jan 30 15:26:29.001914 systemd[1]: run-containerd-runc-k8s.io-ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3-runc.kDT80T.mount: Deactivated successfully. Jan 30 15:26:29.012786 systemd[1]: Started cri-containerd-ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3.scope - libcontainer container ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3. Jan 30 15:26:29.039257 systemd[1]: cri-containerd-ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3.scope: Deactivated successfully. Jan 30 15:26:29.042848 containerd[1476]: time="2025-01-30T15:26:29.040829156Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd67b7e3_7bf4_4174_8e95_118b465075f9.slice/cri-containerd-ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3.scope/memory.events\": no such file or directory" Jan 30 15:26:29.047985 containerd[1476]: time="2025-01-30T15:26:29.047850590Z" level=info msg="StartContainer for \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\" returns successfully" Jan 30 15:26:29.065794 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3-rootfs.mount: Deactivated successfully. Jan 30 15:26:29.070025 containerd[1476]: time="2025-01-30T15:26:29.069931753Z" level=info msg="shim disconnected" id=ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3 namespace=k8s.io Jan 30 15:26:29.070354 containerd[1476]: time="2025-01-30T15:26:29.070097397Z" level=warning msg="cleaning up after shim disconnected" id=ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3 namespace=k8s.io Jan 30 15:26:29.070354 containerd[1476]: time="2025-01-30T15:26:29.070115237Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:26:29.955512 containerd[1476]: time="2025-01-30T15:26:29.955468449Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jan 30 15:26:29.977295 containerd[1476]: time="2025-01-30T15:26:29.977131963Z" level=info msg="CreateContainer within sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\"" Jan 30 15:26:29.981450 containerd[1476]: time="2025-01-30T15:26:29.978337070Z" level=info msg="StartContainer for \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\"" Jan 30 15:26:30.012797 systemd[1]: Started cri-containerd-e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1.scope - libcontainer container e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1. Jan 30 15:26:30.042397 containerd[1476]: time="2025-01-30T15:26:30.042338298Z" level=info msg="StartContainer for \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\" returns successfully" Jan 30 15:26:30.174670 kubelet[2704]: I0130 15:26:30.171840 2704 kubelet_node_status.go:502] "Fast updating node status as it just became ready" Jan 30 15:26:30.217846 systemd[1]: Created slice kubepods-burstable-podacfc3ffa_2506_45ee_bf02_58d1a6b776af.slice - libcontainer container kubepods-burstable-podacfc3ffa_2506_45ee_bf02_58d1a6b776af.slice. Jan 30 15:26:30.230889 systemd[1]: Created slice kubepods-burstable-pod59d8d1db_22ff_4135_8150_fbb3ebe186d7.slice - libcontainer container kubepods-burstable-pod59d8d1db_22ff_4135_8150_fbb3ebe186d7.slice. Jan 30 15:26:30.254415 kubelet[2704]: I0130 15:26:30.254368 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59d8d1db-22ff-4135-8150-fbb3ebe186d7-config-volume\") pod \"coredns-668d6bf9bc-k6gz8\" (UID: \"59d8d1db-22ff-4135-8150-fbb3ebe186d7\") " pod="kube-system/coredns-668d6bf9bc-k6gz8" Jan 30 15:26:30.254415 kubelet[2704]: I0130 15:26:30.254417 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/acfc3ffa-2506-45ee-bf02-58d1a6b776af-config-volume\") pod \"coredns-668d6bf9bc-zxv9q\" (UID: \"acfc3ffa-2506-45ee-bf02-58d1a6b776af\") " pod="kube-system/coredns-668d6bf9bc-zxv9q" Jan 30 15:26:30.254415 kubelet[2704]: I0130 15:26:30.254444 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6rwn\" (UniqueName: \"kubernetes.io/projected/acfc3ffa-2506-45ee-bf02-58d1a6b776af-kube-api-access-q6rwn\") pod \"coredns-668d6bf9bc-zxv9q\" (UID: \"acfc3ffa-2506-45ee-bf02-58d1a6b776af\") " pod="kube-system/coredns-668d6bf9bc-zxv9q" Jan 30 15:26:30.254415 kubelet[2704]: I0130 15:26:30.254523 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tpb4\" (UniqueName: \"kubernetes.io/projected/59d8d1db-22ff-4135-8150-fbb3ebe186d7-kube-api-access-2tpb4\") pod \"coredns-668d6bf9bc-k6gz8\" (UID: \"59d8d1db-22ff-4135-8150-fbb3ebe186d7\") " pod="kube-system/coredns-668d6bf9bc-k6gz8" Jan 30 15:26:30.525757 containerd[1476]: time="2025-01-30T15:26:30.525710534Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-zxv9q,Uid:acfc3ffa-2506-45ee-bf02-58d1a6b776af,Namespace:kube-system,Attempt:0,}" Jan 30 15:26:30.537355 containerd[1476]: time="2025-01-30T15:26:30.537164821Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-k6gz8,Uid:59d8d1db-22ff-4135-8150-fbb3ebe186d7,Namespace:kube-system,Attempt:0,}" Jan 30 15:26:30.990563 kubelet[2704]: I0130 15:26:30.989526 2704 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-76dm8" podStartSLOduration=6.544825712 podStartE2EDuration="13.989506907s" podCreationTimestamp="2025-01-30 15:26:17 +0000 UTC" firstStartedPulling="2025-01-30 15:26:18.50919088 +0000 UTC m=+5.800604710" lastFinishedPulling="2025-01-30 15:26:25.953872075 +0000 UTC m=+13.245285905" observedRunningTime="2025-01-30 15:26:30.986270117 +0000 UTC m=+18.277683947" watchObservedRunningTime="2025-01-30 15:26:30.989506907 +0000 UTC m=+18.280920737" Jan 30 15:26:32.203027 systemd-networkd[1371]: cilium_host: Link UP Jan 30 15:26:32.203174 systemd-networkd[1371]: cilium_net: Link UP Jan 30 15:26:32.203178 systemd-networkd[1371]: cilium_net: Gained carrier Jan 30 15:26:32.203885 systemd-networkd[1371]: cilium_host: Gained carrier Jan 30 15:26:32.337613 systemd-networkd[1371]: cilium_vxlan: Link UP Jan 30 15:26:32.337682 systemd-networkd[1371]: cilium_vxlan: Gained carrier Jan 30 15:26:32.540353 systemd-networkd[1371]: cilium_net: Gained IPv6LL Jan 30 15:26:32.654930 kernel: NET: Registered PF_ALG protocol family Jan 30 15:26:33.083865 systemd-networkd[1371]: cilium_host: Gained IPv6LL Jan 30 15:26:33.442379 systemd-networkd[1371]: lxc_health: Link UP Jan 30 15:26:33.458112 systemd-networkd[1371]: lxc_health: Gained carrier Jan 30 15:26:34.044081 systemd-networkd[1371]: cilium_vxlan: Gained IPv6LL Jan 30 15:26:34.070488 systemd-networkd[1371]: lxcbf9f008b1369: Link UP Jan 30 15:26:34.077583 kernel: eth0: renamed from tmp2c6bc Jan 30 15:26:34.083570 systemd-networkd[1371]: lxcbf9f008b1369: Gained carrier Jan 30 15:26:34.103018 systemd-networkd[1371]: lxc2825e15c3a28: Link UP Jan 30 15:26:34.114585 kernel: eth0: renamed from tmpedc94 Jan 30 15:26:34.120400 systemd-networkd[1371]: lxc2825e15c3a28: Gained carrier Jan 30 15:26:35.387887 systemd-networkd[1371]: lxc_health: Gained IPv6LL Jan 30 15:26:35.451831 systemd-networkd[1371]: lxcbf9f008b1369: Gained IPv6LL Jan 30 15:26:35.835983 systemd-networkd[1371]: lxc2825e15c3a28: Gained IPv6LL Jan 30 15:26:38.248183 containerd[1476]: time="2025-01-30T15:26:38.247714100Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 15:26:38.248959 containerd[1476]: time="2025-01-30T15:26:38.248632878Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 15:26:38.248959 containerd[1476]: time="2025-01-30T15:26:38.248735680Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:38.248959 containerd[1476]: time="2025-01-30T15:26:38.248880402Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:38.280952 systemd[1]: Started cri-containerd-edc9487a447e5db5bbd16c3993b4782cbc87aa33e4484c7d17cebedebc5a7e3a.scope - libcontainer container edc9487a447e5db5bbd16c3993b4782cbc87aa33e4484c7d17cebedebc5a7e3a. Jan 30 15:26:38.302575 containerd[1476]: time="2025-01-30T15:26:38.300894420Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 15:26:38.302575 containerd[1476]: time="2025-01-30T15:26:38.300977422Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 15:26:38.302575 containerd[1476]: time="2025-01-30T15:26:38.301002262Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:38.302575 containerd[1476]: time="2025-01-30T15:26:38.301487072Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:26:38.344776 systemd[1]: Started cri-containerd-2c6bcfcc4129f32641a0e3a95056fde2febb79157c4c911edb80626c22830d3b.scope - libcontainer container 2c6bcfcc4129f32641a0e3a95056fde2febb79157c4c911edb80626c22830d3b. Jan 30 15:26:38.367649 containerd[1476]: time="2025-01-30T15:26:38.367312080Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-k6gz8,Uid:59d8d1db-22ff-4135-8150-fbb3ebe186d7,Namespace:kube-system,Attempt:0,} returns sandbox id \"edc9487a447e5db5bbd16c3993b4782cbc87aa33e4484c7d17cebedebc5a7e3a\"" Jan 30 15:26:38.375927 containerd[1476]: time="2025-01-30T15:26:38.375875487Z" level=info msg="CreateContainer within sandbox \"edc9487a447e5db5bbd16c3993b4782cbc87aa33e4484c7d17cebedebc5a7e3a\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 30 15:26:38.396989 containerd[1476]: time="2025-01-30T15:26:38.396900619Z" level=info msg="CreateContainer within sandbox \"edc9487a447e5db5bbd16c3993b4782cbc87aa33e4484c7d17cebedebc5a7e3a\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"018da6913278bc2070eed84ce4ba957947c2d592795e97bc33b9da66d7b0190b\"" Jan 30 15:26:38.399358 containerd[1476]: time="2025-01-30T15:26:38.399282066Z" level=info msg="StartContainer for \"018da6913278bc2070eed84ce4ba957947c2d592795e97bc33b9da66d7b0190b\"" Jan 30 15:26:38.433094 systemd[1]: Started cri-containerd-018da6913278bc2070eed84ce4ba957947c2d592795e97bc33b9da66d7b0190b.scope - libcontainer container 018da6913278bc2070eed84ce4ba957947c2d592795e97bc33b9da66d7b0190b. Jan 30 15:26:38.445367 containerd[1476]: time="2025-01-30T15:26:38.445311766Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-668d6bf9bc-zxv9q,Uid:acfc3ffa-2506-45ee-bf02-58d1a6b776af,Namespace:kube-system,Attempt:0,} returns sandbox id \"2c6bcfcc4129f32641a0e3a95056fde2febb79157c4c911edb80626c22830d3b\"" Jan 30 15:26:38.452103 containerd[1476]: time="2025-01-30T15:26:38.452046098Z" level=info msg="CreateContainer within sandbox \"2c6bcfcc4129f32641a0e3a95056fde2febb79157c4c911edb80626c22830d3b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 30 15:26:38.478448 containerd[1476]: time="2025-01-30T15:26:38.477188710Z" level=info msg="CreateContainer within sandbox \"2c6bcfcc4129f32641a0e3a95056fde2febb79157c4c911edb80626c22830d3b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"82ff286d9bc0b9d03a75a018a695a775ab41aa36ab0564da131f75d8ceca3480\"" Jan 30 15:26:38.480586 containerd[1476]: time="2025-01-30T15:26:38.478875063Z" level=info msg="StartContainer for \"018da6913278bc2070eed84ce4ba957947c2d592795e97bc33b9da66d7b0190b\" returns successfully" Jan 30 15:26:38.480586 containerd[1476]: time="2025-01-30T15:26:38.479367273Z" level=info msg="StartContainer for \"82ff286d9bc0b9d03a75a018a695a775ab41aa36ab0564da131f75d8ceca3480\"" Jan 30 15:26:38.523055 systemd[1]: Started cri-containerd-82ff286d9bc0b9d03a75a018a695a775ab41aa36ab0564da131f75d8ceca3480.scope - libcontainer container 82ff286d9bc0b9d03a75a018a695a775ab41aa36ab0564da131f75d8ceca3480. Jan 30 15:26:38.562503 containerd[1476]: time="2025-01-30T15:26:38.562447738Z" level=info msg="StartContainer for \"82ff286d9bc0b9d03a75a018a695a775ab41aa36ab0564da131f75d8ceca3480\" returns successfully" Jan 30 15:26:39.027063 kubelet[2704]: I0130 15:26:39.026893 2704 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-k6gz8" podStartSLOduration=22.026868421 podStartE2EDuration="22.026868421s" podCreationTimestamp="2025-01-30 15:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 15:26:39.003339965 +0000 UTC m=+26.294753835" watchObservedRunningTime="2025-01-30 15:26:39.026868421 +0000 UTC m=+26.318282251" Jan 30 15:31:01.437112 systemd[1]: Started sshd@7-168.119.58.127:22-139.178.68.195:35166.service - OpenSSH per-connection server daemon (139.178.68.195:35166). Jan 30 15:31:02.417632 sshd[4116]: Accepted publickey for core from 139.178.68.195 port 35166 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:02.420491 sshd[4116]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:02.426797 systemd-logind[1456]: New session 8 of user core. Jan 30 15:31:02.432834 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 30 15:31:03.196178 sshd[4116]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:03.201203 systemd[1]: sshd@7-168.119.58.127:22-139.178.68.195:35166.service: Deactivated successfully. Jan 30 15:31:03.205521 systemd[1]: session-8.scope: Deactivated successfully. Jan 30 15:31:03.209525 systemd-logind[1456]: Session 8 logged out. Waiting for processes to exit. Jan 30 15:31:03.211211 systemd-logind[1456]: Removed session 8. Jan 30 15:31:08.385737 systemd[1]: Started sshd@8-168.119.58.127:22-139.178.68.195:49494.service - OpenSSH per-connection server daemon (139.178.68.195:49494). Jan 30 15:31:09.381691 sshd[4130]: Accepted publickey for core from 139.178.68.195 port 49494 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:09.384966 sshd[4130]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:09.391016 systemd-logind[1456]: New session 9 of user core. Jan 30 15:31:09.400816 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 30 15:31:10.149056 sshd[4130]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:10.154426 systemd[1]: sshd@8-168.119.58.127:22-139.178.68.195:49494.service: Deactivated successfully. Jan 30 15:31:10.156828 systemd[1]: session-9.scope: Deactivated successfully. Jan 30 15:31:10.158318 systemd-logind[1456]: Session 9 logged out. Waiting for processes to exit. Jan 30 15:31:10.159372 systemd-logind[1456]: Removed session 9. Jan 30 15:31:15.328076 systemd[1]: Started sshd@9-168.119.58.127:22-139.178.68.195:55522.service - OpenSSH per-connection server daemon (139.178.68.195:55522). Jan 30 15:31:16.307835 sshd[4146]: Accepted publickey for core from 139.178.68.195 port 55522 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:16.309925 sshd[4146]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:16.315820 systemd-logind[1456]: New session 10 of user core. Jan 30 15:31:16.320867 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 30 15:31:16.452061 update_engine[1458]: I20250130 15:31:16.451989 1458 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Jan 30 15:31:16.452789 update_engine[1458]: I20250130 15:31:16.452683 1458 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Jan 30 15:31:16.453084 update_engine[1458]: I20250130 15:31:16.453022 1458 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Jan 30 15:31:16.455299 update_engine[1458]: I20250130 15:31:16.455159 1458 omaha_request_params.cc:62] Current group set to lts Jan 30 15:31:16.455454 update_engine[1458]: I20250130 15:31:16.455327 1458 update_attempter.cc:499] Already updated boot flags. Skipping. Jan 30 15:31:16.455454 update_engine[1458]: I20250130 15:31:16.455348 1458 update_attempter.cc:643] Scheduling an action processor start. Jan 30 15:31:16.455454 update_engine[1458]: I20250130 15:31:16.455377 1458 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Jan 30 15:31:16.455454 update_engine[1458]: I20250130 15:31:16.455430 1458 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Jan 30 15:31:16.455647 update_engine[1458]: I20250130 15:31:16.455514 1458 omaha_request_action.cc:271] Posting an Omaha request to disabled Jan 30 15:31:16.455647 update_engine[1458]: I20250130 15:31:16.455529 1458 omaha_request_action.cc:272] Request: Jan 30 15:31:16.455647 update_engine[1458]: Jan 30 15:31:16.455647 update_engine[1458]: Jan 30 15:31:16.455647 update_engine[1458]: Jan 30 15:31:16.455647 update_engine[1458]: Jan 30 15:31:16.455647 update_engine[1458]: Jan 30 15:31:16.455647 update_engine[1458]: Jan 30 15:31:16.455647 update_engine[1458]: Jan 30 15:31:16.455647 update_engine[1458]: Jan 30 15:31:16.455647 update_engine[1458]: I20250130 15:31:16.455570 1458 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jan 30 15:31:16.456435 locksmithd[1501]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Jan 30 15:31:16.457928 update_engine[1458]: I20250130 15:31:16.457890 1458 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jan 30 15:31:16.458273 update_engine[1458]: I20250130 15:31:16.458235 1458 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jan 30 15:31:16.459466 update_engine[1458]: E20250130 15:31:16.459411 1458 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jan 30 15:31:16.459619 update_engine[1458]: I20250130 15:31:16.459489 1458 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Jan 30 15:31:17.067222 sshd[4146]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:17.072232 systemd[1]: sshd@9-168.119.58.127:22-139.178.68.195:55522.service: Deactivated successfully. Jan 30 15:31:17.075228 systemd[1]: session-10.scope: Deactivated successfully. Jan 30 15:31:17.076172 systemd-logind[1456]: Session 10 logged out. Waiting for processes to exit. Jan 30 15:31:17.078599 systemd-logind[1456]: Removed session 10. Jan 30 15:31:17.253692 systemd[1]: Started sshd@10-168.119.58.127:22-139.178.68.195:55528.service - OpenSSH per-connection server daemon (139.178.68.195:55528). Jan 30 15:31:18.244434 sshd[4160]: Accepted publickey for core from 139.178.68.195 port 55528 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:18.247300 sshd[4160]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:18.253678 systemd-logind[1456]: New session 11 of user core. Jan 30 15:31:18.257808 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 30 15:31:19.063795 sshd[4160]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:19.068739 systemd[1]: sshd@10-168.119.58.127:22-139.178.68.195:55528.service: Deactivated successfully. Jan 30 15:31:19.072701 systemd[1]: session-11.scope: Deactivated successfully. Jan 30 15:31:19.075454 systemd-logind[1456]: Session 11 logged out. Waiting for processes to exit. Jan 30 15:31:19.077658 systemd-logind[1456]: Removed session 11. Jan 30 15:31:19.236876 systemd[1]: Started sshd@11-168.119.58.127:22-139.178.68.195:55544.service - OpenSSH per-connection server daemon (139.178.68.195:55544). Jan 30 15:31:20.231593 sshd[4173]: Accepted publickey for core from 139.178.68.195 port 55544 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:20.232609 sshd[4173]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:20.238672 systemd-logind[1456]: New session 12 of user core. Jan 30 15:31:20.247893 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 30 15:31:21.000342 sshd[4173]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:21.006584 systemd[1]: sshd@11-168.119.58.127:22-139.178.68.195:55544.service: Deactivated successfully. Jan 30 15:31:21.009272 systemd[1]: session-12.scope: Deactivated successfully. Jan 30 15:31:21.010490 systemd-logind[1456]: Session 12 logged out. Waiting for processes to exit. Jan 30 15:31:21.012741 systemd-logind[1456]: Removed session 12. Jan 30 15:31:26.176078 systemd[1]: Started sshd@12-168.119.58.127:22-139.178.68.195:47762.service - OpenSSH per-connection server daemon (139.178.68.195:47762). Jan 30 15:31:26.456387 update_engine[1458]: I20250130 15:31:26.456179 1458 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jan 30 15:31:26.456884 update_engine[1458]: I20250130 15:31:26.456782 1458 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jan 30 15:31:26.457133 update_engine[1458]: I20250130 15:31:26.457067 1458 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jan 30 15:31:26.457836 update_engine[1458]: E20250130 15:31:26.457711 1458 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jan 30 15:31:26.457836 update_engine[1458]: I20250130 15:31:26.457804 1458 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Jan 30 15:31:27.154746 sshd[4186]: Accepted publickey for core from 139.178.68.195 port 47762 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:27.156521 sshd[4186]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:27.162816 systemd-logind[1456]: New session 13 of user core. Jan 30 15:31:27.169886 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 30 15:31:27.910816 sshd[4186]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:27.915233 systemd[1]: sshd@12-168.119.58.127:22-139.178.68.195:47762.service: Deactivated successfully. Jan 30 15:31:27.918061 systemd[1]: session-13.scope: Deactivated successfully. Jan 30 15:31:27.919989 systemd-logind[1456]: Session 13 logged out. Waiting for processes to exit. Jan 30 15:31:27.921933 systemd-logind[1456]: Removed session 13. Jan 30 15:31:28.090972 systemd[1]: Started sshd@13-168.119.58.127:22-139.178.68.195:47764.service - OpenSSH per-connection server daemon (139.178.68.195:47764). Jan 30 15:31:29.061605 sshd[4199]: Accepted publickey for core from 139.178.68.195 port 47764 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:29.064742 sshd[4199]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:29.070442 systemd-logind[1456]: New session 14 of user core. Jan 30 15:31:29.078872 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 30 15:31:29.867314 sshd[4199]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:29.874671 systemd[1]: sshd@13-168.119.58.127:22-139.178.68.195:47764.service: Deactivated successfully. Jan 30 15:31:29.876673 systemd[1]: session-14.scope: Deactivated successfully. Jan 30 15:31:29.878489 systemd-logind[1456]: Session 14 logged out. Waiting for processes to exit. Jan 30 15:31:29.880268 systemd-logind[1456]: Removed session 14. Jan 30 15:31:30.046938 systemd[1]: Started sshd@14-168.119.58.127:22-139.178.68.195:47778.service - OpenSSH per-connection server daemon (139.178.68.195:47778). Jan 30 15:31:31.028777 sshd[4210]: Accepted publickey for core from 139.178.68.195 port 47778 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:31.031264 sshd[4210]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:31.037173 systemd-logind[1456]: New session 15 of user core. Jan 30 15:31:31.042808 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 30 15:31:32.786753 sshd[4210]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:32.792758 systemd[1]: sshd@14-168.119.58.127:22-139.178.68.195:47778.service: Deactivated successfully. Jan 30 15:31:32.796693 systemd[1]: session-15.scope: Deactivated successfully. Jan 30 15:31:32.797756 systemd-logind[1456]: Session 15 logged out. Waiting for processes to exit. Jan 30 15:31:32.799089 systemd-logind[1456]: Removed session 15. Jan 30 15:31:32.966943 systemd[1]: Started sshd@15-168.119.58.127:22-139.178.68.195:47794.service - OpenSSH per-connection server daemon (139.178.68.195:47794). Jan 30 15:31:33.956783 sshd[4228]: Accepted publickey for core from 139.178.68.195 port 47794 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:33.958967 sshd[4228]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:33.963642 systemd-logind[1456]: New session 16 of user core. Jan 30 15:31:33.974972 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 30 15:31:34.835403 sshd[4228]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:34.842062 systemd[1]: sshd@15-168.119.58.127:22-139.178.68.195:47794.service: Deactivated successfully. Jan 30 15:31:34.845964 systemd[1]: session-16.scope: Deactivated successfully. Jan 30 15:31:34.847801 systemd-logind[1456]: Session 16 logged out. Waiting for processes to exit. Jan 30 15:31:34.849174 systemd-logind[1456]: Removed session 16. Jan 30 15:31:35.010952 systemd[1]: Started sshd@16-168.119.58.127:22-139.178.68.195:51352.service - OpenSSH per-connection server daemon (139.178.68.195:51352). Jan 30 15:31:35.976925 sshd[4239]: Accepted publickey for core from 139.178.68.195 port 51352 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:35.980670 sshd[4239]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:35.989313 systemd-logind[1456]: New session 17 of user core. Jan 30 15:31:35.997065 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 30 15:31:36.452636 update_engine[1458]: I20250130 15:31:36.452201 1458 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jan 30 15:31:36.453458 update_engine[1458]: I20250130 15:31:36.453284 1458 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jan 30 15:31:36.453756 update_engine[1458]: I20250130 15:31:36.453624 1458 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jan 30 15:31:36.455166 update_engine[1458]: E20250130 15:31:36.454958 1458 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jan 30 15:31:36.455166 update_engine[1458]: I20250130 15:31:36.455071 1458 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Jan 30 15:31:36.731625 sshd[4239]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:36.739415 systemd[1]: sshd@16-168.119.58.127:22-139.178.68.195:51352.service: Deactivated successfully. Jan 30 15:31:36.744648 systemd[1]: session-17.scope: Deactivated successfully. Jan 30 15:31:36.746058 systemd-logind[1456]: Session 17 logged out. Waiting for processes to exit. Jan 30 15:31:36.747175 systemd-logind[1456]: Removed session 17. Jan 30 15:31:41.918327 systemd[1]: Started sshd@17-168.119.58.127:22-139.178.68.195:51366.service - OpenSSH per-connection server daemon (139.178.68.195:51366). Jan 30 15:31:42.905733 sshd[4255]: Accepted publickey for core from 139.178.68.195 port 51366 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:42.908911 sshd[4255]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:42.914149 systemd-logind[1456]: New session 18 of user core. Jan 30 15:31:42.918781 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 30 15:31:43.668934 sshd[4255]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:43.676552 systemd[1]: sshd@17-168.119.58.127:22-139.178.68.195:51366.service: Deactivated successfully. Jan 30 15:31:43.680375 systemd[1]: session-18.scope: Deactivated successfully. Jan 30 15:31:43.682339 systemd-logind[1456]: Session 18 logged out. Waiting for processes to exit. Jan 30 15:31:43.684277 systemd-logind[1456]: Removed session 18. Jan 30 15:31:46.454607 update_engine[1458]: I20250130 15:31:46.454366 1458 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jan 30 15:31:46.455119 update_engine[1458]: I20250130 15:31:46.454849 1458 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jan 30 15:31:46.455255 update_engine[1458]: I20250130 15:31:46.455201 1458 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jan 30 15:31:46.456114 update_engine[1458]: E20250130 15:31:46.456051 1458 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jan 30 15:31:46.456248 update_engine[1458]: I20250130 15:31:46.456132 1458 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Jan 30 15:31:46.456248 update_engine[1458]: I20250130 15:31:46.456146 1458 omaha_request_action.cc:617] Omaha request response: Jan 30 15:31:46.456314 update_engine[1458]: E20250130 15:31:46.456259 1458 omaha_request_action.cc:636] Omaha request network transfer failed. Jan 30 15:31:46.456314 update_engine[1458]: I20250130 15:31:46.456281 1458 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Jan 30 15:31:46.456314 update_engine[1458]: I20250130 15:31:46.456288 1458 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Jan 30 15:31:46.456314 update_engine[1458]: I20250130 15:31:46.456295 1458 update_attempter.cc:306] Processing Done. Jan 30 15:31:46.456314 update_engine[1458]: E20250130 15:31:46.456311 1458 update_attempter.cc:619] Update failed. Jan 30 15:31:46.456748 update_engine[1458]: I20250130 15:31:46.456318 1458 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Jan 30 15:31:46.456748 update_engine[1458]: I20250130 15:31:46.456324 1458 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Jan 30 15:31:46.456748 update_engine[1458]: I20250130 15:31:46.456331 1458 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Jan 30 15:31:46.456748 update_engine[1458]: I20250130 15:31:46.456406 1458 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Jan 30 15:31:46.456748 update_engine[1458]: I20250130 15:31:46.456430 1458 omaha_request_action.cc:271] Posting an Omaha request to disabled Jan 30 15:31:46.456748 update_engine[1458]: I20250130 15:31:46.456437 1458 omaha_request_action.cc:272] Request: Jan 30 15:31:46.456748 update_engine[1458]: Jan 30 15:31:46.456748 update_engine[1458]: Jan 30 15:31:46.456748 update_engine[1458]: Jan 30 15:31:46.456748 update_engine[1458]: Jan 30 15:31:46.456748 update_engine[1458]: Jan 30 15:31:46.456748 update_engine[1458]: Jan 30 15:31:46.456748 update_engine[1458]: I20250130 15:31:46.456444 1458 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Jan 30 15:31:46.457134 locksmithd[1501]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Jan 30 15:31:46.457464 update_engine[1458]: I20250130 15:31:46.456843 1458 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Jan 30 15:31:46.457464 update_engine[1458]: I20250130 15:31:46.457041 1458 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Jan 30 15:31:46.457746 update_engine[1458]: E20250130 15:31:46.457690 1458 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Jan 30 15:31:46.457801 update_engine[1458]: I20250130 15:31:46.457753 1458 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Jan 30 15:31:46.457801 update_engine[1458]: I20250130 15:31:46.457765 1458 omaha_request_action.cc:617] Omaha request response: Jan 30 15:31:46.457801 update_engine[1458]: I20250130 15:31:46.457773 1458 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Jan 30 15:31:46.457801 update_engine[1458]: I20250130 15:31:46.457780 1458 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Jan 30 15:31:46.457801 update_engine[1458]: I20250130 15:31:46.457787 1458 update_attempter.cc:306] Processing Done. Jan 30 15:31:46.457801 update_engine[1458]: I20250130 15:31:46.457793 1458 update_attempter.cc:310] Error event sent. Jan 30 15:31:46.458158 update_engine[1458]: I20250130 15:31:46.457804 1458 update_check_scheduler.cc:74] Next update check in 48m13s Jan 30 15:31:46.458288 locksmithd[1501]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Jan 30 15:31:48.835904 systemd[1]: Started sshd@18-168.119.58.127:22-139.178.68.195:39196.service - OpenSSH per-connection server daemon (139.178.68.195:39196). Jan 30 15:31:49.808886 sshd[4271]: Accepted publickey for core from 139.178.68.195 port 39196 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:49.810644 sshd[4271]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:49.815489 systemd-logind[1456]: New session 19 of user core. Jan 30 15:31:49.822929 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 30 15:31:50.550228 sshd[4271]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:50.554835 systemd[1]: sshd@18-168.119.58.127:22-139.178.68.195:39196.service: Deactivated successfully. Jan 30 15:31:50.557045 systemd[1]: session-19.scope: Deactivated successfully. Jan 30 15:31:50.558673 systemd-logind[1456]: Session 19 logged out. Waiting for processes to exit. Jan 30 15:31:50.560220 systemd-logind[1456]: Removed session 19. Jan 30 15:31:50.726328 systemd[1]: Started sshd@19-168.119.58.127:22-139.178.68.195:39212.service - OpenSSH per-connection server daemon (139.178.68.195:39212). Jan 30 15:31:51.698060 sshd[4284]: Accepted publickey for core from 139.178.68.195 port 39212 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:51.701412 sshd[4284]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:51.708489 systemd-logind[1456]: New session 20 of user core. Jan 30 15:31:51.724911 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 30 15:31:54.427588 kubelet[2704]: I0130 15:31:54.425596 2704 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-668d6bf9bc-zxv9q" podStartSLOduration=337.425565573 podStartE2EDuration="5m37.425565573s" podCreationTimestamp="2025-01-30 15:26:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 15:26:39.0464908 +0000 UTC m=+26.337904630" watchObservedRunningTime="2025-01-30 15:31:54.425565573 +0000 UTC m=+341.716979403" Jan 30 15:31:54.438293 containerd[1476]: time="2025-01-30T15:31:54.438213405Z" level=info msg="StopContainer for \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\" with timeout 30 (s)" Jan 30 15:31:54.440980 containerd[1476]: time="2025-01-30T15:31:54.439842450Z" level=info msg="Stop container \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\" with signal terminated" Jan 30 15:31:54.464941 systemd[1]: run-containerd-runc-k8s.io-e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1-runc.I5Ux3t.mount: Deactivated successfully. Jan 30 15:31:54.466362 systemd[1]: cri-containerd-0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c.scope: Deactivated successfully. Jan 30 15:31:54.482094 containerd[1476]: time="2025-01-30T15:31:54.481776597Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 30 15:31:54.497348 containerd[1476]: time="2025-01-30T15:31:54.497299517Z" level=info msg="StopContainer for \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\" with timeout 2 (s)" Jan 30 15:31:54.499203 containerd[1476]: time="2025-01-30T15:31:54.499077602Z" level=info msg="Stop container \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\" with signal terminated" Jan 30 15:31:54.513651 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c-rootfs.mount: Deactivated successfully. Jan 30 15:31:54.515228 systemd-networkd[1371]: lxc_health: Link DOWN Jan 30 15:31:54.515241 systemd-networkd[1371]: lxc_health: Lost carrier Jan 30 15:31:54.534834 containerd[1476]: time="2025-01-30T15:31:54.534387852Z" level=info msg="shim disconnected" id=0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c namespace=k8s.io Jan 30 15:31:54.534834 containerd[1476]: time="2025-01-30T15:31:54.534464813Z" level=warning msg="cleaning up after shim disconnected" id=0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c namespace=k8s.io Jan 30 15:31:54.534834 containerd[1476]: time="2025-01-30T15:31:54.534478693Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:31:54.536166 systemd[1]: cri-containerd-e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1.scope: Deactivated successfully. Jan 30 15:31:54.536514 systemd[1]: cri-containerd-e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1.scope: Consumed 8.356s CPU time. Jan 30 15:31:54.565930 containerd[1476]: time="2025-01-30T15:31:54.565849733Z" level=info msg="StopContainer for \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\" returns successfully" Jan 30 15:31:54.568805 containerd[1476]: time="2025-01-30T15:31:54.568737141Z" level=info msg="StopPodSandbox for \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\"" Jan 30 15:31:54.568971 containerd[1476]: time="2025-01-30T15:31:54.568811941Z" level=info msg="Container to stop \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 15:31:54.572177 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075-shm.mount: Deactivated successfully. Jan 30 15:31:54.585690 systemd[1]: cri-containerd-fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075.scope: Deactivated successfully. Jan 30 15:31:54.596400 containerd[1476]: time="2025-01-30T15:31:54.596298491Z" level=info msg="shim disconnected" id=e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1 namespace=k8s.io Jan 30 15:31:54.596400 containerd[1476]: time="2025-01-30T15:31:54.596384851Z" level=warning msg="cleaning up after shim disconnected" id=e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1 namespace=k8s.io Jan 30 15:31:54.596400 containerd[1476]: time="2025-01-30T15:31:54.596394292Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:31:54.622229 containerd[1476]: time="2025-01-30T15:31:54.622180838Z" level=info msg="StopContainer for \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\" returns successfully" Jan 30 15:31:54.623325 containerd[1476]: time="2025-01-30T15:31:54.622754559Z" level=info msg="StopPodSandbox for \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\"" Jan 30 15:31:54.623325 containerd[1476]: time="2025-01-30T15:31:54.622810719Z" level=info msg="Container to stop \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 15:31:54.623325 containerd[1476]: time="2025-01-30T15:31:54.622823759Z" level=info msg="Container to stop \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 15:31:54.623325 containerd[1476]: time="2025-01-30T15:31:54.622833839Z" level=info msg="Container to stop \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 15:31:54.623325 containerd[1476]: time="2025-01-30T15:31:54.622844199Z" level=info msg="Container to stop \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 15:31:54.623325 containerd[1476]: time="2025-01-30T15:31:54.622853799Z" level=info msg="Container to stop \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 15:31:54.623325 containerd[1476]: time="2025-01-30T15:31:54.623092240Z" level=info msg="shim disconnected" id=fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075 namespace=k8s.io Jan 30 15:31:54.623325 containerd[1476]: time="2025-01-30T15:31:54.623174800Z" level=warning msg="cleaning up after shim disconnected" id=fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075 namespace=k8s.io Jan 30 15:31:54.623325 containerd[1476]: time="2025-01-30T15:31:54.623188600Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:31:54.633074 systemd[1]: cri-containerd-474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac.scope: Deactivated successfully. Jan 30 15:31:54.649518 containerd[1476]: time="2025-01-30T15:31:54.649351947Z" level=info msg="TearDown network for sandbox \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\" successfully" Jan 30 15:31:54.649518 containerd[1476]: time="2025-01-30T15:31:54.649391988Z" level=info msg="StopPodSandbox for \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\" returns successfully" Jan 30 15:31:54.679499 containerd[1476]: time="2025-01-30T15:31:54.679024744Z" level=info msg="shim disconnected" id=474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac namespace=k8s.io Jan 30 15:31:54.679499 containerd[1476]: time="2025-01-30T15:31:54.679216544Z" level=warning msg="cleaning up after shim disconnected" id=474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac namespace=k8s.io Jan 30 15:31:54.679499 containerd[1476]: time="2025-01-30T15:31:54.679320064Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:31:54.693941 containerd[1476]: time="2025-01-30T15:31:54.693826662Z" level=info msg="TearDown network for sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" successfully" Jan 30 15:31:54.693941 containerd[1476]: time="2025-01-30T15:31:54.693894502Z" level=info msg="StopPodSandbox for \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" returns successfully" Jan 30 15:31:54.789593 kubelet[2704]: I0130 15:31:54.788629 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-run\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.789593 kubelet[2704]: I0130 15:31:54.788684 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.789593 kubelet[2704]: I0130 15:31:54.788726 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/bd67b7e3-7bf4-4174-8e95-118b465075f9-clustermesh-secrets\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.789593 kubelet[2704]: I0130 15:31:54.788765 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-host-proc-sys-net\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.789593 kubelet[2704]: I0130 15:31:54.788805 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-config-path\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.789593 kubelet[2704]: I0130 15:31:54.788846 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-lib-modules\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790238 kubelet[2704]: I0130 15:31:54.788878 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-cgroup\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790238 kubelet[2704]: I0130 15:31:54.788919 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-bpf-maps\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790238 kubelet[2704]: I0130 15:31:54.788960 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-hubble-tls\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790238 kubelet[2704]: I0130 15:31:54.788995 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-xtables-lock\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790238 kubelet[2704]: I0130 15:31:54.789027 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-hostproc\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790238 kubelet[2704]: I0130 15:31:54.789067 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wqfdm\" (UniqueName: \"kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-kube-api-access-wqfdm\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790705 kubelet[2704]: I0130 15:31:54.789099 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cni-path\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790705 kubelet[2704]: I0130 15:31:54.789216 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54b9t\" (UniqueName: \"kubernetes.io/projected/5263f0fa-2944-49ce-8475-80af8b5aa5e2-kube-api-access-54b9t\") pod \"5263f0fa-2944-49ce-8475-80af8b5aa5e2\" (UID: \"5263f0fa-2944-49ce-8475-80af8b5aa5e2\") " Jan 30 15:31:54.790705 kubelet[2704]: I0130 15:31:54.789272 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-etc-cni-netd\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790705 kubelet[2704]: I0130 15:31:54.789307 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-host-proc-sys-kernel\") pod \"bd67b7e3-7bf4-4174-8e95-118b465075f9\" (UID: \"bd67b7e3-7bf4-4174-8e95-118b465075f9\") " Jan 30 15:31:54.790705 kubelet[2704]: I0130 15:31:54.789349 2704 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/5263f0fa-2944-49ce-8475-80af8b5aa5e2-cilium-config-path\") pod \"5263f0fa-2944-49ce-8475-80af8b5aa5e2\" (UID: \"5263f0fa-2944-49ce-8475-80af8b5aa5e2\") " Jan 30 15:31:54.790705 kubelet[2704]: I0130 15:31:54.789430 2704 reconciler_common.go:299] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-run\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.791035 kubelet[2704]: I0130 15:31:54.789660 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.793208 kubelet[2704]: I0130 15:31:54.791912 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.793208 kubelet[2704]: I0130 15:31:54.791983 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.793208 kubelet[2704]: I0130 15:31:54.792000 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.793208 kubelet[2704]: I0130 15:31:54.792327 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cni-path" (OuterVolumeSpecName: "cni-path") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.793208 kubelet[2704]: I0130 15:31:54.792366 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.793432 kubelet[2704]: I0130 15:31:54.792382 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-hostproc" (OuterVolumeSpecName: "hostproc") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.793432 kubelet[2704]: I0130 15:31:54.792851 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jan 30 15:31:54.793842 kubelet[2704]: I0130 15:31:54.793765 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.794170 kubelet[2704]: I0130 15:31:54.794103 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Jan 30 15:31:54.796531 kubelet[2704]: I0130 15:31:54.796447 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd67b7e3-7bf4-4174-8e95-118b465075f9-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Jan 30 15:31:54.798684 kubelet[2704]: I0130 15:31:54.798639 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5263f0fa-2944-49ce-8475-80af8b5aa5e2-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "5263f0fa-2944-49ce-8475-80af8b5aa5e2" (UID: "5263f0fa-2944-49ce-8475-80af8b5aa5e2"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Jan 30 15:31:54.799006 kubelet[2704]: I0130 15:31:54.798972 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5263f0fa-2944-49ce-8475-80af8b5aa5e2-kube-api-access-54b9t" (OuterVolumeSpecName: "kube-api-access-54b9t") pod "5263f0fa-2944-49ce-8475-80af8b5aa5e2" (UID: "5263f0fa-2944-49ce-8475-80af8b5aa5e2"). InnerVolumeSpecName "kube-api-access-54b9t". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jan 30 15:31:54.799220 kubelet[2704]: I0130 15:31:54.799162 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jan 30 15:31:54.799693 kubelet[2704]: I0130 15:31:54.799630 2704 operation_generator.go:780] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-kube-api-access-wqfdm" (OuterVolumeSpecName: "kube-api-access-wqfdm") pod "bd67b7e3-7bf4-4174-8e95-118b465075f9" (UID: "bd67b7e3-7bf4-4174-8e95-118b465075f9"). InnerVolumeSpecName "kube-api-access-wqfdm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Jan 30 15:31:54.822460 kubelet[2704]: I0130 15:31:54.822398 2704 scope.go:117] "RemoveContainer" containerID="e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1" Jan 30 15:31:54.836802 containerd[1476]: time="2025-01-30T15:31:54.836061067Z" level=info msg="RemoveContainer for \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\"" Jan 30 15:31:54.836499 systemd[1]: Removed slice kubepods-burstable-podbd67b7e3_7bf4_4174_8e95_118b465075f9.slice - libcontainer container kubepods-burstable-podbd67b7e3_7bf4_4174_8e95_118b465075f9.slice. Jan 30 15:31:54.837197 systemd[1]: kubepods-burstable-podbd67b7e3_7bf4_4174_8e95_118b465075f9.slice: Consumed 8.454s CPU time. Jan 30 15:31:54.844145 containerd[1476]: time="2025-01-30T15:31:54.843923167Z" level=info msg="RemoveContainer for \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\" returns successfully" Jan 30 15:31:54.852689 systemd[1]: Removed slice kubepods-besteffort-pod5263f0fa_2944_49ce_8475_80af8b5aa5e2.slice - libcontainer container kubepods-besteffort-pod5263f0fa_2944_49ce_8475_80af8b5aa5e2.slice. Jan 30 15:31:54.854107 kubelet[2704]: I0130 15:31:54.853018 2704 scope.go:117] "RemoveContainer" containerID="ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3" Jan 30 15:31:54.860937 containerd[1476]: time="2025-01-30T15:31:54.860195209Z" level=info msg="RemoveContainer for \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\"" Jan 30 15:31:54.866876 containerd[1476]: time="2025-01-30T15:31:54.866818106Z" level=info msg="RemoveContainer for \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\" returns successfully" Jan 30 15:31:54.867429 kubelet[2704]: I0130 15:31:54.867285 2704 scope.go:117] "RemoveContainer" containerID="257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc" Jan 30 15:31:54.869526 containerd[1476]: time="2025-01-30T15:31:54.869479592Z" level=info msg="RemoveContainer for \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\"" Jan 30 15:31:54.877177 containerd[1476]: time="2025-01-30T15:31:54.876437250Z" level=info msg="RemoveContainer for \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\" returns successfully" Jan 30 15:31:54.882595 kubelet[2704]: I0130 15:31:54.877777 2704 scope.go:117] "RemoveContainer" containerID="6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3" Jan 30 15:31:54.885086 containerd[1476]: time="2025-01-30T15:31:54.885046592Z" level=info msg="RemoveContainer for \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\"" Jan 30 15:31:54.890157 kubelet[2704]: I0130 15:31:54.890093 2704 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-54b9t\" (UniqueName: \"kubernetes.io/projected/5263f0fa-2944-49ce-8475-80af8b5aa5e2-kube-api-access-54b9t\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.890378 kubelet[2704]: I0130 15:31:54.890358 2704 reconciler_common.go:299] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-hostproc\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.890452 kubelet[2704]: I0130 15:31:54.890437 2704 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wqfdm\" (UniqueName: \"kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-kube-api-access-wqfdm\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.892170 kubelet[2704]: I0130 15:31:54.892144 2704 reconciler_common.go:299] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cni-path\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.892331 kubelet[2704]: I0130 15:31:54.892317 2704 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-host-proc-sys-kernel\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.892391 kubelet[2704]: I0130 15:31:54.892380 2704 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/5263f0fa-2944-49ce-8475-80af8b5aa5e2-cilium-config-path\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.892509 containerd[1476]: time="2025-01-30T15:31:54.892468491Z" level=info msg="RemoveContainer for \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\" returns successfully" Jan 30 15:31:54.892657 kubelet[2704]: I0130 15:31:54.892639 2704 reconciler_common.go:299] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-etc-cni-netd\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.892723 kubelet[2704]: I0130 15:31:54.892712 2704 reconciler_common.go:299] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-config-path\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.892791 kubelet[2704]: I0130 15:31:54.892778 2704 reconciler_common.go:299] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/bd67b7e3-7bf4-4174-8e95-118b465075f9-clustermesh-secrets\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.892850 kubelet[2704]: I0130 15:31:54.892840 2704 reconciler_common.go:299] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-host-proc-sys-net\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.893019 kubelet[2704]: I0130 15:31:54.893003 2704 reconciler_common.go:299] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-lib-modules\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.893151 kubelet[2704]: I0130 15:31:54.893088 2704 reconciler_common.go:299] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/bd67b7e3-7bf4-4174-8e95-118b465075f9-hubble-tls\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.893231 kubelet[2704]: I0130 15:31:54.893217 2704 reconciler_common.go:299] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-cilium-cgroup\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.893448 kubelet[2704]: I0130 15:31:54.893430 2704 reconciler_common.go:299] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-bpf-maps\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.893552 kubelet[2704]: I0130 15:31:54.893525 2704 reconciler_common.go:299] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/bd67b7e3-7bf4-4174-8e95-118b465075f9-xtables-lock\") on node \"ci-4081-3-0-a-5514912dcd\" DevicePath \"\"" Jan 30 15:31:54.894050 kubelet[2704]: I0130 15:31:54.894024 2704 scope.go:117] "RemoveContainer" containerID="9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b" Jan 30 15:31:54.899604 containerd[1476]: time="2025-01-30T15:31:54.898820748Z" level=info msg="RemoveContainer for \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\"" Jan 30 15:31:54.904526 containerd[1476]: time="2025-01-30T15:31:54.904473602Z" level=info msg="RemoveContainer for \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\" returns successfully" Jan 30 15:31:54.905816 kubelet[2704]: I0130 15:31:54.905732 2704 scope.go:117] "RemoveContainer" containerID="e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1" Jan 30 15:31:54.906752 containerd[1476]: time="2025-01-30T15:31:54.906435527Z" level=error msg="ContainerStatus for \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\": not found" Jan 30 15:31:54.907132 kubelet[2704]: E0130 15:31:54.907083 2704 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\": not found" containerID="e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1" Jan 30 15:31:54.907249 kubelet[2704]: I0130 15:31:54.907158 2704 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1"} err="failed to get container status \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\": rpc error: code = NotFound desc = an error occurred when try to find container \"e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1\": not found" Jan 30 15:31:54.907295 kubelet[2704]: I0130 15:31:54.907254 2704 scope.go:117] "RemoveContainer" containerID="ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3" Jan 30 15:31:54.907610 containerd[1476]: time="2025-01-30T15:31:54.907511850Z" level=error msg="ContainerStatus for \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\": not found" Jan 30 15:31:54.907776 kubelet[2704]: E0130 15:31:54.907726 2704 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\": not found" containerID="ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3" Jan 30 15:31:54.907831 kubelet[2704]: I0130 15:31:54.907784 2704 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3"} err="failed to get container status \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\": rpc error: code = NotFound desc = an error occurred when try to find container \"ad612b10e4a2a6313db832683aa1303ef0b160f4f9aca572cbbe0c832b2642c3\": not found" Jan 30 15:31:54.907831 kubelet[2704]: I0130 15:31:54.907805 2704 scope.go:117] "RemoveContainer" containerID="257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc" Jan 30 15:31:54.908295 containerd[1476]: time="2025-01-30T15:31:54.908112372Z" level=error msg="ContainerStatus for \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\": not found" Jan 30 15:31:54.908422 kubelet[2704]: E0130 15:31:54.908346 2704 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\": not found" containerID="257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc" Jan 30 15:31:54.908422 kubelet[2704]: I0130 15:31:54.908370 2704 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc"} err="failed to get container status \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\": rpc error: code = NotFound desc = an error occurred when try to find container \"257c995b290abc556382dc0405b039251d35699879090d6d9e7d2f4fa8aececc\": not found" Jan 30 15:31:54.908422 kubelet[2704]: I0130 15:31:54.908386 2704 scope.go:117] "RemoveContainer" containerID="6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3" Jan 30 15:31:54.908771 containerd[1476]: time="2025-01-30T15:31:54.908581253Z" level=error msg="ContainerStatus for \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\": not found" Jan 30 15:31:54.909105 kubelet[2704]: E0130 15:31:54.908898 2704 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\": not found" containerID="6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3" Jan 30 15:31:54.909105 kubelet[2704]: I0130 15:31:54.908935 2704 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3"} err="failed to get container status \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\": rpc error: code = NotFound desc = an error occurred when try to find container \"6a39529cfc3fccd6f9ee722d43775277e2fc1f958cf664c08aeedc8b18a591c3\": not found" Jan 30 15:31:54.909105 kubelet[2704]: I0130 15:31:54.908953 2704 scope.go:117] "RemoveContainer" containerID="9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b" Jan 30 15:31:54.909522 containerd[1476]: time="2025-01-30T15:31:54.909430215Z" level=error msg="ContainerStatus for \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\": not found" Jan 30 15:31:54.909654 kubelet[2704]: E0130 15:31:54.909596 2704 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\": not found" containerID="9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b" Jan 30 15:31:54.909654 kubelet[2704]: I0130 15:31:54.909617 2704 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b"} err="failed to get container status \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\": rpc error: code = NotFound desc = an error occurred when try to find container \"9a1c2373dfcdb497d0343d76958c7259ab8204e884815799992e557d037bc75b\": not found" Jan 30 15:31:54.909654 kubelet[2704]: I0130 15:31:54.909646 2704 scope.go:117] "RemoveContainer" containerID="0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c" Jan 30 15:31:54.911106 containerd[1476]: time="2025-01-30T15:31:54.911069139Z" level=info msg="RemoveContainer for \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\"" Jan 30 15:31:54.915203 containerd[1476]: time="2025-01-30T15:31:54.914961909Z" level=info msg="RemoveContainer for \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\" returns successfully" Jan 30 15:31:54.915521 kubelet[2704]: I0130 15:31:54.915456 2704 scope.go:117] "RemoveContainer" containerID="0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c" Jan 30 15:31:54.915995 containerd[1476]: time="2025-01-30T15:31:54.915912672Z" level=error msg="ContainerStatus for \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\": not found" Jan 30 15:31:54.916320 kubelet[2704]: E0130 15:31:54.916273 2704 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\": not found" containerID="0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c" Jan 30 15:31:54.916430 kubelet[2704]: I0130 15:31:54.916308 2704 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c"} err="failed to get container status \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\": rpc error: code = NotFound desc = an error occurred when try to find container \"0da57fb63d16b76e7c229060b833a9d30bcd251514adc6ed8b9b4e9fea95541c\": not found" Jan 30 15:31:55.448511 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e27ba2521826607070f8375c4e7daeeb89c687165e47fa53e7c93dbc22c779f1-rootfs.mount: Deactivated successfully. Jan 30 15:31:55.448864 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac-rootfs.mount: Deactivated successfully. Jan 30 15:31:55.449002 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac-shm.mount: Deactivated successfully. Jan 30 15:31:55.449196 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075-rootfs.mount: Deactivated successfully. Jan 30 15:31:55.449352 systemd[1]: var-lib-kubelet-pods-bd67b7e3\x2d7bf4\x2d4174\x2d8e95\x2d118b465075f9-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dwqfdm.mount: Deactivated successfully. Jan 30 15:31:55.449492 systemd[1]: var-lib-kubelet-pods-5263f0fa\x2d2944\x2d49ce\x2d8475\x2d80af8b5aa5e2-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d54b9t.mount: Deactivated successfully. Jan 30 15:31:55.449657 systemd[1]: var-lib-kubelet-pods-bd67b7e3\x2d7bf4\x2d4174\x2d8e95\x2d118b465075f9-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Jan 30 15:31:55.449718 systemd[1]: var-lib-kubelet-pods-bd67b7e3\x2d7bf4\x2d4174\x2d8e95\x2d118b465075f9-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Jan 30 15:31:56.528250 sshd[4284]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:56.532694 systemd-logind[1456]: Session 20 logged out. Waiting for processes to exit. Jan 30 15:31:56.532967 systemd[1]: sshd@19-168.119.58.127:22-139.178.68.195:39212.service: Deactivated successfully. Jan 30 15:31:56.535807 systemd[1]: session-20.scope: Deactivated successfully. Jan 30 15:31:56.536066 systemd[1]: session-20.scope: Consumed 1.568s CPU time. Jan 30 15:31:56.538626 systemd-logind[1456]: Removed session 20. Jan 30 15:31:56.716978 systemd[1]: Started sshd@20-168.119.58.127:22-139.178.68.195:52540.service - OpenSSH per-connection server daemon (139.178.68.195:52540). Jan 30 15:31:56.827697 kubelet[2704]: I0130 15:31:56.827584 2704 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5263f0fa-2944-49ce-8475-80af8b5aa5e2" path="/var/lib/kubelet/pods/5263f0fa-2944-49ce-8475-80af8b5aa5e2/volumes" Jan 30 15:31:56.828079 kubelet[2704]: I0130 15:31:56.828057 2704 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd67b7e3-7bf4-4174-8e95-118b465075f9" path="/var/lib/kubelet/pods/bd67b7e3-7bf4-4174-8e95-118b465075f9/volumes" Jan 30 15:31:57.706780 sshd[4445]: Accepted publickey for core from 139.178.68.195 port 52540 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:31:57.708572 sshd[4445]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:31:57.714394 systemd-logind[1456]: New session 21 of user core. Jan 30 15:31:57.721832 systemd[1]: Started session-21.scope - Session 21 of User core. Jan 30 15:31:58.033199 kubelet[2704]: E0130 15:31:58.033142 2704 kubelet.go:3008] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jan 30 15:31:59.303346 kubelet[2704]: I0130 15:31:59.303302 2704 memory_manager.go:355] "RemoveStaleState removing state" podUID="bd67b7e3-7bf4-4174-8e95-118b465075f9" containerName="cilium-agent" Jan 30 15:31:59.303346 kubelet[2704]: I0130 15:31:59.303333 2704 memory_manager.go:355] "RemoveStaleState removing state" podUID="5263f0fa-2944-49ce-8475-80af8b5aa5e2" containerName="cilium-operator" Jan 30 15:31:59.312898 systemd[1]: Created slice kubepods-burstable-pod3f4f4074_d671_4cca_b361_e0b0f48a5a3c.slice - libcontainer container kubepods-burstable-pod3f4f4074_d671_4cca_b361_e0b0f48a5a3c.slice. Jan 30 15:31:59.423229 kubelet[2704]: I0130 15:31:59.422591 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-lib-modules\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423229 kubelet[2704]: I0130 15:31:59.422654 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-cilium-ipsec-secrets\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423229 kubelet[2704]: I0130 15:31:59.422691 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhm8s\" (UniqueName: \"kubernetes.io/projected/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-kube-api-access-mhm8s\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423229 kubelet[2704]: I0130 15:31:59.422722 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-hostproc\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423229 kubelet[2704]: I0130 15:31:59.422752 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-host-proc-sys-kernel\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423229 kubelet[2704]: I0130 15:31:59.422797 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-bpf-maps\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423769 kubelet[2704]: I0130 15:31:59.422833 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-cni-path\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423769 kubelet[2704]: I0130 15:31:59.422864 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-cilium-config-path\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423769 kubelet[2704]: I0130 15:31:59.422893 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-xtables-lock\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423769 kubelet[2704]: I0130 15:31:59.422919 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-cilium-run\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423769 kubelet[2704]: I0130 15:31:59.422947 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-cilium-cgroup\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.423769 kubelet[2704]: I0130 15:31:59.422974 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-etc-cni-netd\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.424024 kubelet[2704]: I0130 15:31:59.423081 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-clustermesh-secrets\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.424024 kubelet[2704]: I0130 15:31:59.423191 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-host-proc-sys-net\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.424024 kubelet[2704]: I0130 15:31:59.423223 2704 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/3f4f4074-d671-4cca-b361-e0b0f48a5a3c-hubble-tls\") pod \"cilium-nd2zh\" (UID: \"3f4f4074-d671-4cca-b361-e0b0f48a5a3c\") " pod="kube-system/cilium-nd2zh" Jan 30 15:31:59.486469 sshd[4445]: pam_unix(sshd:session): session closed for user core Jan 30 15:31:59.491579 systemd[1]: sshd@20-168.119.58.127:22-139.178.68.195:52540.service: Deactivated successfully. Jan 30 15:31:59.494004 systemd[1]: session-21.scope: Deactivated successfully. Jan 30 15:31:59.496204 systemd-logind[1456]: Session 21 logged out. Waiting for processes to exit. Jan 30 15:31:59.497273 systemd-logind[1456]: Removed session 21. Jan 30 15:31:59.618184 containerd[1476]: time="2025-01-30T15:31:59.617810124Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-nd2zh,Uid:3f4f4074-d671-4cca-b361-e0b0f48a5a3c,Namespace:kube-system,Attempt:0,}" Jan 30 15:31:59.648834 containerd[1476]: time="2025-01-30T15:31:59.646273287Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 15:31:59.648834 containerd[1476]: time="2025-01-30T15:31:59.646763008Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 15:31:59.648834 containerd[1476]: time="2025-01-30T15:31:59.646781328Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:31:59.648834 containerd[1476]: time="2025-01-30T15:31:59.646930688Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 15:31:59.659997 systemd[1]: Started sshd@21-168.119.58.127:22-139.178.68.195:52554.service - OpenSSH per-connection server daemon (139.178.68.195:52554). Jan 30 15:31:59.666377 systemd[1]: Started cri-containerd-e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb.scope - libcontainer container e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb. Jan 30 15:31:59.692435 containerd[1476]: time="2025-01-30T15:31:59.692394260Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-nd2zh,Uid:3f4f4074-d671-4cca-b361-e0b0f48a5a3c,Namespace:kube-system,Attempt:0,} returns sandbox id \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\"" Jan 30 15:31:59.697453 containerd[1476]: time="2025-01-30T15:31:59.697391235Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Jan 30 15:31:59.708801 containerd[1476]: time="2025-01-30T15:31:59.708723787Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"3eaf6493d6fe31ebf10e7cd3ca5329d03c4e30d8d67a17874d261da73cd1a9a8\"" Jan 30 15:31:59.710582 containerd[1476]: time="2025-01-30T15:31:59.709634750Z" level=info msg="StartContainer for \"3eaf6493d6fe31ebf10e7cd3ca5329d03c4e30d8d67a17874d261da73cd1a9a8\"" Jan 30 15:31:59.738745 systemd[1]: Started cri-containerd-3eaf6493d6fe31ebf10e7cd3ca5329d03c4e30d8d67a17874d261da73cd1a9a8.scope - libcontainer container 3eaf6493d6fe31ebf10e7cd3ca5329d03c4e30d8d67a17874d261da73cd1a9a8. Jan 30 15:31:59.769648 containerd[1476]: time="2025-01-30T15:31:59.769522483Z" level=info msg="StartContainer for \"3eaf6493d6fe31ebf10e7cd3ca5329d03c4e30d8d67a17874d261da73cd1a9a8\" returns successfully" Jan 30 15:31:59.788717 systemd[1]: cri-containerd-3eaf6493d6fe31ebf10e7cd3ca5329d03c4e30d8d67a17874d261da73cd1a9a8.scope: Deactivated successfully. Jan 30 15:31:59.827193 containerd[1476]: time="2025-01-30T15:31:59.827064970Z" level=info msg="shim disconnected" id=3eaf6493d6fe31ebf10e7cd3ca5329d03c4e30d8d67a17874d261da73cd1a9a8 namespace=k8s.io Jan 30 15:31:59.827193 containerd[1476]: time="2025-01-30T15:31:59.827183370Z" level=warning msg="cleaning up after shim disconnected" id=3eaf6493d6fe31ebf10e7cd3ca5329d03c4e30d8d67a17874d261da73cd1a9a8 namespace=k8s.io Jan 30 15:31:59.827193 containerd[1476]: time="2025-01-30T15:31:59.827197330Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:31:59.863252 containerd[1476]: time="2025-01-30T15:31:59.862844154Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Jan 30 15:31:59.887042 containerd[1476]: time="2025-01-30T15:31:59.886875703Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"2a286d08d92c8c0833f955bed0c508ff348b29192bd2971dd10be513681d5e9a\"" Jan 30 15:31:59.888078 containerd[1476]: time="2025-01-30T15:31:59.888040987Z" level=info msg="StartContainer for \"2a286d08d92c8c0833f955bed0c508ff348b29192bd2971dd10be513681d5e9a\"" Jan 30 15:31:59.914781 systemd[1]: Started cri-containerd-2a286d08d92c8c0833f955bed0c508ff348b29192bd2971dd10be513681d5e9a.scope - libcontainer container 2a286d08d92c8c0833f955bed0c508ff348b29192bd2971dd10be513681d5e9a. Jan 30 15:31:59.953038 containerd[1476]: time="2025-01-30T15:31:59.952964334Z" level=info msg="StartContainer for \"2a286d08d92c8c0833f955bed0c508ff348b29192bd2971dd10be513681d5e9a\" returns successfully" Jan 30 15:31:59.961168 systemd[1]: cri-containerd-2a286d08d92c8c0833f955bed0c508ff348b29192bd2971dd10be513681d5e9a.scope: Deactivated successfully. Jan 30 15:31:59.997130 containerd[1476]: time="2025-01-30T15:31:59.996757701Z" level=info msg="shim disconnected" id=2a286d08d92c8c0833f955bed0c508ff348b29192bd2971dd10be513681d5e9a namespace=k8s.io Jan 30 15:31:59.997130 containerd[1476]: time="2025-01-30T15:31:59.996837862Z" level=warning msg="cleaning up after shim disconnected" id=2a286d08d92c8c0833f955bed0c508ff348b29192bd2971dd10be513681d5e9a namespace=k8s.io Jan 30 15:31:59.997130 containerd[1476]: time="2025-01-30T15:31:59.996856142Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:32:00.199893 kubelet[2704]: I0130 15:32:00.199173 2704 setters.go:602] "Node became not ready" node="ci-4081-3-0-a-5514912dcd" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-01-30T15:32:00Z","lastTransitionTime":"2025-01-30T15:32:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Jan 30 15:32:00.630348 sshd[4478]: Accepted publickey for core from 139.178.68.195 port 52554 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:32:00.633280 sshd[4478]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:32:00.638471 systemd-logind[1456]: New session 22 of user core. Jan 30 15:32:00.650935 systemd[1]: Started session-22.scope - Session 22 of User core. Jan 30 15:32:00.870056 containerd[1476]: time="2025-01-30T15:32:00.869259923Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Jan 30 15:32:00.904250 containerd[1476]: time="2025-01-30T15:32:00.904123906Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"79bd1a170d549a0c18685a91a028ea204cf815f0bd60005e0418ff3b581863b5\"" Jan 30 15:32:00.909654 containerd[1476]: time="2025-01-30T15:32:00.906931154Z" level=info msg="StartContainer for \"79bd1a170d549a0c18685a91a028ea204cf815f0bd60005e0418ff3b581863b5\"" Jan 30 15:32:00.952817 systemd[1]: Started cri-containerd-79bd1a170d549a0c18685a91a028ea204cf815f0bd60005e0418ff3b581863b5.scope - libcontainer container 79bd1a170d549a0c18685a91a028ea204cf815f0bd60005e0418ff3b581863b5. Jan 30 15:32:00.989659 containerd[1476]: time="2025-01-30T15:32:00.989576319Z" level=info msg="StartContainer for \"79bd1a170d549a0c18685a91a028ea204cf815f0bd60005e0418ff3b581863b5\" returns successfully" Jan 30 15:32:00.991989 systemd[1]: cri-containerd-79bd1a170d549a0c18685a91a028ea204cf815f0bd60005e0418ff3b581863b5.scope: Deactivated successfully. Jan 30 15:32:01.033599 containerd[1476]: time="2025-01-30T15:32:01.033491971Z" level=info msg="shim disconnected" id=79bd1a170d549a0c18685a91a028ea204cf815f0bd60005e0418ff3b581863b5 namespace=k8s.io Jan 30 15:32:01.033599 containerd[1476]: time="2025-01-30T15:32:01.033587531Z" level=warning msg="cleaning up after shim disconnected" id=79bd1a170d549a0c18685a91a028ea204cf815f0bd60005e0418ff3b581863b5 namespace=k8s.io Jan 30 15:32:01.033599 containerd[1476]: time="2025-01-30T15:32:01.033603091Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:32:01.304330 sshd[4478]: pam_unix(sshd:session): session closed for user core Jan 30 15:32:01.309508 systemd[1]: sshd@21-168.119.58.127:22-139.178.68.195:52554.service: Deactivated successfully. Jan 30 15:32:01.312408 systemd[1]: session-22.scope: Deactivated successfully. Jan 30 15:32:01.313503 systemd-logind[1456]: Session 22 logged out. Waiting for processes to exit. Jan 30 15:32:01.315446 systemd-logind[1456]: Removed session 22. Jan 30 15:32:01.482996 systemd[1]: Started sshd@22-168.119.58.127:22-139.178.68.195:52566.service - OpenSSH per-connection server daemon (139.178.68.195:52566). Jan 30 15:32:01.532863 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-79bd1a170d549a0c18685a91a028ea204cf815f0bd60005e0418ff3b581863b5-rootfs.mount: Deactivated successfully. Jan 30 15:32:01.887872 containerd[1476]: time="2025-01-30T15:32:01.887787513Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Jan 30 15:32:01.911395 containerd[1476]: time="2025-01-30T15:32:01.911263184Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"f6a868d3abe923800219c4eafe1192fcddac10c0a26ce17236b968853b773734\"" Jan 30 15:32:01.912679 containerd[1476]: time="2025-01-30T15:32:01.912140867Z" level=info msg="StartContainer for \"f6a868d3abe923800219c4eafe1192fcddac10c0a26ce17236b968853b773734\"" Jan 30 15:32:01.959852 systemd[1]: Started cri-containerd-f6a868d3abe923800219c4eafe1192fcddac10c0a26ce17236b968853b773734.scope - libcontainer container f6a868d3abe923800219c4eafe1192fcddac10c0a26ce17236b968853b773734. Jan 30 15:32:01.987781 systemd[1]: cri-containerd-f6a868d3abe923800219c4eafe1192fcddac10c0a26ce17236b968853b773734.scope: Deactivated successfully. Jan 30 15:32:01.991151 containerd[1476]: time="2025-01-30T15:32:01.989198340Z" level=info msg="StartContainer for \"f6a868d3abe923800219c4eafe1192fcddac10c0a26ce17236b968853b773734\" returns successfully" Jan 30 15:32:02.021391 containerd[1476]: time="2025-01-30T15:32:02.021312198Z" level=info msg="shim disconnected" id=f6a868d3abe923800219c4eafe1192fcddac10c0a26ce17236b968853b773734 namespace=k8s.io Jan 30 15:32:02.021882 containerd[1476]: time="2025-01-30T15:32:02.021679840Z" level=warning msg="cleaning up after shim disconnected" id=f6a868d3abe923800219c4eafe1192fcddac10c0a26ce17236b968853b773734 namespace=k8s.io Jan 30 15:32:02.021882 containerd[1476]: time="2025-01-30T15:32:02.021701320Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:32:02.465835 sshd[4689]: Accepted publickey for core from 139.178.68.195 port 52566 ssh2: RSA SHA256:sEmXhGFGlwd7KeRcv2oD/pODTHGZASfNUvhka9D+Bx0 Jan 30 15:32:02.468171 sshd[4689]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 15:32:02.474664 systemd-logind[1456]: New session 23 of user core. Jan 30 15:32:02.482907 systemd[1]: Started session-23.scope - Session 23 of User core. Jan 30 15:32:02.534206 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f6a868d3abe923800219c4eafe1192fcddac10c0a26ce17236b968853b773734-rootfs.mount: Deactivated successfully. Jan 30 15:32:02.892121 containerd[1476]: time="2025-01-30T15:32:02.892051926Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Jan 30 15:32:02.920410 containerd[1476]: time="2025-01-30T15:32:02.919932772Z" level=info msg="CreateContainer within sandbox \"e08a3e95ce6e74567b7e38a01d6e76048c3324818c4bf5e05d231271bad29dfb\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"ce5518dd00b7c7dcb7a34994a41fc15050dadba246f0396db92e594bc6597d06\"" Jan 30 15:32:02.929583 containerd[1476]: time="2025-01-30T15:32:02.928419838Z" level=info msg="StartContainer for \"ce5518dd00b7c7dcb7a34994a41fc15050dadba246f0396db92e594bc6597d06\"" Jan 30 15:32:02.969314 systemd[1]: Started cri-containerd-ce5518dd00b7c7dcb7a34994a41fc15050dadba246f0396db92e594bc6597d06.scope - libcontainer container ce5518dd00b7c7dcb7a34994a41fc15050dadba246f0396db92e594bc6597d06. Jan 30 15:32:03.031245 containerd[1476]: time="2025-01-30T15:32:03.031186837Z" level=info msg="StartContainer for \"ce5518dd00b7c7dcb7a34994a41fc15050dadba246f0396db92e594bc6597d06\" returns successfully" Jan 30 15:32:03.037415 kubelet[2704]: E0130 15:32:03.034514 2704 kubelet.go:3008] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Jan 30 15:32:03.446947 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce)) Jan 30 15:32:03.920837 kubelet[2704]: I0130 15:32:03.920759 2704 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-nd2zh" podStartSLOduration=4.920734078 podStartE2EDuration="4.920734078s" podCreationTimestamp="2025-01-30 15:31:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 15:32:03.919192473 +0000 UTC m=+351.210606343" watchObservedRunningTime="2025-01-30 15:32:03.920734078 +0000 UTC m=+351.212147988" Jan 30 15:32:06.484967 systemd-networkd[1371]: lxc_health: Link UP Jan 30 15:32:06.504987 systemd-networkd[1371]: lxc_health: Gained carrier Jan 30 15:32:07.824135 kubelet[2704]: E0130 15:32:07.824034 2704 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="kube-system/coredns-668d6bf9bc-zxv9q" podUID="acfc3ffa-2506-45ee-bf02-58d1a6b776af" Jan 30 15:32:08.059783 systemd-networkd[1371]: lxc_health: Gained IPv6LL Jan 30 15:32:11.989947 sshd[4689]: pam_unix(sshd:session): session closed for user core Jan 30 15:32:11.996687 systemd[1]: sshd@22-168.119.58.127:22-139.178.68.195:52566.service: Deactivated successfully. Jan 30 15:32:11.999571 systemd[1]: session-23.scope: Deactivated successfully. Jan 30 15:32:12.000955 systemd-logind[1456]: Session 23 logged out. Waiting for processes to exit. Jan 30 15:32:12.002159 systemd-logind[1456]: Removed session 23. Jan 30 15:32:12.874626 containerd[1476]: time="2025-01-30T15:32:12.874573541Z" level=info msg="StopPodSandbox for \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\"" Jan 30 15:32:12.875226 containerd[1476]: time="2025-01-30T15:32:12.874749622Z" level=info msg="TearDown network for sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" successfully" Jan 30 15:32:12.875226 containerd[1476]: time="2025-01-30T15:32:12.874775702Z" level=info msg="StopPodSandbox for \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" returns successfully" Jan 30 15:32:12.877586 containerd[1476]: time="2025-01-30T15:32:12.876100307Z" level=info msg="RemovePodSandbox for \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\"" Jan 30 15:32:12.877586 containerd[1476]: time="2025-01-30T15:32:12.876161187Z" level=info msg="Forcibly stopping sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\"" Jan 30 15:32:12.877586 containerd[1476]: time="2025-01-30T15:32:12.876254307Z" level=info msg="TearDown network for sandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" successfully" Jan 30 15:32:12.880848 containerd[1476]: time="2025-01-30T15:32:12.880802764Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 15:32:12.881100 containerd[1476]: time="2025-01-30T15:32:12.881040045Z" level=info msg="RemovePodSandbox \"474c4b3def0dd43b251a386a09e99027f93db5db646521673028bb2864d25aac\" returns successfully" Jan 30 15:32:12.881744 containerd[1476]: time="2025-01-30T15:32:12.881711567Z" level=info msg="StopPodSandbox for \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\"" Jan 30 15:32:12.881822 containerd[1476]: time="2025-01-30T15:32:12.881804448Z" level=info msg="TearDown network for sandbox \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\" successfully" Jan 30 15:32:12.881859 containerd[1476]: time="2025-01-30T15:32:12.881822328Z" level=info msg="StopPodSandbox for \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\" returns successfully" Jan 30 15:32:12.882350 containerd[1476]: time="2025-01-30T15:32:12.882310049Z" level=info msg="RemovePodSandbox for \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\"" Jan 30 15:32:12.882384 containerd[1476]: time="2025-01-30T15:32:12.882358330Z" level=info msg="Forcibly stopping sandbox \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\"" Jan 30 15:32:12.882430 containerd[1476]: time="2025-01-30T15:32:12.882414130Z" level=info msg="TearDown network for sandbox \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\" successfully" Jan 30 15:32:12.885819 containerd[1476]: time="2025-01-30T15:32:12.885767542Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 15:32:12.885920 containerd[1476]: time="2025-01-30T15:32:12.885848503Z" level=info msg="RemovePodSandbox \"fd874104fd000fad3c0c837bb450946195f63f8aabe6bc9229cc8c7258cb1075\" returns successfully" Jan 30 15:32:27.454984 kubelet[2704]: E0130 15:32:27.454927 2704 controller.go:195] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:41268->10.0.0.2:2379: read: connection timed out" Jan 30 15:32:27.462763 systemd[1]: cri-containerd-165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838.scope: Deactivated successfully. Jan 30 15:32:27.463457 systemd[1]: cri-containerd-165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838.scope: Consumed 6.537s CPU time, 16.3M memory peak, 0B memory swap peak. Jan 30 15:32:27.478671 systemd[1]: cri-containerd-9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4.scope: Deactivated successfully. Jan 30 15:32:27.479400 systemd[1]: cri-containerd-9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4.scope: Consumed 6.666s CPU time, 17.4M memory peak, 0B memory swap peak. Jan 30 15:32:27.502722 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838-rootfs.mount: Deactivated successfully. Jan 30 15:32:27.512025 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4-rootfs.mount: Deactivated successfully. Jan 30 15:32:27.513669 containerd[1476]: time="2025-01-30T15:32:27.513340150Z" level=info msg="shim disconnected" id=165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838 namespace=k8s.io Jan 30 15:32:27.513669 containerd[1476]: time="2025-01-30T15:32:27.513399710Z" level=warning msg="cleaning up after shim disconnected" id=165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838 namespace=k8s.io Jan 30 15:32:27.513669 containerd[1476]: time="2025-01-30T15:32:27.513413510Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:32:27.514748 containerd[1476]: time="2025-01-30T15:32:27.514370595Z" level=info msg="shim disconnected" id=9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4 namespace=k8s.io Jan 30 15:32:27.514748 containerd[1476]: time="2025-01-30T15:32:27.514412715Z" level=warning msg="cleaning up after shim disconnected" id=9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4 namespace=k8s.io Jan 30 15:32:27.514748 containerd[1476]: time="2025-01-30T15:32:27.514424155Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 15:32:27.957734 kubelet[2704]: I0130 15:32:27.957697 2704 scope.go:117] "RemoveContainer" containerID="165d1d6840b8537893341c2a4fcca69ffd176805bcc7d9ccf9c59fdf57ce2838" Jan 30 15:32:27.962584 kubelet[2704]: I0130 15:32:27.962304 2704 scope.go:117] "RemoveContainer" containerID="9466bbbf41dbc7c39fcf0ec94677bc3daf82dea512ae4e98530691d1b8757af4" Jan 30 15:32:27.962837 containerd[1476]: time="2025-01-30T15:32:27.962796179Z" level=info msg="CreateContainer within sandbox \"5e530141da586405be65973f8518cd3c9225836d39bcd3d7f238c4b2b2a4b654\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Jan 30 15:32:27.966823 containerd[1476]: time="2025-01-30T15:32:27.966682356Z" level=info msg="CreateContainer within sandbox \"443706cb51f09c76a59b3d00bae25e9f629293fb13424ab2cdbf8798fb8fc8e8\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Jan 30 15:32:27.983823 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1677921218.mount: Deactivated successfully. Jan 30 15:32:27.991882 containerd[1476]: time="2025-01-30T15:32:27.991826670Z" level=info msg="CreateContainer within sandbox \"5e530141da586405be65973f8518cd3c9225836d39bcd3d7f238c4b2b2a4b654\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"9f8db0d589a056ad6d4828d45b5de9c9ddee0a110b5ab21d272595d4ff0195a4\"" Jan 30 15:32:27.993174 containerd[1476]: time="2025-01-30T15:32:27.993127435Z" level=info msg="StartContainer for \"9f8db0d589a056ad6d4828d45b5de9c9ddee0a110b5ab21d272595d4ff0195a4\"" Jan 30 15:32:28.005959 containerd[1476]: time="2025-01-30T15:32:28.005903373Z" level=info msg="CreateContainer within sandbox \"443706cb51f09c76a59b3d00bae25e9f629293fb13424ab2cdbf8798fb8fc8e8\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"b9f459e7fbae6662737c73a030c3521dec9716d541efa6367aa89dedd32b7270\"" Jan 30 15:32:28.008353 containerd[1476]: time="2025-01-30T15:32:28.007078579Z" level=info msg="StartContainer for \"b9f459e7fbae6662737c73a030c3521dec9716d541efa6367aa89dedd32b7270\"" Jan 30 15:32:28.036788 systemd[1]: Started cri-containerd-9f8db0d589a056ad6d4828d45b5de9c9ddee0a110b5ab21d272595d4ff0195a4.scope - libcontainer container 9f8db0d589a056ad6d4828d45b5de9c9ddee0a110b5ab21d272595d4ff0195a4. Jan 30 15:32:28.048793 systemd[1]: Started cri-containerd-b9f459e7fbae6662737c73a030c3521dec9716d541efa6367aa89dedd32b7270.scope - libcontainer container b9f459e7fbae6662737c73a030c3521dec9716d541efa6367aa89dedd32b7270. Jan 30 15:32:28.115154 containerd[1476]: time="2025-01-30T15:32:28.115102112Z" level=info msg="StartContainer for \"9f8db0d589a056ad6d4828d45b5de9c9ddee0a110b5ab21d272595d4ff0195a4\" returns successfully" Jan 30 15:32:28.121719 containerd[1476]: time="2025-01-30T15:32:28.121660062Z" level=info msg="StartContainer for \"b9f459e7fbae6662737c73a030c3521dec9716d541efa6367aa89dedd32b7270\" returns successfully" Jan 30 15:32:32.190723 kubelet[2704]: E0130 15:32:32.190519 2704 event.go:359] "Server rejected event (will not retry!)" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.0.0.3:41082->10.0.0.2:2379: read: connection timed out" event="&Event{ObjectMeta:{kube-apiserver-ci-4081-3-0-a-5514912dcd.181f82342b85c231 kube-system 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:kube-system,Name:kube-apiserver-ci-4081-3-0-a-5514912dcd,UID:971326e260333f76ace84ca47537837d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Readiness probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:ci-4081-3-0-a-5514912dcd,},FirstTimestamp:2025-01-30 15:32:21.768397361 +0000 UTC m=+369.059811231,LastTimestamp:2025-01-30 15:32:21.768397361 +0000 UTC m=+369.059811231,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4081-3-0-a-5514912dcd,}"