Apr 12 18:20:07.726252 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Apr 12 18:20:07.726271 kernel: Linux version 5.15.154-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP PREEMPT Fri Apr 12 17:21:24 -00 2024 Apr 12 18:20:07.726279 kernel: efi: EFI v2.70 by EDK II Apr 12 18:20:07.726285 kernel: efi: SMBIOS 3.0=0xd9260000 ACPI 2.0=0xd9240000 MEMATTR=0xda32b018 RNG=0xd9220018 MEMRESERVE=0xd9521c18 Apr 12 18:20:07.726290 kernel: random: crng init done Apr 12 18:20:07.726295 kernel: ACPI: Early table checksum verification disabled Apr 12 18:20:07.726302 kernel: ACPI: RSDP 0x00000000D9240000 000024 (v02 BOCHS ) Apr 12 18:20:07.726308 kernel: ACPI: XSDT 0x00000000D9230000 000064 (v01 BOCHS BXPC 00000001 01000013) Apr 12 18:20:07.726314 kernel: ACPI: FACP 0x00000000D91E0000 000114 (v06 BOCHS BXPC 00000001 BXPC 00000001) Apr 12 18:20:07.726319 kernel: ACPI: DSDT 0x00000000D91F0000 0014A2 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 12 18:20:07.726325 kernel: ACPI: APIC 0x00000000D91D0000 0001A8 (v04 BOCHS BXPC 00000001 BXPC 00000001) Apr 12 18:20:07.726330 kernel: ACPI: PPTT 0x00000000D91C0000 00009C (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 12 18:20:07.726336 kernel: ACPI: GTDT 0x00000000D91B0000 000060 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 12 18:20:07.726341 kernel: ACPI: MCFG 0x00000000D91A0000 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Apr 12 18:20:07.726349 kernel: ACPI: SPCR 0x00000000D9190000 000050 (v02 BOCHS BXPC 00000001 BXPC 00000001) Apr 12 18:20:07.726355 kernel: ACPI: DBG2 0x00000000D9180000 000057 (v00 BOCHS BXPC 00000001 BXPC 00000001) Apr 12 18:20:07.726361 kernel: ACPI: IORT 0x00000000D9170000 000080 (v03 BOCHS BXPC 00000001 BXPC 00000001) Apr 12 18:20:07.726366 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600 Apr 12 18:20:07.726372 kernel: NUMA: Failed to initialise from firmware Apr 12 18:20:07.726378 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff] Apr 12 18:20:07.726383 kernel: NUMA: NODE_DATA [mem 0xdcb09900-0xdcb0efff] Apr 12 18:20:07.726389 kernel: Zone ranges: Apr 12 18:20:07.726395 kernel: DMA [mem 0x0000000040000000-0x00000000dcffffff] Apr 12 18:20:07.726401 kernel: DMA32 empty Apr 12 18:20:07.726407 kernel: Normal empty Apr 12 18:20:07.726413 kernel: Movable zone start for each node Apr 12 18:20:07.726418 kernel: Early memory node ranges Apr 12 18:20:07.726424 kernel: node 0: [mem 0x0000000040000000-0x00000000d924ffff] Apr 12 18:20:07.726430 kernel: node 0: [mem 0x00000000d9250000-0x00000000d951ffff] Apr 12 18:20:07.726435 kernel: node 0: [mem 0x00000000d9520000-0x00000000dc7fffff] Apr 12 18:20:07.726441 kernel: node 0: [mem 0x00000000dc800000-0x00000000dc88ffff] Apr 12 18:20:07.726447 kernel: node 0: [mem 0x00000000dc890000-0x00000000dc89ffff] Apr 12 18:20:07.726452 kernel: node 0: [mem 0x00000000dc8a0000-0x00000000dc9bffff] Apr 12 18:20:07.726458 kernel: node 0: [mem 0x00000000dc9c0000-0x00000000dcffffff] Apr 12 18:20:07.726464 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff] Apr 12 18:20:07.726471 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges Apr 12 18:20:07.726476 kernel: psci: probing for conduit method from ACPI. Apr 12 18:20:07.726482 kernel: psci: PSCIv1.1 detected in firmware. Apr 12 18:20:07.726488 kernel: psci: Using standard PSCI v0.2 function IDs Apr 12 18:20:07.726494 kernel: psci: Trusted OS migration not required Apr 12 18:20:07.726502 kernel: psci: SMC Calling Convention v1.1 Apr 12 18:20:07.726508 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003) Apr 12 18:20:07.726515 kernel: ACPI: SRAT not present Apr 12 18:20:07.726522 kernel: percpu: Embedded 30 pages/cpu s83032 r8192 d31656 u122880 Apr 12 18:20:07.726528 kernel: pcpu-alloc: s83032 r8192 d31656 u122880 alloc=30*4096 Apr 12 18:20:07.726534 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 Apr 12 18:20:07.726540 kernel: Detected PIPT I-cache on CPU0 Apr 12 18:20:07.726546 kernel: CPU features: detected: GIC system register CPU interface Apr 12 18:20:07.726552 kernel: CPU features: detected: Hardware dirty bit management Apr 12 18:20:07.726558 kernel: CPU features: detected: Spectre-v4 Apr 12 18:20:07.726564 kernel: CPU features: detected: Spectre-BHB Apr 12 18:20:07.726592 kernel: CPU features: kernel page table isolation forced ON by KASLR Apr 12 18:20:07.726599 kernel: CPU features: detected: Kernel page table isolation (KPTI) Apr 12 18:20:07.726605 kernel: CPU features: detected: ARM erratum 1418040 Apr 12 18:20:07.726612 kernel: Built 1 zonelists, mobility grouping on. Total pages: 633024 Apr 12 18:20:07.726618 kernel: Policy zone: DMA Apr 12 18:20:07.726625 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=c0b96868344262519ffdb2dae3782c942008a0fecdbc0bc85d2e170bd2e8b8a8 Apr 12 18:20:07.726632 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Apr 12 18:20:07.726638 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 12 18:20:07.726644 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 12 18:20:07.726650 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 12 18:20:07.726657 kernel: Memory: 2457460K/2572288K available (9792K kernel code, 2092K rwdata, 7568K rodata, 36352K init, 777K bss, 114828K reserved, 0K cma-reserved) Apr 12 18:20:07.726664 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Apr 12 18:20:07.726670 kernel: trace event string verifier disabled Apr 12 18:20:07.726676 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 12 18:20:07.726683 kernel: rcu: RCU event tracing is enabled. Apr 12 18:20:07.726689 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Apr 12 18:20:07.726696 kernel: Trampoline variant of Tasks RCU enabled. Apr 12 18:20:07.726702 kernel: Tracing variant of Tasks RCU enabled. Apr 12 18:20:07.726708 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 12 18:20:07.726714 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Apr 12 18:20:07.726720 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Apr 12 18:20:07.726726 kernel: GICv3: 256 SPIs implemented Apr 12 18:20:07.726734 kernel: GICv3: 0 Extended SPIs implemented Apr 12 18:20:07.726740 kernel: GICv3: Distributor has no Range Selector support Apr 12 18:20:07.726746 kernel: Root IRQ handler: gic_handle_irq Apr 12 18:20:07.726752 kernel: GICv3: 16 PPIs implemented Apr 12 18:20:07.726758 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000 Apr 12 18:20:07.726764 kernel: ACPI: SRAT not present Apr 12 18:20:07.726770 kernel: ITS [mem 0x08080000-0x0809ffff] Apr 12 18:20:07.726776 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400b0000 (indirect, esz 8, psz 64K, shr 1) Apr 12 18:20:07.726783 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400c0000 (flat, esz 8, psz 64K, shr 1) Apr 12 18:20:07.726795 kernel: GICv3: using LPI property table @0x00000000400d0000 Apr 12 18:20:07.726801 kernel: GICv3: CPU0: using allocated LPI pending table @0x00000000400e0000 Apr 12 18:20:07.726807 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 12 18:20:07.726816 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Apr 12 18:20:07.726822 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Apr 12 18:20:07.726828 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Apr 12 18:20:07.726834 kernel: arm-pv: using stolen time PV Apr 12 18:20:07.726841 kernel: Console: colour dummy device 80x25 Apr 12 18:20:07.726847 kernel: ACPI: Core revision 20210730 Apr 12 18:20:07.726854 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Apr 12 18:20:07.726860 kernel: pid_max: default: 32768 minimum: 301 Apr 12 18:20:07.726866 kernel: LSM: Security Framework initializing Apr 12 18:20:07.726872 kernel: SELinux: Initializing. Apr 12 18:20:07.726880 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 12 18:20:07.726886 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 12 18:20:07.726893 kernel: rcu: Hierarchical SRCU implementation. Apr 12 18:20:07.726899 kernel: Platform MSI: ITS@0x8080000 domain created Apr 12 18:20:07.726905 kernel: PCI/MSI: ITS@0x8080000 domain created Apr 12 18:20:07.726911 kernel: Remapping and enabling EFI services. Apr 12 18:20:07.726918 kernel: smp: Bringing up secondary CPUs ... Apr 12 18:20:07.726924 kernel: Detected PIPT I-cache on CPU1 Apr 12 18:20:07.726930 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000 Apr 12 18:20:07.726938 kernel: GICv3: CPU1: using allocated LPI pending table @0x00000000400f0000 Apr 12 18:20:07.726944 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 12 18:20:07.726950 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Apr 12 18:20:07.726957 kernel: Detected PIPT I-cache on CPU2 Apr 12 18:20:07.726963 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000 Apr 12 18:20:07.726969 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040100000 Apr 12 18:20:07.726976 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 12 18:20:07.726982 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1] Apr 12 18:20:07.726988 kernel: Detected PIPT I-cache on CPU3 Apr 12 18:20:07.726994 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000 Apr 12 18:20:07.727001 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040110000 Apr 12 18:20:07.727008 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 12 18:20:07.727014 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1] Apr 12 18:20:07.727020 kernel: smp: Brought up 1 node, 4 CPUs Apr 12 18:20:07.727030 kernel: SMP: Total of 4 processors activated. Apr 12 18:20:07.727038 kernel: CPU features: detected: 32-bit EL0 Support Apr 12 18:20:07.727045 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Apr 12 18:20:07.727051 kernel: CPU features: detected: Common not Private translations Apr 12 18:20:07.727058 kernel: CPU features: detected: CRC32 instructions Apr 12 18:20:07.727065 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Apr 12 18:20:07.727071 kernel: CPU features: detected: LSE atomic instructions Apr 12 18:20:07.727078 kernel: CPU features: detected: Privileged Access Never Apr 12 18:20:07.727086 kernel: CPU features: detected: RAS Extension Support Apr 12 18:20:07.727092 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS) Apr 12 18:20:07.727099 kernel: CPU: All CPU(s) started at EL1 Apr 12 18:20:07.727105 kernel: alternatives: patching kernel code Apr 12 18:20:07.727112 kernel: devtmpfs: initialized Apr 12 18:20:07.727120 kernel: KASLR enabled Apr 12 18:20:07.727126 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 12 18:20:07.727133 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Apr 12 18:20:07.727143 kernel: pinctrl core: initialized pinctrl subsystem Apr 12 18:20:07.727150 kernel: SMBIOS 3.0.0 present. Apr 12 18:20:07.727157 kernel: DMI: QEMU KVM Virtual Machine, BIOS 0.0.0 02/06/2015 Apr 12 18:20:07.727164 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 12 18:20:07.727173 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Apr 12 18:20:07.727180 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Apr 12 18:20:07.727188 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Apr 12 18:20:07.727195 kernel: audit: initializing netlink subsys (disabled) Apr 12 18:20:07.727202 kernel: audit: type=2000 audit(0.033:1): state=initialized audit_enabled=0 res=1 Apr 12 18:20:07.727208 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 12 18:20:07.727215 kernel: cpuidle: using governor menu Apr 12 18:20:07.727222 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Apr 12 18:20:07.727228 kernel: ASID allocator initialised with 32768 entries Apr 12 18:20:07.727235 kernel: ACPI: bus type PCI registered Apr 12 18:20:07.727241 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 12 18:20:07.727249 kernel: Serial: AMBA PL011 UART driver Apr 12 18:20:07.727256 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Apr 12 18:20:07.727262 kernel: HugeTLB registered 32.0 MiB page size, pre-allocated 0 pages Apr 12 18:20:07.727269 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Apr 12 18:20:07.727276 kernel: HugeTLB registered 64.0 KiB page size, pre-allocated 0 pages Apr 12 18:20:07.727282 kernel: cryptd: max_cpu_qlen set to 1000 Apr 12 18:20:07.727289 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Apr 12 18:20:07.727295 kernel: ACPI: Added _OSI(Module Device) Apr 12 18:20:07.727302 kernel: ACPI: Added _OSI(Processor Device) Apr 12 18:20:07.727310 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Apr 12 18:20:07.727316 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 12 18:20:07.727323 kernel: ACPI: Added _OSI(Linux-Dell-Video) Apr 12 18:20:07.727329 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Apr 12 18:20:07.727336 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Apr 12 18:20:07.727343 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 12 18:20:07.727350 kernel: ACPI: Interpreter enabled Apr 12 18:20:07.727357 kernel: ACPI: Using GIC for interrupt routing Apr 12 18:20:07.727363 kernel: ACPI: MCFG table detected, 1 entries Apr 12 18:20:07.727371 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA Apr 12 18:20:07.727378 kernel: printk: console [ttyAMA0] enabled Apr 12 18:20:07.727384 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Apr 12 18:20:07.727502 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Apr 12 18:20:07.727584 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR] Apr 12 18:20:07.727646 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability] Apr 12 18:20:07.727706 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00 Apr 12 18:20:07.727768 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff] Apr 12 18:20:07.727778 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io 0x0000-0xffff window] Apr 12 18:20:07.727785 kernel: PCI host bridge to bus 0000:00 Apr 12 18:20:07.728018 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window] Apr 12 18:20:07.728079 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0xffff window] Apr 12 18:20:07.728134 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window] Apr 12 18:20:07.728190 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Apr 12 18:20:07.728270 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000 Apr 12 18:20:07.728343 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00 Apr 12 18:20:07.728411 kernel: pci 0000:00:01.0: reg 0x10: [io 0x0000-0x001f] Apr 12 18:20:07.728477 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff] Apr 12 18:20:07.728541 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref] Apr 12 18:20:07.728626 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref] Apr 12 18:20:07.728692 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff] Apr 12 18:20:07.728760 kernel: pci 0000:00:01.0: BAR 0: assigned [io 0x1000-0x101f] Apr 12 18:20:07.728831 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window] Apr 12 18:20:07.728891 kernel: pci_bus 0000:00: resource 5 [io 0x0000-0xffff window] Apr 12 18:20:07.728949 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window] Apr 12 18:20:07.728958 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35 Apr 12 18:20:07.728965 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36 Apr 12 18:20:07.728972 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37 Apr 12 18:20:07.728981 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38 Apr 12 18:20:07.728988 kernel: iommu: Default domain type: Translated Apr 12 18:20:07.728995 kernel: iommu: DMA domain TLB invalidation policy: strict mode Apr 12 18:20:07.729001 kernel: vgaarb: loaded Apr 12 18:20:07.729008 kernel: pps_core: LinuxPPS API ver. 1 registered Apr 12 18:20:07.729015 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Apr 12 18:20:07.729022 kernel: PTP clock support registered Apr 12 18:20:07.729029 kernel: Registered efivars operations Apr 12 18:20:07.729036 kernel: clocksource: Switched to clocksource arch_sys_counter Apr 12 18:20:07.729043 kernel: VFS: Disk quotas dquot_6.6.0 Apr 12 18:20:07.729051 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 12 18:20:07.729058 kernel: pnp: PnP ACPI init Apr 12 18:20:07.729130 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved Apr 12 18:20:07.729140 kernel: pnp: PnP ACPI: found 1 devices Apr 12 18:20:07.729147 kernel: NET: Registered PF_INET protocol family Apr 12 18:20:07.729154 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 12 18:20:07.729161 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 12 18:20:07.729168 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 12 18:20:07.729177 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 12 18:20:07.729184 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) Apr 12 18:20:07.729190 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 12 18:20:07.729197 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 12 18:20:07.729204 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 12 18:20:07.729211 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 12 18:20:07.729218 kernel: PCI: CLS 0 bytes, default 64 Apr 12 18:20:07.729225 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available Apr 12 18:20:07.729231 kernel: kvm [1]: HYP mode not available Apr 12 18:20:07.729239 kernel: Initialise system trusted keyrings Apr 12 18:20:07.729246 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 12 18:20:07.729253 kernel: Key type asymmetric registered Apr 12 18:20:07.729260 kernel: Asymmetric key parser 'x509' registered Apr 12 18:20:07.729266 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Apr 12 18:20:07.729273 kernel: io scheduler mq-deadline registered Apr 12 18:20:07.729280 kernel: io scheduler kyber registered Apr 12 18:20:07.729287 kernel: io scheduler bfq registered Apr 12 18:20:07.729294 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0 Apr 12 18:20:07.729301 kernel: ACPI: button: Power Button [PWRB] Apr 12 18:20:07.729309 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36 Apr 12 18:20:07.729374 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007) Apr 12 18:20:07.729383 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 12 18:20:07.729390 kernel: thunder_xcv, ver 1.0 Apr 12 18:20:07.729397 kernel: thunder_bgx, ver 1.0 Apr 12 18:20:07.729404 kernel: nicpf, ver 1.0 Apr 12 18:20:07.729410 kernel: nicvf, ver 1.0 Apr 12 18:20:07.729479 kernel: rtc-efi rtc-efi.0: registered as rtc0 Apr 12 18:20:07.729542 kernel: rtc-efi rtc-efi.0: setting system clock to 2024-04-12T18:20:07 UTC (1712946007) Apr 12 18:20:07.729552 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 12 18:20:07.729559 kernel: NET: Registered PF_INET6 protocol family Apr 12 18:20:07.729566 kernel: Segment Routing with IPv6 Apr 12 18:20:07.729584 kernel: In-situ OAM (IOAM) with IPv6 Apr 12 18:20:07.729590 kernel: NET: Registered PF_PACKET protocol family Apr 12 18:20:07.729597 kernel: Key type dns_resolver registered Apr 12 18:20:07.729604 kernel: registered taskstats version 1 Apr 12 18:20:07.729612 kernel: Loading compiled-in X.509 certificates Apr 12 18:20:07.729619 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.154-flatcar: 8c258d82bbd8df4a9da2c0ea4108142f04be6b34' Apr 12 18:20:07.729626 kernel: Key type .fscrypt registered Apr 12 18:20:07.729633 kernel: Key type fscrypt-provisioning registered Apr 12 18:20:07.729640 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 12 18:20:07.729647 kernel: ima: Allocated hash algorithm: sha1 Apr 12 18:20:07.729653 kernel: ima: No architecture policies found Apr 12 18:20:07.729660 kernel: Freeing unused kernel memory: 36352K Apr 12 18:20:07.729667 kernel: Run /init as init process Apr 12 18:20:07.729675 kernel: with arguments: Apr 12 18:20:07.729681 kernel: /init Apr 12 18:20:07.729688 kernel: with environment: Apr 12 18:20:07.729695 kernel: HOME=/ Apr 12 18:20:07.729701 kernel: TERM=linux Apr 12 18:20:07.729708 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Apr 12 18:20:07.729717 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Apr 12 18:20:07.729726 systemd[1]: Detected virtualization kvm. Apr 12 18:20:07.729735 systemd[1]: Detected architecture arm64. Apr 12 18:20:07.729742 systemd[1]: Running in initrd. Apr 12 18:20:07.729749 systemd[1]: No hostname configured, using default hostname. Apr 12 18:20:07.729756 systemd[1]: Hostname set to . Apr 12 18:20:07.729763 systemd[1]: Initializing machine ID from VM UUID. Apr 12 18:20:07.729770 systemd[1]: Queued start job for default target initrd.target. Apr 12 18:20:07.729778 systemd[1]: Started systemd-ask-password-console.path. Apr 12 18:20:07.729785 systemd[1]: Reached target cryptsetup.target. Apr 12 18:20:07.729801 systemd[1]: Reached target paths.target. Apr 12 18:20:07.729808 systemd[1]: Reached target slices.target. Apr 12 18:20:07.729815 systemd[1]: Reached target swap.target. Apr 12 18:20:07.729822 systemd[1]: Reached target timers.target. Apr 12 18:20:07.729829 systemd[1]: Listening on iscsid.socket. Apr 12 18:20:07.729836 systemd[1]: Listening on iscsiuio.socket. Apr 12 18:20:07.729843 systemd[1]: Listening on systemd-journald-audit.socket. Apr 12 18:20:07.729853 systemd[1]: Listening on systemd-journald-dev-log.socket. Apr 12 18:20:07.729860 systemd[1]: Listening on systemd-journald.socket. Apr 12 18:20:07.729867 systemd[1]: Listening on systemd-networkd.socket. Apr 12 18:20:07.729874 systemd[1]: Listening on systemd-udevd-control.socket. Apr 12 18:20:07.729881 systemd[1]: Listening on systemd-udevd-kernel.socket. Apr 12 18:20:07.729888 systemd[1]: Reached target sockets.target. Apr 12 18:20:07.729895 systemd[1]: Starting kmod-static-nodes.service... Apr 12 18:20:07.729902 systemd[1]: Finished network-cleanup.service. Apr 12 18:20:07.729909 systemd[1]: Starting systemd-fsck-usr.service... Apr 12 18:20:07.729918 systemd[1]: Starting systemd-journald.service... Apr 12 18:20:07.729925 systemd[1]: Starting systemd-modules-load.service... Apr 12 18:20:07.729932 systemd[1]: Starting systemd-resolved.service... Apr 12 18:20:07.729939 systemd[1]: Starting systemd-vconsole-setup.service... Apr 12 18:20:07.729946 systemd[1]: Finished kmod-static-nodes.service. Apr 12 18:20:07.729953 systemd[1]: Finished systemd-fsck-usr.service. Apr 12 18:20:07.729961 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Apr 12 18:20:07.729968 systemd[1]: Finished systemd-vconsole-setup.service. Apr 12 18:20:07.729975 systemd[1]: Starting dracut-cmdline-ask.service... Apr 12 18:20:07.729986 systemd-journald[250]: Journal started Apr 12 18:20:07.730026 systemd-journald[250]: Runtime Journal (/run/log/journal/4532c036a0be47be89214dd3f1adde8f) is 6.0M, max 48.7M, 42.6M free. Apr 12 18:20:07.719313 systemd-modules-load[251]: Inserted module 'overlay' Apr 12 18:20:07.732288 systemd[1]: Started systemd-journald.service. Apr 12 18:20:07.732000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.735629 kernel: audit: type=1130 audit(1712946007.732:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.735658 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Apr 12 18:20:07.735000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.736064 systemd-resolved[252]: Positive Trust Anchors: Apr 12 18:20:07.740287 kernel: audit: type=1130 audit(1712946007.735:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.736071 systemd-resolved[252]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 12 18:20:07.736098 systemd-resolved[252]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Apr 12 18:20:07.750429 kernel: audit: type=1130 audit(1712946007.742:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.750447 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 12 18:20:07.742000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.740173 systemd-resolved[252]: Defaulting to hostname 'linux'. Apr 12 18:20:07.740941 systemd[1]: Started systemd-resolved.service. Apr 12 18:20:07.749042 systemd[1]: Reached target nss-lookup.target. Apr 12 18:20:07.753968 systemd[1]: Finished dracut-cmdline-ask.service. Apr 12 18:20:07.756815 kernel: Bridge firewalling registered Apr 12 18:20:07.756832 kernel: audit: type=1130 audit(1712946007.756:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.756000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.756144 systemd-modules-load[251]: Inserted module 'br_netfilter' Apr 12 18:20:07.757694 systemd[1]: Starting dracut-cmdline.service... Apr 12 18:20:07.766721 dracut-cmdline[268]: dracut-dracut-053 Apr 12 18:20:07.768989 dracut-cmdline[268]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=c0b96868344262519ffdb2dae3782c942008a0fecdbc0bc85d2e170bd2e8b8a8 Apr 12 18:20:07.773066 kernel: SCSI subsystem initialized Apr 12 18:20:07.777709 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 12 18:20:07.777752 kernel: device-mapper: uevent: version 1.0.3 Apr 12 18:20:07.777762 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Apr 12 18:20:07.779642 systemd-modules-load[251]: Inserted module 'dm_multipath' Apr 12 18:20:07.780371 systemd[1]: Finished systemd-modules-load.service. Apr 12 18:20:07.780000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.784407 systemd[1]: Starting systemd-sysctl.service... Apr 12 18:20:07.786268 kernel: audit: type=1130 audit(1712946007.780:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.791702 systemd[1]: Finished systemd-sysctl.service. Apr 12 18:20:07.791000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.795589 kernel: audit: type=1130 audit(1712946007.791:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.831588 kernel: Loading iSCSI transport class v2.0-870. Apr 12 18:20:07.843595 kernel: iscsi: registered transport (tcp) Apr 12 18:20:07.858588 kernel: iscsi: registered transport (qla4xxx) Apr 12 18:20:07.858608 kernel: QLogic iSCSI HBA Driver Apr 12 18:20:07.891481 systemd[1]: Finished dracut-cmdline.service. Apr 12 18:20:07.892000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.893065 systemd[1]: Starting dracut-pre-udev.service... Apr 12 18:20:07.896242 kernel: audit: type=1130 audit(1712946007.892:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:07.938603 kernel: raid6: neonx8 gen() 13749 MB/s Apr 12 18:20:07.955589 kernel: raid6: neonx8 xor() 10774 MB/s Apr 12 18:20:07.972589 kernel: raid6: neonx4 gen() 13498 MB/s Apr 12 18:20:07.989590 kernel: raid6: neonx4 xor() 11207 MB/s Apr 12 18:20:08.006590 kernel: raid6: neonx2 gen() 12871 MB/s Apr 12 18:20:08.023591 kernel: raid6: neonx2 xor() 10256 MB/s Apr 12 18:20:08.040589 kernel: raid6: neonx1 gen() 10470 MB/s Apr 12 18:20:08.057593 kernel: raid6: neonx1 xor() 8700 MB/s Apr 12 18:20:08.074588 kernel: raid6: int64x8 gen() 6224 MB/s Apr 12 18:20:08.091588 kernel: raid6: int64x8 xor() 3525 MB/s Apr 12 18:20:08.108592 kernel: raid6: int64x4 gen() 7163 MB/s Apr 12 18:20:08.125589 kernel: raid6: int64x4 xor() 3831 MB/s Apr 12 18:20:08.142582 kernel: raid6: int64x2 gen() 6111 MB/s Apr 12 18:20:08.159588 kernel: raid6: int64x2 xor() 3288 MB/s Apr 12 18:20:08.176593 kernel: raid6: int64x1 gen() 5020 MB/s Apr 12 18:20:08.193683 kernel: raid6: int64x1 xor() 2633 MB/s Apr 12 18:20:08.193695 kernel: raid6: using algorithm neonx8 gen() 13749 MB/s Apr 12 18:20:08.193705 kernel: raid6: .... xor() 10774 MB/s, rmw enabled Apr 12 18:20:08.194747 kernel: raid6: using neon recovery algorithm Apr 12 18:20:08.206590 kernel: xor: measuring software checksum speed Apr 12 18:20:08.208613 kernel: 8regs : 17199 MB/sec Apr 12 18:20:08.208627 kernel: 32regs : 20562 MB/sec Apr 12 18:20:08.210062 kernel: arm64_neon : 27968 MB/sec Apr 12 18:20:08.210073 kernel: xor: using function: arm64_neon (27968 MB/sec) Apr 12 18:20:08.265589 kernel: Btrfs loaded, crc32c=crc32c-generic, zoned=no, fsverity=no Apr 12 18:20:08.276125 systemd[1]: Finished dracut-pre-udev.service. Apr 12 18:20:08.280633 kernel: audit: type=1130 audit(1712946008.276:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:08.280658 kernel: audit: type=1334 audit(1712946008.279:10): prog-id=7 op=LOAD Apr 12 18:20:08.276000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:08.279000 audit: BPF prog-id=7 op=LOAD Apr 12 18:20:08.279000 audit: BPF prog-id=8 op=LOAD Apr 12 18:20:08.280974 systemd[1]: Starting systemd-udevd.service... Apr 12 18:20:08.294765 systemd-udevd[451]: Using default interface naming scheme 'v252'. Apr 12 18:20:08.298169 systemd[1]: Started systemd-udevd.service. Apr 12 18:20:08.298000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:08.301036 systemd[1]: Starting dracut-pre-trigger.service... Apr 12 18:20:08.312790 dracut-pre-trigger[459]: rd.md=0: removing MD RAID activation Apr 12 18:20:08.339170 systemd[1]: Finished dracut-pre-trigger.service. Apr 12 18:20:08.339000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:08.340808 systemd[1]: Starting systemd-udev-trigger.service... Apr 12 18:20:08.374386 systemd[1]: Finished systemd-udev-trigger.service. Apr 12 18:20:08.374000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:08.403669 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Apr 12 18:20:08.405795 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Apr 12 18:20:08.405819 kernel: GPT:9289727 != 19775487 Apr 12 18:20:08.405828 kernel: GPT:Alternate GPT header not at the end of the disk. Apr 12 18:20:08.405843 kernel: GPT:9289727 != 19775487 Apr 12 18:20:08.405851 kernel: GPT: Use GNU Parted to correct GPT errors. Apr 12 18:20:08.405859 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 12 18:20:08.422584 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (497) Apr 12 18:20:08.426590 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Apr 12 18:20:08.429953 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Apr 12 18:20:08.432802 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Apr 12 18:20:08.433822 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Apr 12 18:20:08.440131 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Apr 12 18:20:08.441815 systemd[1]: Starting disk-uuid.service... Apr 12 18:20:08.447619 disk-uuid[521]: Primary Header is updated. Apr 12 18:20:08.447619 disk-uuid[521]: Secondary Entries is updated. Apr 12 18:20:08.447619 disk-uuid[521]: Secondary Header is updated. Apr 12 18:20:08.450590 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 12 18:20:08.456587 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 12 18:20:09.460076 disk-uuid[522]: The operation has completed successfully. Apr 12 18:20:09.461166 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Apr 12 18:20:09.482423 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 12 18:20:09.483000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.483000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.482519 systemd[1]: Finished disk-uuid.service. Apr 12 18:20:09.486562 systemd[1]: Starting verity-setup.service... Apr 12 18:20:09.505589 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Apr 12 18:20:09.529219 systemd[1]: Found device dev-mapper-usr.device. Apr 12 18:20:09.531359 systemd[1]: Mounting sysusr-usr.mount... Apr 12 18:20:09.533207 systemd[1]: Finished verity-setup.service. Apr 12 18:20:09.533000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.576594 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Apr 12 18:20:09.576959 systemd[1]: Mounted sysusr-usr.mount. Apr 12 18:20:09.577763 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Apr 12 18:20:09.578492 systemd[1]: Starting ignition-setup.service... Apr 12 18:20:09.580683 systemd[1]: Starting parse-ip-for-networkd.service... Apr 12 18:20:09.586878 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Apr 12 18:20:09.586911 kernel: BTRFS info (device vda6): using free space tree Apr 12 18:20:09.586921 kernel: BTRFS info (device vda6): has skinny extents Apr 12 18:20:09.595507 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 12 18:20:09.601347 systemd[1]: Finished ignition-setup.service. Apr 12 18:20:09.602000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.602986 systemd[1]: Starting ignition-fetch-offline.service... Apr 12 18:20:09.666093 systemd[1]: Finished parse-ip-for-networkd.service. Apr 12 18:20:09.666000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.667000 audit: BPF prog-id=9 op=LOAD Apr 12 18:20:09.668414 systemd[1]: Starting systemd-networkd.service... Apr 12 18:20:09.678618 ignition[604]: Ignition 2.14.0 Apr 12 18:20:09.678631 ignition[604]: Stage: fetch-offline Apr 12 18:20:09.678668 ignition[604]: no configs at "/usr/lib/ignition/base.d" Apr 12 18:20:09.678677 ignition[604]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 12 18:20:09.678816 ignition[604]: parsed url from cmdline: "" Apr 12 18:20:09.678819 ignition[604]: no config URL provided Apr 12 18:20:09.678824 ignition[604]: reading system config file "/usr/lib/ignition/user.ign" Apr 12 18:20:09.678831 ignition[604]: no config at "/usr/lib/ignition/user.ign" Apr 12 18:20:09.678849 ignition[604]: op(1): [started] loading QEMU firmware config module Apr 12 18:20:09.678853 ignition[604]: op(1): executing: "modprobe" "qemu_fw_cfg" Apr 12 18:20:09.684114 ignition[604]: op(1): [finished] loading QEMU firmware config module Apr 12 18:20:09.693158 systemd-networkd[698]: lo: Link UP Apr 12 18:20:09.694000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.693173 systemd-networkd[698]: lo: Gained carrier Apr 12 18:20:09.693531 systemd-networkd[698]: Enumeration completed Apr 12 18:20:09.693717 systemd-networkd[698]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 12 18:20:09.694140 systemd[1]: Started systemd-networkd.service. Apr 12 18:20:09.694638 systemd-networkd[698]: eth0: Link UP Apr 12 18:20:09.694641 systemd-networkd[698]: eth0: Gained carrier Apr 12 18:20:09.695373 systemd[1]: Reached target network.target. Apr 12 18:20:09.696906 systemd[1]: Starting iscsiuio.service... Apr 12 18:20:09.706451 systemd[1]: Started iscsiuio.service. Apr 12 18:20:09.706000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.708089 systemd[1]: Starting iscsid.service... Apr 12 18:20:09.711959 iscsid[705]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Apr 12 18:20:09.711959 iscsid[705]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Apr 12 18:20:09.711959 iscsid[705]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Apr 12 18:20:09.711959 iscsid[705]: If using hardware iscsi like qla4xxx this message can be ignored. Apr 12 18:20:09.711959 iscsid[705]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Apr 12 18:20:09.711959 iscsid[705]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Apr 12 18:20:09.718000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.715206 systemd[1]: Started iscsid.service. Apr 12 18:20:09.720161 systemd[1]: Starting dracut-initqueue.service... Apr 12 18:20:09.723654 systemd-networkd[698]: eth0: DHCPv4 address 10.0.0.43/16, gateway 10.0.0.1 acquired from 10.0.0.1 Apr 12 18:20:09.731169 systemd[1]: Finished dracut-initqueue.service. Apr 12 18:20:09.731000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.732185 systemd[1]: Reached target remote-fs-pre.target. Apr 12 18:20:09.733657 systemd[1]: Reached target remote-cryptsetup.target. Apr 12 18:20:09.735238 systemd[1]: Reached target remote-fs.target. Apr 12 18:20:09.737448 systemd[1]: Starting dracut-pre-mount.service... Apr 12 18:20:09.744991 systemd[1]: Finished dracut-pre-mount.service. Apr 12 18:20:09.745000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.776556 ignition[604]: parsing config with SHA512: 1c8d809cfb70a54067bc2fcf856a2076c20f32b09bfbf6e550f720c05e2fefff1e4b2ceb99050d5edc5a63ee5704a8b2c474e22b18502d97eea7f9542e0d7af8 Apr 12 18:20:09.815553 unknown[604]: fetched base config from "system" Apr 12 18:20:09.815581 unknown[604]: fetched user config from "qemu" Apr 12 18:20:09.816259 ignition[604]: fetch-offline: fetch-offline passed Apr 12 18:20:09.816323 ignition[604]: Ignition finished successfully Apr 12 18:20:09.819566 systemd[1]: Finished ignition-fetch-offline.service. Apr 12 18:20:09.819000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.820487 systemd[1]: ignition-fetch.service was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Apr 12 18:20:09.821290 systemd[1]: Starting ignition-kargs.service... Apr 12 18:20:09.830053 ignition[720]: Ignition 2.14.0 Apr 12 18:20:09.830062 ignition[720]: Stage: kargs Apr 12 18:20:09.830147 ignition[720]: no configs at "/usr/lib/ignition/base.d" Apr 12 18:20:09.830156 ignition[720]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 12 18:20:09.832264 systemd[1]: Finished ignition-kargs.service. Apr 12 18:20:09.833000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.831155 ignition[720]: kargs: kargs passed Apr 12 18:20:09.831194 ignition[720]: Ignition finished successfully Apr 12 18:20:09.834597 systemd[1]: Starting ignition-disks.service... Apr 12 18:20:09.840776 ignition[726]: Ignition 2.14.0 Apr 12 18:20:09.840791 ignition[726]: Stage: disks Apr 12 18:20:09.840876 ignition[726]: no configs at "/usr/lib/ignition/base.d" Apr 12 18:20:09.840885 ignition[726]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 12 18:20:09.843792 systemd[1]: Finished ignition-disks.service. Apr 12 18:20:09.844000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.841959 ignition[726]: disks: disks passed Apr 12 18:20:09.844992 systemd[1]: Reached target initrd-root-device.target. Apr 12 18:20:09.841997 ignition[726]: Ignition finished successfully Apr 12 18:20:09.846238 systemd[1]: Reached target local-fs-pre.target. Apr 12 18:20:09.847441 systemd[1]: Reached target local-fs.target. Apr 12 18:20:09.848776 systemd[1]: Reached target sysinit.target. Apr 12 18:20:09.850001 systemd[1]: Reached target basic.target. Apr 12 18:20:09.852773 systemd[1]: Starting systemd-fsck-root.service... Apr 12 18:20:09.863838 systemd-fsck[734]: ROOT: clean, 612/553520 files, 56018/553472 blocks Apr 12 18:20:09.868788 systemd[1]: Finished systemd-fsck-root.service. Apr 12 18:20:09.868000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.870271 systemd[1]: Mounting sysroot.mount... Apr 12 18:20:09.885629 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Apr 12 18:20:09.885341 systemd[1]: Mounted sysroot.mount. Apr 12 18:20:09.886092 systemd[1]: Reached target initrd-root-fs.target. Apr 12 18:20:09.888151 systemd[1]: Mounting sysroot-usr.mount... Apr 12 18:20:09.889059 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Apr 12 18:20:09.889093 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 12 18:20:09.889115 systemd[1]: Reached target ignition-diskful.target. Apr 12 18:20:09.890853 systemd[1]: Mounted sysroot-usr.mount. Apr 12 18:20:09.892698 systemd[1]: Starting initrd-setup-root.service... Apr 12 18:20:09.896969 initrd-setup-root[744]: cut: /sysroot/etc/passwd: No such file or directory Apr 12 18:20:09.902826 initrd-setup-root[752]: cut: /sysroot/etc/group: No such file or directory Apr 12 18:20:09.909011 initrd-setup-root[760]: cut: /sysroot/etc/shadow: No such file or directory Apr 12 18:20:09.913463 initrd-setup-root[768]: cut: /sysroot/etc/gshadow: No such file or directory Apr 12 18:20:09.941952 systemd[1]: Finished initrd-setup-root.service. Apr 12 18:20:09.942000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.943521 systemd[1]: Starting ignition-mount.service... Apr 12 18:20:09.944890 systemd[1]: Starting sysroot-boot.service... Apr 12 18:20:09.949157 bash[785]: umount: /sysroot/usr/share/oem: not mounted. Apr 12 18:20:09.958435 ignition[787]: INFO : Ignition 2.14.0 Apr 12 18:20:09.958435 ignition[787]: INFO : Stage: mount Apr 12 18:20:09.959990 ignition[787]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 12 18:20:09.959990 ignition[787]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 12 18:20:09.959990 ignition[787]: INFO : mount: mount passed Apr 12 18:20:09.959990 ignition[787]: INFO : Ignition finished successfully Apr 12 18:20:09.962000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:09.961026 systemd[1]: Finished ignition-mount.service. Apr 12 18:20:09.966656 systemd[1]: Finished sysroot-boot.service. Apr 12 18:20:09.966000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:10.540261 systemd[1]: Mounting sysroot-usr-share-oem.mount... Apr 12 18:20:10.547292 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (795) Apr 12 18:20:10.547322 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm Apr 12 18:20:10.547332 kernel: BTRFS info (device vda6): using free space tree Apr 12 18:20:10.548579 kernel: BTRFS info (device vda6): has skinny extents Apr 12 18:20:10.551333 systemd[1]: Mounted sysroot-usr-share-oem.mount. Apr 12 18:20:10.552866 systemd[1]: Starting ignition-files.service... Apr 12 18:20:10.566388 ignition[815]: INFO : Ignition 2.14.0 Apr 12 18:20:10.566388 ignition[815]: INFO : Stage: files Apr 12 18:20:10.567883 ignition[815]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 12 18:20:10.567883 ignition[815]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 12 18:20:10.567883 ignition[815]: DEBUG : files: compiled without relabeling support, skipping Apr 12 18:20:10.572211 ignition[815]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 12 18:20:10.572211 ignition[815]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 12 18:20:10.575868 ignition[815]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 12 18:20:10.577160 ignition[815]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 12 18:20:10.577160 ignition[815]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 12 18:20:10.577160 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 12 18:20:10.576609 unknown[815]: wrote ssh authorized keys file for user: core Apr 12 18:20:10.582385 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Apr 12 18:20:10.616901 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 12 18:20:10.656064 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 12 18:20:10.658349 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/crictl-v1.27.0-linux-arm64.tar.gz" Apr 12 18:20:10.658349 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.27.0/crictl-v1.27.0-linux-arm64.tar.gz: attempt #1 Apr 12 18:20:10.989095 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Apr 12 18:20:11.130907 ignition[815]: DEBUG : files: createFilesystemsFiles: createFiles: op(4): file matches expected sum of: db062e43351a63347871e7094115be2ae3853afcd346d47f7b51141da8c3202c2df58d2e17359322f632abcb37474fd7fdb3b7aadbc5cfd5cf6d3bad040b6251 Apr 12 18:20:11.130907 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/crictl-v1.27.0-linux-arm64.tar.gz" Apr 12 18:20:11.135234 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/opt/cni-plugins-linux-arm64-v1.3.0.tgz" Apr 12 18:20:11.135234 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/containernetworking/plugins/releases/download/v1.3.0/cni-plugins-linux-arm64-v1.3.0.tgz: attempt #1 Apr 12 18:20:11.378237 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(5): GET result: OK Apr 12 18:20:11.391743 systemd-networkd[698]: eth0: Gained IPv6LL Apr 12 18:20:11.612484 ignition[815]: DEBUG : files: createFilesystemsFiles: createFiles: op(5): file matches expected sum of: b2b7fb74f1b3cb8928f49e5bf9d4bc686e057e837fac3caf1b366d54757921dba80d70cc010399b274d136e8dee9a25b1ad87cdfdc4ffcf42cf88f3e8f99587a Apr 12 18:20:11.612484 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/cni-plugins-linux-arm64-v1.3.0.tgz" Apr 12 18:20:11.617261 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/etc/docker/daemon.json" Apr 12 18:20:11.617261 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/etc/docker/daemon.json" Apr 12 18:20:11.617261 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/bin/kubectl" Apr 12 18:20:11.617261 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://dl.k8s.io/release/v1.28.1/bin/linux/arm64/kubectl: attempt #1 Apr 12 18:20:11.724953 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Apr 12 18:20:12.299859 ignition[815]: DEBUG : files: createFilesystemsFiles: createFiles: op(7): file matches expected sum of: 6a5c9c02a29126949f096415bb1761a0c0ad44168e2ab3d0409982701da58f96223bec354828ddf958e945ef1ce63c0ad41e77cbcbcce0756163e71b4fbae432 Apr 12 18:20:12.302512 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/bin/kubectl" Apr 12 18:20:12.302512 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/opt/bin/kubelet" Apr 12 18:20:12.302512 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET https://dl.k8s.io/release/v1.28.1/bin/linux/arm64/kubelet: attempt #1 Apr 12 18:20:12.325981 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET result: OK Apr 12 18:20:12.911992 ignition[815]: DEBUG : files: createFilesystemsFiles: createFiles: op(8): file matches expected sum of: 5a898ef543a6482895101ea58e33602e3c0a7682d322aaf08ac3dc8a5a3c8da8f09600d577024549288f8cebb1a86f9c79927796b69a3d8fe989ca8f12b147d6 Apr 12 18:20:12.915009 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/opt/bin/kubelet" Apr 12 18:20:12.915009 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/opt/bin/kubeadm" Apr 12 18:20:12.915009 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(9): GET https://dl.k8s.io/release/v1.28.1/bin/linux/arm64/kubeadm: attempt #1 Apr 12 18:20:12.935763 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(9): GET result: OK Apr 12 18:20:13.209042 ignition[815]: DEBUG : files: createFilesystemsFiles: createFiles: op(9): file matches expected sum of: 5a08b81f9cc82d3cce21130856ca63b8dafca9149d9775dd25b376eb0f18209aa0e4a47c0a6d7e6fb1316aacd5d59dec770f26c09120c866949d70bc415518b3 Apr 12 18:20:13.211705 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/opt/bin/kubeadm" Apr 12 18:20:13.213231 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 12 18:20:13.213231 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Apr 12 18:20:13.501252 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Apr 12 18:20:13.548980 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/home/core/install.sh" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/home/core/install.sh" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(d): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(d): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(e): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(e): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(f): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 12 18:20:13.550689 ignition[815]: INFO : files: createFilesystemsFiles: createFiles: op(f): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(10): [started] processing unit "prepare-cni-plugins.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(10): op(11): [started] writing unit "prepare-cni-plugins.service" at "/sysroot/etc/systemd/system/prepare-cni-plugins.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(10): op(11): [finished] writing unit "prepare-cni-plugins.service" at "/sysroot/etc/systemd/system/prepare-cni-plugins.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(10): [finished] processing unit "prepare-cni-plugins.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(12): [started] processing unit "prepare-critools.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(12): op(13): [started] writing unit "prepare-critools.service" at "/sysroot/etc/systemd/system/prepare-critools.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(12): op(13): [finished] writing unit "prepare-critools.service" at "/sysroot/etc/systemd/system/prepare-critools.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(12): [finished] processing unit "prepare-critools.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(14): [started] processing unit "prepare-helm.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(14): op(15): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(14): op(15): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(14): [finished] processing unit "prepare-helm.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(16): [started] processing unit "coreos-metadata.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(16): op(17): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(16): op(17): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(16): [finished] processing unit "coreos-metadata.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(18): [started] setting preset to disabled for "coreos-metadata.service" Apr 12 18:20:13.568369 ignition[815]: INFO : files: op(18): op(19): [started] removing enablement symlink(s) for "coreos-metadata.service" Apr 12 18:20:13.596613 ignition[815]: INFO : files: op(18): op(19): [finished] removing enablement symlink(s) for "coreos-metadata.service" Apr 12 18:20:13.596613 ignition[815]: INFO : files: op(18): [finished] setting preset to disabled for "coreos-metadata.service" Apr 12 18:20:13.596613 ignition[815]: INFO : files: op(1a): [started] setting preset to enabled for "prepare-cni-plugins.service" Apr 12 18:20:13.596613 ignition[815]: INFO : files: op(1a): [finished] setting preset to enabled for "prepare-cni-plugins.service" Apr 12 18:20:13.596613 ignition[815]: INFO : files: op(1b): [started] setting preset to enabled for "prepare-critools.service" Apr 12 18:20:13.596613 ignition[815]: INFO : files: op(1b): [finished] setting preset to enabled for "prepare-critools.service" Apr 12 18:20:13.596613 ignition[815]: INFO : files: op(1c): [started] setting preset to enabled for "prepare-helm.service" Apr 12 18:20:13.596613 ignition[815]: INFO : files: op(1c): [finished] setting preset to enabled for "prepare-helm.service" Apr 12 18:20:13.596613 ignition[815]: INFO : files: createResultFile: createFiles: op(1d): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 12 18:20:13.596613 ignition[815]: INFO : files: createResultFile: createFiles: op(1d): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 12 18:20:13.596613 ignition[815]: INFO : files: files passed Apr 12 18:20:13.596613 ignition[815]: INFO : Ignition finished successfully Apr 12 18:20:13.625224 kernel: kauditd_printk_skb: 22 callbacks suppressed Apr 12 18:20:13.625247 kernel: audit: type=1130 audit(1712946013.596:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.625258 kernel: audit: type=1130 audit(1712946013.611:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.625268 kernel: audit: type=1130 audit(1712946013.618:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.625277 kernel: audit: type=1131 audit(1712946013.618:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.596000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.611000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.618000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.618000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.595637 systemd[1]: Finished ignition-files.service. Apr 12 18:20:13.598279 systemd[1]: Starting initrd-setup-root-after-ignition.service... Apr 12 18:20:13.607271 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Apr 12 18:20:13.629060 initrd-setup-root-after-ignition[839]: grep: /sysroot/usr/share/oem/oem-release: No such file or directory Apr 12 18:20:13.635121 kernel: audit: type=1130 audit(1712946013.629:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.635140 kernel: audit: type=1131 audit(1712946013.629:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.629000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.629000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.608032 systemd[1]: Starting ignition-quench.service... Apr 12 18:20:13.636522 initrd-setup-root-after-ignition[842]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 12 18:20:13.609542 systemd[1]: Finished initrd-setup-root-after-ignition.service. Apr 12 18:20:13.611237 systemd[1]: Reached target ignition-complete.target. Apr 12 18:20:13.616207 systemd[1]: Starting initrd-parse-etc.service... Apr 12 18:20:13.617641 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 12 18:20:13.617730 systemd[1]: Finished ignition-quench.service. Apr 12 18:20:13.628469 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 12 18:20:13.628550 systemd[1]: Finished initrd-parse-etc.service. Apr 12 18:20:13.629805 systemd[1]: Reached target initrd-fs.target. Apr 12 18:20:13.635829 systemd[1]: Reached target initrd.target. Apr 12 18:20:13.637096 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Apr 12 18:20:13.637763 systemd[1]: Starting dracut-pre-pivot.service... Apr 12 18:20:13.647504 systemd[1]: Finished dracut-pre-pivot.service. Apr 12 18:20:13.648000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.649270 systemd[1]: Starting initrd-cleanup.service... Apr 12 18:20:13.652565 kernel: audit: type=1130 audit(1712946013.648:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.656688 systemd[1]: Stopped target nss-lookup.target. Apr 12 18:20:13.657479 systemd[1]: Stopped target remote-cryptsetup.target. Apr 12 18:20:13.658852 systemd[1]: Stopped target timers.target. Apr 12 18:20:13.660126 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 12 18:20:13.660000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.660223 systemd[1]: Stopped dracut-pre-pivot.service. Apr 12 18:20:13.665380 kernel: audit: type=1131 audit(1712946013.660:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.661446 systemd[1]: Stopped target initrd.target. Apr 12 18:20:13.664892 systemd[1]: Stopped target basic.target. Apr 12 18:20:13.666069 systemd[1]: Stopped target ignition-complete.target. Apr 12 18:20:13.667327 systemd[1]: Stopped target ignition-diskful.target. Apr 12 18:20:13.668540 systemd[1]: Stopped target initrd-root-device.target. Apr 12 18:20:13.669950 systemd[1]: Stopped target remote-fs.target. Apr 12 18:20:13.671231 systemd[1]: Stopped target remote-fs-pre.target. Apr 12 18:20:13.672561 systemd[1]: Stopped target sysinit.target. Apr 12 18:20:13.673788 systemd[1]: Stopped target local-fs.target. Apr 12 18:20:13.675011 systemd[1]: Stopped target local-fs-pre.target. Apr 12 18:20:13.676242 systemd[1]: Stopped target swap.target. Apr 12 18:20:13.678000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.677360 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 12 18:20:13.682630 kernel: audit: type=1131 audit(1712946013.678:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.677457 systemd[1]: Stopped dracut-pre-mount.service. Apr 12 18:20:13.682000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.678721 systemd[1]: Stopped target cryptsetup.target. Apr 12 18:20:13.687383 kernel: audit: type=1131 audit(1712946013.682:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.686000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.681938 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 12 18:20:13.682034 systemd[1]: Stopped dracut-initqueue.service. Apr 12 18:20:13.683428 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 12 18:20:13.683517 systemd[1]: Stopped ignition-fetch-offline.service. Apr 12 18:20:13.686948 systemd[1]: Stopped target paths.target. Apr 12 18:20:13.688042 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 12 18:20:13.691607 systemd[1]: Stopped systemd-ask-password-console.path. Apr 12 18:20:13.692640 systemd[1]: Stopped target slices.target. Apr 12 18:20:13.694130 systemd[1]: Stopped target sockets.target. Apr 12 18:20:13.696000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.695494 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 12 18:20:13.697000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.695624 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Apr 12 18:20:13.700338 iscsid[705]: iscsid shutting down. Apr 12 18:20:13.696947 systemd[1]: ignition-files.service: Deactivated successfully. Apr 12 18:20:13.701000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.697041 systemd[1]: Stopped ignition-files.service. Apr 12 18:20:13.698865 systemd[1]: Stopping ignition-mount.service... Apr 12 18:20:13.699670 systemd[1]: Stopping iscsid.service... Apr 12 18:20:13.700752 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 12 18:20:13.704000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.705000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.707497 ignition[856]: INFO : Ignition 2.14.0 Apr 12 18:20:13.707497 ignition[856]: INFO : Stage: umount Apr 12 18:20:13.707497 ignition[856]: INFO : no configs at "/usr/lib/ignition/base.d" Apr 12 18:20:13.707497 ignition[856]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Apr 12 18:20:13.707497 ignition[856]: INFO : umount: umount passed Apr 12 18:20:13.707497 ignition[856]: INFO : Ignition finished successfully Apr 12 18:20:13.709000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.710000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.700881 systemd[1]: Stopped kmod-static-nodes.service. Apr 12 18:20:13.714000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.702810 systemd[1]: Stopping sysroot-boot.service... Apr 12 18:20:13.715000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.703958 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 12 18:20:13.717000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.704088 systemd[1]: Stopped systemd-udev-trigger.service. Apr 12 18:20:13.705490 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 12 18:20:13.720000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.720000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.705604 systemd[1]: Stopped dracut-pre-trigger.service. Apr 12 18:20:13.721000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.708621 systemd[1]: iscsid.service: Deactivated successfully. Apr 12 18:20:13.722000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.708713 systemd[1]: Stopped iscsid.service. Apr 12 18:20:13.710036 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 12 18:20:13.710107 systemd[1]: Stopped ignition-mount.service. Apr 12 18:20:13.726000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.712275 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 12 18:20:13.712793 systemd[1]: iscsid.socket: Deactivated successfully. Apr 12 18:20:13.712861 systemd[1]: Closed iscsid.socket. Apr 12 18:20:13.713850 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 12 18:20:13.713889 systemd[1]: Stopped ignition-disks.service. Apr 12 18:20:13.715269 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 12 18:20:13.715324 systemd[1]: Stopped ignition-kargs.service. Apr 12 18:20:13.716512 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 12 18:20:13.716552 systemd[1]: Stopped ignition-setup.service. Apr 12 18:20:13.717831 systemd[1]: Stopping iscsiuio.service... Apr 12 18:20:13.719762 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 12 18:20:13.719849 systemd[1]: Finished initrd-cleanup.service. Apr 12 18:20:13.721111 systemd[1]: iscsiuio.service: Deactivated successfully. Apr 12 18:20:13.739000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.721186 systemd[1]: Stopped iscsiuio.service. Apr 12 18:20:13.741000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.722433 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 12 18:20:13.722499 systemd[1]: Stopped sysroot-boot.service. Apr 12 18:20:13.724175 systemd[1]: Stopped target network.target. Apr 12 18:20:13.744000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.724910 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 12 18:20:13.746000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.724942 systemd[1]: Closed iscsiuio.socket. Apr 12 18:20:13.746000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.750000 audit: BPF prog-id=6 op=UNLOAD Apr 12 18:20:13.725953 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 12 18:20:13.725991 systemd[1]: Stopped initrd-setup-root.service. Apr 12 18:20:13.727468 systemd[1]: Stopping systemd-networkd.service... Apr 12 18:20:13.728897 systemd[1]: Stopping systemd-resolved.service... Apr 12 18:20:13.738180 systemd-networkd[698]: eth0: DHCPv6 lease lost Apr 12 18:20:13.753000 audit: BPF prog-id=9 op=UNLOAD Apr 12 18:20:13.738449 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 12 18:20:13.755000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.738533 systemd[1]: Stopped systemd-resolved.service. Apr 12 18:20:13.740174 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 12 18:20:13.740259 systemd[1]: Stopped systemd-networkd.service. Apr 12 18:20:13.741399 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 12 18:20:13.741424 systemd[1]: Closed systemd-networkd.socket. Apr 12 18:20:13.759000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.742944 systemd[1]: Stopping network-cleanup.service... Apr 12 18:20:13.743583 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 12 18:20:13.763000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.743641 systemd[1]: Stopped parse-ip-for-networkd.service. Apr 12 18:20:13.764000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.744968 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 12 18:20:13.766000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.745008 systemd[1]: Stopped systemd-sysctl.service. Apr 12 18:20:13.746843 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 12 18:20:13.768000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.746880 systemd[1]: Stopped systemd-modules-load.service. Apr 12 18:20:13.747727 systemd[1]: Stopping systemd-udevd.service... Apr 12 18:20:13.752322 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Apr 12 18:20:13.754728 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 12 18:20:13.773000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.773000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:13.754822 systemd[1]: Stopped network-cleanup.service. Apr 12 18:20:13.758525 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 12 18:20:13.758718 systemd[1]: Stopped systemd-udevd.service. Apr 12 18:20:13.760080 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 12 18:20:13.760115 systemd[1]: Closed systemd-udevd-control.socket. Apr 12 18:20:13.761415 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 12 18:20:13.761446 systemd[1]: Closed systemd-udevd-kernel.socket. Apr 12 18:20:13.762656 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 12 18:20:13.762696 systemd[1]: Stopped dracut-pre-udev.service. Apr 12 18:20:13.763927 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 12 18:20:13.763963 systemd[1]: Stopped dracut-cmdline.service. Apr 12 18:20:13.765411 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 12 18:20:13.765448 systemd[1]: Stopped dracut-cmdline-ask.service. Apr 12 18:20:13.767380 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Apr 12 18:20:13.768196 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 12 18:20:13.768248 systemd[1]: Stopped systemd-vconsole-setup.service. Apr 12 18:20:13.772300 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 12 18:20:13.772379 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Apr 12 18:20:13.773728 systemd[1]: Reached target initrd-switch-root.target. Apr 12 18:20:13.775866 systemd[1]: Starting initrd-switch-root.service... Apr 12 18:20:13.781680 systemd[1]: Switching root. Apr 12 18:20:13.798889 systemd-journald[250]: Journal stopped Apr 12 18:20:15.878327 systemd-journald[250]: Received SIGTERM from PID 1 (systemd). Apr 12 18:20:15.878381 kernel: SELinux: Class mctp_socket not defined in policy. Apr 12 18:20:15.878397 kernel: SELinux: Class anon_inode not defined in policy. Apr 12 18:20:15.878408 kernel: SELinux: the above unknown classes and permissions will be allowed Apr 12 18:20:15.878418 kernel: SELinux: policy capability network_peer_controls=1 Apr 12 18:20:15.878428 kernel: SELinux: policy capability open_perms=1 Apr 12 18:20:15.878437 kernel: SELinux: policy capability extended_socket_class=1 Apr 12 18:20:15.878447 kernel: SELinux: policy capability always_check_network=0 Apr 12 18:20:15.878457 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 12 18:20:15.878469 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 12 18:20:15.878479 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 12 18:20:15.878491 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 12 18:20:15.878502 systemd[1]: Successfully loaded SELinux policy in 32.308ms. Apr 12 18:20:15.878518 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 8.588ms. Apr 12 18:20:15.878530 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Apr 12 18:20:15.878542 systemd[1]: Detected virtualization kvm. Apr 12 18:20:15.878553 systemd[1]: Detected architecture arm64. Apr 12 18:20:15.878564 systemd[1]: Detected first boot. Apr 12 18:20:15.878586 systemd[1]: Initializing machine ID from VM UUID. Apr 12 18:20:15.878598 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Apr 12 18:20:15.878608 systemd[1]: Populated /etc with preset unit settings. Apr 12 18:20:15.878619 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Apr 12 18:20:15.878631 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Apr 12 18:20:15.878643 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 12 18:20:15.878656 systemd[1]: initrd-switch-root.service: Deactivated successfully. Apr 12 18:20:15.878667 systemd[1]: Stopped initrd-switch-root.service. Apr 12 18:20:15.878678 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Apr 12 18:20:15.878689 systemd[1]: Created slice system-addon\x2dconfig.slice. Apr 12 18:20:15.878701 systemd[1]: Created slice system-addon\x2drun.slice. Apr 12 18:20:15.878715 systemd[1]: Created slice system-getty.slice. Apr 12 18:20:15.878726 systemd[1]: Created slice system-modprobe.slice. Apr 12 18:20:15.878737 systemd[1]: Created slice system-serial\x2dgetty.slice. Apr 12 18:20:15.878749 systemd[1]: Created slice system-system\x2dcloudinit.slice. Apr 12 18:20:15.878760 systemd[1]: Created slice system-systemd\x2dfsck.slice. Apr 12 18:20:15.878778 systemd[1]: Created slice user.slice. Apr 12 18:20:15.878790 systemd[1]: Started systemd-ask-password-console.path. Apr 12 18:20:15.878801 systemd[1]: Started systemd-ask-password-wall.path. Apr 12 18:20:15.878812 systemd[1]: Set up automount boot.automount. Apr 12 18:20:15.878823 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Apr 12 18:20:15.878833 systemd[1]: Stopped target initrd-switch-root.target. Apr 12 18:20:15.878844 systemd[1]: Stopped target initrd-fs.target. Apr 12 18:20:15.878857 systemd[1]: Stopped target initrd-root-fs.target. Apr 12 18:20:15.878868 systemd[1]: Reached target integritysetup.target. Apr 12 18:20:15.878879 systemd[1]: Reached target remote-cryptsetup.target. Apr 12 18:20:15.878889 systemd[1]: Reached target remote-fs.target. Apr 12 18:20:15.878900 systemd[1]: Reached target slices.target. Apr 12 18:20:15.878910 systemd[1]: Reached target swap.target. Apr 12 18:20:15.878921 systemd[1]: Reached target torcx.target. Apr 12 18:20:15.878932 systemd[1]: Reached target veritysetup.target. Apr 12 18:20:15.878946 systemd[1]: Listening on systemd-coredump.socket. Apr 12 18:20:15.878957 systemd[1]: Listening on systemd-initctl.socket. Apr 12 18:20:15.878968 systemd[1]: Listening on systemd-networkd.socket. Apr 12 18:20:15.878979 systemd[1]: Listening on systemd-udevd-control.socket. Apr 12 18:20:15.878989 systemd[1]: Listening on systemd-udevd-kernel.socket. Apr 12 18:20:15.879000 systemd[1]: Listening on systemd-userdbd.socket. Apr 12 18:20:15.879011 systemd[1]: Mounting dev-hugepages.mount... Apr 12 18:20:15.879021 systemd[1]: Mounting dev-mqueue.mount... Apr 12 18:20:15.879032 systemd[1]: Mounting media.mount... Apr 12 18:20:15.879042 systemd[1]: Mounting sys-kernel-debug.mount... Apr 12 18:20:15.879055 systemd[1]: Mounting sys-kernel-tracing.mount... Apr 12 18:20:15.879065 systemd[1]: Mounting tmp.mount... Apr 12 18:20:15.879077 systemd[1]: Starting flatcar-tmpfiles.service... Apr 12 18:20:15.879088 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Apr 12 18:20:15.879099 systemd[1]: Starting kmod-static-nodes.service... Apr 12 18:20:15.879110 systemd[1]: Starting modprobe@configfs.service... Apr 12 18:20:15.879121 systemd[1]: Starting modprobe@dm_mod.service... Apr 12 18:20:15.879132 systemd[1]: Starting modprobe@drm.service... Apr 12 18:20:15.879142 systemd[1]: Starting modprobe@efi_pstore.service... Apr 12 18:20:15.879156 systemd[1]: Starting modprobe@fuse.service... Apr 12 18:20:15.879167 systemd[1]: Starting modprobe@loop.service... Apr 12 18:20:15.879178 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 12 18:20:15.879189 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Apr 12 18:20:15.879200 systemd[1]: Stopped systemd-fsck-root.service. Apr 12 18:20:15.879210 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Apr 12 18:20:15.879220 kernel: loop: module loaded Apr 12 18:20:15.879232 systemd[1]: Stopped systemd-fsck-usr.service. Apr 12 18:20:15.879244 systemd[1]: Stopped systemd-journald.service. Apr 12 18:20:15.879254 systemd[1]: Starting systemd-journald.service... Apr 12 18:20:15.879265 systemd[1]: Starting systemd-modules-load.service... Apr 12 18:20:15.879277 systemd[1]: Starting systemd-network-generator.service... Apr 12 18:20:15.879287 kernel: fuse: init (API version 7.34) Apr 12 18:20:15.879297 systemd[1]: Starting systemd-remount-fs.service... Apr 12 18:20:15.879307 systemd[1]: Starting systemd-udev-trigger.service... Apr 12 18:20:15.879318 systemd[1]: verity-setup.service: Deactivated successfully. Apr 12 18:20:15.879328 systemd[1]: Stopped verity-setup.service. Apr 12 18:20:15.879339 systemd[1]: Mounted dev-hugepages.mount. Apr 12 18:20:15.879351 systemd-journald[951]: Journal started Apr 12 18:20:15.879391 systemd-journald[951]: Runtime Journal (/run/log/journal/4532c036a0be47be89214dd3f1adde8f) is 6.0M, max 48.7M, 42.6M free. Apr 12 18:20:15.879421 systemd[1]: Mounted dev-mqueue.mount. Apr 12 18:20:13.871000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Apr 12 18:20:14.043000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Apr 12 18:20:14.043000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Apr 12 18:20:14.043000 audit: BPF prog-id=10 op=LOAD Apr 12 18:20:14.043000 audit: BPF prog-id=10 op=UNLOAD Apr 12 18:20:14.043000 audit: BPF prog-id=11 op=LOAD Apr 12 18:20:14.043000 audit: BPF prog-id=11 op=UNLOAD Apr 12 18:20:14.088000 audit[889]: AVC avc: denied { associate } for pid=889 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Apr 12 18:20:14.088000 audit[889]: SYSCALL arch=c00000b7 syscall=5 success=yes exit=0 a0=40001c58ac a1=40000c8de0 a2=40000cf0c0 a3=32 items=0 ppid=872 pid=889 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:20:14.088000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Apr 12 18:20:14.089000 audit[889]: AVC avc: denied { associate } for pid=889 comm="torcx-generator" name="bin" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Apr 12 18:20:14.089000 audit[889]: SYSCALL arch=c00000b7 syscall=34 success=yes exit=0 a0=ffffffffffffff9c a1=40001c5989 a2=1ed a3=0 items=2 ppid=872 pid=889 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:20:14.089000 audit: CWD cwd="/" Apr 12 18:20:14.089000 audit: PATH item=0 name=(null) inode=2 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:20:14.089000 audit: PATH item=1 name=(null) inode=3 dev=00:1c mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:20:14.089000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Apr 12 18:20:15.754000 audit: BPF prog-id=12 op=LOAD Apr 12 18:20:15.754000 audit: BPF prog-id=3 op=UNLOAD Apr 12 18:20:15.754000 audit: BPF prog-id=13 op=LOAD Apr 12 18:20:15.754000 audit: BPF prog-id=14 op=LOAD Apr 12 18:20:15.754000 audit: BPF prog-id=4 op=UNLOAD Apr 12 18:20:15.754000 audit: BPF prog-id=5 op=UNLOAD Apr 12 18:20:15.755000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.759000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.759000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.766000 audit: BPF prog-id=12 op=UNLOAD Apr 12 18:20:15.849000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.851000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.854000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.854000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.856000 audit: BPF prog-id=15 op=LOAD Apr 12 18:20:15.856000 audit: BPF prog-id=16 op=LOAD Apr 12 18:20:15.856000 audit: BPF prog-id=17 op=LOAD Apr 12 18:20:15.856000 audit: BPF prog-id=13 op=UNLOAD Apr 12 18:20:15.856000 audit: BPF prog-id=14 op=UNLOAD Apr 12 18:20:15.875000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.876000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Apr 12 18:20:15.876000 audit[951]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=60 a0=6 a1=fffff95122a0 a2=4000 a3=1 items=0 ppid=1 pid=951 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:20:15.876000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Apr 12 18:20:15.753402 systemd[1]: Queued start job for default target multi-user.target. Apr 12 18:20:14.086961 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.3 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.3 /var/lib/torcx/store]" Apr 12 18:20:15.753413 systemd[1]: Unnecessary job was removed for dev-vda6.device. Apr 12 18:20:14.087189 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Apr 12 18:20:15.755997 systemd[1]: systemd-journald.service: Deactivated successfully. Apr 12 18:20:14.087206 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Apr 12 18:20:14.087234 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Apr 12 18:20:14.087243 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="skipped missing lower profile" missing profile=oem Apr 12 18:20:14.087269 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Apr 12 18:20:14.087280 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Apr 12 18:20:14.087474 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Apr 12 18:20:14.087506 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Apr 12 18:20:14.087518 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Apr 12 18:20:14.088307 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Apr 12 18:20:14.088343 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Apr 12 18:20:14.088361 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.3: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.3 Apr 12 18:20:14.088375 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Apr 12 18:20:15.881585 systemd[1]: Started systemd-journald.service. Apr 12 18:20:14.088392 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.3: no such file or directory" path=/var/lib/torcx/store/3510.3.3 Apr 12 18:20:14.088404 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:14Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Apr 12 18:20:15.505904 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:15Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Apr 12 18:20:15.506155 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:15Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Apr 12 18:20:15.506248 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:15Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Apr 12 18:20:15.506407 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:15Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Apr 12 18:20:15.506453 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:15Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Apr 12 18:20:15.506508 /usr/lib/systemd/system-generators/torcx-generator[889]: time="2024-04-12T18:20:15Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Apr 12 18:20:15.881000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.882893 systemd[1]: Mounted media.mount. Apr 12 18:20:15.883639 systemd[1]: Mounted sys-kernel-debug.mount. Apr 12 18:20:15.884479 systemd[1]: Mounted sys-kernel-tracing.mount. Apr 12 18:20:15.885410 systemd[1]: Mounted tmp.mount. Apr 12 18:20:15.886329 systemd[1]: Finished kmod-static-nodes.service. Apr 12 18:20:15.887000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.887401 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 12 18:20:15.887529 systemd[1]: Finished modprobe@configfs.service. Apr 12 18:20:15.888000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.888000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.888641 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 12 18:20:15.889282 systemd[1]: Finished modprobe@dm_mod.service. Apr 12 18:20:15.889000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.889000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.890402 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 12 18:20:15.890562 systemd[1]: Finished modprobe@drm.service. Apr 12 18:20:15.890000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.890000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.891544 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 12 18:20:15.891692 systemd[1]: Finished modprobe@efi_pstore.service. Apr 12 18:20:15.891000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.891000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.893000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.893000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.892877 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 12 18:20:15.893148 systemd[1]: Finished modprobe@fuse.service. Apr 12 18:20:15.894185 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 12 18:20:15.894316 systemd[1]: Finished modprobe@loop.service. Apr 12 18:20:15.894000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.894000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.895380 systemd[1]: Finished systemd-modules-load.service. Apr 12 18:20:15.896000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.896608 systemd[1]: Finished systemd-network-generator.service. Apr 12 18:20:15.896000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.897000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.897705 systemd[1]: Finished systemd-remount-fs.service. Apr 12 18:20:15.899001 systemd[1]: Reached target network-pre.target. Apr 12 18:20:15.900973 systemd[1]: Mounting sys-fs-fuse-connections.mount... Apr 12 18:20:15.902683 systemd[1]: Mounting sys-kernel-config.mount... Apr 12 18:20:15.903375 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 12 18:20:15.905011 systemd[1]: Starting systemd-hwdb-update.service... Apr 12 18:20:15.906903 systemd[1]: Starting systemd-journal-flush.service... Apr 12 18:20:15.907757 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 12 18:20:15.908729 systemd[1]: Starting systemd-random-seed.service... Apr 12 18:20:15.909507 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Apr 12 18:20:15.910513 systemd[1]: Starting systemd-sysctl.service... Apr 12 18:20:15.916994 systemd-journald[951]: Time spent on flushing to /var/log/journal/4532c036a0be47be89214dd3f1adde8f is 22.372ms for 1019 entries. Apr 12 18:20:15.916994 systemd-journald[951]: System Journal (/var/log/journal/4532c036a0be47be89214dd3f1adde8f) is 8.0M, max 195.6M, 187.6M free. Apr 12 18:20:15.950553 systemd-journald[951]: Received client request to flush runtime journal. Apr 12 18:20:15.923000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.924000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.929000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.937000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.945000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:15.914136 systemd[1]: Mounted sys-fs-fuse-connections.mount. Apr 12 18:20:15.915123 systemd[1]: Mounted sys-kernel-config.mount. Apr 12 18:20:15.922980 systemd[1]: Finished flatcar-tmpfiles.service. Apr 12 18:20:15.951238 udevadm[990]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Apr 12 18:20:15.924266 systemd[1]: Finished systemd-random-seed.service. Apr 12 18:20:15.925185 systemd[1]: Reached target first-boot-complete.target. Apr 12 18:20:15.927038 systemd[1]: Starting systemd-sysusers.service... Apr 12 18:20:15.929098 systemd[1]: Finished systemd-udev-trigger.service. Apr 12 18:20:15.931137 systemd[1]: Starting systemd-udev-settle.service... Apr 12 18:20:15.937096 systemd[1]: Finished systemd-sysctl.service. Apr 12 18:20:15.944829 systemd[1]: Finished systemd-sysusers.service. Apr 12 18:20:15.951326 systemd[1]: Finished systemd-journal-flush.service. Apr 12 18:20:15.952000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.300589 systemd[1]: Finished systemd-hwdb-update.service. Apr 12 18:20:16.300000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.301000 audit: BPF prog-id=18 op=LOAD Apr 12 18:20:16.301000 audit: BPF prog-id=19 op=LOAD Apr 12 18:20:16.301000 audit: BPF prog-id=7 op=UNLOAD Apr 12 18:20:16.301000 audit: BPF prog-id=8 op=UNLOAD Apr 12 18:20:16.302877 systemd[1]: Starting systemd-udevd.service... Apr 12 18:20:16.319644 systemd-udevd[992]: Using default interface naming scheme 'v252'. Apr 12 18:20:16.330669 systemd[1]: Started systemd-udevd.service. Apr 12 18:20:16.330000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.331000 audit: BPF prog-id=20 op=LOAD Apr 12 18:20:16.333109 systemd[1]: Starting systemd-networkd.service... Apr 12 18:20:16.340000 audit: BPF prog-id=21 op=LOAD Apr 12 18:20:16.340000 audit: BPF prog-id=22 op=LOAD Apr 12 18:20:16.340000 audit: BPF prog-id=23 op=LOAD Apr 12 18:20:16.342136 systemd[1]: Starting systemd-userdbd.service... Apr 12 18:20:16.347811 systemd[1]: Condition check resulted in dev-ttyAMA0.device being skipped. Apr 12 18:20:16.377703 systemd[1]: Started systemd-userdbd.service. Apr 12 18:20:16.377000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.379700 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Apr 12 18:20:16.416923 systemd[1]: Finished systemd-udev-settle.service. Apr 12 18:20:16.417000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.419084 systemd[1]: Starting lvm2-activation-early.service... Apr 12 18:20:16.438684 lvm[1024]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 12 18:20:16.443692 systemd-networkd[1000]: lo: Link UP Apr 12 18:20:16.443927 systemd-networkd[1000]: lo: Gained carrier Apr 12 18:20:16.444339 systemd-networkd[1000]: Enumeration completed Apr 12 18:20:16.444519 systemd[1]: Started systemd-networkd.service. Apr 12 18:20:16.444619 systemd-networkd[1000]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 12 18:20:16.445000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.449336 systemd-networkd[1000]: eth0: Link UP Apr 12 18:20:16.449424 systemd-networkd[1000]: eth0: Gained carrier Apr 12 18:20:16.474483 systemd[1]: Finished lvm2-activation-early.service. Apr 12 18:20:16.474000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.475555 systemd[1]: Reached target cryptsetup.target. Apr 12 18:20:16.477605 systemd[1]: Starting lvm2-activation.service... Apr 12 18:20:16.481295 lvm[1026]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 12 18:20:16.482696 systemd-networkd[1000]: eth0: DHCPv4 address 10.0.0.43/16, gateway 10.0.0.1 acquired from 10.0.0.1 Apr 12 18:20:16.515333 systemd[1]: Finished lvm2-activation.service. Apr 12 18:20:16.515000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.516278 systemd[1]: Reached target local-fs-pre.target. Apr 12 18:20:16.517130 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 12 18:20:16.517164 systemd[1]: Reached target local-fs.target. Apr 12 18:20:16.517931 systemd[1]: Reached target machines.target. Apr 12 18:20:16.519863 systemd[1]: Starting ldconfig.service... Apr 12 18:20:16.520878 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Apr 12 18:20:16.520942 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Apr 12 18:20:16.522045 systemd[1]: Starting systemd-boot-update.service... Apr 12 18:20:16.523981 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Apr 12 18:20:16.526134 systemd[1]: Starting systemd-machine-id-commit.service... Apr 12 18:20:16.527297 systemd[1]: systemd-sysext.service was skipped because no trigger condition checks were met. Apr 12 18:20:16.527357 systemd[1]: ensure-sysext.service was skipped because no trigger condition checks were met. Apr 12 18:20:16.528314 systemd[1]: Starting systemd-tmpfiles-setup.service... Apr 12 18:20:16.529376 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1028 (bootctl) Apr 12 18:20:16.530785 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Apr 12 18:20:16.537276 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Apr 12 18:20:16.540000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.653521 systemd-tmpfiles[1031]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Apr 12 18:20:16.658250 systemd-tmpfiles[1031]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 12 18:20:16.660200 systemd[1]: Finished systemd-machine-id-commit.service. Apr 12 18:20:16.660000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.663999 systemd-tmpfiles[1031]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 12 18:20:16.673090 systemd-fsck[1037]: fsck.fat 4.2 (2021-01-31) Apr 12 18:20:16.673090 systemd-fsck[1037]: /dev/vda1: 236 files, 117047/258078 clusters Apr 12 18:20:16.675684 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Apr 12 18:20:16.676000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.747443 ldconfig[1027]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 12 18:20:16.751000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.750902 systemd[1]: Finished ldconfig.service. Apr 12 18:20:16.877341 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 12 18:20:16.878786 systemd[1]: Mounting boot.mount... Apr 12 18:20:16.885566 systemd[1]: Mounted boot.mount. Apr 12 18:20:16.892468 systemd[1]: Finished systemd-boot-update.service. Apr 12 18:20:16.893000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.945030 systemd[1]: Finished systemd-tmpfiles-setup.service. Apr 12 18:20:16.945000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.947171 systemd[1]: Starting audit-rules.service... Apr 12 18:20:16.948928 systemd[1]: Starting clean-ca-certificates.service... Apr 12 18:20:16.950902 systemd[1]: Starting systemd-journal-catalog-update.service... Apr 12 18:20:16.954000 audit: BPF prog-id=24 op=LOAD Apr 12 18:20:16.955671 systemd[1]: Starting systemd-resolved.service... Apr 12 18:20:16.956000 audit: BPF prog-id=25 op=LOAD Apr 12 18:20:16.962000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.959660 systemd[1]: Starting systemd-timesyncd.service... Apr 12 18:20:16.961335 systemd[1]: Starting systemd-update-utmp.service... Apr 12 18:20:16.962655 systemd[1]: Finished clean-ca-certificates.service. Apr 12 18:20:16.963772 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 12 18:20:16.965000 audit[1052]: SYSTEM_BOOT pid=1052 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.968687 systemd[1]: Finished systemd-update-utmp.service. Apr 12 18:20:16.968000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.976557 systemd[1]: Finished systemd-journal-catalog-update.service. Apr 12 18:20:16.976000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.978784 systemd[1]: Starting systemd-update-done.service... Apr 12 18:20:16.987628 systemd[1]: Finished systemd-update-done.service. Apr 12 18:20:16.987000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-done comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:20:16.996053 augenrules[1063]: No rules Apr 12 18:20:16.995000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Apr 12 18:20:16.995000 audit[1063]: SYSCALL arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffd1f38f20 a2=420 a3=0 items=0 ppid=1041 pid=1063 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:20:16.995000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Apr 12 18:20:16.997043 systemd[1]: Finished audit-rules.service. Apr 12 18:20:17.005141 systemd-resolved[1045]: Positive Trust Anchors: Apr 12 18:20:17.005151 systemd-resolved[1045]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 12 18:20:17.005177 systemd-resolved[1045]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Apr 12 18:20:17.017667 systemd-resolved[1045]: Defaulting to hostname 'linux'. Apr 12 18:20:17.018098 systemd[1]: Started systemd-timesyncd.service. Apr 12 18:20:17.019227 systemd-timesyncd[1051]: Contacted time server 10.0.0.1:123 (10.0.0.1). Apr 12 18:20:17.019278 systemd-timesyncd[1051]: Initial clock synchronization to Fri 2024-04-12 18:20:16.954231 UTC. Apr 12 18:20:17.019304 systemd[1]: Started systemd-resolved.service. Apr 12 18:20:17.020166 systemd[1]: Reached target network.target. Apr 12 18:20:17.020903 systemd[1]: Reached target nss-lookup.target. Apr 12 18:20:17.021641 systemd[1]: Reached target sysinit.target. Apr 12 18:20:17.022429 systemd[1]: Started motdgen.path. Apr 12 18:20:17.023283 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Apr 12 18:20:17.024331 systemd[1]: Started systemd-tmpfiles-clean.timer. Apr 12 18:20:17.025148 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 12 18:20:17.025180 systemd[1]: Reached target paths.target. Apr 12 18:20:17.025895 systemd[1]: Reached target time-set.target. Apr 12 18:20:17.026792 systemd[1]: Started logrotate.timer. Apr 12 18:20:17.027547 systemd[1]: Started mdadm.timer. Apr 12 18:20:17.028193 systemd[1]: Reached target timers.target. Apr 12 18:20:17.029258 systemd[1]: Listening on dbus.socket. Apr 12 18:20:17.030982 systemd[1]: Starting docker.socket... Apr 12 18:20:17.033918 systemd[1]: Listening on sshd.socket. Apr 12 18:20:17.034718 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Apr 12 18:20:17.035125 systemd[1]: Listening on docker.socket. Apr 12 18:20:17.035935 systemd[1]: Reached target sockets.target. Apr 12 18:20:17.036667 systemd[1]: Reached target basic.target. Apr 12 18:20:17.037387 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Apr 12 18:20:17.037419 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Apr 12 18:20:17.038379 systemd[1]: Starting containerd.service... Apr 12 18:20:17.039996 systemd[1]: Starting dbus.service... Apr 12 18:20:17.041724 systemd[1]: Starting enable-oem-cloudinit.service... Apr 12 18:20:17.043672 systemd[1]: Starting extend-filesystems.service... Apr 12 18:20:17.044539 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Apr 12 18:20:17.045769 systemd[1]: Starting motdgen.service... Apr 12 18:20:17.048268 systemd[1]: Starting prepare-cni-plugins.service... Apr 12 18:20:17.051393 jq[1073]: false Apr 12 18:20:17.050155 systemd[1]: Starting prepare-critools.service... Apr 12 18:20:17.052553 systemd[1]: Starting prepare-helm.service... Apr 12 18:20:17.054143 systemd[1]: Starting ssh-key-proc-cmdline.service... Apr 12 18:20:17.056011 systemd[1]: Starting sshd-keygen.service... Apr 12 18:20:17.058592 systemd[1]: Starting systemd-logind.service... Apr 12 18:20:17.059290 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Apr 12 18:20:17.059353 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 12 18:20:17.059878 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Apr 12 18:20:17.060475 systemd[1]: Starting update-engine.service... Apr 12 18:20:17.062424 systemd[1]: Starting update-ssh-keys-after-ignition.service... Apr 12 18:20:17.064814 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 12 18:20:17.065803 jq[1093]: true Apr 12 18:20:17.064972 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Apr 12 18:20:17.073554 extend-filesystems[1074]: Found vda Apr 12 18:20:17.073554 extend-filesystems[1074]: Found vda1 Apr 12 18:20:17.073554 extend-filesystems[1074]: Found vda2 Apr 12 18:20:17.073554 extend-filesystems[1074]: Found vda3 Apr 12 18:20:17.073554 extend-filesystems[1074]: Found usr Apr 12 18:20:17.073554 extend-filesystems[1074]: Found vda4 Apr 12 18:20:17.073554 extend-filesystems[1074]: Found vda6 Apr 12 18:20:17.073554 extend-filesystems[1074]: Found vda7 Apr 12 18:20:17.073554 extend-filesystems[1074]: Found vda9 Apr 12 18:20:17.073554 extend-filesystems[1074]: Checking size of /dev/vda9 Apr 12 18:20:17.067545 systemd[1]: motdgen.service: Deactivated successfully. Apr 12 18:20:17.067715 systemd[1]: Finished motdgen.service. Apr 12 18:20:17.069792 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 12 18:20:17.069961 systemd[1]: Finished ssh-key-proc-cmdline.service. Apr 12 18:20:17.092528 jq[1100]: true Apr 12 18:20:17.092652 tar[1095]: ./ Apr 12 18:20:17.092652 tar[1095]: ./loopback Apr 12 18:20:17.093046 tar[1096]: crictl Apr 12 18:20:17.108645 dbus-daemon[1072]: [system] SELinux support is enabled Apr 12 18:20:17.108800 systemd[1]: Started dbus.service. Apr 12 18:20:17.111394 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 12 18:20:17.111421 systemd[1]: Reached target system-config.target. Apr 12 18:20:17.111684 tar[1099]: linux-arm64/helm Apr 12 18:20:17.112282 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 12 18:20:17.112302 systemd[1]: Reached target user-config.target. Apr 12 18:20:17.139229 extend-filesystems[1074]: Resized partition /dev/vda9 Apr 12 18:20:17.152403 extend-filesystems[1128]: resize2fs 1.46.5 (30-Dec-2021) Apr 12 18:20:17.161155 tar[1095]: ./bandwidth Apr 12 18:20:17.157686 systemd-logind[1089]: Watching system buttons on /dev/input/event0 (Power Button) Apr 12 18:20:17.158628 systemd-logind[1089]: New seat seat0. Apr 12 18:20:17.160054 systemd[1]: Started systemd-logind.service. Apr 12 18:20:17.168597 bash[1129]: Updated "/home/core/.ssh/authorized_keys" Apr 12 18:20:17.169260 systemd[1]: Finished update-ssh-keys-after-ignition.service. Apr 12 18:20:17.201585 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Apr 12 18:20:17.216786 env[1102]: time="2024-04-12T18:20:17.216730000Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Apr 12 18:20:17.218592 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Apr 12 18:20:17.230641 extend-filesystems[1128]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Apr 12 18:20:17.230641 extend-filesystems[1128]: old_desc_blocks = 1, new_desc_blocks = 1 Apr 12 18:20:17.230641 extend-filesystems[1128]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Apr 12 18:20:17.234682 extend-filesystems[1074]: Resized filesystem in /dev/vda9 Apr 12 18:20:17.231358 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 12 18:20:17.231511 systemd[1]: Finished extend-filesystems.service. Apr 12 18:20:17.255644 env[1102]: time="2024-04-12T18:20:17.255596840Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 12 18:20:17.255787 env[1102]: time="2024-04-12T18:20:17.255759080Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 12 18:20:17.257377 env[1102]: time="2024-04-12T18:20:17.257338000Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.154-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 12 18:20:17.257377 env[1102]: time="2024-04-12T18:20:17.257371200Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 12 18:20:17.257632 env[1102]: time="2024-04-12T18:20:17.257606920Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 12 18:20:17.257664 env[1102]: time="2024-04-12T18:20:17.257631560Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 12 18:20:17.257664 env[1102]: time="2024-04-12T18:20:17.257646040Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Apr 12 18:20:17.257664 env[1102]: time="2024-04-12T18:20:17.257655680Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 12 18:20:17.257769 env[1102]: time="2024-04-12T18:20:17.257741960Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 12 18:20:17.257983 env[1102]: time="2024-04-12T18:20:17.257961760Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 12 18:20:17.258120 env[1102]: time="2024-04-12T18:20:17.258098000Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 12 18:20:17.258152 env[1102]: time="2024-04-12T18:20:17.258119240Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 12 18:20:17.258194 env[1102]: time="2024-04-12T18:20:17.258176040Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Apr 12 18:20:17.258235 env[1102]: time="2024-04-12T18:20:17.258193520Z" level=info msg="metadata content store policy set" policy=shared Apr 12 18:20:17.259722 tar[1095]: ./ptp Apr 12 18:20:17.266447 update_engine[1092]: I0412 18:20:17.258913 1092 main.cc:92] Flatcar Update Engine starting Apr 12 18:20:17.282714 systemd[1]: Started update-engine.service. Apr 12 18:20:17.283755 update_engine[1092]: I0412 18:20:17.283723 1092 update_check_scheduler.cc:74] Next update check in 8m0s Apr 12 18:20:17.285353 systemd[1]: Started locksmithd.service. Apr 12 18:20:17.285506 env[1102]: time="2024-04-12T18:20:17.285468840Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 12 18:20:17.285547 env[1102]: time="2024-04-12T18:20:17.285514200Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 12 18:20:17.285547 env[1102]: time="2024-04-12T18:20:17.285529400Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 12 18:20:17.285603 env[1102]: time="2024-04-12T18:20:17.285579440Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 12 18:20:17.285626 env[1102]: time="2024-04-12T18:20:17.285600920Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 12 18:20:17.285626 env[1102]: time="2024-04-12T18:20:17.285616160Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 12 18:20:17.285673 env[1102]: time="2024-04-12T18:20:17.285630200Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 12 18:20:17.286018 env[1102]: time="2024-04-12T18:20:17.285995040Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 12 18:20:17.286049 env[1102]: time="2024-04-12T18:20:17.286023960Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Apr 12 18:20:17.286049 env[1102]: time="2024-04-12T18:20:17.286039880Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 12 18:20:17.286088 env[1102]: time="2024-04-12T18:20:17.286055000Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 12 18:20:17.286088 env[1102]: time="2024-04-12T18:20:17.286068880Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 12 18:20:17.286235 env[1102]: time="2024-04-12T18:20:17.286214840Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 12 18:20:17.286318 env[1102]: time="2024-04-12T18:20:17.286300600Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 12 18:20:17.286547 env[1102]: time="2024-04-12T18:20:17.286529560Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 12 18:20:17.286585 env[1102]: time="2024-04-12T18:20:17.286557960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.286611 env[1102]: time="2024-04-12T18:20:17.286585480Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 12 18:20:17.286834 env[1102]: time="2024-04-12T18:20:17.286818200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.286863 env[1102]: time="2024-04-12T18:20:17.286836320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.286863 env[1102]: time="2024-04-12T18:20:17.286849320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.286863 env[1102]: time="2024-04-12T18:20:17.286860800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.286926 env[1102]: time="2024-04-12T18:20:17.286873320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.286926 env[1102]: time="2024-04-12T18:20:17.286886560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.286926 env[1102]: time="2024-04-12T18:20:17.286897960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.286926 env[1102]: time="2024-04-12T18:20:17.286909080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.286926 env[1102]: time="2024-04-12T18:20:17.286922520Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 12 18:20:17.287055 env[1102]: time="2024-04-12T18:20:17.287037320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.287157 env[1102]: time="2024-04-12T18:20:17.287053600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.287157 env[1102]: time="2024-04-12T18:20:17.287066320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.287157 env[1102]: time="2024-04-12T18:20:17.287079200Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 12 18:20:17.287157 env[1102]: time="2024-04-12T18:20:17.287095000Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Apr 12 18:20:17.287157 env[1102]: time="2024-04-12T18:20:17.287105960Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 12 18:20:17.287157 env[1102]: time="2024-04-12T18:20:17.287123440Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Apr 12 18:20:17.287157 env[1102]: time="2024-04-12T18:20:17.287156560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 12 18:20:17.287634 env[1102]: time="2024-04-12T18:20:17.287562440Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.287642000Z" level=info msg="Connect containerd service" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.287677280Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.288396520Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.289212360Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.289254880Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.289301640Z" level=info msg="containerd successfully booted in 0.073310s" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.289808320Z" level=info msg="Start subscribing containerd event" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.289864880Z" level=info msg="Start recovering state" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.289928520Z" level=info msg="Start event monitor" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.289948800Z" level=info msg="Start snapshots syncer" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.289958720Z" level=info msg="Start cni network conf syncer for default" Apr 12 18:20:17.292609 env[1102]: time="2024-04-12T18:20:17.289966080Z" level=info msg="Start streaming server" Apr 12 18:20:17.289375 systemd[1]: Started containerd.service. Apr 12 18:20:17.317836 tar[1095]: ./vlan Apr 12 18:20:17.363590 tar[1095]: ./host-device Apr 12 18:20:17.410242 tar[1095]: ./tuning Apr 12 18:20:17.453091 tar[1095]: ./vrf Apr 12 18:20:17.493058 tar[1095]: ./sbr Apr 12 18:20:17.533171 tar[1095]: ./tap Apr 12 18:20:17.543892 tar[1099]: linux-arm64/LICENSE Apr 12 18:20:17.543892 tar[1099]: linux-arm64/README.md Apr 12 18:20:17.548821 systemd[1]: Finished prepare-helm.service. Apr 12 18:20:17.572233 tar[1095]: ./dhcp Apr 12 18:20:17.591068 locksmithd[1136]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 12 18:20:17.595842 systemd[1]: Finished prepare-critools.service. Apr 12 18:20:17.654755 tar[1095]: ./static Apr 12 18:20:17.675631 tar[1095]: ./firewall Apr 12 18:20:17.707169 tar[1095]: ./macvlan Apr 12 18:20:17.735829 tar[1095]: ./dummy Apr 12 18:20:17.764044 tar[1095]: ./bridge Apr 12 18:20:17.795138 tar[1095]: ./ipvlan Apr 12 18:20:17.823357 tar[1095]: ./portmap Apr 12 18:20:17.850209 tar[1095]: ./host-local Apr 12 18:20:17.884845 systemd[1]: Finished prepare-cni-plugins.service. Apr 12 18:20:18.011788 sshd_keygen[1098]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 12 18:20:18.030081 systemd[1]: Finished sshd-keygen.service. Apr 12 18:20:18.032119 systemd[1]: Starting issuegen.service... Apr 12 18:20:18.036079 systemd[1]: issuegen.service: Deactivated successfully. Apr 12 18:20:18.036214 systemd[1]: Finished issuegen.service. Apr 12 18:20:18.038083 systemd[1]: Starting systemd-user-sessions.service... Apr 12 18:20:18.043434 systemd[1]: Finished systemd-user-sessions.service. Apr 12 18:20:18.045439 systemd[1]: Started getty@tty1.service. Apr 12 18:20:18.047331 systemd[1]: Started serial-getty@ttyAMA0.service. Apr 12 18:20:18.048383 systemd[1]: Reached target getty.target. Apr 12 18:20:18.049159 systemd[1]: Reached target multi-user.target. Apr 12 18:20:18.050957 systemd[1]: Starting systemd-update-utmp-runlevel.service... Apr 12 18:20:18.056974 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Apr 12 18:20:18.057112 systemd[1]: Finished systemd-update-utmp-runlevel.service. Apr 12 18:20:18.058068 systemd[1]: Startup finished in 561ms (kernel) + 6.263s (initrd) + 4.224s (userspace) = 11.049s. Apr 12 18:20:18.111759 systemd-networkd[1000]: eth0: Gained IPv6LL Apr 12 18:20:19.891207 systemd[1]: Created slice system-sshd.slice. Apr 12 18:20:19.892249 systemd[1]: Started sshd@0-10.0.0.43:22-10.0.0.1:34318.service. Apr 12 18:20:19.942117 sshd[1161]: Accepted publickey for core from 10.0.0.1 port 34318 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:20:19.944037 sshd[1161]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:20:19.951830 systemd[1]: Created slice user-500.slice. Apr 12 18:20:19.952875 systemd[1]: Starting user-runtime-dir@500.service... Apr 12 18:20:19.954615 systemd-logind[1089]: New session 1 of user core. Apr 12 18:20:19.960335 systemd[1]: Finished user-runtime-dir@500.service. Apr 12 18:20:19.961676 systemd[1]: Starting user@500.service... Apr 12 18:20:19.964001 (systemd)[1164]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:20:20.021554 systemd[1164]: Queued start job for default target default.target. Apr 12 18:20:20.022014 systemd[1164]: Reached target paths.target. Apr 12 18:20:20.022041 systemd[1164]: Reached target sockets.target. Apr 12 18:20:20.022051 systemd[1164]: Reached target timers.target. Apr 12 18:20:20.022062 systemd[1164]: Reached target basic.target. Apr 12 18:20:20.022111 systemd[1164]: Reached target default.target. Apr 12 18:20:20.022134 systemd[1164]: Startup finished in 52ms. Apr 12 18:20:20.022305 systemd[1]: Started user@500.service. Apr 12 18:20:20.023163 systemd[1]: Started session-1.scope. Apr 12 18:20:20.074344 systemd[1]: Started sshd@1-10.0.0.43:22-10.0.0.1:34324.service. Apr 12 18:20:20.117437 sshd[1173]: Accepted publickey for core from 10.0.0.1 port 34324 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:20:20.118487 sshd[1173]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:20:20.121620 systemd-logind[1089]: New session 2 of user core. Apr 12 18:20:20.122390 systemd[1]: Started session-2.scope. Apr 12 18:20:20.178332 sshd[1173]: pam_unix(sshd:session): session closed for user core Apr 12 18:20:20.182170 systemd[1]: sshd@1-10.0.0.43:22-10.0.0.1:34324.service: Deactivated successfully. Apr 12 18:20:20.182771 systemd[1]: session-2.scope: Deactivated successfully. Apr 12 18:20:20.183745 systemd-logind[1089]: Session 2 logged out. Waiting for processes to exit. Apr 12 18:20:20.185113 systemd[1]: Started sshd@2-10.0.0.43:22-10.0.0.1:34328.service. Apr 12 18:20:20.186020 systemd-logind[1089]: Removed session 2. Apr 12 18:20:20.228629 sshd[1179]: Accepted publickey for core from 10.0.0.1 port 34328 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:20:20.229719 sshd[1179]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:20:20.232748 systemd-logind[1089]: New session 3 of user core. Apr 12 18:20:20.233488 systemd[1]: Started session-3.scope. Apr 12 18:20:20.283448 sshd[1179]: pam_unix(sshd:session): session closed for user core Apr 12 18:20:20.285959 systemd[1]: sshd@2-10.0.0.43:22-10.0.0.1:34328.service: Deactivated successfully. Apr 12 18:20:20.286512 systemd[1]: session-3.scope: Deactivated successfully. Apr 12 18:20:20.287010 systemd-logind[1089]: Session 3 logged out. Waiting for processes to exit. Apr 12 18:20:20.287977 systemd[1]: Started sshd@3-10.0.0.43:22-10.0.0.1:34342.service. Apr 12 18:20:20.288667 systemd-logind[1089]: Removed session 3. Apr 12 18:20:20.331248 sshd[1185]: Accepted publickey for core from 10.0.0.1 port 34342 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:20:20.332298 sshd[1185]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:20:20.336242 systemd[1]: Started session-4.scope. Apr 12 18:20:20.337430 systemd-logind[1089]: New session 4 of user core. Apr 12 18:20:20.390441 sshd[1185]: pam_unix(sshd:session): session closed for user core Apr 12 18:20:20.394231 systemd[1]: sshd@3-10.0.0.43:22-10.0.0.1:34342.service: Deactivated successfully. Apr 12 18:20:20.394772 systemd[1]: session-4.scope: Deactivated successfully. Apr 12 18:20:20.395308 systemd-logind[1089]: Session 4 logged out. Waiting for processes to exit. Apr 12 18:20:20.396302 systemd[1]: Started sshd@4-10.0.0.43:22-10.0.0.1:34354.service. Apr 12 18:20:20.396923 systemd-logind[1089]: Removed session 4. Apr 12 18:20:20.440243 sshd[1191]: Accepted publickey for core from 10.0.0.1 port 34354 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:20:20.441384 sshd[1191]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:20:20.444592 systemd-logind[1089]: New session 5 of user core. Apr 12 18:20:20.445337 systemd[1]: Started session-5.scope. Apr 12 18:20:20.504913 sudo[1194]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 12 18:20:20.505115 sudo[1194]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Apr 12 18:20:21.247286 systemd[1]: Starting systemd-networkd-wait-online.service... Apr 12 18:20:21.377165 systemd[1]: Finished systemd-networkd-wait-online.service. Apr 12 18:20:21.377469 systemd[1]: Reached target network-online.target. Apr 12 18:20:21.378869 systemd[1]: Starting docker.service... Apr 12 18:20:21.455650 env[1211]: time="2024-04-12T18:20:21.455589859Z" level=info msg="Starting up" Apr 12 18:20:21.457135 env[1211]: time="2024-04-12T18:20:21.457106250Z" level=info msg="parsed scheme: \"unix\"" module=grpc Apr 12 18:20:21.457230 env[1211]: time="2024-04-12T18:20:21.457216289Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Apr 12 18:20:21.457299 env[1211]: time="2024-04-12T18:20:21.457285271Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Apr 12 18:20:21.457349 env[1211]: time="2024-04-12T18:20:21.457337347Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Apr 12 18:20:21.459535 env[1211]: time="2024-04-12T18:20:21.459500164Z" level=info msg="parsed scheme: \"unix\"" module=grpc Apr 12 18:20:21.459535 env[1211]: time="2024-04-12T18:20:21.459525028Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Apr 12 18:20:21.459649 env[1211]: time="2024-04-12T18:20:21.459539668Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Apr 12 18:20:21.459649 env[1211]: time="2024-04-12T18:20:21.459548937Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Apr 12 18:20:21.794836 env[1211]: time="2024-04-12T18:20:21.794797551Z" level=info msg="Loading containers: start." Apr 12 18:20:21.901587 kernel: Initializing XFRM netlink socket Apr 12 18:20:21.925078 env[1211]: time="2024-04-12T18:20:21.925029418Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Apr 12 18:20:21.976058 systemd-networkd[1000]: docker0: Link UP Apr 12 18:20:21.985102 env[1211]: time="2024-04-12T18:20:21.985063571Z" level=info msg="Loading containers: done." Apr 12 18:20:22.007744 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck898775287-merged.mount: Deactivated successfully. Apr 12 18:20:22.012597 env[1211]: time="2024-04-12T18:20:22.012539904Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 12 18:20:22.012915 env[1211]: time="2024-04-12T18:20:22.012894808Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Apr 12 18:20:22.013082 env[1211]: time="2024-04-12T18:20:22.013065751Z" level=info msg="Daemon has completed initialization" Apr 12 18:20:22.025166 systemd[1]: Started docker.service. Apr 12 18:20:22.031409 env[1211]: time="2024-04-12T18:20:22.031303140Z" level=info msg="API listen on /run/docker.sock" Apr 12 18:20:22.046832 systemd[1]: Reloading. Apr 12 18:20:22.087978 /usr/lib/systemd/system-generators/torcx-generator[1355]: time="2024-04-12T18:20:22Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.3 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.3 /var/lib/torcx/store]" Apr 12 18:20:22.088010 /usr/lib/systemd/system-generators/torcx-generator[1355]: time="2024-04-12T18:20:22Z" level=info msg="torcx already run" Apr 12 18:20:22.143331 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Apr 12 18:20:22.143351 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Apr 12 18:20:22.158280 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 12 18:20:22.218970 systemd[1]: Started kubelet.service. Apr 12 18:20:22.342489 kubelet[1391]: E0412 18:20:22.342360 1391 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 12 18:20:22.344340 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 12 18:20:22.344469 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 12 18:20:22.604623 env[1102]: time="2024-04-12T18:20:22.604327006Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.28.8\"" Apr 12 18:20:23.103063 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount293902962.mount: Deactivated successfully. Apr 12 18:20:25.150255 env[1102]: time="2024-04-12T18:20:25.150180396Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.28.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:25.151413 env[1102]: time="2024-04-12T18:20:25.151378087Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:883d43b86efe04c7ca7bcd566f873179fa9c1dbceb67e32cd5d30213c3bc17de,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:25.152915 env[1102]: time="2024-04-12T18:20:25.152887661Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.28.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:25.154475 env[1102]: time="2024-04-12T18:20:25.154447634Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:7e7f3c806333528451a1e0bfdf17da0341adaea7d50a703db9c2005c474a97b9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:25.155259 env[1102]: time="2024-04-12T18:20:25.155228497Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.28.8\" returns image reference \"sha256:883d43b86efe04c7ca7bcd566f873179fa9c1dbceb67e32cd5d30213c3bc17de\"" Apr 12 18:20:25.163902 env[1102]: time="2024-04-12T18:20:25.163859823Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.28.8\"" Apr 12 18:20:27.931774 env[1102]: time="2024-04-12T18:20:27.931721330Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.28.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:27.933141 env[1102]: time="2024-04-12T18:20:27.933118114Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7beedd93d8e53aab4b98613d37758450bbbac01a94f42cdb7670da900d1e11d8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:27.934816 env[1102]: time="2024-04-12T18:20:27.934788548Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.28.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:27.936750 env[1102]: time="2024-04-12T18:20:27.936728124Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:f3d0e8da9d1532e081e719a985e89a0cfe1a29d127773ad8e2c2fee1dd10fd00,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:27.937431 env[1102]: time="2024-04-12T18:20:27.937403791Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.28.8\" returns image reference \"sha256:7beedd93d8e53aab4b98613d37758450bbbac01a94f42cdb7670da900d1e11d8\"" Apr 12 18:20:27.946432 env[1102]: time="2024-04-12T18:20:27.946405294Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.28.8\"" Apr 12 18:20:29.766962 env[1102]: time="2024-04-12T18:20:29.766912773Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.28.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:29.768467 env[1102]: time="2024-04-12T18:20:29.768440833Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:36dcd04414a4b2b645ad6da4fd60a5d1479f6eb9da01a928082abb025958a687,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:29.770097 env[1102]: time="2024-04-12T18:20:29.770071781Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.28.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:29.772248 env[1102]: time="2024-04-12T18:20:29.772200876Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:4d61604f259d3c91d8b3ec7a6a999f5eae9aff371567151cd5165eaa698c6d7b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:29.774237 env[1102]: time="2024-04-12T18:20:29.774206522Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.28.8\" returns image reference \"sha256:36dcd04414a4b2b645ad6da4fd60a5d1479f6eb9da01a928082abb025958a687\"" Apr 12 18:20:29.786509 env[1102]: time="2024-04-12T18:20:29.786478993Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.8\"" Apr 12 18:20:30.805315 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1766532985.mount: Deactivated successfully. Apr 12 18:20:31.339908 env[1102]: time="2024-04-12T18:20:31.339854862Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.28.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:31.341284 env[1102]: time="2024-04-12T18:20:31.341254498Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:837f825eec6c170d5e5bbfbd7bb0a4afac97759d0f0c57b80e4712d417fd690b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:31.342702 env[1102]: time="2024-04-12T18:20:31.342673865Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.28.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:31.344308 env[1102]: time="2024-04-12T18:20:31.344277089Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:9e9dd46799712c58e1a49f973374ffa9ad4e5a6175896e5d805a8738bf5c5865,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:31.344873 env[1102]: time="2024-04-12T18:20:31.344841001Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.8\" returns image reference \"sha256:837f825eec6c170d5e5bbfbd7bb0a4afac97759d0f0c57b80e4712d417fd690b\"" Apr 12 18:20:31.353671 env[1102]: time="2024-04-12T18:20:31.353643154Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Apr 12 18:20:31.779461 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4241322182.mount: Deactivated successfully. Apr 12 18:20:31.783520 env[1102]: time="2024-04-12T18:20:31.783480221Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:31.785126 env[1102]: time="2024-04-12T18:20:31.785088038Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:31.786834 env[1102]: time="2024-04-12T18:20:31.786795113Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:31.788949 env[1102]: time="2024-04-12T18:20:31.788914278Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:31.789607 env[1102]: time="2024-04-12T18:20:31.789581162Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" Apr 12 18:20:31.799032 env[1102]: time="2024-04-12T18:20:31.798999712Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.9-0\"" Apr 12 18:20:32.322803 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2135786265.mount: Deactivated successfully. Apr 12 18:20:32.595209 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 12 18:20:32.595388 systemd[1]: Stopped kubelet.service. Apr 12 18:20:32.596936 systemd[1]: Started kubelet.service. Apr 12 18:20:32.640642 kubelet[1445]: E0412 18:20:32.640600 1445 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Apr 12 18:20:32.646243 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 12 18:20:32.646375 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 12 18:20:36.129337 env[1102]: time="2024-04-12T18:20:36.129288748Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.9-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:36.132148 env[1102]: time="2024-04-12T18:20:36.131699698Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:9cdd6470f48c8b127530b7ce6ea4b3524137984481e48bcde619735890840ace,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:36.133860 env[1102]: time="2024-04-12T18:20:36.133818623Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.9-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:36.136094 env[1102]: time="2024-04-12T18:20:36.136059737Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:e013d0d5e4e25d00c61a7ff839927a1f36479678f11e49502b53a5e0b14f10c3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:36.136987 env[1102]: time="2024-04-12T18:20:36.136952362Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.9-0\" returns image reference \"sha256:9cdd6470f48c8b127530b7ce6ea4b3524137984481e48bcde619735890840ace\"" Apr 12 18:20:36.147361 env[1102]: time="2024-04-12T18:20:36.147322709Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\"" Apr 12 18:20:36.780814 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3072349332.mount: Deactivated successfully. Apr 12 18:20:37.534087 env[1102]: time="2024-04-12T18:20:37.534016895Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.10.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:37.536782 env[1102]: time="2024-04-12T18:20:37.536747781Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:97e04611ad43405a2e5863ae17c6f1bc9181bdefdaa78627c432ef754a4eb108,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:37.538044 env[1102]: time="2024-04-12T18:20:37.538013048Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.10.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:37.539725 env[1102]: time="2024-04-12T18:20:37.539638484Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:37.540199 env[1102]: time="2024-04-12T18:20:37.540148756Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\" returns image reference \"sha256:97e04611ad43405a2e5863ae17c6f1bc9181bdefdaa78627c432ef754a4eb108\"" Apr 12 18:20:41.731200 systemd[1]: Stopped kubelet.service. Apr 12 18:20:41.747233 systemd[1]: Reloading. Apr 12 18:20:41.791678 /usr/lib/systemd/system-generators/torcx-generator[1556]: time="2024-04-12T18:20:41Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.3 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.3 /var/lib/torcx/store]" Apr 12 18:20:41.791706 /usr/lib/systemd/system-generators/torcx-generator[1556]: time="2024-04-12T18:20:41Z" level=info msg="torcx already run" Apr 12 18:20:41.850657 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Apr 12 18:20:41.850677 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Apr 12 18:20:41.865885 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 12 18:20:41.935866 systemd[1]: Started kubelet.service. Apr 12 18:20:41.975653 kubelet[1593]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 12 18:20:41.975653 kubelet[1593]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Apr 12 18:20:41.975653 kubelet[1593]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 12 18:20:41.976021 kubelet[1593]: I0412 18:20:41.975693 1593 server.go:203] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 12 18:20:42.707609 kubelet[1593]: I0412 18:20:42.707556 1593 server.go:467] "Kubelet version" kubeletVersion="v1.28.1" Apr 12 18:20:42.707609 kubelet[1593]: I0412 18:20:42.707606 1593 server.go:469] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 12 18:20:42.707877 kubelet[1593]: I0412 18:20:42.707860 1593 server.go:895] "Client rotation is on, will bootstrap in background" Apr 12 18:20:42.712637 kubelet[1593]: I0412 18:20:42.712606 1593 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 12 18:20:42.713018 kubelet[1593]: E0412 18:20:42.712961 1593 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.43:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:42.718542 kubelet[1593]: W0412 18:20:42.718518 1593 machine.go:65] Cannot read vendor id correctly, set empty. Apr 12 18:20:42.719197 kubelet[1593]: I0412 18:20:42.719171 1593 server.go:725] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 12 18:20:42.719380 kubelet[1593]: I0412 18:20:42.719364 1593 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 12 18:20:42.719525 kubelet[1593]: I0412 18:20:42.719504 1593 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Apr 12 18:20:42.719525 kubelet[1593]: I0412 18:20:42.719525 1593 topology_manager.go:138] "Creating topology manager with none policy" Apr 12 18:20:42.719658 kubelet[1593]: I0412 18:20:42.719533 1593 container_manager_linux.go:301] "Creating device plugin manager" Apr 12 18:20:42.719658 kubelet[1593]: I0412 18:20:42.719633 1593 state_mem.go:36] "Initialized new in-memory state store" Apr 12 18:20:42.719858 kubelet[1593]: I0412 18:20:42.719831 1593 kubelet.go:393] "Attempting to sync node with API server" Apr 12 18:20:42.719858 kubelet[1593]: I0412 18:20:42.719856 1593 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 12 18:20:42.719913 kubelet[1593]: I0412 18:20:42.719878 1593 kubelet.go:309] "Adding apiserver pod source" Apr 12 18:20:42.719913 kubelet[1593]: I0412 18:20:42.719892 1593 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 12 18:20:42.720295 kubelet[1593]: W0412 18:20:42.720245 1593 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://10.0.0.43:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:42.720328 kubelet[1593]: E0412 18:20:42.720304 1593 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.43:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:42.720642 kubelet[1593]: I0412 18:20:42.720622 1593 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Apr 12 18:20:42.720765 kubelet[1593]: W0412 18:20:42.720736 1593 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://10.0.0.43:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:42.720862 kubelet[1593]: E0412 18:20:42.720850 1593 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.43:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:42.721020 kubelet[1593]: W0412 18:20:42.721003 1593 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 12 18:20:42.721711 kubelet[1593]: I0412 18:20:42.721696 1593 server.go:1232] "Started kubelet" Apr 12 18:20:42.722108 kubelet[1593]: I0412 18:20:42.722093 1593 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10 Apr 12 18:20:42.722409 kubelet[1593]: E0412 18:20:42.722377 1593 cri_stats_provider.go:448] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Apr 12 18:20:42.722409 kubelet[1593]: E0412 18:20:42.722407 1593 kubelet.go:1431] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 12 18:20:42.722519 kubelet[1593]: I0412 18:20:42.722503 1593 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 12 18:20:42.722597 kubelet[1593]: I0412 18:20:42.722105 1593 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Apr 12 18:20:42.722776 kubelet[1593]: E0412 18:20:42.722682 1593 event.go:289] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"localhost.17c59b59abd61b41", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"localhost", UID:"localhost", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"localhost"}, FirstTimestamp:time.Date(2024, time.April, 12, 18, 20, 42, 721672001, time.Local), LastTimestamp:time.Date(2024, time.April, 12, 18, 20, 42, 721672001, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"localhost"}': 'Post "https://10.0.0.43:6443/api/v1/namespaces/default/events": dial tcp 10.0.0.43:6443: connect: connection refused'(may retry after sleeping) Apr 12 18:20:42.723407 kubelet[1593]: I0412 18:20:42.723386 1593 server.go:462] "Adding debug handlers to kubelet server" Apr 12 18:20:42.723891 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Apr 12 18:20:42.724039 kubelet[1593]: I0412 18:20:42.724019 1593 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 12 18:20:42.724879 kubelet[1593]: E0412 18:20:42.724858 1593 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found" Apr 12 18:20:42.725151 kubelet[1593]: I0412 18:20:42.725130 1593 volume_manager.go:291] "Starting Kubelet Volume Manager" Apr 12 18:20:42.725274 kubelet[1593]: I0412 18:20:42.725257 1593 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Apr 12 18:20:42.725333 kubelet[1593]: I0412 18:20:42.725319 1593 reconciler_new.go:29] "Reconciler: start to sync state" Apr 12 18:20:42.725631 kubelet[1593]: W0412 18:20:42.725590 1593 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://10.0.0.43:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:42.725701 kubelet[1593]: E0412 18:20:42.725632 1593 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.43:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:42.725893 kubelet[1593]: E0412 18:20:42.725869 1593 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.43:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.43:6443: connect: connection refused" interval="200ms" Apr 12 18:20:42.742231 kubelet[1593]: I0412 18:20:42.742202 1593 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Apr 12 18:20:42.743728 kubelet[1593]: I0412 18:20:42.743624 1593 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Apr 12 18:20:42.743728 kubelet[1593]: I0412 18:20:42.743653 1593 status_manager.go:217] "Starting to sync pod status with apiserver" Apr 12 18:20:42.743728 kubelet[1593]: I0412 18:20:42.743668 1593 kubelet.go:2303] "Starting kubelet main sync loop" Apr 12 18:20:42.743878 kubelet[1593]: I0412 18:20:42.743760 1593 cpu_manager.go:214] "Starting CPU manager" policy="none" Apr 12 18:20:42.743878 kubelet[1593]: I0412 18:20:42.743772 1593 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Apr 12 18:20:42.743878 kubelet[1593]: I0412 18:20:42.743789 1593 state_mem.go:36] "Initialized new in-memory state store" Apr 12 18:20:42.744694 kubelet[1593]: E0412 18:20:42.744665 1593 kubelet.go:2327] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 12 18:20:42.744694 kubelet[1593]: W0412 18:20:42.744650 1593 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://10.0.0.43:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:42.744832 kubelet[1593]: E0412 18:20:42.744709 1593 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.43:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:42.751802 kubelet[1593]: I0412 18:20:42.751768 1593 policy_none.go:49] "None policy: Start" Apr 12 18:20:42.752425 kubelet[1593]: I0412 18:20:42.752404 1593 memory_manager.go:169] "Starting memorymanager" policy="None" Apr 12 18:20:42.752480 kubelet[1593]: I0412 18:20:42.752430 1593 state_mem.go:35] "Initializing new in-memory state store" Apr 12 18:20:42.757727 systemd[1]: Created slice kubepods.slice. Apr 12 18:20:42.762784 systemd[1]: Created slice kubepods-burstable.slice. Apr 12 18:20:42.765440 systemd[1]: Created slice kubepods-besteffort.slice. Apr 12 18:20:42.779211 kubelet[1593]: I0412 18:20:42.779174 1593 manager.go:471] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 12 18:20:42.779514 kubelet[1593]: I0412 18:20:42.779442 1593 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 12 18:20:42.780090 kubelet[1593]: E0412 18:20:42.780021 1593 eviction_manager.go:258] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Apr 12 18:20:42.826942 kubelet[1593]: I0412 18:20:42.826903 1593 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Apr 12 18:20:42.827333 kubelet[1593]: E0412 18:20:42.827301 1593 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.43:6443/api/v1/nodes\": dial tcp 10.0.0.43:6443: connect: connection refused" node="localhost" Apr 12 18:20:42.845472 kubelet[1593]: I0412 18:20:42.845429 1593 topology_manager.go:215] "Topology Admit Handler" podUID="d562bbd47681de03357399071ce5c29e" podNamespace="kube-system" podName="kube-apiserver-localhost" Apr 12 18:20:42.846448 kubelet[1593]: I0412 18:20:42.846423 1593 topology_manager.go:215] "Topology Admit Handler" podUID="7ea520b1e9b7ab03da4e0d60de4cd055" podNamespace="kube-system" podName="kube-controller-manager-localhost" Apr 12 18:20:42.847188 kubelet[1593]: I0412 18:20:42.847157 1593 topology_manager.go:215] "Topology Admit Handler" podUID="7bb06e2cba933e89936d947f3ee7de1d" podNamespace="kube-system" podName="kube-scheduler-localhost" Apr 12 18:20:42.851469 systemd[1]: Created slice kubepods-burstable-podd562bbd47681de03357399071ce5c29e.slice. Apr 12 18:20:42.862606 systemd[1]: Created slice kubepods-burstable-pod7ea520b1e9b7ab03da4e0d60de4cd055.slice. Apr 12 18:20:42.876989 systemd[1]: Created slice kubepods-burstable-pod7bb06e2cba933e89936d947f3ee7de1d.slice. Apr 12 18:20:42.926124 kubelet[1593]: I0412 18:20:42.926075 1593 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:42.926124 kubelet[1593]: I0412 18:20:42.926118 1593 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:42.926286 kubelet[1593]: I0412 18:20:42.926141 1593 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:42.926286 kubelet[1593]: I0412 18:20:42.926163 1593 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7bb06e2cba933e89936d947f3ee7de1d-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"7bb06e2cba933e89936d947f3ee7de1d\") " pod="kube-system/kube-scheduler-localhost" Apr 12 18:20:42.926286 kubelet[1593]: I0412 18:20:42.926183 1593 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d562bbd47681de03357399071ce5c29e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"d562bbd47681de03357399071ce5c29e\") " pod="kube-system/kube-apiserver-localhost" Apr 12 18:20:42.926286 kubelet[1593]: I0412 18:20:42.926203 1593 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d562bbd47681de03357399071ce5c29e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"d562bbd47681de03357399071ce5c29e\") " pod="kube-system/kube-apiserver-localhost" Apr 12 18:20:42.926286 kubelet[1593]: I0412 18:20:42.926223 1593 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d562bbd47681de03357399071ce5c29e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"d562bbd47681de03357399071ce5c29e\") " pod="kube-system/kube-apiserver-localhost" Apr 12 18:20:42.926405 kubelet[1593]: I0412 18:20:42.926272 1593 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:42.926405 kubelet[1593]: E0412 18:20:42.926285 1593 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.43:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.43:6443: connect: connection refused" interval="400ms" Apr 12 18:20:42.926405 kubelet[1593]: I0412 18:20:42.926307 1593 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:43.029756 kubelet[1593]: I0412 18:20:43.029638 1593 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Apr 12 18:20:43.030045 kubelet[1593]: E0412 18:20:43.029993 1593 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.43:6443/api/v1/nodes\": dial tcp 10.0.0.43:6443: connect: connection refused" node="localhost" Apr 12 18:20:43.162020 kubelet[1593]: E0412 18:20:43.161990 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:43.162803 env[1102]: time="2024-04-12T18:20:43.162751607Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:d562bbd47681de03357399071ce5c29e,Namespace:kube-system,Attempt:0,}" Apr 12 18:20:43.175033 kubelet[1593]: E0412 18:20:43.175007 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:43.175621 env[1102]: time="2024-04-12T18:20:43.175388286Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:7ea520b1e9b7ab03da4e0d60de4cd055,Namespace:kube-system,Attempt:0,}" Apr 12 18:20:43.178678 kubelet[1593]: E0412 18:20:43.178640 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:43.179024 env[1102]: time="2024-04-12T18:20:43.178994167Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:7bb06e2cba933e89936d947f3ee7de1d,Namespace:kube-system,Attempt:0,}" Apr 12 18:20:43.327534 kubelet[1593]: E0412 18:20:43.327422 1593 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.43:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.43:6443: connect: connection refused" interval="800ms" Apr 12 18:20:43.431243 kubelet[1593]: I0412 18:20:43.431192 1593 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Apr 12 18:20:43.431619 kubelet[1593]: E0412 18:20:43.431565 1593 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.43:6443/api/v1/nodes\": dial tcp 10.0.0.43:6443: connect: connection refused" node="localhost" Apr 12 18:20:43.629974 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2123860296.mount: Deactivated successfully. Apr 12 18:20:43.631928 kubelet[1593]: W0412 18:20:43.631876 1593 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://10.0.0.43:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:43.632107 kubelet[1593]: E0412 18:20:43.632090 1593 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.43:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:43.634795 env[1102]: time="2024-04-12T18:20:43.634713807Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.637950 env[1102]: time="2024-04-12T18:20:43.637915645Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.639281 env[1102]: time="2024-04-12T18:20:43.639250820Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.640371 env[1102]: time="2024-04-12T18:20:43.640345185Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.641452 env[1102]: time="2024-04-12T18:20:43.641394362Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.643026 env[1102]: time="2024-04-12T18:20:43.642995661Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.644318 env[1102]: time="2024-04-12T18:20:43.644290608Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.646519 env[1102]: time="2024-04-12T18:20:43.646495012Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.647284 env[1102]: time="2024-04-12T18:20:43.647262511Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.649817 env[1102]: time="2024-04-12T18:20:43.649787344Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.654053 env[1102]: time="2024-04-12T18:20:43.654019964Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.655004 env[1102]: time="2024-04-12T18:20:43.654980287Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:20:43.680069 env[1102]: time="2024-04-12T18:20:43.680003757Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:20:43.680069 env[1102]: time="2024-04-12T18:20:43.680044665Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:20:43.680227 env[1102]: time="2024-04-12T18:20:43.680058981Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:20:43.680275 env[1102]: time="2024-04-12T18:20:43.680230291Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/2b5fa32c46a3576be869375db1691dc7b649e9b7d3947f5e7a70dd4128813134 pid=1651 runtime=io.containerd.runc.v2 Apr 12 18:20:43.680528 env[1102]: time="2024-04-12T18:20:43.680484658Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:20:43.680693 env[1102]: time="2024-04-12T18:20:43.680520967Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:20:43.680693 env[1102]: time="2024-04-12T18:20:43.680533844Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:20:43.680782 env[1102]: time="2024-04-12T18:20:43.680707914Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ef7a0f42226bc6a52dce0a3486d687475d68c4f656aaf345f7bdb8ebe528031d pid=1652 runtime=io.containerd.runc.v2 Apr 12 18:20:43.682036 env[1102]: time="2024-04-12T18:20:43.681983186Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:20:43.682124 env[1102]: time="2024-04-12T18:20:43.682022815Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:20:43.682124 env[1102]: time="2024-04-12T18:20:43.682033292Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:20:43.682190 env[1102]: time="2024-04-12T18:20:43.682149858Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/5a1ef719d22973e743356a302aa509d71d7938415bb42233fc8b906b2f95b3ad pid=1650 runtime=io.containerd.runc.v2 Apr 12 18:20:43.692279 systemd[1]: Started cri-containerd-5a1ef719d22973e743356a302aa509d71d7938415bb42233fc8b906b2f95b3ad.scope. Apr 12 18:20:43.700277 systemd[1]: Started cri-containerd-ef7a0f42226bc6a52dce0a3486d687475d68c4f656aaf345f7bdb8ebe528031d.scope. Apr 12 18:20:43.710106 systemd[1]: Started cri-containerd-2b5fa32c46a3576be869375db1691dc7b649e9b7d3947f5e7a70dd4128813134.scope. Apr 12 18:20:43.721984 kubelet[1593]: W0412 18:20:43.721902 1593 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://10.0.0.43:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:43.721984 kubelet[1593]: E0412 18:20:43.721961 1593 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.43:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:43.765868 env[1102]: time="2024-04-12T18:20:43.764816837Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:7bb06e2cba933e89936d947f3ee7de1d,Namespace:kube-system,Attempt:0,} returns sandbox id \"5a1ef719d22973e743356a302aa509d71d7938415bb42233fc8b906b2f95b3ad\"" Apr 12 18:20:43.765965 kubelet[1593]: E0412 18:20:43.765648 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:43.767816 env[1102]: time="2024-04-12T18:20:43.767769026Z" level=info msg="CreateContainer within sandbox \"5a1ef719d22973e743356a302aa509d71d7938415bb42233fc8b906b2f95b3ad\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 12 18:20:43.772937 env[1102]: time="2024-04-12T18:20:43.772903347Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:7ea520b1e9b7ab03da4e0d60de4cd055,Namespace:kube-system,Attempt:0,} returns sandbox id \"2b5fa32c46a3576be869375db1691dc7b649e9b7d3947f5e7a70dd4128813134\"" Apr 12 18:20:43.773909 env[1102]: time="2024-04-12T18:20:43.773882145Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:d562bbd47681de03357399071ce5c29e,Namespace:kube-system,Attempt:0,} returns sandbox id \"ef7a0f42226bc6a52dce0a3486d687475d68c4f656aaf345f7bdb8ebe528031d\"" Apr 12 18:20:43.774775 kubelet[1593]: E0412 18:20:43.774756 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:43.774859 kubelet[1593]: E0412 18:20:43.774826 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:43.777099 kubelet[1593]: W0412 18:20:43.776867 1593 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://10.0.0.43:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:43.777099 kubelet[1593]: E0412 18:20:43.777003 1593 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.43:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.43:6443: connect: connection refused Apr 12 18:20:43.777720 env[1102]: time="2024-04-12T18:20:43.777687168Z" level=info msg="CreateContainer within sandbox \"ef7a0f42226bc6a52dce0a3486d687475d68c4f656aaf345f7bdb8ebe528031d\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 12 18:20:43.777902 env[1102]: time="2024-04-12T18:20:43.777685409Z" level=info msg="CreateContainer within sandbox \"2b5fa32c46a3576be869375db1691dc7b649e9b7d3947f5e7a70dd4128813134\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 12 18:20:43.784992 env[1102]: time="2024-04-12T18:20:43.784946676Z" level=info msg="CreateContainer within sandbox \"5a1ef719d22973e743356a302aa509d71d7938415bb42233fc8b906b2f95b3ad\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"3872e8d48a21b5a7b5d9ecd0ca8c6af9923154bd0013c23b2c053f9fd2f6f26a\"" Apr 12 18:20:43.785623 env[1102]: time="2024-04-12T18:20:43.785594889Z" level=info msg="StartContainer for \"3872e8d48a21b5a7b5d9ecd0ca8c6af9923154bd0013c23b2c053f9fd2f6f26a\"" Apr 12 18:20:43.794538 env[1102]: time="2024-04-12T18:20:43.794495925Z" level=info msg="CreateContainer within sandbox \"2b5fa32c46a3576be869375db1691dc7b649e9b7d3947f5e7a70dd4128813134\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"f9c105b01e4068f4d6fab44af50f31c23758a5bd2920849b85c0a0aa3405034e\"" Apr 12 18:20:43.795096 env[1102]: time="2024-04-12T18:20:43.795047326Z" level=info msg="StartContainer for \"f9c105b01e4068f4d6fab44af50f31c23758a5bd2920849b85c0a0aa3405034e\"" Apr 12 18:20:43.795809 env[1102]: time="2024-04-12T18:20:43.795780714Z" level=info msg="CreateContainer within sandbox \"ef7a0f42226bc6a52dce0a3486d687475d68c4f656aaf345f7bdb8ebe528031d\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"b7af115603159355076a62903f0433b9d2a1e8fc9eb090c61b116e4f44d99099\"" Apr 12 18:20:43.796149 env[1102]: time="2024-04-12T18:20:43.796121736Z" level=info msg="StartContainer for \"b7af115603159355076a62903f0433b9d2a1e8fc9eb090c61b116e4f44d99099\"" Apr 12 18:20:43.802836 systemd[1]: Started cri-containerd-3872e8d48a21b5a7b5d9ecd0ca8c6af9923154bd0013c23b2c053f9fd2f6f26a.scope. Apr 12 18:20:43.811369 systemd[1]: Started cri-containerd-b7af115603159355076a62903f0433b9d2a1e8fc9eb090c61b116e4f44d99099.scope. Apr 12 18:20:43.825543 systemd[1]: Started cri-containerd-f9c105b01e4068f4d6fab44af50f31c23758a5bd2920849b85c0a0aa3405034e.scope. Apr 12 18:20:43.864742 env[1102]: time="2024-04-12T18:20:43.864691057Z" level=info msg="StartContainer for \"b7af115603159355076a62903f0433b9d2a1e8fc9eb090c61b116e4f44d99099\" returns successfully" Apr 12 18:20:43.877156 env[1102]: time="2024-04-12T18:20:43.877106200Z" level=info msg="StartContainer for \"f9c105b01e4068f4d6fab44af50f31c23758a5bd2920849b85c0a0aa3405034e\" returns successfully" Apr 12 18:20:43.928538 env[1102]: time="2024-04-12T18:20:43.928394900Z" level=info msg="StartContainer for \"3872e8d48a21b5a7b5d9ecd0ca8c6af9923154bd0013c23b2c053f9fd2f6f26a\" returns successfully" Apr 12 18:20:44.232905 kubelet[1593]: I0412 18:20:44.232799 1593 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Apr 12 18:20:44.751191 kubelet[1593]: E0412 18:20:44.751164 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:44.752740 kubelet[1593]: E0412 18:20:44.752720 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:44.754967 kubelet[1593]: E0412 18:20:44.754943 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:45.756319 kubelet[1593]: E0412 18:20:45.756285 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:45.769961 kubelet[1593]: E0412 18:20:45.769931 1593 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:45.869083 kubelet[1593]: E0412 18:20:45.869048 1593 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Apr 12 18:20:45.950218 kubelet[1593]: I0412 18:20:45.950111 1593 kubelet_node_status.go:73] "Successfully registered node" node="localhost" Apr 12 18:20:46.724609 kubelet[1593]: I0412 18:20:46.724480 1593 apiserver.go:52] "Watching apiserver" Apr 12 18:20:46.825933 kubelet[1593]: I0412 18:20:46.825898 1593 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Apr 12 18:20:48.513842 systemd[1]: Reloading. Apr 12 18:20:48.565928 /usr/lib/systemd/system-generators/torcx-generator[1895]: time="2024-04-12T18:20:48Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.3 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.3 /var/lib/torcx/store]" Apr 12 18:20:48.566270 /usr/lib/systemd/system-generators/torcx-generator[1895]: time="2024-04-12T18:20:48Z" level=info msg="torcx already run" Apr 12 18:20:48.624675 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Apr 12 18:20:48.624693 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Apr 12 18:20:48.640555 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 12 18:20:48.729480 systemd[1]: Stopping kubelet.service... Apr 12 18:20:48.743988 systemd[1]: kubelet.service: Deactivated successfully. Apr 12 18:20:48.744193 systemd[1]: Stopped kubelet.service. Apr 12 18:20:48.744241 systemd[1]: kubelet.service: Consumed 1.075s CPU time. Apr 12 18:20:48.745917 systemd[1]: Started kubelet.service. Apr 12 18:20:48.803262 kubelet[1933]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 12 18:20:48.803262 kubelet[1933]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Apr 12 18:20:48.803262 kubelet[1933]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 12 18:20:48.803262 kubelet[1933]: I0412 18:20:48.802800 1933 server.go:203] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 12 18:20:48.807309 kubelet[1933]: I0412 18:20:48.807284 1933 server.go:467] "Kubelet version" kubeletVersion="v1.28.1" Apr 12 18:20:48.807414 kubelet[1933]: I0412 18:20:48.807403 1933 server.go:469] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 12 18:20:48.807689 kubelet[1933]: I0412 18:20:48.807668 1933 server.go:895] "Client rotation is on, will bootstrap in background" Apr 12 18:20:48.809194 kubelet[1933]: I0412 18:20:48.809171 1933 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Apr 12 18:20:48.810189 kubelet[1933]: I0412 18:20:48.810155 1933 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 12 18:20:48.816025 kubelet[1933]: W0412 18:20:48.816001 1933 machine.go:65] Cannot read vendor id correctly, set empty. Apr 12 18:20:48.816719 kubelet[1933]: I0412 18:20:48.816703 1933 server.go:725] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 12 18:20:48.816942 kubelet[1933]: I0412 18:20:48.816930 1933 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 12 18:20:48.817108 kubelet[1933]: I0412 18:20:48.817070 1933 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Apr 12 18:20:48.817108 kubelet[1933]: I0412 18:20:48.817107 1933 topology_manager.go:138] "Creating topology manager with none policy" Apr 12 18:20:48.817222 kubelet[1933]: I0412 18:20:48.817116 1933 container_manager_linux.go:301] "Creating device plugin manager" Apr 12 18:20:48.817222 kubelet[1933]: I0412 18:20:48.817142 1933 state_mem.go:36] "Initialized new in-memory state store" Apr 12 18:20:48.817278 kubelet[1933]: I0412 18:20:48.817226 1933 kubelet.go:393] "Attempting to sync node with API server" Apr 12 18:20:48.817278 kubelet[1933]: I0412 18:20:48.817245 1933 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 12 18:20:48.817278 kubelet[1933]: I0412 18:20:48.817263 1933 kubelet.go:309] "Adding apiserver pod source" Apr 12 18:20:48.817278 kubelet[1933]: I0412 18:20:48.817277 1933 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 12 18:20:48.818047 kubelet[1933]: I0412 18:20:48.818027 1933 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Apr 12 18:20:48.818604 kubelet[1933]: I0412 18:20:48.818584 1933 server.go:1232] "Started kubelet" Apr 12 18:20:48.819554 kubelet[1933]: I0412 18:20:48.819529 1933 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 12 18:20:48.820537 kubelet[1933]: I0412 18:20:48.820509 1933 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Apr 12 18:20:48.821133 kubelet[1933]: I0412 18:20:48.821094 1933 server.go:462] "Adding debug handlers to kubelet server" Apr 12 18:20:48.822557 kubelet[1933]: E0412 18:20:48.822334 1933 cri_stats_provider.go:448] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Apr 12 18:20:48.822693 kubelet[1933]: E0412 18:20:48.822562 1933 kubelet.go:1431] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 12 18:20:48.830240 kubelet[1933]: I0412 18:20:48.830200 1933 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10 Apr 12 18:20:48.830432 kubelet[1933]: I0412 18:20:48.830403 1933 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Apr 12 18:20:48.831136 kubelet[1933]: I0412 18:20:48.830602 1933 volume_manager.go:291] "Starting Kubelet Volume Manager" Apr 12 18:20:48.831136 kubelet[1933]: I0412 18:20:48.830696 1933 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Apr 12 18:20:48.831136 kubelet[1933]: I0412 18:20:48.830810 1933 reconciler_new.go:29] "Reconciler: start to sync state" Apr 12 18:20:48.850072 kubelet[1933]: I0412 18:20:48.850029 1933 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Apr 12 18:20:48.851862 kubelet[1933]: I0412 18:20:48.851820 1933 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Apr 12 18:20:48.851862 kubelet[1933]: I0412 18:20:48.851850 1933 status_manager.go:217] "Starting to sync pod status with apiserver" Apr 12 18:20:48.851862 kubelet[1933]: I0412 18:20:48.851867 1933 kubelet.go:2303] "Starting kubelet main sync loop" Apr 12 18:20:48.852004 kubelet[1933]: E0412 18:20:48.851919 1933 kubelet.go:2327] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 12 18:20:48.877919 kubelet[1933]: I0412 18:20:48.877894 1933 cpu_manager.go:214] "Starting CPU manager" policy="none" Apr 12 18:20:48.877919 kubelet[1933]: I0412 18:20:48.877917 1933 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Apr 12 18:20:48.878069 kubelet[1933]: I0412 18:20:48.877936 1933 state_mem.go:36] "Initialized new in-memory state store" Apr 12 18:20:48.878096 kubelet[1933]: I0412 18:20:48.878073 1933 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 12 18:20:48.878096 kubelet[1933]: I0412 18:20:48.878092 1933 state_mem.go:96] "Updated CPUSet assignments" assignments={} Apr 12 18:20:48.878145 kubelet[1933]: I0412 18:20:48.878099 1933 policy_none.go:49] "None policy: Start" Apr 12 18:20:48.878803 kubelet[1933]: I0412 18:20:48.878786 1933 memory_manager.go:169] "Starting memorymanager" policy="None" Apr 12 18:20:48.878877 kubelet[1933]: I0412 18:20:48.878812 1933 state_mem.go:35] "Initializing new in-memory state store" Apr 12 18:20:48.878952 kubelet[1933]: I0412 18:20:48.878938 1933 state_mem.go:75] "Updated machine memory state" Apr 12 18:20:48.883878 kubelet[1933]: I0412 18:20:48.883812 1933 manager.go:471] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 12 18:20:48.884185 kubelet[1933]: I0412 18:20:48.884167 1933 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 12 18:20:48.896436 sudo[1965]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Apr 12 18:20:48.896677 sudo[1965]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Apr 12 18:20:48.933915 kubelet[1933]: I0412 18:20:48.933883 1933 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Apr 12 18:20:48.941903 kubelet[1933]: I0412 18:20:48.941734 1933 kubelet_node_status.go:108] "Node was previously registered" node="localhost" Apr 12 18:20:48.941903 kubelet[1933]: I0412 18:20:48.941815 1933 kubelet_node_status.go:73] "Successfully registered node" node="localhost" Apr 12 18:20:48.952063 kubelet[1933]: I0412 18:20:48.952039 1933 topology_manager.go:215] "Topology Admit Handler" podUID="d562bbd47681de03357399071ce5c29e" podNamespace="kube-system" podName="kube-apiserver-localhost" Apr 12 18:20:48.952284 kubelet[1933]: I0412 18:20:48.952268 1933 topology_manager.go:215] "Topology Admit Handler" podUID="7ea520b1e9b7ab03da4e0d60de4cd055" podNamespace="kube-system" podName="kube-controller-manager-localhost" Apr 12 18:20:48.952386 kubelet[1933]: I0412 18:20:48.952373 1933 topology_manager.go:215] "Topology Admit Handler" podUID="7bb06e2cba933e89936d947f3ee7de1d" podNamespace="kube-system" podName="kube-scheduler-localhost" Apr 12 18:20:49.132721 kubelet[1933]: I0412 18:20:49.132631 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:49.132949 kubelet[1933]: I0412 18:20:49.132933 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:49.133079 kubelet[1933]: I0412 18:20:49.133069 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/7bb06e2cba933e89936d947f3ee7de1d-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"7bb06e2cba933e89936d947f3ee7de1d\") " pod="kube-system/kube-scheduler-localhost" Apr 12 18:20:49.133201 kubelet[1933]: I0412 18:20:49.133189 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d562bbd47681de03357399071ce5c29e-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"d562bbd47681de03357399071ce5c29e\") " pod="kube-system/kube-apiserver-localhost" Apr 12 18:20:49.133288 kubelet[1933]: I0412 18:20:49.133277 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d562bbd47681de03357399071ce5c29e-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"d562bbd47681de03357399071ce5c29e\") " pod="kube-system/kube-apiserver-localhost" Apr 12 18:20:49.133362 kubelet[1933]: I0412 18:20:49.133354 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:49.133492 kubelet[1933]: I0412 18:20:49.133479 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:49.133624 kubelet[1933]: I0412 18:20:49.133611 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7ea520b1e9b7ab03da4e0d60de4cd055-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"7ea520b1e9b7ab03da4e0d60de4cd055\") " pod="kube-system/kube-controller-manager-localhost" Apr 12 18:20:49.133743 kubelet[1933]: I0412 18:20:49.133731 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d562bbd47681de03357399071ce5c29e-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"d562bbd47681de03357399071ce5c29e\") " pod="kube-system/kube-apiserver-localhost" Apr 12 18:20:49.263042 kubelet[1933]: E0412 18:20:49.263013 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:49.263602 kubelet[1933]: E0412 18:20:49.263584 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:49.264330 kubelet[1933]: E0412 18:20:49.264310 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:49.336551 sudo[1965]: pam_unix(sudo:session): session closed for user root Apr 12 18:20:49.817612 kubelet[1933]: I0412 18:20:49.817560 1933 apiserver.go:52] "Watching apiserver" Apr 12 18:20:49.831842 kubelet[1933]: I0412 18:20:49.831808 1933 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Apr 12 18:20:49.860902 kubelet[1933]: E0412 18:20:49.860877 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:49.861036 kubelet[1933]: E0412 18:20:49.861018 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:49.866225 kubelet[1933]: E0412 18:20:49.866160 1933 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Apr 12 18:20:49.866631 kubelet[1933]: E0412 18:20:49.866619 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:49.878927 kubelet[1933]: I0412 18:20:49.878896 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.878852363 podCreationTimestamp="2024-04-12 18:20:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:20:49.87865559 +0000 UTC m=+1.129290937" watchObservedRunningTime="2024-04-12 18:20:49.878852363 +0000 UTC m=+1.129487750" Apr 12 18:20:49.890861 kubelet[1933]: I0412 18:20:49.890821 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.890790495 podCreationTimestamp="2024-04-12 18:20:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:20:49.884690727 +0000 UTC m=+1.135326074" watchObservedRunningTime="2024-04-12 18:20:49.890790495 +0000 UTC m=+1.141425842" Apr 12 18:20:49.890969 kubelet[1933]: I0412 18:20:49.890926 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.890910359 podCreationTimestamp="2024-04-12 18:20:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:20:49.890884602 +0000 UTC m=+1.141519909" watchObservedRunningTime="2024-04-12 18:20:49.890910359 +0000 UTC m=+1.141545666" Apr 12 18:20:50.706489 sudo[1194]: pam_unix(sudo:session): session closed for user root Apr 12 18:20:50.707977 sshd[1191]: pam_unix(sshd:session): session closed for user core Apr 12 18:20:50.710789 systemd[1]: sshd@4-10.0.0.43:22-10.0.0.1:34354.service: Deactivated successfully. Apr 12 18:20:50.711748 systemd[1]: session-5.scope: Deactivated successfully. Apr 12 18:20:50.711931 systemd[1]: session-5.scope: Consumed 6.069s CPU time. Apr 12 18:20:50.712386 systemd-logind[1089]: Session 5 logged out. Waiting for processes to exit. Apr 12 18:20:50.713114 systemd-logind[1089]: Removed session 5. Apr 12 18:20:50.862501 kubelet[1933]: E0412 18:20:50.862468 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:56.064331 kubelet[1933]: E0412 18:20:56.064291 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:56.870412 kubelet[1933]: E0412 18:20:56.870379 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:56.888429 kubelet[1933]: E0412 18:20:56.888405 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:57.871934 kubelet[1933]: E0412 18:20:57.871895 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:59.234695 kubelet[1933]: E0412 18:20:59.234665 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:20:59.875279 kubelet[1933]: E0412 18:20:59.875251 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:00.790131 kubelet[1933]: I0412 18:21:00.790097 1933 kuberuntime_manager.go:1463] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 12 18:21:00.790564 env[1102]: time="2024-04-12T18:21:00.790508886Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 12 18:21:00.790983 kubelet[1933]: I0412 18:21:00.790955 1933 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 12 18:21:01.851994 kubelet[1933]: I0412 18:21:01.851957 1933 topology_manager.go:215] "Topology Admit Handler" podUID="45fb15a9-3f8e-4c17-ab63-be7da6d35da7" podNamespace="kube-system" podName="cilium-4pzmr" Apr 12 18:21:01.852612 kubelet[1933]: I0412 18:21:01.852594 1933 topology_manager.go:215] "Topology Admit Handler" podUID="3be6ac03-a531-4110-89f0-142150c26fcb" podNamespace="kube-system" podName="kube-proxy-5jhjg" Apr 12 18:21:01.857368 systemd[1]: Created slice kubepods-burstable-pod45fb15a9_3f8e_4c17_ab63_be7da6d35da7.slice. Apr 12 18:21:01.861302 systemd[1]: Created slice kubepods-besteffort-pod3be6ac03_a531_4110_89f0_142150c26fcb.slice. Apr 12 18:21:01.914490 kubelet[1933]: I0412 18:21:01.914448 1933 topology_manager.go:215] "Topology Admit Handler" podUID="03731d2d-2199-4d64-a3aa-c64459041e8b" podNamespace="kube-system" podName="cilium-operator-6bc8ccdb58-brd4z" Apr 12 18:21:01.914984 kubelet[1933]: I0412 18:21:01.914959 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-run\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915090 kubelet[1933]: I0412 18:21:01.914996 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cni-path\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915090 kubelet[1933]: I0412 18:21:01.915019 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-clustermesh-secrets\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915090 kubelet[1933]: I0412 18:21:01.915038 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-hubble-tls\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915090 kubelet[1933]: I0412 18:21:01.915066 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-bpf-maps\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915090 kubelet[1933]: I0412 18:21:01.915085 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-host-proc-sys-net\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915090 kubelet[1933]: I0412 18:21:01.915107 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-etc-cni-netd\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915288 kubelet[1933]: I0412 18:21:01.915132 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-config-path\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915805 kubelet[1933]: I0412 18:21:01.915784 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hwzf\" (UniqueName: \"kubernetes.io/projected/3be6ac03-a531-4110-89f0-142150c26fcb-kube-api-access-9hwzf\") pod \"kube-proxy-5jhjg\" (UID: \"3be6ac03-a531-4110-89f0-142150c26fcb\") " pod="kube-system/kube-proxy-5jhjg" Apr 12 18:21:01.915878 kubelet[1933]: I0412 18:21:01.915819 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/3be6ac03-a531-4110-89f0-142150c26fcb-lib-modules\") pod \"kube-proxy-5jhjg\" (UID: \"3be6ac03-a531-4110-89f0-142150c26fcb\") " pod="kube-system/kube-proxy-5jhjg" Apr 12 18:21:01.915878 kubelet[1933]: I0412 18:21:01.915844 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-host-proc-sys-kernel\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915878 kubelet[1933]: I0412 18:21:01.915865 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-cgroup\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915957 kubelet[1933]: I0412 18:21:01.915882 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-lib-modules\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915957 kubelet[1933]: I0412 18:21:01.915908 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-hostproc\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915957 kubelet[1933]: I0412 18:21:01.915926 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-xtables-lock\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.915957 kubelet[1933]: I0412 18:21:01.915944 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm5gm\" (UniqueName: \"kubernetes.io/projected/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-kube-api-access-nm5gm\") pod \"cilium-4pzmr\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " pod="kube-system/cilium-4pzmr" Apr 12 18:21:01.916051 kubelet[1933]: I0412 18:21:01.915962 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/3be6ac03-a531-4110-89f0-142150c26fcb-kube-proxy\") pod \"kube-proxy-5jhjg\" (UID: \"3be6ac03-a531-4110-89f0-142150c26fcb\") " pod="kube-system/kube-proxy-5jhjg" Apr 12 18:21:01.916051 kubelet[1933]: I0412 18:21:01.915980 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/3be6ac03-a531-4110-89f0-142150c26fcb-xtables-lock\") pod \"kube-proxy-5jhjg\" (UID: \"3be6ac03-a531-4110-89f0-142150c26fcb\") " pod="kube-system/kube-proxy-5jhjg" Apr 12 18:21:01.919464 systemd[1]: Created slice kubepods-besteffort-pod03731d2d_2199_4d64_a3aa_c64459041e8b.slice. Apr 12 18:21:02.016335 kubelet[1933]: I0412 18:21:02.016257 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq6t7\" (UniqueName: \"kubernetes.io/projected/03731d2d-2199-4d64-a3aa-c64459041e8b-kube-api-access-nq6t7\") pod \"cilium-operator-6bc8ccdb58-brd4z\" (UID: \"03731d2d-2199-4d64-a3aa-c64459041e8b\") " pod="kube-system/cilium-operator-6bc8ccdb58-brd4z" Apr 12 18:21:02.016471 kubelet[1933]: I0412 18:21:02.016456 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/03731d2d-2199-4d64-a3aa-c64459041e8b-cilium-config-path\") pod \"cilium-operator-6bc8ccdb58-brd4z\" (UID: \"03731d2d-2199-4d64-a3aa-c64459041e8b\") " pod="kube-system/cilium-operator-6bc8ccdb58-brd4z" Apr 12 18:21:02.160455 kubelet[1933]: E0412 18:21:02.159680 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:02.161010 env[1102]: time="2024-04-12T18:21:02.160955992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4pzmr,Uid:45fb15a9-3f8e-4c17-ab63-be7da6d35da7,Namespace:kube-system,Attempt:0,}" Apr 12 18:21:02.169575 kubelet[1933]: E0412 18:21:02.169542 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:02.171442 env[1102]: time="2024-04-12T18:21:02.170150008Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-5jhjg,Uid:3be6ac03-a531-4110-89f0-142150c26fcb,Namespace:kube-system,Attempt:0,}" Apr 12 18:21:02.175113 env[1102]: time="2024-04-12T18:21:02.175059525Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:21:02.175194 env[1102]: time="2024-04-12T18:21:02.175096244Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:21:02.175194 env[1102]: time="2024-04-12T18:21:02.175106244Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:21:02.175588 env[1102]: time="2024-04-12T18:21:02.175542269Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846 pid=2025 runtime=io.containerd.runc.v2 Apr 12 18:21:02.184929 env[1102]: time="2024-04-12T18:21:02.184757244Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:21:02.184929 env[1102]: time="2024-04-12T18:21:02.184791563Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:21:02.184929 env[1102]: time="2024-04-12T18:21:02.184801643Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:21:02.186868 systemd[1]: Started cri-containerd-9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846.scope. Apr 12 18:21:02.187776 env[1102]: time="2024-04-12T18:21:02.187696107Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/dc97a7d8fe197b8d563975782ea52c3ee159c24d7748a234f50f1161a545d9bb pid=2052 runtime=io.containerd.runc.v2 Apr 12 18:21:02.203750 systemd[1]: Started cri-containerd-dc97a7d8fe197b8d563975782ea52c3ee159c24d7748a234f50f1161a545d9bb.scope. Apr 12 18:21:02.222592 kubelet[1933]: E0412 18:21:02.221908 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:02.222947 env[1102]: time="2024-04-12T18:21:02.222901182Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6bc8ccdb58-brd4z,Uid:03731d2d-2199-4d64-a3aa-c64459041e8b,Namespace:kube-system,Attempt:0,}" Apr 12 18:21:02.233781 env[1102]: time="2024-04-12T18:21:02.233745223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-4pzmr,Uid:45fb15a9-3f8e-4c17-ab63-be7da6d35da7,Namespace:kube-system,Attempt:0,} returns sandbox id \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\"" Apr 12 18:21:02.235867 kubelet[1933]: E0412 18:21:02.235846 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:02.237529 env[1102]: time="2024-04-12T18:21:02.237496739Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Apr 12 18:21:02.241292 env[1102]: time="2024-04-12T18:21:02.241232815Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:21:02.241392 env[1102]: time="2024-04-12T18:21:02.241268374Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:21:02.241392 env[1102]: time="2024-04-12T18:21:02.241278693Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:21:02.241617 env[1102]: time="2024-04-12T18:21:02.241541205Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b pid=2109 runtime=io.containerd.runc.v2 Apr 12 18:21:02.243749 env[1102]: time="2024-04-12T18:21:02.243703853Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-5jhjg,Uid:3be6ac03-a531-4110-89f0-142150c26fcb,Namespace:kube-system,Attempt:0,} returns sandbox id \"dc97a7d8fe197b8d563975782ea52c3ee159c24d7748a234f50f1161a545d9bb\"" Apr 12 18:21:02.244591 kubelet[1933]: E0412 18:21:02.244370 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:02.246171 env[1102]: time="2024-04-12T18:21:02.246140692Z" level=info msg="CreateContainer within sandbox \"dc97a7d8fe197b8d563975782ea52c3ee159c24d7748a234f50f1161a545d9bb\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 12 18:21:02.259200 env[1102]: time="2024-04-12T18:21:02.259157982Z" level=info msg="CreateContainer within sandbox \"dc97a7d8fe197b8d563975782ea52c3ee159c24d7748a234f50f1161a545d9bb\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"fc3f98df79d01b13273008eeb9550bceeb1a7b042de019c050c3524c452917ec\"" Apr 12 18:21:02.259513 systemd[1]: Started cri-containerd-a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b.scope. Apr 12 18:21:02.261462 env[1102]: time="2024-04-12T18:21:02.261421427Z" level=info msg="StartContainer for \"fc3f98df79d01b13273008eeb9550bceeb1a7b042de019c050c3524c452917ec\"" Apr 12 18:21:02.279840 systemd[1]: Started cri-containerd-fc3f98df79d01b13273008eeb9550bceeb1a7b042de019c050c3524c452917ec.scope. Apr 12 18:21:02.314271 env[1102]: time="2024-04-12T18:21:02.314104123Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6bc8ccdb58-brd4z,Uid:03731d2d-2199-4d64-a3aa-c64459041e8b,Namespace:kube-system,Attempt:0,} returns sandbox id \"a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b\"" Apr 12 18:21:02.315248 kubelet[1933]: E0412 18:21:02.314853 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:02.319515 env[1102]: time="2024-04-12T18:21:02.319479625Z" level=info msg="StartContainer for \"fc3f98df79d01b13273008eeb9550bceeb1a7b042de019c050c3524c452917ec\" returns successfully" Apr 12 18:21:02.880720 kubelet[1933]: E0412 18:21:02.880692 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:02.890524 kubelet[1933]: I0412 18:21:02.890479 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-5jhjg" podStartSLOduration=1.8904403250000001 podCreationTimestamp="2024-04-12 18:21:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:21:02.889278564 +0000 UTC m=+14.139913871" watchObservedRunningTime="2024-04-12 18:21:02.890440325 +0000 UTC m=+14.141075672" Apr 12 18:21:03.010290 update_engine[1092]: I0412 18:21:03.010248 1092 update_attempter.cc:509] Updating boot flags... Apr 12 18:21:05.966183 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2112298441.mount: Deactivated successfully. Apr 12 18:21:08.260218 env[1102]: time="2024-04-12T18:21:08.260166375Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:21:08.261573 env[1102]: time="2024-04-12T18:21:08.261500422Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:21:08.263602 env[1102]: time="2024-04-12T18:21:08.263558531Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:21:08.264096 env[1102]: time="2024-04-12T18:21:08.264058478Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Apr 12 18:21:08.265006 env[1102]: time="2024-04-12T18:21:08.264972336Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Apr 12 18:21:08.266998 env[1102]: time="2024-04-12T18:21:08.266966286Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 12 18:21:08.276772 env[1102]: time="2024-04-12T18:21:08.276724125Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\"" Apr 12 18:21:08.277343 env[1102]: time="2024-04-12T18:21:08.277179834Z" level=info msg="StartContainer for \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\"" Apr 12 18:21:08.296278 systemd[1]: Started cri-containerd-ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f.scope. Apr 12 18:21:08.348729 env[1102]: time="2024-04-12T18:21:08.348683305Z" level=info msg="StartContainer for \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\" returns successfully" Apr 12 18:21:08.383526 systemd[1]: cri-containerd-ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f.scope: Deactivated successfully. Apr 12 18:21:08.564416 env[1102]: time="2024-04-12T18:21:08.564296330Z" level=info msg="shim disconnected" id=ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f Apr 12 18:21:08.564817 env[1102]: time="2024-04-12T18:21:08.564795478Z" level=warning msg="cleaning up after shim disconnected" id=ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f namespace=k8s.io Apr 12 18:21:08.564957 env[1102]: time="2024-04-12T18:21:08.564940874Z" level=info msg="cleaning up dead shim" Apr 12 18:21:08.571324 env[1102]: time="2024-04-12T18:21:08.571296837Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:21:08Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2362 runtime=io.containerd.runc.v2\n" Apr 12 18:21:08.891950 kubelet[1933]: E0412 18:21:08.891867 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:08.901250 env[1102]: time="2024-04-12T18:21:08.901194715Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Apr 12 18:21:08.916222 env[1102]: time="2024-04-12T18:21:08.916078387Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\"" Apr 12 18:21:08.916783 env[1102]: time="2024-04-12T18:21:08.916739210Z" level=info msg="StartContainer for \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\"" Apr 12 18:21:08.935276 systemd[1]: Started cri-containerd-3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191.scope. Apr 12 18:21:08.969680 env[1102]: time="2024-04-12T18:21:08.969628542Z" level=info msg="StartContainer for \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\" returns successfully" Apr 12 18:21:08.977830 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 12 18:21:08.978061 systemd[1]: Stopped systemd-sysctl.service. Apr 12 18:21:08.978852 systemd[1]: Stopping systemd-sysctl.service... Apr 12 18:21:08.980342 systemd[1]: Starting systemd-sysctl.service... Apr 12 18:21:08.980598 systemd[1]: cri-containerd-3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191.scope: Deactivated successfully. Apr 12 18:21:08.990690 systemd[1]: Finished systemd-sysctl.service. Apr 12 18:21:09.017069 env[1102]: time="2024-04-12T18:21:09.017013107Z" level=info msg="shim disconnected" id=3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191 Apr 12 18:21:09.017069 env[1102]: time="2024-04-12T18:21:09.017065025Z" level=warning msg="cleaning up after shim disconnected" id=3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191 namespace=k8s.io Apr 12 18:21:09.017069 env[1102]: time="2024-04-12T18:21:09.017075545Z" level=info msg="cleaning up dead shim" Apr 12 18:21:09.023860 env[1102]: time="2024-04-12T18:21:09.023822906Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:21:09Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2428 runtime=io.containerd.runc.v2\n" Apr 12 18:21:09.274621 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f-rootfs.mount: Deactivated successfully. Apr 12 18:21:09.680402 env[1102]: time="2024-04-12T18:21:09.680360148Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:21:09.681822 env[1102]: time="2024-04-12T18:21:09.681793554Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:21:09.683391 env[1102]: time="2024-04-12T18:21:09.683366317Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:21:09.684035 env[1102]: time="2024-04-12T18:21:09.684008982Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Apr 12 18:21:09.688043 env[1102]: time="2024-04-12T18:21:09.688012287Z" level=info msg="CreateContainer within sandbox \"a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Apr 12 18:21:09.699582 env[1102]: time="2024-04-12T18:21:09.699526255Z" level=info msg="CreateContainer within sandbox \"a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\"" Apr 12 18:21:09.699921 env[1102]: time="2024-04-12T18:21:09.699891166Z" level=info msg="StartContainer for \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\"" Apr 12 18:21:09.717437 systemd[1]: Started cri-containerd-9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72.scope. Apr 12 18:21:09.764887 env[1102]: time="2024-04-12T18:21:09.764842631Z" level=info msg="StartContainer for \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\" returns successfully" Apr 12 18:21:09.895218 kubelet[1933]: E0412 18:21:09.895190 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:09.897320 kubelet[1933]: E0412 18:21:09.897183 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:09.899075 env[1102]: time="2024-04-12T18:21:09.899029859Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Apr 12 18:21:09.916519 env[1102]: time="2024-04-12T18:21:09.916464807Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\"" Apr 12 18:21:09.917133 env[1102]: time="2024-04-12T18:21:09.917106872Z" level=info msg="StartContainer for \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\"" Apr 12 18:21:09.933326 kubelet[1933]: I0412 18:21:09.933224 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-6bc8ccdb58-brd4z" podStartSLOduration=1.563923565 podCreationTimestamp="2024-04-12 18:21:01 +0000 UTC" firstStartedPulling="2024-04-12 18:21:02.315484157 +0000 UTC m=+13.566119504" lastFinishedPulling="2024-04-12 18:21:09.684745604 +0000 UTC m=+20.935380951" observedRunningTime="2024-04-12 18:21:09.904136459 +0000 UTC m=+21.154771766" watchObservedRunningTime="2024-04-12 18:21:09.933185012 +0000 UTC m=+21.183820359" Apr 12 18:21:09.944764 systemd[1]: Started cri-containerd-5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f.scope. Apr 12 18:21:10.010046 systemd[1]: cri-containerd-5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f.scope: Deactivated successfully. Apr 12 18:21:10.014026 env[1102]: time="2024-04-12T18:21:10.013989075Z" level=info msg="StartContainer for \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\" returns successfully" Apr 12 18:21:10.073393 env[1102]: time="2024-04-12T18:21:10.073347493Z" level=info msg="shim disconnected" id=5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f Apr 12 18:21:10.073393 env[1102]: time="2024-04-12T18:21:10.073392212Z" level=warning msg="cleaning up after shim disconnected" id=5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f namespace=k8s.io Apr 12 18:21:10.073393 env[1102]: time="2024-04-12T18:21:10.073401292Z" level=info msg="cleaning up dead shim" Apr 12 18:21:10.079983 env[1102]: time="2024-04-12T18:21:10.079937345Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:21:10Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2524 runtime=io.containerd.runc.v2\n" Apr 12 18:21:10.901001 kubelet[1933]: E0412 18:21:10.900962 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:10.901353 kubelet[1933]: E0412 18:21:10.901060 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:10.902744 env[1102]: time="2024-04-12T18:21:10.902691550Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Apr 12 18:21:10.915356 env[1102]: time="2024-04-12T18:21:10.915312025Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\"" Apr 12 18:21:10.917151 env[1102]: time="2024-04-12T18:21:10.917119184Z" level=info msg="StartContainer for \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\"" Apr 12 18:21:10.938115 systemd[1]: Started cri-containerd-57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3.scope. Apr 12 18:21:10.972967 systemd[1]: cri-containerd-57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3.scope: Deactivated successfully. Apr 12 18:21:10.976269 env[1102]: time="2024-04-12T18:21:10.976217609Z" level=info msg="StartContainer for \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\" returns successfully" Apr 12 18:21:10.993281 env[1102]: time="2024-04-12T18:21:10.993232624Z" level=info msg="shim disconnected" id=57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3 Apr 12 18:21:10.993281 env[1102]: time="2024-04-12T18:21:10.993276383Z" level=warning msg="cleaning up after shim disconnected" id=57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3 namespace=k8s.io Apr 12 18:21:10.993281 env[1102]: time="2024-04-12T18:21:10.993285383Z" level=info msg="cleaning up dead shim" Apr 12 18:21:10.999211 env[1102]: time="2024-04-12T18:21:10.999174530Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:21:10Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2580 runtime=io.containerd.runc.v2\n" Apr 12 18:21:11.274681 systemd[1]: run-containerd-runc-k8s.io-57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3-runc.KHrqJf.mount: Deactivated successfully. Apr 12 18:21:11.274781 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3-rootfs.mount: Deactivated successfully. Apr 12 18:21:11.904907 kubelet[1933]: E0412 18:21:11.904842 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:11.907799 env[1102]: time="2024-04-12T18:21:11.907759596Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Apr 12 18:21:11.921448 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2863773804.mount: Deactivated successfully. Apr 12 18:21:11.925445 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2189440076.mount: Deactivated successfully. Apr 12 18:21:11.928934 env[1102]: time="2024-04-12T18:21:11.928880259Z" level=info msg="CreateContainer within sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\"" Apr 12 18:21:11.929607 env[1102]: time="2024-04-12T18:21:11.929578804Z" level=info msg="StartContainer for \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\"" Apr 12 18:21:11.943700 systemd[1]: Started cri-containerd-40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5.scope. Apr 12 18:21:11.998125 env[1102]: time="2024-04-12T18:21:11.997728610Z" level=info msg="StartContainer for \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\" returns successfully" Apr 12 18:21:12.136459 kubelet[1933]: I0412 18:21:12.136412 1933 kubelet_node_status.go:493] "Fast updating node status as it just became ready" Apr 12 18:21:12.157039 kubelet[1933]: I0412 18:21:12.156913 1933 topology_manager.go:215] "Topology Admit Handler" podUID="3f9010c6-7077-452f-9867-e01ab14fbf7b" podNamespace="kube-system" podName="coredns-5dd5756b68-ks4m8" Apr 12 18:21:12.157259 kubelet[1933]: I0412 18:21:12.157237 1933 topology_manager.go:215] "Topology Admit Handler" podUID="41692888-d57c-4631-a66c-92ae4db03554" podNamespace="kube-system" podName="coredns-5dd5756b68-8wfk7" Apr 12 18:21:12.162384 systemd[1]: Created slice kubepods-burstable-pod3f9010c6_7077_452f_9867_e01ab14fbf7b.slice. Apr 12 18:21:12.166464 systemd[1]: Created slice kubepods-burstable-pod41692888_d57c_4631_a66c_92ae4db03554.slice. Apr 12 18:21:12.192861 kubelet[1933]: I0412 18:21:12.192810 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6g5x\" (UniqueName: \"kubernetes.io/projected/41692888-d57c-4631-a66c-92ae4db03554-kube-api-access-p6g5x\") pod \"coredns-5dd5756b68-8wfk7\" (UID: \"41692888-d57c-4631-a66c-92ae4db03554\") " pod="kube-system/coredns-5dd5756b68-8wfk7" Apr 12 18:21:12.192861 kubelet[1933]: I0412 18:21:12.192867 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxgrx\" (UniqueName: \"kubernetes.io/projected/3f9010c6-7077-452f-9867-e01ab14fbf7b-kube-api-access-gxgrx\") pod \"coredns-5dd5756b68-ks4m8\" (UID: \"3f9010c6-7077-452f-9867-e01ab14fbf7b\") " pod="kube-system/coredns-5dd5756b68-ks4m8" Apr 12 18:21:12.193044 kubelet[1933]: I0412 18:21:12.192893 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f9010c6-7077-452f-9867-e01ab14fbf7b-config-volume\") pod \"coredns-5dd5756b68-ks4m8\" (UID: \"3f9010c6-7077-452f-9867-e01ab14fbf7b\") " pod="kube-system/coredns-5dd5756b68-ks4m8" Apr 12 18:21:12.193044 kubelet[1933]: I0412 18:21:12.192919 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/41692888-d57c-4631-a66c-92ae4db03554-config-volume\") pod \"coredns-5dd5756b68-8wfk7\" (UID: \"41692888-d57c-4631-a66c-92ae4db03554\") " pod="kube-system/coredns-5dd5756b68-8wfk7" Apr 12 18:21:12.277596 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Apr 12 18:21:12.465529 kubelet[1933]: E0412 18:21:12.465495 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:12.466314 env[1102]: time="2024-04-12T18:21:12.466268099Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-ks4m8,Uid:3f9010c6-7077-452f-9867-e01ab14fbf7b,Namespace:kube-system,Attempt:0,}" Apr 12 18:21:12.468935 kubelet[1933]: E0412 18:21:12.468912 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:12.469305 env[1102]: time="2024-04-12T18:21:12.469266717Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-8wfk7,Uid:41692888-d57c-4631-a66c-92ae4db03554,Namespace:kube-system,Attempt:0,}" Apr 12 18:21:12.556655 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Apr 12 18:21:12.909414 kubelet[1933]: E0412 18:21:12.909307 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:13.911451 kubelet[1933]: E0412 18:21:13.911420 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:14.171640 systemd-networkd[1000]: cilium_host: Link UP Apr 12 18:21:14.174611 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Apr 12 18:21:14.174695 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Apr 12 18:21:14.174781 systemd-networkd[1000]: cilium_net: Link UP Apr 12 18:21:14.174966 systemd-networkd[1000]: cilium_net: Gained carrier Apr 12 18:21:14.175080 systemd-networkd[1000]: cilium_host: Gained carrier Apr 12 18:21:14.175166 systemd-networkd[1000]: cilium_net: Gained IPv6LL Apr 12 18:21:14.175278 systemd-networkd[1000]: cilium_host: Gained IPv6LL Apr 12 18:21:14.247197 systemd-networkd[1000]: cilium_vxlan: Link UP Apr 12 18:21:14.247203 systemd-networkd[1000]: cilium_vxlan: Gained carrier Apr 12 18:21:14.552607 kernel: NET: Registered PF_ALG protocol family Apr 12 18:21:14.912933 kubelet[1933]: E0412 18:21:14.912886 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:15.131982 systemd-networkd[1000]: lxc_health: Link UP Apr 12 18:21:15.145873 systemd-networkd[1000]: lxc_health: Gained carrier Apr 12 18:21:15.146597 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Apr 12 18:21:15.563746 systemd-networkd[1000]: lxce59c2bbf7cac: Link UP Apr 12 18:21:15.575601 systemd-networkd[1000]: lxc684797e7ba1e: Link UP Apr 12 18:21:15.585653 kernel: eth0: renamed from tmp23c9e Apr 12 18:21:15.592599 kernel: eth0: renamed from tmp0cf1e Apr 12 18:21:15.596173 systemd-networkd[1000]: cilium_vxlan: Gained IPv6LL Apr 12 18:21:15.598195 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Apr 12 18:21:15.598252 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxce59c2bbf7cac: link becomes ready Apr 12 18:21:15.598566 systemd-networkd[1000]: lxce59c2bbf7cac: Gained carrier Apr 12 18:21:15.599766 systemd-networkd[1000]: lxc684797e7ba1e: Gained carrier Apr 12 18:21:15.600606 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc684797e7ba1e: link becomes ready Apr 12 18:21:15.919228 kubelet[1933]: E0412 18:21:15.919195 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:16.173342 kubelet[1933]: I0412 18:21:16.173237 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-4pzmr" podStartSLOduration=9.145138011 podCreationTimestamp="2024-04-12 18:21:01 +0000 UTC" firstStartedPulling="2024-04-12 18:21:02.236271059 +0000 UTC m=+13.486906406" lastFinishedPulling="2024-04-12 18:21:08.264333791 +0000 UTC m=+19.514969178" observedRunningTime="2024-04-12 18:21:12.924095814 +0000 UTC m=+24.174731241" watchObservedRunningTime="2024-04-12 18:21:16.173200783 +0000 UTC m=+27.423836130" Apr 12 18:21:16.198094 systemd[1]: Started sshd@5-10.0.0.43:22-10.0.0.1:41152.service. Apr 12 18:21:16.224657 systemd-networkd[1000]: lxc_health: Gained IPv6LL Apr 12 18:21:16.247632 sshd[3133]: Accepted publickey for core from 10.0.0.1 port 41152 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:16.249187 sshd[3133]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:16.252411 systemd-logind[1089]: New session 6 of user core. Apr 12 18:21:16.253239 systemd[1]: Started session-6.scope. Apr 12 18:21:16.428998 sshd[3133]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:16.431501 systemd[1]: sshd@5-10.0.0.43:22-10.0.0.1:41152.service: Deactivated successfully. Apr 12 18:21:16.432235 systemd[1]: session-6.scope: Deactivated successfully. Apr 12 18:21:16.432809 systemd-logind[1089]: Session 6 logged out. Waiting for processes to exit. Apr 12 18:21:16.433655 systemd-logind[1089]: Removed session 6. Apr 12 18:21:16.864733 systemd-networkd[1000]: lxc684797e7ba1e: Gained IPv6LL Apr 12 18:21:16.915454 kubelet[1933]: E0412 18:21:16.915422 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:17.439756 systemd-networkd[1000]: lxce59c2bbf7cac: Gained IPv6LL Apr 12 18:21:17.917203 kubelet[1933]: E0412 18:21:17.917177 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:19.177527 env[1102]: time="2024-04-12T18:21:19.177354380Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:21:19.177527 env[1102]: time="2024-04-12T18:21:19.177393699Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:21:19.177527 env[1102]: time="2024-04-12T18:21:19.177404099Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:21:19.178060 env[1102]: time="2024-04-12T18:21:19.177611496Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/23c9e2caa65953a9ba7b7ca780c89f81a01e683e41d0b7aa96e912aeb1861f3f pid=3170 runtime=io.containerd.runc.v2 Apr 12 18:21:19.178132 env[1102]: time="2024-04-12T18:21:19.177649175Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:21:19.178132 env[1102]: time="2024-04-12T18:21:19.177675935Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:21:19.178132 env[1102]: time="2024-04-12T18:21:19.177685815Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:21:19.178132 env[1102]: time="2024-04-12T18:21:19.177840332Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0cf1e05d145e958002d960dec6a39061a70231849e887f914047f35b793256c8 pid=3173 runtime=io.containerd.runc.v2 Apr 12 18:21:19.189006 systemd[1]: Started cri-containerd-23c9e2caa65953a9ba7b7ca780c89f81a01e683e41d0b7aa96e912aeb1861f3f.scope. Apr 12 18:21:19.204812 systemd[1]: Started cri-containerd-0cf1e05d145e958002d960dec6a39061a70231849e887f914047f35b793256c8.scope. Apr 12 18:21:19.220748 systemd-resolved[1045]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 12 18:21:19.225050 systemd-resolved[1045]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Apr 12 18:21:19.239835 env[1102]: time="2024-04-12T18:21:19.239798516Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-8wfk7,Uid:41692888-d57c-4631-a66c-92ae4db03554,Namespace:kube-system,Attempt:0,} returns sandbox id \"23c9e2caa65953a9ba7b7ca780c89f81a01e683e41d0b7aa96e912aeb1861f3f\"" Apr 12 18:21:19.240473 kubelet[1933]: E0412 18:21:19.240453 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:19.247584 env[1102]: time="2024-04-12T18:21:19.247535954Z" level=info msg="CreateContainer within sandbox \"23c9e2caa65953a9ba7b7ca780c89f81a01e683e41d0b7aa96e912aeb1861f3f\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 12 18:21:19.248363 env[1102]: time="2024-04-12T18:21:19.247980227Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-ks4m8,Uid:3f9010c6-7077-452f-9867-e01ab14fbf7b,Namespace:kube-system,Attempt:0,} returns sandbox id \"0cf1e05d145e958002d960dec6a39061a70231849e887f914047f35b793256c8\"" Apr 12 18:21:19.249738 kubelet[1933]: E0412 18:21:19.249236 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:19.251181 env[1102]: time="2024-04-12T18:21:19.251136217Z" level=info msg="CreateContainer within sandbox \"0cf1e05d145e958002d960dec6a39061a70231849e887f914047f35b793256c8\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 12 18:21:19.265350 env[1102]: time="2024-04-12T18:21:19.265315074Z" level=info msg="CreateContainer within sandbox \"0cf1e05d145e958002d960dec6a39061a70231849e887f914047f35b793256c8\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"0cd116e1d2f10f3c04b316ef3aab3063cf64ff5ee9d239671178be85bc149169\"" Apr 12 18:21:19.265993 env[1102]: time="2024-04-12T18:21:19.265954504Z" level=info msg="StartContainer for \"0cd116e1d2f10f3c04b316ef3aab3063cf64ff5ee9d239671178be85bc149169\"" Apr 12 18:21:19.268181 env[1102]: time="2024-04-12T18:21:19.268138390Z" level=info msg="CreateContainer within sandbox \"23c9e2caa65953a9ba7b7ca780c89f81a01e683e41d0b7aa96e912aeb1861f3f\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"3f2b0e5e2a161a45f3bff5063742b250a17393c76e0332b86381860acc8e88a5\"" Apr 12 18:21:19.269608 env[1102]: time="2024-04-12T18:21:19.268563303Z" level=info msg="StartContainer for \"3f2b0e5e2a161a45f3bff5063742b250a17393c76e0332b86381860acc8e88a5\"" Apr 12 18:21:19.282425 systemd[1]: Started cri-containerd-0cd116e1d2f10f3c04b316ef3aab3063cf64ff5ee9d239671178be85bc149169.scope. Apr 12 18:21:19.289144 systemd[1]: Started cri-containerd-3f2b0e5e2a161a45f3bff5063742b250a17393c76e0332b86381860acc8e88a5.scope. Apr 12 18:21:19.313040 env[1102]: time="2024-04-12T18:21:19.312119857Z" level=info msg="StartContainer for \"0cd116e1d2f10f3c04b316ef3aab3063cf64ff5ee9d239671178be85bc149169\" returns successfully" Apr 12 18:21:19.328446 env[1102]: time="2024-04-12T18:21:19.328404040Z" level=info msg="StartContainer for \"3f2b0e5e2a161a45f3bff5063742b250a17393c76e0332b86381860acc8e88a5\" returns successfully" Apr 12 18:21:19.921946 kubelet[1933]: E0412 18:21:19.921922 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:19.924204 kubelet[1933]: E0412 18:21:19.924173 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:19.939504 kubelet[1933]: I0412 18:21:19.939480 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5dd5756b68-ks4m8" podStartSLOduration=18.939442694 podCreationTimestamp="2024-04-12 18:21:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:21:19.939131699 +0000 UTC m=+31.189767046" watchObservedRunningTime="2024-04-12 18:21:19.939442694 +0000 UTC m=+31.190078001" Apr 12 18:21:19.939737 kubelet[1933]: I0412 18:21:19.939723 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5dd5756b68-8wfk7" podStartSLOduration=18.93968769 podCreationTimestamp="2024-04-12 18:21:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:21:19.931557178 +0000 UTC m=+31.182192525" watchObservedRunningTime="2024-04-12 18:21:19.93968769 +0000 UTC m=+31.190323037" Apr 12 18:21:20.925767 kubelet[1933]: E0412 18:21:20.925725 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:20.926124 kubelet[1933]: E0412 18:21:20.926105 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:21.433114 systemd[1]: Started sshd@6-10.0.0.43:22-10.0.0.1:34132.service. Apr 12 18:21:21.479080 sshd[3326]: Accepted publickey for core from 10.0.0.1 port 34132 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:21.480644 sshd[3326]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:21.483957 systemd-logind[1089]: New session 7 of user core. Apr 12 18:21:21.484784 systemd[1]: Started session-7.scope. Apr 12 18:21:21.599133 sshd[3326]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:21.601507 systemd[1]: sshd@6-10.0.0.43:22-10.0.0.1:34132.service: Deactivated successfully. Apr 12 18:21:21.602289 systemd[1]: session-7.scope: Deactivated successfully. Apr 12 18:21:21.602804 systemd-logind[1089]: Session 7 logged out. Waiting for processes to exit. Apr 12 18:21:21.603539 systemd-logind[1089]: Removed session 7. Apr 12 18:21:21.927194 kubelet[1933]: E0412 18:21:21.927154 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:21.927993 kubelet[1933]: E0412 18:21:21.927964 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:21:26.603941 systemd[1]: Started sshd@7-10.0.0.43:22-10.0.0.1:34146.service. Apr 12 18:21:26.659665 sshd[3340]: Accepted publickey for core from 10.0.0.1 port 34146 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:26.661245 sshd[3340]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:26.664827 systemd-logind[1089]: New session 8 of user core. Apr 12 18:21:26.665715 systemd[1]: Started session-8.scope. Apr 12 18:21:26.779299 sshd[3340]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:26.781581 systemd[1]: session-8.scope: Deactivated successfully. Apr 12 18:21:26.782191 systemd[1]: sshd@7-10.0.0.43:22-10.0.0.1:34146.service: Deactivated successfully. Apr 12 18:21:26.783050 systemd-logind[1089]: Session 8 logged out. Waiting for processes to exit. Apr 12 18:21:26.783603 systemd-logind[1089]: Removed session 8. Apr 12 18:21:31.784016 systemd[1]: Started sshd@8-10.0.0.43:22-10.0.0.1:36162.service. Apr 12 18:21:31.827955 sshd[3354]: Accepted publickey for core from 10.0.0.1 port 36162 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:31.829137 sshd[3354]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:31.832342 systemd-logind[1089]: New session 9 of user core. Apr 12 18:21:31.833279 systemd[1]: Started session-9.scope. Apr 12 18:21:31.947521 sshd[3354]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:31.950516 systemd[1]: sshd@8-10.0.0.43:22-10.0.0.1:36162.service: Deactivated successfully. Apr 12 18:21:31.951274 systemd[1]: session-9.scope: Deactivated successfully. Apr 12 18:21:31.951931 systemd-logind[1089]: Session 9 logged out. Waiting for processes to exit. Apr 12 18:21:31.953393 systemd[1]: Started sshd@9-10.0.0.43:22-10.0.0.1:36164.service. Apr 12 18:21:31.954003 systemd-logind[1089]: Removed session 9. Apr 12 18:21:31.998683 sshd[3368]: Accepted publickey for core from 10.0.0.1 port 36164 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:32.000002 sshd[3368]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:32.003148 systemd-logind[1089]: New session 10 of user core. Apr 12 18:21:32.004076 systemd[1]: Started session-10.scope. Apr 12 18:21:32.684651 sshd[3368]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:32.686357 systemd[1]: Started sshd@10-10.0.0.43:22-10.0.0.1:36170.service. Apr 12 18:21:32.688615 systemd[1]: sshd@9-10.0.0.43:22-10.0.0.1:36164.service: Deactivated successfully. Apr 12 18:21:32.689429 systemd[1]: session-10.scope: Deactivated successfully. Apr 12 18:21:32.692404 systemd-logind[1089]: Session 10 logged out. Waiting for processes to exit. Apr 12 18:21:32.693193 systemd-logind[1089]: Removed session 10. Apr 12 18:21:32.735805 sshd[3382]: Accepted publickey for core from 10.0.0.1 port 36170 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:32.737002 sshd[3382]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:32.740382 systemd-logind[1089]: New session 11 of user core. Apr 12 18:21:32.741279 systemd[1]: Started session-11.scope. Apr 12 18:21:32.851543 sshd[3382]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:32.854134 systemd[1]: sshd@10-10.0.0.43:22-10.0.0.1:36170.service: Deactivated successfully. Apr 12 18:21:32.855017 systemd[1]: session-11.scope: Deactivated successfully. Apr 12 18:21:32.855765 systemd-logind[1089]: Session 11 logged out. Waiting for processes to exit. Apr 12 18:21:32.856630 systemd-logind[1089]: Removed session 11. Apr 12 18:21:37.856603 systemd[1]: Started sshd@11-10.0.0.43:22-10.0.0.1:36174.service. Apr 12 18:21:37.900566 sshd[3396]: Accepted publickey for core from 10.0.0.1 port 36174 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:37.902148 sshd[3396]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:37.906120 systemd-logind[1089]: New session 12 of user core. Apr 12 18:21:37.906613 systemd[1]: Started session-12.scope. Apr 12 18:21:38.015546 sshd[3396]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:38.018458 systemd[1]: sshd@11-10.0.0.43:22-10.0.0.1:36174.service: Deactivated successfully. Apr 12 18:21:38.019221 systemd[1]: session-12.scope: Deactivated successfully. Apr 12 18:21:38.019795 systemd-logind[1089]: Session 12 logged out. Waiting for processes to exit. Apr 12 18:21:38.021176 systemd[1]: Started sshd@12-10.0.0.43:22-10.0.0.1:36184.service. Apr 12 18:21:38.021831 systemd-logind[1089]: Removed session 12. Apr 12 18:21:38.065687 sshd[3409]: Accepted publickey for core from 10.0.0.1 port 36184 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:38.067033 sshd[3409]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:38.070137 systemd-logind[1089]: New session 13 of user core. Apr 12 18:21:38.071236 systemd[1]: Started session-13.scope. Apr 12 18:21:38.248870 sshd[3409]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:38.251933 systemd[1]: sshd@12-10.0.0.43:22-10.0.0.1:36184.service: Deactivated successfully. Apr 12 18:21:38.252688 systemd[1]: session-13.scope: Deactivated successfully. Apr 12 18:21:38.253252 systemd-logind[1089]: Session 13 logged out. Waiting for processes to exit. Apr 12 18:21:38.254450 systemd[1]: Started sshd@13-10.0.0.43:22-10.0.0.1:36192.service. Apr 12 18:21:38.255289 systemd-logind[1089]: Removed session 13. Apr 12 18:21:38.300043 sshd[3420]: Accepted publickey for core from 10.0.0.1 port 36192 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:38.301141 sshd[3420]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:38.304241 systemd-logind[1089]: New session 14 of user core. Apr 12 18:21:38.305108 systemd[1]: Started session-14.scope. Apr 12 18:21:39.012033 sshd[3420]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:39.012752 systemd[1]: Started sshd@14-10.0.0.43:22-10.0.0.1:36194.service. Apr 12 18:21:39.016283 systemd[1]: sshd@13-10.0.0.43:22-10.0.0.1:36192.service: Deactivated successfully. Apr 12 18:21:39.017165 systemd[1]: session-14.scope: Deactivated successfully. Apr 12 18:21:39.017696 systemd-logind[1089]: Session 14 logged out. Waiting for processes to exit. Apr 12 18:21:39.018440 systemd-logind[1089]: Removed session 14. Apr 12 18:21:39.059830 sshd[3439]: Accepted publickey for core from 10.0.0.1 port 36194 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:39.061206 sshd[3439]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:39.064496 systemd-logind[1089]: New session 15 of user core. Apr 12 18:21:39.065302 systemd[1]: Started session-15.scope. Apr 12 18:21:39.330864 sshd[3439]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:39.330851 systemd[1]: Started sshd@15-10.0.0.43:22-10.0.0.1:48646.service. Apr 12 18:21:39.334937 systemd[1]: sshd@14-10.0.0.43:22-10.0.0.1:36194.service: Deactivated successfully. Apr 12 18:21:39.335859 systemd[1]: session-15.scope: Deactivated successfully. Apr 12 18:21:39.337141 systemd-logind[1089]: Session 15 logged out. Waiting for processes to exit. Apr 12 18:21:39.337910 systemd-logind[1089]: Removed session 15. Apr 12 18:21:39.376429 sshd[3452]: Accepted publickey for core from 10.0.0.1 port 48646 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:39.381080 sshd[3452]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:39.385467 systemd-logind[1089]: New session 16 of user core. Apr 12 18:21:39.386056 systemd[1]: Started session-16.scope. Apr 12 18:21:39.496182 sshd[3452]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:39.500692 systemd[1]: sshd@15-10.0.0.43:22-10.0.0.1:48646.service: Deactivated successfully. Apr 12 18:21:39.500988 systemd-logind[1089]: Session 16 logged out. Waiting for processes to exit. Apr 12 18:21:39.501429 systemd[1]: session-16.scope: Deactivated successfully. Apr 12 18:21:39.502757 systemd-logind[1089]: Removed session 16. Apr 12 18:21:44.500179 systemd[1]: Started sshd@16-10.0.0.43:22-10.0.0.1:48648.service. Apr 12 18:21:44.543830 sshd[3471]: Accepted publickey for core from 10.0.0.1 port 48648 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:44.545510 sshd[3471]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:44.549641 systemd-logind[1089]: New session 17 of user core. Apr 12 18:21:44.550004 systemd[1]: Started session-17.scope. Apr 12 18:21:44.667601 sshd[3471]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:44.670084 systemd[1]: sshd@16-10.0.0.43:22-10.0.0.1:48648.service: Deactivated successfully. Apr 12 18:21:44.670808 systemd[1]: session-17.scope: Deactivated successfully. Apr 12 18:21:44.671351 systemd-logind[1089]: Session 17 logged out. Waiting for processes to exit. Apr 12 18:21:44.672094 systemd-logind[1089]: Removed session 17. Apr 12 18:21:49.672058 systemd[1]: Started sshd@17-10.0.0.43:22-10.0.0.1:60124.service. Apr 12 18:21:49.715756 sshd[3486]: Accepted publickey for core from 10.0.0.1 port 60124 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:49.716890 sshd[3486]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:49.719909 systemd-logind[1089]: New session 18 of user core. Apr 12 18:21:49.720784 systemd[1]: Started session-18.scope. Apr 12 18:21:49.830539 sshd[3486]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:49.832812 systemd[1]: sshd@17-10.0.0.43:22-10.0.0.1:60124.service: Deactivated successfully. Apr 12 18:21:49.833585 systemd[1]: session-18.scope: Deactivated successfully. Apr 12 18:21:49.834121 systemd-logind[1089]: Session 18 logged out. Waiting for processes to exit. Apr 12 18:21:49.834860 systemd-logind[1089]: Removed session 18. Apr 12 18:21:54.837110 systemd[1]: Started sshd@18-10.0.0.43:22-10.0.0.1:60130.service. Apr 12 18:21:54.880689 sshd[3500]: Accepted publickey for core from 10.0.0.1 port 60130 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:21:54.881857 sshd[3500]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:21:54.884937 systemd-logind[1089]: New session 19 of user core. Apr 12 18:21:54.885783 systemd[1]: Started session-19.scope. Apr 12 18:21:54.990940 sshd[3500]: pam_unix(sshd:session): session closed for user core Apr 12 18:21:54.993176 systemd[1]: sshd@18-10.0.0.43:22-10.0.0.1:60130.service: Deactivated successfully. Apr 12 18:21:54.993913 systemd[1]: session-19.scope: Deactivated successfully. Apr 12 18:21:54.994408 systemd-logind[1089]: Session 19 logged out. Waiting for processes to exit. Apr 12 18:21:54.995092 systemd-logind[1089]: Removed session 19. Apr 12 18:21:59.995546 systemd[1]: Started sshd@19-10.0.0.43:22-10.0.0.1:44426.service. Apr 12 18:22:00.041462 sshd[3513]: Accepted publickey for core from 10.0.0.1 port 44426 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:22:00.043013 sshd[3513]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:22:00.048620 systemd-logind[1089]: New session 20 of user core. Apr 12 18:22:00.049064 systemd[1]: Started session-20.scope. Apr 12 18:22:00.158759 sshd[3513]: pam_unix(sshd:session): session closed for user core Apr 12 18:22:00.163074 systemd[1]: Started sshd@20-10.0.0.43:22-10.0.0.1:44440.service. Apr 12 18:22:00.165291 systemd[1]: sshd@19-10.0.0.43:22-10.0.0.1:44426.service: Deactivated successfully. Apr 12 18:22:00.166099 systemd[1]: session-20.scope: Deactivated successfully. Apr 12 18:22:00.167061 systemd-logind[1089]: Session 20 logged out. Waiting for processes to exit. Apr 12 18:22:00.169712 systemd-logind[1089]: Removed session 20. Apr 12 18:22:00.211689 sshd[3525]: Accepted publickey for core from 10.0.0.1 port 44440 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:22:00.213133 sshd[3525]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:22:00.218112 systemd-logind[1089]: New session 21 of user core. Apr 12 18:22:00.218535 systemd[1]: Started session-21.scope. Apr 12 18:22:02.248759 env[1102]: time="2024-04-12T18:22:02.248710523Z" level=info msg="StopContainer for \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\" with timeout 30 (s)" Apr 12 18:22:02.249162 env[1102]: time="2024-04-12T18:22:02.249044448Z" level=info msg="Stop container \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\" with signal terminated" Apr 12 18:22:02.260424 systemd[1]: cri-containerd-9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72.scope: Deactivated successfully. Apr 12 18:22:02.280411 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72-rootfs.mount: Deactivated successfully. Apr 12 18:22:02.289226 env[1102]: time="2024-04-12T18:22:02.289181156Z" level=info msg="shim disconnected" id=9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72 Apr 12 18:22:02.289226 env[1102]: time="2024-04-12T18:22:02.289228277Z" level=warning msg="cleaning up after shim disconnected" id=9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72 namespace=k8s.io Apr 12 18:22:02.289402 env[1102]: time="2024-04-12T18:22:02.289238557Z" level=info msg="cleaning up dead shim" Apr 12 18:22:02.289996 env[1102]: time="2024-04-12T18:22:02.289957049Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 12 18:22:02.294984 env[1102]: time="2024-04-12T18:22:02.294951492Z" level=info msg="StopContainer for \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\" with timeout 2 (s)" Apr 12 18:22:02.295243 env[1102]: time="2024-04-12T18:22:02.295220497Z" level=info msg="Stop container \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\" with signal terminated" Apr 12 18:22:02.299202 env[1102]: time="2024-04-12T18:22:02.299147402Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:02Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3572 runtime=io.containerd.runc.v2\n" Apr 12 18:22:02.301161 systemd-networkd[1000]: lxc_health: Link DOWN Apr 12 18:22:02.301169 systemd-networkd[1000]: lxc_health: Lost carrier Apr 12 18:22:02.301828 env[1102]: time="2024-04-12T18:22:02.301794726Z" level=info msg="StopContainer for \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\" returns successfully" Apr 12 18:22:02.302386 env[1102]: time="2024-04-12T18:22:02.302351855Z" level=info msg="StopPodSandbox for \"a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b\"" Apr 12 18:22:02.302464 env[1102]: time="2024-04-12T18:22:02.302418217Z" level=info msg="Container to stop \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:22:02.303847 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b-shm.mount: Deactivated successfully. Apr 12 18:22:02.309621 systemd[1]: cri-containerd-a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b.scope: Deactivated successfully. Apr 12 18:22:02.329084 systemd[1]: cri-containerd-40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5.scope: Deactivated successfully. Apr 12 18:22:02.329389 systemd[1]: cri-containerd-40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5.scope: Consumed 6.526s CPU time. Apr 12 18:22:02.337242 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b-rootfs.mount: Deactivated successfully. Apr 12 18:22:02.343068 env[1102]: time="2024-04-12T18:22:02.343021452Z" level=info msg="shim disconnected" id=a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b Apr 12 18:22:02.343650 env[1102]: time="2024-04-12T18:22:02.343621262Z" level=warning msg="cleaning up after shim disconnected" id=a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b namespace=k8s.io Apr 12 18:22:02.343740 env[1102]: time="2024-04-12T18:22:02.343725424Z" level=info msg="cleaning up dead shim" Apr 12 18:22:02.348304 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5-rootfs.mount: Deactivated successfully. Apr 12 18:22:02.352587 env[1102]: time="2024-04-12T18:22:02.352544811Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:02Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3627 runtime=io.containerd.runc.v2\n" Apr 12 18:22:02.352872 env[1102]: time="2024-04-12T18:22:02.352840976Z" level=info msg="TearDown network for sandbox \"a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b\" successfully" Apr 12 18:22:02.352872 env[1102]: time="2024-04-12T18:22:02.352865576Z" level=info msg="StopPodSandbox for \"a3eb84da17ef6ae092dba2c3a88025bfb241fca60f839883d3df4c991e6fb74b\" returns successfully" Apr 12 18:22:02.353502 env[1102]: time="2024-04-12T18:22:02.353471146Z" level=info msg="shim disconnected" id=40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5 Apr 12 18:22:02.353924 env[1102]: time="2024-04-12T18:22:02.353895793Z" level=warning msg="cleaning up after shim disconnected" id=40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5 namespace=k8s.io Apr 12 18:22:02.354035 env[1102]: time="2024-04-12T18:22:02.354018435Z" level=info msg="cleaning up dead shim" Apr 12 18:22:02.364611 env[1102]: time="2024-04-12T18:22:02.364547610Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:02Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3640 runtime=io.containerd.runc.v2\ntime=\"2024-04-12T18:22:02Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" Apr 12 18:22:02.368496 env[1102]: time="2024-04-12T18:22:02.368438595Z" level=info msg="StopContainer for \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\" returns successfully" Apr 12 18:22:02.368904 env[1102]: time="2024-04-12T18:22:02.368871802Z" level=info msg="StopPodSandbox for \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\"" Apr 12 18:22:02.368960 env[1102]: time="2024-04-12T18:22:02.368942764Z" level=info msg="Container to stop \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:22:02.368994 env[1102]: time="2024-04-12T18:22:02.368960964Z" level=info msg="Container to stop \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:22:02.368994 env[1102]: time="2024-04-12T18:22:02.368972444Z" level=info msg="Container to stop \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:22:02.368994 env[1102]: time="2024-04-12T18:22:02.368986324Z" level=info msg="Container to stop \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:22:02.369138 env[1102]: time="2024-04-12T18:22:02.368996164Z" level=info msg="Container to stop \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:22:02.373970 systemd[1]: cri-containerd-9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846.scope: Deactivated successfully. Apr 12 18:22:02.374361 kubelet[1933]: I0412 18:22:02.374334 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/03731d2d-2199-4d64-a3aa-c64459041e8b-cilium-config-path\") pod \"03731d2d-2199-4d64-a3aa-c64459041e8b\" (UID: \"03731d2d-2199-4d64-a3aa-c64459041e8b\") " Apr 12 18:22:02.374648 kubelet[1933]: I0412 18:22:02.374386 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq6t7\" (UniqueName: \"kubernetes.io/projected/03731d2d-2199-4d64-a3aa-c64459041e8b-kube-api-access-nq6t7\") pod \"03731d2d-2199-4d64-a3aa-c64459041e8b\" (UID: \"03731d2d-2199-4d64-a3aa-c64459041e8b\") " Apr 12 18:22:02.377563 kubelet[1933]: I0412 18:22:02.377274 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03731d2d-2199-4d64-a3aa-c64459041e8b-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "03731d2d-2199-4d64-a3aa-c64459041e8b" (UID: "03731d2d-2199-4d64-a3aa-c64459041e8b"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Apr 12 18:22:02.378182 kubelet[1933]: I0412 18:22:02.378130 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03731d2d-2199-4d64-a3aa-c64459041e8b-kube-api-access-nq6t7" (OuterVolumeSpecName: "kube-api-access-nq6t7") pod "03731d2d-2199-4d64-a3aa-c64459041e8b" (UID: "03731d2d-2199-4d64-a3aa-c64459041e8b"). InnerVolumeSpecName "kube-api-access-nq6t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:22:02.406142 env[1102]: time="2024-04-12T18:22:02.406080422Z" level=info msg="shim disconnected" id=9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846 Apr 12 18:22:02.406142 env[1102]: time="2024-04-12T18:22:02.406136022Z" level=warning msg="cleaning up after shim disconnected" id=9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846 namespace=k8s.io Apr 12 18:22:02.406142 env[1102]: time="2024-04-12T18:22:02.406147103Z" level=info msg="cleaning up dead shim" Apr 12 18:22:02.416091 env[1102]: time="2024-04-12T18:22:02.416017427Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:02Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3672 runtime=io.containerd.runc.v2\n" Apr 12 18:22:02.416405 env[1102]: time="2024-04-12T18:22:02.416378153Z" level=info msg="TearDown network for sandbox \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" successfully" Apr 12 18:22:02.416457 env[1102]: time="2024-04-12T18:22:02.416406713Z" level=info msg="StopPodSandbox for \"9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846\" returns successfully" Apr 12 18:22:02.475643 kubelet[1933]: I0412 18:22:02.475052 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-hubble-tls\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475643 kubelet[1933]: I0412 18:22:02.475092 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-clustermesh-secrets\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475643 kubelet[1933]: I0412 18:22:02.475111 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-etc-cni-netd\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475643 kubelet[1933]: I0412 18:22:02.475127 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-hostproc\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475643 kubelet[1933]: I0412 18:22:02.475149 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm5gm\" (UniqueName: \"kubernetes.io/projected/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-kube-api-access-nm5gm\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475643 kubelet[1933]: I0412 18:22:02.475168 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-cgroup\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475923 kubelet[1933]: I0412 18:22:02.475187 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cni-path\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475923 kubelet[1933]: I0412 18:22:02.475207 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-config-path\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475923 kubelet[1933]: I0412 18:22:02.475224 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-bpf-maps\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475923 kubelet[1933]: I0412 18:22:02.475240 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-lib-modules\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475923 kubelet[1933]: I0412 18:22:02.475260 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-host-proc-sys-net\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.475923 kubelet[1933]: I0412 18:22:02.475280 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-xtables-lock\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.476088 kubelet[1933]: I0412 18:22:02.475297 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-run\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.476088 kubelet[1933]: I0412 18:22:02.475317 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-host-proc-sys-kernel\") pod \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\" (UID: \"45fb15a9-3f8e-4c17-ab63-be7da6d35da7\") " Apr 12 18:22:02.476088 kubelet[1933]: I0412 18:22:02.475347 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/03731d2d-2199-4d64-a3aa-c64459041e8b-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.476088 kubelet[1933]: I0412 18:22:02.475361 1933 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-nq6t7\" (UniqueName: \"kubernetes.io/projected/03731d2d-2199-4d64-a3aa-c64459041e8b-kube-api-access-nq6t7\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.476088 kubelet[1933]: I0412 18:22:02.475406 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.476088 kubelet[1933]: I0412 18:22:02.475667 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cni-path" (OuterVolumeSpecName: "cni-path") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.476237 kubelet[1933]: I0412 18:22:02.475776 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.476237 kubelet[1933]: I0412 18:22:02.476115 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.476237 kubelet[1933]: I0412 18:22:02.476147 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.476237 kubelet[1933]: I0412 18:22:02.476164 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.476237 kubelet[1933]: I0412 18:22:02.476180 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.476355 kubelet[1933]: I0412 18:22:02.476205 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.476355 kubelet[1933]: I0412 18:22:02.476221 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.477642 kubelet[1933]: I0412 18:22:02.477601 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Apr 12 18:22:02.477699 kubelet[1933]: I0412 18:22:02.477665 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-hostproc" (OuterVolumeSpecName: "hostproc") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:02.478179 kubelet[1933]: I0412 18:22:02.478071 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:22:02.480207 kubelet[1933]: I0412 18:22:02.480089 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Apr 12 18:22:02.480737 kubelet[1933]: I0412 18:22:02.480683 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-kube-api-access-nm5gm" (OuterVolumeSpecName: "kube-api-access-nm5gm") pod "45fb15a9-3f8e-4c17-ab63-be7da6d35da7" (UID: "45fb15a9-3f8e-4c17-ab63-be7da6d35da7"). InnerVolumeSpecName "kube-api-access-nm5gm". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:22:02.576216 kubelet[1933]: I0412 18:22:02.576088 1933 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cni-path\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576216 kubelet[1933]: I0412 18:22:02.576129 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576216 kubelet[1933]: I0412 18:22:02.576143 1933 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-bpf-maps\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576216 kubelet[1933]: I0412 18:22:02.576153 1933 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-lib-modules\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576216 kubelet[1933]: I0412 18:22:02.576163 1933 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576216 kubelet[1933]: I0412 18:22:02.576172 1933 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-xtables-lock\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576216 kubelet[1933]: I0412 18:22:02.576181 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-run\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576216 kubelet[1933]: I0412 18:22:02.576190 1933 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576524 kubelet[1933]: I0412 18:22:02.576200 1933 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-hubble-tls\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576524 kubelet[1933]: I0412 18:22:02.576209 1933 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576524 kubelet[1933]: I0412 18:22:02.576218 1933 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576524 kubelet[1933]: I0412 18:22:02.576227 1933 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-hostproc\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576524 kubelet[1933]: I0412 18:22:02.576236 1933 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-nm5gm\" (UniqueName: \"kubernetes.io/projected/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-kube-api-access-nm5gm\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.576524 kubelet[1933]: I0412 18:22:02.576246 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/45fb15a9-3f8e-4c17-ab63-be7da6d35da7-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:02.853350 kubelet[1933]: E0412 18:22:02.853248 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:02.859907 systemd[1]: Removed slice kubepods-besteffort-pod03731d2d_2199_4d64_a3aa_c64459041e8b.slice. Apr 12 18:22:02.861284 systemd[1]: Removed slice kubepods-burstable-pod45fb15a9_3f8e_4c17_ab63_be7da6d35da7.slice. Apr 12 18:22:02.861361 systemd[1]: kubepods-burstable-pod45fb15a9_3f8e_4c17_ab63_be7da6d35da7.slice: Consumed 6.719s CPU time. Apr 12 18:22:03.004942 kubelet[1933]: I0412 18:22:03.004910 1933 scope.go:117] "RemoveContainer" containerID="9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72" Apr 12 18:22:03.007779 env[1102]: time="2024-04-12T18:22:03.007703988Z" level=info msg="RemoveContainer for \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\"" Apr 12 18:22:03.012430 env[1102]: time="2024-04-12T18:22:03.012388943Z" level=info msg="RemoveContainer for \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\" returns successfully" Apr 12 18:22:03.012654 kubelet[1933]: I0412 18:22:03.012629 1933 scope.go:117] "RemoveContainer" containerID="9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72" Apr 12 18:22:03.012885 env[1102]: time="2024-04-12T18:22:03.012808430Z" level=error msg="ContainerStatus for \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\": not found" Apr 12 18:22:03.013204 kubelet[1933]: E0412 18:22:03.013181 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\": not found" containerID="9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72" Apr 12 18:22:03.013405 kubelet[1933]: I0412 18:22:03.013376 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72"} err="failed to get container status \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\": rpc error: code = NotFound desc = an error occurred when try to find container \"9c309a05dbd6474954658e2f8664e3d2a1613fba2fcf42270377ca2ffbaace72\": not found" Apr 12 18:22:03.013405 kubelet[1933]: I0412 18:22:03.013395 1933 scope.go:117] "RemoveContainer" containerID="40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5" Apr 12 18:22:03.014469 env[1102]: time="2024-04-12T18:22:03.014441096Z" level=info msg="RemoveContainer for \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\"" Apr 12 18:22:03.021612 env[1102]: time="2024-04-12T18:22:03.021557290Z" level=info msg="RemoveContainer for \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\" returns successfully" Apr 12 18:22:03.021853 kubelet[1933]: I0412 18:22:03.021828 1933 scope.go:117] "RemoveContainer" containerID="57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3" Apr 12 18:22:03.024490 env[1102]: time="2024-04-12T18:22:03.024459897Z" level=info msg="RemoveContainer for \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\"" Apr 12 18:22:03.027359 env[1102]: time="2024-04-12T18:22:03.027312822Z" level=info msg="RemoveContainer for \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\" returns successfully" Apr 12 18:22:03.027529 kubelet[1933]: I0412 18:22:03.027500 1933 scope.go:117] "RemoveContainer" containerID="5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f" Apr 12 18:22:03.028817 env[1102]: time="2024-04-12T18:22:03.028789606Z" level=info msg="RemoveContainer for \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\"" Apr 12 18:22:03.032133 env[1102]: time="2024-04-12T18:22:03.032092859Z" level=info msg="RemoveContainer for \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\" returns successfully" Apr 12 18:22:03.032264 kubelet[1933]: I0412 18:22:03.032243 1933 scope.go:117] "RemoveContainer" containerID="3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191" Apr 12 18:22:03.033067 env[1102]: time="2024-04-12T18:22:03.033038114Z" level=info msg="RemoveContainer for \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\"" Apr 12 18:22:03.035306 env[1102]: time="2024-04-12T18:22:03.035276350Z" level=info msg="RemoveContainer for \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\" returns successfully" Apr 12 18:22:03.035428 kubelet[1933]: I0412 18:22:03.035410 1933 scope.go:117] "RemoveContainer" containerID="ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f" Apr 12 18:22:03.036249 env[1102]: time="2024-04-12T18:22:03.036225405Z" level=info msg="RemoveContainer for \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\"" Apr 12 18:22:03.038338 env[1102]: time="2024-04-12T18:22:03.038304518Z" level=info msg="RemoveContainer for \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\" returns successfully" Apr 12 18:22:03.038448 kubelet[1933]: I0412 18:22:03.038430 1933 scope.go:117] "RemoveContainer" containerID="40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5" Apr 12 18:22:03.038698 env[1102]: time="2024-04-12T18:22:03.038633283Z" level=error msg="ContainerStatus for \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\": not found" Apr 12 18:22:03.038886 kubelet[1933]: E0412 18:22:03.038872 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\": not found" containerID="40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5" Apr 12 18:22:03.038958 kubelet[1933]: I0412 18:22:03.038903 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5"} err="failed to get container status \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\": rpc error: code = NotFound desc = an error occurred when try to find container \"40b1558e8456578ccb01f7a72c45806beb85065949c574e2fd857e676ff537b5\": not found" Apr 12 18:22:03.038958 kubelet[1933]: I0412 18:22:03.038914 1933 scope.go:117] "RemoveContainer" containerID="57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3" Apr 12 18:22:03.039100 env[1102]: time="2024-04-12T18:22:03.039053370Z" level=error msg="ContainerStatus for \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\": not found" Apr 12 18:22:03.039170 kubelet[1933]: E0412 18:22:03.039157 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\": not found" containerID="57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3" Apr 12 18:22:03.039205 kubelet[1933]: I0412 18:22:03.039176 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3"} err="failed to get container status \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\": rpc error: code = NotFound desc = an error occurred when try to find container \"57d4ba1c67da504001295c693db3de0870489230c69a9600bb7e27cbeac9adf3\": not found" Apr 12 18:22:03.039205 kubelet[1933]: I0412 18:22:03.039184 1933 scope.go:117] "RemoveContainer" containerID="5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f" Apr 12 18:22:03.039494 env[1102]: time="2024-04-12T18:22:03.039428376Z" level=error msg="ContainerStatus for \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\": not found" Apr 12 18:22:03.039689 kubelet[1933]: E0412 18:22:03.039674 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\": not found" containerID="5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f" Apr 12 18:22:03.039761 kubelet[1933]: I0412 18:22:03.039697 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f"} err="failed to get container status \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\": rpc error: code = NotFound desc = an error occurred when try to find container \"5d6baff110cc9dd16c780daf25928d5f353cb3ae95abe8c98c77c2e9db760f0f\": not found" Apr 12 18:22:03.039761 kubelet[1933]: I0412 18:22:03.039706 1933 scope.go:117] "RemoveContainer" containerID="3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191" Apr 12 18:22:03.039875 env[1102]: time="2024-04-12T18:22:03.039821542Z" level=error msg="ContainerStatus for \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\": not found" Apr 12 18:22:03.039942 kubelet[1933]: E0412 18:22:03.039926 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\": not found" containerID="3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191" Apr 12 18:22:03.039986 kubelet[1933]: I0412 18:22:03.039962 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191"} err="failed to get container status \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\": rpc error: code = NotFound desc = an error occurred when try to find container \"3170efc893b5df7f1ba4a8fa380499e323e2fd35060354dd9896700648a72191\": not found" Apr 12 18:22:03.039986 kubelet[1933]: I0412 18:22:03.039972 1933 scope.go:117] "RemoveContainer" containerID="ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f" Apr 12 18:22:03.040227 env[1102]: time="2024-04-12T18:22:03.040166348Z" level=error msg="ContainerStatus for \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\": not found" Apr 12 18:22:03.040391 kubelet[1933]: E0412 18:22:03.040379 1933 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\": not found" containerID="ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f" Apr 12 18:22:03.040450 kubelet[1933]: I0412 18:22:03.040398 1933 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f"} err="failed to get container status \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\": rpc error: code = NotFound desc = an error occurred when try to find container \"ede9f43dc03555c7777bc168f08567505d2637ff93850dcdf4f82651bd607a2f\": not found" Apr 12 18:22:03.257398 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846-rootfs.mount: Deactivated successfully. Apr 12 18:22:03.257501 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-9826e4d204ff18dcd6ed8cbcedcfd2062ca00613241888e01fe5a0d8f68d1846-shm.mount: Deactivated successfully. Apr 12 18:22:03.257563 systemd[1]: var-lib-kubelet-pods-03731d2d\x2d2199\x2d4d64\x2da3aa\x2dc64459041e8b-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dnq6t7.mount: Deactivated successfully. Apr 12 18:22:03.257652 systemd[1]: var-lib-kubelet-pods-45fb15a9\x2d3f8e\x2d4c17\x2dab63\x2dbe7da6d35da7-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dnm5gm.mount: Deactivated successfully. Apr 12 18:22:03.257705 systemd[1]: var-lib-kubelet-pods-45fb15a9\x2d3f8e\x2d4c17\x2dab63\x2dbe7da6d35da7-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Apr 12 18:22:03.257753 systemd[1]: var-lib-kubelet-pods-45fb15a9\x2d3f8e\x2d4c17\x2dab63\x2dbe7da6d35da7-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Apr 12 18:22:03.901419 kubelet[1933]: E0412 18:22:03.901391 1933 kubelet.go:2855] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 12 18:22:04.215852 sshd[3525]: pam_unix(sshd:session): session closed for user core Apr 12 18:22:04.219423 systemd[1]: Started sshd@21-10.0.0.43:22-10.0.0.1:44448.service. Apr 12 18:22:04.219948 systemd[1]: sshd@20-10.0.0.43:22-10.0.0.1:44440.service: Deactivated successfully. Apr 12 18:22:04.220759 systemd[1]: session-21.scope: Deactivated successfully. Apr 12 18:22:04.220952 systemd[1]: session-21.scope: Consumed 1.369s CPU time. Apr 12 18:22:04.221372 systemd-logind[1089]: Session 21 logged out. Waiting for processes to exit. Apr 12 18:22:04.222197 systemd-logind[1089]: Removed session 21. Apr 12 18:22:04.263895 sshd[3691]: Accepted publickey for core from 10.0.0.1 port 44448 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:22:04.265143 sshd[3691]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:22:04.269046 systemd-logind[1089]: New session 22 of user core. Apr 12 18:22:04.269501 systemd[1]: Started session-22.scope. Apr 12 18:22:04.855371 kubelet[1933]: I0412 18:22:04.855332 1933 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="03731d2d-2199-4d64-a3aa-c64459041e8b" path="/var/lib/kubelet/pods/03731d2d-2199-4d64-a3aa-c64459041e8b/volumes" Apr 12 18:22:04.855787 kubelet[1933]: I0412 18:22:04.855767 1933 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="45fb15a9-3f8e-4c17-ab63-be7da6d35da7" path="/var/lib/kubelet/pods/45fb15a9-3f8e-4c17-ab63-be7da6d35da7/volumes" Apr 12 18:22:05.331968 sshd[3691]: pam_unix(sshd:session): session closed for user core Apr 12 18:22:05.336270 systemd[1]: Started sshd@22-10.0.0.43:22-10.0.0.1:44464.service. Apr 12 18:22:05.341685 systemd[1]: session-22.scope: Deactivated successfully. Apr 12 18:22:05.342346 systemd-logind[1089]: Session 22 logged out. Waiting for processes to exit. Apr 12 18:22:05.342432 systemd[1]: sshd@21-10.0.0.43:22-10.0.0.1:44448.service: Deactivated successfully. Apr 12 18:22:05.344517 systemd-logind[1089]: Removed session 22. Apr 12 18:22:05.353084 kubelet[1933]: I0412 18:22:05.353047 1933 topology_manager.go:215] "Topology Admit Handler" podUID="6465158a-1aff-4d53-9fed-c2ccde6c6928" podNamespace="kube-system" podName="cilium-6q8ll" Apr 12 18:22:05.353359 kubelet[1933]: E0412 18:22:05.353107 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="45fb15a9-3f8e-4c17-ab63-be7da6d35da7" containerName="mount-cgroup" Apr 12 18:22:05.353359 kubelet[1933]: E0412 18:22:05.353118 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="45fb15a9-3f8e-4c17-ab63-be7da6d35da7" containerName="apply-sysctl-overwrites" Apr 12 18:22:05.353359 kubelet[1933]: E0412 18:22:05.353125 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="45fb15a9-3f8e-4c17-ab63-be7da6d35da7" containerName="clean-cilium-state" Apr 12 18:22:05.353359 kubelet[1933]: E0412 18:22:05.353132 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="45fb15a9-3f8e-4c17-ab63-be7da6d35da7" containerName="cilium-agent" Apr 12 18:22:05.353359 kubelet[1933]: E0412 18:22:05.353139 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="03731d2d-2199-4d64-a3aa-c64459041e8b" containerName="cilium-operator" Apr 12 18:22:05.353359 kubelet[1933]: E0412 18:22:05.353146 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="45fb15a9-3f8e-4c17-ab63-be7da6d35da7" containerName="mount-bpf-fs" Apr 12 18:22:05.353359 kubelet[1933]: I0412 18:22:05.353165 1933 memory_manager.go:346] "RemoveStaleState removing state" podUID="03731d2d-2199-4d64-a3aa-c64459041e8b" containerName="cilium-operator" Apr 12 18:22:05.353359 kubelet[1933]: I0412 18:22:05.353172 1933 memory_manager.go:346] "RemoveStaleState removing state" podUID="45fb15a9-3f8e-4c17-ab63-be7da6d35da7" containerName="cilium-agent" Apr 12 18:22:05.362748 systemd[1]: Created slice kubepods-burstable-pod6465158a_1aff_4d53_9fed_c2ccde6c6928.slice. Apr 12 18:22:05.390559 kubelet[1933]: I0412 18:22:05.390528 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzn7t\" (UniqueName: \"kubernetes.io/projected/6465158a-1aff-4d53-9fed-c2ccde6c6928-kube-api-access-tzn7t\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.390758 kubelet[1933]: I0412 18:22:05.390745 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-host-proc-sys-net\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.390833 kubelet[1933]: I0412 18:22:05.390823 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-ipsec-secrets\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.390935 kubelet[1933]: I0412 18:22:05.390924 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-hostproc\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391023 kubelet[1933]: I0412 18:22:05.391011 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-cgroup\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391097 kubelet[1933]: I0412 18:22:05.391087 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-etc-cni-netd\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391173 kubelet[1933]: I0412 18:22:05.391164 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/6465158a-1aff-4d53-9fed-c2ccde6c6928-clustermesh-secrets\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391247 kubelet[1933]: I0412 18:22:05.391237 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-config-path\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391331 kubelet[1933]: I0412 18:22:05.391320 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-lib-modules\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391409 kubelet[1933]: I0412 18:22:05.391398 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-host-proc-sys-kernel\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391503 kubelet[1933]: I0412 18:22:05.391493 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/6465158a-1aff-4d53-9fed-c2ccde6c6928-hubble-tls\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391603 kubelet[1933]: I0412 18:22:05.391591 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-run\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391679 kubelet[1933]: I0412 18:22:05.391667 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-bpf-maps\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391768 kubelet[1933]: I0412 18:22:05.391756 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cni-path\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.391842 kubelet[1933]: I0412 18:22:05.391832 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-xtables-lock\") pod \"cilium-6q8ll\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " pod="kube-system/cilium-6q8ll" Apr 12 18:22:05.397258 sshd[3703]: Accepted publickey for core from 10.0.0.1 port 44464 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:22:05.398759 sshd[3703]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:22:05.402404 systemd-logind[1089]: New session 23 of user core. Apr 12 18:22:05.402903 systemd[1]: Started session-23.scope. Apr 12 18:22:05.545953 sshd[3703]: pam_unix(sshd:session): session closed for user core Apr 12 18:22:05.549331 systemd[1]: sshd@22-10.0.0.43:22-10.0.0.1:44464.service: Deactivated successfully. Apr 12 18:22:05.551294 systemd[1]: session-23.scope: Deactivated successfully. Apr 12 18:22:05.551942 systemd-logind[1089]: Session 23 logged out. Waiting for processes to exit. Apr 12 18:22:05.554555 systemd[1]: Started sshd@23-10.0.0.43:22-10.0.0.1:44476.service. Apr 12 18:22:05.555065 systemd-logind[1089]: Removed session 23. Apr 12 18:22:05.562335 kubelet[1933]: E0412 18:22:05.562312 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:05.565182 env[1102]: time="2024-04-12T18:22:05.565139170Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6q8ll,Uid:6465158a-1aff-4d53-9fed-c2ccde6c6928,Namespace:kube-system,Attempt:0,}" Apr 12 18:22:05.583041 env[1102]: time="2024-04-12T18:22:05.582713910Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:22:05.583041 env[1102]: time="2024-04-12T18:22:05.582755111Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:22:05.583041 env[1102]: time="2024-04-12T18:22:05.582765111Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:22:05.583921 env[1102]: time="2024-04-12T18:22:05.583741485Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7 pid=3732 runtime=io.containerd.runc.v2 Apr 12 18:22:05.603915 systemd[1]: Started cri-containerd-6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7.scope. Apr 12 18:22:05.606332 sshd[3721]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:22:05.607558 sshd[3721]: Accepted publickey for core from 10.0.0.1 port 44476 ssh2: RSA SHA256:QUhY8l8fo09wOQgBdU1SXiqM8N1XKRTa5W0hOYR625c Apr 12 18:22:05.617638 systemd[1]: Started session-24.scope. Apr 12 18:22:05.617964 systemd-logind[1089]: New session 24 of user core. Apr 12 18:22:05.634414 env[1102]: time="2024-04-12T18:22:05.634374834Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-6q8ll,Uid:6465158a-1aff-4d53-9fed-c2ccde6c6928,Namespace:kube-system,Attempt:0,} returns sandbox id \"6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7\"" Apr 12 18:22:05.635412 kubelet[1933]: E0412 18:22:05.635216 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:05.637332 env[1102]: time="2024-04-12T18:22:05.637297797Z" level=info msg="CreateContainer within sandbox \"6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 12 18:22:05.647698 env[1102]: time="2024-04-12T18:22:05.647659351Z" level=info msg="CreateContainer within sandbox \"6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8\"" Apr 12 18:22:05.649481 env[1102]: time="2024-04-12T18:22:05.649420377Z" level=info msg="StartContainer for \"73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8\"" Apr 12 18:22:05.662447 systemd[1]: Started cri-containerd-73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8.scope. Apr 12 18:22:05.679783 systemd[1]: cri-containerd-73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8.scope: Deactivated successfully. Apr 12 18:22:05.699007 env[1102]: time="2024-04-12T18:22:05.698948349Z" level=info msg="shim disconnected" id=73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8 Apr 12 18:22:05.699274 env[1102]: time="2024-04-12T18:22:05.699254673Z" level=warning msg="cleaning up after shim disconnected" id=73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8 namespace=k8s.io Apr 12 18:22:05.699348 env[1102]: time="2024-04-12T18:22:05.699332155Z" level=info msg="cleaning up dead shim" Apr 12 18:22:05.708965 env[1102]: time="2024-04-12T18:22:05.708922336Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:05Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3797 runtime=io.containerd.runc.v2\ntime=\"2024-04-12T18:22:05Z\" level=warning msg=\"failed to read init pid file\" error=\"open /run/containerd/io.containerd.runtime.v2.task/k8s.io/73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8/init.pid: no such file or directory\" runtime=io.containerd.runc.v2\n" Apr 12 18:22:05.709421 env[1102]: time="2024-04-12T18:22:05.709308582Z" level=error msg="copy shim log" error="read /proc/self/fd/39: file already closed" Apr 12 18:22:05.709847 env[1102]: time="2024-04-12T18:22:05.709759309Z" level=error msg="Failed to pipe stderr of container \"73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8\"" error="reading from a closed fifo" Apr 12 18:22:05.713128 env[1102]: time="2024-04-12T18:22:05.713090838Z" level=error msg="Failed to pipe stdout of container \"73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8\"" error="reading from a closed fifo" Apr 12 18:22:05.715652 env[1102]: time="2024-04-12T18:22:05.715597835Z" level=error msg="StartContainer for \"73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8\" failed" error="failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" Apr 12 18:22:05.716028 kubelet[1933]: E0412 18:22:05.715986 1933 remote_runtime.go:343] "StartContainer from runtime service failed" err="rpc error: code = Unknown desc = failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown" containerID="73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8" Apr 12 18:22:05.718102 kubelet[1933]: E0412 18:22:05.718071 1933 kuberuntime_manager.go:1209] init container &Container{Name:mount-cgroup,Image:quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Command:[sh -ec cp /usr/bin/cilium-mount /hostbin/cilium-mount; Apr 12 18:22:05.718102 kubelet[1933]: nsenter --cgroup=/hostproc/1/ns/cgroup --mount=/hostproc/1/ns/mnt "${BIN_PATH}/cilium-mount" $CGROUP_ROOT; Apr 12 18:22:05.718102 kubelet[1933]: rm /hostbin/cilium-mount Apr 12 18:22:05.718229 kubelet[1933]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CGROUP_ROOT,Value:/run/cilium/cgroupv2,ValueFrom:nil,},EnvVar{Name:BIN_PATH,Value:/opt/cni/bin,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hostproc,ReadOnly:false,MountPath:/hostproc,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:cni-path,ReadOnly:false,MountPath:/hostbin,SubPath:,MountPropagation:nil,SubPathExpr:,},VolumeMount{Name:kube-api-access-tzn7t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[SYS_ADMIN SYS_CHROOT SYS_PTRACE],Drop:[ALL],},Privileged:nil,SELinuxOptions:&SELinuxOptions{User:,Role:,Type:spc_t,Level:s0,},RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cilium-6q8ll_kube-system(6465158a-1aff-4d53-9fed-c2ccde6c6928): RunContainerError: failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown Apr 12 18:22:05.718229 kubelet[1933]: E0412 18:22:05.718152 1933 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mount-cgroup\" with RunContainerError: \"failed to create containerd task: failed to create shim task: OCI runtime create failed: runc create failed: unable to start container process: error during container init: write /proc/self/attr/keycreate: invalid argument: unknown\"" pod="kube-system/cilium-6q8ll" podUID="6465158a-1aff-4d53-9fed-c2ccde6c6928" Apr 12 18:22:06.020537 env[1102]: time="2024-04-12T18:22:06.020497452Z" level=info msg="StopPodSandbox for \"6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7\"" Apr 12 18:22:06.021799 env[1102]: time="2024-04-12T18:22:06.021768950Z" level=info msg="Container to stop \"73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:22:06.028120 systemd[1]: cri-containerd-6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7.scope: Deactivated successfully. Apr 12 18:22:06.054538 env[1102]: time="2024-04-12T18:22:06.054484015Z" level=info msg="shim disconnected" id=6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7 Apr 12 18:22:06.054874 env[1102]: time="2024-04-12T18:22:06.054854380Z" level=warning msg="cleaning up after shim disconnected" id=6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7 namespace=k8s.io Apr 12 18:22:06.054946 env[1102]: time="2024-04-12T18:22:06.054932461Z" level=info msg="cleaning up dead shim" Apr 12 18:22:06.062424 env[1102]: time="2024-04-12T18:22:06.062386047Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:06Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3827 runtime=io.containerd.runc.v2\n" Apr 12 18:22:06.062843 env[1102]: time="2024-04-12T18:22:06.062811333Z" level=info msg="TearDown network for sandbox \"6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7\" successfully" Apr 12 18:22:06.062954 env[1102]: time="2024-04-12T18:22:06.062934335Z" level=info msg="StopPodSandbox for \"6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7\" returns successfully" Apr 12 18:22:06.098144 kubelet[1933]: I0412 18:22:06.098078 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-run\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098144 kubelet[1933]: I0412 18:22:06.098128 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-host-proc-sys-kernel\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098144 kubelet[1933]: I0412 18:22:06.098146 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cni-path\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098170 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/6465158a-1aff-4d53-9fed-c2ccde6c6928-clustermesh-secrets\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098161 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098196 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-lib-modules\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098216 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-cgroup\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098216 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cni-path" (OuterVolumeSpecName: "cni-path") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098235 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098240 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzn7t\" (UniqueName: \"kubernetes.io/projected/6465158a-1aff-4d53-9fed-c2ccde6c6928-kube-api-access-tzn7t\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098284 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-etc-cni-netd\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098311 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/6465158a-1aff-4d53-9fed-c2ccde6c6928-hubble-tls\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098339 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-bpf-maps\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098358 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-host-proc-sys-net\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098379 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-ipsec-secrets\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098395 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-hostproc\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098426 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-xtables-lock\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.098444 kubelet[1933]: I0412 18:22:06.098451 1933 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-config-path\") pod \"6465158a-1aff-4d53-9fed-c2ccde6c6928\" (UID: \"6465158a-1aff-4d53-9fed-c2ccde6c6928\") " Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098492 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-run\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098504 1933 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098514 1933 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cni-path\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098506 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098543 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098631 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098870 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098896 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-hostproc" (OuterVolumeSpecName: "hostproc") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098920 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.099376 kubelet[1933]: I0412 18:22:06.098955 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:22:06.100751 kubelet[1933]: I0412 18:22:06.100690 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Apr 12 18:22:06.101616 kubelet[1933]: I0412 18:22:06.101343 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6465158a-1aff-4d53-9fed-c2ccde6c6928-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:22:06.101616 kubelet[1933]: I0412 18:22:06.101398 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6465158a-1aff-4d53-9fed-c2ccde6c6928-kube-api-access-tzn7t" (OuterVolumeSpecName: "kube-api-access-tzn7t") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "kube-api-access-tzn7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:22:06.102679 kubelet[1933]: I0412 18:22:06.102649 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6465158a-1aff-4d53-9fed-c2ccde6c6928-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Apr 12 18:22:06.103697 kubelet[1933]: I0412 18:22:06.103667 1933 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "6465158a-1aff-4d53-9fed-c2ccde6c6928" (UID: "6465158a-1aff-4d53-9fed-c2ccde6c6928"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Apr 12 18:22:06.199068 kubelet[1933]: I0412 18:22:06.199028 1933 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/6465158a-1aff-4d53-9fed-c2ccde6c6928-clustermesh-secrets\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199068 kubelet[1933]: I0412 18:22:06.199060 1933 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-lib-modules\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199068 kubelet[1933]: I0412 18:22:06.199071 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-cgroup\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199068 kubelet[1933]: I0412 18:22:06.199082 1933 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-etc-cni-netd\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199312 kubelet[1933]: I0412 18:22:06.199094 1933 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-tzn7t\" (UniqueName: \"kubernetes.io/projected/6465158a-1aff-4d53-9fed-c2ccde6c6928-kube-api-access-tzn7t\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199312 kubelet[1933]: I0412 18:22:06.199103 1933 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-bpf-maps\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199312 kubelet[1933]: I0412 18:22:06.199115 1933 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/6465158a-1aff-4d53-9fed-c2ccde6c6928-hubble-tls\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199312 kubelet[1933]: I0412 18:22:06.199125 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-ipsec-secrets\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199312 kubelet[1933]: I0412 18:22:06.199135 1933 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-hostproc\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199312 kubelet[1933]: I0412 18:22:06.199144 1933 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-host-proc-sys-net\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199312 kubelet[1933]: I0412 18:22:06.199153 1933 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/6465158a-1aff-4d53-9fed-c2ccde6c6928-xtables-lock\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.199312 kubelet[1933]: I0412 18:22:06.199162 1933 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/6465158a-1aff-4d53-9fed-c2ccde6c6928-cilium-config-path\") on node \"localhost\" DevicePath \"\"" Apr 12 18:22:06.499735 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-6615085f149239b4ae0d0c11050d7d4d75bb7f9da9f8c2309a66793547eeabc7-shm.mount: Deactivated successfully. Apr 12 18:22:06.499848 systemd[1]: var-lib-kubelet-pods-6465158a\x2d1aff\x2d4d53\x2d9fed\x2dc2ccde6c6928-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dtzn7t.mount: Deactivated successfully. Apr 12 18:22:06.499912 systemd[1]: var-lib-kubelet-pods-6465158a\x2d1aff\x2d4d53\x2d9fed\x2dc2ccde6c6928-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Apr 12 18:22:06.499967 systemd[1]: var-lib-kubelet-pods-6465158a\x2d1aff\x2d4d53\x2d9fed\x2dc2ccde6c6928-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Apr 12 18:22:06.500037 systemd[1]: var-lib-kubelet-pods-6465158a\x2d1aff\x2d4d53\x2d9fed\x2dc2ccde6c6928-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Apr 12 18:22:06.858219 systemd[1]: Removed slice kubepods-burstable-pod6465158a_1aff_4d53_9fed_c2ccde6c6928.slice. Apr 12 18:22:07.023495 kubelet[1933]: I0412 18:22:07.023456 1933 scope.go:117] "RemoveContainer" containerID="73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8" Apr 12 18:22:07.025304 env[1102]: time="2024-04-12T18:22:07.025265712Z" level=info msg="RemoveContainer for \"73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8\"" Apr 12 18:22:07.028851 env[1102]: time="2024-04-12T18:22:07.028797960Z" level=info msg="RemoveContainer for \"73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8\" returns successfully" Apr 12 18:22:07.056951 kubelet[1933]: I0412 18:22:07.056894 1933 topology_manager.go:215] "Topology Admit Handler" podUID="eca89c7e-61be-4395-b2bf-c030e73c8f24" podNamespace="kube-system" podName="cilium-v4qkc" Apr 12 18:22:07.056951 kubelet[1933]: E0412 18:22:07.056954 1933 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="6465158a-1aff-4d53-9fed-c2ccde6c6928" containerName="mount-cgroup" Apr 12 18:22:07.057162 kubelet[1933]: I0412 18:22:07.056982 1933 memory_manager.go:346] "RemoveStaleState removing state" podUID="6465158a-1aff-4d53-9fed-c2ccde6c6928" containerName="mount-cgroup" Apr 12 18:22:07.064750 systemd[1]: Created slice kubepods-burstable-podeca89c7e_61be_4395_b2bf_c030e73c8f24.slice. Apr 12 18:22:07.106009 kubelet[1933]: I0412 18:22:07.105964 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-xtables-lock\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106172 kubelet[1933]: I0412 18:22:07.106076 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-987z5\" (UniqueName: \"kubernetes.io/projected/eca89c7e-61be-4395-b2bf-c030e73c8f24-kube-api-access-987z5\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106172 kubelet[1933]: I0412 18:22:07.106102 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-cilium-cgroup\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106172 kubelet[1933]: I0412 18:22:07.106151 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/eca89c7e-61be-4395-b2bf-c030e73c8f24-cilium-config-path\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106262 kubelet[1933]: I0412 18:22:07.106213 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-cilium-run\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106262 kubelet[1933]: I0412 18:22:07.106255 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-host-proc-sys-net\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106315 kubelet[1933]: I0412 18:22:07.106275 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-host-proc-sys-kernel\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106339 kubelet[1933]: I0412 18:22:07.106322 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-hostproc\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106366 kubelet[1933]: I0412 18:22:07.106349 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-lib-modules\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106394 kubelet[1933]: I0412 18:22:07.106371 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/eca89c7e-61be-4395-b2bf-c030e73c8f24-clustermesh-secrets\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106394 kubelet[1933]: I0412 18:22:07.106392 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-cni-path\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106445 kubelet[1933]: I0412 18:22:07.106416 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/eca89c7e-61be-4395-b2bf-c030e73c8f24-hubble-tls\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106472 kubelet[1933]: I0412 18:22:07.106452 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-bpf-maps\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106472 kubelet[1933]: I0412 18:22:07.106471 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/eca89c7e-61be-4395-b2bf-c030e73c8f24-etc-cni-netd\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.106535 kubelet[1933]: I0412 18:22:07.106502 1933 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/eca89c7e-61be-4395-b2bf-c030e73c8f24-cilium-ipsec-secrets\") pod \"cilium-v4qkc\" (UID: \"eca89c7e-61be-4395-b2bf-c030e73c8f24\") " pod="kube-system/cilium-v4qkc" Apr 12 18:22:07.366902 kubelet[1933]: E0412 18:22:07.366849 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:07.367390 env[1102]: time="2024-04-12T18:22:07.367347180Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-v4qkc,Uid:eca89c7e-61be-4395-b2bf-c030e73c8f24,Namespace:kube-system,Attempt:0,}" Apr 12 18:22:07.380696 env[1102]: time="2024-04-12T18:22:07.380639081Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:22:07.380822 env[1102]: time="2024-04-12T18:22:07.380705202Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:22:07.380822 env[1102]: time="2024-04-12T18:22:07.380732242Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:22:07.381049 env[1102]: time="2024-04-12T18:22:07.380994526Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2 pid=3855 runtime=io.containerd.runc.v2 Apr 12 18:22:07.390986 systemd[1]: Started cri-containerd-81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2.scope. Apr 12 18:22:07.421031 env[1102]: time="2024-04-12T18:22:07.420981711Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-v4qkc,Uid:eca89c7e-61be-4395-b2bf-c030e73c8f24,Namespace:kube-system,Attempt:0,} returns sandbox id \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\"" Apr 12 18:22:07.421840 kubelet[1933]: E0412 18:22:07.421817 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:07.424951 env[1102]: time="2024-04-12T18:22:07.424911685Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 12 18:22:07.434720 env[1102]: time="2024-04-12T18:22:07.434667858Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe\"" Apr 12 18:22:07.435187 env[1102]: time="2024-04-12T18:22:07.435153825Z" level=info msg="StartContainer for \"8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe\"" Apr 12 18:22:07.449502 systemd[1]: Started cri-containerd-8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe.scope. Apr 12 18:22:07.485631 env[1102]: time="2024-04-12T18:22:07.483144800Z" level=info msg="StartContainer for \"8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe\" returns successfully" Apr 12 18:22:07.495139 systemd[1]: cri-containerd-8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe.scope: Deactivated successfully. Apr 12 18:22:07.515540 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe-rootfs.mount: Deactivated successfully. Apr 12 18:22:07.524969 env[1102]: time="2024-04-12T18:22:07.524921130Z" level=info msg="shim disconnected" id=8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe Apr 12 18:22:07.524969 env[1102]: time="2024-04-12T18:22:07.524966330Z" level=warning msg="cleaning up after shim disconnected" id=8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe namespace=k8s.io Apr 12 18:22:07.524969 env[1102]: time="2024-04-12T18:22:07.524977610Z" level=info msg="cleaning up dead shim" Apr 12 18:22:07.532830 env[1102]: time="2024-04-12T18:22:07.532768317Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:07Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3937 runtime=io.containerd.runc.v2\n" Apr 12 18:22:08.026925 kubelet[1933]: E0412 18:22:08.026870 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:08.028956 env[1102]: time="2024-04-12T18:22:08.028888351Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Apr 12 18:22:08.039048 env[1102]: time="2024-04-12T18:22:08.038978803Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2\"" Apr 12 18:22:08.039498 env[1102]: time="2024-04-12T18:22:08.039459049Z" level=info msg="StartContainer for \"8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2\"" Apr 12 18:22:08.057914 systemd[1]: Started cri-containerd-8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2.scope. Apr 12 18:22:08.084841 env[1102]: time="2024-04-12T18:22:08.084797123Z" level=info msg="StartContainer for \"8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2\" returns successfully" Apr 12 18:22:08.091138 systemd[1]: cri-containerd-8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2.scope: Deactivated successfully. Apr 12 18:22:08.109439 env[1102]: time="2024-04-12T18:22:08.109390926Z" level=info msg="shim disconnected" id=8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2 Apr 12 18:22:08.109439 env[1102]: time="2024-04-12T18:22:08.109436566Z" level=warning msg="cleaning up after shim disconnected" id=8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2 namespace=k8s.io Apr 12 18:22:08.109656 env[1102]: time="2024-04-12T18:22:08.109445246Z" level=info msg="cleaning up dead shim" Apr 12 18:22:08.117523 env[1102]: time="2024-04-12T18:22:08.117484992Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:08Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3999 runtime=io.containerd.runc.v2\n" Apr 12 18:22:08.499897 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2-rootfs.mount: Deactivated successfully. Apr 12 18:22:08.804699 kubelet[1933]: W0412 18:22:08.804106 1933 manager.go:1159] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6465158a_1aff_4d53_9fed_c2ccde6c6928.slice/cri-containerd-73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8.scope WatchSource:0}: container "73266651e666b33f925ec1d5692e7db2755ecc0a46f8f6912ae0ef6dea407be8" in namespace "k8s.io": not found Apr 12 18:22:08.852970 kubelet[1933]: E0412 18:22:08.852919 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:08.855300 kubelet[1933]: I0412 18:22:08.854604 1933 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="6465158a-1aff-4d53-9fed-c2ccde6c6928" path="/var/lib/kubelet/pods/6465158a-1aff-4d53-9fed-c2ccde6c6928/volumes" Apr 12 18:22:08.902350 kubelet[1933]: E0412 18:22:08.902309 1933 kubelet.go:2855] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 12 18:22:09.034354 kubelet[1933]: E0412 18:22:09.030928 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:09.037615 env[1102]: time="2024-04-12T18:22:09.037549505Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Apr 12 18:22:09.048334 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3689450752.mount: Deactivated successfully. Apr 12 18:22:09.051611 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1829814435.mount: Deactivated successfully. Apr 12 18:22:09.053887 env[1102]: time="2024-04-12T18:22:09.053832390Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"f05a8757179e5b13203e10caa099c739cad43e3bf2533b35a5a50d3af097cc1d\"" Apr 12 18:22:09.056041 env[1102]: time="2024-04-12T18:22:09.054727041Z" level=info msg="StartContainer for \"f05a8757179e5b13203e10caa099c739cad43e3bf2533b35a5a50d3af097cc1d\"" Apr 12 18:22:09.070605 systemd[1]: Started cri-containerd-f05a8757179e5b13203e10caa099c739cad43e3bf2533b35a5a50d3af097cc1d.scope. Apr 12 18:22:09.103603 systemd[1]: cri-containerd-f05a8757179e5b13203e10caa099c739cad43e3bf2533b35a5a50d3af097cc1d.scope: Deactivated successfully. Apr 12 18:22:09.105228 env[1102]: time="2024-04-12T18:22:09.105176636Z" level=info msg="StartContainer for \"f05a8757179e5b13203e10caa099c739cad43e3bf2533b35a5a50d3af097cc1d\" returns successfully" Apr 12 18:22:09.125523 env[1102]: time="2024-04-12T18:22:09.125464451Z" level=info msg="shim disconnected" id=f05a8757179e5b13203e10caa099c739cad43e3bf2533b35a5a50d3af097cc1d Apr 12 18:22:09.125523 env[1102]: time="2024-04-12T18:22:09.125511251Z" level=warning msg="cleaning up after shim disconnected" id=f05a8757179e5b13203e10caa099c739cad43e3bf2533b35a5a50d3af097cc1d namespace=k8s.io Apr 12 18:22:09.125523 env[1102]: time="2024-04-12T18:22:09.125520931Z" level=info msg="cleaning up dead shim" Apr 12 18:22:09.132012 env[1102]: time="2024-04-12T18:22:09.131971653Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:09Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4056 runtime=io.containerd.runc.v2\n" Apr 12 18:22:10.035124 kubelet[1933]: E0412 18:22:10.035091 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:10.036900 env[1102]: time="2024-04-12T18:22:10.036859652Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Apr 12 18:22:10.047465 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount918918408.mount: Deactivated successfully. Apr 12 18:22:10.051432 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2984350036.mount: Deactivated successfully. Apr 12 18:22:10.055525 env[1102]: time="2024-04-12T18:22:10.055473116Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62\"" Apr 12 18:22:10.056709 env[1102]: time="2024-04-12T18:22:10.056657571Z" level=info msg="StartContainer for \"eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62\"" Apr 12 18:22:10.070065 systemd[1]: Started cri-containerd-eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62.scope. Apr 12 18:22:10.100489 systemd[1]: cri-containerd-eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62.scope: Deactivated successfully. Apr 12 18:22:10.101722 env[1102]: time="2024-04-12T18:22:10.101562232Z" level=warning msg="error from *cgroupsv2.Manager.EventChan" error="failed to add inotify watch for \"/sys/fs/cgroup/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeca89c7e_61be_4395_b2bf_c030e73c8f24.slice/cri-containerd-eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62.scope/memory.events\": no such file or directory" Apr 12 18:22:10.104449 env[1102]: time="2024-04-12T18:22:10.104400266Z" level=info msg="StartContainer for \"eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62\" returns successfully" Apr 12 18:22:10.125175 env[1102]: time="2024-04-12T18:22:10.124939554Z" level=info msg="shim disconnected" id=eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62 Apr 12 18:22:10.125175 env[1102]: time="2024-04-12T18:22:10.124980755Z" level=warning msg="cleaning up after shim disconnected" id=eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62 namespace=k8s.io Apr 12 18:22:10.125175 env[1102]: time="2024-04-12T18:22:10.124989915Z" level=info msg="cleaning up dead shim" Apr 12 18:22:10.131699 env[1102]: time="2024-04-12T18:22:10.131661555Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:22:10Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4111 runtime=io.containerd.runc.v2\n" Apr 12 18:22:10.500047 kubelet[1933]: I0412 18:22:10.499380 1933 setters.go:552] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-04-12T18:22:10Z","lastTransitionTime":"2024-04-12T18:22:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Apr 12 18:22:11.038433 kubelet[1933]: E0412 18:22:11.038395 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:11.042817 env[1102]: time="2024-04-12T18:22:11.042775405Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Apr 12 18:22:11.058546 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount536412298.mount: Deactivated successfully. Apr 12 18:22:11.061829 env[1102]: time="2024-04-12T18:22:11.061777505Z" level=info msg="CreateContainer within sandbox \"81de4292559c997dcd265c5a9c6758f4c18d96eb3824cfaaecee5bef6d3f1ad2\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"3b6938726808c55b96cb6ab556ba56f9371c3e496452474306c039eaa5d59256\"" Apr 12 18:22:11.062261 env[1102]: time="2024-04-12T18:22:11.062233270Z" level=info msg="StartContainer for \"3b6938726808c55b96cb6ab556ba56f9371c3e496452474306c039eaa5d59256\"" Apr 12 18:22:11.076552 systemd[1]: Started cri-containerd-3b6938726808c55b96cb6ab556ba56f9371c3e496452474306c039eaa5d59256.scope. Apr 12 18:22:11.114238 env[1102]: time="2024-04-12T18:22:11.114196231Z" level=info msg="StartContainer for \"3b6938726808c55b96cb6ab556ba56f9371c3e496452474306c039eaa5d59256\" returns successfully" Apr 12 18:22:11.340591 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106(gcm-aes-ce))) Apr 12 18:22:11.912787 kubelet[1933]: W0412 18:22:11.912749 1933 manager.go:1159] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeca89c7e_61be_4395_b2bf_c030e73c8f24.slice/cri-containerd-8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe.scope WatchSource:0}: task 8116082cfb46e08f49cf30ab29c71e1caf4ef170b22a035dd231f18256b2fdbe not found: not found Apr 12 18:22:12.042883 kubelet[1933]: E0412 18:22:12.042833 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:12.056732 kubelet[1933]: I0412 18:22:12.056694 1933 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-v4qkc" podStartSLOduration=5.056651946 podCreationTimestamp="2024-04-12 18:22:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:22:12.055960179 +0000 UTC m=+83.306595526" watchObservedRunningTime="2024-04-12 18:22:12.056651946 +0000 UTC m=+83.307287253" Apr 12 18:22:13.368166 kubelet[1933]: E0412 18:22:13.368138 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:14.013129 systemd[1]: run-containerd-runc-k8s.io-3b6938726808c55b96cb6ab556ba56f9371c3e496452474306c039eaa5d59256-runc.76Uzpd.mount: Deactivated successfully. Apr 12 18:22:14.040280 systemd-networkd[1000]: lxc_health: Link UP Apr 12 18:22:14.053268 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Apr 12 18:22:14.052889 systemd-networkd[1000]: lxc_health: Gained carrier Apr 12 18:22:15.018644 kubelet[1933]: W0412 18:22:15.018480 1933 manager.go:1159] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeca89c7e_61be_4395_b2bf_c030e73c8f24.slice/cri-containerd-8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2.scope WatchSource:0}: task 8d354557d1862b7c1958fd9f7b1df0fa592a2fa9796fefa11aae3f835785bfb2 not found: not found Apr 12 18:22:15.368843 kubelet[1933]: E0412 18:22:15.368743 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:16.049931 kubelet[1933]: E0412 18:22:16.049901 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:16.063712 systemd-networkd[1000]: lxc_health: Gained IPv6LL Apr 12 18:22:16.223393 kubelet[1933]: E0412 18:22:16.223360 1933 upgradeaware.go:439] Error proxying data from backend to client: read tcp 127.0.0.1:44408->127.0.0.1:35851: read: connection reset by peer Apr 12 18:22:17.853178 kubelet[1933]: E0412 18:22:17.853134 1933 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Apr 12 18:22:18.125955 kubelet[1933]: W0412 18:22:18.125837 1933 manager.go:1159] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeca89c7e_61be_4395_b2bf_c030e73c8f24.slice/cri-containerd-f05a8757179e5b13203e10caa099c739cad43e3bf2533b35a5a50d3af097cc1d.scope WatchSource:0}: task f05a8757179e5b13203e10caa099c739cad43e3bf2533b35a5a50d3af097cc1d not found: not found Apr 12 18:22:20.500375 sshd[3721]: pam_unix(sshd:session): session closed for user core Apr 12 18:22:20.502838 systemd[1]: sshd@23-10.0.0.43:22-10.0.0.1:44476.service: Deactivated successfully. Apr 12 18:22:20.503537 systemd[1]: session-24.scope: Deactivated successfully. Apr 12 18:22:20.504033 systemd-logind[1089]: Session 24 logged out. Waiting for processes to exit. Apr 12 18:22:20.504649 systemd-logind[1089]: Removed session 24. Apr 12 18:22:21.234091 kubelet[1933]: W0412 18:22:21.234035 1933 manager.go:1159] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeca89c7e_61be_4395_b2bf_c030e73c8f24.slice/cri-containerd-eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62.scope WatchSource:0}: task eb44ecf917073dc4c5b157496b3fdc2eb8b5562aceae4804be212c7e92f35b62 not found: not found