Jul  2 00:00:58.043183 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1]
Jul  2 00:00:58.043203 kernel: Linux version 6.6.36-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.2.1_p20240210 p14) 13.2.1 20240210, GNU ld (Gentoo 2.41 p5) 2.41.0) #1 SMP PREEMPT Mon Jul  1 22:48:46 -00 2024
Jul  2 00:00:58.043213 kernel: KASLR enabled
Jul  2 00:00:58.043219 kernel: efi: EFI v2.7 by EDK II
Jul  2 00:00:58.043225 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb900018 ACPI 2.0=0xd9710018 RNG=0xd971e498 MEMRESERVE=0xd9b43d18 
Jul  2 00:00:58.043230 kernel: random: crng init done
Jul  2 00:00:58.043237 kernel: ACPI: Early table checksum verification disabled
Jul  2 00:00:58.043243 kernel: ACPI: RSDP 0x00000000D9710018 000024 (v02 BOCHS )
Jul  2 00:00:58.043250 kernel: ACPI: XSDT 0x00000000D971FE98 000064 (v01 BOCHS  BXPC     00000001      01000013)
Jul  2 00:00:58.043257 kernel: ACPI: FACP 0x00000000D971FA98 000114 (v06 BOCHS  BXPC     00000001 BXPC 00000001)
Jul  2 00:00:58.043263 kernel: ACPI: DSDT 0x00000000D9717518 0014A2 (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jul  2 00:00:58.043269 kernel: ACPI: APIC 0x00000000D971FC18 0001A8 (v04 BOCHS  BXPC     00000001 BXPC 00000001)
Jul  2 00:00:58.043275 kernel: ACPI: PPTT 0x00000000D971D898 00009C (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jul  2 00:00:58.043281 kernel: ACPI: GTDT 0x00000000D971E818 000060 (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jul  2 00:00:58.043288 kernel: ACPI: MCFG 0x00000000D971E918 00003C (v01 BOCHS  BXPC     00000001 BXPC 00000001)
Jul  2 00:00:58.043310 kernel: ACPI: SPCR 0x00000000D971FF98 000050 (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jul  2 00:00:58.043316 kernel: ACPI: DBG2 0x00000000D971E418 000057 (v00 BOCHS  BXPC     00000001 BXPC 00000001)
Jul  2 00:00:58.043323 kernel: ACPI: IORT 0x00000000D971E718 000080 (v03 BOCHS  BXPC     00000001 BXPC 00000001)
Jul  2 00:00:58.043329 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600
Jul  2 00:00:58.043335 kernel: NUMA: Failed to initialise from firmware
Jul  2 00:00:58.043342 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff]
Jul  2 00:00:58.043348 kernel: NUMA: NODE_DATA [mem 0xdc959800-0xdc95efff]
Jul  2 00:00:58.043360 kernel: Zone ranges:
Jul  2 00:00:58.043366 kernel:   DMA      [mem 0x0000000040000000-0x00000000dcffffff]
Jul  2 00:00:58.043372 kernel:   DMA32    empty
Jul  2 00:00:58.043380 kernel:   Normal   empty
Jul  2 00:00:58.043386 kernel: Movable zone start for each node
Jul  2 00:00:58.043392 kernel: Early memory node ranges
Jul  2 00:00:58.043398 kernel:   node   0: [mem 0x0000000040000000-0x00000000d976ffff]
Jul  2 00:00:58.043405 kernel:   node   0: [mem 0x00000000d9770000-0x00000000d9b3ffff]
Jul  2 00:00:58.043411 kernel:   node   0: [mem 0x00000000d9b40000-0x00000000dce1ffff]
Jul  2 00:00:58.043417 kernel:   node   0: [mem 0x00000000dce20000-0x00000000dceaffff]
Jul  2 00:00:58.043424 kernel:   node   0: [mem 0x00000000dceb0000-0x00000000dcebffff]
Jul  2 00:00:58.043431 kernel:   node   0: [mem 0x00000000dcec0000-0x00000000dcfdffff]
Jul  2 00:00:58.043437 kernel:   node   0: [mem 0x00000000dcfe0000-0x00000000dcffffff]
Jul  2 00:00:58.043443 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff]
Jul  2 00:00:58.043450 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges
Jul  2 00:00:58.043458 kernel: psci: probing for conduit method from ACPI.
Jul  2 00:00:58.043464 kernel: psci: PSCIv1.1 detected in firmware.
Jul  2 00:00:58.043471 kernel: psci: Using standard PSCI v0.2 function IDs
Jul  2 00:00:58.043480 kernel: psci: Trusted OS migration not required
Jul  2 00:00:58.043487 kernel: psci: SMC Calling Convention v1.1
Jul  2 00:00:58.043494 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003)
Jul  2 00:00:58.043502 kernel: percpu: Embedded 31 pages/cpu s86632 r8192 d32152 u126976
Jul  2 00:00:58.043509 kernel: pcpu-alloc: s86632 r8192 d32152 u126976 alloc=31*4096
Jul  2 00:00:58.043516 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 
Jul  2 00:00:58.043523 kernel: Detected PIPT I-cache on CPU0
Jul  2 00:00:58.043529 kernel: CPU features: detected: GIC system register CPU interface
Jul  2 00:00:58.043536 kernel: CPU features: detected: Hardware dirty bit management
Jul  2 00:00:58.043543 kernel: CPU features: detected: Spectre-v4
Jul  2 00:00:58.043549 kernel: CPU features: detected: Spectre-BHB
Jul  2 00:00:58.043556 kernel: CPU features: kernel page table isolation forced ON by KASLR
Jul  2 00:00:58.043563 kernel: CPU features: detected: Kernel page table isolation (KPTI)
Jul  2 00:00:58.043571 kernel: CPU features: detected: ARM erratum 1418040
Jul  2 00:00:58.043577 kernel: alternatives: applying boot alternatives
Jul  2 00:00:58.043585 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=894d8ea3debe01ca4faf80384c3adbf31dc72d8c1b6ccdad26befbaf28696295
Jul  2 00:00:58.043592 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space.
Jul  2 00:00:58.043599 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear)
Jul  2 00:00:58.043606 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear)
Jul  2 00:00:58.043613 kernel: Fallback order for Node 0: 0 
Jul  2 00:00:58.043619 kernel: Built 1 zonelists, mobility grouping on.  Total pages: 633024
Jul  2 00:00:58.043626 kernel: Policy zone: DMA
Jul  2 00:00:58.043633 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off
Jul  2 00:00:58.043639 kernel: software IO TLB: area num 4.
Jul  2 00:00:58.043648 kernel: software IO TLB: mapped [mem 0x00000000d2e00000-0x00000000d6e00000] (64MB)
Jul  2 00:00:58.043655 kernel: Memory: 2386856K/2572288K available (10240K kernel code, 2182K rwdata, 8072K rodata, 39040K init, 897K bss, 185432K reserved, 0K cma-reserved)
Jul  2 00:00:58.043662 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1
Jul  2 00:00:58.043668 kernel: trace event string verifier disabled
Jul  2 00:00:58.043675 kernel: rcu: Preemptible hierarchical RCU implementation.
Jul  2 00:00:58.043682 kernel: rcu:         RCU event tracing is enabled.
Jul  2 00:00:58.043689 kernel: rcu:         RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4.
Jul  2 00:00:58.043696 kernel:         Trampoline variant of Tasks RCU enabled.
Jul  2 00:00:58.043703 kernel:         Tracing variant of Tasks RCU enabled.
Jul  2 00:00:58.043710 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies.
Jul  2 00:00:58.043717 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4
Jul  2 00:00:58.043724 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0
Jul  2 00:00:58.043732 kernel: GICv3: 256 SPIs implemented
Jul  2 00:00:58.043739 kernel: GICv3: 0 Extended SPIs implemented
Jul  2 00:00:58.043745 kernel: Root IRQ handler: gic_handle_irq
Jul  2 00:00:58.043752 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI
Jul  2 00:00:58.043759 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000
Jul  2 00:00:58.043765 kernel: ITS [mem 0x08080000-0x0809ffff]
Jul  2 00:00:58.043772 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400d0000 (indirect, esz 8, psz 64K, shr 1)
Jul  2 00:00:58.043779 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400e0000 (flat, esz 8, psz 64K, shr 1)
Jul  2 00:00:58.043786 kernel: GICv3: using LPI property table @0x00000000400f0000
Jul  2 00:00:58.043793 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040100000
Jul  2 00:00:58.043800 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention.
Jul  2 00:00:58.043808 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jul  2 00:00:58.043815 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt).
Jul  2 00:00:58.043821 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns
Jul  2 00:00:58.043828 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns
Jul  2 00:00:58.043835 kernel: arm-pv: using stolen time PV
Jul  2 00:00:58.043842 kernel: Console: colour dummy device 80x25
Jul  2 00:00:58.043849 kernel: ACPI: Core revision 20230628
Jul  2 00:00:58.043856 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000)
Jul  2 00:00:58.043863 kernel: pid_max: default: 32768 minimum: 301
Jul  2 00:00:58.043870 kernel: LSM: initializing lsm=lockdown,capability,selinux,integrity
Jul  2 00:00:58.043879 kernel: SELinux:  Initializing.
Jul  2 00:00:58.043885 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear)
Jul  2 00:00:58.043898 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear)
Jul  2 00:00:58.043906 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1.
Jul  2 00:00:58.043913 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1.
Jul  2 00:00:58.043920 kernel: rcu: Hierarchical SRCU implementation.
Jul  2 00:00:58.043927 kernel: rcu:         Max phase no-delay instances is 400.
Jul  2 00:00:58.043934 kernel: Platform MSI: ITS@0x8080000 domain created
Jul  2 00:00:58.043941 kernel: PCI/MSI: ITS@0x8080000 domain created
Jul  2 00:00:58.043964 kernel: Remapping and enabling EFI services.
Jul  2 00:00:58.043971 kernel: smp: Bringing up secondary CPUs ...
Jul  2 00:00:58.043978 kernel: Detected PIPT I-cache on CPU1
Jul  2 00:00:58.043985 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000
Jul  2 00:00:58.043992 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040110000
Jul  2 00:00:58.043999 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jul  2 00:00:58.044006 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1]
Jul  2 00:00:58.044013 kernel: Detected PIPT I-cache on CPU2
Jul  2 00:00:58.044019 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000
Jul  2 00:00:58.044027 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040120000
Jul  2 00:00:58.044035 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jul  2 00:00:58.044042 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1]
Jul  2 00:00:58.044054 kernel: Detected PIPT I-cache on CPU3
Jul  2 00:00:58.044063 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000
Jul  2 00:00:58.044070 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040130000
Jul  2 00:00:58.044078 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jul  2 00:00:58.044084 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1]
Jul  2 00:00:58.044092 kernel: smp: Brought up 1 node, 4 CPUs
Jul  2 00:00:58.044099 kernel: SMP: Total of 4 processors activated.
Jul  2 00:00:58.044108 kernel: CPU features: detected: 32-bit EL0 Support
Jul  2 00:00:58.044115 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence
Jul  2 00:00:58.044122 kernel: CPU features: detected: Common not Private translations
Jul  2 00:00:58.044130 kernel: CPU features: detected: CRC32 instructions
Jul  2 00:00:58.044137 kernel: CPU features: detected: Enhanced Virtualization Traps
Jul  2 00:00:58.044144 kernel: CPU features: detected: RCpc load-acquire (LDAPR)
Jul  2 00:00:58.044152 kernel: CPU features: detected: LSE atomic instructions
Jul  2 00:00:58.044159 kernel: CPU features: detected: Privileged Access Never
Jul  2 00:00:58.044168 kernel: CPU features: detected: RAS Extension Support
Jul  2 00:00:58.044175 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS)
Jul  2 00:00:58.044182 kernel: CPU: All CPU(s) started at EL1
Jul  2 00:00:58.044190 kernel: alternatives: applying system-wide alternatives
Jul  2 00:00:58.044197 kernel: devtmpfs: initialized
Jul  2 00:00:58.044205 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns
Jul  2 00:00:58.044212 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear)
Jul  2 00:00:58.044220 kernel: pinctrl core: initialized pinctrl subsystem
Jul  2 00:00:58.044227 kernel: SMBIOS 3.0.0 present.
Jul  2 00:00:58.044236 kernel: DMI: QEMU KVM Virtual Machine, BIOS edk2-20230524-3.fc38 05/24/2023
Jul  2 00:00:58.044244 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family
Jul  2 00:00:58.044251 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations
Jul  2 00:00:58.044259 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations
Jul  2 00:00:58.044266 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations
Jul  2 00:00:58.044274 kernel: audit: initializing netlink subsys (disabled)
Jul  2 00:00:58.044281 kernel: audit: type=2000 audit(0.024:1): state=initialized audit_enabled=0 res=1
Jul  2 00:00:58.044289 kernel: thermal_sys: Registered thermal governor 'step_wise'
Jul  2 00:00:58.044397 kernel: cpuidle: using governor menu
Jul  2 00:00:58.044407 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers.
Jul  2 00:00:58.044415 kernel: ASID allocator initialised with 32768 entries
Jul  2 00:00:58.044422 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5
Jul  2 00:00:58.044430 kernel: Serial: AMBA PL011 UART driver
Jul  2 00:00:58.044437 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL
Jul  2 00:00:58.044445 kernel: Modules: 0 pages in range for non-PLT usage
Jul  2 00:00:58.044452 kernel: Modules: 509120 pages in range for PLT usage
Jul  2 00:00:58.044460 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages
Jul  2 00:00:58.044467 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page
Jul  2 00:00:58.044476 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages
Jul  2 00:00:58.044484 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page
Jul  2 00:00:58.044491 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages
Jul  2 00:00:58.044499 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page
Jul  2 00:00:58.044506 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages
Jul  2 00:00:58.044514 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page
Jul  2 00:00:58.044521 kernel: ACPI: Added _OSI(Module Device)
Jul  2 00:00:58.044528 kernel: ACPI: Added _OSI(Processor Device)
Jul  2 00:00:58.044536 kernel: ACPI: Added _OSI(3.0 _SCP Extensions)
Jul  2 00:00:58.044545 kernel: ACPI: Added _OSI(Processor Aggregator Device)
Jul  2 00:00:58.044552 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded
Jul  2 00:00:58.044560 kernel: ACPI: Interpreter enabled
Jul  2 00:00:58.044567 kernel: ACPI: Using GIC for interrupt routing
Jul  2 00:00:58.044574 kernel: ACPI: MCFG table detected, 1 entries
Jul  2 00:00:58.044582 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA
Jul  2 00:00:58.044589 kernel: printk: console [ttyAMA0] enabled
Jul  2 00:00:58.044597 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff])
Jul  2 00:00:58.044746 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3]
Jul  2 00:00:58.044843 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR]
Jul  2 00:00:58.044932 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability]
Jul  2 00:00:58.044997 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00
Jul  2 00:00:58.045059 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff]
Jul  2 00:00:58.045069 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io  0x0000-0xffff window]
Jul  2 00:00:58.045077 kernel: PCI host bridge to bus 0000:00
Jul  2 00:00:58.045150 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window]
Jul  2 00:00:58.045218 kernel: pci_bus 0000:00: root bus resource [io  0x0000-0xffff window]
Jul  2 00:00:58.045278 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window]
Jul  2 00:00:58.045354 kernel: pci_bus 0000:00: root bus resource [bus 00-ff]
Jul  2 00:00:58.045436 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000
Jul  2 00:00:58.045516 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00
Jul  2 00:00:58.045587 kernel: pci 0000:00:01.0: reg 0x10: [io  0x0000-0x001f]
Jul  2 00:00:58.045661 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff]
Jul  2 00:00:58.045731 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref]
Jul  2 00:00:58.045800 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref]
Jul  2 00:00:58.045870 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff]
Jul  2 00:00:58.045947 kernel: pci 0000:00:01.0: BAR 0: assigned [io  0x1000-0x101f]
Jul  2 00:00:58.046010 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window]
Jul  2 00:00:58.046072 kernel: pci_bus 0000:00: resource 5 [io  0x0000-0xffff window]
Jul  2 00:00:58.046145 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window]
Jul  2 00:00:58.046157 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35
Jul  2 00:00:58.046165 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36
Jul  2 00:00:58.046172 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37
Jul  2 00:00:58.046180 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38
Jul  2 00:00:58.046187 kernel: iommu: Default domain type: Translated
Jul  2 00:00:58.046195 kernel: iommu: DMA domain TLB invalidation policy: strict mode
Jul  2 00:00:58.046202 kernel: efivars: Registered efivars operations
Jul  2 00:00:58.046210 kernel: vgaarb: loaded
Jul  2 00:00:58.046220 kernel: clocksource: Switched to clocksource arch_sys_counter
Jul  2 00:00:58.046228 kernel: VFS: Disk quotas dquot_6.6.0
Jul  2 00:00:58.046235 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes)
Jul  2 00:00:58.046243 kernel: pnp: PnP ACPI init
Jul  2 00:00:58.046443 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved
Jul  2 00:00:58.046459 kernel: pnp: PnP ACPI: found 1 devices
Jul  2 00:00:58.046467 kernel: NET: Registered PF_INET protocol family
Jul  2 00:00:58.046475 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear)
Jul  2 00:00:58.046487 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear)
Jul  2 00:00:58.046495 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear)
Jul  2 00:00:58.046503 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear)
Jul  2 00:00:58.046510 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear)
Jul  2 00:00:58.046518 kernel: TCP: Hash tables configured (established 32768 bind 32768)
Jul  2 00:00:58.046526 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear)
Jul  2 00:00:58.046534 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear)
Jul  2 00:00:58.046541 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family
Jul  2 00:00:58.046549 kernel: PCI: CLS 0 bytes, default 64
Jul  2 00:00:58.046558 kernel: kvm [1]: HYP mode not available
Jul  2 00:00:58.046565 kernel: Initialise system trusted keyrings
Jul  2 00:00:58.046573 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0
Jul  2 00:00:58.046581 kernel: Key type asymmetric registered
Jul  2 00:00:58.046588 kernel: Asymmetric key parser 'x509' registered
Jul  2 00:00:58.046596 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250)
Jul  2 00:00:58.046604 kernel: io scheduler mq-deadline registered
Jul  2 00:00:58.046611 kernel: io scheduler kyber registered
Jul  2 00:00:58.046619 kernel: io scheduler bfq registered
Jul  2 00:00:58.046628 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0
Jul  2 00:00:58.046636 kernel: ACPI: button: Power Button [PWRB]
Jul  2 00:00:58.046703 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36
Jul  2 00:00:58.046802 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007)
Jul  2 00:00:58.046814 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled
Jul  2 00:00:58.046821 kernel: thunder_xcv, ver 1.0
Jul  2 00:00:58.046829 kernel: thunder_bgx, ver 1.0
Jul  2 00:00:58.046837 kernel: nicpf, ver 1.0
Jul  2 00:00:58.046844 kernel: nicvf, ver 1.0
Jul  2 00:00:58.046959 kernel: rtc-efi rtc-efi.0: registered as rtc0
Jul  2 00:00:58.047027 kernel: rtc-efi rtc-efi.0: setting system clock to 2024-07-02T00:00:57 UTC (1719878457)
Jul  2 00:00:58.047037 kernel: hid: raw HID events driver (C) Jiri Kosina
Jul  2 00:00:58.047045 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available
Jul  2 00:00:58.047053 kernel: watchdog: Delayed init of the lockup detector failed: -19
Jul  2 00:00:58.047060 kernel: watchdog: Hard watchdog permanently disabled
Jul  2 00:00:58.047068 kernel: NET: Registered PF_INET6 protocol family
Jul  2 00:00:58.047076 kernel: Segment Routing with IPv6
Jul  2 00:00:58.047086 kernel: In-situ OAM (IOAM) with IPv6
Jul  2 00:00:58.047093 kernel: NET: Registered PF_PACKET protocol family
Jul  2 00:00:58.047101 kernel: Key type dns_resolver registered
Jul  2 00:00:58.047109 kernel: registered taskstats version 1
Jul  2 00:00:58.047116 kernel: Loading compiled-in X.509 certificates
Jul  2 00:00:58.047124 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.36-flatcar: 60660d9c77cbf90f55b5b3c47931cf5941193eaf'
Jul  2 00:00:58.047132 kernel: Key type .fscrypt registered
Jul  2 00:00:58.047140 kernel: Key type fscrypt-provisioning registered
Jul  2 00:00:58.047148 kernel: ima: No TPM chip found, activating TPM-bypass!
Jul  2 00:00:58.047157 kernel: ima: Allocated hash algorithm: sha1
Jul  2 00:00:58.047165 kernel: ima: No architecture policies found
Jul  2 00:00:58.047173 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng)
Jul  2 00:00:58.047194 kernel: clk: Disabling unused clocks
Jul  2 00:00:58.047202 kernel: Freeing unused kernel memory: 39040K
Jul  2 00:00:58.047210 kernel: Run /init as init process
Jul  2 00:00:58.047217 kernel:   with arguments:
Jul  2 00:00:58.047225 kernel:     /init
Jul  2 00:00:58.047232 kernel:   with environment:
Jul  2 00:00:58.047242 kernel:     HOME=/
Jul  2 00:00:58.047249 kernel:     TERM=linux
Jul  2 00:00:58.047258 kernel:     BOOT_IMAGE=/flatcar/vmlinuz-a
Jul  2 00:00:58.047267 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified)
Jul  2 00:00:58.047277 systemd[1]: Detected virtualization kvm.
Jul  2 00:00:58.047285 systemd[1]: Detected architecture arm64.
Jul  2 00:00:58.047306 systemd[1]: Running in initrd.
Jul  2 00:00:58.047314 systemd[1]: No hostname configured, using default hostname.
Jul  2 00:00:58.047324 systemd[1]: Hostname set to <localhost>.
Jul  2 00:00:58.047332 systemd[1]: Initializing machine ID from VM UUID.
Jul  2 00:00:58.047340 systemd[1]: Queued start job for default target initrd.target.
Jul  2 00:00:58.047349 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Jul  2 00:00:58.047358 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Jul  2 00:00:58.047367 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM...
Jul  2 00:00:58.047375 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM...
Jul  2 00:00:58.047383 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT...
Jul  2 00:00:58.047393 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A...
Jul  2 00:00:58.047403 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132...
Jul  2 00:00:58.047411 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr...
Jul  2 00:00:58.047420 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Jul  2 00:00:58.047428 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes.
Jul  2 00:00:58.047436 systemd[1]: Reached target paths.target - Path Units.
Jul  2 00:00:58.047446 systemd[1]: Reached target slices.target - Slice Units.
Jul  2 00:00:58.047454 systemd[1]: Reached target swap.target - Swaps.
Jul  2 00:00:58.047462 systemd[1]: Reached target timers.target - Timer Units.
Jul  2 00:00:58.047470 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket.
Jul  2 00:00:58.047479 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket.
Jul  2 00:00:58.047487 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log).
Jul  2 00:00:58.047496 systemd[1]: Listening on systemd-journald.socket - Journal Socket.
Jul  2 00:00:58.047504 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket.
Jul  2 00:00:58.047513 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket.
Jul  2 00:00:58.047522 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket.
Jul  2 00:00:58.047531 systemd[1]: Reached target sockets.target - Socket Units.
Jul  2 00:00:58.047539 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup...
Jul  2 00:00:58.047548 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes...
Jul  2 00:00:58.047556 systemd[1]: Finished network-cleanup.service - Network Cleanup.
Jul  2 00:00:58.047564 systemd[1]: Starting systemd-fsck-usr.service...
Jul  2 00:00:58.047573 systemd[1]: Starting systemd-journald.service - Journal Service...
Jul  2 00:00:58.047581 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules...
Jul  2 00:00:58.047590 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Jul  2 00:00:58.047599 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup.
Jul  2 00:00:58.047612 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes.
Jul  2 00:00:58.047621 systemd[1]: Finished systemd-fsck-usr.service.
Jul  2 00:00:58.047630 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully...
Jul  2 00:00:58.047661 systemd-journald[238]: Collecting audit messages is disabled.
Jul  2 00:00:58.047685 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully.
Jul  2 00:00:58.047696 systemd-journald[238]: Journal started
Jul  2 00:00:58.047722 systemd-journald[238]: Runtime Journal (/run/log/journal/9b34d472248a4422b5ae1e477ecfad8c) is 5.9M, max 47.3M, 41.4M free.
Jul  2 00:00:58.052435 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this.
Jul  2 00:00:58.037828 systemd-modules-load[240]: Inserted module 'overlay'
Jul  2 00:00:58.054801 kernel: Bridge firewalling registered
Jul  2 00:00:58.054821 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev...
Jul  2 00:00:58.053825 systemd-modules-load[240]: Inserted module 'br_netfilter'
Jul  2 00:00:58.057363 systemd[1]: Started systemd-journald.service - Journal Service.
Jul  2 00:00:58.059350 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules.
Jul  2 00:00:58.060500 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Jul  2 00:00:58.063122 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Jul  2 00:00:58.080526 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Jul  2 00:00:58.082220 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Jul  2 00:00:58.086513 systemd[1]: Starting systemd-tmpfiles-setup.service - Create Volatile Files and Directories...
Jul  2 00:00:58.097820 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Jul  2 00:00:58.098993 systemd[1]: Finished systemd-tmpfiles-setup.service - Create Volatile Files and Directories.
Jul  2 00:00:58.101957 systemd[1]: Starting systemd-resolved.service - Network Name Resolution...
Jul  2 00:00:58.104559 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jul  2 00:00:58.114499 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook...
Jul  2 00:00:58.124765 dracut-cmdline[279]: dracut-dracut-053
Jul  2 00:00:58.127389 dracut-cmdline[279]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=894d8ea3debe01ca4faf80384c3adbf31dc72d8c1b6ccdad26befbaf28696295
Jul  2 00:00:58.143154 systemd-resolved[277]: Positive Trust Anchors:
Jul  2 00:00:58.143173 systemd-resolved[277]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Jul  2 00:00:58.143203 systemd-resolved[277]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa corp home internal intranet lan local private test
Jul  2 00:00:58.148507 systemd-resolved[277]: Defaulting to hostname 'linux'.
Jul  2 00:00:58.149976 systemd[1]: Started systemd-resolved.service - Network Name Resolution.
Jul  2 00:00:58.151199 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups.
Jul  2 00:00:58.208335 kernel: SCSI subsystem initialized
Jul  2 00:00:58.215378 kernel: Loading iSCSI transport class v2.0-870.
Jul  2 00:00:58.224454 kernel: iscsi: registered transport (tcp)
Jul  2 00:00:58.240340 kernel: iscsi: registered transport (qla4xxx)
Jul  2 00:00:58.240400 kernel: QLogic iSCSI HBA Driver
Jul  2 00:00:58.287647 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook.
Jul  2 00:00:58.297446 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook...
Jul  2 00:00:58.315403 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log.
Jul  2 00:00:58.315496 kernel: device-mapper: uevent: version 1.0.3
Jul  2 00:00:58.315512 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com
Jul  2 00:00:58.366342 kernel: raid6: neonx8   gen() 13716 MB/s
Jul  2 00:00:58.383323 kernel: raid6: neonx4   gen() 14288 MB/s
Jul  2 00:00:58.400314 kernel: raid6: neonx2   gen() 12109 MB/s
Jul  2 00:00:58.417317 kernel: raid6: neonx1   gen()  9736 MB/s
Jul  2 00:00:58.434314 kernel: raid6: int64x8  gen()  6695 MB/s
Jul  2 00:00:58.451323 kernel: raid6: int64x4  gen()  7253 MB/s
Jul  2 00:00:58.468310 kernel: raid6: int64x2  gen()  6114 MB/s
Jul  2 00:00:58.485312 kernel: raid6: int64x1  gen()  5043 MB/s
Jul  2 00:00:58.485329 kernel: raid6: using algorithm neonx4 gen() 14288 MB/s
Jul  2 00:00:58.502316 kernel: raid6: .... xor() 12369 MB/s, rmw enabled
Jul  2 00:00:58.502330 kernel: raid6: using neon recovery algorithm
Jul  2 00:00:58.507314 kernel: xor: measuring software checksum speed
Jul  2 00:00:58.508310 kernel:    8regs           : 19854 MB/sec
Jul  2 00:00:58.509730 kernel:    32regs          : 19673 MB/sec
Jul  2 00:00:58.509742 kernel:    arm64_neon      : 27215 MB/sec
Jul  2 00:00:58.509751 kernel: xor: using function: arm64_neon (27215 MB/sec)
Jul  2 00:00:58.570347 kernel: Btrfs loaded, zoned=no, fsverity=no
Jul  2 00:00:58.582372 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook.
Jul  2 00:00:58.595537 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files...
Jul  2 00:00:58.609543 systemd-udevd[463]: Using default interface naming scheme 'v255'.
Jul  2 00:00:58.612813 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files.
Jul  2 00:00:58.623520 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook...
Jul  2 00:00:58.636944 dracut-pre-trigger[469]: rd.md=0: removing MD RAID activation
Jul  2 00:00:58.668433 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook.
Jul  2 00:00:58.683554 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices...
Jul  2 00:00:58.729593 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices.
Jul  2 00:00:58.737527 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook...
Jul  2 00:00:58.751480 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook.
Jul  2 00:00:58.754768 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems.
Jul  2 00:00:58.756614 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes.
Jul  2 00:00:58.758747 systemd[1]: Reached target remote-fs.target - Remote File Systems.
Jul  2 00:00:58.771110 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook...
Jul  2 00:00:58.782354 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues
Jul  2 00:00:58.797367 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB)
Jul  2 00:00:58.797490 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk.
Jul  2 00:00:58.797502 kernel: GPT:9289727 != 19775487
Jul  2 00:00:58.797512 kernel: GPT:Alternate GPT header not at the end of the disk.
Jul  2 00:00:58.797522 kernel: GPT:9289727 != 19775487
Jul  2 00:00:58.797531 kernel: GPT: Use GNU Parted to correct GPT errors.
Jul  2 00:00:58.797541 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Jul  2 00:00:58.786069 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook.
Jul  2 00:00:58.795440 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully.
Jul  2 00:00:58.795562 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jul  2 00:00:58.799230 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Jul  2 00:00:58.800204 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully.
Jul  2 00:00:58.800839 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup.
Jul  2 00:00:58.802028 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup...
Jul  2 00:00:58.813612 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Jul  2 00:00:58.823558 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM.
Jul  2 00:00:58.828745 kernel: BTRFS: device fsid 2e7aff7f-b51e-4094-8f16-54690a62fb17 devid 1 transid 38 /dev/vda3 scanned by (udev-worker) (516)
Jul  2 00:00:58.830908 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Jul  2 00:00:58.833937 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (511)
Jul  2 00:00:58.840368 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT.
Jul  2 00:00:58.848005 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132.
Jul  2 00:00:58.849119 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A.
Jul  2 00:00:58.855733 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM.
Jul  2 00:00:58.868499 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary...
Jul  2 00:00:58.870218 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Jul  2 00:00:58.879120 disk-uuid[551]: Primary Header is updated.
Jul  2 00:00:58.879120 disk-uuid[551]: Secondary Entries is updated.
Jul  2 00:00:58.879120 disk-uuid[551]: Secondary Header is updated.
Jul  2 00:00:58.884803 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Jul  2 00:00:58.903079 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jul  2 00:00:59.911379 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Jul  2 00:00:59.912614 disk-uuid[553]: The operation has completed successfully.
Jul  2 00:00:59.936456 systemd[1]: disk-uuid.service: Deactivated successfully.
Jul  2 00:00:59.936572 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary.
Jul  2 00:00:59.962508 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr...
Jul  2 00:00:59.965616 sh[570]: Success
Jul  2 00:00:59.984324 kernel: device-mapper: verity: sha256 using implementation "sha256-ce"
Jul  2 00:01:00.070898 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr.
Jul  2 00:01:00.072985 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr...
Jul  2 00:01:00.074082 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr.
Jul  2 00:01:00.089636 kernel: BTRFS info (device dm-0): first mount of filesystem 2e7aff7f-b51e-4094-8f16-54690a62fb17
Jul  2 00:01:00.089679 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm
Jul  2 00:01:00.089690 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead
Jul  2 00:01:00.091814 kernel: BTRFS info (device dm-0): disabling log replay at mount time
Jul  2 00:01:00.091834 kernel: BTRFS info (device dm-0): using free space tree
Jul  2 00:01:00.095388 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr.
Jul  2 00:01:00.096674 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met.
Jul  2 00:01:00.105502 systemd[1]: Starting ignition-setup.service - Ignition (setup)...
Jul  2 00:01:00.110169 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline...
Jul  2 00:01:00.117987 kernel: BTRFS info (device vda6): first mount of filesystem f333e8f9-4cd9-418a-86af-1531564c69c1
Jul  2 00:01:00.118037 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm
Jul  2 00:01:00.118047 kernel: BTRFS info (device vda6): using free space tree
Jul  2 00:01:00.120304 kernel: BTRFS info (device vda6): auto enabling async discard
Jul  2 00:01:00.129220 systemd[1]: mnt-oem.mount: Deactivated successfully.
Jul  2 00:01:00.131191 kernel: BTRFS info (device vda6): last unmount of filesystem f333e8f9-4cd9-418a-86af-1531564c69c1
Jul  2 00:01:00.138168 systemd[1]: Finished ignition-setup.service - Ignition (setup).
Jul  2 00:01:00.145479 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)...
Jul  2 00:01:00.228810 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline.
Jul  2 00:01:00.246587 systemd[1]: Starting systemd-networkd.service - Network Configuration...
Jul  2 00:01:00.283786 systemd-networkd[755]: lo: Link UP
Jul  2 00:01:00.283798 systemd-networkd[755]: lo: Gained carrier
Jul  2 00:01:00.284524 systemd-networkd[755]: Enumeration completed
Jul  2 00:01:00.285435 systemd-networkd[755]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jul  2 00:01:00.285438 systemd-networkd[755]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network.
Jul  2 00:01:00.286534 systemd[1]: Started systemd-networkd.service - Network Configuration.
Jul  2 00:01:00.286580 systemd-networkd[755]: eth0: Link UP
Jul  2 00:01:00.286584 systemd-networkd[755]: eth0: Gained carrier
Jul  2 00:01:00.286592 systemd-networkd[755]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jul  2 00:01:00.288360 systemd[1]: Reached target network.target - Network.
Jul  2 00:01:00.320187 ignition[662]: Ignition 2.18.0
Jul  2 00:01:00.320210 ignition[662]: Stage: fetch-offline
Jul  2 00:01:00.320249 ignition[662]: no configs at "/usr/lib/ignition/base.d"
Jul  2 00:01:00.320258 ignition[662]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jul  2 00:01:00.320362 ignition[662]: parsed url from cmdline: ""
Jul  2 00:01:00.324380 systemd-networkd[755]: eth0: DHCPv4 address 10.0.0.39/16, gateway 10.0.0.1 acquired from 10.0.0.1
Jul  2 00:01:00.320365 ignition[662]: no config URL provided
Jul  2 00:01:00.320371 ignition[662]: reading system config file "/usr/lib/ignition/user.ign"
Jul  2 00:01:00.320379 ignition[662]: no config at "/usr/lib/ignition/user.ign"
Jul  2 00:01:00.320407 ignition[662]: op(1): [started]  loading QEMU firmware config module
Jul  2 00:01:00.320411 ignition[662]: op(1): executing: "modprobe" "qemu_fw_cfg"
Jul  2 00:01:00.335249 ignition[662]: op(1): [finished] loading QEMU firmware config module
Jul  2 00:01:00.373808 ignition[662]: parsing config with SHA512: 6c82501c730f966e1b8e648b3d3af89a21bbac0ff933f8dbd1afd2d82c18b3024b03edcab1330e24ca93a86faea276c984ae6489a6a9775416af74d0ac17c2e8
Jul  2 00:01:00.378561 unknown[662]: fetched base config from "system"
Jul  2 00:01:00.379036 ignition[662]: fetch-offline: fetch-offline passed
Jul  2 00:01:00.378569 unknown[662]: fetched user config from "qemu"
Jul  2 00:01:00.379092 ignition[662]: Ignition finished successfully
Jul  2 00:01:00.388466 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline).
Jul  2 00:01:00.389932 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json).
Jul  2 00:01:00.404821 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)...
Jul  2 00:01:00.417399 ignition[768]: Ignition 2.18.0
Jul  2 00:01:00.417411 ignition[768]: Stage: kargs
Jul  2 00:01:00.417595 ignition[768]: no configs at "/usr/lib/ignition/base.d"
Jul  2 00:01:00.417605 ignition[768]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jul  2 00:01:00.418584 ignition[768]: kargs: kargs passed
Jul  2 00:01:00.421991 systemd[1]: Finished ignition-kargs.service - Ignition (kargs).
Jul  2 00:01:00.418634 ignition[768]: Ignition finished successfully
Jul  2 00:01:00.438561 systemd[1]: Starting ignition-disks.service - Ignition (disks)...
Jul  2 00:01:00.450143 ignition[778]: Ignition 2.18.0
Jul  2 00:01:00.450154 ignition[778]: Stage: disks
Jul  2 00:01:00.450377 ignition[778]: no configs at "/usr/lib/ignition/base.d"
Jul  2 00:01:00.450388 ignition[778]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jul  2 00:01:00.452970 systemd[1]: Finished ignition-disks.service - Ignition (disks).
Jul  2 00:01:00.451286 ignition[778]: disks: disks passed
Jul  2 00:01:00.454942 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device.
Jul  2 00:01:00.451349 ignition[778]: Ignition finished successfully
Jul  2 00:01:00.456614 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems.
Jul  2 00:01:00.458415 systemd[1]: Reached target local-fs.target - Local File Systems.
Jul  2 00:01:00.460445 systemd[1]: Reached target sysinit.target - System Initialization.
Jul  2 00:01:00.462236 systemd[1]: Reached target basic.target - Basic System.
Jul  2 00:01:00.478485 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT...
Jul  2 00:01:00.491697 systemd-resolved[277]: Detected conflict on linux IN A 10.0.0.39
Jul  2 00:01:00.491715 systemd-resolved[277]: Hostname conflict, changing published hostname from 'linux' to 'linux8'.
Jul  2 00:01:00.494593 systemd-fsck[788]: ROOT: clean, 14/553520 files, 52654/553472 blocks
Jul  2 00:01:00.503064 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT.
Jul  2 00:01:00.509429 systemd[1]: Mounting sysroot.mount - /sysroot...
Jul  2 00:01:00.561315 kernel: EXT4-fs (vda9): mounted filesystem 95038baa-e9f1-4207-86a5-38a4ce3cff7d r/w with ordered data mode. Quota mode: none.
Jul  2 00:01:00.561621 systemd[1]: Mounted sysroot.mount - /sysroot.
Jul  2 00:01:00.563571 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System.
Jul  2 00:01:00.578425 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem...
Jul  2 00:01:00.580493 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr...
Jul  2 00:01:00.581789 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met.
Jul  2 00:01:00.581838 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot).
Jul  2 00:01:00.581917 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup.
Jul  2 00:01:00.588374 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr.
Jul  2 00:01:00.590725 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (796)
Jul  2 00:01:00.590748 kernel: BTRFS info (device vda6): first mount of filesystem f333e8f9-4cd9-418a-86af-1531564c69c1
Jul  2 00:01:00.591185 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup...
Jul  2 00:01:00.594492 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm
Jul  2 00:01:00.594513 kernel: BTRFS info (device vda6): using free space tree
Jul  2 00:01:00.596329 kernel: BTRFS info (device vda6): auto enabling async discard
Jul  2 00:01:00.598331 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem.
Jul  2 00:01:00.648520 initrd-setup-root[820]: cut: /sysroot/etc/passwd: No such file or directory
Jul  2 00:01:00.653817 initrd-setup-root[827]: cut: /sysroot/etc/group: No such file or directory
Jul  2 00:01:00.660122 initrd-setup-root[834]: cut: /sysroot/etc/shadow: No such file or directory
Jul  2 00:01:00.665642 initrd-setup-root[841]: cut: /sysroot/etc/gshadow: No such file or directory
Jul  2 00:01:00.752508 systemd-resolved[277]: Detected conflict on linux8 IN A 10.0.0.39
Jul  2 00:01:00.752527 systemd-resolved[277]: Hostname conflict, changing published hostname from 'linux8' to 'linux15'.
Jul  2 00:01:00.753050 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup.
Jul  2 00:01:00.759405 systemd[1]: Starting ignition-mount.service - Ignition (mount)...
Jul  2 00:01:00.761837 systemd[1]: Starting sysroot-boot.service - /sysroot/boot...
Jul  2 00:01:00.767317 kernel: BTRFS info (device vda6): last unmount of filesystem f333e8f9-4cd9-418a-86af-1531564c69c1
Jul  2 00:01:00.791510 ignition[909]: INFO     : Ignition 2.18.0
Jul  2 00:01:00.791510 ignition[909]: INFO     : Stage: mount
Jul  2 00:01:00.793307 ignition[909]: INFO     : no configs at "/usr/lib/ignition/base.d"
Jul  2 00:01:00.793307 ignition[909]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jul  2 00:01:00.793307 ignition[909]: INFO     : mount: mount passed
Jul  2 00:01:00.793307 ignition[909]: INFO     : Ignition finished successfully
Jul  2 00:01:00.793111 systemd[1]: Finished sysroot-boot.service - /sysroot/boot.
Jul  2 00:01:00.794398 systemd[1]: Finished ignition-mount.service - Ignition (mount).
Jul  2 00:01:00.806470 systemd[1]: Starting ignition-files.service - Ignition (files)...
Jul  2 00:01:01.088513 systemd[1]: sysroot-oem.mount: Deactivated successfully.
Jul  2 00:01:01.102535 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem...
Jul  2 00:01:01.120322 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (923)
Jul  2 00:01:01.122425 kernel: BTRFS info (device vda6): first mount of filesystem f333e8f9-4cd9-418a-86af-1531564c69c1
Jul  2 00:01:01.122452 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm
Jul  2 00:01:01.122463 kernel: BTRFS info (device vda6): using free space tree
Jul  2 00:01:01.126309 kernel: BTRFS info (device vda6): auto enabling async discard
Jul  2 00:01:01.127539 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem.
Jul  2 00:01:01.156298 ignition[940]: INFO     : Ignition 2.18.0
Jul  2 00:01:01.156298 ignition[940]: INFO     : Stage: files
Jul  2 00:01:01.157774 ignition[940]: INFO     : no configs at "/usr/lib/ignition/base.d"
Jul  2 00:01:01.157774 ignition[940]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jul  2 00:01:01.157774 ignition[940]: DEBUG    : files: compiled without relabeling support, skipping
Jul  2 00:01:01.161066 ignition[940]: INFO     : files: ensureUsers: op(1): [started]  creating or modifying user "core"
Jul  2 00:01:01.161066 ignition[940]: DEBUG    : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core"
Jul  2 00:01:01.161066 ignition[940]: INFO     : files: ensureUsers: op(1): [finished] creating or modifying user "core"
Jul  2 00:01:01.164701 ignition[940]: INFO     : files: ensureUsers: op(2): [started]  adding ssh keys to user "core"
Jul  2 00:01:01.164701 ignition[940]: INFO     : files: ensureUsers: op(2): [finished] adding ssh keys to user "core"
Jul  2 00:01:01.164701 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(3): [started]  writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz"
Jul  2 00:01:01.164701 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1
Jul  2 00:01:01.161539 unknown[940]: wrote ssh authorized keys file for user: core
Jul  2 00:01:01.203463 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(3): GET result: OK
Jul  2 00:01:01.254281 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz"
Jul  2 00:01:01.254281 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(4): [started]  writing file "/sysroot/opt/bin/cilium.tar.gz"
Jul  2 00:01:01.257547 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1
Jul  2 00:01:01.572967 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(4): GET result: OK
Jul  2 00:01:01.681212 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz"
Jul  2 00:01:01.681212 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(5): [started]  writing file "/sysroot/home/core/install.sh"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(6): [started]  writing file "/sysroot/home/core/nginx.yaml"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(7): [started]  writing file "/sysroot/home/core/nfs-pod.yaml"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(8): [started]  writing file "/sysroot/home/core/nfs-pvc.yaml"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(9): [started]  writing file "/sysroot/etc/flatcar/update.conf"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(a): [started]  writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(b): [started]  writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw"
Jul  2 00:01:01.684513 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1
Jul  2 00:01:01.878042 systemd-networkd[755]: eth0: Gained IPv6LL
Jul  2 00:01:01.962242 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(b): GET result: OK
Jul  2 00:01:02.220756 ignition[940]: INFO     : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw"
Jul  2 00:01:02.220756 ignition[940]: INFO     : files: op(c): [started]  processing unit "prepare-helm.service"
Jul  2 00:01:02.224383 ignition[940]: INFO     : files: op(c): op(d): [started]  writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service"
Jul  2 00:01:02.224383 ignition[940]: INFO     : files: op(c): op(d): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service"
Jul  2 00:01:02.224383 ignition[940]: INFO     : files: op(c): [finished] processing unit "prepare-helm.service"
Jul  2 00:01:02.224383 ignition[940]: INFO     : files: op(e): [started]  processing unit "coreos-metadata.service"
Jul  2 00:01:02.224383 ignition[940]: INFO     : files: op(e): op(f): [started]  writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service"
Jul  2 00:01:02.224383 ignition[940]: INFO     : files: op(e): op(f): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service"
Jul  2 00:01:02.224383 ignition[940]: INFO     : files: op(e): [finished] processing unit "coreos-metadata.service"
Jul  2 00:01:02.224383 ignition[940]: INFO     : files: op(10): [started]  setting preset to disabled for "coreos-metadata.service"
Jul  2 00:01:02.264945 ignition[940]: INFO     : files: op(10): op(11): [started]  removing enablement symlink(s) for "coreos-metadata.service"
Jul  2 00:01:02.269818 ignition[940]: INFO     : files: op(10): op(11): [finished] removing enablement symlink(s) for "coreos-metadata.service"
Jul  2 00:01:02.271518 ignition[940]: INFO     : files: op(10): [finished] setting preset to disabled for "coreos-metadata.service"
Jul  2 00:01:02.271518 ignition[940]: INFO     : files: op(12): [started]  setting preset to enabled for "prepare-helm.service"
Jul  2 00:01:02.271518 ignition[940]: INFO     : files: op(12): [finished] setting preset to enabled for "prepare-helm.service"
Jul  2 00:01:02.271518 ignition[940]: INFO     : files: createResultFile: createFiles: op(13): [started]  writing file "/sysroot/etc/.ignition-result.json"
Jul  2 00:01:02.271518 ignition[940]: INFO     : files: createResultFile: createFiles: op(13): [finished] writing file "/sysroot/etc/.ignition-result.json"
Jul  2 00:01:02.271518 ignition[940]: INFO     : files: files passed
Jul  2 00:01:02.271518 ignition[940]: INFO     : Ignition finished successfully
Jul  2 00:01:02.273574 systemd[1]: Finished ignition-files.service - Ignition (files).
Jul  2 00:01:02.285643 systemd[1]: Starting ignition-quench.service - Ignition (record completion)...
Jul  2 00:01:02.288559 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion...
Jul  2 00:01:02.293917 systemd[1]: ignition-quench.service: Deactivated successfully.
Jul  2 00:01:02.294110 systemd[1]: Finished ignition-quench.service - Ignition (record completion).
Jul  2 00:01:02.305737 initrd-setup-root-after-ignition[969]: grep: /sysroot/oem/oem-release: No such file or directory
Jul  2 00:01:02.311790 initrd-setup-root-after-ignition[971]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory
Jul  2 00:01:02.311790 initrd-setup-root-after-ignition[971]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory
Jul  2 00:01:02.314760 initrd-setup-root-after-ignition[975]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory
Jul  2 00:01:02.317280 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion.
Jul  2 00:01:02.321689 systemd[1]: Reached target ignition-complete.target - Ignition Complete.
Jul  2 00:01:02.333558 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root...
Jul  2 00:01:02.377509 systemd[1]: initrd-parse-etc.service: Deactivated successfully.
Jul  2 00:01:02.378485 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root.
Jul  2 00:01:02.379929 systemd[1]: Reached target initrd-fs.target - Initrd File Systems.
Jul  2 00:01:02.381621 systemd[1]: Reached target initrd.target - Initrd Default Target.
Jul  2 00:01:02.383465 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met.
Jul  2 00:01:02.393500 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook...
Jul  2 00:01:02.408283 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook.
Jul  2 00:01:02.428569 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons...
Jul  2 00:01:02.439358 systemd[1]: Stopped target network.target - Network.
Jul  2 00:01:02.440158 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups.
Jul  2 00:01:02.441572 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes.
Jul  2 00:01:02.442584 systemd[1]: Stopped target timers.target - Timer Units.
Jul  2 00:01:02.444002 systemd[1]: dracut-pre-pivot.service: Deactivated successfully.
Jul  2 00:01:02.444141 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook.
Jul  2 00:01:02.446474 systemd[1]: Stopped target initrd.target - Initrd Default Target.
Jul  2 00:01:02.448046 systemd[1]: Stopped target basic.target - Basic System.
Jul  2 00:01:02.449544 systemd[1]: Stopped target ignition-complete.target - Ignition Complete.
Jul  2 00:01:02.451106 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup.
Jul  2 00:01:02.452557 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device.
Jul  2 00:01:02.454237 systemd[1]: Stopped target remote-fs.target - Remote File Systems.
Jul  2 00:01:02.456115 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems.
Jul  2 00:01:02.457819 systemd[1]: Stopped target sysinit.target - System Initialization.
Jul  2 00:01:02.459248 systemd[1]: Stopped target local-fs.target - Local File Systems.
Jul  2 00:01:02.461011 systemd[1]: Stopped target swap.target - Swaps.
Jul  2 00:01:02.462270 systemd[1]: dracut-pre-mount.service: Deactivated successfully.
Jul  2 00:01:02.462438 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook.
Jul  2 00:01:02.464388 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes.
Jul  2 00:01:02.465923 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Jul  2 00:01:02.467472 systemd[1]: clevis-luks-askpass.path: Deactivated successfully.
Jul  2 00:01:02.467646 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Jul  2 00:01:02.469156 systemd[1]: dracut-initqueue.service: Deactivated successfully.
Jul  2 00:01:02.469314 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook.
Jul  2 00:01:02.471931 systemd[1]: ignition-fetch-offline.service: Deactivated successfully.
Jul  2 00:01:02.472052 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline).
Jul  2 00:01:02.474012 systemd[1]: Stopped target paths.target - Path Units.
Jul  2 00:01:02.475359 systemd[1]: systemd-ask-password-console.path: Deactivated successfully.
Jul  2 00:01:02.480335 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Jul  2 00:01:02.481458 systemd[1]: Stopped target slices.target - Slice Units.
Jul  2 00:01:02.483265 systemd[1]: Stopped target sockets.target - Socket Units.
Jul  2 00:01:02.485419 systemd[1]: iscsid.socket: Deactivated successfully.
Jul  2 00:01:02.485525 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket.
Jul  2 00:01:02.486789 systemd[1]: iscsiuio.socket: Deactivated successfully.
Jul  2 00:01:02.486864 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket.
Jul  2 00:01:02.488388 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully.
Jul  2 00:01:02.488505 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion.
Jul  2 00:01:02.489977 systemd[1]: ignition-files.service: Deactivated successfully.
Jul  2 00:01:02.490087 systemd[1]: Stopped ignition-files.service - Ignition (files).
Jul  2 00:01:02.500572 systemd[1]: Stopping ignition-mount.service - Ignition (mount)...
Jul  2 00:01:02.503078 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot...
Jul  2 00:01:02.504854 systemd[1]: Stopping systemd-networkd.service - Network Configuration...
Jul  2 00:01:02.506819 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution...
Jul  2 00:01:02.507665 systemd[1]: systemd-udev-trigger.service: Deactivated successfully.
Jul  2 00:01:02.507805 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices.
Jul  2 00:01:02.508844 systemd[1]: dracut-pre-trigger.service: Deactivated successfully.
Jul  2 00:01:02.508969 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook.
Jul  2 00:01:02.516677 systemd[1]: initrd-cleanup.service: Deactivated successfully.
Jul  2 00:01:02.516799 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons.
Jul  2 00:01:02.517354 systemd-networkd[755]: eth0: DHCPv6 lease lost
Jul  2 00:01:02.521224 systemd[1]: sysroot-boot.mount: Deactivated successfully.
Jul  2 00:01:02.527064 ignition[996]: INFO     : Ignition 2.18.0
Jul  2 00:01:02.527064 ignition[996]: INFO     : Stage: umount
Jul  2 00:01:02.527064 ignition[996]: INFO     : no configs at "/usr/lib/ignition/base.d"
Jul  2 00:01:02.527064 ignition[996]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jul  2 00:01:02.527064 ignition[996]: INFO     : umount: umount passed
Jul  2 00:01:02.527064 ignition[996]: INFO     : Ignition finished successfully
Jul  2 00:01:02.521858 systemd[1]: systemd-resolved.service: Deactivated successfully.
Jul  2 00:01:02.521993 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution.
Jul  2 00:01:02.527476 systemd[1]: systemd-networkd.service: Deactivated successfully.
Jul  2 00:01:02.527712 systemd[1]: Stopped systemd-networkd.service - Network Configuration.
Jul  2 00:01:02.529241 systemd[1]: ignition-mount.service: Deactivated successfully.
Jul  2 00:01:02.529416 systemd[1]: Stopped ignition-mount.service - Ignition (mount).
Jul  2 00:01:02.531316 systemd[1]: sysroot-boot.service: Deactivated successfully.
Jul  2 00:01:02.531417 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot.
Jul  2 00:01:02.537569 systemd[1]: systemd-networkd.socket: Deactivated successfully.
Jul  2 00:01:02.537629 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket.
Jul  2 00:01:02.539155 systemd[1]: ignition-disks.service: Deactivated successfully.
Jul  2 00:01:02.539213 systemd[1]: Stopped ignition-disks.service - Ignition (disks).
Jul  2 00:01:02.541046 systemd[1]: ignition-kargs.service: Deactivated successfully.
Jul  2 00:01:02.541099 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs).
Jul  2 00:01:02.542660 systemd[1]: ignition-setup.service: Deactivated successfully.
Jul  2 00:01:02.542707 systemd[1]: Stopped ignition-setup.service - Ignition (setup).
Jul  2 00:01:02.544115 systemd[1]: ignition-setup-pre.service: Deactivated successfully.
Jul  2 00:01:02.544154 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup.
Jul  2 00:01:02.546567 systemd[1]: initrd-setup-root.service: Deactivated successfully.
Jul  2 00:01:02.546608 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup.
Jul  2 00:01:02.561473 systemd[1]: Stopping network-cleanup.service - Network Cleanup...
Jul  2 00:01:02.562547 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully.
Jul  2 00:01:02.562624 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline.
Jul  2 00:01:02.564721 systemd[1]: systemd-sysctl.service: Deactivated successfully.
Jul  2 00:01:02.564784 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables.
Jul  2 00:01:02.566723 systemd[1]: systemd-modules-load.service: Deactivated successfully.
Jul  2 00:01:02.566787 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules.
Jul  2 00:01:02.569177 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully.
Jul  2 00:01:02.569230 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create Volatile Files and Directories.
Jul  2 00:01:02.571510 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files...
Jul  2 00:01:02.583566 systemd[1]: network-cleanup.service: Deactivated successfully.
Jul  2 00:01:02.583735 systemd[1]: Stopped network-cleanup.service - Network Cleanup.
Jul  2 00:01:02.590027 systemd[1]: systemd-udevd.service: Deactivated successfully.
Jul  2 00:01:02.590203 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files.
Jul  2 00:01:02.592907 systemd[1]: systemd-udevd-control.socket: Deactivated successfully.
Jul  2 00:01:02.592968 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket.
Jul  2 00:01:02.594876 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully.
Jul  2 00:01:02.594927 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket.
Jul  2 00:01:02.597167 systemd[1]: dracut-pre-udev.service: Deactivated successfully.
Jul  2 00:01:02.597234 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook.
Jul  2 00:01:02.599992 systemd[1]: dracut-cmdline.service: Deactivated successfully.
Jul  2 00:01:02.600052 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook.
Jul  2 00:01:02.602994 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully.
Jul  2 00:01:02.603068 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jul  2 00:01:02.617523 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database...
Jul  2 00:01:02.618482 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully.
Jul  2 00:01:02.618550 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Jul  2 00:01:02.620716 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully.
Jul  2 00:01:02.620765 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully.
Jul  2 00:01:02.622888 systemd[1]: kmod-static-nodes.service: Deactivated successfully.
Jul  2 00:01:02.622947 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes.
Jul  2 00:01:02.624944 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully.
Jul  2 00:01:02.624984 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup.
Jul  2 00:01:02.627461 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully.
Jul  2 00:01:02.627552 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database.
Jul  2 00:01:02.629555 systemd[1]: Reached target initrd-switch-root.target - Switch Root.
Jul  2 00:01:02.631573 systemd[1]: Starting initrd-switch-root.service - Switch Root...
Jul  2 00:01:02.644118 systemd[1]: Switching root.
Jul  2 00:01:02.671507 systemd-journald[238]: Journal stopped
Jul  2 00:01:03.606732 systemd-journald[238]: Received SIGTERM from PID 1 (systemd).
Jul  2 00:01:03.606790 kernel: SELinux:  policy capability network_peer_controls=1
Jul  2 00:01:03.606802 kernel: SELinux:  policy capability open_perms=1
Jul  2 00:01:03.606812 kernel: SELinux:  policy capability extended_socket_class=1
Jul  2 00:01:03.606821 kernel: SELinux:  policy capability always_check_network=0
Jul  2 00:01:03.606831 kernel: SELinux:  policy capability cgroup_seclabel=1
Jul  2 00:01:03.606840 kernel: SELinux:  policy capability nnp_nosuid_transition=1
Jul  2 00:01:03.606850 kernel: SELinux:  policy capability genfs_seclabel_symlinks=0
Jul  2 00:01:03.606859 kernel: SELinux:  policy capability ioctl_skip_cloexec=0
Jul  2 00:01:03.606871 kernel: audit: type=1403 audit(1719878462.853:2): auid=4294967295 ses=4294967295 lsm=selinux res=1
Jul  2 00:01:03.606902 systemd[1]: Successfully loaded SELinux policy in 44.908ms.
Jul  2 00:01:03.606928 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 13.604ms.
Jul  2 00:01:03.606940 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified)
Jul  2 00:01:03.606951 systemd[1]: Detected virtualization kvm.
Jul  2 00:01:03.606961 systemd[1]: Detected architecture arm64.
Jul  2 00:01:03.606972 systemd[1]: Detected first boot.
Jul  2 00:01:03.606983 systemd[1]: Initializing machine ID from VM UUID.
Jul  2 00:01:03.606994 zram_generator::config[1042]: No configuration found.
Jul  2 00:01:03.607007 systemd[1]: Populated /etc with preset unit settings.
Jul  2 00:01:03.607019 systemd[1]: initrd-switch-root.service: Deactivated successfully.
Jul  2 00:01:03.607030 systemd[1]: Stopped initrd-switch-root.service - Switch Root.
Jul  2 00:01:03.607043 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1.
Jul  2 00:01:03.607055 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config.
Jul  2 00:01:03.607066 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run.
Jul  2 00:01:03.607076 systemd[1]: Created slice system-getty.slice - Slice /system/getty.
Jul  2 00:01:03.607087 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe.
Jul  2 00:01:03.607100 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty.
Jul  2 00:01:03.607111 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit.
Jul  2 00:01:03.607122 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck.
Jul  2 00:01:03.607132 systemd[1]: Created slice user.slice - User and Session Slice.
Jul  2 00:01:03.607143 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Jul  2 00:01:03.607155 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Jul  2 00:01:03.607168 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch.
Jul  2 00:01:03.607180 systemd[1]: Set up automount boot.automount - Boot partition Automount Point.
Jul  2 00:01:03.607191 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point.
Jul  2 00:01:03.607204 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM...
Jul  2 00:01:03.607215 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0...
Jul  2 00:01:03.607225 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Jul  2 00:01:03.607236 systemd[1]: Stopped target initrd-switch-root.target - Switch Root.
Jul  2 00:01:03.607247 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems.
Jul  2 00:01:03.607258 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System.
Jul  2 00:01:03.607268 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes.
Jul  2 00:01:03.607280 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes.
Jul  2 00:01:03.609501 systemd[1]: Reached target remote-fs.target - Remote File Systems.
Jul  2 00:01:03.609546 systemd[1]: Reached target slices.target - Slice Units.
Jul  2 00:01:03.609558 systemd[1]: Reached target swap.target - Swaps.
Jul  2 00:01:03.609569 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes.
Jul  2 00:01:03.609581 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket.
Jul  2 00:01:03.609591 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket.
Jul  2 00:01:03.609602 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket.
Jul  2 00:01:03.609613 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket.
Jul  2 00:01:03.609625 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket.
Jul  2 00:01:03.609643 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System...
Jul  2 00:01:03.609654 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System...
Jul  2 00:01:03.609664 systemd[1]: Mounting media.mount - External Media Directory...
Jul  2 00:01:03.609676 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System...
Jul  2 00:01:03.609686 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System...
Jul  2 00:01:03.609697 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp...
Jul  2 00:01:03.609708 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw).
Jul  2 00:01:03.609719 systemd[1]: Reached target machines.target - Containers.
Jul  2 00:01:03.609731 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files...
Jul  2 00:01:03.609742 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jul  2 00:01:03.609753 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes...
Jul  2 00:01:03.609764 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs...
Jul  2 00:01:03.609775 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jul  2 00:01:03.609789 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm...
Jul  2 00:01:03.609800 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jul  2 00:01:03.609810 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse...
Jul  2 00:01:03.609820 kernel: ACPI: bus type drm_connector registered
Jul  2 00:01:03.609834 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jul  2 00:01:03.609845 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf).
Jul  2 00:01:03.609856 systemd[1]: systemd-fsck-root.service: Deactivated successfully.
Jul  2 00:01:03.609866 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device.
Jul  2 00:01:03.609888 systemd[1]: systemd-fsck-usr.service: Deactivated successfully.
Jul  2 00:01:03.609899 systemd[1]: Stopped systemd-fsck-usr.service.
Jul  2 00:01:03.609909 kernel: fuse: init (API version 7.39)
Jul  2 00:01:03.609919 kernel: loop: module loaded
Jul  2 00:01:03.609931 systemd[1]: Starting systemd-journald.service - Journal Service...
Jul  2 00:01:03.609943 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules...
Jul  2 00:01:03.609956 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line...
Jul  2 00:01:03.610001 systemd-journald[1108]: Collecting audit messages is disabled.
Jul  2 00:01:03.610030 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems...
Jul  2 00:01:03.610042 systemd-journald[1108]: Journal started
Jul  2 00:01:03.610066 systemd-journald[1108]: Runtime Journal (/run/log/journal/9b34d472248a4422b5ae1e477ecfad8c) is 5.9M, max 47.3M, 41.4M free.
Jul  2 00:01:03.367766 systemd[1]: Queued start job for default target multi-user.target.
Jul  2 00:01:03.389822 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6.
Jul  2 00:01:03.390222 systemd[1]: systemd-journald.service: Deactivated successfully.
Jul  2 00:01:03.614415 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices...
Jul  2 00:01:03.615734 systemd[1]: verity-setup.service: Deactivated successfully.
Jul  2 00:01:03.615789 systemd[1]: Stopped verity-setup.service.
Jul  2 00:01:03.619933 systemd[1]: Started systemd-journald.service - Journal Service.
Jul  2 00:01:03.620648 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System.
Jul  2 00:01:03.621929 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System.
Jul  2 00:01:03.623363 systemd[1]: Mounted media.mount - External Media Directory.
Jul  2 00:01:03.624673 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System.
Jul  2 00:01:03.625971 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System.
Jul  2 00:01:03.627271 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp.
Jul  2 00:01:03.628690 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files.
Jul  2 00:01:03.630206 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes.
Jul  2 00:01:03.631844 systemd[1]: modprobe@configfs.service: Deactivated successfully.
Jul  2 00:01:03.632001 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs.
Jul  2 00:01:03.633508 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jul  2 00:01:03.633653 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jul  2 00:01:03.635043 systemd[1]: modprobe@drm.service: Deactivated successfully.
Jul  2 00:01:03.635188 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm.
Jul  2 00:01:03.636779 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jul  2 00:01:03.636941 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jul  2 00:01:03.638515 systemd[1]: modprobe@fuse.service: Deactivated successfully.
Jul  2 00:01:03.638670 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse.
Jul  2 00:01:03.640179 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jul  2 00:01:03.640352 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jul  2 00:01:03.641754 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules.
Jul  2 00:01:03.643187 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line.
Jul  2 00:01:03.644999 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems.
Jul  2 00:01:03.659009 systemd[1]: Reached target network-pre.target - Preparation for Network.
Jul  2 00:01:03.670415 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System...
Jul  2 00:01:03.672820 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System...
Jul  2 00:01:03.674076 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/).
Jul  2 00:01:03.674127 systemd[1]: Reached target local-fs.target - Local File Systems.
Jul  2 00:01:03.676271 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink).
Jul  2 00:01:03.678792 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown...
Jul  2 00:01:03.681244 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache...
Jul  2 00:01:03.682473 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jul  2 00:01:03.684313 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database...
Jul  2 00:01:03.686523 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage...
Jul  2 00:01:03.687808 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore).
Jul  2 00:01:03.691500 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed...
Jul  2 00:01:03.692800 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met.
Jul  2 00:01:03.696533 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Jul  2 00:01:03.701508 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/...
Jul  2 00:01:03.703810 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully...
Jul  2 00:01:03.707646 systemd-journald[1108]: Time spent on flushing to /var/log/journal/9b34d472248a4422b5ae1e477ecfad8c is 15.095ms for 863 entries.
Jul  2 00:01:03.707646 systemd-journald[1108]: System Journal (/var/log/journal/9b34d472248a4422b5ae1e477ecfad8c) is 8.0M, max 195.6M, 187.6M free.
Jul  2 00:01:03.738273 systemd-journald[1108]: Received client request to flush runtime journal.
Jul  2 00:01:03.738348 kernel: loop0: detected capacity change from 0 to 194096
Jul  2 00:01:03.709509 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices.
Jul  2 00:01:03.710821 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System.
Jul  2 00:01:03.712065 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System.
Jul  2 00:01:03.714481 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown.
Jul  2 00:01:03.715782 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed.
Jul  2 00:01:03.719869 systemd[1]: Reached target first-boot-complete.target - First Boot Complete.
Jul  2 00:01:03.723316 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk...
Jul  2 00:01:03.726052 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization...
Jul  2 00:01:03.740861 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage.
Jul  2 00:01:03.746350 kernel: block loop0: the capability attribute has been deprecated.
Jul  2 00:01:03.748335 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Jul  2 00:01:03.758130 systemd[1]: etc-machine\x2did.mount: Deactivated successfully.
Jul  2 00:01:03.758851 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk.
Jul  2 00:01:03.763851 udevadm[1161]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in.
Jul  2 00:01:03.765328 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher
Jul  2 00:01:03.769667 systemd-tmpfiles[1153]: ACLs are not supported, ignoring.
Jul  2 00:01:03.769690 systemd-tmpfiles[1153]: ACLs are not supported, ignoring.
Jul  2 00:01:03.775016 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully.
Jul  2 00:01:03.782537 systemd[1]: Starting systemd-sysusers.service - Create System Users...
Jul  2 00:01:03.793391 kernel: loop1: detected capacity change from 0 to 113672
Jul  2 00:01:03.808497 systemd[1]: Finished systemd-sysusers.service - Create System Users.
Jul  2 00:01:03.814560 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev...
Jul  2 00:01:03.830113 systemd-tmpfiles[1175]: ACLs are not supported, ignoring.
Jul  2 00:01:03.830133 systemd-tmpfiles[1175]: ACLs are not supported, ignoring.
Jul  2 00:01:03.836884 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Jul  2 00:01:03.837458 kernel: loop2: detected capacity change from 0 to 59672
Jul  2 00:01:03.887333 kernel: loop3: detected capacity change from 0 to 194096
Jul  2 00:01:03.897353 kernel: loop4: detected capacity change from 0 to 113672
Jul  2 00:01:03.903343 kernel: loop5: detected capacity change from 0 to 59672
Jul  2 00:01:03.908695 (sd-merge)[1180]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'.
Jul  2 00:01:03.909103 (sd-merge)[1180]: Merged extensions into '/usr'.
Jul  2 00:01:03.913734 systemd[1]: Reloading requested from client PID 1152 ('systemd-sysext') (unit systemd-sysext.service)...
Jul  2 00:01:03.913753 systemd[1]: Reloading...
Jul  2 00:01:03.973643 zram_generator::config[1207]: No configuration found.
Jul  2 00:01:04.011304 ldconfig[1147]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start.
Jul  2 00:01:04.071477 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jul  2 00:01:04.111036 systemd[1]: Reloading finished in 196 ms.
Jul  2 00:01:04.153341 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache.
Jul  2 00:01:04.154591 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/.
Jul  2 00:01:04.170668 systemd[1]: Starting ensure-sysext.service...
Jul  2 00:01:04.172891 systemd[1]: Starting systemd-tmpfiles-setup.service - Create Volatile Files and Directories...
Jul  2 00:01:04.185123 systemd[1]: Reloading requested from client PID 1239 ('systemctl') (unit ensure-sysext.service)...
Jul  2 00:01:04.185139 systemd[1]: Reloading...
Jul  2 00:01:04.202450 systemd-tmpfiles[1241]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring.
Jul  2 00:01:04.202705 systemd-tmpfiles[1241]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring.
Jul  2 00:01:04.203366 systemd-tmpfiles[1241]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring.
Jul  2 00:01:04.203581 systemd-tmpfiles[1241]: ACLs are not supported, ignoring.
Jul  2 00:01:04.203624 systemd-tmpfiles[1241]: ACLs are not supported, ignoring.
Jul  2 00:01:04.207270 systemd-tmpfiles[1241]: Detected autofs mount point /boot during canonicalization of boot.
Jul  2 00:01:04.207512 systemd-tmpfiles[1241]: Skipping /boot
Jul  2 00:01:04.214200 systemd-tmpfiles[1241]: Detected autofs mount point /boot during canonicalization of boot.
Jul  2 00:01:04.214348 systemd-tmpfiles[1241]: Skipping /boot
Jul  2 00:01:04.242553 zram_generator::config[1264]: No configuration found.
Jul  2 00:01:04.335746 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jul  2 00:01:04.376680 systemd[1]: Reloading finished in 191 ms.
Jul  2 00:01:04.390764 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database.
Jul  2 00:01:04.408256 systemd[1]: Finished systemd-tmpfiles-setup.service - Create Volatile Files and Directories.
Jul  2 00:01:04.416158 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules...
Jul  2 00:01:04.418916 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs...
Jul  2 00:01:04.421083 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog...
Jul  2 00:01:04.426560 systemd[1]: Starting systemd-resolved.service - Network Name Resolution...
Jul  2 00:01:04.443765 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files...
Jul  2 00:01:04.450749 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP...
Jul  2 00:01:04.454701 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog.
Jul  2 00:01:04.458548 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jul  2 00:01:04.461415 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jul  2 00:01:04.465590 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jul  2 00:01:04.470851 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jul  2 00:01:04.471958 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jul  2 00:01:04.474678 systemd[1]: Starting systemd-update-done.service - Update is Completed...
Jul  2 00:01:04.476156 systemd-udevd[1313]: Using default interface naming scheme 'v255'.
Jul  2 00:01:04.481288 systemd[1]: Starting systemd-userdbd.service - User Database Manager...
Jul  2 00:01:04.484630 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jul  2 00:01:04.486682 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jul  2 00:01:04.489681 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs.
Jul  2 00:01:04.491714 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jul  2 00:01:04.491869 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jul  2 00:01:04.493715 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jul  2 00:01:04.493851 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jul  2 00:01:04.495666 systemd[1]: Finished systemd-update-done.service - Update is Completed.
Jul  2 00:01:04.499217 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files.
Jul  2 00:01:04.519095 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP.
Jul  2 00:01:04.522998 systemd[1]: Finished ensure-sysext.service.
Jul  2 00:01:04.525958 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jul  2 00:01:04.535626 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jul  2 00:01:04.538668 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm...
Jul  2 00:01:04.541354 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jul  2 00:01:04.544217 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jul  2 00:01:04.545661 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jul  2 00:01:04.547671 augenrules[1356]: No rules
Jul  2 00:01:04.550745 systemd[1]: Starting systemd-networkd.service - Network Configuration...
Jul  2 00:01:04.558012 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization...
Jul  2 00:01:04.559247 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt).
Jul  2 00:01:04.559565 systemd[1]: Started systemd-userdbd.service - User Database Manager.
Jul  2 00:01:04.561396 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules.
Jul  2 00:01:04.563120 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jul  2 00:01:04.563343 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jul  2 00:01:04.564888 systemd[1]: modprobe@drm.service: Deactivated successfully.
Jul  2 00:01:04.565042 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm.
Jul  2 00:01:04.566602 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jul  2 00:01:04.566735 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jul  2 00:01:04.569890 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jul  2 00:01:04.570100 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jul  2 00:01:04.584996 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped.
Jul  2 00:01:04.588030 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore).
Jul  2 00:01:04.588102 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met.
Jul  2 00:01:04.588623 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1334)
Jul  2 00:01:04.598313 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1348)
Jul  2 00:01:04.630620 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM.
Jul  2 00:01:04.639560 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM...
Jul  2 00:01:04.672508 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization.
Jul  2 00:01:04.673903 systemd[1]: Reached target time-set.target - System Time Set.
Jul  2 00:01:04.682508 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM.
Jul  2 00:01:04.724647 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Jul  2 00:01:04.730305 systemd-resolved[1306]: Positive Trust Anchors:
Jul  2 00:01:04.730322 systemd-resolved[1306]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Jul  2 00:01:04.730354 systemd-resolved[1306]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa corp home internal intranet lan local private test
Jul  2 00:01:04.732465 systemd-networkd[1368]: lo: Link UP
Jul  2 00:01:04.732472 systemd-networkd[1368]: lo: Gained carrier
Jul  2 00:01:04.733244 systemd-networkd[1368]: Enumeration completed
Jul  2 00:01:04.734878 systemd[1]: Started systemd-networkd.service - Network Configuration.
Jul  2 00:01:04.738007 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured...
Jul  2 00:01:04.744838 systemd-resolved[1306]: Defaulting to hostname 'linux'.
Jul  2 00:01:04.755458 systemd-networkd[1368]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jul  2 00:01:04.755468 systemd-networkd[1368]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network.
Jul  2 00:01:04.755510 systemd[1]: Started systemd-resolved.service - Network Name Resolution.
Jul  2 00:01:04.757053 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization.
Jul  2 00:01:04.758452 systemd-networkd[1368]: eth0: Link UP
Jul  2 00:01:04.758459 systemd-networkd[1368]: eth0: Gained carrier
Jul  2 00:01:04.758478 systemd-networkd[1368]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jul  2 00:01:04.759892 systemd[1]: Reached target network.target - Network.
Jul  2 00:01:04.760969 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups.
Jul  2 00:01:04.765012 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes...
Jul  2 00:01:04.805591 lvm[1393]:   WARNING: Failed to connect to lvmetad. Falling back to device scanning.
Jul  2 00:01:04.811365 systemd-networkd[1368]: eth0: DHCPv4 address 10.0.0.39/16, gateway 10.0.0.1 acquired from 10.0.0.1
Jul  2 00:01:04.812228 systemd-timesyncd[1369]: Network configuration changed, trying to establish connection.
Jul  2 00:01:04.812957 systemd-timesyncd[1369]: Contacted time server 10.0.0.1:123 (10.0.0.1).
Jul  2 00:01:04.813009 systemd-timesyncd[1369]: Initial clock synchronization to Tue 2024-07-02 00:01:04.933913 UTC.
Jul  2 00:01:04.824494 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Jul  2 00:01:04.841315 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes.
Jul  2 00:01:04.843044 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes.
Jul  2 00:01:04.844218 systemd[1]: Reached target sysinit.target - System Initialization.
Jul  2 00:01:04.845743 systemd[1]: Started motdgen.path - Watch for update engine configuration changes.
Jul  2 00:01:04.847067 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data.
Jul  2 00:01:04.848595 systemd[1]: Started logrotate.timer - Daily rotation of log files.
Jul  2 00:01:04.849829 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information..
Jul  2 00:01:04.851335 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories.
Jul  2 00:01:04.852559 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate).
Jul  2 00:01:04.852601 systemd[1]: Reached target paths.target - Path Units.
Jul  2 00:01:04.853560 systemd[1]: Reached target timers.target - Timer Units.
Jul  2 00:01:04.856477 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket.
Jul  2 00:01:04.859172 systemd[1]: Starting docker.socket - Docker Socket for the API...
Jul  2 00:01:04.868461 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket.
Jul  2 00:01:04.871626 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes...
Jul  2 00:01:04.873275 systemd[1]: Listening on docker.socket - Docker Socket for the API.
Jul  2 00:01:04.874225 systemd[1]: Reached target sockets.target - Socket Units.
Jul  2 00:01:04.875054 systemd[1]: Reached target basic.target - Basic System.
Jul  2 00:01:04.875811 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met.
Jul  2 00:01:04.875841 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met.
Jul  2 00:01:04.877036 systemd[1]: Starting containerd.service - containerd container runtime...
Jul  2 00:01:04.879094 systemd[1]: Starting dbus.service - D-Bus System Message Bus...
Jul  2 00:01:04.880632 lvm[1401]:   WARNING: Failed to connect to lvmetad. Falling back to device scanning.
Jul  2 00:01:04.883450 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit...
Jul  2 00:01:04.887651 systemd[1]: Starting extend-filesystems.service - Extend Filesystems...
Jul  2 00:01:04.896504 jq[1404]: false
Jul  2 00:01:04.891535 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment).
Jul  2 00:01:04.895154 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd...
Jul  2 00:01:04.899456 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin...
Jul  2 00:01:04.903712 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline...
Jul  2 00:01:04.906362 extend-filesystems[1405]: Found loop3
Jul  2 00:01:04.906362 extend-filesystems[1405]: Found loop4
Jul  2 00:01:04.906362 extend-filesystems[1405]: Found loop5
Jul  2 00:01:04.906362 extend-filesystems[1405]: Found vda
Jul  2 00:01:04.906362 extend-filesystems[1405]: Found vda1
Jul  2 00:01:04.914566 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys...
Jul  2 00:01:04.920113 extend-filesystems[1405]: Found vda2
Jul  2 00:01:04.920113 extend-filesystems[1405]: Found vda3
Jul  2 00:01:04.920113 extend-filesystems[1405]: Found usr
Jul  2 00:01:04.920113 extend-filesystems[1405]: Found vda4
Jul  2 00:01:04.920113 extend-filesystems[1405]: Found vda6
Jul  2 00:01:04.920113 extend-filesystems[1405]: Found vda7
Jul  2 00:01:04.920113 extend-filesystems[1405]: Found vda9
Jul  2 00:01:04.920113 extend-filesystems[1405]: Checking size of /dev/vda9
Jul  2 00:01:04.969624 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (1334)
Jul  2 00:01:04.969655 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks
Jul  2 00:01:04.926147 systemd[1]: Starting systemd-logind.service - User Login Management...
Jul  2 00:01:04.969920 extend-filesystems[1405]: Resized partition /dev/vda9
Jul  2 00:01:04.925648 dbus-daemon[1403]: [system] SELinux support is enabled
Jul  2 00:01:04.933766 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0).
Jul  2 00:01:04.974904 extend-filesystems[1421]: resize2fs 1.47.0 (5-Feb-2023)
Jul  2 00:01:05.022390 kernel: EXT4-fs (vda9): resized filesystem to 1864699
Jul  2 00:01:05.022439 update_engine[1425]: I0702 00:01:05.003409  1425 main.cc:92] Flatcar Update Engine starting
Jul  2 00:01:05.022439 update_engine[1425]: I0702 00:01:05.005972  1425 update_check_scheduler.cc:74] Next update check in 2m23s
Jul  2 00:01:04.934397 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details.
Jul  2 00:01:05.022838 extend-filesystems[1421]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required
Jul  2 00:01:05.022838 extend-filesystems[1421]: old_desc_blocks = 1, new_desc_blocks = 1
Jul  2 00:01:05.022838 extend-filesystems[1421]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long.
Jul  2 00:01:04.947759 systemd[1]: Starting update-engine.service - Update Engine...
Jul  2 00:01:05.032954 extend-filesystems[1405]: Resized filesystem in /dev/vda9
Jul  2 00:01:05.035323 jq[1427]: true
Jul  2 00:01:04.952097 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition...
Jul  2 00:01:04.956922 systemd[1]: Started dbus.service - D-Bus System Message Bus.
Jul  2 00:01:05.036545 tar[1429]: linux-arm64/helm
Jul  2 00:01:04.963325 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes.
Jul  2 00:01:05.036911 jq[1430]: true
Jul  2 00:01:04.971575 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'.
Jul  2 00:01:04.971747 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped.
Jul  2 00:01:04.972036 systemd[1]: motdgen.service: Deactivated successfully.
Jul  2 00:01:04.972183 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd.
Jul  2 00:01:04.979835 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully.
Jul  2 00:01:04.982288 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline.
Jul  2 00:01:05.001481 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml).
Jul  2 00:01:05.001510 systemd[1]: Reached target system-config.target - Load system-provided cloud configs.
Jul  2 00:01:05.002395 (ntainerd)[1431]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR
Jul  2 00:01:05.007316 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url).
Jul  2 00:01:05.007342 systemd[1]: Reached target user-config.target - Load user-provided cloud configs.
Jul  2 00:01:05.010111 systemd[1]: Started update-engine.service - Update Engine.
Jul  2 00:01:05.012972 systemd[1]: Started locksmithd.service - Cluster reboot manager.
Jul  2 00:01:05.021725 systemd[1]: extend-filesystems.service: Deactivated successfully.
Jul  2 00:01:05.021949 systemd[1]: Finished extend-filesystems.service - Extend Filesystems.
Jul  2 00:01:05.046900 systemd-logind[1423]: Watching system buttons on /dev/input/event0 (Power Button)
Jul  2 00:01:05.047963 systemd-logind[1423]: New seat seat0.
Jul  2 00:01:05.050013 systemd[1]: Started systemd-logind.service - User Login Management.
Jul  2 00:01:05.097516 bash[1458]: Updated "/home/core/.ssh/authorized_keys"
Jul  2 00:01:05.099192 locksmithd[1439]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot"
Jul  2 00:01:05.099368 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition.
Jul  2 00:01:05.101807 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met.
Jul  2 00:01:05.282881 containerd[1431]: time="2024-07-02T00:01:05.280161924Z" level=info msg="starting containerd" revision=1fbfc07f8d28210e62bdbcbf7b950bac8028afbf version=v1.7.17
Jul  2 00:01:05.316728 containerd[1431]: time="2024-07-02T00:01:05.316677225Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1
Jul  2 00:01:05.316728 containerd[1431]: time="2024-07-02T00:01:05.316727442Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1
Jul  2 00:01:05.318172 containerd[1431]: time="2024-07-02T00:01:05.318134043Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.36-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1
Jul  2 00:01:05.318172 containerd[1431]: time="2024-07-02T00:01:05.318171162Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1
Jul  2 00:01:05.318703 containerd[1431]: time="2024-07-02T00:01:05.318586360Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jul  2 00:01:05.318703 containerd[1431]: time="2024-07-02T00:01:05.318662975Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1
Jul  2 00:01:05.318785 containerd[1431]: time="2024-07-02T00:01:05.318764014Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1
Jul  2 00:01:05.318897 containerd[1431]: time="2024-07-02T00:01:05.318874523Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1
Jul  2 00:01:05.318925 containerd[1431]: time="2024-07-02T00:01:05.318897456Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1
Jul  2 00:01:05.318981 containerd[1431]: time="2024-07-02T00:01:05.318965406Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1
Jul  2 00:01:05.319251 containerd[1431]: time="2024-07-02T00:01:05.319229388Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1
Jul  2 00:01:05.319288 containerd[1431]: time="2024-07-02T00:01:05.319257076Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured"
Jul  2 00:01:05.319288 containerd[1431]: time="2024-07-02T00:01:05.319277993Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1
Jul  2 00:01:05.319538 containerd[1431]: time="2024-07-02T00:01:05.319457823Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jul  2 00:01:05.319538 containerd[1431]: time="2024-07-02T00:01:05.319482327Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1
Jul  2 00:01:05.319693 containerd[1431]: time="2024-07-02T00:01:05.319540927Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured"
Jul  2 00:01:05.319693 containerd[1431]: time="2024-07-02T00:01:05.319615044Z" level=info msg="metadata content store policy set" policy=shared
Jul  2 00:01:05.327692 containerd[1431]: time="2024-07-02T00:01:05.327651466Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.327700192Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.327714419Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.327769593Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.327785070Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.327796395Z" level=info msg="NRI interface is disabled by configuration."
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.327809009Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.328031842Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.328051309Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.328067591Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.328081455Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.328095641Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.328113294Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.328129576Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1
Jul  2 00:01:05.328347 containerd[1431]: time="2024-07-02T00:01:05.328142393Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1
Jul  2 00:01:05.328661 containerd[1431]: time="2024-07-02T00:01:05.328156095Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1
Jul  2 00:01:05.328661 containerd[1431]: time="2024-07-02T00:01:05.328171048Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1
Jul  2 00:01:05.328661 containerd[1431]: time="2024-07-02T00:01:05.328184307Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1
Jul  2 00:01:05.328661 containerd[1431]: time="2024-07-02T00:01:05.328197285Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1
Jul  2 00:01:05.328661 containerd[1431]: time="2024-07-02T00:01:05.328292117Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1
Jul  2 00:01:05.328896 containerd[1431]: time="2024-07-02T00:01:05.328874933Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1
Jul  2 00:01:05.328931 containerd[1431]: time="2024-07-02T00:01:05.328913664Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.328952 containerd[1431]: time="2024-07-02T00:01:05.328930228Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1
Jul  2 00:01:05.328974 containerd[1431]: time="2024-07-02T00:01:05.328954692Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1
Jul  2 00:01:05.329207 containerd[1431]: time="2024-07-02T00:01:05.329191308Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329237 containerd[1431]: time="2024-07-02T00:01:05.329212387Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329237 containerd[1431]: time="2024-07-02T00:01:05.329227984Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329280 containerd[1431]: time="2024-07-02T00:01:05.329240276Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329280 containerd[1431]: time="2024-07-02T00:01:05.329254060Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329280 containerd[1431]: time="2024-07-02T00:01:05.329267480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329362 containerd[1431]: time="2024-07-02T00:01:05.329280780Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329470 containerd[1431]: time="2024-07-02T00:01:05.329292911Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329470 containerd[1431]: time="2024-07-02T00:01:05.329437920Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1
Jul  2 00:01:05.329661 containerd[1431]: time="2024-07-02T00:01:05.329593851Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329693 containerd[1431]: time="2024-07-02T00:01:05.329677318Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329727 containerd[1431]: time="2024-07-02T00:01:05.329698396Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329727 containerd[1431]: time="2024-07-02T00:01:05.329713227Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329768 containerd[1431]: time="2024-07-02T00:01:05.329726084Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329768 containerd[1431]: time="2024-07-02T00:01:05.329740432Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329768 containerd[1431]: time="2024-07-02T00:01:05.329753490Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.329768 containerd[1431]: time="2024-07-02T00:01:05.329765177Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1
Jul  2 00:01:05.330301 containerd[1431]: time="2024-07-02T00:01:05.330241111Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}"
Jul  2 00:01:05.330426 containerd[1431]: time="2024-07-02T00:01:05.330319500Z" level=info msg="Connect containerd service"
Jul  2 00:01:05.330426 containerd[1431]: time="2024-07-02T00:01:05.330361092Z" level=info msg="using legacy CRI server"
Jul  2 00:01:05.330426 containerd[1431]: time="2024-07-02T00:01:05.330369636Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this"
Jul  2 00:01:05.330613 containerd[1431]: time="2024-07-02T00:01:05.330596056Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\""
Jul  2 00:01:05.331632 containerd[1431]: time="2024-07-02T00:01:05.331601969Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config"
Jul  2 00:01:05.331686 containerd[1431]: time="2024-07-02T00:01:05.331670282Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1
Jul  2 00:01:05.331708 containerd[1431]: time="2024-07-02T00:01:05.331688459Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1
Jul  2 00:01:05.331708 containerd[1431]: time="2024-07-02T00:01:05.331699945Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1
Jul  2 00:01:05.331745 containerd[1431]: time="2024-07-02T00:01:05.331713124Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1
Jul  2 00:01:05.334320 containerd[1431]: time="2024-07-02T00:01:05.332011807Z" level=info msg="Start subscribing containerd event"
Jul  2 00:01:05.334320 containerd[1431]: time="2024-07-02T00:01:05.332211224Z" level=info msg="Start recovering state"
Jul  2 00:01:05.334320 containerd[1431]: time="2024-07-02T00:01:05.332290862Z" level=info msg="Start event monitor"
Jul  2 00:01:05.334320 containerd[1431]: time="2024-07-02T00:01:05.332320444Z" level=info msg="Start snapshots syncer"
Jul  2 00:01:05.334320 containerd[1431]: time="2024-07-02T00:01:05.332329754Z" level=info msg="Start cni network conf syncer for default"
Jul  2 00:01:05.334320 containerd[1431]: time="2024-07-02T00:01:05.332338097Z" level=info msg="Start streaming server"
Jul  2 00:01:05.334320 containerd[1431]: time="2024-07-02T00:01:05.332760790Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc
Jul  2 00:01:05.334320 containerd[1431]: time="2024-07-02T00:01:05.332804358Z" level=info msg=serving... address=/run/containerd/containerd.sock
Jul  2 00:01:05.334320 containerd[1431]: time="2024-07-02T00:01:05.332850988Z" level=info msg="containerd successfully booted in 0.055839s"
Jul  2 00:01:05.332963 systemd[1]: Started containerd.service - containerd container runtime.
Jul  2 00:01:05.382522 tar[1429]: linux-arm64/LICENSE
Jul  2 00:01:05.382522 tar[1429]: linux-arm64/README.md
Jul  2 00:01:05.388938 sshd_keygen[1424]: ssh-keygen: generating new host keys: RSA ECDSA ED25519
Jul  2 00:01:05.396766 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin.
Jul  2 00:01:05.410168 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys.
Jul  2 00:01:05.418601 systemd[1]: Starting issuegen.service - Generate /run/issue...
Jul  2 00:01:05.425846 systemd[1]: issuegen.service: Deactivated successfully.
Jul  2 00:01:05.427056 systemd[1]: Finished issuegen.service - Generate /run/issue.
Jul  2 00:01:05.430325 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions...
Jul  2 00:01:05.443434 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions.
Jul  2 00:01:05.446444 systemd[1]: Started getty@tty1.service - Getty on tty1.
Jul  2 00:01:05.448706 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0.
Jul  2 00:01:05.450225 systemd[1]: Reached target getty.target - Login Prompts.
Jul  2 00:01:06.039508 systemd-networkd[1368]: eth0: Gained IPv6LL
Jul  2 00:01:06.045530 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured.
Jul  2 00:01:06.047554 systemd[1]: Reached target network-online.target - Network is Online.
Jul  2 00:01:06.067667 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent...
Jul  2 00:01:06.070861 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jul  2 00:01:06.073716 systemd[1]: Starting nvidia.service - NVIDIA Configure Service...
Jul  2 00:01:06.092371 systemd[1]: coreos-metadata.service: Deactivated successfully.
Jul  2 00:01:06.092574 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent.
Jul  2 00:01:06.094996 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met.
Jul  2 00:01:06.099780 systemd[1]: Finished nvidia.service - NVIDIA Configure Service.
Jul  2 00:01:06.685637 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jul  2 00:01:06.687895 systemd[1]: Reached target multi-user.target - Multi-User System.
Jul  2 00:01:06.691197 (kubelet)[1513]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Jul  2 00:01:06.692408 systemd[1]: Startup finished in 706ms (kernel) + 5.055s (initrd) + 3.886s (userspace) = 9.648s.
Jul  2 00:01:07.275228 kubelet[1513]: E0702 00:01:07.275155    1513 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Jul  2 00:01:07.278532 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Jul  2 00:01:07.278684 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Jul  2 00:01:10.588450 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd.
Jul  2 00:01:10.593069 systemd[1]: Started sshd@0-10.0.0.39:22-10.0.0.1:59354.service - OpenSSH per-connection server daemon (10.0.0.1:59354).
Jul  2 00:01:10.663737 sshd[1529]: Accepted publickey for core from 10.0.0.1 port 59354 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:01:10.671463 sshd[1529]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:01:10.696228 systemd[1]: Created slice user-500.slice - User Slice of UID 500.
Jul  2 00:01:10.711659 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500...
Jul  2 00:01:10.714940 systemd-logind[1423]: New session 1 of user core.
Jul  2 00:01:10.727773 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500.
Jul  2 00:01:10.731578 systemd[1]: Starting user@500.service - User Manager for UID 500...
Jul  2 00:01:10.742256 (systemd)[1533]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:01:10.860393 systemd[1533]: Queued start job for default target default.target.
Jul  2 00:01:10.869397 systemd[1533]: Created slice app.slice - User Application Slice.
Jul  2 00:01:10.869426 systemd[1533]: Reached target paths.target - Paths.
Jul  2 00:01:10.869439 systemd[1533]: Reached target timers.target - Timers.
Jul  2 00:01:10.870988 systemd[1533]: Starting dbus.socket - D-Bus User Message Bus Socket...
Jul  2 00:01:10.886960 systemd[1533]: Listening on dbus.socket - D-Bus User Message Bus Socket.
Jul  2 00:01:10.887362 systemd[1533]: Reached target sockets.target - Sockets.
Jul  2 00:01:10.887592 systemd[1533]: Reached target basic.target - Basic System.
Jul  2 00:01:10.887642 systemd[1533]: Reached target default.target - Main User Target.
Jul  2 00:01:10.887672 systemd[1533]: Startup finished in 127ms.
Jul  2 00:01:10.888040 systemd[1]: Started user@500.service - User Manager for UID 500.
Jul  2 00:01:10.890882 systemd[1]: Started session-1.scope - Session 1 of User core.
Jul  2 00:01:10.957131 systemd[1]: Started sshd@1-10.0.0.39:22-10.0.0.1:59368.service - OpenSSH per-connection server daemon (10.0.0.1:59368).
Jul  2 00:01:11.010405 sshd[1544]: Accepted publickey for core from 10.0.0.1 port 59368 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:01:11.011936 sshd[1544]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:01:11.017647 systemd-logind[1423]: New session 2 of user core.
Jul  2 00:01:11.031548 systemd[1]: Started session-2.scope - Session 2 of User core.
Jul  2 00:01:11.094887 sshd[1544]: pam_unix(sshd:session): session closed for user core
Jul  2 00:01:11.113551 systemd[1]: sshd@1-10.0.0.39:22-10.0.0.1:59368.service: Deactivated successfully.
Jul  2 00:01:11.115465 systemd[1]: session-2.scope: Deactivated successfully.
Jul  2 00:01:11.119532 systemd-logind[1423]: Session 2 logged out. Waiting for processes to exit.
Jul  2 00:01:11.130712 systemd[1]: Started sshd@2-10.0.0.39:22-10.0.0.1:59378.service - OpenSSH per-connection server daemon (10.0.0.1:59378).
Jul  2 00:01:11.131792 systemd-logind[1423]: Removed session 2.
Jul  2 00:01:11.173927 sshd[1551]: Accepted publickey for core from 10.0.0.1 port 59378 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:01:11.176137 sshd[1551]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:01:11.182597 systemd-logind[1423]: New session 3 of user core.
Jul  2 00:01:11.196567 systemd[1]: Started session-3.scope - Session 3 of User core.
Jul  2 00:01:11.249586 sshd[1551]: pam_unix(sshd:session): session closed for user core
Jul  2 00:01:11.263013 systemd[1]: sshd@2-10.0.0.39:22-10.0.0.1:59378.service: Deactivated successfully.
Jul  2 00:01:11.267858 systemd[1]: session-3.scope: Deactivated successfully.
Jul  2 00:01:11.272757 systemd-logind[1423]: Session 3 logged out. Waiting for processes to exit.
Jul  2 00:01:11.282973 systemd[1]: Started sshd@3-10.0.0.39:22-10.0.0.1:59388.service - OpenSSH per-connection server daemon (10.0.0.1:59388).
Jul  2 00:01:11.284594 systemd-logind[1423]: Removed session 3.
Jul  2 00:01:11.322024 sshd[1558]: Accepted publickey for core from 10.0.0.1 port 59388 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:01:11.323361 sshd[1558]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:01:11.327990 systemd-logind[1423]: New session 4 of user core.
Jul  2 00:01:11.342561 systemd[1]: Started session-4.scope - Session 4 of User core.
Jul  2 00:01:11.404611 sshd[1558]: pam_unix(sshd:session): session closed for user core
Jul  2 00:01:11.412996 systemd[1]: sshd@3-10.0.0.39:22-10.0.0.1:59388.service: Deactivated successfully.
Jul  2 00:01:11.417883 systemd[1]: session-4.scope: Deactivated successfully.
Jul  2 00:01:11.420756 systemd-logind[1423]: Session 4 logged out. Waiting for processes to exit.
Jul  2 00:01:11.429738 systemd[1]: Started sshd@4-10.0.0.39:22-10.0.0.1:59400.service - OpenSSH per-connection server daemon (10.0.0.1:59400).
Jul  2 00:01:11.432586 systemd-logind[1423]: Removed session 4.
Jul  2 00:01:11.463395 sshd[1565]: Accepted publickey for core from 10.0.0.1 port 59400 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:01:11.464803 sshd[1565]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:01:11.469656 systemd-logind[1423]: New session 5 of user core.
Jul  2 00:01:11.478531 systemd[1]: Started session-5.scope - Session 5 of User core.
Jul  2 00:01:11.575089 sudo[1568]:     core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1
Jul  2 00:01:11.582006 sudo[1568]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500)
Jul  2 00:01:11.597389 sudo[1568]: pam_unix(sudo:session): session closed for user root
Jul  2 00:01:11.602626 sshd[1565]: pam_unix(sshd:session): session closed for user core
Jul  2 00:01:11.617670 systemd[1]: sshd@4-10.0.0.39:22-10.0.0.1:59400.service: Deactivated successfully.
Jul  2 00:01:11.619715 systemd[1]: session-5.scope: Deactivated successfully.
Jul  2 00:01:11.621535 systemd-logind[1423]: Session 5 logged out. Waiting for processes to exit.
Jul  2 00:01:11.637660 systemd[1]: Started sshd@5-10.0.0.39:22-10.0.0.1:59416.service - OpenSSH per-connection server daemon (10.0.0.1:59416).
Jul  2 00:01:11.638573 systemd-logind[1423]: Removed session 5.
Jul  2 00:01:11.668806 sshd[1573]: Accepted publickey for core from 10.0.0.1 port 59416 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:01:11.671161 sshd[1573]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:01:11.675435 systemd-logind[1423]: New session 6 of user core.
Jul  2 00:01:11.682504 systemd[1]: Started session-6.scope - Session 6 of User core.
Jul  2 00:01:11.735881 sudo[1577]:     core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules
Jul  2 00:01:11.736157 sudo[1577]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500)
Jul  2 00:01:11.740259 sudo[1577]: pam_unix(sudo:session): session closed for user root
Jul  2 00:01:11.745517 sudo[1576]:     core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules
Jul  2 00:01:11.745767 sudo[1576]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500)
Jul  2 00:01:11.764864 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules...
Jul  2 00:01:11.766246 auditctl[1580]: No rules
Jul  2 00:01:11.766750 systemd[1]: audit-rules.service: Deactivated successfully.
Jul  2 00:01:11.767021 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules.
Jul  2 00:01:11.770969 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules...
Jul  2 00:01:11.803339 augenrules[1598]: No rules
Jul  2 00:01:11.805417 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules.
Jul  2 00:01:11.806631 sudo[1576]: pam_unix(sudo:session): session closed for user root
Jul  2 00:01:11.808725 sshd[1573]: pam_unix(sshd:session): session closed for user core
Jul  2 00:01:11.822010 systemd[1]: sshd@5-10.0.0.39:22-10.0.0.1:59416.service: Deactivated successfully.
Jul  2 00:01:11.824075 systemd[1]: session-6.scope: Deactivated successfully.
Jul  2 00:01:11.826429 systemd-logind[1423]: Session 6 logged out. Waiting for processes to exit.
Jul  2 00:01:11.836675 systemd[1]: Started sshd@6-10.0.0.39:22-10.0.0.1:59426.service - OpenSSH per-connection server daemon (10.0.0.1:59426).
Jul  2 00:01:11.837843 systemd-logind[1423]: Removed session 6.
Jul  2 00:01:11.872944 sshd[1606]: Accepted publickey for core from 10.0.0.1 port 59426 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:01:11.874805 sshd[1606]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:01:11.881387 systemd-logind[1423]: New session 7 of user core.
Jul  2 00:01:11.897558 systemd[1]: Started session-7.scope - Session 7 of User core.
Jul  2 00:01:11.950335 sudo[1609]:     core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh
Jul  2 00:01:11.950597 sudo[1609]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500)
Jul  2 00:01:12.079605 systemd[1]: Starting docker.service - Docker Application Container Engine...
Jul  2 00:01:12.079719 (dockerd)[1619]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU
Jul  2 00:01:12.365587 dockerd[1619]: time="2024-07-02T00:01:12.365448106Z" level=info msg="Starting up"
Jul  2 00:01:12.577331 dockerd[1619]: time="2024-07-02T00:01:12.577206283Z" level=info msg="Loading containers: start."
Jul  2 00:01:12.723348 kernel: Initializing XFRM netlink socket
Jul  2 00:01:12.815864 systemd-networkd[1368]: docker0: Link UP
Jul  2 00:01:12.841419 dockerd[1619]: time="2024-07-02T00:01:12.841364981Z" level=info msg="Loading containers: done."
Jul  2 00:01:12.928371 dockerd[1619]: time="2024-07-02T00:01:12.927976413Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2
Jul  2 00:01:12.928371 dockerd[1619]: time="2024-07-02T00:01:12.928193292Z" level=info msg="Docker daemon" commit=fca702de7f71362c8d103073c7e4a1d0a467fadd graphdriver=overlay2 version=24.0.9
Jul  2 00:01:12.928567 dockerd[1619]: time="2024-07-02T00:01:12.928405308Z" level=info msg="Daemon has completed initialization"
Jul  2 00:01:12.967174 dockerd[1619]: time="2024-07-02T00:01:12.967113008Z" level=info msg="API listen on /run/docker.sock"
Jul  2 00:01:12.967510 systemd[1]: Started docker.service - Docker Application Container Engine.
Jul  2 00:01:13.748987 containerd[1431]: time="2024-07-02T00:01:13.748898675Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.2\""
Jul  2 00:01:14.474593 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount567805693.mount: Deactivated successfully.
Jul  2 00:01:15.516484 containerd[1431]: time="2024-07-02T00:01:15.516424953Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.2\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:15.520036 containerd[1431]: time="2024-07-02T00:01:15.519984621Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.2: active requests=0, bytes read=29940432"
Jul  2 00:01:15.521063 containerd[1431]: time="2024-07-02T00:01:15.521022202Z" level=info msg="ImageCreate event name:\"sha256:84c601f3f72c87776cdcf77a73329d1f45297e43a92508b0f289fa2fcf8872a0\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:15.524110 containerd[1431]: time="2024-07-02T00:01:15.524067638Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:340ab4a1d66a60630a7a298aa0b2576fcd82e51ecdddb751cf61e5d3846fde2d\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:15.525726 containerd[1431]: time="2024-07-02T00:01:15.525521584Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.2\" with image id \"sha256:84c601f3f72c87776cdcf77a73329d1f45297e43a92508b0f289fa2fcf8872a0\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.2\", repo digest \"registry.k8s.io/kube-apiserver@sha256:340ab4a1d66a60630a7a298aa0b2576fcd82e51ecdddb751cf61e5d3846fde2d\", size \"29937230\" in 1.776574498s"
Jul  2 00:01:15.525726 containerd[1431]: time="2024-07-02T00:01:15.525561983Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.2\" returns image reference \"sha256:84c601f3f72c87776cdcf77a73329d1f45297e43a92508b0f289fa2fcf8872a0\""
Jul  2 00:01:15.545528 containerd[1431]: time="2024-07-02T00:01:15.545441750Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.2\""
Jul  2 00:01:17.220598 containerd[1431]: time="2024-07-02T00:01:17.220525930Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.2\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:17.221159 containerd[1431]: time="2024-07-02T00:01:17.221121399Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.2: active requests=0, bytes read=26881373"
Jul  2 00:01:17.222479 containerd[1431]: time="2024-07-02T00:01:17.222433639Z" level=info msg="ImageCreate event name:\"sha256:e1dcc3400d3ea6a268c7ea6e66c3a196703770a8e346b695f54344ab53a47567\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:17.225675 containerd[1431]: time="2024-07-02T00:01:17.225616942Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:4c412bc1fc585ddeba10d34a02e7507ea787ec2c57256d4c18fd230377ab048e\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:17.227313 containerd[1431]: time="2024-07-02T00:01:17.226763968Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.2\" with image id \"sha256:e1dcc3400d3ea6a268c7ea6e66c3a196703770a8e346b695f54344ab53a47567\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.2\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:4c412bc1fc585ddeba10d34a02e7507ea787ec2c57256d4c18fd230377ab048e\", size \"28368865\" in 1.681278892s"
Jul  2 00:01:17.227313 containerd[1431]: time="2024-07-02T00:01:17.226817434Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.2\" returns image reference \"sha256:e1dcc3400d3ea6a268c7ea6e66c3a196703770a8e346b695f54344ab53a47567\""
Jul  2 00:01:17.247941 containerd[1431]: time="2024-07-02T00:01:17.247872605Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.2\""
Jul  2 00:01:17.528987 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1.
Jul  2 00:01:17.539532 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jul  2 00:01:17.636238 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jul  2 00:01:17.641462 (kubelet)[1838]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Jul  2 00:01:17.761505 kubelet[1838]: E0702 00:01:17.761441    1838 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Jul  2 00:01:17.765739 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Jul  2 00:01:17.765897 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Jul  2 00:01:19.174350 containerd[1431]: time="2024-07-02T00:01:19.174262016Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.2\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:19.175186 containerd[1431]: time="2024-07-02T00:01:19.175141584Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.2: active requests=0, bytes read=16155690"
Jul  2 00:01:19.176143 containerd[1431]: time="2024-07-02T00:01:19.176115239Z" level=info msg="ImageCreate event name:\"sha256:c7dd04b1bafeb51c650fde7f34ac0fdafa96030e77ea7a822135ff302d895dd5\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:19.179664 containerd[1431]: time="2024-07-02T00:01:19.179618744Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:0ed75a333704f5d315395c6ec04d7af7405715537069b65d40b43ec1c8e030bc\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:19.181270 containerd[1431]: time="2024-07-02T00:01:19.181222846Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.2\" with image id \"sha256:c7dd04b1bafeb51c650fde7f34ac0fdafa96030e77ea7a822135ff302d895dd5\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.2\", repo digest \"registry.k8s.io/kube-scheduler@sha256:0ed75a333704f5d315395c6ec04d7af7405715537069b65d40b43ec1c8e030bc\", size \"17643200\" in 1.933130884s"
Jul  2 00:01:19.181270 containerd[1431]: time="2024-07-02T00:01:19.181268225Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.2\" returns image reference \"sha256:c7dd04b1bafeb51c650fde7f34ac0fdafa96030e77ea7a822135ff302d895dd5\""
Jul  2 00:01:19.200973 containerd[1431]: time="2024-07-02T00:01:19.200934839Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.2\""
Jul  2 00:01:22.306893 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4006099054.mount: Deactivated successfully.
Jul  2 00:01:22.577896 containerd[1431]: time="2024-07-02T00:01:22.577712312Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.2\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:22.578890 containerd[1431]: time="2024-07-02T00:01:22.578749324Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.2: active requests=0, bytes read=25634094"
Jul  2 00:01:22.579756 containerd[1431]: time="2024-07-02T00:01:22.579712751Z" level=info msg="ImageCreate event name:\"sha256:66dbb96a9149f69913ff817f696be766014cacdffc2ce0889a76c81165415fae\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:22.582136 containerd[1431]: time="2024-07-02T00:01:22.582093869Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:8a44c6e094af3dea3de57fa967e201608a358a3bd8b4e3f31ab905bbe4108aec\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:22.582834 containerd[1431]: time="2024-07-02T00:01:22.582787416Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.2\" with image id \"sha256:66dbb96a9149f69913ff817f696be766014cacdffc2ce0889a76c81165415fae\", repo tag \"registry.k8s.io/kube-proxy:v1.30.2\", repo digest \"registry.k8s.io/kube-proxy@sha256:8a44c6e094af3dea3de57fa967e201608a358a3bd8b4e3f31ab905bbe4108aec\", size \"25633111\" in 3.381810615s"
Jul  2 00:01:22.582834 containerd[1431]: time="2024-07-02T00:01:22.582826876Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.2\" returns image reference \"sha256:66dbb96a9149f69913ff817f696be766014cacdffc2ce0889a76c81165415fae\""
Jul  2 00:01:22.604367 containerd[1431]: time="2024-07-02T00:01:22.604286375Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\""
Jul  2 00:01:23.147461 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount561669151.mount: Deactivated successfully.
Jul  2 00:01:23.740611 containerd[1431]: time="2024-07-02T00:01:23.740540512Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:23.742350 containerd[1431]: time="2024-07-02T00:01:23.742269396Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485383"
Jul  2 00:01:23.744942 containerd[1431]: time="2024-07-02T00:01:23.744905863Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:23.751227 containerd[1431]: time="2024-07-02T00:01:23.751177357Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:23.752375 containerd[1431]: time="2024-07-02T00:01:23.752332004Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 1.147982192s"
Jul  2 00:01:23.752665 containerd[1431]: time="2024-07-02T00:01:23.752459225Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\""
Jul  2 00:01:23.774130 containerd[1431]: time="2024-07-02T00:01:23.774085186Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\""
Jul  2 00:01:24.234333 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1569474465.mount: Deactivated successfully.
Jul  2 00:01:24.238588 containerd[1431]: time="2024-07-02T00:01:24.238550290Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:24.239052 containerd[1431]: time="2024-07-02T00:01:24.239024059Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=268823"
Jul  2 00:01:24.239822 containerd[1431]: time="2024-07-02T00:01:24.239785345Z" level=info msg="ImageCreate event name:\"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:24.242047 containerd[1431]: time="2024-07-02T00:01:24.241991190Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:24.242881 containerd[1431]: time="2024-07-02T00:01:24.242846324Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"268051\" in 468.717356ms"
Jul  2 00:01:24.242956 containerd[1431]: time="2024-07-02T00:01:24.242884649Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\""
Jul  2 00:01:24.261302 containerd[1431]: time="2024-07-02T00:01:24.261238098Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\""
Jul  2 00:01:24.896437 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1233382258.mount: Deactivated successfully.
Jul  2 00:01:26.538018 containerd[1431]: time="2024-07-02T00:01:26.537954330Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:26.539000 containerd[1431]: time="2024-07-02T00:01:26.538749799Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=66191474"
Jul  2 00:01:26.539875 containerd[1431]: time="2024-07-02T00:01:26.539813550Z" level=info msg="ImageCreate event name:\"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:26.542664 containerd[1431]: time="2024-07-02T00:01:26.542631276Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:01:26.544976 containerd[1431]: time="2024-07-02T00:01:26.544393347Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"66189079\" in 2.283116526s"
Jul  2 00:01:26.544976 containerd[1431]: time="2024-07-02T00:01:26.544429377Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\""
Jul  2 00:01:28.016245 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2.
Jul  2 00:01:28.026557 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jul  2 00:01:28.118247 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jul  2 00:01:28.123089 (kubelet)[2061]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Jul  2 00:01:28.164055 kubelet[2061]: E0702 00:01:28.163896    2061 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Jul  2 00:01:28.166781 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Jul  2 00:01:28.166929 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Jul  2 00:01:31.463360 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Jul  2 00:01:31.473605 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jul  2 00:01:31.493614 systemd[1]: Reloading requested from client PID 2076 ('systemctl') (unit session-7.scope)...
Jul  2 00:01:31.493630 systemd[1]: Reloading...
Jul  2 00:01:31.583350 zram_generator::config[2113]: No configuration found.
Jul  2 00:01:31.714562 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jul  2 00:01:31.773183 systemd[1]: Reloading finished in 279 ms.
Jul  2 00:01:31.820794 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM
Jul  2 00:01:31.820858 systemd[1]: kubelet.service: Failed with result 'signal'.
Jul  2 00:01:31.821083 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Jul  2 00:01:31.824188 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jul  2 00:01:31.923315 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jul  2 00:01:31.927582 (kubelet)[2159]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS
Jul  2 00:01:31.974878 kubelet[2159]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jul  2 00:01:31.974878 kubelet[2159]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI.
Jul  2 00:01:31.974878 kubelet[2159]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jul  2 00:01:31.976028 kubelet[2159]: I0702 00:01:31.975971    2159 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Jul  2 00:01:32.455743 kubelet[2159]: I0702 00:01:32.455633    2159 server.go:484] "Kubelet version" kubeletVersion="v1.30.1"
Jul  2 00:01:32.455743 kubelet[2159]: I0702 00:01:32.455668    2159 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Jul  2 00:01:32.455910 kubelet[2159]: I0702 00:01:32.455891    2159 server.go:927] "Client rotation is on, will bootstrap in background"
Jul  2 00:01:32.511058 kubelet[2159]: E0702 00:01:32.511020    2159 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.39:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:32.511606 kubelet[2159]: I0702 00:01:32.511526    2159 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt"
Jul  2 00:01:32.525336 kubelet[2159]: I0702 00:01:32.525207    2159 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Jul  2 00:01:32.526389 kubelet[2159]: I0702 00:01:32.526334    2159 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Jul  2 00:01:32.526620 kubelet[2159]: I0702 00:01:32.526390    2159 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null}
Jul  2 00:01:32.526717 kubelet[2159]: I0702 00:01:32.526672    2159 topology_manager.go:138] "Creating topology manager with none policy"
Jul  2 00:01:32.526717 kubelet[2159]: I0702 00:01:32.526682    2159 container_manager_linux.go:301] "Creating device plugin manager"
Jul  2 00:01:32.526980 kubelet[2159]: I0702 00:01:32.526947    2159 state_mem.go:36] "Initialized new in-memory state store"
Jul  2 00:01:32.531718 kubelet[2159]: I0702 00:01:32.531690    2159 kubelet.go:400] "Attempting to sync node with API server"
Jul  2 00:01:32.531767 kubelet[2159]: I0702 00:01:32.531726    2159 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests"
Jul  2 00:01:32.532704 kubelet[2159]: I0702 00:01:32.531993    2159 kubelet.go:312] "Adding apiserver pod source"
Jul  2 00:01:32.532704 kubelet[2159]: I0702 00:01:32.532135    2159 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Jul  2 00:01:32.533133 kubelet[2159]: W0702 00:01:32.532919    2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.39:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:32.533133 kubelet[2159]: E0702 00:01:32.533021    2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.39:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:32.533133 kubelet[2159]: W0702 00:01:32.533058    2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.39:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:32.533133 kubelet[2159]: E0702 00:01:32.533104    2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.39:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:32.539734 kubelet[2159]: I0702 00:01:32.539700    2159 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.17" apiVersion="v1"
Jul  2 00:01:32.540320 kubelet[2159]: I0702 00:01:32.540285    2159 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode"
Jul  2 00:01:32.540475 kubelet[2159]: W0702 00:01:32.540462    2159 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating.
Jul  2 00:01:32.541547 kubelet[2159]: I0702 00:01:32.541526    2159 server.go:1264] "Started kubelet"
Jul  2 00:01:32.542488 kubelet[2159]: I0702 00:01:32.541777    2159 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10
Jul  2 00:01:32.542488 kubelet[2159]: I0702 00:01:32.542108    2159 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock"
Jul  2 00:01:32.542488 kubelet[2159]: I0702 00:01:32.542155    2159 server.go:163] "Starting to listen" address="0.0.0.0" port=10250
Jul  2 00:01:32.542619 kubelet[2159]: I0702 00:01:32.542591    2159 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Jul  2 00:01:32.543321 kubelet[2159]: I0702 00:01:32.543177    2159 server.go:455] "Adding debug handlers to kubelet server"
Jul  2 00:01:32.544945 kubelet[2159]: E0702 00:01:32.544906    2159 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found"
Jul  2 00:01:32.545578 kubelet[2159]: I0702 00:01:32.545098    2159 volume_manager.go:291] "Starting Kubelet Volume Manager"
Jul  2 00:01:32.545578 kubelet[2159]: I0702 00:01:32.545234    2159 desired_state_of_world_populator.go:149] "Desired state populator starts to run"
Jul  2 00:01:32.546277 kubelet[2159]: I0702 00:01:32.546244    2159 reconciler.go:26] "Reconciler: start to sync state"
Jul  2 00:01:32.546685 kubelet[2159]: W0702 00:01:32.546625    2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.39:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:32.546685 kubelet[2159]: E0702 00:01:32.546683    2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.39:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:32.547571 kubelet[2159]: E0702 00:01:32.547510    2159 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.39:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.39:6443: connect: connection refused" interval="200ms"
Jul  2 00:01:32.554820 kubelet[2159]: E0702 00:01:32.549391    2159 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.39:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.39:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.17de3c606cc53365  default    0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2024-07-02 00:01:32.541498213 +0000 UTC m=+0.610796401,LastTimestamp:2024-07-02 00:01:32.541498213 +0000 UTC m=+0.610796401,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}"
Jul  2 00:01:32.554820 kubelet[2159]: I0702 00:01:32.554751    2159 factory.go:221] Registration of the systemd container factory successfully
Jul  2 00:01:32.555057 kubelet[2159]: I0702 00:01:32.554843    2159 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory
Jul  2 00:01:32.556657 kubelet[2159]: E0702 00:01:32.556523    2159 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem"
Jul  2 00:01:32.556869 kubelet[2159]: I0702 00:01:32.556843    2159 factory.go:221] Registration of the containerd container factory successfully
Jul  2 00:01:32.560135 kubelet[2159]: I0702 00:01:32.559402    2159 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4"
Jul  2 00:01:32.560621 kubelet[2159]: I0702 00:01:32.560581    2159 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6"
Jul  2 00:01:32.560621 kubelet[2159]: I0702 00:01:32.560620    2159 status_manager.go:217] "Starting to sync pod status with apiserver"
Jul  2 00:01:32.560699 kubelet[2159]: I0702 00:01:32.560639    2159 kubelet.go:2337] "Starting kubelet main sync loop"
Jul  2 00:01:32.560699 kubelet[2159]: E0702 00:01:32.560689    2159 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]"
Jul  2 00:01:32.564198 kubelet[2159]: W0702 00:01:32.564149    2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.39:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:32.564678 kubelet[2159]: E0702 00:01:32.564537    2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.39:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:32.570825 kubelet[2159]: I0702 00:01:32.570784    2159 cpu_manager.go:214] "Starting CPU manager" policy="none"
Jul  2 00:01:32.570825 kubelet[2159]: I0702 00:01:32.570805    2159 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s"
Jul  2 00:01:32.570825 kubelet[2159]: I0702 00:01:32.570826    2159 state_mem.go:36] "Initialized new in-memory state store"
Jul  2 00:01:32.646745 kubelet[2159]: I0702 00:01:32.646454    2159 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jul  2 00:01:32.646856 kubelet[2159]: E0702 00:01:32.646796    2159 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.39:6443/api/v1/nodes\": dial tcp 10.0.0.39:6443: connect: connection refused" node="localhost"
Jul  2 00:01:32.660488 kubelet[2159]: I0702 00:01:32.660449    2159 policy_none.go:49] "None policy: Start"
Jul  2 00:01:32.660905 kubelet[2159]: E0702 00:01:32.660763    2159 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet"
Jul  2 00:01:32.661228 kubelet[2159]: I0702 00:01:32.661194    2159 memory_manager.go:170] "Starting memorymanager" policy="None"
Jul  2 00:01:32.661228 kubelet[2159]: I0702 00:01:32.661225    2159 state_mem.go:35] "Initializing new in-memory state store"
Jul  2 00:01:32.667357 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice.
Jul  2 00:01:32.678376 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice.
Jul  2 00:01:32.681272 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice.
Jul  2 00:01:32.692154 kubelet[2159]: I0702 00:01:32.692118    2159 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found"
Jul  2 00:01:32.692415 kubelet[2159]: I0702 00:01:32.692369    2159 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s"
Jul  2 00:01:32.692524 kubelet[2159]: I0702 00:01:32.692498    2159 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Jul  2 00:01:32.694382 kubelet[2159]: E0702 00:01:32.694348    2159 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found"
Jul  2 00:01:32.748469 kubelet[2159]: E0702 00:01:32.748343    2159 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.39:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.39:6443: connect: connection refused" interval="400ms"
Jul  2 00:01:32.848887 kubelet[2159]: I0702 00:01:32.848857    2159 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jul  2 00:01:32.849220 kubelet[2159]: E0702 00:01:32.849190    2159 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.39:6443/api/v1/nodes\": dial tcp 10.0.0.39:6443: connect: connection refused" node="localhost"
Jul  2 00:01:32.861397 kubelet[2159]: I0702 00:01:32.861325    2159 topology_manager.go:215] "Topology Admit Handler" podUID="7a00b1c456d28d9cad91da69ad03d518" podNamespace="kube-system" podName="kube-apiserver-localhost"
Jul  2 00:01:32.862462 kubelet[2159]: I0702 00:01:32.862430    2159 topology_manager.go:215] "Topology Admit Handler" podUID="fd87124bd1ab6d9b01dedf07aaa171f7" podNamespace="kube-system" podName="kube-controller-manager-localhost"
Jul  2 00:01:32.863281 kubelet[2159]: I0702 00:01:32.863252    2159 topology_manager.go:215] "Topology Admit Handler" podUID="5df30d679156d9b860331584e2d47675" podNamespace="kube-system" podName="kube-scheduler-localhost"
Jul  2 00:01:32.868684 systemd[1]: Created slice kubepods-burstable-pod7a00b1c456d28d9cad91da69ad03d518.slice - libcontainer container kubepods-burstable-pod7a00b1c456d28d9cad91da69ad03d518.slice.
Jul  2 00:01:32.889343 systemd[1]: Created slice kubepods-burstable-podfd87124bd1ab6d9b01dedf07aaa171f7.slice - libcontainer container kubepods-burstable-podfd87124bd1ab6d9b01dedf07aaa171f7.slice.
Jul  2 00:01:32.908363 systemd[1]: Created slice kubepods-burstable-pod5df30d679156d9b860331584e2d47675.slice - libcontainer container kubepods-burstable-pod5df30d679156d9b860331584e2d47675.slice.
Jul  2 00:01:32.948439 kubelet[2159]: I0702 00:01:32.948395    2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:32.948439 kubelet[2159]: I0702 00:01:32.948435    2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:32.948606 kubelet[2159]: I0702 00:01:32.948456    2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7a00b1c456d28d9cad91da69ad03d518-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"7a00b1c456d28d9cad91da69ad03d518\") " pod="kube-system/kube-apiserver-localhost"
Jul  2 00:01:32.948606 kubelet[2159]: I0702 00:01:32.948474    2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7a00b1c456d28d9cad91da69ad03d518-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"7a00b1c456d28d9cad91da69ad03d518\") " pod="kube-system/kube-apiserver-localhost"
Jul  2 00:01:32.948606 kubelet[2159]: I0702 00:01:32.948515    2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:32.948606 kubelet[2159]: I0702 00:01:32.948532    2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:32.948606 kubelet[2159]: I0702 00:01:32.948551    2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:32.948709 kubelet[2159]: I0702 00:01:32.948566    2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5df30d679156d9b860331584e2d47675-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"5df30d679156d9b860331584e2d47675\") " pod="kube-system/kube-scheduler-localhost"
Jul  2 00:01:32.948709 kubelet[2159]: I0702 00:01:32.948587    2159 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7a00b1c456d28d9cad91da69ad03d518-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"7a00b1c456d28d9cad91da69ad03d518\") " pod="kube-system/kube-apiserver-localhost"
Jul  2 00:01:33.149080 kubelet[2159]: E0702 00:01:33.148953    2159 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.39:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.39:6443: connect: connection refused" interval="800ms"
Jul  2 00:01:33.186604 kubelet[2159]: E0702 00:01:33.186548    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:33.189155 containerd[1431]: time="2024-07-02T00:01:33.188850512Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:7a00b1c456d28d9cad91da69ad03d518,Namespace:kube-system,Attempt:0,}"
Jul  2 00:01:33.206766 kubelet[2159]: E0702 00:01:33.206725    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:33.207239 containerd[1431]: time="2024-07-02T00:01:33.207197826Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:fd87124bd1ab6d9b01dedf07aaa171f7,Namespace:kube-system,Attempt:0,}"
Jul  2 00:01:33.212828 kubelet[2159]: E0702 00:01:33.212794    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:33.213228 containerd[1431]: time="2024-07-02T00:01:33.213180646Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:5df30d679156d9b860331584e2d47675,Namespace:kube-system,Attempt:0,}"
Jul  2 00:01:33.251153 kubelet[2159]: I0702 00:01:33.250883    2159 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jul  2 00:01:33.253471 kubelet[2159]: E0702 00:01:33.253425    2159 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.39:6443/api/v1/nodes\": dial tcp 10.0.0.39:6443: connect: connection refused" node="localhost"
Jul  2 00:01:33.493249 kubelet[2159]: W0702 00:01:33.493089    2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.0.0.39:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:33.493249 kubelet[2159]: E0702 00:01:33.493162    2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.39:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:33.612469 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2062818107.mount: Deactivated successfully.
Jul  2 00:01:33.620506 containerd[1431]: time="2024-07-02T00:01:33.620401001Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}  labels:{key:\"io.cri-containerd.pinned\"  value:\"pinned\"}"
Jul  2 00:01:33.622181 containerd[1431]: time="2024-07-02T00:01:33.622130265Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}  labels:{key:\"io.cri-containerd.pinned\"  value:\"pinned\"}"
Jul  2 00:01:33.623218 containerd[1431]: time="2024-07-02T00:01:33.623185734Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}  labels:{key:\"io.cri-containerd.pinned\"  value:\"pinned\"}"
Jul  2 00:01:33.624324 containerd[1431]: time="2024-07-02T00:01:33.624270119Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0"
Jul  2 00:01:33.625008 containerd[1431]: time="2024-07-02T00:01:33.624979239Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0"
Jul  2 00:01:33.628408 containerd[1431]: time="2024-07-02T00:01:33.628360032Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269175"
Jul  2 00:01:33.629235 containerd[1431]: time="2024-07-02T00:01:33.629176524Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}  labels:{key:\"io.cri-containerd.pinned\"  value:\"pinned\"}"
Jul  2 00:01:33.632669 containerd[1431]: time="2024-07-02T00:01:33.632608340Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}  labels:{key:\"io.cri-containerd.pinned\"  value:\"pinned\"}"
Jul  2 00:01:33.633594 containerd[1431]: time="2024-07-02T00:01:33.633545583Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 420.275666ms"
Jul  2 00:01:33.636373 containerd[1431]: time="2024-07-02T00:01:33.636322146Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 447.375155ms"
Jul  2 00:01:33.638251 containerd[1431]: time="2024-07-02T00:01:33.638076202Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 430.780775ms"
Jul  2 00:01:33.714373 kubelet[2159]: W0702 00:01:33.714252    2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.0.0.39:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:33.714373 kubelet[2159]: E0702 00:01:33.714342    2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.39:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:33.822671 containerd[1431]: time="2024-07-02T00:01:33.822410413Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul  2 00:01:33.823594 containerd[1431]: time="2024-07-02T00:01:33.823485106Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:33.823724 containerd[1431]: time="2024-07-02T00:01:33.823657039Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul  2 00:01:33.823879 containerd[1431]: time="2024-07-02T00:01:33.823679146Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:33.824908 containerd[1431]: time="2024-07-02T00:01:33.824827971Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul  2 00:01:33.824908 containerd[1431]: time="2024-07-02T00:01:33.824879475Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:33.825033 containerd[1431]: time="2024-07-02T00:01:33.824904987Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul  2 00:01:33.825033 containerd[1431]: time="2024-07-02T00:01:33.824916481Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:33.826284 containerd[1431]: time="2024-07-02T00:01:33.825692563Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul  2 00:01:33.826284 containerd[1431]: time="2024-07-02T00:01:33.826235196Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:33.826284 containerd[1431]: time="2024-07-02T00:01:33.826260948Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul  2 00:01:33.826499 containerd[1431]: time="2024-07-02T00:01:33.826272563Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:33.856556 systemd[1]: Started cri-containerd-07222cf47c974d13fde251a2eeab728013e9f406c619edabd0353e5d891f93ea.scope - libcontainer container 07222cf47c974d13fde251a2eeab728013e9f406c619edabd0353e5d891f93ea.
Jul  2 00:01:33.858228 systemd[1]: Started cri-containerd-724c7085449a290084f0c701ac1ec4e5b37eca189de410333ed2e465d3eb4db4.scope - libcontainer container 724c7085449a290084f0c701ac1ec4e5b37eca189de410333ed2e465d3eb4db4.
Jul  2 00:01:33.860039 systemd[1]: Started cri-containerd-d094d2328866e9bf6ac287fddf4db30568f3da0c4cdbf5f31fee82058825820a.scope - libcontainer container d094d2328866e9bf6ac287fddf4db30568f3da0c4cdbf5f31fee82058825820a.
Jul  2 00:01:33.901218 containerd[1431]: time="2024-07-02T00:01:33.900550082Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:7a00b1c456d28d9cad91da69ad03d518,Namespace:kube-system,Attempt:0,} returns sandbox id \"07222cf47c974d13fde251a2eeab728013e9f406c619edabd0353e5d891f93ea\""
Jul  2 00:01:33.903218 kubelet[2159]: E0702 00:01:33.902928    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:33.907487 containerd[1431]: time="2024-07-02T00:01:33.906228124Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:fd87124bd1ab6d9b01dedf07aaa171f7,Namespace:kube-system,Attempt:0,} returns sandbox id \"724c7085449a290084f0c701ac1ec4e5b37eca189de410333ed2e465d3eb4db4\""
Jul  2 00:01:33.907665 kubelet[2159]: E0702 00:01:33.907175    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:33.912458 containerd[1431]: time="2024-07-02T00:01:33.912273741Z" level=info msg="CreateContainer within sandbox \"07222cf47c974d13fde251a2eeab728013e9f406c619edabd0353e5d891f93ea\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Jul  2 00:01:33.912658 containerd[1431]: time="2024-07-02T00:01:33.912371382Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:5df30d679156d9b860331584e2d47675,Namespace:kube-system,Attempt:0,} returns sandbox id \"d094d2328866e9bf6ac287fddf4db30568f3da0c4cdbf5f31fee82058825820a\""
Jul  2 00:01:33.912795 containerd[1431]: time="2024-07-02T00:01:33.912770477Z" level=info msg="CreateContainer within sandbox \"724c7085449a290084f0c701ac1ec4e5b37eca189de410333ed2e465d3eb4db4\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Jul  2 00:01:33.913335 kubelet[2159]: E0702 00:01:33.913160    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:33.922708 containerd[1431]: time="2024-07-02T00:01:33.922045660Z" level=info msg="CreateContainer within sandbox \"d094d2328866e9bf6ac287fddf4db30568f3da0c4cdbf5f31fee82058825820a\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Jul  2 00:01:33.941207 containerd[1431]: time="2024-07-02T00:01:33.941087276Z" level=info msg="CreateContainer within sandbox \"724c7085449a290084f0c701ac1ec4e5b37eca189de410333ed2e465d3eb4db4\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"7e2457ba7c4501733072144fe1f5d7c391bf4c7526da6513afef1361e01402b0\""
Jul  2 00:01:33.941912 containerd[1431]: time="2024-07-02T00:01:33.941876054Z" level=info msg="StartContainer for \"7e2457ba7c4501733072144fe1f5d7c391bf4c7526da6513afef1361e01402b0\""
Jul  2 00:01:33.950365 kubelet[2159]: E0702 00:01:33.950310    2159 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.39:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.39:6443: connect: connection refused" interval="1.6s"
Jul  2 00:01:33.956279 containerd[1431]: time="2024-07-02T00:01:33.955254166Z" level=info msg="CreateContainer within sandbox \"07222cf47c974d13fde251a2eeab728013e9f406c619edabd0353e5d891f93ea\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"9f4174f2d0ba775065eadb3bbbd4292f369cf74b2d8320c54e8e54d7bff09cec\""
Jul  2 00:01:33.956279 containerd[1431]: time="2024-07-02T00:01:33.955912142Z" level=info msg="StartContainer for \"9f4174f2d0ba775065eadb3bbbd4292f369cf74b2d8320c54e8e54d7bff09cec\""
Jul  2 00:01:33.959129 containerd[1431]: time="2024-07-02T00:01:33.959065933Z" level=info msg="CreateContainer within sandbox \"d094d2328866e9bf6ac287fddf4db30568f3da0c4cdbf5f31fee82058825820a\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"07968fabdcd1b0087989325657eeed13995999a1340e6d9e83bcf8250d06a1ed\""
Jul  2 00:01:33.959893 containerd[1431]: time="2024-07-02T00:01:33.959855192Z" level=info msg="StartContainer for \"07968fabdcd1b0087989325657eeed13995999a1340e6d9e83bcf8250d06a1ed\""
Jul  2 00:01:33.971550 systemd[1]: Started cri-containerd-7e2457ba7c4501733072144fe1f5d7c391bf4c7526da6513afef1361e01402b0.scope - libcontainer container 7e2457ba7c4501733072144fe1f5d7c391bf4c7526da6513afef1361e01402b0.
Jul  2 00:01:33.994534 systemd[1]: Started cri-containerd-9f4174f2d0ba775065eadb3bbbd4292f369cf74b2d8320c54e8e54d7bff09cec.scope - libcontainer container 9f4174f2d0ba775065eadb3bbbd4292f369cf74b2d8320c54e8e54d7bff09cec.
Jul  2 00:01:33.999501 systemd[1]: Started cri-containerd-07968fabdcd1b0087989325657eeed13995999a1340e6d9e83bcf8250d06a1ed.scope - libcontainer container 07968fabdcd1b0087989325657eeed13995999a1340e6d9e83bcf8250d06a1ed.
Jul  2 00:01:34.011939 kubelet[2159]: W0702 00:01:34.011389    2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.0.0.39:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:34.011939 kubelet[2159]: E0702 00:01:34.011893    2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.39:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:34.027012 containerd[1431]: time="2024-07-02T00:01:34.026957553Z" level=info msg="StartContainer for \"7e2457ba7c4501733072144fe1f5d7c391bf4c7526da6513afef1361e01402b0\" returns successfully"
Jul  2 00:01:34.027985 kubelet[2159]: W0702 00:01:34.027709    2159 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.0.0.39:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:34.027985 kubelet[2159]: E0702 00:01:34.027898    2159 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.39:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.39:6443: connect: connection refused
Jul  2 00:01:34.058540 kubelet[2159]: I0702 00:01:34.058341    2159 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jul  2 00:01:34.058943 kubelet[2159]: E0702 00:01:34.058908    2159 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.39:6443/api/v1/nodes\": dial tcp 10.0.0.39:6443: connect: connection refused" node="localhost"
Jul  2 00:01:34.073972 containerd[1431]: time="2024-07-02T00:01:34.073827092Z" level=info msg="StartContainer for \"07968fabdcd1b0087989325657eeed13995999a1340e6d9e83bcf8250d06a1ed\" returns successfully"
Jul  2 00:01:34.073972 containerd[1431]: time="2024-07-02T00:01:34.073933576Z" level=info msg="StartContainer for \"9f4174f2d0ba775065eadb3bbbd4292f369cf74b2d8320c54e8e54d7bff09cec\" returns successfully"
Jul  2 00:01:34.572506 kubelet[2159]: E0702 00:01:34.572377    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:34.580206 kubelet[2159]: E0702 00:01:34.580172    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:34.581840 kubelet[2159]: E0702 00:01:34.581813    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:35.583344 kubelet[2159]: E0702 00:01:35.583312    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:35.661348 kubelet[2159]: I0702 00:01:35.661318    2159 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jul  2 00:01:35.837404 kubelet[2159]: E0702 00:01:35.837278    2159 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost"
Jul  2 00:01:35.930818 kubelet[2159]: I0702 00:01:35.930778    2159 kubelet_node_status.go:76] "Successfully registered node" node="localhost"
Jul  2 00:01:36.534260 kubelet[2159]: I0702 00:01:36.534177    2159 apiserver.go:52] "Watching apiserver"
Jul  2 00:01:36.546203 kubelet[2159]: I0702 00:01:36.546139    2159 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world"
Jul  2 00:01:37.389940 kubelet[2159]: E0702 00:01:37.389896    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:37.585502 kubelet[2159]: E0702 00:01:37.585440    2159 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:37.890464 systemd[1]: Reloading requested from client PID 2439 ('systemctl') (unit session-7.scope)...
Jul  2 00:01:37.890779 systemd[1]: Reloading...
Jul  2 00:01:37.971330 zram_generator::config[2476]: No configuration found.
Jul  2 00:01:38.066815 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jul  2 00:01:38.139628 systemd[1]: Reloading finished in 248 ms.
Jul  2 00:01:38.181009 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent...
Jul  2 00:01:38.198377 systemd[1]: kubelet.service: Deactivated successfully.
Jul  2 00:01:38.198657 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Jul  2 00:01:38.198804 systemd[1]: kubelet.service: Consumed 1.065s CPU time, 116.5M memory peak, 0B memory swap peak.
Jul  2 00:01:38.212585 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jul  2 00:01:38.314350 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jul  2 00:01:38.319646 (kubelet)[2518]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS
Jul  2 00:01:38.370668 kubelet[2518]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jul  2 00:01:38.370668 kubelet[2518]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI.
Jul  2 00:01:38.370668 kubelet[2518]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jul  2 00:01:38.371135 kubelet[2518]: I0702 00:01:38.370688    2518 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Jul  2 00:01:38.375741 kubelet[2518]: I0702 00:01:38.375704    2518 server.go:484] "Kubelet version" kubeletVersion="v1.30.1"
Jul  2 00:01:38.375741 kubelet[2518]: I0702 00:01:38.375738    2518 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Jul  2 00:01:38.375975 kubelet[2518]: I0702 00:01:38.375955    2518 server.go:927] "Client rotation is on, will bootstrap in background"
Jul  2 00:01:38.377476 kubelet[2518]: I0702 00:01:38.377450    2518 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem".
Jul  2 00:01:38.378720 kubelet[2518]: I0702 00:01:38.378687    2518 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt"
Jul  2 00:01:38.384670 kubelet[2518]: I0702 00:01:38.384630    2518 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Jul  2 00:01:38.384993 kubelet[2518]: I0702 00:01:38.384952    2518 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Jul  2 00:01:38.387315 kubelet[2518]: I0702 00:01:38.384995    2518 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"localhost","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null}
Jul  2 00:01:38.387315 kubelet[2518]: I0702 00:01:38.387318    2518 topology_manager.go:138] "Creating topology manager with none policy"
Jul  2 00:01:38.387519 kubelet[2518]: I0702 00:01:38.387332    2518 container_manager_linux.go:301] "Creating device plugin manager"
Jul  2 00:01:38.387519 kubelet[2518]: I0702 00:01:38.387400    2518 state_mem.go:36] "Initialized new in-memory state store"
Jul  2 00:01:38.387639 kubelet[2518]: I0702 00:01:38.387618    2518 kubelet.go:400] "Attempting to sync node with API server"
Jul  2 00:01:38.387691 kubelet[2518]: I0702 00:01:38.387641    2518 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests"
Jul  2 00:01:38.387691 kubelet[2518]: I0702 00:01:38.387671    2518 kubelet.go:312] "Adding apiserver pod source"
Jul  2 00:01:38.387691 kubelet[2518]: I0702 00:01:38.387682    2518 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Jul  2 00:01:38.388981 kubelet[2518]: I0702 00:01:38.388825    2518 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.17" apiVersion="v1"
Jul  2 00:01:38.391507 kubelet[2518]: I0702 00:01:38.391479    2518 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode"
Jul  2 00:01:38.392653 kubelet[2518]: I0702 00:01:38.391959    2518 server.go:1264] "Started kubelet"
Jul  2 00:01:38.392653 kubelet[2518]: I0702 00:01:38.392055    2518 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10
Jul  2 00:01:38.392653 kubelet[2518]: I0702 00:01:38.392403    2518 server.go:163] "Starting to listen" address="0.0.0.0" port=10250
Jul  2 00:01:38.392653 kubelet[2518]: I0702 00:01:38.392426    2518 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock"
Jul  2 00:01:38.393000 kubelet[2518]: I0702 00:01:38.392980    2518 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Jul  2 00:01:38.393349 kubelet[2518]: I0702 00:01:38.393323    2518 server.go:455] "Adding debug handlers to kubelet server"
Jul  2 00:01:38.394549 kubelet[2518]: I0702 00:01:38.394531    2518 volume_manager.go:291] "Starting Kubelet Volume Manager"
Jul  2 00:01:38.394898 kubelet[2518]: I0702 00:01:38.394877    2518 desired_state_of_world_populator.go:149] "Desired state populator starts to run"
Jul  2 00:01:38.395182 kubelet[2518]: I0702 00:01:38.395138    2518 reconciler.go:26] "Reconciler: start to sync state"
Jul  2 00:01:38.409708 kubelet[2518]: E0702 00:01:38.409496    2518 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem"
Jul  2 00:01:38.409856 kubelet[2518]: I0702 00:01:38.409825    2518 factory.go:221] Registration of the systemd container factory successfully
Jul  2 00:01:38.409965 kubelet[2518]: I0702 00:01:38.409942    2518 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory
Jul  2 00:01:38.412973 kubelet[2518]: I0702 00:01:38.412900    2518 factory.go:221] Registration of the containerd container factory successfully
Jul  2 00:01:38.418673 kubelet[2518]: I0702 00:01:38.418628    2518 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4"
Jul  2 00:01:38.421450 kubelet[2518]: I0702 00:01:38.421408    2518 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6"
Jul  2 00:01:38.421621 kubelet[2518]: I0702 00:01:38.421470    2518 status_manager.go:217] "Starting to sync pod status with apiserver"
Jul  2 00:01:38.421621 kubelet[2518]: I0702 00:01:38.421491    2518 kubelet.go:2337] "Starting kubelet main sync loop"
Jul  2 00:01:38.421621 kubelet[2518]: E0702 00:01:38.421542    2518 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]"
Jul  2 00:01:38.451554 kubelet[2518]: I0702 00:01:38.451438    2518 cpu_manager.go:214] "Starting CPU manager" policy="none"
Jul  2 00:01:38.451554 kubelet[2518]: I0702 00:01:38.451459    2518 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s"
Jul  2 00:01:38.451554 kubelet[2518]: I0702 00:01:38.451483    2518 state_mem.go:36] "Initialized new in-memory state store"
Jul  2 00:01:38.451705 kubelet[2518]: I0702 00:01:38.451684    2518 state_mem.go:88] "Updated default CPUSet" cpuSet=""
Jul  2 00:01:38.451734 kubelet[2518]: I0702 00:01:38.451697    2518 state_mem.go:96] "Updated CPUSet assignments" assignments={}
Jul  2 00:01:38.451734 kubelet[2518]: I0702 00:01:38.451718    2518 policy_none.go:49] "None policy: Start"
Jul  2 00:01:38.452358 kubelet[2518]: I0702 00:01:38.452339    2518 memory_manager.go:170] "Starting memorymanager" policy="None"
Jul  2 00:01:38.452442 kubelet[2518]: I0702 00:01:38.452365    2518 state_mem.go:35] "Initializing new in-memory state store"
Jul  2 00:01:38.452598 kubelet[2518]: I0702 00:01:38.452558    2518 state_mem.go:75] "Updated machine memory state"
Jul  2 00:01:38.456740 kubelet[2518]: I0702 00:01:38.456702    2518 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found"
Jul  2 00:01:38.456941 kubelet[2518]: I0702 00:01:38.456886    2518 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s"
Jul  2 00:01:38.457030 kubelet[2518]: I0702 00:01:38.457018    2518 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Jul  2 00:01:38.498220 kubelet[2518]: I0702 00:01:38.498162    2518 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jul  2 00:01:38.512240 kubelet[2518]: I0702 00:01:38.512184    2518 kubelet_node_status.go:112] "Node was previously registered" node="localhost"
Jul  2 00:01:38.512384 kubelet[2518]: I0702 00:01:38.512282    2518 kubelet_node_status.go:76] "Successfully registered node" node="localhost"
Jul  2 00:01:38.521933 kubelet[2518]: I0702 00:01:38.521879    2518 topology_manager.go:215] "Topology Admit Handler" podUID="fd87124bd1ab6d9b01dedf07aaa171f7" podNamespace="kube-system" podName="kube-controller-manager-localhost"
Jul  2 00:01:38.522048 kubelet[2518]: I0702 00:01:38.522028    2518 topology_manager.go:215] "Topology Admit Handler" podUID="5df30d679156d9b860331584e2d47675" podNamespace="kube-system" podName="kube-scheduler-localhost"
Jul  2 00:01:38.522242 kubelet[2518]: I0702 00:01:38.522199    2518 topology_manager.go:215] "Topology Admit Handler" podUID="7a00b1c456d28d9cad91da69ad03d518" podNamespace="kube-system" podName="kube-apiserver-localhost"
Jul  2 00:01:38.529155 kubelet[2518]: E0702 00:01:38.529099    2518 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:38.596397 kubelet[2518]: I0702 00:01:38.596349    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:38.596397 kubelet[2518]: I0702 00:01:38.596399    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/7a00b1c456d28d9cad91da69ad03d518-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"7a00b1c456d28d9cad91da69ad03d518\") " pod="kube-system/kube-apiserver-localhost"
Jul  2 00:01:38.596580 kubelet[2518]: I0702 00:01:38.596424    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/7a00b1c456d28d9cad91da69ad03d518-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"7a00b1c456d28d9cad91da69ad03d518\") " pod="kube-system/kube-apiserver-localhost"
Jul  2 00:01:38.596580 kubelet[2518]: I0702 00:01:38.596447    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/7a00b1c456d28d9cad91da69ad03d518-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"7a00b1c456d28d9cad91da69ad03d518\") " pod="kube-system/kube-apiserver-localhost"
Jul  2 00:01:38.596580 kubelet[2518]: I0702 00:01:38.596479    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:38.596580 kubelet[2518]: I0702 00:01:38.596507    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:38.596580 kubelet[2518]: I0702 00:01:38.596525    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:38.596695 kubelet[2518]: I0702 00:01:38.596555    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/fd87124bd1ab6d9b01dedf07aaa171f7-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"fd87124bd1ab6d9b01dedf07aaa171f7\") " pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:38.596695 kubelet[2518]: I0702 00:01:38.596576    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/5df30d679156d9b860331584e2d47675-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"5df30d679156d9b860331584e2d47675\") " pod="kube-system/kube-scheduler-localhost"
Jul  2 00:01:38.830345 kubelet[2518]: E0702 00:01:38.829964    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:38.830345 kubelet[2518]: E0702 00:01:38.830077    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:38.830345 kubelet[2518]: E0702 00:01:38.830139    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:38.891194 sudo[2552]:     root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin
Jul  2 00:01:38.891473 sudo[2552]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0)
Jul  2 00:01:39.344056 sudo[2552]: pam_unix(sudo:session): session closed for user root
Jul  2 00:01:39.389425 kubelet[2518]: I0702 00:01:39.389370    2518 apiserver.go:52] "Watching apiserver"
Jul  2 00:01:39.446606 kubelet[2518]: E0702 00:01:39.446558    2518 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost"
Jul  2 00:01:39.446749 kubelet[2518]: E0702 00:01:39.446732    2518 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost"
Jul  2 00:01:39.447144 kubelet[2518]: E0702 00:01:39.447110    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:39.447191 kubelet[2518]: E0702 00:01:39.447117    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:39.450182 kubelet[2518]: E0702 00:01:39.450135    2518 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost"
Jul  2 00:01:39.450559 kubelet[2518]: E0702 00:01:39.450529    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:39.478506 kubelet[2518]: I0702 00:01:39.478435    2518 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=2.478416648 podStartE2EDuration="2.478416648s" podCreationTimestamp="2024-07-02 00:01:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 00:01:39.471378519 +0000 UTC m=+1.147795173" watchObservedRunningTime="2024-07-02 00:01:39.478416648 +0000 UTC m=+1.154833302"
Jul  2 00:01:39.478689 kubelet[2518]: I0702 00:01:39.478557    2518 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.478552082 podStartE2EDuration="1.478552082s" podCreationTimestamp="2024-07-02 00:01:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 00:01:39.47839555 +0000 UTC m=+1.154812204" watchObservedRunningTime="2024-07-02 00:01:39.478552082 +0000 UTC m=+1.154968696"
Jul  2 00:01:39.489219 kubelet[2518]: I0702 00:01:39.489146    2518 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.489125789 podStartE2EDuration="1.489125789s" podCreationTimestamp="2024-07-02 00:01:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 00:01:39.488410947 +0000 UTC m=+1.164827601" watchObservedRunningTime="2024-07-02 00:01:39.489125789 +0000 UTC m=+1.165542443"
Jul  2 00:01:39.495418 kubelet[2518]: I0702 00:01:39.495370    2518 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world"
Jul  2 00:01:40.437952 kubelet[2518]: E0702 00:01:40.437812    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:40.437952 kubelet[2518]: E0702 00:01:40.437867    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:40.439038 kubelet[2518]: E0702 00:01:40.438995    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:40.928098 sudo[1609]: pam_unix(sudo:session): session closed for user root
Jul  2 00:01:40.931152 sshd[1606]: pam_unix(sshd:session): session closed for user core
Jul  2 00:01:40.935182 systemd[1]: sshd@6-10.0.0.39:22-10.0.0.1:59426.service: Deactivated successfully.
Jul  2 00:01:40.937112 systemd[1]: session-7.scope: Deactivated successfully.
Jul  2 00:01:40.937336 systemd[1]: session-7.scope: Consumed 7.440s CPU time, 139.3M memory peak, 0B memory swap peak.
Jul  2 00:01:40.939159 systemd-logind[1423]: Session 7 logged out. Waiting for processes to exit.
Jul  2 00:01:40.940682 systemd-logind[1423]: Removed session 7.
Jul  2 00:01:41.439742 kubelet[2518]: E0702 00:01:41.439493    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:43.195994 kubelet[2518]: E0702 00:01:43.195921    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:45.487903 kubelet[2518]: E0702 00:01:45.487864    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:46.447357 kubelet[2518]: E0702 00:01:46.447327    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:50.053512 update_engine[1425]: I0702 00:01:50.053447  1425 update_attempter.cc:509] Updating boot flags...
Jul  2 00:01:50.089334 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2599)
Jul  2 00:01:50.114398 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 38 scanned by (udev-worker) (2598)
Jul  2 00:01:50.634250 kubelet[2518]: E0702 00:01:50.634210    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:52.995208 kubelet[2518]: I0702 00:01:52.994380    2518 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24"
Jul  2 00:01:52.995208 kubelet[2518]: I0702 00:01:52.994904    2518 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24"
Jul  2 00:01:52.995658 containerd[1431]: time="2024-07-02T00:01:52.994701063Z" level=info msg="No cni config template is specified, wait for other system components to drop the config."
Jul  2 00:01:53.208635 kubelet[2518]: E0702 00:01:53.208603    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:54.031860 kubelet[2518]: I0702 00:01:54.031799    2518 topology_manager.go:215] "Topology Admit Handler" podUID="1b7e26a8-aa49-4041-b853-71b7189b25fa" podNamespace="kube-system" podName="kube-proxy-7cl5g"
Jul  2 00:01:54.038269 kubelet[2518]: I0702 00:01:54.038196    2518 topology_manager.go:215] "Topology Admit Handler" podUID="1f20980c-f700-4772-9986-b99108fa3c3d" podNamespace="kube-system" podName="cilium-r2tjl"
Jul  2 00:01:54.047411 systemd[1]: Created slice kubepods-besteffort-pod1b7e26a8_aa49_4041_b853_71b7189b25fa.slice - libcontainer container kubepods-besteffort-pod1b7e26a8_aa49_4041_b853_71b7189b25fa.slice.
Jul  2 00:01:54.064088 systemd[1]: Created slice kubepods-burstable-pod1f20980c_f700_4772_9986_b99108fa3c3d.slice - libcontainer container kubepods-burstable-pod1f20980c_f700_4772_9986_b99108fa3c3d.slice.
Jul  2 00:01:54.075752 kubelet[2518]: I0702 00:01:54.074676    2518 topology_manager.go:215] "Topology Admit Handler" podUID="5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33" podNamespace="kube-system" podName="cilium-operator-599987898-xch4f"
Jul  2 00:01:54.083968 systemd[1]: Created slice kubepods-besteffort-pod5be9b8b1_bfbc_4ece_9980_31aa7ff1bf33.slice - libcontainer container kubepods-besteffort-pod5be9b8b1_bfbc_4ece_9980_31aa7ff1bf33.slice.
Jul  2 00:01:54.094522 kubelet[2518]: I0702 00:01:54.093419    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-run\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.094522 kubelet[2518]: I0702 00:01:54.093463    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpgbd\" (UniqueName: \"kubernetes.io/projected/1f20980c-f700-4772-9986-b99108fa3c3d-kube-api-access-qpgbd\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.094522 kubelet[2518]: I0702 00:01:54.093486    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqbbj\" (UniqueName: \"kubernetes.io/projected/5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33-kube-api-access-fqbbj\") pod \"cilium-operator-599987898-xch4f\" (UID: \"5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33\") " pod="kube-system/cilium-operator-599987898-xch4f"
Jul  2 00:01:54.094522 kubelet[2518]: I0702 00:01:54.093503    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1f20980c-f700-4772-9986-b99108fa3c3d-clustermesh-secrets\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.094522 kubelet[2518]: I0702 00:01:54.093520    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-host-proc-sys-net\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095047 kubelet[2518]: I0702 00:01:54.093536    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33-cilium-config-path\") pod \"cilium-operator-599987898-xch4f\" (UID: \"5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33\") " pod="kube-system/cilium-operator-599987898-xch4f"
Jul  2 00:01:54.095047 kubelet[2518]: I0702 00:01:54.093552    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1b7e26a8-aa49-4041-b853-71b7189b25fa-lib-modules\") pod \"kube-proxy-7cl5g\" (UID: \"1b7e26a8-aa49-4041-b853-71b7189b25fa\") " pod="kube-system/kube-proxy-7cl5g"
Jul  2 00:01:54.095047 kubelet[2518]: I0702 00:01:54.093566    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-bpf-maps\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095047 kubelet[2518]: I0702 00:01:54.093580    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-etc-cni-netd\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095047 kubelet[2518]: I0702 00:01:54.093597    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-lib-modules\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095209 kubelet[2518]: I0702 00:01:54.093611    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-xtables-lock\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095209 kubelet[2518]: I0702 00:01:54.093630    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jg6q\" (UniqueName: \"kubernetes.io/projected/1b7e26a8-aa49-4041-b853-71b7189b25fa-kube-api-access-5jg6q\") pod \"kube-proxy-7cl5g\" (UID: \"1b7e26a8-aa49-4041-b853-71b7189b25fa\") " pod="kube-system/kube-proxy-7cl5g"
Jul  2 00:01:54.095209 kubelet[2518]: I0702 00:01:54.093646    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-host-proc-sys-kernel\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095209 kubelet[2518]: I0702 00:01:54.093662    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1f20980c-f700-4772-9986-b99108fa3c3d-hubble-tls\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095209 kubelet[2518]: I0702 00:01:54.093726    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-config-path\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095397 kubelet[2518]: I0702 00:01:54.093744    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1b7e26a8-aa49-4041-b853-71b7189b25fa-xtables-lock\") pod \"kube-proxy-7cl5g\" (UID: \"1b7e26a8-aa49-4041-b853-71b7189b25fa\") " pod="kube-system/kube-proxy-7cl5g"
Jul  2 00:01:54.095397 kubelet[2518]: I0702 00:01:54.093760    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-hostproc\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095397 kubelet[2518]: I0702 00:01:54.093779    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-cgroup\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.095397 kubelet[2518]: I0702 00:01:54.093795    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/1b7e26a8-aa49-4041-b853-71b7189b25fa-kube-proxy\") pod \"kube-proxy-7cl5g\" (UID: \"1b7e26a8-aa49-4041-b853-71b7189b25fa\") " pod="kube-system/kube-proxy-7cl5g"
Jul  2 00:01:54.095397 kubelet[2518]: I0702 00:01:54.093809    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cni-path\") pod \"cilium-r2tjl\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") " pod="kube-system/cilium-r2tjl"
Jul  2 00:01:54.358859 kubelet[2518]: E0702 00:01:54.358739    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:54.359553 containerd[1431]: time="2024-07-02T00:01:54.359515528Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-7cl5g,Uid:1b7e26a8-aa49-4041-b853-71b7189b25fa,Namespace:kube-system,Attempt:0,}"
Jul  2 00:01:54.368647 kubelet[2518]: E0702 00:01:54.368603    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:54.369248 containerd[1431]: time="2024-07-02T00:01:54.369022651Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-r2tjl,Uid:1f20980c-f700-4772-9986-b99108fa3c3d,Namespace:kube-system,Attempt:0,}"
Jul  2 00:01:54.384470 containerd[1431]: time="2024-07-02T00:01:54.382632808Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul  2 00:01:54.384470 containerd[1431]: time="2024-07-02T00:01:54.382763209Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:54.384470 containerd[1431]: time="2024-07-02T00:01:54.382790858Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul  2 00:01:54.384470 containerd[1431]: time="2024-07-02T00:01:54.382806263Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:54.390028 kubelet[2518]: E0702 00:01:54.389735    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:54.392716 containerd[1431]: time="2024-07-02T00:01:54.392367884Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-xch4f,Uid:5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33,Namespace:kube-system,Attempt:0,}"
Jul  2 00:01:54.406900 containerd[1431]: time="2024-07-02T00:01:54.406589436Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul  2 00:01:54.406900 containerd[1431]: time="2024-07-02T00:01:54.406654577Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:54.406900 containerd[1431]: time="2024-07-02T00:01:54.406668942Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul  2 00:01:54.406900 containerd[1431]: time="2024-07-02T00:01:54.406679265Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:54.415557 systemd[1]: Started cri-containerd-6d7359ae8da86841a0228fcd585a742d6169587cd1fe62604544ec3b07fc2810.scope - libcontainer container 6d7359ae8da86841a0228fcd585a742d6169587cd1fe62604544ec3b07fc2810.
Jul  2 00:01:54.425445 containerd[1431]: time="2024-07-02T00:01:54.424686829Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul  2 00:01:54.425445 containerd[1431]: time="2024-07-02T00:01:54.424779739Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:54.425445 containerd[1431]: time="2024-07-02T00:01:54.424800305Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul  2 00:01:54.425445 containerd[1431]: time="2024-07-02T00:01:54.424815790Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:01:54.440535 systemd[1]: Started cri-containerd-4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7.scope - libcontainer container 4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7.
Jul  2 00:01:54.445676 containerd[1431]: time="2024-07-02T00:01:54.445562952Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-7cl5g,Uid:1b7e26a8-aa49-4041-b853-71b7189b25fa,Namespace:kube-system,Attempt:0,} returns sandbox id \"6d7359ae8da86841a0228fcd585a742d6169587cd1fe62604544ec3b07fc2810\""
Jul  2 00:01:54.447085 kubelet[2518]: E0702 00:01:54.447055    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:54.447142 systemd[1]: Started cri-containerd-e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040.scope - libcontainer container e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040.
Jul  2 00:01:54.450953 containerd[1431]: time="2024-07-02T00:01:54.450839321Z" level=info msg="CreateContainer within sandbox \"6d7359ae8da86841a0228fcd585a742d6169587cd1fe62604544ec3b07fc2810\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}"
Jul  2 00:01:54.482636 containerd[1431]: time="2024-07-02T00:01:54.482555273Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-r2tjl,Uid:1f20980c-f700-4772-9986-b99108fa3c3d,Namespace:kube-system,Attempt:0,} returns sandbox id \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\""
Jul  2 00:01:54.484280 containerd[1431]: time="2024-07-02T00:01:54.483561435Z" level=info msg="CreateContainer within sandbox \"6d7359ae8da86841a0228fcd585a742d6169587cd1fe62604544ec3b07fc2810\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"4b20ad4563b2f0a27484c94e274743de384949a865dc4911d18cb9eec76b2cab\""
Jul  2 00:01:54.484424 kubelet[2518]: E0702 00:01:54.483686    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:54.484483 containerd[1431]: time="2024-07-02T00:01:54.484442637Z" level=info msg="StartContainer for \"4b20ad4563b2f0a27484c94e274743de384949a865dc4911d18cb9eec76b2cab\""
Jul  2 00:01:54.484712 containerd[1431]: time="2024-07-02T00:01:54.484672671Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\""
Jul  2 00:01:54.497636 containerd[1431]: time="2024-07-02T00:01:54.497598208Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-599987898-xch4f,Uid:5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33,Namespace:kube-system,Attempt:0,} returns sandbox id \"e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040\""
Jul  2 00:01:54.498955 kubelet[2518]: E0702 00:01:54.498930    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:54.517518 systemd[1]: Started cri-containerd-4b20ad4563b2f0a27484c94e274743de384949a865dc4911d18cb9eec76b2cab.scope - libcontainer container 4b20ad4563b2f0a27484c94e274743de384949a865dc4911d18cb9eec76b2cab.
Jul  2 00:01:54.551460 containerd[1431]: time="2024-07-02T00:01:54.551410954Z" level=info msg="StartContainer for \"4b20ad4563b2f0a27484c94e274743de384949a865dc4911d18cb9eec76b2cab\" returns successfully"
Jul  2 00:01:55.475066 kubelet[2518]: E0702 00:01:55.475014    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:01:55.486636 kubelet[2518]: I0702 00:01:55.486530    2518 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-7cl5g" podStartSLOduration=1.486493924 podStartE2EDuration="1.486493924s" podCreationTimestamp="2024-07-02 00:01:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 00:01:55.486222642 +0000 UTC m=+17.162639296" watchObservedRunningTime="2024-07-02 00:01:55.486493924 +0000 UTC m=+17.162910578"
Jul  2 00:01:56.482014 kubelet[2518]: E0702 00:01:56.481981    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:01.214817 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3371340922.mount: Deactivated successfully.
Jul  2 00:02:02.716689 containerd[1431]: time="2024-07-02T00:02:02.716630092Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:02:02.717945 containerd[1431]: time="2024-07-02T00:02:02.717701536Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157651558"
Jul  2 00:02:02.718932 containerd[1431]: time="2024-07-02T00:02:02.718868199Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:02:02.721132 containerd[1431]: time="2024-07-02T00:02:02.720995766Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 8.236277002s"
Jul  2 00:02:02.721132 containerd[1431]: time="2024-07-02T00:02:02.721042695Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\""
Jul  2 00:02:02.729118 containerd[1431]: time="2024-07-02T00:02:02.729073349Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\""
Jul  2 00:02:02.736796 containerd[1431]: time="2024-07-02T00:02:02.736623791Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}"
Jul  2 00:02:02.756696 containerd[1431]: time="2024-07-02T00:02:02.756644096Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\""
Jul  2 00:02:02.758820 containerd[1431]: time="2024-07-02T00:02:02.757635165Z" level=info msg="StartContainer for \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\""
Jul  2 00:02:02.782145 systemd[1]: run-containerd-runc-k8s.io-776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303-runc.hZw6It.mount: Deactivated successfully.
Jul  2 00:02:02.795535 systemd[1]: Started cri-containerd-776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303.scope - libcontainer container 776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303.
Jul  2 00:02:02.899560 systemd[1]: cri-containerd-776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303.scope: Deactivated successfully.
Jul  2 00:02:02.921049 containerd[1431]: time="2024-07-02T00:02:02.920907756Z" level=info msg="StartContainer for \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\" returns successfully"
Jul  2 00:02:02.961919 containerd[1431]: time="2024-07-02T00:02:02.961807289Z" level=info msg="shim disconnected" id=776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303 namespace=k8s.io
Jul  2 00:02:02.961919 containerd[1431]: time="2024-07-02T00:02:02.961912509Z" level=warning msg="cleaning up after shim disconnected" id=776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303 namespace=k8s.io
Jul  2 00:02:02.962210 containerd[1431]: time="2024-07-02T00:02:02.961929633Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:02:03.503246 kubelet[2518]: E0702 00:02:03.503199    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:03.506802 containerd[1431]: time="2024-07-02T00:02:03.506624779Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}"
Jul  2 00:02:03.531843 containerd[1431]: time="2024-07-02T00:02:03.531744558Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\""
Jul  2 00:02:03.534401 containerd[1431]: time="2024-07-02T00:02:03.534320619Z" level=info msg="StartContainer for \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\""
Jul  2 00:02:03.565513 systemd[1]: Started cri-containerd-ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14.scope - libcontainer container ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14.
Jul  2 00:02:03.623606 containerd[1431]: time="2024-07-02T00:02:03.623532277Z" level=info msg="StartContainer for \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\" returns successfully"
Jul  2 00:02:03.630968 systemd[1]: systemd-sysctl.service: Deactivated successfully.
Jul  2 00:02:03.631491 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables.
Jul  2 00:02:03.631704 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables...
Jul  2 00:02:03.638684 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Jul  2 00:02:03.639731 systemd[1]: cri-containerd-ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14.scope: Deactivated successfully.
Jul  2 00:02:03.684160 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Jul  2 00:02:03.751180 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303-rootfs.mount: Deactivated successfully.
Jul  2 00:02:03.796675 containerd[1431]: time="2024-07-02T00:02:03.796373992Z" level=info msg="shim disconnected" id=ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14 namespace=k8s.io
Jul  2 00:02:03.796675 containerd[1431]: time="2024-07-02T00:02:03.796433323Z" level=warning msg="cleaning up after shim disconnected" id=ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14 namespace=k8s.io
Jul  2 00:02:03.796675 containerd[1431]: time="2024-07-02T00:02:03.796444445Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:02:04.243280 containerd[1431]: time="2024-07-02T00:02:04.243185348Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:02:04.243760 containerd[1431]: time="2024-07-02T00:02:04.243708316Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17138282"
Jul  2 00:02:04.247520 containerd[1431]: time="2024-07-02T00:02:04.247465707Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"  labels:{key:\"io.cri-containerd.image\"  value:\"managed\"}"
Jul  2 00:02:04.249236 containerd[1431]: time="2024-07-02T00:02:04.248879544Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 1.519760587s"
Jul  2 00:02:04.249236 containerd[1431]: time="2024-07-02T00:02:04.248928752Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\""
Jul  2 00:02:04.252817 containerd[1431]: time="2024-07-02T00:02:04.252767037Z" level=info msg="CreateContainer within sandbox \"e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}"
Jul  2 00:02:04.264446 containerd[1431]: time="2024-07-02T00:02:04.264395909Z" level=info msg="CreateContainer within sandbox \"e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\""
Jul  2 00:02:04.264883 containerd[1431]: time="2024-07-02T00:02:04.264863188Z" level=info msg="StartContainer for \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\""
Jul  2 00:02:04.299525 systemd[1]: Started cri-containerd-b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4.scope - libcontainer container b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4.
Jul  2 00:02:04.319450 containerd[1431]: time="2024-07-02T00:02:04.319403986Z" level=info msg="StartContainer for \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\" returns successfully"
Jul  2 00:02:04.508204 kubelet[2518]: E0702 00:02:04.508083    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:04.513642 kubelet[2518]: E0702 00:02:04.513603    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:04.515730 containerd[1431]: time="2024-07-02T00:02:04.515678541Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}"
Jul  2 00:02:04.562562 kubelet[2518]: I0702 00:02:04.562476    2518 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-operator-599987898-xch4f" podStartSLOduration=0.812083258 podStartE2EDuration="10.562335816s" podCreationTimestamp="2024-07-02 00:01:54 +0000 UTC" firstStartedPulling="2024-07-02 00:01:54.499472048 +0000 UTC m=+16.175888702" lastFinishedPulling="2024-07-02 00:02:04.249724646 +0000 UTC m=+25.926141260" observedRunningTime="2024-07-02 00:02:04.533924445 +0000 UTC m=+26.210341059" watchObservedRunningTime="2024-07-02 00:02:04.562335816 +0000 UTC m=+26.238752470"
Jul  2 00:02:04.597215 containerd[1431]: time="2024-07-02T00:02:04.597133938Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\""
Jul  2 00:02:04.598246 containerd[1431]: time="2024-07-02T00:02:04.598135026Z" level=info msg="StartContainer for \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\""
Jul  2 00:02:04.633698 systemd[1]: Started cri-containerd-0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783.scope - libcontainer container 0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783.
Jul  2 00:02:04.667544 containerd[1431]: time="2024-07-02T00:02:04.667459066Z" level=info msg="StartContainer for \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\" returns successfully"
Jul  2 00:02:04.683059 systemd[1]: cri-containerd-0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783.scope: Deactivated successfully.
Jul  2 00:02:04.738794 containerd[1431]: time="2024-07-02T00:02:04.738708510Z" level=info msg="shim disconnected" id=0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783 namespace=k8s.io
Jul  2 00:02:04.738794 containerd[1431]: time="2024-07-02T00:02:04.738786203Z" level=warning msg="cleaning up after shim disconnected" id=0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783 namespace=k8s.io
Jul  2 00:02:04.738794 containerd[1431]: time="2024-07-02T00:02:04.738795804Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:02:05.516899 kubelet[2518]: E0702 00:02:05.516847    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:05.517649 kubelet[2518]: E0702 00:02:05.517624    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:05.520467 containerd[1431]: time="2024-07-02T00:02:05.520391862Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}"
Jul  2 00:02:05.538081 containerd[1431]: time="2024-07-02T00:02:05.538025318Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\""
Jul  2 00:02:05.539751 containerd[1431]: time="2024-07-02T00:02:05.538753393Z" level=info msg="StartContainer for \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\""
Jul  2 00:02:05.567517 systemd[1]: Started cri-containerd-1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803.scope - libcontainer container 1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803.
Jul  2 00:02:05.588660 systemd[1]: cri-containerd-1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803.scope: Deactivated successfully.
Jul  2 00:02:05.591479 containerd[1431]: time="2024-07-02T00:02:05.591434405Z" level=info msg="StartContainer for \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\" returns successfully"
Jul  2 00:02:05.610250 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803-rootfs.mount: Deactivated successfully.
Jul  2 00:02:05.617825 containerd[1431]: time="2024-07-02T00:02:05.617767390Z" level=info msg="shim disconnected" id=1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803 namespace=k8s.io
Jul  2 00:02:05.617825 containerd[1431]: time="2024-07-02T00:02:05.617823239Z" level=warning msg="cleaning up after shim disconnected" id=1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803 namespace=k8s.io
Jul  2 00:02:05.617825 containerd[1431]: time="2024-07-02T00:02:05.617832481Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:02:06.520045 kubelet[2518]: E0702 00:02:06.520004    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:06.524722 containerd[1431]: time="2024-07-02T00:02:06.524614783Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}"
Jul  2 00:02:06.543051 containerd[1431]: time="2024-07-02T00:02:06.542991215Z" level=info msg="CreateContainer within sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\""
Jul  2 00:02:06.543777 containerd[1431]: time="2024-07-02T00:02:06.543720043Z" level=info msg="StartContainer for \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\""
Jul  2 00:02:06.581460 systemd[1]: Started cri-containerd-1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd.scope - libcontainer container 1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd.
Jul  2 00:02:06.603436 containerd[1431]: time="2024-07-02T00:02:06.603318598Z" level=info msg="StartContainer for \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\" returns successfully"
Jul  2 00:02:06.735781 kubelet[2518]: I0702 00:02:06.735588    2518 kubelet_node_status.go:497] "Fast updating node status as it just became ready"
Jul  2 00:02:06.757369 kubelet[2518]: I0702 00:02:06.757324    2518 topology_manager.go:215] "Topology Admit Handler" podUID="8a2decdf-e2c1-47f3-9adf-693e93cf20d4" podNamespace="kube-system" podName="coredns-7db6d8ff4d-pbtzr"
Jul  2 00:02:06.761334 kubelet[2518]: I0702 00:02:06.760610    2518 topology_manager.go:215] "Topology Admit Handler" podUID="a551c430-36e4-49f1-88a8-3a8040567b13" podNamespace="kube-system" podName="coredns-7db6d8ff4d-kj4f8"
Jul  2 00:02:06.769107 systemd[1]: Created slice kubepods-burstable-pod8a2decdf_e2c1_47f3_9adf_693e93cf20d4.slice - libcontainer container kubepods-burstable-pod8a2decdf_e2c1_47f3_9adf_693e93cf20d4.slice.
Jul  2 00:02:06.774861 systemd[1]: Created slice kubepods-burstable-poda551c430_36e4_49f1_88a8_3a8040567b13.slice - libcontainer container kubepods-burstable-poda551c430_36e4_49f1_88a8_3a8040567b13.slice.
Jul  2 00:02:06.884012 kubelet[2518]: I0702 00:02:06.883960    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a551c430-36e4-49f1-88a8-3a8040567b13-config-volume\") pod \"coredns-7db6d8ff4d-kj4f8\" (UID: \"a551c430-36e4-49f1-88a8-3a8040567b13\") " pod="kube-system/coredns-7db6d8ff4d-kj4f8"
Jul  2 00:02:06.884012 kubelet[2518]: I0702 00:02:06.884011    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpvqq\" (UniqueName: \"kubernetes.io/projected/a551c430-36e4-49f1-88a8-3a8040567b13-kube-api-access-cpvqq\") pod \"coredns-7db6d8ff4d-kj4f8\" (UID: \"a551c430-36e4-49f1-88a8-3a8040567b13\") " pod="kube-system/coredns-7db6d8ff4d-kj4f8"
Jul  2 00:02:06.884191 kubelet[2518]: I0702 00:02:06.884035    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8a2decdf-e2c1-47f3-9adf-693e93cf20d4-config-volume\") pod \"coredns-7db6d8ff4d-pbtzr\" (UID: \"8a2decdf-e2c1-47f3-9adf-693e93cf20d4\") " pod="kube-system/coredns-7db6d8ff4d-pbtzr"
Jul  2 00:02:06.884191 kubelet[2518]: I0702 00:02:06.884059    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k74t\" (UniqueName: \"kubernetes.io/projected/8a2decdf-e2c1-47f3-9adf-693e93cf20d4-kube-api-access-7k74t\") pod \"coredns-7db6d8ff4d-pbtzr\" (UID: \"8a2decdf-e2c1-47f3-9adf-693e93cf20d4\") " pod="kube-system/coredns-7db6d8ff4d-pbtzr"
Jul  2 00:02:07.074547 kubelet[2518]: E0702 00:02:07.074438    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:07.075661 containerd[1431]: time="2024-07-02T00:02:07.075610132Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-pbtzr,Uid:8a2decdf-e2c1-47f3-9adf-693e93cf20d4,Namespace:kube-system,Attempt:0,}"
Jul  2 00:02:07.084265 kubelet[2518]: E0702 00:02:07.083774    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:07.084654 containerd[1431]: time="2024-07-02T00:02:07.084619499Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-kj4f8,Uid:a551c430-36e4-49f1-88a8-3a8040567b13,Namespace:kube-system,Attempt:0,}"
Jul  2 00:02:07.525475 kubelet[2518]: E0702 00:02:07.525143    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:07.542328 kubelet[2518]: I0702 00:02:07.541470    2518 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-r2tjl" podStartSLOduration=5.296761629 podStartE2EDuration="13.541453224s" podCreationTimestamp="2024-07-02 00:01:54 +0000 UTC" firstStartedPulling="2024-07-02 00:01:54.484191437 +0000 UTC m=+16.160608091" lastFinishedPulling="2024-07-02 00:02:02.728883032 +0000 UTC m=+24.405299686" observedRunningTime="2024-07-02 00:02:07.541086093 +0000 UTC m=+29.217502747" watchObservedRunningTime="2024-07-02 00:02:07.541453224 +0000 UTC m=+29.217869838"
Jul  2 00:02:08.529919 kubelet[2518]: E0702 00:02:08.529871    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:08.953287 systemd-networkd[1368]: cilium_host: Link UP
Jul  2 00:02:08.953443 systemd-networkd[1368]: cilium_net: Link UP
Jul  2 00:02:08.953446 systemd-networkd[1368]: cilium_net: Gained carrier
Jul  2 00:02:08.954046 systemd-networkd[1368]: cilium_host: Gained carrier
Jul  2 00:02:08.954248 systemd-networkd[1368]: cilium_host: Gained IPv6LL
Jul  2 00:02:09.048262 systemd-networkd[1368]: cilium_vxlan: Link UP
Jul  2 00:02:09.048268 systemd-networkd[1368]: cilium_vxlan: Gained carrier
Jul  2 00:02:09.351457 systemd-networkd[1368]: cilium_net: Gained IPv6LL
Jul  2 00:02:09.387326 kernel: NET: Registered PF_ALG protocol family
Jul  2 00:02:09.532009 kubelet[2518]: E0702 00:02:09.531950    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:10.030131 systemd-networkd[1368]: lxc_health: Link UP
Jul  2 00:02:10.041518 systemd-networkd[1368]: lxc_health: Gained carrier
Jul  2 00:02:10.216769 systemd-networkd[1368]: lxc22301661d778: Link UP
Jul  2 00:02:10.217332 kernel: eth0: renamed from tmp6000f
Jul  2 00:02:10.225698 systemd-networkd[1368]: lxc85b05e7531fe: Link UP
Jul  2 00:02:10.238433 kernel: eth0: renamed from tmp5ce95
Jul  2 00:02:10.239842 systemd-networkd[1368]: lxc22301661d778: Gained carrier
Jul  2 00:02:10.246257 systemd-networkd[1368]: lxc85b05e7531fe: Gained carrier
Jul  2 00:02:10.420434 systemd-networkd[1368]: cilium_vxlan: Gained IPv6LL
Jul  2 00:02:10.535350 kubelet[2518]: E0702 00:02:10.535285    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:11.639779 systemd-networkd[1368]: lxc22301661d778: Gained IPv6LL
Jul  2 00:02:11.700614 systemd-networkd[1368]: lxc_health: Gained IPv6LL
Jul  2 00:02:12.021847 systemd-networkd[1368]: lxc85b05e7531fe: Gained IPv6LL
Jul  2 00:02:13.117196 systemd[1]: Started sshd@7-10.0.0.39:22-10.0.0.1:52336.service - OpenSSH per-connection server daemon (10.0.0.1:52336).
Jul  2 00:02:13.163677 sshd[3749]: Accepted publickey for core from 10.0.0.1 port 52336 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:13.166355 sshd[3749]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:13.176942 systemd-logind[1423]: New session 8 of user core.
Jul  2 00:02:13.190905 systemd[1]: Started session-8.scope - Session 8 of User core.
Jul  2 00:02:13.376641 sshd[3749]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:13.380212 systemd-logind[1423]: Session 8 logged out. Waiting for processes to exit.
Jul  2 00:02:13.381036 systemd[1]: sshd@7-10.0.0.39:22-10.0.0.1:52336.service: Deactivated successfully.
Jul  2 00:02:13.384243 systemd[1]: session-8.scope: Deactivated successfully.
Jul  2 00:02:13.387141 systemd-logind[1423]: Removed session 8.
Jul  2 00:02:13.984096 containerd[1431]: time="2024-07-02T00:02:13.984000788Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul  2 00:02:13.984096 containerd[1431]: time="2024-07-02T00:02:13.984061033Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:02:13.984631 containerd[1431]: time="2024-07-02T00:02:13.984074634Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul  2 00:02:13.984631 containerd[1431]: time="2024-07-02T00:02:13.984086035Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:02:13.984631 containerd[1431]: time="2024-07-02T00:02:13.984236847Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul  2 00:02:13.984631 containerd[1431]: time="2024-07-02T00:02:13.984288331Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:02:13.984631 containerd[1431]: time="2024-07-02T00:02:13.984320254Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul  2 00:02:13.984631 containerd[1431]: time="2024-07-02T00:02:13.984339215Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:02:14.017538 systemd[1]: Started cri-containerd-5ce95416bb29a1c4703d5f06d53be9a659b975081ff0e01bc723945073c6d042.scope - libcontainer container 5ce95416bb29a1c4703d5f06d53be9a659b975081ff0e01bc723945073c6d042.
Jul  2 00:02:14.018830 systemd[1]: Started cri-containerd-6000fda9b2fb61f25e07f24a299831c7e3d36acd7f3f878208da2351f38cccc4.scope - libcontainer container 6000fda9b2fb61f25e07f24a299831c7e3d36acd7f3f878208da2351f38cccc4.
Jul  2 00:02:14.030033 systemd-resolved[1306]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address
Jul  2 00:02:14.032254 systemd-resolved[1306]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address
Jul  2 00:02:14.053952 containerd[1431]: time="2024-07-02T00:02:14.053912992Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-kj4f8,Uid:a551c430-36e4-49f1-88a8-3a8040567b13,Namespace:kube-system,Attempt:0,} returns sandbox id \"6000fda9b2fb61f25e07f24a299831c7e3d36acd7f3f878208da2351f38cccc4\""
Jul  2 00:02:14.055021 kubelet[2518]: E0702 00:02:14.054996    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:14.059420 containerd[1431]: time="2024-07-02T00:02:14.058870102Z" level=info msg="CreateContainer within sandbox \"6000fda9b2fb61f25e07f24a299831c7e3d36acd7f3f878208da2351f38cccc4\" for container &ContainerMetadata{Name:coredns,Attempt:0,}"
Jul  2 00:02:14.059880 containerd[1431]: time="2024-07-02T00:02:14.059838658Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-pbtzr,Uid:8a2decdf-e2c1-47f3-9adf-693e93cf20d4,Namespace:kube-system,Attempt:0,} returns sandbox id \"5ce95416bb29a1c4703d5f06d53be9a659b975081ff0e01bc723945073c6d042\""
Jul  2 00:02:14.060708 kubelet[2518]: E0702 00:02:14.060683    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:14.062594 containerd[1431]: time="2024-07-02T00:02:14.062564352Z" level=info msg="CreateContainer within sandbox \"5ce95416bb29a1c4703d5f06d53be9a659b975081ff0e01bc723945073c6d042\" for container &ContainerMetadata{Name:coredns,Attempt:0,}"
Jul  2 00:02:14.081936 containerd[1431]: time="2024-07-02T00:02:14.081876988Z" level=info msg="CreateContainer within sandbox \"5ce95416bb29a1c4703d5f06d53be9a659b975081ff0e01bc723945073c6d042\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"c2c3129ee4623d4fe087dbb73d978bf8060b043e0f11b91fc4d490cf85dcd30b\""
Jul  2 00:02:14.082724 containerd[1431]: time="2024-07-02T00:02:14.082694092Z" level=info msg="StartContainer for \"c2c3129ee4623d4fe087dbb73d978bf8060b043e0f11b91fc4d490cf85dcd30b\""
Jul  2 00:02:14.093494 containerd[1431]: time="2024-07-02T00:02:14.093436775Z" level=info msg="CreateContainer within sandbox \"6000fda9b2fb61f25e07f24a299831c7e3d36acd7f3f878208da2351f38cccc4\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"0efb46cdef4ccfbcb4884b53b478e7d75cd525254931bf8f3c3f751215dc37ce\""
Jul  2 00:02:14.095645 containerd[1431]: time="2024-07-02T00:02:14.095595185Z" level=info msg="StartContainer for \"0efb46cdef4ccfbcb4884b53b478e7d75cd525254931bf8f3c3f751215dc37ce\""
Jul  2 00:02:14.119515 systemd[1]: Started cri-containerd-c2c3129ee4623d4fe087dbb73d978bf8060b043e0f11b91fc4d490cf85dcd30b.scope - libcontainer container c2c3129ee4623d4fe087dbb73d978bf8060b043e0f11b91fc4d490cf85dcd30b.
Jul  2 00:02:14.142469 systemd[1]: Started cri-containerd-0efb46cdef4ccfbcb4884b53b478e7d75cd525254931bf8f3c3f751215dc37ce.scope - libcontainer container 0efb46cdef4ccfbcb4884b53b478e7d75cd525254931bf8f3c3f751215dc37ce.
Jul  2 00:02:14.146959 containerd[1431]: time="2024-07-02T00:02:14.146915853Z" level=info msg="StartContainer for \"c2c3129ee4623d4fe087dbb73d978bf8060b043e0f11b91fc4d490cf85dcd30b\" returns successfully"
Jul  2 00:02:14.170309 containerd[1431]: time="2024-07-02T00:02:14.170257566Z" level=info msg="StartContainer for \"0efb46cdef4ccfbcb4884b53b478e7d75cd525254931bf8f3c3f751215dc37ce\" returns successfully"
Jul  2 00:02:14.551192 kubelet[2518]: E0702 00:02:14.551065    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:14.558124 kubelet[2518]: E0702 00:02:14.557594    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:14.563068 kubelet[2518]: I0702 00:02:14.562217    2518 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-kj4f8" podStartSLOduration=20.562199814 podStartE2EDuration="20.562199814s" podCreationTimestamp="2024-07-02 00:01:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 00:02:14.561898031 +0000 UTC m=+36.238314685" watchObservedRunningTime="2024-07-02 00:02:14.562199814 +0000 UTC m=+36.238616468"
Jul  2 00:02:14.576419 kubelet[2518]: I0702 00:02:14.576052    2518 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-pbtzr" podStartSLOduration=20.5760341 podStartE2EDuration="20.5760341s" podCreationTimestamp="2024-07-02 00:01:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 00:02:14.575444894 +0000 UTC m=+36.251861548" watchObservedRunningTime="2024-07-02 00:02:14.5760341 +0000 UTC m=+36.252450714"
Jul  2 00:02:15.558943 kubelet[2518]: E0702 00:02:15.557622    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:15.558943 kubelet[2518]: E0702 00:02:15.557740    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:16.559651 kubelet[2518]: E0702 00:02:16.559610    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:16.560019 kubelet[2518]: E0702 00:02:16.559770    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:17.603310 kubelet[2518]: I0702 00:02:17.603253    2518 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness"
Jul  2 00:02:17.604227 kubelet[2518]: E0702 00:02:17.604057    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:18.393259 systemd[1]: Started sshd@8-10.0.0.39:22-10.0.0.1:52344.service - OpenSSH per-connection server daemon (10.0.0.1:52344).
Jul  2 00:02:18.453173 sshd[3936]: Accepted publickey for core from 10.0.0.1 port 52344 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:18.454831 sshd[3936]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:18.467865 systemd-logind[1423]: New session 9 of user core.
Jul  2 00:02:18.478133 systemd[1]: Started session-9.scope - Session 9 of User core.
Jul  2 00:02:18.563502 kubelet[2518]: E0702 00:02:18.563465    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:02:18.647512 sshd[3936]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:18.650906 systemd[1]: sshd@8-10.0.0.39:22-10.0.0.1:52344.service: Deactivated successfully.
Jul  2 00:02:18.652770 systemd[1]: session-9.scope: Deactivated successfully.
Jul  2 00:02:18.656019 systemd-logind[1423]: Session 9 logged out. Waiting for processes to exit.
Jul  2 00:02:18.657224 systemd-logind[1423]: Removed session 9.
Jul  2 00:02:23.663110 systemd[1]: Started sshd@9-10.0.0.39:22-10.0.0.1:53472.service - OpenSSH per-connection server daemon (10.0.0.1:53472).
Jul  2 00:02:23.695888 sshd[3954]: Accepted publickey for core from 10.0.0.1 port 53472 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:23.697282 sshd[3954]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:23.701471 systemd-logind[1423]: New session 10 of user core.
Jul  2 00:02:23.707492 systemd[1]: Started session-10.scope - Session 10 of User core.
Jul  2 00:02:23.820721 sshd[3954]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:23.823623 systemd[1]: sshd@9-10.0.0.39:22-10.0.0.1:53472.service: Deactivated successfully.
Jul  2 00:02:23.825817 systemd[1]: session-10.scope: Deactivated successfully.
Jul  2 00:02:23.829593 systemd-logind[1423]: Session 10 logged out. Waiting for processes to exit.
Jul  2 00:02:23.830512 systemd-logind[1423]: Removed session 10.
Jul  2 00:02:28.842332 systemd[1]: Started sshd@10-10.0.0.39:22-10.0.0.1:53488.service - OpenSSH per-connection server daemon (10.0.0.1:53488).
Jul  2 00:02:28.877007 sshd[3972]: Accepted publickey for core from 10.0.0.1 port 53488 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:28.878381 sshd[3972]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:28.882608 systemd-logind[1423]: New session 11 of user core.
Jul  2 00:02:28.894542 systemd[1]: Started session-11.scope - Session 11 of User core.
Jul  2 00:02:29.040919 sshd[3972]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:29.050483 systemd[1]: sshd@10-10.0.0.39:22-10.0.0.1:53488.service: Deactivated successfully.
Jul  2 00:02:29.052866 systemd[1]: session-11.scope: Deactivated successfully.
Jul  2 00:02:29.055600 systemd-logind[1423]: Session 11 logged out. Waiting for processes to exit.
Jul  2 00:02:29.062601 systemd[1]: Started sshd@11-10.0.0.39:22-10.0.0.1:53496.service - OpenSSH per-connection server daemon (10.0.0.1:53496).
Jul  2 00:02:29.065404 systemd-logind[1423]: Removed session 11.
Jul  2 00:02:29.101057 sshd[3988]: Accepted publickey for core from 10.0.0.1 port 53496 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:29.102867 sshd[3988]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:29.107151 systemd-logind[1423]: New session 12 of user core.
Jul  2 00:02:29.116526 systemd[1]: Started session-12.scope - Session 12 of User core.
Jul  2 00:02:29.287811 sshd[3988]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:29.297156 systemd[1]: sshd@11-10.0.0.39:22-10.0.0.1:53496.service: Deactivated successfully.
Jul  2 00:02:29.300151 systemd[1]: session-12.scope: Deactivated successfully.
Jul  2 00:02:29.304547 systemd-logind[1423]: Session 12 logged out. Waiting for processes to exit.
Jul  2 00:02:29.314664 systemd[1]: Started sshd@12-10.0.0.39:22-10.0.0.1:53500.service - OpenSSH per-connection server daemon (10.0.0.1:53500).
Jul  2 00:02:29.316157 systemd-logind[1423]: Removed session 12.
Jul  2 00:02:29.357575 sshd[4000]: Accepted publickey for core from 10.0.0.1 port 53500 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:29.358932 sshd[4000]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:29.367394 systemd-logind[1423]: New session 13 of user core.
Jul  2 00:02:29.379617 systemd[1]: Started session-13.scope - Session 13 of User core.
Jul  2 00:02:29.510040 sshd[4000]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:29.517483 systemd[1]: sshd@12-10.0.0.39:22-10.0.0.1:53500.service: Deactivated successfully.
Jul  2 00:02:29.519580 systemd[1]: session-13.scope: Deactivated successfully.
Jul  2 00:02:29.523211 systemd-logind[1423]: Session 13 logged out. Waiting for processes to exit.
Jul  2 00:02:29.525014 systemd-logind[1423]: Removed session 13.
Jul  2 00:02:34.528639 systemd[1]: Started sshd@13-10.0.0.39:22-10.0.0.1:38634.service - OpenSSH per-connection server daemon (10.0.0.1:38634).
Jul  2 00:02:34.566927 sshd[4015]: Accepted publickey for core from 10.0.0.1 port 38634 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:34.568378 sshd[4015]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:34.573048 systemd-logind[1423]: New session 14 of user core.
Jul  2 00:02:34.580688 systemd[1]: Started session-14.scope - Session 14 of User core.
Jul  2 00:02:34.693815 sshd[4015]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:34.697621 systemd[1]: sshd@13-10.0.0.39:22-10.0.0.1:38634.service: Deactivated successfully.
Jul  2 00:02:34.699744 systemd[1]: session-14.scope: Deactivated successfully.
Jul  2 00:02:34.702207 systemd-logind[1423]: Session 14 logged out. Waiting for processes to exit.
Jul  2 00:02:34.703162 systemd-logind[1423]: Removed session 14.
Jul  2 00:02:39.705048 systemd[1]: Started sshd@14-10.0.0.39:22-10.0.0.1:38648.service - OpenSSH per-connection server daemon (10.0.0.1:38648).
Jul  2 00:02:39.746166 sshd[4032]: Accepted publickey for core from 10.0.0.1 port 38648 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:39.747513 sshd[4032]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:39.751305 systemd-logind[1423]: New session 15 of user core.
Jul  2 00:02:39.764535 systemd[1]: Started session-15.scope - Session 15 of User core.
Jul  2 00:02:39.877983 sshd[4032]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:39.889967 systemd[1]: sshd@14-10.0.0.39:22-10.0.0.1:38648.service: Deactivated successfully.
Jul  2 00:02:39.891644 systemd[1]: session-15.scope: Deactivated successfully.
Jul  2 00:02:39.893733 systemd-logind[1423]: Session 15 logged out. Waiting for processes to exit.
Jul  2 00:02:39.905671 systemd[1]: Started sshd@15-10.0.0.39:22-10.0.0.1:38656.service - OpenSSH per-connection server daemon (10.0.0.1:38656).
Jul  2 00:02:39.907083 systemd-logind[1423]: Removed session 15.
Jul  2 00:02:39.946250 sshd[4047]: Accepted publickey for core from 10.0.0.1 port 38656 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:39.947720 sshd[4047]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:39.952362 systemd-logind[1423]: New session 16 of user core.
Jul  2 00:02:39.959513 systemd[1]: Started session-16.scope - Session 16 of User core.
Jul  2 00:02:40.153483 sshd[4047]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:40.164895 systemd[1]: sshd@15-10.0.0.39:22-10.0.0.1:38656.service: Deactivated successfully.
Jul  2 00:02:40.166723 systemd[1]: session-16.scope: Deactivated successfully.
Jul  2 00:02:40.167848 systemd-logind[1423]: Session 16 logged out. Waiting for processes to exit.
Jul  2 00:02:40.179639 systemd[1]: Started sshd@16-10.0.0.39:22-10.0.0.1:38662.service - OpenSSH per-connection server daemon (10.0.0.1:38662).
Jul  2 00:02:40.180964 systemd-logind[1423]: Removed session 16.
Jul  2 00:02:40.213174 sshd[4059]: Accepted publickey for core from 10.0.0.1 port 38662 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:40.214587 sshd[4059]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:40.219698 systemd-logind[1423]: New session 17 of user core.
Jul  2 00:02:40.230512 systemd[1]: Started session-17.scope - Session 17 of User core.
Jul  2 00:02:41.571243 sshd[4059]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:41.584812 systemd[1]: sshd@16-10.0.0.39:22-10.0.0.1:38662.service: Deactivated successfully.
Jul  2 00:02:41.590020 systemd[1]: session-17.scope: Deactivated successfully.
Jul  2 00:02:41.594190 systemd-logind[1423]: Session 17 logged out. Waiting for processes to exit.
Jul  2 00:02:41.602656 systemd[1]: Started sshd@17-10.0.0.39:22-10.0.0.1:42048.service - OpenSSH per-connection server daemon (10.0.0.1:42048).
Jul  2 00:02:41.603598 systemd-logind[1423]: Removed session 17.
Jul  2 00:02:41.633830 sshd[4079]: Accepted publickey for core from 10.0.0.1 port 42048 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:41.635473 sshd[4079]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:41.639374 systemd-logind[1423]: New session 18 of user core.
Jul  2 00:02:41.646461 systemd[1]: Started session-18.scope - Session 18 of User core.
Jul  2 00:02:41.918453 sshd[4079]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:41.927208 systemd[1]: sshd@17-10.0.0.39:22-10.0.0.1:42048.service: Deactivated successfully.
Jul  2 00:02:41.929432 systemd[1]: session-18.scope: Deactivated successfully.
Jul  2 00:02:41.932585 systemd-logind[1423]: Session 18 logged out. Waiting for processes to exit.
Jul  2 00:02:41.943352 systemd[1]: Started sshd@18-10.0.0.39:22-10.0.0.1:42060.service - OpenSSH per-connection server daemon (10.0.0.1:42060).
Jul  2 00:02:41.944467 systemd-logind[1423]: Removed session 18.
Jul  2 00:02:41.975320 sshd[4092]: Accepted publickey for core from 10.0.0.1 port 42060 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:41.976929 sshd[4092]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:41.981875 systemd-logind[1423]: New session 19 of user core.
Jul  2 00:02:41.999597 systemd[1]: Started session-19.scope - Session 19 of User core.
Jul  2 00:02:42.119586 sshd[4092]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:42.128035 systemd[1]: sshd@18-10.0.0.39:22-10.0.0.1:42060.service: Deactivated successfully.
Jul  2 00:02:42.131843 systemd[1]: session-19.scope: Deactivated successfully.
Jul  2 00:02:42.133274 systemd-logind[1423]: Session 19 logged out. Waiting for processes to exit.
Jul  2 00:02:42.134129 systemd-logind[1423]: Removed session 19.
Jul  2 00:02:47.129265 systemd[1]: Started sshd@19-10.0.0.39:22-10.0.0.1:42068.service - OpenSSH per-connection server daemon (10.0.0.1:42068).
Jul  2 00:02:47.164980 sshd[4109]: Accepted publickey for core from 10.0.0.1 port 42068 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:47.166488 sshd[4109]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:47.170811 systemd-logind[1423]: New session 20 of user core.
Jul  2 00:02:47.178522 systemd[1]: Started session-20.scope - Session 20 of User core.
Jul  2 00:02:47.296247 sshd[4109]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:47.300048 systemd[1]: sshd@19-10.0.0.39:22-10.0.0.1:42068.service: Deactivated successfully.
Jul  2 00:02:47.302531 systemd[1]: session-20.scope: Deactivated successfully.
Jul  2 00:02:47.303526 systemd-logind[1423]: Session 20 logged out. Waiting for processes to exit.
Jul  2 00:02:47.304661 systemd-logind[1423]: Removed session 20.
Jul  2 00:02:52.307267 systemd[1]: Started sshd@20-10.0.0.39:22-10.0.0.1:43608.service - OpenSSH per-connection server daemon (10.0.0.1:43608).
Jul  2 00:02:52.343895 sshd[4124]: Accepted publickey for core from 10.0.0.1 port 43608 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:52.345206 sshd[4124]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:52.349453 systemd-logind[1423]: New session 21 of user core.
Jul  2 00:02:52.359519 systemd[1]: Started session-21.scope - Session 21 of User core.
Jul  2 00:02:52.470969 sshd[4124]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:52.474658 systemd[1]: sshd@20-10.0.0.39:22-10.0.0.1:43608.service: Deactivated successfully.
Jul  2 00:02:52.476855 systemd[1]: session-21.scope: Deactivated successfully.
Jul  2 00:02:52.478151 systemd-logind[1423]: Session 21 logged out. Waiting for processes to exit.
Jul  2 00:02:52.479672 systemd-logind[1423]: Removed session 21.
Jul  2 00:02:57.482220 systemd[1]: Started sshd@21-10.0.0.39:22-10.0.0.1:43614.service - OpenSSH per-connection server daemon (10.0.0.1:43614).
Jul  2 00:02:57.515652 sshd[4140]: Accepted publickey for core from 10.0.0.1 port 43614 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:02:57.517269 sshd[4140]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:02:57.521513 systemd-logind[1423]: New session 22 of user core.
Jul  2 00:02:57.530487 systemd[1]: Started session-22.scope - Session 22 of User core.
Jul  2 00:02:57.636524 sshd[4140]: pam_unix(sshd:session): session closed for user core
Jul  2 00:02:57.640042 systemd[1]: sshd@21-10.0.0.39:22-10.0.0.1:43614.service: Deactivated successfully.
Jul  2 00:02:57.642071 systemd[1]: session-22.scope: Deactivated successfully.
Jul  2 00:02:57.643042 systemd-logind[1423]: Session 22 logged out. Waiting for processes to exit.
Jul  2 00:02:57.644044 systemd-logind[1423]: Removed session 22.
Jul  2 00:03:01.423413 kubelet[2518]: E0702 00:03:01.423364    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:02.650091 systemd[1]: Started sshd@22-10.0.0.39:22-10.0.0.1:50958.service - OpenSSH per-connection server daemon (10.0.0.1:50958).
Jul  2 00:03:02.684531 sshd[4154]: Accepted publickey for core from 10.0.0.1 port 50958 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:03:02.685922 sshd[4154]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:03:02.689936 systemd-logind[1423]: New session 23 of user core.
Jul  2 00:03:02.699483 systemd[1]: Started session-23.scope - Session 23 of User core.
Jul  2 00:03:02.816125 sshd[4154]: pam_unix(sshd:session): session closed for user core
Jul  2 00:03:02.827746 systemd[1]: sshd@22-10.0.0.39:22-10.0.0.1:50958.service: Deactivated successfully.
Jul  2 00:03:02.829466 systemd[1]: session-23.scope: Deactivated successfully.
Jul  2 00:03:02.830743 systemd-logind[1423]: Session 23 logged out. Waiting for processes to exit.
Jul  2 00:03:02.837206 systemd[1]: Started sshd@23-10.0.0.39:22-10.0.0.1:50968.service - OpenSSH per-connection server daemon (10.0.0.1:50968).
Jul  2 00:03:02.838064 systemd-logind[1423]: Removed session 23.
Jul  2 00:03:02.872272 sshd[4168]: Accepted publickey for core from 10.0.0.1 port 50968 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:03:02.873668 sshd[4168]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:03:02.877605 systemd-logind[1423]: New session 24 of user core.
Jul  2 00:03:02.886541 systemd[1]: Started session-24.scope - Session 24 of User core.
Jul  2 00:03:05.327536 containerd[1431]: time="2024-07-02T00:03:05.327486950Z" level=info msg="StopContainer for \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\" with timeout 30 (s)"
Jul  2 00:03:05.328673 containerd[1431]: time="2024-07-02T00:03:05.328037523Z" level=info msg="Stop container \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\" with signal terminated"
Jul  2 00:03:05.342223 systemd[1]: cri-containerd-b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4.scope: Deactivated successfully.
Jul  2 00:03:05.358272 containerd[1431]: time="2024-07-02T00:03:05.358209115Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE        \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config"
Jul  2 00:03:05.364202 containerd[1431]: time="2024-07-02T00:03:05.364008292Z" level=info msg="StopContainer for \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\" with timeout 2 (s)"
Jul  2 00:03:05.364829 containerd[1431]: time="2024-07-02T00:03:05.364800511Z" level=info msg="Stop container \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\" with signal terminated"
Jul  2 00:03:05.370118 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4-rootfs.mount: Deactivated successfully.
Jul  2 00:03:05.375022 systemd-networkd[1368]: lxc_health: Link DOWN
Jul  2 00:03:05.375030 systemd-networkd[1368]: lxc_health: Lost carrier
Jul  2 00:03:05.381939 containerd[1431]: time="2024-07-02T00:03:05.381876834Z" level=info msg="shim disconnected" id=b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4 namespace=k8s.io
Jul  2 00:03:05.381939 containerd[1431]: time="2024-07-02T00:03:05.381935155Z" level=warning msg="cleaning up after shim disconnected" id=b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4 namespace=k8s.io
Jul  2 00:03:05.382130 containerd[1431]: time="2024-07-02T00:03:05.381949315Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:03:05.400870 containerd[1431]: time="2024-07-02T00:03:05.400799080Z" level=info msg="StopContainer for \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\" returns successfully"
Jul  2 00:03:05.403649 containerd[1431]: time="2024-07-02T00:03:05.403595026Z" level=info msg="StopPodSandbox for \"e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040\""
Jul  2 00:03:05.403842 containerd[1431]: time="2024-07-02T00:03:05.403657587Z" level=info msg="Container to stop \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jul  2 00:03:05.405639 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040-shm.mount: Deactivated successfully.
Jul  2 00:03:05.406671 systemd[1]: cri-containerd-1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd.scope: Deactivated successfully.
Jul  2 00:03:05.408725 systemd[1]: cri-containerd-1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd.scope: Consumed 6.980s CPU time.
Jul  2 00:03:05.418681 systemd[1]: cri-containerd-e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040.scope: Deactivated successfully.
Jul  2 00:03:05.429806 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd-rootfs.mount: Deactivated successfully.
Jul  2 00:03:05.438285 containerd[1431]: time="2024-07-02T00:03:05.438067639Z" level=info msg="shim disconnected" id=1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd namespace=k8s.io
Jul  2 00:03:05.438285 containerd[1431]: time="2024-07-02T00:03:05.438126401Z" level=warning msg="cleaning up after shim disconnected" id=1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd namespace=k8s.io
Jul  2 00:03:05.438285 containerd[1431]: time="2024-07-02T00:03:05.438135001Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:03:05.442586 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040-rootfs.mount: Deactivated successfully.
Jul  2 00:03:05.444576 containerd[1431]: time="2024-07-02T00:03:05.444451350Z" level=info msg="shim disconnected" id=e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040 namespace=k8s.io
Jul  2 00:03:05.444576 containerd[1431]: time="2024-07-02T00:03:05.444546952Z" level=warning msg="cleaning up after shim disconnected" id=e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040 namespace=k8s.io
Jul  2 00:03:05.444576 containerd[1431]: time="2024-07-02T00:03:05.444558152Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:03:05.459545 containerd[1431]: time="2024-07-02T00:03:05.459430743Z" level=info msg="StopContainer for \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\" returns successfully"
Jul  2 00:03:05.460142 containerd[1431]: time="2024-07-02T00:03:05.459901394Z" level=info msg="StopPodSandbox for \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\""
Jul  2 00:03:05.460142 containerd[1431]: time="2024-07-02T00:03:05.459942355Z" level=info msg="Container to stop \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jul  2 00:03:05.460142 containerd[1431]: time="2024-07-02T00:03:05.459978476Z" level=info msg="Container to stop \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jul  2 00:03:05.460142 containerd[1431]: time="2024-07-02T00:03:05.459996517Z" level=info msg="Container to stop \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jul  2 00:03:05.460142 containerd[1431]: time="2024-07-02T00:03:05.460010597Z" level=info msg="Container to stop \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jul  2 00:03:05.460142 containerd[1431]: time="2024-07-02T00:03:05.460019557Z" level=info msg="Container to stop \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jul  2 00:03:05.466180 systemd[1]: cri-containerd-4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7.scope: Deactivated successfully.
Jul  2 00:03:05.481516 containerd[1431]: time="2024-07-02T00:03:05.481457303Z" level=info msg="TearDown network for sandbox \"e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040\" successfully"
Jul  2 00:03:05.481516 containerd[1431]: time="2024-07-02T00:03:05.481502744Z" level=info msg="StopPodSandbox for \"e318a6eff014101661c5e968fc60298a8150600701b1274f8777d271c7938040\" returns successfully"
Jul  2 00:03:05.496702 containerd[1431]: time="2024-07-02T00:03:05.496639141Z" level=info msg="shim disconnected" id=4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7 namespace=k8s.io
Jul  2 00:03:05.496702 containerd[1431]: time="2024-07-02T00:03:05.496701503Z" level=warning msg="cleaning up after shim disconnected" id=4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7 namespace=k8s.io
Jul  2 00:03:05.496702 containerd[1431]: time="2024-07-02T00:03:05.496709903Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:03:05.510015 containerd[1431]: time="2024-07-02T00:03:05.509880774Z" level=info msg="TearDown network for sandbox \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" successfully"
Jul  2 00:03:05.510015 containerd[1431]: time="2024-07-02T00:03:05.509922455Z" level=info msg="StopPodSandbox for \"4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7\" returns successfully"
Jul  2 00:03:05.563337 kubelet[2518]: I0702 00:03:05.562802    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqbbj\" (UniqueName: \"kubernetes.io/projected/5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33-kube-api-access-fqbbj\") pod \"5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33\" (UID: \"5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33\") "
Jul  2 00:03:05.563337 kubelet[2518]: I0702 00:03:05.562861    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33-cilium-config-path\") pod \"5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33\" (UID: \"5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33\") "
Jul  2 00:03:05.564831 kubelet[2518]: I0702 00:03:05.564787    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33" (UID: "5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jul  2 00:03:05.565933 kubelet[2518]: I0702 00:03:05.565880    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33-kube-api-access-fqbbj" (OuterVolumeSpecName: "kube-api-access-fqbbj") pod "5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33" (UID: "5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33"). InnerVolumeSpecName "kube-api-access-fqbbj". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jul  2 00:03:05.662183 kubelet[2518]: I0702 00:03:05.661915    2518 scope.go:117] "RemoveContainer" containerID="b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4"
Jul  2 00:03:05.662987 kubelet[2518]: I0702 00:03:05.662968    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-xtables-lock\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.663087 kubelet[2518]: I0702 00:03:05.663072    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-lib-modules\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.663160 kubelet[2518]: I0702 00:03:05.663144    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-host-proc-sys-kernel\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.663233 kubelet[2518]: I0702 00:03:05.663222    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-config-path\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665338 kubelet[2518]: I0702 00:03:05.663309    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-cgroup\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665338 kubelet[2518]: I0702 00:03:05.663334    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpgbd\" (UniqueName: \"kubernetes.io/projected/1f20980c-f700-4772-9986-b99108fa3c3d-kube-api-access-qpgbd\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665338 kubelet[2518]: I0702 00:03:05.663352    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1f20980c-f700-4772-9986-b99108fa3c3d-clustermesh-secrets\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665338 kubelet[2518]: I0702 00:03:05.663369    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-host-proc-sys-net\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665338 kubelet[2518]: I0702 00:03:05.663386    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-run\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665338 kubelet[2518]: I0702 00:03:05.663403    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1f20980c-f700-4772-9986-b99108fa3c3d-hubble-tls\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665542 kubelet[2518]: I0702 00:03:05.663419    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-hostproc\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665542 kubelet[2518]: I0702 00:03:05.663434    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cni-path\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665542 kubelet[2518]: I0702 00:03:05.663448    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-bpf-maps\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665542 kubelet[2518]: I0702 00:03:05.663461    2518 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-etc-cni-netd\") pod \"1f20980c-f700-4772-9986-b99108fa3c3d\" (UID: \"1f20980c-f700-4772-9986-b99108fa3c3d\") "
Jul  2 00:03:05.665542 kubelet[2518]: I0702 00:03:05.663493    2518 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-fqbbj\" (UniqueName: \"kubernetes.io/projected/5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33-kube-api-access-fqbbj\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.665542 kubelet[2518]: I0702 00:03:05.663503    2518 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33-cilium-config-path\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.665667 kubelet[2518]: I0702 00:03:05.663544    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.665667 kubelet[2518]: I0702 00:03:05.663573    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.665667 kubelet[2518]: I0702 00:03:05.663600    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.665667 kubelet[2518]: I0702 00:03:05.663613    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.665667 kubelet[2518]: I0702 00:03:05.664745    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.665772 kubelet[2518]: I0702 00:03:05.664774    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.665772 kubelet[2518]: I0702 00:03:05.664842    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-hostproc" (OuterVolumeSpecName: "hostproc") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.666627 containerd[1431]: time="2024-07-02T00:03:05.666523750Z" level=info msg="RemoveContainer for \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\""
Jul  2 00:03:05.667905 kubelet[2518]: I0702 00:03:05.667390    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f20980c-f700-4772-9986-b99108fa3c3d-kube-api-access-qpgbd" (OuterVolumeSpecName: "kube-api-access-qpgbd") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "kube-api-access-qpgbd". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jul  2 00:03:05.667905 kubelet[2518]: I0702 00:03:05.667449    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.667632 systemd[1]: Removed slice kubepods-besteffort-pod5be9b8b1_bfbc_4ece_9980_31aa7ff1bf33.slice - libcontainer container kubepods-besteffort-pod5be9b8b1_bfbc_4ece_9980_31aa7ff1bf33.slice.
Jul  2 00:03:05.668592 kubelet[2518]: I0702 00:03:05.668533    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f20980c-f700-4772-9986-b99108fa3c3d-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jul  2 00:03:05.668592 kubelet[2518]: I0702 00:03:05.668544    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jul  2 00:03:05.668592 kubelet[2518]: I0702 00:03:05.668575    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cni-path" (OuterVolumeSpecName: "cni-path") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.668731 kubelet[2518]: I0702 00:03:05.668624    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jul  2 00:03:05.671187 containerd[1431]: time="2024-07-02T00:03:05.671144219Z" level=info msg="RemoveContainer for \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\" returns successfully"
Jul  2 00:03:05.671529 kubelet[2518]: I0702 00:03:05.671490    2518 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f20980c-f700-4772-9986-b99108fa3c3d-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "1f20980c-f700-4772-9986-b99108fa3c3d" (UID: "1f20980c-f700-4772-9986-b99108fa3c3d"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jul  2 00:03:05.672664 kubelet[2518]: I0702 00:03:05.672046    2518 scope.go:117] "RemoveContainer" containerID="b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4"
Jul  2 00:03:05.682664 containerd[1431]: time="2024-07-02T00:03:05.672795898Z" level=error msg="ContainerStatus for \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\": not found"
Jul  2 00:03:05.682763 kubelet[2518]: E0702 00:03:05.682486    2518 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\": not found" containerID="b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4"
Jul  2 00:03:05.684807 kubelet[2518]: I0702 00:03:05.682534    2518 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4"} err="failed to get container status \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\": rpc error: code = NotFound desc = an error occurred when try to find container \"b8608ff79c71885ba58f81a2a33bab3abb28d9ecddf18b5da9e39732cec8e8b4\": not found"
Jul  2 00:03:05.684807 kubelet[2518]: I0702 00:03:05.684804    2518 scope.go:117] "RemoveContainer" containerID="1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd"
Jul  2 00:03:05.687134 containerd[1431]: time="2024-07-02T00:03:05.687086635Z" level=info msg="RemoveContainer for \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\""
Jul  2 00:03:05.690162 containerd[1431]: time="2024-07-02T00:03:05.690102146Z" level=info msg="RemoveContainer for \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\" returns successfully"
Jul  2 00:03:05.690429 kubelet[2518]: I0702 00:03:05.690322    2518 scope.go:117] "RemoveContainer" containerID="1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803"
Jul  2 00:03:05.691956 containerd[1431]: time="2024-07-02T00:03:05.691400977Z" level=info msg="RemoveContainer for \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\""
Jul  2 00:03:05.692062 systemd[1]: Removed slice kubepods-burstable-pod1f20980c_f700_4772_9986_b99108fa3c3d.slice - libcontainer container kubepods-burstable-pod1f20980c_f700_4772_9986_b99108fa3c3d.slice.
Jul  2 00:03:05.692259 systemd[1]: kubepods-burstable-pod1f20980c_f700_4772_9986_b99108fa3c3d.slice: Consumed 7.143s CPU time.
Jul  2 00:03:05.694324 containerd[1431]: time="2024-07-02T00:03:05.694267004Z" level=info msg="RemoveContainer for \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\" returns successfully"
Jul  2 00:03:05.694573 kubelet[2518]: I0702 00:03:05.694535    2518 scope.go:117] "RemoveContainer" containerID="0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783"
Jul  2 00:03:05.697182 containerd[1431]: time="2024-07-02T00:03:05.696745583Z" level=info msg="RemoveContainer for \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\""
Jul  2 00:03:05.699227 containerd[1431]: time="2024-07-02T00:03:05.699177320Z" level=info msg="RemoveContainer for \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\" returns successfully"
Jul  2 00:03:05.699426 kubelet[2518]: I0702 00:03:05.699390    2518 scope.go:117] "RemoveContainer" containerID="ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14"
Jul  2 00:03:05.700725 containerd[1431]: time="2024-07-02T00:03:05.700684876Z" level=info msg="RemoveContainer for \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\""
Jul  2 00:03:05.704738 containerd[1431]: time="2024-07-02T00:03:05.704703410Z" level=info msg="RemoveContainer for \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\" returns successfully"
Jul  2 00:03:05.704963 kubelet[2518]: I0702 00:03:05.704893    2518 scope.go:117] "RemoveContainer" containerID="776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303"
Jul  2 00:03:05.705989 containerd[1431]: time="2024-07-02T00:03:05.705962720Z" level=info msg="RemoveContainer for \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\""
Jul  2 00:03:05.708346 containerd[1431]: time="2024-07-02T00:03:05.708281815Z" level=info msg="RemoveContainer for \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\" returns successfully"
Jul  2 00:03:05.708568 kubelet[2518]: I0702 00:03:05.708539    2518 scope.go:117] "RemoveContainer" containerID="1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd"
Jul  2 00:03:05.708934 containerd[1431]: time="2024-07-02T00:03:05.708765826Z" level=error msg="ContainerStatus for \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\": not found"
Jul  2 00:03:05.709090 kubelet[2518]: E0702 00:03:05.709066    2518 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\": not found" containerID="1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd"
Jul  2 00:03:05.709127 kubelet[2518]: I0702 00:03:05.709100    2518 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd"} err="failed to get container status \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\": rpc error: code = NotFound desc = an error occurred when try to find container \"1960092a7dd63e6e802ae41e440b4491e3b093867f26dca23efdb85b3ee6d0fd\": not found"
Jul  2 00:03:05.709127 kubelet[2518]: I0702 00:03:05.709122    2518 scope.go:117] "RemoveContainer" containerID="1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803"
Jul  2 00:03:05.709328 containerd[1431]: time="2024-07-02T00:03:05.709280318Z" level=error msg="ContainerStatus for \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\": not found"
Jul  2 00:03:05.709436 kubelet[2518]: E0702 00:03:05.709415    2518 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\": not found" containerID="1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803"
Jul  2 00:03:05.709469 kubelet[2518]: I0702 00:03:05.709441    2518 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803"} err="failed to get container status \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\": rpc error: code = NotFound desc = an error occurred when try to find container \"1a7c6b9da601fb8a2951e2adfe0bc8d1bf5b8270bfb40354ddd1849744ce9803\": not found"
Jul  2 00:03:05.709469 kubelet[2518]: I0702 00:03:05.709457    2518 scope.go:117] "RemoveContainer" containerID="0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783"
Jul  2 00:03:05.709800 containerd[1431]: time="2024-07-02T00:03:05.709766050Z" level=error msg="ContainerStatus for \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\": not found"
Jul  2 00:03:05.709966 kubelet[2518]: E0702 00:03:05.709946    2518 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\": not found" containerID="0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783"
Jul  2 00:03:05.710008 kubelet[2518]: I0702 00:03:05.709986    2518 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783"} err="failed to get container status \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\": rpc error: code = NotFound desc = an error occurred when try to find container \"0cc9a3425c7cb4b9e2aaedb17d24b5216bef71bfc750eb325b59379b4cd77783\": not found"
Jul  2 00:03:05.710008 kubelet[2518]: I0702 00:03:05.710003    2518 scope.go:117] "RemoveContainer" containerID="ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14"
Jul  2 00:03:05.710186 containerd[1431]: time="2024-07-02T00:03:05.710156219Z" level=error msg="ContainerStatus for \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\": not found"
Jul  2 00:03:05.710298 kubelet[2518]: E0702 00:03:05.710273    2518 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\": not found" containerID="ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14"
Jul  2 00:03:05.710332 kubelet[2518]: I0702 00:03:05.710309    2518 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14"} err="failed to get container status \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\": rpc error: code = NotFound desc = an error occurred when try to find container \"ee944f6d1f393d04c54b9addff1947dd87791120775274703b4752fc33a11f14\": not found"
Jul  2 00:03:05.710332 kubelet[2518]: I0702 00:03:05.710325    2518 scope.go:117] "RemoveContainer" containerID="776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303"
Jul  2 00:03:05.710618 containerd[1431]: time="2024-07-02T00:03:05.710587389Z" level=error msg="ContainerStatus for \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\": not found"
Jul  2 00:03:05.710779 kubelet[2518]: E0702 00:03:05.710743    2518 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\": not found" containerID="776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303"
Jul  2 00:03:05.710779 kubelet[2518]: I0702 00:03:05.710765    2518 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303"} err="failed to get container status \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\": rpc error: code = NotFound desc = an error occurred when try to find container \"776d47d0064b6fac545777709640eb52133db5105511380705c2c4c8f9207303\": not found"
Jul  2 00:03:05.764077 kubelet[2518]: I0702 00:03:05.764038    2518 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-xtables-lock\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764077 kubelet[2518]: I0702 00:03:05.764072    2518 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-lib-modules\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764282 kubelet[2518]: I0702 00:03:05.764083    2518 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764282 kubelet[2518]: I0702 00:03:05.764105    2518 reconciler_common.go:289] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-config-path\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764282 kubelet[2518]: I0702 00:03:05.764114    2518 reconciler_common.go:289] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-cgroup\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764282 kubelet[2518]: I0702 00:03:05.764121    2518 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-qpgbd\" (UniqueName: \"kubernetes.io/projected/1f20980c-f700-4772-9986-b99108fa3c3d-kube-api-access-qpgbd\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764282 kubelet[2518]: I0702 00:03:05.764131    2518 reconciler_common.go:289] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/1f20980c-f700-4772-9986-b99108fa3c3d-clustermesh-secrets\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764282 kubelet[2518]: I0702 00:03:05.764138    2518 reconciler_common.go:289] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-host-proc-sys-net\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764282 kubelet[2518]: I0702 00:03:05.764145    2518 reconciler_common.go:289] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cilium-run\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764282 kubelet[2518]: I0702 00:03:05.764154    2518 reconciler_common.go:289] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-bpf-maps\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764462 kubelet[2518]: I0702 00:03:05.764161    2518 reconciler_common.go:289] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-etc-cni-netd\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764462 kubelet[2518]: I0702 00:03:05.764169    2518 reconciler_common.go:289] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/1f20980c-f700-4772-9986-b99108fa3c3d-hubble-tls\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764462 kubelet[2518]: I0702 00:03:05.764176    2518 reconciler_common.go:289] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-hostproc\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:05.764462 kubelet[2518]: I0702 00:03:05.764183    2518 reconciler_common.go:289] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/1f20980c-f700-4772-9986-b99108fa3c3d-cni-path\") on node \"localhost\" DevicePath \"\""
Jul  2 00:03:06.345757 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7-rootfs.mount: Deactivated successfully.
Jul  2 00:03:06.345849 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-4123df72d0ca4ff0a5121545c4514c44776504020a5ee33445291cbd4dba3bb7-shm.mount: Deactivated successfully.
Jul  2 00:03:06.345918 systemd[1]: var-lib-kubelet-pods-1f20980c\x2df700\x2d4772\x2d9986\x2db99108fa3c3d-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dqpgbd.mount: Deactivated successfully.
Jul  2 00:03:06.345982 systemd[1]: var-lib-kubelet-pods-5be9b8b1\x2dbfbc\x2d4ece\x2d9980\x2d31aa7ff1bf33-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dfqbbj.mount: Deactivated successfully.
Jul  2 00:03:06.346033 systemd[1]: var-lib-kubelet-pods-1f20980c\x2df700\x2d4772\x2d9986\x2db99108fa3c3d-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully.
Jul  2 00:03:06.346084 systemd[1]: var-lib-kubelet-pods-1f20980c\x2df700\x2d4772\x2d9986\x2db99108fa3c3d-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully.
Jul  2 00:03:06.425132 kubelet[2518]: I0702 00:03:06.425096    2518 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f20980c-f700-4772-9986-b99108fa3c3d" path="/var/lib/kubelet/pods/1f20980c-f700-4772-9986-b99108fa3c3d/volumes"
Jul  2 00:03:06.425642 kubelet[2518]: I0702 00:03:06.425625    2518 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33" path="/var/lib/kubelet/pods/5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33/volumes"
Jul  2 00:03:07.279169 sshd[4168]: pam_unix(sshd:session): session closed for user core
Jul  2 00:03:07.293032 systemd[1]: sshd@23-10.0.0.39:22-10.0.0.1:50968.service: Deactivated successfully.
Jul  2 00:03:07.295741 systemd[1]: session-24.scope: Deactivated successfully.
Jul  2 00:03:07.295920 systemd[1]: session-24.scope: Consumed 1.743s CPU time.
Jul  2 00:03:07.299897 systemd-logind[1423]: Session 24 logged out. Waiting for processes to exit.
Jul  2 00:03:07.317743 systemd[1]: Started sshd@24-10.0.0.39:22-10.0.0.1:50978.service - OpenSSH per-connection server daemon (10.0.0.1:50978).
Jul  2 00:03:07.320651 systemd-logind[1423]: Removed session 24.
Jul  2 00:03:07.349751 sshd[4330]: Accepted publickey for core from 10.0.0.1 port 50978 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:03:07.351130 sshd[4330]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:03:07.355842 systemd-logind[1423]: New session 25 of user core.
Jul  2 00:03:07.368507 systemd[1]: Started session-25.scope - Session 25 of User core.
Jul  2 00:03:07.423224 kubelet[2518]: E0702 00:03:07.423189    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:08.200517 sshd[4330]: pam_unix(sshd:session): session closed for user core
Jul  2 00:03:08.212405 systemd[1]: sshd@24-10.0.0.39:22-10.0.0.1:50978.service: Deactivated successfully.
Jul  2 00:03:08.214954 systemd[1]: session-25.scope: Deactivated successfully.
Jul  2 00:03:08.218729 systemd-logind[1423]: Session 25 logged out. Waiting for processes to exit.
Jul  2 00:03:08.230027 kubelet[2518]: I0702 00:03:08.229612    2518 topology_manager.go:215] "Topology Admit Handler" podUID="395ee682-451c-453b-a9d1-833f7952e747" podNamespace="kube-system" podName="cilium-h4mwg"
Jul  2 00:03:08.230027 kubelet[2518]: E0702 00:03:08.229732    2518 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="1f20980c-f700-4772-9986-b99108fa3c3d" containerName="cilium-agent"
Jul  2 00:03:08.230027 kubelet[2518]: E0702 00:03:08.229742    2518 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="1f20980c-f700-4772-9986-b99108fa3c3d" containerName="mount-cgroup"
Jul  2 00:03:08.230027 kubelet[2518]: E0702 00:03:08.229758    2518 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="1f20980c-f700-4772-9986-b99108fa3c3d" containerName="apply-sysctl-overwrites"
Jul  2 00:03:08.230027 kubelet[2518]: E0702 00:03:08.229764    2518 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="1f20980c-f700-4772-9986-b99108fa3c3d" containerName="mount-bpf-fs"
Jul  2 00:03:08.230027 kubelet[2518]: E0702 00:03:08.229769    2518 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="1f20980c-f700-4772-9986-b99108fa3c3d" containerName="clean-cilium-state"
Jul  2 00:03:08.230027 kubelet[2518]: E0702 00:03:08.229779    2518 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33" containerName="cilium-operator"
Jul  2 00:03:08.230027 kubelet[2518]: I0702 00:03:08.229802    2518 memory_manager.go:354] "RemoveStaleState removing state" podUID="5be9b8b1-bfbc-4ece-9980-31aa7ff1bf33" containerName="cilium-operator"
Jul  2 00:03:08.230027 kubelet[2518]: I0702 00:03:08.229808    2518 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f20980c-f700-4772-9986-b99108fa3c3d" containerName="cilium-agent"
Jul  2 00:03:08.230599 systemd[1]: Started sshd@25-10.0.0.39:22-10.0.0.1:50982.service - OpenSSH per-connection server daemon (10.0.0.1:50982).
Jul  2 00:03:08.237307 systemd-logind[1423]: Removed session 25.
Jul  2 00:03:08.250942 systemd[1]: Created slice kubepods-burstable-pod395ee682_451c_453b_a9d1_833f7952e747.slice - libcontainer container kubepods-burstable-pod395ee682_451c_453b_a9d1_833f7952e747.slice.
Jul  2 00:03:08.261700 sshd[4343]: Accepted publickey for core from 10.0.0.1 port 50982 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:03:08.263178 sshd[4343]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:03:08.271021 systemd-logind[1423]: New session 26 of user core.
Jul  2 00:03:08.279538 systemd[1]: Started session-26.scope - Session 26 of User core.
Jul  2 00:03:08.330653 sshd[4343]: pam_unix(sshd:session): session closed for user core
Jul  2 00:03:08.339799 systemd[1]: sshd@25-10.0.0.39:22-10.0.0.1:50982.service: Deactivated successfully.
Jul  2 00:03:08.342822 systemd[1]: session-26.scope: Deactivated successfully.
Jul  2 00:03:08.344467 systemd-logind[1423]: Session 26 logged out. Waiting for processes to exit.
Jul  2 00:03:08.353656 systemd[1]: Started sshd@26-10.0.0.39:22-10.0.0.1:50998.service - OpenSSH per-connection server daemon (10.0.0.1:50998).
Jul  2 00:03:08.354892 systemd-logind[1423]: Removed session 26.
Jul  2 00:03:08.383316 kubelet[2518]: I0702 00:03:08.383191    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-xtables-lock\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383316 kubelet[2518]: I0702 00:03:08.383238    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-hostproc\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383316 kubelet[2518]: I0702 00:03:08.383265    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/395ee682-451c-453b-a9d1-833f7952e747-hubble-tls\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383492 kubelet[2518]: I0702 00:03:08.383334    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-cilium-run\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383492 kubelet[2518]: I0702 00:03:08.383379    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-cilium-cgroup\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383492 kubelet[2518]: I0702 00:03:08.383409    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-lib-modules\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383492 kubelet[2518]: I0702 00:03:08.383425    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/395ee682-451c-453b-a9d1-833f7952e747-cilium-config-path\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383492 kubelet[2518]: I0702 00:03:08.383451    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-host-proc-sys-kernel\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383492 kubelet[2518]: I0702 00:03:08.383473    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-cni-path\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383778 kubelet[2518]: I0702 00:03:08.383491    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-host-proc-sys-net\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383778 kubelet[2518]: I0702 00:03:08.383510    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-etc-cni-netd\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383778 kubelet[2518]: I0702 00:03:08.383534    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/395ee682-451c-453b-a9d1-833f7952e747-cilium-ipsec-secrets\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383778 kubelet[2518]: I0702 00:03:08.383553    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/395ee682-451c-453b-a9d1-833f7952e747-bpf-maps\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383778 kubelet[2518]: I0702 00:03:08.383568    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqmh4\" (UniqueName: \"kubernetes.io/projected/395ee682-451c-453b-a9d1-833f7952e747-kube-api-access-mqmh4\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.383886 sshd[4351]: Accepted publickey for core from 10.0.0.1 port 50998 ssh2: RSA SHA256:Et/UiMXmFMbY2cyXsriYvaFlh38PhzkKrD1eNEeM82U
Jul  2 00:03:08.384150 kubelet[2518]: I0702 00:03:08.383637    2518 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/395ee682-451c-453b-a9d1-833f7952e747-clustermesh-secrets\") pod \"cilium-h4mwg\" (UID: \"395ee682-451c-453b-a9d1-833f7952e747\") " pod="kube-system/cilium-h4mwg"
Jul  2 00:03:08.385207 sshd[4351]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jul  2 00:03:08.389156 systemd-logind[1423]: New session 27 of user core.
Jul  2 00:03:08.400449 systemd[1]: Started session-27.scope - Session 27 of User core.
Jul  2 00:03:08.481865 kubelet[2518]: E0702 00:03:08.481754    2518 kubelet.go:2900] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"
Jul  2 00:03:08.554657 kubelet[2518]: E0702 00:03:08.554619    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:08.555252 containerd[1431]: time="2024-07-02T00:03:08.555212765Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-h4mwg,Uid:395ee682-451c-453b-a9d1-833f7952e747,Namespace:kube-system,Attempt:0,}"
Jul  2 00:03:08.572104 containerd[1431]: time="2024-07-02T00:03:08.572006181Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jul  2 00:03:08.572104 containerd[1431]: time="2024-07-02T00:03:08.572068102Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:03:08.572104 containerd[1431]: time="2024-07-02T00:03:08.572094063Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jul  2 00:03:08.572268 containerd[1431]: time="2024-07-02T00:03:08.572110343Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jul  2 00:03:08.589536 systemd[1]: Started cri-containerd-af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304.scope - libcontainer container af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304.
Jul  2 00:03:08.610061 containerd[1431]: time="2024-07-02T00:03:08.610020951Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-h4mwg,Uid:395ee682-451c-453b-a9d1-833f7952e747,Namespace:kube-system,Attempt:0,} returns sandbox id \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\""
Jul  2 00:03:08.610680 kubelet[2518]: E0702 00:03:08.610657    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:08.612688 containerd[1431]: time="2024-07-02T00:03:08.612572928Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}"
Jul  2 00:03:08.622056 containerd[1431]: time="2024-07-02T00:03:08.621998739Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"7f5393e4a346ca9d53649d6abc80a98de09adaf114b1638c90bce7ab054cc05d\""
Jul  2 00:03:08.622472 containerd[1431]: time="2024-07-02T00:03:08.622448829Z" level=info msg="StartContainer for \"7f5393e4a346ca9d53649d6abc80a98de09adaf114b1638c90bce7ab054cc05d\""
Jul  2 00:03:08.647473 systemd[1]: Started cri-containerd-7f5393e4a346ca9d53649d6abc80a98de09adaf114b1638c90bce7ab054cc05d.scope - libcontainer container 7f5393e4a346ca9d53649d6abc80a98de09adaf114b1638c90bce7ab054cc05d.
Jul  2 00:03:08.673344 containerd[1431]: time="2024-07-02T00:03:08.673232644Z" level=info msg="StartContainer for \"7f5393e4a346ca9d53649d6abc80a98de09adaf114b1638c90bce7ab054cc05d\" returns successfully"
Jul  2 00:03:08.686465 systemd[1]: cri-containerd-7f5393e4a346ca9d53649d6abc80a98de09adaf114b1638c90bce7ab054cc05d.scope: Deactivated successfully.
Jul  2 00:03:08.688319 kubelet[2518]: E0702 00:03:08.687423    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:08.720205 containerd[1431]: time="2024-07-02T00:03:08.720124573Z" level=info msg="shim disconnected" id=7f5393e4a346ca9d53649d6abc80a98de09adaf114b1638c90bce7ab054cc05d namespace=k8s.io
Jul  2 00:03:08.720205 containerd[1431]: time="2024-07-02T00:03:08.720189654Z" level=warning msg="cleaning up after shim disconnected" id=7f5393e4a346ca9d53649d6abc80a98de09adaf114b1638c90bce7ab054cc05d namespace=k8s.io
Jul  2 00:03:08.720205 containerd[1431]: time="2024-07-02T00:03:08.720198215Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:03:09.690335 kubelet[2518]: E0702 00:03:09.690175    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:09.692975 containerd[1431]: time="2024-07-02T00:03:09.692928660Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}"
Jul  2 00:03:09.705015 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2372384415.mount: Deactivated successfully.
Jul  2 00:03:09.707111 containerd[1431]: time="2024-07-02T00:03:09.706939368Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"62af89aec6a5034e034fb9f6cd516132013e6dbc1c218ab2d69147cf95c9d1e6\""
Jul  2 00:03:09.707544 containerd[1431]: time="2024-07-02T00:03:09.707520061Z" level=info msg="StartContainer for \"62af89aec6a5034e034fb9f6cd516132013e6dbc1c218ab2d69147cf95c9d1e6\""
Jul  2 00:03:09.741522 systemd[1]: Started cri-containerd-62af89aec6a5034e034fb9f6cd516132013e6dbc1c218ab2d69147cf95c9d1e6.scope - libcontainer container 62af89aec6a5034e034fb9f6cd516132013e6dbc1c218ab2d69147cf95c9d1e6.
Jul  2 00:03:09.765270 containerd[1431]: time="2024-07-02T00:03:09.765224409Z" level=info msg="StartContainer for \"62af89aec6a5034e034fb9f6cd516132013e6dbc1c218ab2d69147cf95c9d1e6\" returns successfully"
Jul  2 00:03:09.774840 systemd[1]: cri-containerd-62af89aec6a5034e034fb9f6cd516132013e6dbc1c218ab2d69147cf95c9d1e6.scope: Deactivated successfully.
Jul  2 00:03:09.816864 containerd[1431]: time="2024-07-02T00:03:09.816796862Z" level=info msg="shim disconnected" id=62af89aec6a5034e034fb9f6cd516132013e6dbc1c218ab2d69147cf95c9d1e6 namespace=k8s.io
Jul  2 00:03:09.817281 containerd[1431]: time="2024-07-02T00:03:09.817104669Z" level=warning msg="cleaning up after shim disconnected" id=62af89aec6a5034e034fb9f6cd516132013e6dbc1c218ab2d69147cf95c9d1e6 namespace=k8s.io
Jul  2 00:03:09.817281 containerd[1431]: time="2024-07-02T00:03:09.817122190Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:03:10.410559 kubelet[2518]: I0702 00:03:10.410503    2518 setters.go:580] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-07-02T00:03:10Z","lastTransitionTime":"2024-07-02T00:03:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"}
Jul  2 00:03:10.491702 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-62af89aec6a5034e034fb9f6cd516132013e6dbc1c218ab2d69147cf95c9d1e6-rootfs.mount: Deactivated successfully.
Jul  2 00:03:10.693956 kubelet[2518]: E0702 00:03:10.693838    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:10.696645 containerd[1431]: time="2024-07-02T00:03:10.695993565Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}"
Jul  2 00:03:10.714424 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount168690435.mount: Deactivated successfully.
Jul  2 00:03:10.718028 containerd[1431]: time="2024-07-02T00:03:10.717941039Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"35dae686ae5f3e00a620ecef52d5696b7d235f7f34e5288ebee42feab3abdba4\""
Jul  2 00:03:10.718558 containerd[1431]: time="2024-07-02T00:03:10.718516931Z" level=info msg="StartContainer for \"35dae686ae5f3e00a620ecef52d5696b7d235f7f34e5288ebee42feab3abdba4\""
Jul  2 00:03:10.751532 systemd[1]: Started cri-containerd-35dae686ae5f3e00a620ecef52d5696b7d235f7f34e5288ebee42feab3abdba4.scope - libcontainer container 35dae686ae5f3e00a620ecef52d5696b7d235f7f34e5288ebee42feab3abdba4.
Jul  2 00:03:10.777976 systemd[1]: cri-containerd-35dae686ae5f3e00a620ecef52d5696b7d235f7f34e5288ebee42feab3abdba4.scope: Deactivated successfully.
Jul  2 00:03:10.779583 containerd[1431]: time="2024-07-02T00:03:10.779454008Z" level=info msg="StartContainer for \"35dae686ae5f3e00a620ecef52d5696b7d235f7f34e5288ebee42feab3abdba4\" returns successfully"
Jul  2 00:03:10.803898 containerd[1431]: time="2024-07-02T00:03:10.803824134Z" level=info msg="shim disconnected" id=35dae686ae5f3e00a620ecef52d5696b7d235f7f34e5288ebee42feab3abdba4 namespace=k8s.io
Jul  2 00:03:10.803898 containerd[1431]: time="2024-07-02T00:03:10.803879815Z" level=warning msg="cleaning up after shim disconnected" id=35dae686ae5f3e00a620ecef52d5696b7d235f7f34e5288ebee42feab3abdba4 namespace=k8s.io
Jul  2 00:03:10.803898 containerd[1431]: time="2024-07-02T00:03:10.803892216Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:03:11.423023 kubelet[2518]: E0702 00:03:11.422937    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:11.491829 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-35dae686ae5f3e00a620ecef52d5696b7d235f7f34e5288ebee42feab3abdba4-rootfs.mount: Deactivated successfully.
Jul  2 00:03:11.703696 kubelet[2518]: E0702 00:03:11.701798    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:11.706414 containerd[1431]: time="2024-07-02T00:03:11.706024690Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}"
Jul  2 00:03:11.729868 containerd[1431]: time="2024-07-02T00:03:11.729692433Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"5825a721d38bd90e17aa0237a0b8180da9b78ebb55a5a31f9333e5c939a12b0b\""
Jul  2 00:03:11.730699 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount291728973.mount: Deactivated successfully.
Jul  2 00:03:11.733575 containerd[1431]: time="2024-07-02T00:03:11.733520514Z" level=info msg="StartContainer for \"5825a721d38bd90e17aa0237a0b8180da9b78ebb55a5a31f9333e5c939a12b0b\""
Jul  2 00:03:11.760489 systemd[1]: Started cri-containerd-5825a721d38bd90e17aa0237a0b8180da9b78ebb55a5a31f9333e5c939a12b0b.scope - libcontainer container 5825a721d38bd90e17aa0237a0b8180da9b78ebb55a5a31f9333e5c939a12b0b.
Jul  2 00:03:11.782507 systemd[1]: cri-containerd-5825a721d38bd90e17aa0237a0b8180da9b78ebb55a5a31f9333e5c939a12b0b.scope: Deactivated successfully.
Jul  2 00:03:11.785316 containerd[1431]: time="2024-07-02T00:03:11.784961727Z" level=info msg="StartContainer for \"5825a721d38bd90e17aa0237a0b8180da9b78ebb55a5a31f9333e5c939a12b0b\" returns successfully"
Jul  2 00:03:11.813723 containerd[1431]: time="2024-07-02T00:03:11.813661217Z" level=info msg="shim disconnected" id=5825a721d38bd90e17aa0237a0b8180da9b78ebb55a5a31f9333e5c939a12b0b namespace=k8s.io
Jul  2 00:03:11.813723 containerd[1431]: time="2024-07-02T00:03:11.813718538Z" level=warning msg="cleaning up after shim disconnected" id=5825a721d38bd90e17aa0237a0b8180da9b78ebb55a5a31f9333e5c939a12b0b namespace=k8s.io
Jul  2 00:03:11.813723 containerd[1431]: time="2024-07-02T00:03:11.813726538Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jul  2 00:03:12.492119 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5825a721d38bd90e17aa0237a0b8180da9b78ebb55a5a31f9333e5c939a12b0b-rootfs.mount: Deactivated successfully.
Jul  2 00:03:12.703648 kubelet[2518]: E0702 00:03:12.703605    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:12.706854 containerd[1431]: time="2024-07-02T00:03:12.706797741Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}"
Jul  2 00:03:12.725241 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3443727438.mount: Deactivated successfully.
Jul  2 00:03:12.735782 containerd[1431]: time="2024-07-02T00:03:12.735727666Z" level=info msg="CreateContainer within sandbox \"af0de69845bf354dc60205fac43a4caee3847b2d5d8e07ae6917ff6035130304\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"5b3c2fc52a5e15f89ff0d557d1dc96d65521f4524fcee8fccc80ccfadebde1ef\""
Jul  2 00:03:12.736769 containerd[1431]: time="2024-07-02T00:03:12.736670885Z" level=info msg="StartContainer for \"5b3c2fc52a5e15f89ff0d557d1dc96d65521f4524fcee8fccc80ccfadebde1ef\""
Jul  2 00:03:12.769532 systemd[1]: Started cri-containerd-5b3c2fc52a5e15f89ff0d557d1dc96d65521f4524fcee8fccc80ccfadebde1ef.scope - libcontainer container 5b3c2fc52a5e15f89ff0d557d1dc96d65521f4524fcee8fccc80ccfadebde1ef.
Jul  2 00:03:12.795903 containerd[1431]: time="2024-07-02T00:03:12.795854162Z" level=info msg="StartContainer for \"5b3c2fc52a5e15f89ff0d557d1dc96d65521f4524fcee8fccc80ccfadebde1ef\" returns successfully"
Jul  2 00:03:13.092340 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce))
Jul  2 00:03:13.708660 kubelet[2518]: E0702 00:03:13.708625    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:14.711076 kubelet[2518]: E0702 00:03:14.710957    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:16.027899 systemd-networkd[1368]: lxc_health: Link UP
Jul  2 00:03:16.046198 systemd-networkd[1368]: lxc_health: Gained carrier
Jul  2 00:03:16.557649 kubelet[2518]: E0702 00:03:16.557593    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:16.575609 kubelet[2518]: I0702 00:03:16.575474    2518 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/cilium-h4mwg" podStartSLOduration=8.575454467 podStartE2EDuration="8.575454467s" podCreationTimestamp="2024-07-02 00:03:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-07-02 00:03:13.726280834 +0000 UTC m=+95.402697528" watchObservedRunningTime="2024-07-02 00:03:16.575454467 +0000 UTC m=+98.251871121"
Jul  2 00:03:16.717856 kubelet[2518]: E0702 00:03:16.717812    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:17.423136 kubelet[2518]: E0702 00:03:17.423084    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:17.686459 systemd-networkd[1368]: lxc_health: Gained IPv6LL
Jul  2 00:03:17.719624 kubelet[2518]: E0702 00:03:17.719584    2518 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jul  2 00:03:21.264123 sshd[4351]: pam_unix(sshd:session): session closed for user core
Jul  2 00:03:21.270414 systemd[1]: sshd@26-10.0.0.39:22-10.0.0.1:50998.service: Deactivated successfully.
Jul  2 00:03:21.274629 systemd[1]: session-27.scope: Deactivated successfully.
Jul  2 00:03:21.276529 systemd-logind[1423]: Session 27 logged out. Waiting for processes to exit.
Jul  2 00:03:21.277698 systemd-logind[1423]: Removed session 27.