Jun 25 18:46:37.920493 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1]
Jun 25 18:46:37.920513 kernel: Linux version 6.6.35-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.2.1_p20240210 p14) 13.2.1 20240210, GNU ld (Gentoo 2.41 p5) 2.41.0) #1 SMP PREEMPT Tue Jun 25 17:19:03 -00 2024
Jun 25 18:46:37.920523 kernel: KASLR enabled
Jun 25 18:46:37.920529 kernel: efi: EFI v2.7 by EDK II
Jun 25 18:46:37.920534 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdb8fd018 ACPI 2.0=0xd9710018 RNG=0xd971e498 MEMRESERVE=0xd9b43d18 
Jun 25 18:46:37.920540 kernel: random: crng init done
Jun 25 18:46:37.920547 kernel: ACPI: Early table checksum verification disabled
Jun 25 18:46:37.920553 kernel: ACPI: RSDP 0x00000000D9710018 000024 (v02 BOCHS )
Jun 25 18:46:37.920559 kernel: ACPI: XSDT 0x00000000D971FE98 000064 (v01 BOCHS  BXPC     00000001      01000013)
Jun 25 18:46:37.920566 kernel: ACPI: FACP 0x00000000D971FA98 000114 (v06 BOCHS  BXPC     00000001 BXPC 00000001)
Jun 25 18:46:37.920572 kernel: ACPI: DSDT 0x00000000D9717518 0014A2 (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jun 25 18:46:37.920578 kernel: ACPI: APIC 0x00000000D971FC18 0001A8 (v04 BOCHS  BXPC     00000001 BXPC 00000001)
Jun 25 18:46:37.920584 kernel: ACPI: PPTT 0x00000000D971D898 00009C (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jun 25 18:46:37.920590 kernel: ACPI: GTDT 0x00000000D971E818 000060 (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jun 25 18:46:37.920598 kernel: ACPI: MCFG 0x00000000D971E918 00003C (v01 BOCHS  BXPC     00000001 BXPC 00000001)
Jun 25 18:46:37.920605 kernel: ACPI: SPCR 0x00000000D971FF98 000050 (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jun 25 18:46:37.920612 kernel: ACPI: DBG2 0x00000000D971E418 000057 (v00 BOCHS  BXPC     00000001 BXPC 00000001)
Jun 25 18:46:37.920618 kernel: ACPI: IORT 0x00000000D971E718 000080 (v03 BOCHS  BXPC     00000001 BXPC 00000001)
Jun 25 18:46:37.920624 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600
Jun 25 18:46:37.920630 kernel: NUMA: Failed to initialise from firmware
Jun 25 18:46:37.920637 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff]
Jun 25 18:46:37.920643 kernel: NUMA: NODE_DATA [mem 0xdc957800-0xdc95cfff]
Jun 25 18:46:37.920649 kernel: Zone ranges:
Jun 25 18:46:37.920655 kernel:   DMA      [mem 0x0000000040000000-0x00000000dcffffff]
Jun 25 18:46:37.920662 kernel:   DMA32    empty
Jun 25 18:46:37.920669 kernel:   Normal   empty
Jun 25 18:46:37.920675 kernel: Movable zone start for each node
Jun 25 18:46:37.920682 kernel: Early memory node ranges
Jun 25 18:46:37.920688 kernel:   node   0: [mem 0x0000000040000000-0x00000000d976ffff]
Jun 25 18:46:37.920694 kernel:   node   0: [mem 0x00000000d9770000-0x00000000d9b3ffff]
Jun 25 18:46:37.920700 kernel:   node   0: [mem 0x00000000d9b40000-0x00000000dce1ffff]
Jun 25 18:46:37.920706 kernel:   node   0: [mem 0x00000000dce20000-0x00000000dceaffff]
Jun 25 18:46:37.920713 kernel:   node   0: [mem 0x00000000dceb0000-0x00000000dcebffff]
Jun 25 18:46:37.920719 kernel:   node   0: [mem 0x00000000dcec0000-0x00000000dcfdffff]
Jun 25 18:46:37.920725 kernel:   node   0: [mem 0x00000000dcfe0000-0x00000000dcffffff]
Jun 25 18:46:37.920731 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff]
Jun 25 18:46:37.920738 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges
Jun 25 18:46:37.920745 kernel: psci: probing for conduit method from ACPI.
Jun 25 18:46:37.920752 kernel: psci: PSCIv1.1 detected in firmware.
Jun 25 18:46:37.920758 kernel: psci: Using standard PSCI v0.2 function IDs
Jun 25 18:46:37.920767 kernel: psci: Trusted OS migration not required
Jun 25 18:46:37.920774 kernel: psci: SMC Calling Convention v1.1
Jun 25 18:46:37.920781 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003)
Jun 25 18:46:37.920789 kernel: percpu: Embedded 31 pages/cpu s86632 r8192 d32152 u126976
Jun 25 18:46:37.920796 kernel: pcpu-alloc: s86632 r8192 d32152 u126976 alloc=31*4096
Jun 25 18:46:37.920802 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 
Jun 25 18:46:37.920809 kernel: Detected PIPT I-cache on CPU0
Jun 25 18:46:37.920816 kernel: CPU features: detected: GIC system register CPU interface
Jun 25 18:46:37.920823 kernel: CPU features: detected: Hardware dirty bit management
Jun 25 18:46:37.920829 kernel: CPU features: detected: Spectre-v4
Jun 25 18:46:37.920836 kernel: CPU features: detected: Spectre-BHB
Jun 25 18:46:37.920843 kernel: CPU features: kernel page table isolation forced ON by KASLR
Jun 25 18:46:37.920849 kernel: CPU features: detected: Kernel page table isolation (KPTI)
Jun 25 18:46:37.920858 kernel: CPU features: detected: ARM erratum 1418040
Jun 25 18:46:37.920865 kernel: alternatives: applying boot alternatives
Jun 25 18:46:37.920872 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=e6069a8408a0ca7e7bc40a0bde7fe3ef89df2f98c4bdd2e7e7f9f8f3f8ad207f
Jun 25 18:46:37.920880 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space.
Jun 25 18:46:37.920886 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear)
Jun 25 18:46:37.920893 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear)
Jun 25 18:46:37.920900 kernel: Fallback order for Node 0: 0 
Jun 25 18:46:37.920907 kernel: Built 1 zonelists, mobility grouping on.  Total pages: 633024
Jun 25 18:46:37.920913 kernel: Policy zone: DMA
Jun 25 18:46:37.920920 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off
Jun 25 18:46:37.920927 kernel: software IO TLB: area num 4.
Jun 25 18:46:37.920935 kernel: software IO TLB: mapped [mem 0x00000000d2e00000-0x00000000d6e00000] (64MB)
Jun 25 18:46:37.920942 kernel: Memory: 2386848K/2572288K available (10240K kernel code, 2182K rwdata, 8072K rodata, 39040K init, 897K bss, 185440K reserved, 0K cma-reserved)
Jun 25 18:46:37.920949 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1
Jun 25 18:46:37.920956 kernel: trace event string verifier disabled
Jun 25 18:46:37.920962 kernel: rcu: Preemptible hierarchical RCU implementation.
Jun 25 18:46:37.920970 kernel: rcu:         RCU event tracing is enabled.
Jun 25 18:46:37.920976 kernel: rcu:         RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4.
Jun 25 18:46:37.920983 kernel:         Trampoline variant of Tasks RCU enabled.
Jun 25 18:46:37.920990 kernel:         Tracing variant of Tasks RCU enabled.
Jun 25 18:46:37.920997 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies.
Jun 25 18:46:37.921004 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4
Jun 25 18:46:37.921011 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0
Jun 25 18:46:37.921019 kernel: GICv3: 256 SPIs implemented
Jun 25 18:46:37.921026 kernel: GICv3: 0 Extended SPIs implemented
Jun 25 18:46:37.921032 kernel: Root IRQ handler: gic_handle_irq
Jun 25 18:46:37.921039 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI
Jun 25 18:46:37.921046 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000
Jun 25 18:46:37.921052 kernel: ITS [mem 0x08080000-0x0809ffff]
Jun 25 18:46:37.921059 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400d0000 (indirect, esz 8, psz 64K, shr 1)
Jun 25 18:46:37.921066 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400e0000 (flat, esz 8, psz 64K, shr 1)
Jun 25 18:46:37.921073 kernel: GICv3: using LPI property table @0x00000000400f0000
Jun 25 18:46:37.921079 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040100000
Jun 25 18:46:37.921086 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention.
Jun 25 18:46:37.921094 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jun 25 18:46:37.921101 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt).
Jun 25 18:46:37.921108 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns
Jun 25 18:46:37.921115 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns
Jun 25 18:46:37.921122 kernel: arm-pv: using stolen time PV
Jun 25 18:46:37.921128 kernel: Console: colour dummy device 80x25
Jun 25 18:46:37.921136 kernel: ACPI: Core revision 20230628
Jun 25 18:46:37.921150 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000)
Jun 25 18:46:37.921159 kernel: pid_max: default: 32768 minimum: 301
Jun 25 18:46:37.921166 kernel: LSM: initializing lsm=lockdown,capability,selinux,integrity
Jun 25 18:46:37.921174 kernel: SELinux:  Initializing.
Jun 25 18:46:37.921181 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear)
Jun 25 18:46:37.921189 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear)
Jun 25 18:46:37.921196 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1.
Jun 25 18:46:37.921203 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1.
Jun 25 18:46:37.921223 kernel: rcu: Hierarchical SRCU implementation.
Jun 25 18:46:37.921230 kernel: rcu:         Max phase no-delay instances is 400.
Jun 25 18:46:37.921237 kernel: Platform MSI: ITS@0x8080000 domain created
Jun 25 18:46:37.921244 kernel: PCI/MSI: ITS@0x8080000 domain created
Jun 25 18:46:37.921253 kernel: Remapping and enabling EFI services.
Jun 25 18:46:37.921260 kernel: smp: Bringing up secondary CPUs ...
Jun 25 18:46:37.921266 kernel: Detected PIPT I-cache on CPU1
Jun 25 18:46:37.921273 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000
Jun 25 18:46:37.921280 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040110000
Jun 25 18:46:37.921287 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jun 25 18:46:37.921294 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1]
Jun 25 18:46:37.921301 kernel: Detected PIPT I-cache on CPU2
Jun 25 18:46:37.921308 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000
Jun 25 18:46:37.921315 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040120000
Jun 25 18:46:37.921323 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jun 25 18:46:37.921330 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1]
Jun 25 18:46:37.921341 kernel: Detected PIPT I-cache on CPU3
Jun 25 18:46:37.921350 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000
Jun 25 18:46:37.921357 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040130000
Jun 25 18:46:37.921364 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jun 25 18:46:37.921371 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1]
Jun 25 18:46:37.921379 kernel: smp: Brought up 1 node, 4 CPUs
Jun 25 18:46:37.921386 kernel: SMP: Total of 4 processors activated.
Jun 25 18:46:37.921395 kernel: CPU features: detected: 32-bit EL0 Support
Jun 25 18:46:37.921402 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence
Jun 25 18:46:37.921409 kernel: CPU features: detected: Common not Private translations
Jun 25 18:46:37.921417 kernel: CPU features: detected: CRC32 instructions
Jun 25 18:46:37.921424 kernel: CPU features: detected: Enhanced Virtualization Traps
Jun 25 18:46:37.921431 kernel: CPU features: detected: RCpc load-acquire (LDAPR)
Jun 25 18:46:37.921438 kernel: CPU features: detected: LSE atomic instructions
Jun 25 18:46:37.921446 kernel: CPU features: detected: Privileged Access Never
Jun 25 18:46:37.921454 kernel: CPU features: detected: RAS Extension Support
Jun 25 18:46:37.921462 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS)
Jun 25 18:46:37.921469 kernel: CPU: All CPU(s) started at EL1
Jun 25 18:46:37.921476 kernel: alternatives: applying system-wide alternatives
Jun 25 18:46:37.921483 kernel: devtmpfs: initialized
Jun 25 18:46:37.921491 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns
Jun 25 18:46:37.921498 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear)
Jun 25 18:46:37.921505 kernel: pinctrl core: initialized pinctrl subsystem
Jun 25 18:46:37.921512 kernel: SMBIOS 3.0.0 present.
Jun 25 18:46:37.921521 kernel: DMI: QEMU KVM Virtual Machine, BIOS edk2-20230524-3.fc38 05/24/2023
Jun 25 18:46:37.921528 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family
Jun 25 18:46:37.921535 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations
Jun 25 18:46:37.921543 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations
Jun 25 18:46:37.921550 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations
Jun 25 18:46:37.921557 kernel: audit: initializing netlink subsys (disabled)
Jun 25 18:46:37.921565 kernel: audit: type=2000 audit(0.024:1): state=initialized audit_enabled=0 res=1
Jun 25 18:46:37.921572 kernel: thermal_sys: Registered thermal governor 'step_wise'
Jun 25 18:46:37.921579 kernel: cpuidle: using governor menu
Jun 25 18:46:37.921588 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers.
Jun 25 18:46:37.921595 kernel: ASID allocator initialised with 32768 entries
Jun 25 18:46:37.921602 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5
Jun 25 18:46:37.921610 kernel: Serial: AMBA PL011 UART driver
Jun 25 18:46:37.921617 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL
Jun 25 18:46:37.921624 kernel: Modules: 0 pages in range for non-PLT usage
Jun 25 18:46:37.921631 kernel: Modules: 509120 pages in range for PLT usage
Jun 25 18:46:37.921638 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages
Jun 25 18:46:37.921646 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page
Jun 25 18:46:37.921655 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages
Jun 25 18:46:37.921662 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page
Jun 25 18:46:37.921669 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages
Jun 25 18:46:37.921676 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page
Jun 25 18:46:37.921684 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages
Jun 25 18:46:37.921691 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page
Jun 25 18:46:37.921698 kernel: ACPI: Added _OSI(Module Device)
Jun 25 18:46:37.921705 kernel: ACPI: Added _OSI(Processor Device)
Jun 25 18:46:37.921712 kernel: ACPI: Added _OSI(3.0 _SCP Extensions)
Jun 25 18:46:37.921721 kernel: ACPI: Added _OSI(Processor Aggregator Device)
Jun 25 18:46:37.921728 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded
Jun 25 18:46:37.921735 kernel: ACPI: Interpreter enabled
Jun 25 18:46:37.921742 kernel: ACPI: Using GIC for interrupt routing
Jun 25 18:46:37.921750 kernel: ACPI: MCFG table detected, 1 entries
Jun 25 18:46:37.921757 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA
Jun 25 18:46:37.921764 kernel: printk: console [ttyAMA0] enabled
Jun 25 18:46:37.921772 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff])
Jun 25 18:46:37.921903 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3]
Jun 25 18:46:37.921978 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR]
Jun 25 18:46:37.922044 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability]
Jun 25 18:46:37.922107 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00
Jun 25 18:46:37.922186 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff]
Jun 25 18:46:37.922197 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io  0x0000-0xffff window]
Jun 25 18:46:37.922251 kernel: PCI host bridge to bus 0000:00
Jun 25 18:46:37.922330 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window]
Jun 25 18:46:37.922396 kernel: pci_bus 0000:00: root bus resource [io  0x0000-0xffff window]
Jun 25 18:46:37.922454 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window]
Jun 25 18:46:37.922511 kernel: pci_bus 0000:00: root bus resource [bus 00-ff]
Jun 25 18:46:37.922588 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000
Jun 25 18:46:37.922661 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00
Jun 25 18:46:37.922728 kernel: pci 0000:00:01.0: reg 0x10: [io  0x0000-0x001f]
Jun 25 18:46:37.922795 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff]
Jun 25 18:46:37.922861 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref]
Jun 25 18:46:37.922928 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref]
Jun 25 18:46:37.922992 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff]
Jun 25 18:46:37.923057 kernel: pci 0000:00:01.0: BAR 0: assigned [io  0x1000-0x101f]
Jun 25 18:46:37.923115 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window]
Jun 25 18:46:37.923185 kernel: pci_bus 0000:00: resource 5 [io  0x0000-0xffff window]
Jun 25 18:46:37.923257 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window]
Jun 25 18:46:37.923267 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35
Jun 25 18:46:37.923274 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36
Jun 25 18:46:37.923281 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37
Jun 25 18:46:37.923289 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38
Jun 25 18:46:37.923296 kernel: iommu: Default domain type: Translated
Jun 25 18:46:37.923303 kernel: iommu: DMA domain TLB invalidation policy: strict mode
Jun 25 18:46:37.923310 kernel: efivars: Registered efivars operations
Jun 25 18:46:37.923318 kernel: vgaarb: loaded
Jun 25 18:46:37.923327 kernel: clocksource: Switched to clocksource arch_sys_counter
Jun 25 18:46:37.923334 kernel: VFS: Disk quotas dquot_6.6.0
Jun 25 18:46:37.923341 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes)
Jun 25 18:46:37.923348 kernel: pnp: PnP ACPI init
Jun 25 18:46:37.923424 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved
Jun 25 18:46:37.923435 kernel: pnp: PnP ACPI: found 1 devices
Jun 25 18:46:37.923442 kernel: NET: Registered PF_INET protocol family
Jun 25 18:46:37.923450 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear)
Jun 25 18:46:37.923459 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear)
Jun 25 18:46:37.923466 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear)
Jun 25 18:46:37.923474 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear)
Jun 25 18:46:37.923481 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear)
Jun 25 18:46:37.923489 kernel: TCP: Hash tables configured (established 32768 bind 32768)
Jun 25 18:46:37.923496 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear)
Jun 25 18:46:37.923503 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear)
Jun 25 18:46:37.923511 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family
Jun 25 18:46:37.923518 kernel: PCI: CLS 0 bytes, default 64
Jun 25 18:46:37.923527 kernel: kvm [1]: HYP mode not available
Jun 25 18:46:37.923534 kernel: Initialise system trusted keyrings
Jun 25 18:46:37.923541 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0
Jun 25 18:46:37.923548 kernel: Key type asymmetric registered
Jun 25 18:46:37.923555 kernel: Asymmetric key parser 'x509' registered
Jun 25 18:46:37.923563 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250)
Jun 25 18:46:37.923570 kernel: io scheduler mq-deadline registered
Jun 25 18:46:37.923577 kernel: io scheduler kyber registered
Jun 25 18:46:37.923584 kernel: io scheduler bfq registered
Jun 25 18:46:37.923593 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0
Jun 25 18:46:37.923600 kernel: ACPI: button: Power Button [PWRB]
Jun 25 18:46:37.923607 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36
Jun 25 18:46:37.923674 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007)
Jun 25 18:46:37.923684 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled
Jun 25 18:46:37.923691 kernel: thunder_xcv, ver 1.0
Jun 25 18:46:37.923698 kernel: thunder_bgx, ver 1.0
Jun 25 18:46:37.923705 kernel: nicpf, ver 1.0
Jun 25 18:46:37.923712 kernel: nicvf, ver 1.0
Jun 25 18:46:37.923787 kernel: rtc-efi rtc-efi.0: registered as rtc0
Jun 25 18:46:37.923849 kernel: rtc-efi rtc-efi.0: setting system clock to 2024-06-25T18:46:37 UTC (1719341197)
Jun 25 18:46:37.923858 kernel: hid: raw HID events driver (C) Jiri Kosina
Jun 25 18:46:37.923866 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available
Jun 25 18:46:37.923873 kernel: watchdog: Delayed init of the lockup detector failed: -19
Jun 25 18:46:37.923881 kernel: watchdog: Hard watchdog permanently disabled
Jun 25 18:46:37.923888 kernel: NET: Registered PF_INET6 protocol family
Jun 25 18:46:37.923895 kernel: Segment Routing with IPv6
Jun 25 18:46:37.923904 kernel: In-situ OAM (IOAM) with IPv6
Jun 25 18:46:37.923912 kernel: NET: Registered PF_PACKET protocol family
Jun 25 18:46:37.923919 kernel: Key type dns_resolver registered
Jun 25 18:46:37.923926 kernel: registered taskstats version 1
Jun 25 18:46:37.923933 kernel: Loading compiled-in X.509 certificates
Jun 25 18:46:37.923941 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.35-flatcar: 751918e575d02f96b0daadd44b8f442a8c39ecd3'
Jun 25 18:46:37.923948 kernel: Key type .fscrypt registered
Jun 25 18:46:37.923955 kernel: Key type fscrypt-provisioning registered
Jun 25 18:46:37.923963 kernel: ima: No TPM chip found, activating TPM-bypass!
Jun 25 18:46:37.923971 kernel: ima: Allocated hash algorithm: sha1
Jun 25 18:46:37.923979 kernel: ima: No architecture policies found
Jun 25 18:46:37.923987 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng)
Jun 25 18:46:37.923994 kernel: clk: Disabling unused clocks
Jun 25 18:46:37.924001 kernel: Freeing unused kernel memory: 39040K
Jun 25 18:46:37.924008 kernel: Run /init as init process
Jun 25 18:46:37.924015 kernel:   with arguments:
Jun 25 18:46:37.924022 kernel:     /init
Jun 25 18:46:37.924030 kernel:   with environment:
Jun 25 18:46:37.924038 kernel:     HOME=/
Jun 25 18:46:37.924045 kernel:     TERM=linux
Jun 25 18:46:37.924052 kernel:     BOOT_IMAGE=/flatcar/vmlinuz-a
Jun 25 18:46:37.924061 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified)
Jun 25 18:46:37.924070 systemd[1]: Detected virtualization kvm.
Jun 25 18:46:37.924078 systemd[1]: Detected architecture arm64.
Jun 25 18:46:37.924086 systemd[1]: Running in initrd.
Jun 25 18:46:37.924095 systemd[1]: No hostname configured, using default hostname.
Jun 25 18:46:37.924102 systemd[1]: Hostname set to <localhost>.
Jun 25 18:46:37.924110 systemd[1]: Initializing machine ID from VM UUID.
Jun 25 18:46:37.924118 systemd[1]: Queued start job for default target initrd.target.
Jun 25 18:46:37.924126 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Jun 25 18:46:37.924134 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Jun 25 18:46:37.924142 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM...
Jun 25 18:46:37.924156 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM...
Jun 25 18:46:37.924166 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT...
Jun 25 18:46:37.924175 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A...
Jun 25 18:46:37.924184 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132...
Jun 25 18:46:37.924192 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr...
Jun 25 18:46:37.924200 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Jun 25 18:46:37.924215 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes.
Jun 25 18:46:37.924223 systemd[1]: Reached target paths.target - Path Units.
Jun 25 18:46:37.924233 systemd[1]: Reached target slices.target - Slice Units.
Jun 25 18:46:37.924240 systemd[1]: Reached target swap.target - Swaps.
Jun 25 18:46:37.924248 systemd[1]: Reached target timers.target - Timer Units.
Jun 25 18:46:37.924256 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket.
Jun 25 18:46:37.924264 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket.
Jun 25 18:46:37.924272 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log).
Jun 25 18:46:37.924280 systemd[1]: Listening on systemd-journald.socket - Journal Socket.
Jun 25 18:46:37.924288 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket.
Jun 25 18:46:37.924296 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket.
Jun 25 18:46:37.924305 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket.
Jun 25 18:46:37.924313 systemd[1]: Reached target sockets.target - Socket Units.
Jun 25 18:46:37.924321 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup...
Jun 25 18:46:37.924328 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes...
Jun 25 18:46:37.924336 systemd[1]: Finished network-cleanup.service - Network Cleanup.
Jun 25 18:46:37.924344 systemd[1]: Starting systemd-fsck-usr.service...
Jun 25 18:46:37.924352 systemd[1]: Starting systemd-journald.service - Journal Service...
Jun 25 18:46:37.924359 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules...
Jun 25 18:46:37.924369 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Jun 25 18:46:37.924376 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup.
Jun 25 18:46:37.924384 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes.
Jun 25 18:46:37.924392 systemd[1]: Finished systemd-fsck-usr.service.
Jun 25 18:46:37.924400 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully...
Jun 25 18:46:37.924427 systemd-journald[237]: Collecting audit messages is disabled.
Jun 25 18:46:37.924446 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Jun 25 18:46:37.924455 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully.
Jun 25 18:46:37.924463 systemd-journald[237]: Journal started
Jun 25 18:46:37.924483 systemd-journald[237]: Runtime Journal (/run/log/journal/a78f43e1c1ce473b961e557db24d7181) is 5.9M, max 47.3M, 41.4M free.
Jun 25 18:46:37.924517 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Jun 25 18:46:37.915977 systemd-modules-load[238]: Inserted module 'overlay'
Jun 25 18:46:37.930225 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev...
Jun 25 18:46:37.930262 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this.
Jun 25 18:46:37.930280 systemd[1]: Started systemd-journald.service - Journal Service.
Jun 25 18:46:37.932446 systemd-modules-load[238]: Inserted module 'br_netfilter'
Jun 25 18:46:37.933262 kernel: Bridge firewalling registered
Jun 25 18:46:37.933562 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules.
Jun 25 18:46:37.935923 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Jun 25 18:46:37.938469 systemd[1]: Starting systemd-tmpfiles-setup.service - Create Volatile Files and Directories...
Jun 25 18:46:37.939463 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Jun 25 18:46:37.948291 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Jun 25 18:46:37.949395 systemd[1]: Finished systemd-tmpfiles-setup.service - Create Volatile Files and Directories.
Jun 25 18:46:37.951345 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jun 25 18:46:37.966401 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook...
Jun 25 18:46:37.968473 systemd[1]: Starting systemd-resolved.service - Network Name Resolution...
Jun 25 18:46:37.975897 dracut-cmdline[275]: dracut-dracut-053
Jun 25 18:46:37.978234 dracut-cmdline[275]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=e6069a8408a0ca7e7bc40a0bde7fe3ef89df2f98c4bdd2e7e7f9f8f3f8ad207f
Jun 25 18:46:37.994313 systemd-resolved[277]: Positive Trust Anchors:
Jun 25 18:46:37.994331 systemd-resolved[277]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Jun 25 18:46:37.994361 systemd-resolved[277]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa corp home internal intranet lan local private test
Jun 25 18:46:37.998859 systemd-resolved[277]: Defaulting to hostname 'linux'.
Jun 25 18:46:38.000057 systemd[1]: Started systemd-resolved.service - Network Name Resolution.
Jun 25 18:46:38.002456 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups.
Jun 25 18:46:38.044233 kernel: SCSI subsystem initialized
Jun 25 18:46:38.049221 kernel: Loading iSCSI transport class v2.0-870.
Jun 25 18:46:38.056251 kernel: iscsi: registered transport (tcp)
Jun 25 18:46:38.069264 kernel: iscsi: registered transport (qla4xxx)
Jun 25 18:46:38.069307 kernel: QLogic iSCSI HBA Driver
Jun 25 18:46:38.111252 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook.
Jun 25 18:46:38.119375 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook...
Jun 25 18:46:38.136828 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log.
Jun 25 18:46:38.136907 kernel: device-mapper: uevent: version 1.0.3
Jun 25 18:46:38.136934 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com
Jun 25 18:46:38.185236 kernel: raid6: neonx8   gen() 15766 MB/s
Jun 25 18:46:38.202231 kernel: raid6: neonx4   gen() 15647 MB/s
Jun 25 18:46:38.219225 kernel: raid6: neonx2   gen() 13220 MB/s
Jun 25 18:46:38.236232 kernel: raid6: neonx1   gen() 10463 MB/s
Jun 25 18:46:38.253231 kernel: raid6: int64x8  gen()  6974 MB/s
Jun 25 18:46:38.270223 kernel: raid6: int64x4  gen()  7330 MB/s
Jun 25 18:46:38.287222 kernel: raid6: int64x2  gen()  6128 MB/s
Jun 25 18:46:38.304224 kernel: raid6: int64x1  gen()  5052 MB/s
Jun 25 18:46:38.304242 kernel: raid6: using algorithm neonx8 gen() 15766 MB/s
Jun 25 18:46:38.321227 kernel: raid6: .... xor() 11917 MB/s, rmw enabled
Jun 25 18:46:38.321241 kernel: raid6: using neon recovery algorithm
Jun 25 18:46:38.326502 kernel: xor: measuring software checksum speed
Jun 25 18:46:38.326518 kernel:    8regs           : 19864 MB/sec
Jun 25 18:46:38.327370 kernel:    32regs          : 19635 MB/sec
Jun 25 18:46:38.328535 kernel:    arm64_neon      : 27224 MB/sec
Jun 25 18:46:38.328549 kernel: xor: using function: arm64_neon (27224 MB/sec)
Jun 25 18:46:38.378233 kernel: Btrfs loaded, zoned=no, fsverity=no
Jun 25 18:46:38.389274 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook.
Jun 25 18:46:38.400346 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files...
Jun 25 18:46:38.411645 systemd-udevd[459]: Using default interface naming scheme 'v255'.
Jun 25 18:46:38.414732 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files.
Jun 25 18:46:38.417116 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook...
Jun 25 18:46:38.431621 dracut-pre-trigger[467]: rd.md=0: removing MD RAID activation
Jun 25 18:46:38.456831 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook.
Jun 25 18:46:38.466331 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices...
Jun 25 18:46:38.503773 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices.
Jun 25 18:46:38.511351 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook...
Jun 25 18:46:38.524995 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook.
Jun 25 18:46:38.527527 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems.
Jun 25 18:46:38.528412 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes.
Jun 25 18:46:38.530873 systemd[1]: Reached target remote-fs.target - Remote File Systems.
Jun 25 18:46:38.538334 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook...
Jun 25 18:46:38.549802 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook.
Jun 25 18:46:38.556785 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues
Jun 25 18:46:38.570487 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB)
Jun 25 18:46:38.570588 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk.
Jun 25 18:46:38.570599 kernel: GPT:9289727 != 19775487
Jun 25 18:46:38.570608 kernel: GPT:Alternate GPT header not at the end of the disk.
Jun 25 18:46:38.570618 kernel: GPT:9289727 != 19775487
Jun 25 18:46:38.570626 kernel: GPT: Use GNU Parted to correct GPT errors.
Jun 25 18:46:38.570638 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Jun 25 18:46:38.564677 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully.
Jun 25 18:46:38.564781 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jun 25 18:46:38.571023 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Jun 25 18:46:38.572677 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully.
Jun 25 18:46:38.572812 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup.
Jun 25 18:46:38.574768 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup...
Jun 25 18:46:38.583427 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Jun 25 18:46:38.590026 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (502)
Jun 25 18:46:38.595246 kernel: BTRFS: device fsid c80091a6-4bf3-4ad3-8e1c-e6eb918765f9 devid 1 transid 36 /dev/vda3 scanned by (udev-worker) (522)
Jun 25 18:46:38.597657 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM.
Jun 25 18:46:38.600249 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Jun 25 18:46:38.609481 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT.
Jun 25 18:46:38.616387 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM.
Jun 25 18:46:38.620013 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A.
Jun 25 18:46:38.621159 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132.
Jun 25 18:46:38.632371 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary...
Jun 25 18:46:38.633976 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Jun 25 18:46:38.638919 disk-uuid[551]: Primary Header is updated.
Jun 25 18:46:38.638919 disk-uuid[551]: Secondary Entries is updated.
Jun 25 18:46:38.638919 disk-uuid[551]: Secondary Header is updated.
Jun 25 18:46:38.642226 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Jun 25 18:46:38.653013 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jun 25 18:46:39.656246 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Jun 25 18:46:39.656722 disk-uuid[552]: The operation has completed successfully.
Jun 25 18:46:39.676480 systemd[1]: disk-uuid.service: Deactivated successfully.
Jun 25 18:46:39.676601 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary.
Jun 25 18:46:39.699350 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr...
Jun 25 18:46:39.702116 sh[573]: Success
Jun 25 18:46:39.717241 kernel: device-mapper: verity: sha256 using implementation "sha256-ce"
Jun 25 18:46:39.752625 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr.
Jun 25 18:46:39.764501 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr...
Jun 25 18:46:39.768088 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr.
Jun 25 18:46:39.781396 kernel: BTRFS info (device dm-0): first mount of filesystem c80091a6-4bf3-4ad3-8e1c-e6eb918765f9
Jun 25 18:46:39.781428 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm
Jun 25 18:46:39.781447 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead
Jun 25 18:46:39.782710 kernel: BTRFS info (device dm-0): disabling log replay at mount time
Jun 25 18:46:39.782725 kernel: BTRFS info (device dm-0): using free space tree
Jun 25 18:46:39.789162 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr.
Jun 25 18:46:39.790229 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met.
Jun 25 18:46:39.809354 systemd[1]: Starting ignition-setup.service - Ignition (setup)...
Jun 25 18:46:39.810794 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline...
Jun 25 18:46:39.821401 kernel: BTRFS info (device vda6): first mount of filesystem 0ee4f8d8-9b37-4f6c-84aa-681a87076704
Jun 25 18:46:39.821437 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm
Jun 25 18:46:39.821448 kernel: BTRFS info (device vda6): using free space tree
Jun 25 18:46:39.824301 kernel: BTRFS info (device vda6): auto enabling async discard
Jun 25 18:46:39.831946 systemd[1]: mnt-oem.mount: Deactivated successfully.
Jun 25 18:46:39.834232 kernel: BTRFS info (device vda6): last unmount of filesystem 0ee4f8d8-9b37-4f6c-84aa-681a87076704
Jun 25 18:46:39.839161 systemd[1]: Finished ignition-setup.service - Ignition (setup).
Jun 25 18:46:39.845399 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)...
Jun 25 18:46:39.903923 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline.
Jun 25 18:46:39.914455 systemd[1]: Starting systemd-networkd.service - Network Configuration...
Jun 25 18:46:39.942747 systemd-networkd[760]: lo: Link UP
Jun 25 18:46:39.942759 systemd-networkd[760]: lo: Gained carrier
Jun 25 18:46:39.943420 systemd-networkd[760]: Enumeration completed
Jun 25 18:46:39.943875 systemd-networkd[760]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jun 25 18:46:39.943878 systemd-networkd[760]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network.
Jun 25 18:46:39.943940 systemd[1]: Started systemd-networkd.service - Network Configuration.
Jun 25 18:46:39.944726 systemd-networkd[760]: eth0: Link UP
Jun 25 18:46:39.944729 systemd-networkd[760]: eth0: Gained carrier
Jun 25 18:46:39.944736 systemd-networkd[760]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jun 25 18:46:39.945226 systemd[1]: Reached target network.target - Network.
Jun 25 18:46:39.955115 ignition[668]: Ignition 2.19.0
Jun 25 18:46:39.955125 ignition[668]: Stage: fetch-offline
Jun 25 18:46:39.955163 ignition[668]: no configs at "/usr/lib/ignition/base.d"
Jun 25 18:46:39.955171 ignition[668]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jun 25 18:46:39.955262 ignition[668]: parsed url from cmdline: ""
Jun 25 18:46:39.955266 ignition[668]: no config URL provided
Jun 25 18:46:39.955270 ignition[668]: reading system config file "/usr/lib/ignition/user.ign"
Jun 25 18:46:39.955277 ignition[668]: no config at "/usr/lib/ignition/user.ign"
Jun 25 18:46:39.955299 ignition[668]: op(1): [started]  loading QEMU firmware config module
Jun 25 18:46:39.955303 ignition[668]: op(1): executing: "modprobe" "qemu_fw_cfg"
Jun 25 18:46:39.968441 ignition[668]: op(1): [finished] loading QEMU firmware config module
Jun 25 18:46:39.970263 systemd-networkd[760]: eth0: DHCPv4 address 10.0.0.141/16, gateway 10.0.0.1 acquired from 10.0.0.1
Jun 25 18:46:40.006821 ignition[668]: parsing config with SHA512: 6238c74591ee0b8eb1b7c3d273e3ef2b7fc8e0a962c9fdffffd798fe04de0565e6e82355686f646b9461c951b252db907bb4de7b0036c7ad676da9634ad72d7a
Jun 25 18:46:40.010880 unknown[668]: fetched base config from "system"
Jun 25 18:46:40.010890 unknown[668]: fetched user config from "qemu"
Jun 25 18:46:40.011353 ignition[668]: fetch-offline: fetch-offline passed
Jun 25 18:46:40.012612 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline).
Jun 25 18:46:40.011410 ignition[668]: Ignition finished successfully
Jun 25 18:46:40.014185 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json).
Jun 25 18:46:40.023390 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)...
Jun 25 18:46:40.034447 ignition[772]: Ignition 2.19.0
Jun 25 18:46:40.034455 ignition[772]: Stage: kargs
Jun 25 18:46:40.034603 ignition[772]: no configs at "/usr/lib/ignition/base.d"
Jun 25 18:46:40.034613 ignition[772]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jun 25 18:46:40.035514 ignition[772]: kargs: kargs passed
Jun 25 18:46:40.035558 ignition[772]: Ignition finished successfully
Jun 25 18:46:40.038381 systemd[1]: Finished ignition-kargs.service - Ignition (kargs).
Jun 25 18:46:40.048332 systemd[1]: Starting ignition-disks.service - Ignition (disks)...
Jun 25 18:46:40.057721 ignition[781]: Ignition 2.19.0
Jun 25 18:46:40.057739 ignition[781]: Stage: disks
Jun 25 18:46:40.058234 ignition[781]: no configs at "/usr/lib/ignition/base.d"
Jun 25 18:46:40.058246 ignition[781]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jun 25 18:46:40.059180 ignition[781]: disks: disks passed
Jun 25 18:46:40.061415 systemd[1]: Finished ignition-disks.service - Ignition (disks).
Jun 25 18:46:40.059255 ignition[781]: Ignition finished successfully
Jun 25 18:46:40.062879 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device.
Jun 25 18:46:40.064184 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems.
Jun 25 18:46:40.065647 systemd[1]: Reached target local-fs.target - Local File Systems.
Jun 25 18:46:40.067093 systemd[1]: Reached target sysinit.target - System Initialization.
Jun 25 18:46:40.068691 systemd[1]: Reached target basic.target - Basic System.
Jun 25 18:46:40.079394 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT...
Jun 25 18:46:40.090751 systemd-fsck[792]: ROOT: clean, 14/553520 files, 52654/553472 blocks
Jun 25 18:46:40.094520 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT.
Jun 25 18:46:40.096334 systemd[1]: Mounting sysroot.mount - /sysroot...
Jun 25 18:46:40.143981 systemd[1]: Mounted sysroot.mount - /sysroot.
Jun 25 18:46:40.145323 kernel: EXT4-fs (vda9): mounted filesystem 91548e21-ce72-437e-94b9-d3fed380163a r/w with ordered data mode. Quota mode: none.
Jun 25 18:46:40.145110 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System.
Jun 25 18:46:40.153333 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem...
Jun 25 18:46:40.154874 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr...
Jun 25 18:46:40.156200 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met.
Jun 25 18:46:40.156245 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot).
Jun 25 18:46:40.162806 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (800)
Jun 25 18:46:40.162827 kernel: BTRFS info (device vda6): first mount of filesystem 0ee4f8d8-9b37-4f6c-84aa-681a87076704
Jun 25 18:46:40.162837 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm
Jun 25 18:46:40.156266 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup.
Jun 25 18:46:40.165550 kernel: BTRFS info (device vda6): using free space tree
Jun 25 18:46:40.160541 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr.
Jun 25 18:46:40.164828 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup...
Jun 25 18:46:40.169410 kernel: BTRFS info (device vda6): auto enabling async discard
Jun 25 18:46:40.170146 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem.
Jun 25 18:46:40.206590 initrd-setup-root[824]: cut: /sysroot/etc/passwd: No such file or directory
Jun 25 18:46:40.209659 initrd-setup-root[831]: cut: /sysroot/etc/group: No such file or directory
Jun 25 18:46:40.212595 initrd-setup-root[838]: cut: /sysroot/etc/shadow: No such file or directory
Jun 25 18:46:40.215549 initrd-setup-root[845]: cut: /sysroot/etc/gshadow: No such file or directory
Jun 25 18:46:40.282852 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup.
Jun 25 18:46:40.294351 systemd[1]: Starting ignition-mount.service - Ignition (mount)...
Jun 25 18:46:40.295728 systemd[1]: Starting sysroot-boot.service - /sysroot/boot...
Jun 25 18:46:40.301226 kernel: BTRFS info (device vda6): last unmount of filesystem 0ee4f8d8-9b37-4f6c-84aa-681a87076704
Jun 25 18:46:40.314293 systemd[1]: Finished sysroot-boot.service - /sysroot/boot.
Jun 25 18:46:40.317892 ignition[912]: INFO     : Ignition 2.19.0
Jun 25 18:46:40.317892 ignition[912]: INFO     : Stage: mount
Jun 25 18:46:40.317892 ignition[912]: INFO     : no configs at "/usr/lib/ignition/base.d"
Jun 25 18:46:40.317892 ignition[912]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jun 25 18:46:40.317892 ignition[912]: INFO     : mount: mount passed
Jun 25 18:46:40.317892 ignition[912]: INFO     : Ignition finished successfully
Jun 25 18:46:40.319148 systemd[1]: Finished ignition-mount.service - Ignition (mount).
Jun 25 18:46:40.327332 systemd[1]: Starting ignition-files.service - Ignition (files)...
Jun 25 18:46:40.780565 systemd[1]: sysroot-oem.mount: Deactivated successfully.
Jun 25 18:46:40.791384 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem...
Jun 25 18:46:40.797906 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (926)
Jun 25 18:46:40.797936 kernel: BTRFS info (device vda6): first mount of filesystem 0ee4f8d8-9b37-4f6c-84aa-681a87076704
Jun 25 18:46:40.797947 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm
Jun 25 18:46:40.799234 kernel: BTRFS info (device vda6): using free space tree
Jun 25 18:46:40.801241 kernel: BTRFS info (device vda6): auto enabling async discard
Jun 25 18:46:40.802335 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem.
Jun 25 18:46:40.823557 ignition[943]: INFO     : Ignition 2.19.0
Jun 25 18:46:40.823557 ignition[943]: INFO     : Stage: files
Jun 25 18:46:40.824953 ignition[943]: INFO     : no configs at "/usr/lib/ignition/base.d"
Jun 25 18:46:40.824953 ignition[943]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jun 25 18:46:40.827255 ignition[943]: DEBUG    : files: compiled without relabeling support, skipping
Jun 25 18:46:40.827255 ignition[943]: INFO     : files: ensureUsers: op(1): [started]  creating or modifying user "core"
Jun 25 18:46:40.827255 ignition[943]: DEBUG    : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core"
Jun 25 18:46:40.830755 ignition[943]: INFO     : files: ensureUsers: op(1): [finished] creating or modifying user "core"
Jun 25 18:46:40.830755 ignition[943]: INFO     : files: ensureUsers: op(2): [started]  adding ssh keys to user "core"
Jun 25 18:46:40.830755 ignition[943]: INFO     : files: ensureUsers: op(2): [finished] adding ssh keys to user "core"
Jun 25 18:46:40.830755 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(3): [started]  writing file "/sysroot/etc/flatcar-cgroupv1"
Jun 25 18:46:40.830755 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1"
Jun 25 18:46:40.830755 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(4): [started]  writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz"
Jun 25 18:46:40.830755 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1
Jun 25 18:46:40.828168 unknown[943]: wrote ssh authorized keys file for user: core
Jun 25 18:46:41.077661 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(4): GET result: OK
Jun 25 18:46:41.129810 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz"
Jun 25 18:46:41.131713 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(5): [started]  writing file "/sysroot/opt/bin/cilium.tar.gz"
Jun 25 18:46:41.131713 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(5): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1
Jun 25 18:46:41.437639 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(5): GET result: OK
Jun 25 18:46:41.498549 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(6): [started]  writing file "/sysroot/home/core/install.sh"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/install.sh"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(7): [started]  writing file "/sysroot/home/core/nginx.yaml"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nginx.yaml"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(8): [started]  writing file "/sysroot/home/core/nfs-pod.yaml"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pod.yaml"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(9): [started]  writing file "/sysroot/home/core/nfs-pvc.yaml"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(a): [started]  writing file "/sysroot/etc/flatcar/update.conf"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/flatcar/update.conf"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(b): [started]  writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.28.7-arm64.raw"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(b): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.28.7-arm64.raw"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(c): [started]  writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.28.7-arm64.raw"
Jun 25 18:46:41.500148 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(c): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.28.7-arm64.raw: attempt #1
Jun 25 18:46:41.575336 systemd-networkd[760]: eth0: Gained IPv6LL
Jun 25 18:46:41.709789 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(c): GET result: OK
Jun 25 18:46:41.891990 ignition[943]: INFO     : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.28.7-arm64.raw"
Jun 25 18:46:41.891990 ignition[943]: INFO     : files: op(d): [started]  processing unit "containerd.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(d): op(e): [started]  writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(d): op(e): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(d): [finished] processing unit "containerd.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(f): [started]  processing unit "prepare-helm.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(f): op(10): [started]  writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(f): op(10): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(f): [finished] processing unit "prepare-helm.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(11): [started]  processing unit "coreos-metadata.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(11): op(12): [started]  writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(11): op(12): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(11): [finished] processing unit "coreos-metadata.service"
Jun 25 18:46:41.895629 ignition[943]: INFO     : files: op(13): [started]  setting preset to disabled for "coreos-metadata.service"
Jun 25 18:46:41.916700 ignition[943]: INFO     : files: op(13): op(14): [started]  removing enablement symlink(s) for "coreos-metadata.service"
Jun 25 18:46:41.920347 ignition[943]: INFO     : files: op(13): op(14): [finished] removing enablement symlink(s) for "coreos-metadata.service"
Jun 25 18:46:41.923058 ignition[943]: INFO     : files: op(13): [finished] setting preset to disabled for "coreos-metadata.service"
Jun 25 18:46:41.923058 ignition[943]: INFO     : files: op(15): [started]  setting preset to enabled for "prepare-helm.service"
Jun 25 18:46:41.923058 ignition[943]: INFO     : files: op(15): [finished] setting preset to enabled for "prepare-helm.service"
Jun 25 18:46:41.923058 ignition[943]: INFO     : files: createResultFile: createFiles: op(16): [started]  writing file "/sysroot/etc/.ignition-result.json"
Jun 25 18:46:41.923058 ignition[943]: INFO     : files: createResultFile: createFiles: op(16): [finished] writing file "/sysroot/etc/.ignition-result.json"
Jun 25 18:46:41.923058 ignition[943]: INFO     : files: files passed
Jun 25 18:46:41.923058 ignition[943]: INFO     : Ignition finished successfully
Jun 25 18:46:41.923446 systemd[1]: Finished ignition-files.service - Ignition (files).
Jun 25 18:46:41.938469 systemd[1]: Starting ignition-quench.service - Ignition (record completion)...
Jun 25 18:46:41.940265 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion...
Jun 25 18:46:41.942763 systemd[1]: ignition-quench.service: Deactivated successfully.
Jun 25 18:46:41.942872 systemd[1]: Finished ignition-quench.service - Ignition (record completion).
Jun 25 18:46:41.948303 initrd-setup-root-after-ignition[972]: grep: /sysroot/oem/oem-release: No such file or directory
Jun 25 18:46:41.951671 initrd-setup-root-after-ignition[974]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory
Jun 25 18:46:41.951671 initrd-setup-root-after-ignition[974]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory
Jun 25 18:46:41.954767 initrd-setup-root-after-ignition[978]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory
Jun 25 18:46:41.955737 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion.
Jun 25 18:46:41.957343 systemd[1]: Reached target ignition-complete.target - Ignition Complete.
Jun 25 18:46:41.967344 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root...
Jun 25 18:46:41.986673 systemd[1]: initrd-parse-etc.service: Deactivated successfully.
Jun 25 18:46:41.987547 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root.
Jun 25 18:46:41.988902 systemd[1]: Reached target initrd-fs.target - Initrd File Systems.
Jun 25 18:46:41.990604 systemd[1]: Reached target initrd.target - Initrd Default Target.
Jun 25 18:46:41.992201 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met.
Jun 25 18:46:41.992967 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook...
Jun 25 18:46:42.009271 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook.
Jun 25 18:46:42.013729 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons...
Jun 25 18:46:42.025737 systemd[1]: Stopped target network.target - Network.
Jun 25 18:46:42.026637 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups.
Jun 25 18:46:42.028263 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes.
Jun 25 18:46:42.030242 systemd[1]: Stopped target timers.target - Timer Units.
Jun 25 18:46:42.031875 systemd[1]: dracut-pre-pivot.service: Deactivated successfully.
Jun 25 18:46:42.032002 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook.
Jun 25 18:46:42.034182 systemd[1]: Stopped target initrd.target - Initrd Default Target.
Jun 25 18:46:42.036054 systemd[1]: Stopped target basic.target - Basic System.
Jun 25 18:46:42.037756 systemd[1]: Stopped target ignition-complete.target - Ignition Complete.
Jun 25 18:46:42.039326 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup.
Jun 25 18:46:42.041077 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device.
Jun 25 18:46:42.042906 systemd[1]: Stopped target remote-fs.target - Remote File Systems.
Jun 25 18:46:42.044587 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems.
Jun 25 18:46:42.046395 systemd[1]: Stopped target sysinit.target - System Initialization.
Jun 25 18:46:42.048183 systemd[1]: Stopped target local-fs.target - Local File Systems.
Jun 25 18:46:42.049778 systemd[1]: Stopped target swap.target - Swaps.
Jun 25 18:46:42.051426 systemd[1]: dracut-pre-mount.service: Deactivated successfully.
Jun 25 18:46:42.051550 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook.
Jun 25 18:46:42.053806 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes.
Jun 25 18:46:42.055698 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Jun 25 18:46:42.057436 systemd[1]: clevis-luks-askpass.path: Deactivated successfully.
Jun 25 18:46:42.057548 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Jun 25 18:46:42.059349 systemd[1]: dracut-initqueue.service: Deactivated successfully.
Jun 25 18:46:42.059463 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook.
Jun 25 18:46:42.061925 systemd[1]: ignition-fetch-offline.service: Deactivated successfully.
Jun 25 18:46:42.062042 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline).
Jun 25 18:46:42.063798 systemd[1]: Stopped target paths.target - Path Units.
Jun 25 18:46:42.065199 systemd[1]: systemd-ask-password-console.path: Deactivated successfully.
Jun 25 18:46:42.066329 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Jun 25 18:46:42.068224 systemd[1]: Stopped target slices.target - Slice Units.
Jun 25 18:46:42.069589 systemd[1]: Stopped target sockets.target - Socket Units.
Jun 25 18:46:42.071056 systemd[1]: iscsid.socket: Deactivated successfully.
Jun 25 18:46:42.071158 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket.
Jun 25 18:46:42.073123 systemd[1]: iscsiuio.socket: Deactivated successfully.
Jun 25 18:46:42.073225 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket.
Jun 25 18:46:42.074613 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully.
Jun 25 18:46:42.074722 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion.
Jun 25 18:46:42.076271 systemd[1]: ignition-files.service: Deactivated successfully.
Jun 25 18:46:42.076376 systemd[1]: Stopped ignition-files.service - Ignition (files).
Jun 25 18:46:42.087381 systemd[1]: Stopping ignition-mount.service - Ignition (mount)...
Jun 25 18:46:42.088916 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot...
Jun 25 18:46:42.089913 systemd[1]: Stopping systemd-networkd.service - Network Configuration...
Jun 25 18:46:42.091559 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution...
Jun 25 18:46:42.093461 systemd[1]: systemd-udev-trigger.service: Deactivated successfully.
Jun 25 18:46:42.093593 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices.
Jun 25 18:46:42.095630 systemd[1]: dracut-pre-trigger.service: Deactivated successfully.
Jun 25 18:46:42.095745 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook.
Jun 25 18:46:42.103605 ignition[999]: INFO     : Ignition 2.19.0
Jun 25 18:46:42.103605 ignition[999]: INFO     : Stage: umount
Jun 25 18:46:42.103605 ignition[999]: INFO     : no configs at "/usr/lib/ignition/base.d"
Jun 25 18:46:42.103605 ignition[999]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jun 25 18:46:42.097249 systemd-networkd[760]: eth0: DHCPv6 lease lost
Jun 25 18:46:42.110745 ignition[999]: INFO     : umount: umount passed
Jun 25 18:46:42.110745 ignition[999]: INFO     : Ignition finished successfully
Jun 25 18:46:42.100529 systemd[1]: systemd-networkd.service: Deactivated successfully.
Jun 25 18:46:42.100623 systemd[1]: Stopped systemd-networkd.service - Network Configuration.
Jun 25 18:46:42.102641 systemd[1]: systemd-resolved.service: Deactivated successfully.
Jun 25 18:46:42.102768 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution.
Jun 25 18:46:42.107510 systemd[1]: sysroot-boot.mount: Deactivated successfully.
Jun 25 18:46:42.108147 systemd[1]: ignition-mount.service: Deactivated successfully.
Jun 25 18:46:42.108248 systemd[1]: Stopped ignition-mount.service - Ignition (mount).
Jun 25 18:46:42.112463 systemd[1]: initrd-cleanup.service: Deactivated successfully.
Jun 25 18:46:42.112551 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons.
Jun 25 18:46:42.117610 systemd[1]: sysroot-boot.service: Deactivated successfully.
Jun 25 18:46:42.117701 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot.
Jun 25 18:46:42.121848 systemd[1]: systemd-networkd.socket: Deactivated successfully.
Jun 25 18:46:42.121886 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket.
Jun 25 18:46:42.123347 systemd[1]: ignition-disks.service: Deactivated successfully.
Jun 25 18:46:42.123397 systemd[1]: Stopped ignition-disks.service - Ignition (disks).
Jun 25 18:46:42.124719 systemd[1]: ignition-kargs.service: Deactivated successfully.
Jun 25 18:46:42.124762 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs).
Jun 25 18:46:42.126086 systemd[1]: ignition-setup.service: Deactivated successfully.
Jun 25 18:46:42.126136 systemd[1]: Stopped ignition-setup.service - Ignition (setup).
Jun 25 18:46:42.127481 systemd[1]: ignition-setup-pre.service: Deactivated successfully.
Jun 25 18:46:42.127524 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup.
Jun 25 18:46:42.128986 systemd[1]: initrd-setup-root.service: Deactivated successfully.
Jun 25 18:46:42.129031 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup.
Jun 25 18:46:42.136310 systemd[1]: Stopping network-cleanup.service - Network Cleanup...
Jun 25 18:46:42.137652 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully.
Jun 25 18:46:42.137712 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline.
Jun 25 18:46:42.139331 systemd[1]: systemd-sysctl.service: Deactivated successfully.
Jun 25 18:46:42.139375 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables.
Jun 25 18:46:42.140795 systemd[1]: systemd-modules-load.service: Deactivated successfully.
Jun 25 18:46:42.140840 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules.
Jun 25 18:46:42.142644 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully.
Jun 25 18:46:42.142685 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create Volatile Files and Directories.
Jun 25 18:46:42.144301 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files...
Jun 25 18:46:42.154592 systemd[1]: network-cleanup.service: Deactivated successfully.
Jun 25 18:46:42.155678 systemd[1]: Stopped network-cleanup.service - Network Cleanup.
Jun 25 18:46:42.162860 systemd[1]: systemd-udevd.service: Deactivated successfully.
Jun 25 18:46:42.163034 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files.
Jun 25 18:46:42.165261 systemd[1]: systemd-udevd-control.socket: Deactivated successfully.
Jun 25 18:46:42.165302 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket.
Jun 25 18:46:42.166985 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully.
Jun 25 18:46:42.167015 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket.
Jun 25 18:46:42.169532 systemd[1]: dracut-pre-udev.service: Deactivated successfully.
Jun 25 18:46:42.169581 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook.
Jun 25 18:46:42.171880 systemd[1]: dracut-cmdline.service: Deactivated successfully.
Jun 25 18:46:42.171924 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook.
Jun 25 18:46:42.174215 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully.
Jun 25 18:46:42.174263 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jun 25 18:46:42.187363 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database...
Jun 25 18:46:42.188317 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully.
Jun 25 18:46:42.188372 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Jun 25 18:46:42.190380 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully.
Jun 25 18:46:42.190424 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully.
Jun 25 18:46:42.192235 systemd[1]: kmod-static-nodes.service: Deactivated successfully.
Jun 25 18:46:42.192281 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes.
Jun 25 18:46:42.194342 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully.
Jun 25 18:46:42.194393 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup.
Jun 25 18:46:42.196490 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully.
Jun 25 18:46:42.196605 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database.
Jun 25 18:46:42.199756 systemd[1]: Reached target initrd-switch-root.target - Switch Root.
Jun 25 18:46:42.202075 systemd[1]: Starting initrd-switch-root.service - Switch Root...
Jun 25 18:46:42.212027 systemd[1]: Switching root.
Jun 25 18:46:42.245273 systemd-journald[237]: Journal stopped
Jun 25 18:46:42.946633 systemd-journald[237]: Received SIGTERM from PID 1 (systemd).
Jun 25 18:46:42.946690 kernel: SELinux:  policy capability network_peer_controls=1
Jun 25 18:46:42.946707 kernel: SELinux:  policy capability open_perms=1
Jun 25 18:46:42.946719 kernel: SELinux:  policy capability extended_socket_class=1
Jun 25 18:46:42.946729 kernel: SELinux:  policy capability always_check_network=0
Jun 25 18:46:42.946738 kernel: SELinux:  policy capability cgroup_seclabel=1
Jun 25 18:46:42.946749 kernel: SELinux:  policy capability nnp_nosuid_transition=1
Jun 25 18:46:42.946764 kernel: SELinux:  policy capability genfs_seclabel_symlinks=0
Jun 25 18:46:42.946774 kernel: SELinux:  policy capability ioctl_skip_cloexec=0
Jun 25 18:46:42.946784 kernel: audit: type=1403 audit(1719341202.437:2): auid=4294967295 ses=4294967295 lsm=selinux res=1
Jun 25 18:46:42.946794 systemd[1]: Successfully loaded SELinux policy in 31.569ms.
Jun 25 18:46:42.946811 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 9.131ms.
Jun 25 18:46:42.946825 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified)
Jun 25 18:46:42.946836 systemd[1]: Detected virtualization kvm.
Jun 25 18:46:42.946847 systemd[1]: Detected architecture arm64.
Jun 25 18:46:42.946857 systemd[1]: Detected first boot.
Jun 25 18:46:42.946867 systemd[1]: Initializing machine ID from VM UUID.
Jun 25 18:46:42.946878 zram_generator::config[1063]: No configuration found.
Jun 25 18:46:42.946889 systemd[1]: Populated /etc with preset unit settings.
Jun 25 18:46:42.946900 systemd[1]: Queued start job for default target multi-user.target.
Jun 25 18:46:42.946912 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6.
Jun 25 18:46:42.946923 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config.
Jun 25 18:46:42.946935 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run.
Jun 25 18:46:42.946946 systemd[1]: Created slice system-getty.slice - Slice /system/getty.
Jun 25 18:46:42.946956 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe.
Jun 25 18:46:42.946967 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty.
Jun 25 18:46:42.946978 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit.
Jun 25 18:46:42.946988 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck.
Jun 25 18:46:42.947000 systemd[1]: Created slice user.slice - User and Session Slice.
Jun 25 18:46:42.947012 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Jun 25 18:46:42.947023 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Jun 25 18:46:42.947033 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch.
Jun 25 18:46:42.947044 systemd[1]: Set up automount boot.automount - Boot partition Automount Point.
Jun 25 18:46:42.947055 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point.
Jun 25 18:46:42.947065 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM...
Jun 25 18:46:42.947076 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0...
Jun 25 18:46:42.947086 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Jun 25 18:46:42.947097 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes.
Jun 25 18:46:42.947109 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes.
Jun 25 18:46:42.947129 systemd[1]: Reached target remote-fs.target - Remote File Systems.
Jun 25 18:46:42.947140 systemd[1]: Reached target slices.target - Slice Units.
Jun 25 18:46:42.947151 systemd[1]: Reached target swap.target - Swaps.
Jun 25 18:46:42.947161 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes.
Jun 25 18:46:42.947172 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket.
Jun 25 18:46:42.947183 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log).
Jun 25 18:46:42.947193 systemd[1]: Listening on systemd-journald.socket - Journal Socket.
Jun 25 18:46:42.947217 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket.
Jun 25 18:46:42.947229 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket.
Jun 25 18:46:42.947239 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket.
Jun 25 18:46:42.947252 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket.
Jun 25 18:46:42.947262 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System...
Jun 25 18:46:42.947272 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System...
Jun 25 18:46:42.947283 systemd[1]: Mounting media.mount - External Media Directory...
Jun 25 18:46:42.947294 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System...
Jun 25 18:46:42.947305 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System...
Jun 25 18:46:42.947317 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp...
Jun 25 18:46:42.947328 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files...
Jun 25 18:46:42.947339 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jun 25 18:46:42.947350 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes...
Jun 25 18:46:42.947360 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs...
Jun 25 18:46:42.947371 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jun 25 18:46:42.947381 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm...
Jun 25 18:46:42.947392 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jun 25 18:46:42.947403 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse...
Jun 25 18:46:42.947414 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jun 25 18:46:42.947425 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf).
Jun 25 18:46:42.947437 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling.
Jun 25 18:46:42.947448 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.)
Jun 25 18:46:42.947457 kernel: loop: module loaded
Jun 25 18:46:42.947468 systemd[1]: Starting systemd-journald.service - Journal Service...
Jun 25 18:46:42.947479 kernel: fuse: init (API version 7.39)
Jun 25 18:46:42.947490 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules...
Jun 25 18:46:42.947502 kernel: ACPI: bus type drm_connector registered
Jun 25 18:46:42.947512 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line...
Jun 25 18:46:42.947523 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems...
Jun 25 18:46:42.947548 systemd-journald[1138]: Collecting audit messages is disabled.
Jun 25 18:46:42.947569 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices...
Jun 25 18:46:42.947580 systemd-journald[1138]: Journal started
Jun 25 18:46:42.947603 systemd-journald[1138]: Runtime Journal (/run/log/journal/a78f43e1c1ce473b961e557db24d7181) is 5.9M, max 47.3M, 41.4M free.
Jun 25 18:46:42.952234 systemd[1]: Started systemd-journald.service - Journal Service.
Jun 25 18:46:42.953098 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System.
Jun 25 18:46:42.954221 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System.
Jun 25 18:46:42.955308 systemd[1]: Mounted media.mount - External Media Directory.
Jun 25 18:46:42.956486 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System.
Jun 25 18:46:42.957726 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System.
Jun 25 18:46:42.958824 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp.
Jun 25 18:46:42.959948 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes.
Jun 25 18:46:42.961387 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files.
Jun 25 18:46:42.962683 systemd[1]: modprobe@configfs.service: Deactivated successfully.
Jun 25 18:46:42.962840 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs.
Jun 25 18:46:42.964327 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jun 25 18:46:42.964476 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jun 25 18:46:42.965833 systemd[1]: modprobe@drm.service: Deactivated successfully.
Jun 25 18:46:42.965989 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm.
Jun 25 18:46:42.967182 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jun 25 18:46:42.967338 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jun 25 18:46:42.968717 systemd[1]: modprobe@fuse.service: Deactivated successfully.
Jun 25 18:46:42.968868 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse.
Jun 25 18:46:42.970229 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jun 25 18:46:42.970415 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jun 25 18:46:42.971669 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules.
Jun 25 18:46:42.972951 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line.
Jun 25 18:46:42.974504 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems.
Jun 25 18:46:42.985884 systemd[1]: Reached target network-pre.target - Preparation for Network.
Jun 25 18:46:42.993356 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System...
Jun 25 18:46:42.995331 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System...
Jun 25 18:46:42.996321 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/).
Jun 25 18:46:43.000068 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database...
Jun 25 18:46:43.004458 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage...
Jun 25 18:46:43.005489 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore).
Jun 25 18:46:43.006436 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed...
Jun 25 18:46:43.009891 systemd-journald[1138]: Time spent on flushing to /var/log/journal/a78f43e1c1ce473b961e557db24d7181 is 16.036ms for 845 entries.
Jun 25 18:46:43.009891 systemd-journald[1138]: System Journal (/var/log/journal/a78f43e1c1ce473b961e557db24d7181) is 8.0M, max 195.6M, 187.6M free.
Jun 25 18:46:43.029567 systemd-journald[1138]: Received client request to flush runtime journal.
Jun 25 18:46:43.007582 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met.
Jun 25 18:46:43.011360 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Jun 25 18:46:43.016334 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully...
Jun 25 18:46:43.018530 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices.
Jun 25 18:46:43.019929 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System.
Jun 25 18:46:43.021067 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System.
Jun 25 18:46:43.022428 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed.
Jun 25 18:46:43.025000 systemd[1]: Reached target first-boot-complete.target - First Boot Complete.
Jun 25 18:46:43.032491 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization...
Jun 25 18:46:43.033841 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage.
Jun 25 18:46:43.039815 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Jun 25 18:46:43.044390 udevadm[1200]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in.
Jun 25 18:46:43.054233 systemd-tmpfiles[1194]: ACLs are not supported, ignoring.
Jun 25 18:46:43.054253 systemd-tmpfiles[1194]: ACLs are not supported, ignoring.
Jun 25 18:46:43.059612 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully.
Jun 25 18:46:43.071375 systemd[1]: Starting systemd-sysusers.service - Create System Users...
Jun 25 18:46:43.090409 systemd[1]: Finished systemd-sysusers.service - Create System Users.
Jun 25 18:46:43.104391 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev...
Jun 25 18:46:43.115580 systemd-tmpfiles[1224]: ACLs are not supported, ignoring.
Jun 25 18:46:43.115822 systemd-tmpfiles[1224]: ACLs are not supported, ignoring.
Jun 25 18:46:43.119451 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Jun 25 18:46:43.405767 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database.
Jun 25 18:46:43.416402 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files...
Jun 25 18:46:43.436199 systemd-udevd[1230]: Using default interface naming scheme 'v255'.
Jun 25 18:46:43.447928 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files.
Jun 25 18:46:43.460104 systemd[1]: Starting systemd-networkd.service - Network Configuration...
Jun 25 18:46:43.470186 systemd[1]: Found device dev-ttyAMA0.device - /dev/ttyAMA0.
Jun 25 18:46:43.480245 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1231)
Jun 25 18:46:43.488243 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (1246)
Jun 25 18:46:43.488756 systemd[1]: Starting systemd-userdbd.service - User Database Manager...
Jun 25 18:46:43.548940 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM.
Jun 25 18:46:43.560409 systemd[1]: Started systemd-userdbd.service - User Database Manager.
Jun 25 18:46:43.567527 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Jun 25 18:46:43.575560 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization.
Jun 25 18:46:43.579368 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes...
Jun 25 18:46:43.610417 lvm[1266]:   WARNING: Failed to connect to lvmetad. Falling back to device scanning.
Jun 25 18:46:43.623707 systemd-networkd[1237]: lo: Link UP
Jun 25 18:46:43.623714 systemd-networkd[1237]: lo: Gained carrier
Jun 25 18:46:43.624401 systemd-networkd[1237]: Enumeration completed
Jun 25 18:46:43.624698 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Jun 25 18:46:43.624866 systemd-networkd[1237]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jun 25 18:46:43.624874 systemd-networkd[1237]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network.
Jun 25 18:46:43.625482 systemd-networkd[1237]: eth0: Link UP
Jun 25 18:46:43.625495 systemd-networkd[1237]: eth0: Gained carrier
Jun 25 18:46:43.625507 systemd-networkd[1237]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jun 25 18:46:43.626232 systemd[1]: Started systemd-networkd.service - Network Configuration.
Jun 25 18:46:43.635249 systemd-networkd[1237]: eth0: DHCPv4 address 10.0.0.141/16, gateway 10.0.0.1 acquired from 10.0.0.1
Jun 25 18:46:43.635345 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured...
Jun 25 18:46:43.636623 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes.
Jun 25 18:46:43.638490 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes.
Jun 25 18:46:43.640715 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes...
Jun 25 18:46:43.647813 lvm[1281]:   WARNING: Failed to connect to lvmetad. Falling back to device scanning.
Jun 25 18:46:43.682554 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes.
Jun 25 18:46:43.683847 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems.
Jun 25 18:46:43.685024 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw).
Jun 25 18:46:43.685058 systemd[1]: Reached target local-fs.target - Local File Systems.
Jun 25 18:46:43.686000 systemd[1]: Reached target machines.target - Containers.
Jun 25 18:46:43.688062 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink).
Jun 25 18:46:43.708358 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown...
Jun 25 18:46:43.710662 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache...
Jun 25 18:46:43.711567 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jun 25 18:46:43.712454 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM...
Jun 25 18:46:43.714307 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk...
Jun 25 18:46:43.718806 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/...
Jun 25 18:46:43.720522 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown.
Jun 25 18:46:43.726685 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM.
Jun 25 18:46:43.730946 kernel: loop0: detected capacity change from 0 to 59688
Jun 25 18:46:43.731047 kernel: block loop0: the capability attribute has been deprecated.
Jun 25 18:46:43.739870 systemd[1]: etc-machine\x2did.mount: Deactivated successfully.
Jun 25 18:46:43.740751 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk.
Jun 25 18:46:43.747240 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher
Jun 25 18:46:43.772229 kernel: loop1: detected capacity change from 0 to 113712
Jun 25 18:46:43.809230 kernel: loop2: detected capacity change from 0 to 193208
Jun 25 18:46:43.846236 kernel: loop3: detected capacity change from 0 to 59688
Jun 25 18:46:43.852240 kernel: loop4: detected capacity change from 0 to 113712
Jun 25 18:46:43.859243 kernel: loop5: detected capacity change from 0 to 193208
Jun 25 18:46:43.864218 (sd-merge)[1302]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'.
Jun 25 18:46:43.864591 (sd-merge)[1302]: Merged extensions into '/usr'.
Jun 25 18:46:43.870135 systemd[1]: Reloading requested from client PID 1289 ('systemd-sysext') (unit systemd-sysext.service)...
Jun 25 18:46:43.870150 systemd[1]: Reloading...
Jun 25 18:46:43.920282 zram_generator::config[1328]: No configuration found.
Jun 25 18:46:43.949195 ldconfig[1286]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start.
Jun 25 18:46:44.013972 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jun 25 18:46:44.058065 systemd[1]: Reloading finished in 187 ms.
Jun 25 18:46:44.072019 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache.
Jun 25 18:46:44.073225 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/.
Jun 25 18:46:44.088336 systemd[1]: Starting ensure-sysext.service...
Jun 25 18:46:44.089874 systemd[1]: Starting systemd-tmpfiles-setup.service - Create Volatile Files and Directories...
Jun 25 18:46:44.093068 systemd[1]: Reloading requested from client PID 1371 ('systemctl') (unit ensure-sysext.service)...
Jun 25 18:46:44.093082 systemd[1]: Reloading...
Jun 25 18:46:44.105470 systemd-tmpfiles[1377]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring.
Jun 25 18:46:44.105736 systemd-tmpfiles[1377]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring.
Jun 25 18:46:44.106392 systemd-tmpfiles[1377]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring.
Jun 25 18:46:44.106608 systemd-tmpfiles[1377]: ACLs are not supported, ignoring.
Jun 25 18:46:44.106659 systemd-tmpfiles[1377]: ACLs are not supported, ignoring.
Jun 25 18:46:44.108696 systemd-tmpfiles[1377]: Detected autofs mount point /boot during canonicalization of boot.
Jun 25 18:46:44.108711 systemd-tmpfiles[1377]: Skipping /boot
Jun 25 18:46:44.115014 systemd-tmpfiles[1377]: Detected autofs mount point /boot during canonicalization of boot.
Jun 25 18:46:44.115034 systemd-tmpfiles[1377]: Skipping /boot
Jun 25 18:46:44.130361 zram_generator::config[1403]: No configuration found.
Jun 25 18:46:44.220413 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jun 25 18:46:44.264394 systemd[1]: Reloading finished in 171 ms.
Jun 25 18:46:44.278920 systemd[1]: Finished systemd-tmpfiles-setup.service - Create Volatile Files and Directories.
Jun 25 18:46:44.295931 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules...
Jun 25 18:46:44.297960 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs...
Jun 25 18:46:44.300422 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog...
Jun 25 18:46:44.304326 systemd[1]: Starting systemd-resolved.service - Network Name Resolution...
Jun 25 18:46:44.309329 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP...
Jun 25 18:46:44.315001 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jun 25 18:46:44.323373 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jun 25 18:46:44.325683 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jun 25 18:46:44.332056 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jun 25 18:46:44.333201 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jun 25 18:46:44.334904 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog.
Jun 25 18:46:44.337853 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jun 25 18:46:44.338009 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jun 25 18:46:44.339535 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jun 25 18:46:44.339732 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jun 25 18:46:44.341588 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jun 25 18:46:44.341748 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jun 25 18:46:44.349748 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs.
Jun 25 18:46:44.352266 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jun 25 18:46:44.357903 augenrules[1479]: No rules
Jun 25 18:46:44.359398 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jun 25 18:46:44.361248 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jun 25 18:46:44.365424 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jun 25 18:46:44.366365 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jun 25 18:46:44.367531 systemd[1]: Starting systemd-update-done.service - Update is Completed...
Jun 25 18:46:44.370360 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt).
Jun 25 18:46:44.371487 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules.
Jun 25 18:46:44.373076 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP.
Jun 25 18:46:44.374887 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jun 25 18:46:44.375022 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jun 25 18:46:44.376419 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jun 25 18:46:44.376554 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jun 25 18:46:44.378143 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jun 25 18:46:44.378414 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jun 25 18:46:44.380535 systemd[1]: Finished systemd-update-done.service - Update is Completed.
Jun 25 18:46:44.388916 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jun 25 18:46:44.391169 systemd-resolved[1450]: Positive Trust Anchors:
Jun 25 18:46:44.391187 systemd-resolved[1450]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Jun 25 18:46:44.391230 systemd-resolved[1450]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa corp home internal intranet lan local private test
Jun 25 18:46:44.398911 systemd-resolved[1450]: Defaulting to hostname 'linux'.
Jun 25 18:46:44.406604 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jun 25 18:46:44.408495 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm...
Jun 25 18:46:44.410371 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jun 25 18:46:44.412419 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jun 25 18:46:44.414421 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jun 25 18:46:44.414554 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt).
Jun 25 18:46:44.415185 systemd[1]: Started systemd-resolved.service - Network Name Resolution.
Jun 25 18:46:44.416702 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jun 25 18:46:44.416840 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jun 25 18:46:44.418302 systemd[1]: modprobe@drm.service: Deactivated successfully.
Jun 25 18:46:44.418430 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm.
Jun 25 18:46:44.419954 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jun 25 18:46:44.420086 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jun 25 18:46:44.421569 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jun 25 18:46:44.421757 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jun 25 18:46:44.424578 systemd[1]: Finished ensure-sysext.service.
Jun 25 18:46:44.428637 systemd[1]: Reached target network.target - Network.
Jun 25 18:46:44.429624 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups.
Jun 25 18:46:44.430682 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore).
Jun 25 18:46:44.430745 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met.
Jun 25 18:46:44.441353 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization...
Jun 25 18:46:44.482328 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization.
Jun 25 18:46:44.483006 systemd-timesyncd[1517]: Contacted time server 10.0.0.1:123 (10.0.0.1).
Jun 25 18:46:44.483055 systemd-timesyncd[1517]: Initial clock synchronization to Tue 2024-06-25 18:46:44.723225 UTC.
Jun 25 18:46:44.483703 systemd[1]: Reached target sysinit.target - System Initialization.
Jun 25 18:46:44.484744 systemd[1]: Started motdgen.path - Watch for update engine configuration changes.
Jun 25 18:46:44.485852 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data.
Jun 25 18:46:44.486974 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories.
Jun 25 18:46:44.488114 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate).
Jun 25 18:46:44.488152 systemd[1]: Reached target paths.target - Path Units.
Jun 25 18:46:44.488985 systemd[1]: Reached target time-set.target - System Time Set.
Jun 25 18:46:44.490041 systemd[1]: Started logrotate.timer - Daily rotation of log files.
Jun 25 18:46:44.491135 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information..
Jun 25 18:46:44.492239 systemd[1]: Reached target timers.target - Timer Units.
Jun 25 18:46:44.495268 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket.
Jun 25 18:46:44.497470 systemd[1]: Starting docker.socket - Docker Socket for the API...
Jun 25 18:46:44.499359 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket.
Jun 25 18:46:44.505047 systemd[1]: Listening on docker.socket - Docker Socket for the API.
Jun 25 18:46:44.506036 systemd[1]: Reached target sockets.target - Socket Units.
Jun 25 18:46:44.506936 systemd[1]: Reached target basic.target - Basic System.
Jun 25 18:46:44.507908 systemd[1]: System is tainted: cgroupsv1
Jun 25 18:46:44.507955 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met.
Jun 25 18:46:44.507976 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met.
Jun 25 18:46:44.509057 systemd[1]: Starting containerd.service - containerd container runtime...
Jun 25 18:46:44.510981 systemd[1]: Starting dbus.service - D-Bus System Message Bus...
Jun 25 18:46:44.512804 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit...
Jun 25 18:46:44.517368 systemd[1]: Starting extend-filesystems.service - Extend Filesystems...
Jun 25 18:46:44.518259 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment).
Jun 25 18:46:44.519278 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd...
Jun 25 18:46:44.525413 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin...
Jun 25 18:46:44.528315 jq[1523]: false
Jun 25 18:46:44.527236 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline...
Jun 25 18:46:44.534714 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys...
Jun 25 18:46:44.538128 systemd[1]: Starting systemd-logind.service - User Login Management...
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found loop3
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found loop4
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found loop5
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found vda
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found vda1
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found vda2
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found vda3
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found usr
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found vda4
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found vda6
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found vda7
Jun 25 18:46:44.544004 extend-filesystems[1525]: Found vda9
Jun 25 18:46:44.544004 extend-filesystems[1525]: Checking size of /dev/vda9
Jun 25 18:46:44.543822 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0).
Jun 25 18:46:44.571526 extend-filesystems[1525]: Resized partition /dev/vda9
Jun 25 18:46:44.556345 dbus-daemon[1522]: [system] SELinux support is enabled
Jun 25 18:46:44.623586 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (1244)
Jun 25 18:46:44.623658 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks
Jun 25 18:46:44.623674 kernel: EXT4-fs (vda9): resized filesystem to 1864699
Jun 25 18:46:44.548187 systemd[1]: Starting update-engine.service - Update Engine...
Jun 25 18:46:44.623881 extend-filesystems[1551]: resize2fs 1.47.0 (5-Feb-2023)
Jun 25 18:46:44.555461 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition...
Jun 25 18:46:44.628809 extend-filesystems[1551]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required
Jun 25 18:46:44.628809 extend-filesystems[1551]: old_desc_blocks = 1, new_desc_blocks = 1
Jun 25 18:46:44.628809 extend-filesystems[1551]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long.
Jun 25 18:46:44.636272 update_engine[1544]: I0625 18:46:44.622584  1544 main.cc:92] Flatcar Update Engine starting
Jun 25 18:46:44.636272 update_engine[1544]: I0625 18:46:44.628742  1544 update_check_scheduler.cc:74] Next update check in 2m20s
Jun 25 18:46:44.562012 systemd[1]: Started dbus.service - D-Bus System Message Bus.
Jun 25 18:46:44.639588 extend-filesystems[1525]: Resized filesystem in /dev/vda9
Jun 25 18:46:44.640367 jq[1546]: true
Jun 25 18:46:44.581791 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'.
Jun 25 18:46:44.582005 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped.
Jun 25 18:46:44.640766 tar[1560]: linux-arm64/helm
Jun 25 18:46:44.582298 systemd[1]: motdgen.service: Deactivated successfully.
Jun 25 18:46:44.582492 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd.
Jun 25 18:46:44.641146 jq[1562]: true
Jun 25 18:46:44.597203 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully.
Jun 25 18:46:44.599450 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline.
Jun 25 18:46:44.607846 (ntainerd)[1563]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR
Jun 25 18:46:44.623714 systemd-logind[1538]: Watching system buttons on /dev/input/event0 (Power Button)
Jun 25 18:46:44.623889 systemd-logind[1538]: New seat seat0.
Jun 25 18:46:44.625177 systemd[1]: Started systemd-logind.service - User Login Management.
Jun 25 18:46:44.631564 systemd[1]: extend-filesystems.service: Deactivated successfully.
Jun 25 18:46:44.631792 systemd[1]: Finished extend-filesystems.service - Extend Filesystems.
Jun 25 18:46:44.643446 systemd[1]: Started update-engine.service - Update Engine.
Jun 25 18:46:44.645365 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml).
Jun 25 18:46:44.645503 systemd[1]: Reached target system-config.target - Load system-provided cloud configs.
Jun 25 18:46:44.646918 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url).
Jun 25 18:46:44.647367 systemd[1]: Reached target user-config.target - Load user-provided cloud configs.
Jun 25 18:46:44.649026 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details.
Jun 25 18:46:44.650423 systemd[1]: Started locksmithd.service - Cluster reboot manager.
Jun 25 18:46:44.670034 bash[1591]: Updated "/home/core/.ssh/authorized_keys"
Jun 25 18:46:44.673465 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition.
Jun 25 18:46:44.676689 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met.
Jun 25 18:46:44.705204 locksmithd[1592]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot"
Jun 25 18:46:44.775387 systemd-networkd[1237]: eth0: Gained IPv6LL
Jun 25 18:46:44.778391 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured.
Jun 25 18:46:44.779963 systemd[1]: Reached target network-online.target - Network is Online.
Jun 25 18:46:44.786737 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent...
Jun 25 18:46:44.791384 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jun 25 18:46:44.794443 systemd[1]: Starting nvidia.service - NVIDIA Configure Service...
Jun 25 18:46:44.823365 systemd[1]: coreos-metadata.service: Deactivated successfully.
Jun 25 18:46:44.823585 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent.
Jun 25 18:46:44.825450 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met.
Jun 25 18:46:44.832047 systemd[1]: Finished nvidia.service - NVIDIA Configure Service.
Jun 25 18:46:44.840927 containerd[1563]: time="2024-06-25T18:46:44.840862240Z" level=info msg="starting containerd" revision=cd7148ac666309abf41fd4a49a8a5895b905e7f3 version=v1.7.18
Jun 25 18:46:44.866399 containerd[1563]: time="2024-06-25T18:46:44.866361840Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1
Jun 25 18:46:44.866610 containerd[1563]: time="2024-06-25T18:46:44.866595040Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1
Jun 25 18:46:44.868077 containerd[1563]: time="2024-06-25T18:46:44.868033480Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.35-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1
Jun 25 18:46:44.868077 containerd[1563]: time="2024-06-25T18:46:44.868074280Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1
Jun 25 18:46:44.869297 containerd[1563]: time="2024-06-25T18:46:44.869153160Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jun 25 18:46:44.869297 containerd[1563]: time="2024-06-25T18:46:44.869293640Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1
Jun 25 18:46:44.870563 containerd[1563]: time="2024-06-25T18:46:44.870527000Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1
Jun 25 18:46:44.871298 containerd[1563]: time="2024-06-25T18:46:44.871266440Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1
Jun 25 18:46:44.871298 containerd[1563]: time="2024-06-25T18:46:44.871295160Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1
Jun 25 18:46:44.871437 containerd[1563]: time="2024-06-25T18:46:44.871411760Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1
Jun 25 18:46:44.871641 containerd[1563]: time="2024-06-25T18:46:44.871613520Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1
Jun 25 18:46:44.871641 containerd[1563]: time="2024-06-25T18:46:44.871638440Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured"
Jun 25 18:46:44.871685 containerd[1563]: time="2024-06-25T18:46:44.871649520Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1
Jun 25 18:46:44.871806 containerd[1563]: time="2024-06-25T18:46:44.871778360Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jun 25 18:46:44.871806 containerd[1563]: time="2024-06-25T18:46:44.871798200Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1
Jun 25 18:46:44.871863 containerd[1563]: time="2024-06-25T18:46:44.871849000Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured"
Jun 25 18:46:44.871884 containerd[1563]: time="2024-06-25T18:46:44.871865560Z" level=info msg="metadata content store policy set" policy=shared
Jun 25 18:46:44.876225 containerd[1563]: time="2024-06-25T18:46:44.876117320Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1
Jun 25 18:46:44.876225 containerd[1563]: time="2024-06-25T18:46:44.876148000Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1
Jun 25 18:46:44.876225 containerd[1563]: time="2024-06-25T18:46:44.876160560Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1
Jun 25 18:46:44.876225 containerd[1563]: time="2024-06-25T18:46:44.876190120Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1
Jun 25 18:46:44.876380 containerd[1563]: time="2024-06-25T18:46:44.876203520Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1
Jun 25 18:46:44.876380 containerd[1563]: time="2024-06-25T18:46:44.876340880Z" level=info msg="NRI interface is disabled by configuration."
Jun 25 18:46:44.876380 containerd[1563]: time="2024-06-25T18:46:44.876356760Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876485040Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876509120Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876522640Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876536240Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876549760Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876565360Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876597080Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876610120Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876623680Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876636560Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876647880Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876658960Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1
Jun 25 18:46:44.877183 containerd[1563]: time="2024-06-25T18:46:44.876750160Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1
Jun 25 18:46:44.877576 containerd[1563]: time="2024-06-25T18:46:44.877544440Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1
Jun 25 18:46:44.877610 containerd[1563]: time="2024-06-25T18:46:44.877586440Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.877610 containerd[1563]: time="2024-06-25T18:46:44.877600600Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1
Jun 25 18:46:44.877649 containerd[1563]: time="2024-06-25T18:46:44.877624280Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1
Jun 25 18:46:44.877823 containerd[1563]: time="2024-06-25T18:46:44.877801840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.877851 containerd[1563]: time="2024-06-25T18:46:44.877826640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.877851 containerd[1563]: time="2024-06-25T18:46:44.877839960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.877889 containerd[1563]: time="2024-06-25T18:46:44.877852080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.877889 containerd[1563]: time="2024-06-25T18:46:44.877864960Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.877889 containerd[1563]: time="2024-06-25T18:46:44.877877360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.877939 containerd[1563]: time="2024-06-25T18:46:44.877889040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.877939 containerd[1563]: time="2024-06-25T18:46:44.877900080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.877939 containerd[1563]: time="2024-06-25T18:46:44.877913360Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1
Jun 25 18:46:44.878620 containerd[1563]: time="2024-06-25T18:46:44.878051200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.878620 containerd[1563]: time="2024-06-25T18:46:44.878074040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.878620 containerd[1563]: time="2024-06-25T18:46:44.878089600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.878620 containerd[1563]: time="2024-06-25T18:46:44.878109040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.878620 containerd[1563]: time="2024-06-25T18:46:44.878123840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.878620 containerd[1563]: time="2024-06-25T18:46:44.878138840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.878620 containerd[1563]: time="2024-06-25T18:46:44.878150120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.878620 containerd[1563]: time="2024-06-25T18:46:44.878160200Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1
Jun 25 18:46:44.878842 containerd[1563]: time="2024-06-25T18:46:44.878529400Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}"
Jun 25 18:46:44.878842 containerd[1563]: time="2024-06-25T18:46:44.878593080Z" level=info msg="Connect containerd service"
Jun 25 18:46:44.878842 containerd[1563]: time="2024-06-25T18:46:44.878622560Z" level=info msg="using legacy CRI server"
Jun 25 18:46:44.878842 containerd[1563]: time="2024-06-25T18:46:44.878630240Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this"
Jun 25 18:46:44.878842 containerd[1563]: time="2024-06-25T18:46:44.878771160Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\""
Jun 25 18:46:44.879591 containerd[1563]: time="2024-06-25T18:46:44.879562200Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config"
Jun 25 18:46:44.879642 containerd[1563]: time="2024-06-25T18:46:44.879626000Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1
Jun 25 18:46:44.879667 containerd[1563]: time="2024-06-25T18:46:44.879646960Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1
Jun 25 18:46:44.879770 containerd[1563]: time="2024-06-25T18:46:44.879658520Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1
Jun 25 18:46:44.879806 containerd[1563]: time="2024-06-25T18:46:44.879776840Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1
Jun 25 18:46:44.879907 containerd[1563]: time="2024-06-25T18:46:44.879747880Z" level=info msg="Start subscribing containerd event"
Jun 25 18:46:44.879907 containerd[1563]: time="2024-06-25T18:46:44.879863800Z" level=info msg="Start recovering state"
Jun 25 18:46:44.879952 containerd[1563]: time="2024-06-25T18:46:44.879919400Z" level=info msg="Start event monitor"
Jun 25 18:46:44.879952 containerd[1563]: time="2024-06-25T18:46:44.879930200Z" level=info msg="Start snapshots syncer"
Jun 25 18:46:44.879952 containerd[1563]: time="2024-06-25T18:46:44.879938240Z" level=info msg="Start cni network conf syncer for default"
Jun 25 18:46:44.879952 containerd[1563]: time="2024-06-25T18:46:44.879944960Z" level=info msg="Start streaming server"
Jun 25 18:46:44.879952 containerd[1563]: time="2024-06-25T18:46:44.880586720Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc
Jun 25 18:46:44.879952 containerd[1563]: time="2024-06-25T18:46:44.880632600Z" level=info msg=serving... address=/run/containerd/containerd.sock
Jun 25 18:46:44.884508 containerd[1563]: time="2024-06-25T18:46:44.883260640Z" level=info msg="containerd successfully booted in 0.043089s"
Jun 25 18:46:44.883367 systemd[1]: Started containerd.service - containerd container runtime.
Jun 25 18:46:45.012875 tar[1560]: linux-arm64/LICENSE
Jun 25 18:46:45.012980 tar[1560]: linux-arm64/README.md
Jun 25 18:46:45.025725 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin.
Jun 25 18:46:45.072998 sshd_keygen[1549]: ssh-keygen: generating new host keys: RSA ECDSA ED25519
Jun 25 18:46:45.092546 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys.
Jun 25 18:46:45.100532 systemd[1]: Starting issuegen.service - Generate /run/issue...
Jun 25 18:46:45.107478 systemd[1]: issuegen.service: Deactivated successfully.
Jun 25 18:46:45.107708 systemd[1]: Finished issuegen.service - Generate /run/issue.
Jun 25 18:46:45.118498 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions...
Jun 25 18:46:45.127877 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions.
Jun 25 18:46:45.130655 systemd[1]: Started getty@tty1.service - Getty on tty1.
Jun 25 18:46:45.132703 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0.
Jun 25 18:46:45.134131 systemd[1]: Reached target getty.target - Login Prompts.
Jun 25 18:46:45.301956 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jun 25 18:46:45.303519 systemd[1]: Reached target multi-user.target - Multi-User System.
Jun 25 18:46:45.304635 systemd[1]: Startup finished in 5.274s (kernel) + 2.901s (userspace) = 8.176s.
Jun 25 18:46:45.305797 (kubelet)[1664]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Jun 25 18:46:45.781194 kubelet[1664]: E0625 18:46:45.779136    1664 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Jun 25 18:46:45.784401 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Jun 25 18:46:45.784589 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Jun 25 18:46:50.941969 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd.
Jun 25 18:46:50.951456 systemd[1]: Started sshd@0-10.0.0.141:22-10.0.0.1:43178.service - OpenSSH per-connection server daemon (10.0.0.1:43178).
Jun 25 18:46:51.005300 sshd[1678]: Accepted publickey for core from 10.0.0.1 port 43178 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:46:51.006911 sshd[1678]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:46:51.026974 systemd[1]: Created slice user-500.slice - User Slice of UID 500.
Jun 25 18:46:51.037494 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500...
Jun 25 18:46:51.039335 systemd-logind[1538]: New session 1 of user core.
Jun 25 18:46:51.046868 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500.
Jun 25 18:46:51.048992 systemd[1]: Starting user@500.service - User Manager for UID 500...
Jun 25 18:46:51.054977 (systemd)[1684]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:46:51.127934 systemd[1684]: Queued start job for default target default.target.
Jun 25 18:46:51.128307 systemd[1684]: Created slice app.slice - User Application Slice.
Jun 25 18:46:51.128347 systemd[1684]: Reached target paths.target - Paths.
Jun 25 18:46:51.128358 systemd[1684]: Reached target timers.target - Timers.
Jun 25 18:46:51.138313 systemd[1684]: Starting dbus.socket - D-Bus User Message Bus Socket...
Jun 25 18:46:51.143841 systemd[1684]: Listening on dbus.socket - D-Bus User Message Bus Socket.
Jun 25 18:46:51.143903 systemd[1684]: Reached target sockets.target - Sockets.
Jun 25 18:46:51.143914 systemd[1684]: Reached target basic.target - Basic System.
Jun 25 18:46:51.143948 systemd[1684]: Reached target default.target - Main User Target.
Jun 25 18:46:51.143972 systemd[1684]: Startup finished in 84ms.
Jun 25 18:46:51.144320 systemd[1]: Started user@500.service - User Manager for UID 500.
Jun 25 18:46:51.145615 systemd[1]: Started session-1.scope - Session 1 of User core.
Jun 25 18:46:51.201455 systemd[1]: Started sshd@1-10.0.0.141:22-10.0.0.1:43194.service - OpenSSH per-connection server daemon (10.0.0.1:43194).
Jun 25 18:46:51.234725 sshd[1696]: Accepted publickey for core from 10.0.0.1 port 43194 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:46:51.235898 sshd[1696]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:46:51.240507 systemd-logind[1538]: New session 2 of user core.
Jun 25 18:46:51.248537 systemd[1]: Started session-2.scope - Session 2 of User core.
Jun 25 18:46:51.299850 sshd[1696]: pam_unix(sshd:session): session closed for user core
Jun 25 18:46:51.311502 systemd[1]: Started sshd@2-10.0.0.141:22-10.0.0.1:43202.service - OpenSSH per-connection server daemon (10.0.0.1:43202).
Jun 25 18:46:51.311856 systemd[1]: sshd@1-10.0.0.141:22-10.0.0.1:43194.service: Deactivated successfully.
Jun 25 18:46:51.314145 systemd[1]: session-2.scope: Deactivated successfully.
Jun 25 18:46:51.314785 systemd-logind[1538]: Session 2 logged out. Waiting for processes to exit.
Jun 25 18:46:51.315677 systemd-logind[1538]: Removed session 2.
Jun 25 18:46:51.342113 sshd[1701]: Accepted publickey for core from 10.0.0.1 port 43202 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:46:51.343201 sshd[1701]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:46:51.346935 systemd-logind[1538]: New session 3 of user core.
Jun 25 18:46:51.362463 systemd[1]: Started session-3.scope - Session 3 of User core.
Jun 25 18:46:51.411392 sshd[1701]: pam_unix(sshd:session): session closed for user core
Jun 25 18:46:51.428432 systemd[1]: Started sshd@3-10.0.0.141:22-10.0.0.1:43206.service - OpenSSH per-connection server daemon (10.0.0.1:43206).
Jun 25 18:46:51.429103 systemd[1]: sshd@2-10.0.0.141:22-10.0.0.1:43202.service: Deactivated successfully.
Jun 25 18:46:51.431539 systemd-logind[1538]: Session 3 logged out. Waiting for processes to exit.
Jun 25 18:46:51.431666 systemd[1]: session-3.scope: Deactivated successfully.
Jun 25 18:46:51.432833 systemd-logind[1538]: Removed session 3.
Jun 25 18:46:51.458832 sshd[1709]: Accepted publickey for core from 10.0.0.1 port 43206 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:46:51.460396 sshd[1709]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:46:51.464164 systemd-logind[1538]: New session 4 of user core.
Jun 25 18:46:51.470451 systemd[1]: Started session-4.scope - Session 4 of User core.
Jun 25 18:46:51.522201 sshd[1709]: pam_unix(sshd:session): session closed for user core
Jun 25 18:46:51.533545 systemd[1]: Started sshd@4-10.0.0.141:22-10.0.0.1:43208.service - OpenSSH per-connection server daemon (10.0.0.1:43208).
Jun 25 18:46:51.533957 systemd[1]: sshd@3-10.0.0.141:22-10.0.0.1:43206.service: Deactivated successfully.
Jun 25 18:46:51.535614 systemd-logind[1538]: Session 4 logged out. Waiting for processes to exit.
Jun 25 18:46:51.536311 systemd[1]: session-4.scope: Deactivated successfully.
Jun 25 18:46:51.537514 systemd-logind[1538]: Removed session 4.
Jun 25 18:46:51.564301 sshd[1717]: Accepted publickey for core from 10.0.0.1 port 43208 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:46:51.565520 sshd[1717]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:46:51.569326 systemd-logind[1538]: New session 5 of user core.
Jun 25 18:46:51.580568 systemd[1]: Started session-5.scope - Session 5 of User core.
Jun 25 18:46:51.641004 sudo[1724]:     core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1
Jun 25 18:46:51.641254 sudo[1724]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500)
Jun 25 18:46:51.655944 sudo[1724]: pam_unix(sudo:session): session closed for user root
Jun 25 18:46:51.657606 sshd[1717]: pam_unix(sshd:session): session closed for user core
Jun 25 18:46:51.675506 systemd[1]: Started sshd@5-10.0.0.141:22-10.0.0.1:43218.service - OpenSSH per-connection server daemon (10.0.0.1:43218).
Jun 25 18:46:51.675852 systemd[1]: sshd@4-10.0.0.141:22-10.0.0.1:43208.service: Deactivated successfully.
Jun 25 18:46:51.678110 systemd[1]: session-5.scope: Deactivated successfully.
Jun 25 18:46:51.678278 systemd-logind[1538]: Session 5 logged out. Waiting for processes to exit.
Jun 25 18:46:51.679565 systemd-logind[1538]: Removed session 5.
Jun 25 18:46:51.706209 sshd[1726]: Accepted publickey for core from 10.0.0.1 port 43218 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:46:51.707690 sshd[1726]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:46:51.712173 systemd-logind[1538]: New session 6 of user core.
Jun 25 18:46:51.722493 systemd[1]: Started session-6.scope - Session 6 of User core.
Jun 25 18:46:51.774307 sudo[1734]:     core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules
Jun 25 18:46:51.774554 sudo[1734]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500)
Jun 25 18:46:51.778146 sudo[1734]: pam_unix(sudo:session): session closed for user root
Jun 25 18:46:51.782861 sudo[1733]:     core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules
Jun 25 18:46:51.783098 sudo[1733]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500)
Jun 25 18:46:51.802444 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules...
Jun 25 18:46:51.804113 auditctl[1737]: No rules
Jun 25 18:46:51.804985 systemd[1]: audit-rules.service: Deactivated successfully.
Jun 25 18:46:51.805246 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules.
Jun 25 18:46:51.806949 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules...
Jun 25 18:46:51.829457 augenrules[1756]: No rules
Jun 25 18:46:51.830118 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules.
Jun 25 18:46:51.831365 sudo[1733]: pam_unix(sudo:session): session closed for user root
Jun 25 18:46:51.832717 sshd[1726]: pam_unix(sshd:session): session closed for user core
Jun 25 18:46:51.848623 systemd[1]: Started sshd@6-10.0.0.141:22-10.0.0.1:43230.service - OpenSSH per-connection server daemon (10.0.0.1:43230).
Jun 25 18:46:51.849026 systemd[1]: sshd@5-10.0.0.141:22-10.0.0.1:43218.service: Deactivated successfully.
Jun 25 18:46:51.850758 systemd-logind[1538]: Session 6 logged out. Waiting for processes to exit.
Jun 25 18:46:51.851390 systemd[1]: session-6.scope: Deactivated successfully.
Jun 25 18:46:51.852649 systemd-logind[1538]: Removed session 6.
Jun 25 18:46:51.879096 sshd[1762]: Accepted publickey for core from 10.0.0.1 port 43230 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:46:51.880177 sshd[1762]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:46:51.883929 systemd-logind[1538]: New session 7 of user core.
Jun 25 18:46:51.895449 systemd[1]: Started session-7.scope - Session 7 of User core.
Jun 25 18:46:51.945376 sudo[1769]:     core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh
Jun 25 18:46:51.945612 sudo[1769]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500)
Jun 25 18:46:52.052611 (dockerd)[1779]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU
Jun 25 18:46:52.052984 systemd[1]: Starting docker.service - Docker Application Container Engine...
Jun 25 18:46:52.289606 dockerd[1779]: time="2024-06-25T18:46:52.289544753Z" level=info msg="Starting up"
Jun 25 18:46:52.467350 dockerd[1779]: time="2024-06-25T18:46:52.467195123Z" level=info msg="Loading containers: start."
Jun 25 18:46:52.540299 kernel: Initializing XFRM netlink socket
Jun 25 18:46:52.603383 systemd-networkd[1237]: docker0: Link UP
Jun 25 18:46:52.611551 dockerd[1779]: time="2024-06-25T18:46:52.611503237Z" level=info msg="Loading containers: done."
Jun 25 18:46:52.669174 dockerd[1779]: time="2024-06-25T18:46:52.669126760Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2
Jun 25 18:46:52.669363 dockerd[1779]: time="2024-06-25T18:46:52.669344903Z" level=info msg="Docker daemon" commit=fca702de7f71362c8d103073c7e4a1d0a467fadd graphdriver=overlay2 version=24.0.9
Jun 25 18:46:52.669484 dockerd[1779]: time="2024-06-25T18:46:52.669469839Z" level=info msg="Daemon has completed initialization"
Jun 25 18:46:52.694536 dockerd[1779]: time="2024-06-25T18:46:52.694470411Z" level=info msg="API listen on /run/docker.sock"
Jun 25 18:46:52.694643 systemd[1]: Started docker.service - Docker Application Container Engine.
Jun 25 18:46:53.267407 containerd[1563]: time="2024-06-25T18:46:53.267355817Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.28.11\""
Jun 25 18:46:53.902857 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3589837370.mount: Deactivated successfully.
Jun 25 18:46:55.619376 containerd[1563]: time="2024-06-25T18:46:55.619302916Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.28.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:55.620647 containerd[1563]: time="2024-06-25T18:46:55.620615245Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.28.11: active requests=0, bytes read=31671540"
Jun 25 18:46:55.621732 containerd[1563]: time="2024-06-25T18:46:55.621671163Z" level=info msg="ImageCreate event name:\"sha256:d2b5500cdb8d455434ebcaa569918eb0c5e68e82d75d4c85c509519786f24a8d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:55.624472 containerd[1563]: time="2024-06-25T18:46:55.624441321Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:aec9d1701c304eee8607d728a39baaa511d65bef6dd9861010618f63fbadeb10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:55.625694 containerd[1563]: time="2024-06-25T18:46:55.625663866Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.28.11\" with image id \"sha256:d2b5500cdb8d455434ebcaa569918eb0c5e68e82d75d4c85c509519786f24a8d\", repo tag \"registry.k8s.io/kube-apiserver:v1.28.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:aec9d1701c304eee8607d728a39baaa511d65bef6dd9861010618f63fbadeb10\", size \"31668338\" in 2.358257278s"
Jun 25 18:46:55.625746 containerd[1563]: time="2024-06-25T18:46:55.625696925Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.28.11\" returns image reference \"sha256:d2b5500cdb8d455434ebcaa569918eb0c5e68e82d75d4c85c509519786f24a8d\""
Jun 25 18:46:55.667536 containerd[1563]: time="2024-06-25T18:46:55.667275016Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.28.11\""
Jun 25 18:46:56.034925 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1.
Jun 25 18:46:56.048402 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jun 25 18:46:56.147960 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jun 25 18:46:56.150786 (kubelet)[1992]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Jun 25 18:46:56.199999 kubelet[1992]: E0625 18:46:56.199898    1992 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Jun 25 18:46:56.204118 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Jun 25 18:46:56.204312 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Jun 25 18:46:57.246558 containerd[1563]: time="2024-06-25T18:46:57.246501030Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.28.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:57.247728 containerd[1563]: time="2024-06-25T18:46:57.247692958Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.28.11: active requests=0, bytes read=28893120"
Jun 25 18:46:57.249272 containerd[1563]: time="2024-06-25T18:46:57.248874261Z" level=info msg="ImageCreate event name:\"sha256:24cd2c3bd254238005fcc2fcc15e9e56347b218c10b8399a28d1bf813800266a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:57.251389 containerd[1563]: time="2024-06-25T18:46:57.251341297Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:6014c3572ec683841bbb16f87b94da28ee0254b95e2dba2d1850d62bd0111f09\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:57.252638 containerd[1563]: time="2024-06-25T18:46:57.252593869Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.28.11\" with image id \"sha256:24cd2c3bd254238005fcc2fcc15e9e56347b218c10b8399a28d1bf813800266a\", repo tag \"registry.k8s.io/kube-controller-manager:v1.28.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:6014c3572ec683841bbb16f87b94da28ee0254b95e2dba2d1850d62bd0111f09\", size \"30445463\" in 1.585275128s"
Jun 25 18:46:57.252638 containerd[1563]: time="2024-06-25T18:46:57.252631938Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.28.11\" returns image reference \"sha256:24cd2c3bd254238005fcc2fcc15e9e56347b218c10b8399a28d1bf813800266a\""
Jun 25 18:46:57.273368 containerd[1563]: time="2024-06-25T18:46:57.273328276Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.28.11\""
Jun 25 18:46:58.299219 containerd[1563]: time="2024-06-25T18:46:58.299162486Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.28.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:58.300604 containerd[1563]: time="2024-06-25T18:46:58.300562016Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.28.11: active requests=0, bytes read=15358440"
Jun 25 18:46:58.301919 containerd[1563]: time="2024-06-25T18:46:58.301854622Z" level=info msg="ImageCreate event name:\"sha256:fdf13db9a96001adee7d1c69fd6849d6cd45fc3c138c95c8240d353eb79acf50\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:58.304456 containerd[1563]: time="2024-06-25T18:46:58.304418885Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:46cf7475c8daffb743c856a1aea0ddea35e5acd2418be18b1e22cf98d9c9b445\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:58.305652 containerd[1563]: time="2024-06-25T18:46:58.305619126Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.28.11\" with image id \"sha256:fdf13db9a96001adee7d1c69fd6849d6cd45fc3c138c95c8240d353eb79acf50\", repo tag \"registry.k8s.io/kube-scheduler:v1.28.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:46cf7475c8daffb743c856a1aea0ddea35e5acd2418be18b1e22cf98d9c9b445\", size \"16910801\" in 1.032251459s"
Jun 25 18:46:58.305689 containerd[1563]: time="2024-06-25T18:46:58.305650973Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.28.11\" returns image reference \"sha256:fdf13db9a96001adee7d1c69fd6849d6cd45fc3c138c95c8240d353eb79acf50\""
Jun 25 18:46:58.325249 containerd[1563]: time="2024-06-25T18:46:58.325194284Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.11\""
Jun 25 18:46:59.388145 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3370236975.mount: Deactivated successfully.
Jun 25 18:46:59.659850 containerd[1563]: time="2024-06-25T18:46:59.659732393Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.28.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:59.660430 containerd[1563]: time="2024-06-25T18:46:59.660401438Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.28.11: active requests=0, bytes read=24772463"
Jun 25 18:46:59.661802 containerd[1563]: time="2024-06-25T18:46:59.661761870Z" level=info msg="ImageCreate event name:\"sha256:e195d3cf134bc9d64104f5e82e95fce811d55b1cdc9cb26fb8f52c8d107d1661\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:59.663659 containerd[1563]: time="2024-06-25T18:46:59.663610916Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ae4b671d4cfc23dd75030bb4490207cd939b3b11a799bcb4119698cd712eb5b4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:46:59.664357 containerd[1563]: time="2024-06-25T18:46:59.664163986Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.28.11\" with image id \"sha256:e195d3cf134bc9d64104f5e82e95fce811d55b1cdc9cb26fb8f52c8d107d1661\", repo tag \"registry.k8s.io/kube-proxy:v1.28.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:ae4b671d4cfc23dd75030bb4490207cd939b3b11a799bcb4119698cd712eb5b4\", size \"24771480\" in 1.338913934s"
Jun 25 18:46:59.664357 containerd[1563]: time="2024-06-25T18:46:59.664202282Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.11\" returns image reference \"sha256:e195d3cf134bc9d64104f5e82e95fce811d55b1cdc9cb26fb8f52c8d107d1661\""
Jun 25 18:46:59.682426 containerd[1563]: time="2024-06-25T18:46:59.682392430Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\""
Jun 25 18:47:00.092040 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4124151985.mount: Deactivated successfully.
Jun 25 18:47:00.096356 containerd[1563]: time="2024-06-25T18:47:00.096291015Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:00.097454 containerd[1563]: time="2024-06-25T18:47:00.097418325Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=268823"
Jun 25 18:47:00.098178 containerd[1563]: time="2024-06-25T18:47:00.098149156Z" level=info msg="ImageCreate event name:\"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:00.100458 containerd[1563]: time="2024-06-25T18:47:00.100421046Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:00.101395 containerd[1563]: time="2024-06-25T18:47:00.101299994Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"268051\" in 418.866502ms"
Jun 25 18:47:00.101395 containerd[1563]: time="2024-06-25T18:47:00.101338470Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\""
Jun 25 18:47:00.121159 containerd[1563]: time="2024-06-25T18:47:00.120850070Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\""
Jun 25 18:47:00.638992 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2051362064.mount: Deactivated successfully.
Jun 25 18:47:02.288298 containerd[1563]: time="2024-06-25T18:47:02.287642162Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.10-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:02.288988 containerd[1563]: time="2024-06-25T18:47:02.288953097Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.10-0: active requests=0, bytes read=65200788"
Jun 25 18:47:02.290041 containerd[1563]: time="2024-06-25T18:47:02.290014661Z" level=info msg="ImageCreate event name:\"sha256:79f8d13ae8b8839cadfb2f83416935f5184206d386028e2d1263577f0ab3620b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:02.293002 containerd[1563]: time="2024-06-25T18:47:02.292970364Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:02.295284 containerd[1563]: time="2024-06-25T18:47:02.295247729Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.10-0\" with image id \"sha256:79f8d13ae8b8839cadfb2f83416935f5184206d386028e2d1263577f0ab3620b\", repo tag \"registry.k8s.io/etcd:3.5.10-0\", repo digest \"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\", size \"65198393\" in 2.174358308s"
Jun 25 18:47:02.295284 containerd[1563]: time="2024-06-25T18:47:02.295280270Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\" returns image reference \"sha256:79f8d13ae8b8839cadfb2f83416935f5184206d386028e2d1263577f0ab3620b\""
Jun 25 18:47:02.313595 containerd[1563]: time="2024-06-25T18:47:02.313562184Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\""
Jun 25 18:47:02.827796 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount183970541.mount: Deactivated successfully.
Jun 25 18:47:03.140826 containerd[1563]: time="2024-06-25T18:47:03.140717071Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:03.141410 containerd[1563]: time="2024-06-25T18:47:03.141376898Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.10.1: active requests=0, bytes read=14558464"
Jun 25 18:47:03.142258 containerd[1563]: time="2024-06-25T18:47:03.142200127Z" level=info msg="ImageCreate event name:\"sha256:97e04611ad43405a2e5863ae17c6f1bc9181bdefdaa78627c432ef754a4eb108\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:03.144346 containerd[1563]: time="2024-06-25T18:47:03.144317541Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:03.145298 containerd[1563]: time="2024-06-25T18:47:03.145265828Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.10.1\" with image id \"sha256:97e04611ad43405a2e5863ae17c6f1bc9181bdefdaa78627c432ef754a4eb108\", repo tag \"registry.k8s.io/coredns/coredns:v1.10.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e\", size \"14557471\" in 831.665169ms"
Jun 25 18:47:03.145360 containerd[1563]: time="2024-06-25T18:47:03.145299720Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\" returns image reference \"sha256:97e04611ad43405a2e5863ae17c6f1bc9181bdefdaa78627c432ef754a4eb108\""
Jun 25 18:47:06.454758 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2.
Jun 25 18:47:06.464444 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jun 25 18:47:06.552705 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jun 25 18:47:06.556581 (kubelet)[2189]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Jun 25 18:47:06.597176 kubelet[2189]: E0625 18:47:06.597120    2189 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Jun 25 18:47:06.599581 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Jun 25 18:47:06.599717 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Jun 25 18:47:06.906175 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Jun 25 18:47:06.918466 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jun 25 18:47:06.934424 systemd[1]: Reloading requested from client PID 2207 ('systemctl') (unit session-7.scope)...
Jun 25 18:47:06.934443 systemd[1]: Reloading...
Jun 25 18:47:07.013252 zram_generator::config[2245]: No configuration found.
Jun 25 18:47:07.145769 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jun 25 18:47:07.198242 systemd[1]: Reloading finished in 263 ms.
Jun 25 18:47:07.245784 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM
Jun 25 18:47:07.245854 systemd[1]: kubelet.service: Failed with result 'signal'.
Jun 25 18:47:07.246126 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Jun 25 18:47:07.247744 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jun 25 18:47:07.344202 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jun 25 18:47:07.350306 (kubelet)[2301]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS
Jun 25 18:47:07.392000 kubelet[2301]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jun 25 18:47:07.392000 kubelet[2301]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI.
Jun 25 18:47:07.392000 kubelet[2301]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jun 25 18:47:07.392394 kubelet[2301]: I0625 18:47:07.392045    2301 server.go:203] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Jun 25 18:47:07.976404 kubelet[2301]: I0625 18:47:07.975432    2301 server.go:467] "Kubelet version" kubeletVersion="v1.28.7"
Jun 25 18:47:07.976404 kubelet[2301]: I0625 18:47:07.975461    2301 server.go:469] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Jun 25 18:47:07.976404 kubelet[2301]: I0625 18:47:07.975659    2301 server.go:895] "Client rotation is on, will bootstrap in background"
Jun 25 18:47:07.993541 kubelet[2301]: I0625 18:47:07.993503    2301 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt"
Jun 25 18:47:07.994351 kubelet[2301]: E0625 18:47:07.994327    2301 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.141:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.007748 kubelet[2301]: W0625 18:47:08.007707    2301 machine.go:65] Cannot read vendor id correctly, set empty.
Jun 25 18:47:08.008451 kubelet[2301]: I0625 18:47:08.008422    2301 server.go:725] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Jun 25 18:47:08.008758 kubelet[2301]: I0625 18:47:08.008732    2301 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Jun 25 18:47:08.008923 kubelet[2301]: I0625 18:47:08.008898    2301 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null}
Jun 25 18:47:08.009010 kubelet[2301]: I0625 18:47:08.008927    2301 topology_manager.go:138] "Creating topology manager with none policy"
Jun 25 18:47:08.009010 kubelet[2301]: I0625 18:47:08.008936    2301 container_manager_linux.go:301] "Creating device plugin manager"
Jun 25 18:47:08.009130 kubelet[2301]: I0625 18:47:08.009103    2301 state_mem.go:36] "Initialized new in-memory state store"
Jun 25 18:47:08.010239 kubelet[2301]: I0625 18:47:08.010216    2301 kubelet.go:393] "Attempting to sync node with API server"
Jun 25 18:47:08.010284 kubelet[2301]: I0625 18:47:08.010243    2301 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests"
Jun 25 18:47:08.010675 kubelet[2301]: I0625 18:47:08.010327    2301 kubelet.go:309] "Adding apiserver pod source"
Jun 25 18:47:08.010675 kubelet[2301]: I0625 18:47:08.010341    2301 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Jun 25 18:47:08.010925 kubelet[2301]: W0625 18:47:08.010881    2301 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://10.0.0.141:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.011021 kubelet[2301]: E0625 18:47:08.011008    2301 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.141:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.011102 kubelet[2301]: W0625 18:47:08.010894    2301 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://10.0.0.141:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.011192 kubelet[2301]: E0625 18:47:08.011180    2301 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.141:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.013389 kubelet[2301]: I0625 18:47:08.013368    2301 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="v1.7.18" apiVersion="v1"
Jun 25 18:47:08.018114 kubelet[2301]: W0625 18:47:08.018075    2301 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating.
Jun 25 18:47:08.018754 kubelet[2301]: I0625 18:47:08.018687    2301 server.go:1232] "Started kubelet"
Jun 25 18:47:08.019169 kubelet[2301]: I0625 18:47:08.018864    2301 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10
Jun 25 18:47:08.019169 kubelet[2301]: I0625 18:47:08.019116    2301 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock"
Jun 25 18:47:08.019169 kubelet[2301]: I0625 18:47:08.019171    2301 server.go:162] "Starting to listen" address="0.0.0.0" port=10250
Jun 25 18:47:08.019974 kubelet[2301]: I0625 18:47:08.019812    2301 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Jun 25 18:47:08.019974 kubelet[2301]: I0625 18:47:08.019900    2301 server.go:462] "Adding debug handlers to kubelet server"
Jun 25 18:47:08.022223 kubelet[2301]: E0625 18:47:08.021304    2301 cri_stats_provider.go:448] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs"
Jun 25 18:47:08.022223 kubelet[2301]: E0625 18:47:08.021332    2301 kubelet.go:1431] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem"
Jun 25 18:47:08.024034 kubelet[2301]: E0625 18:47:08.023946    2301 event.go:289] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"localhost.17dc53bcc7bc7d0d", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:<nil>, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"localhost", UID:"localhost", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"localhost"}, FirstTimestamp:time.Date(2024, time.June, 25, 18, 47, 8, 18662669, time.Local), LastTimestamp:time.Date(2024, time.June, 25, 18, 47, 8, 18662669, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"localhost"}': 'Post "https://10.0.0.141:6443/api/v1/namespaces/default/events": dial tcp 10.0.0.141:6443: connect: connection refused'(may retry after sleeping)
Jun 25 18:47:08.024757 kubelet[2301]: I0625 18:47:08.024180    2301 volume_manager.go:291] "Starting Kubelet Volume Manager"
Jun 25 18:47:08.024879 kubelet[2301]: I0625 18:47:08.024205    2301 desired_state_of_world_populator.go:151] "Desired state populator starts to run"
Jun 25 18:47:08.024985 kubelet[2301]: I0625 18:47:08.024973    2301 reconciler_new.go:29] "Reconciler: start to sync state"
Jun 25 18:47:08.025113 kubelet[2301]: E0625 18:47:08.024459    2301 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.141:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.141:6443: connect: connection refused" interval="200ms"
Jun 25 18:47:08.025167 kubelet[2301]: W0625 18:47:08.024690    2301 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://10.0.0.141:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.025250 kubelet[2301]: E0625 18:47:08.025238    2301 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.141:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.038770 kubelet[2301]: I0625 18:47:08.038744    2301 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4"
Jun 25 18:47:08.039845 kubelet[2301]: I0625 18:47:08.039830    2301 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6"
Jun 25 18:47:08.039947 kubelet[2301]: I0625 18:47:08.039937    2301 status_manager.go:217] "Starting to sync pod status with apiserver"
Jun 25 18:47:08.041892 kubelet[2301]: I0625 18:47:08.041873    2301 kubelet.go:2303] "Starting kubelet main sync loop"
Jun 25 18:47:08.042039 kubelet[2301]: E0625 18:47:08.042026    2301 kubelet.go:2327] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]"
Jun 25 18:47:08.042634 kubelet[2301]: W0625 18:47:08.042594    2301 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://10.0.0.141:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.042960 kubelet[2301]: E0625 18:47:08.042938    2301 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.141:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.061554 kubelet[2301]: I0625 18:47:08.061530    2301 cpu_manager.go:214] "Starting CPU manager" policy="none"
Jun 25 18:47:08.061694 kubelet[2301]: I0625 18:47:08.061682    2301 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s"
Jun 25 18:47:08.061755 kubelet[2301]: I0625 18:47:08.061747    2301 state_mem.go:36] "Initialized new in-memory state store"
Jun 25 18:47:08.125855 kubelet[2301]: I0625 18:47:08.125829    2301 kubelet_node_status.go:70] "Attempting to register node" node="localhost"
Jun 25 18:47:08.126320 kubelet[2301]: E0625 18:47:08.126273    2301 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.141:6443/api/v1/nodes\": dial tcp 10.0.0.141:6443: connect: connection refused" node="localhost"
Jun 25 18:47:08.142520 kubelet[2301]: E0625 18:47:08.142497    2301 kubelet.go:2327] "Skipping pod synchronization" err="container runtime status check may not have completed yet"
Jun 25 18:47:08.161118 kubelet[2301]: I0625 18:47:08.160983    2301 policy_none.go:49] "None policy: Start"
Jun 25 18:47:08.161734 kubelet[2301]: I0625 18:47:08.161702    2301 memory_manager.go:169] "Starting memorymanager" policy="None"
Jun 25 18:47:08.161734 kubelet[2301]: I0625 18:47:08.161730    2301 state_mem.go:35] "Initializing new in-memory state store"
Jun 25 18:47:08.172504 kubelet[2301]: I0625 18:47:08.172476    2301 manager.go:471] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found"
Jun 25 18:47:08.172744 kubelet[2301]: I0625 18:47:08.172727    2301 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Jun 25 18:47:08.174374 kubelet[2301]: E0625 18:47:08.174354    2301 eviction_manager.go:258] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found"
Jun 25 18:47:08.226272 kubelet[2301]: E0625 18:47:08.226248    2301 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.141:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.141:6443: connect: connection refused" interval="400ms"
Jun 25 18:47:08.327589 kubelet[2301]: I0625 18:47:08.327488    2301 kubelet_node_status.go:70] "Attempting to register node" node="localhost"
Jun 25 18:47:08.328373 kubelet[2301]: E0625 18:47:08.328347    2301 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.141:6443/api/v1/nodes\": dial tcp 10.0.0.141:6443: connect: connection refused" node="localhost"
Jun 25 18:47:08.343532 kubelet[2301]: I0625 18:47:08.343499    2301 topology_manager.go:215] "Topology Admit Handler" podUID="cf2738f0339c495ab6929b1dd88bcc6c" podNamespace="kube-system" podName="kube-apiserver-localhost"
Jun 25 18:47:08.345036 kubelet[2301]: I0625 18:47:08.344569    2301 topology_manager.go:215] "Topology Admit Handler" podUID="d27baad490d2d4f748c86b318d7d74ef" podNamespace="kube-system" podName="kube-controller-manager-localhost"
Jun 25 18:47:08.345451 kubelet[2301]: I0625 18:47:08.345434    2301 topology_manager.go:215] "Topology Admit Handler" podUID="9c3207d669e00aa24ded52617c0d65d0" podNamespace="kube-system" podName="kube-scheduler-localhost"
Jun 25 18:47:08.427848 kubelet[2301]: I0625 18:47:08.427814    2301 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:08.428363 kubelet[2301]: I0625 18:47:08.428262    2301 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/9c3207d669e00aa24ded52617c0d65d0-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"9c3207d669e00aa24ded52617c0d65d0\") " pod="kube-system/kube-scheduler-localhost"
Jun 25 18:47:08.428363 kubelet[2301]: I0625 18:47:08.428295    2301 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/cf2738f0339c495ab6929b1dd88bcc6c-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"cf2738f0339c495ab6929b1dd88bcc6c\") " pod="kube-system/kube-apiserver-localhost"
Jun 25 18:47:08.428363 kubelet[2301]: I0625 18:47:08.428316    2301 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/cf2738f0339c495ab6929b1dd88bcc6c-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"cf2738f0339c495ab6929b1dd88bcc6c\") " pod="kube-system/kube-apiserver-localhost"
Jun 25 18:47:08.428363 kubelet[2301]: I0625 18:47:08.428337    2301 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:08.428562 kubelet[2301]: I0625 18:47:08.428386    2301 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:08.428562 kubelet[2301]: I0625 18:47:08.428422    2301 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/cf2738f0339c495ab6929b1dd88bcc6c-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"cf2738f0339c495ab6929b1dd88bcc6c\") " pod="kube-system/kube-apiserver-localhost"
Jun 25 18:47:08.428562 kubelet[2301]: I0625 18:47:08.428447    2301 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:08.428562 kubelet[2301]: I0625 18:47:08.428468    2301 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:08.627607 kubelet[2301]: E0625 18:47:08.627175    2301 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.141:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.141:6443: connect: connection refused" interval="800ms"
Jun 25 18:47:08.649247 kubelet[2301]: E0625 18:47:08.649198    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:08.649865 kubelet[2301]: E0625 18:47:08.649667    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:08.649940 containerd[1563]: time="2024-06-25T18:47:08.649799871Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:cf2738f0339c495ab6929b1dd88bcc6c,Namespace:kube-system,Attempt:0,}"
Jun 25 18:47:08.650244 containerd[1563]: time="2024-06-25T18:47:08.649973072Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:d27baad490d2d4f748c86b318d7d74ef,Namespace:kube-system,Attempt:0,}"
Jun 25 18:47:08.650283 kubelet[2301]: E0625 18:47:08.650195    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:08.650526 containerd[1563]: time="2024-06-25T18:47:08.650495879Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:9c3207d669e00aa24ded52617c0d65d0,Namespace:kube-system,Attempt:0,}"
Jun 25 18:47:08.729841 kubelet[2301]: I0625 18:47:08.729817    2301 kubelet_node_status.go:70] "Attempting to register node" node="localhost"
Jun 25 18:47:08.730108 kubelet[2301]: E0625 18:47:08.730093    2301 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.141:6443/api/v1/nodes\": dial tcp 10.0.0.141:6443: connect: connection refused" node="localhost"
Jun 25 18:47:08.820730 kubelet[2301]: W0625 18:47:08.820671    2301 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://10.0.0.141:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:08.820730 kubelet[2301]: E0625 18:47:08.820732    2301 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.141:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:09.027899 kubelet[2301]: W0625 18:47:09.027838    2301 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://10.0.0.141:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:09.027899 kubelet[2301]: E0625 18:47:09.027900    2301 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.141:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:09.155683 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount491919248.mount: Deactivated successfully.
Jun 25 18:47:09.161455 containerd[1563]: time="2024-06-25T18:47:09.161400668Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jun 25 18:47:09.162255 containerd[1563]: time="2024-06-25T18:47:09.162202804Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0"
Jun 25 18:47:09.162939 containerd[1563]: time="2024-06-25T18:47:09.162899893Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jun 25 18:47:09.164729 containerd[1563]: time="2024-06-25T18:47:09.164691273Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0"
Jun 25 18:47:09.164818 containerd[1563]: time="2024-06-25T18:47:09.164783185Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jun 25 18:47:09.166336 containerd[1563]: time="2024-06-25T18:47:09.166302955Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jun 25 18:47:09.166445 containerd[1563]: time="2024-06-25T18:47:09.166419136Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269175"
Jun 25 18:47:09.167962 containerd[1563]: time="2024-06-25T18:47:09.167542143Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jun 25 18:47:09.173343 containerd[1563]: time="2024-06-25T18:47:09.173303595Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 522.730051ms"
Jun 25 18:47:09.174587 containerd[1563]: time="2024-06-25T18:47:09.174364005Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 524.46568ms"
Jun 25 18:47:09.176425 containerd[1563]: time="2024-06-25T18:47:09.176323671Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 526.279782ms"
Jun 25 18:47:09.315292 kubelet[2301]: W0625 18:47:09.314994    2301 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://10.0.0.141:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:09.315292 kubelet[2301]: E0625 18:47:09.315040    2301 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.141:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:09.320339 containerd[1563]: time="2024-06-25T18:47:09.320224849Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 25 18:47:09.320339 containerd[1563]: time="2024-06-25T18:47:09.320272267Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:09.320339 containerd[1563]: time="2024-06-25T18:47:09.319930771Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 25 18:47:09.320339 containerd[1563]: time="2024-06-25T18:47:09.319987840Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:09.320339 containerd[1563]: time="2024-06-25T18:47:09.320006103Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 25 18:47:09.320339 containerd[1563]: time="2024-06-25T18:47:09.320019799Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:09.320821 containerd[1563]: time="2024-06-25T18:47:09.320412397Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 25 18:47:09.320821 containerd[1563]: time="2024-06-25T18:47:09.320478838Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:09.325266 containerd[1563]: time="2024-06-25T18:47:09.325158013Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 25 18:47:09.325266 containerd[1563]: time="2024-06-25T18:47:09.325204550Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:09.325427 containerd[1563]: time="2024-06-25T18:47:09.325279681Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 25 18:47:09.325427 containerd[1563]: time="2024-06-25T18:47:09.325312321Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:09.360752 kubelet[2301]: W0625 18:47:09.360256    2301 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://10.0.0.141:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:09.360752 kubelet[2301]: E0625 18:47:09.360311    2301 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.141:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.141:6443: connect: connection refused
Jun 25 18:47:09.369471 containerd[1563]: time="2024-06-25T18:47:09.369430175Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:9c3207d669e00aa24ded52617c0d65d0,Namespace:kube-system,Attempt:0,} returns sandbox id \"e835d7aae80310a5d2ccaa6303c0a5364d9de4c281aec2045c52b364702b732f\""
Jun 25 18:47:09.375565 containerd[1563]: time="2024-06-25T18:47:09.375536247Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:cf2738f0339c495ab6929b1dd88bcc6c,Namespace:kube-system,Attempt:0,} returns sandbox id \"acd73f701066cce8947ac3616de54a9125ccfd556bdae3194b978c2192f82552\""
Jun 25 18:47:09.379337 kubelet[2301]: E0625 18:47:09.379307    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:09.380006 containerd[1563]: time="2024-06-25T18:47:09.379973848Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:d27baad490d2d4f748c86b318d7d74ef,Namespace:kube-system,Attempt:0,} returns sandbox id \"03ab353affa22daf0ef6b0009a434af04cb4f767374faad7bc4b6bc0ae8c160f\""
Jun 25 18:47:09.380200 kubelet[2301]: E0625 18:47:09.380180    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:09.381441 kubelet[2301]: E0625 18:47:09.381416    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:09.383395 containerd[1563]: time="2024-06-25T18:47:09.383363453Z" level=info msg="CreateContainer within sandbox \"e835d7aae80310a5d2ccaa6303c0a5364d9de4c281aec2045c52b364702b732f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Jun 25 18:47:09.384144 containerd[1563]: time="2024-06-25T18:47:09.384085252Z" level=info msg="CreateContainer within sandbox \"acd73f701066cce8947ac3616de54a9125ccfd556bdae3194b978c2192f82552\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Jun 25 18:47:09.385052 containerd[1563]: time="2024-06-25T18:47:09.385016665Z" level=info msg="CreateContainer within sandbox \"03ab353affa22daf0ef6b0009a434af04cb4f767374faad7bc4b6bc0ae8c160f\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Jun 25 18:47:09.400988 containerd[1563]: time="2024-06-25T18:47:09.400815253Z" level=info msg="CreateContainer within sandbox \"e835d7aae80310a5d2ccaa6303c0a5364d9de4c281aec2045c52b364702b732f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"7ad67926fab39fe8292031c15b4ec7b58362357d577e0293c39c8d19767cb63e\""
Jun 25 18:47:09.401642 containerd[1563]: time="2024-06-25T18:47:09.401612624Z" level=info msg="StartContainer for \"7ad67926fab39fe8292031c15b4ec7b58362357d577e0293c39c8d19767cb63e\""
Jun 25 18:47:09.403819 containerd[1563]: time="2024-06-25T18:47:09.403773173Z" level=info msg="CreateContainer within sandbox \"acd73f701066cce8947ac3616de54a9125ccfd556bdae3194b978c2192f82552\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"e14237bc69066d6c2129dd2668cecce1554320f3b30040d80bad3399f8ac3dd8\""
Jun 25 18:47:09.404878 containerd[1563]: time="2024-06-25T18:47:09.404175383Z" level=info msg="CreateContainer within sandbox \"03ab353affa22daf0ef6b0009a434af04cb4f767374faad7bc4b6bc0ae8c160f\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"11215e7b618e70308920fc2e1ff528bfe810c22129ec044ee5825f23a4b0f41b\""
Jun 25 18:47:09.404878 containerd[1563]: time="2024-06-25T18:47:09.404176304Z" level=info msg="StartContainer for \"e14237bc69066d6c2129dd2668cecce1554320f3b30040d80bad3399f8ac3dd8\""
Jun 25 18:47:09.404878 containerd[1563]: time="2024-06-25T18:47:09.404451479Z" level=info msg="StartContainer for \"11215e7b618e70308920fc2e1ff528bfe810c22129ec044ee5825f23a4b0f41b\""
Jun 25 18:47:09.429227 kubelet[2301]: E0625 18:47:09.428443    2301 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.141:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.141:6443: connect: connection refused" interval="1.6s"
Jun 25 18:47:09.458765 containerd[1563]: time="2024-06-25T18:47:09.458711637Z" level=info msg="StartContainer for \"11215e7b618e70308920fc2e1ff528bfe810c22129ec044ee5825f23a4b0f41b\" returns successfully"
Jun 25 18:47:09.474162 containerd[1563]: time="2024-06-25T18:47:09.468673121Z" level=info msg="StartContainer for \"7ad67926fab39fe8292031c15b4ec7b58362357d577e0293c39c8d19767cb63e\" returns successfully"
Jun 25 18:47:09.474162 containerd[1563]: time="2024-06-25T18:47:09.468746010Z" level=info msg="StartContainer for \"e14237bc69066d6c2129dd2668cecce1554320f3b30040d80bad3399f8ac3dd8\" returns successfully"
Jun 25 18:47:09.536761 kubelet[2301]: I0625 18:47:09.536522    2301 kubelet_node_status.go:70] "Attempting to register node" node="localhost"
Jun 25 18:47:09.536963 kubelet[2301]: E0625 18:47:09.536948    2301 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.141:6443/api/v1/nodes\": dial tcp 10.0.0.141:6443: connect: connection refused" node="localhost"
Jun 25 18:47:10.050970 kubelet[2301]: E0625 18:47:10.050938    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:10.057306 kubelet[2301]: E0625 18:47:10.057278    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:10.057771 kubelet[2301]: E0625 18:47:10.057750    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:11.064260 kubelet[2301]: E0625 18:47:11.062546    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:11.140220 kubelet[2301]: I0625 18:47:11.138685    2301 kubelet_node_status.go:70] "Attempting to register node" node="localhost"
Jun 25 18:47:11.262579 kubelet[2301]: E0625 18:47:11.262542    2301 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost"
Jun 25 18:47:11.321711 kubelet[2301]: I0625 18:47:11.321598    2301 kubelet_node_status.go:73] "Successfully registered node" node="localhost"
Jun 25 18:47:11.329269 kubelet[2301]: E0625 18:47:11.329007    2301 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found"
Jun 25 18:47:11.429391 kubelet[2301]: E0625 18:47:11.429357    2301 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found"
Jun 25 18:47:11.530046 kubelet[2301]: E0625 18:47:11.530016    2301 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found"
Jun 25 18:47:11.631339 kubelet[2301]: E0625 18:47:11.631091    2301 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found"
Jun 25 18:47:11.732105 kubelet[2301]: E0625 18:47:11.732076    2301 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found"
Jun 25 18:47:11.833001 kubelet[2301]: E0625 18:47:11.832976    2301 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found"
Jun 25 18:47:11.933649 kubelet[2301]: E0625 18:47:11.933574    2301 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found"
Jun 25 18:47:12.033859 kubelet[2301]: E0625 18:47:12.033837    2301 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found"
Jun 25 18:47:12.063619 kubelet[2301]: E0625 18:47:12.063488    2301 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:12.134293 kubelet[2301]: E0625 18:47:12.134271    2301 kubelet_node_status.go:458] "Error getting the current node from lister" err="node \"localhost\" not found"
Jun 25 18:47:13.013120 kubelet[2301]: I0625 18:47:13.012897    2301 apiserver.go:52] "Watching apiserver"
Jun 25 18:47:13.025587 kubelet[2301]: I0625 18:47:13.025555    2301 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world"
Jun 25 18:47:13.959882 systemd[1]: Reloading requested from client PID 2575 ('systemctl') (unit session-7.scope)...
Jun 25 18:47:13.959898 systemd[1]: Reloading...
Jun 25 18:47:14.014488 zram_generator::config[2613]: No configuration found.
Jun 25 18:47:14.199650 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jun 25 18:47:14.255584 systemd[1]: Reloading finished in 295 ms.
Jun 25 18:47:14.284427 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent...
Jun 25 18:47:14.298177 systemd[1]: kubelet.service: Deactivated successfully.
Jun 25 18:47:14.298590 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Jun 25 18:47:14.309536 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jun 25 18:47:14.392418 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jun 25 18:47:14.395918 (kubelet)[2664]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS
Jun 25 18:47:14.436336 kubelet[2664]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jun 25 18:47:14.436336 kubelet[2664]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI.
Jun 25 18:47:14.436336 kubelet[2664]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jun 25 18:47:14.436761 kubelet[2664]: I0625 18:47:14.436378    2664 server.go:203] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Jun 25 18:47:14.441452 kubelet[2664]: I0625 18:47:14.441428    2664 server.go:467] "Kubelet version" kubeletVersion="v1.28.7"
Jun 25 18:47:14.441452 kubelet[2664]: I0625 18:47:14.441452    2664 server.go:469] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Jun 25 18:47:14.442220 kubelet[2664]: I0625 18:47:14.441872    2664 server.go:895] "Client rotation is on, will bootstrap in background"
Jun 25 18:47:14.443957 kubelet[2664]: I0625 18:47:14.443934    2664 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem".
Jun 25 18:47:14.444857 kubelet[2664]: I0625 18:47:14.444833    2664 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt"
Jun 25 18:47:14.448449 kubelet[2664]: W0625 18:47:14.448428    2664 machine.go:65] Cannot read vendor id correctly, set empty.
Jun 25 18:47:14.449286 kubelet[2664]: I0625 18:47:14.449268    2664 server.go:725] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Jun 25 18:47:14.449701 kubelet[2664]: I0625 18:47:14.449687    2664 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Jun 25 18:47:14.449846 kubelet[2664]: I0625 18:47:14.449832    2664 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null}
Jun 25 18:47:14.449916 kubelet[2664]: I0625 18:47:14.449861    2664 topology_manager.go:138] "Creating topology manager with none policy"
Jun 25 18:47:14.449916 kubelet[2664]: I0625 18:47:14.449870    2664 container_manager_linux.go:301] "Creating device plugin manager"
Jun 25 18:47:14.449916 kubelet[2664]: I0625 18:47:14.449901    2664 state_mem.go:36] "Initialized new in-memory state store"
Jun 25 18:47:14.449984 kubelet[2664]: I0625 18:47:14.449978    2664 kubelet.go:393] "Attempting to sync node with API server"
Jun 25 18:47:14.450007 kubelet[2664]: I0625 18:47:14.449991    2664 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests"
Jun 25 18:47:14.450032 kubelet[2664]: I0625 18:47:14.450013    2664 kubelet.go:309] "Adding apiserver pod source"
Jun 25 18:47:14.450032 kubelet[2664]: I0625 18:47:14.450023    2664 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Jun 25 18:47:14.454219 kubelet[2664]: I0625 18:47:14.451185    2664 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="v1.7.18" apiVersion="v1"
Jun 25 18:47:14.454219 kubelet[2664]: I0625 18:47:14.451611    2664 server.go:1232] "Started kubelet"
Jun 25 18:47:14.454219 kubelet[2664]: I0625 18:47:14.453028    2664 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10
Jun 25 18:47:14.454219 kubelet[2664]: I0625 18:47:14.453222    2664 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock"
Jun 25 18:47:14.454219 kubelet[2664]: I0625 18:47:14.453262    2664 server.go:162] "Starting to listen" address="0.0.0.0" port=10250
Jun 25 18:47:14.454219 kubelet[2664]: I0625 18:47:14.453913    2664 server.go:462] "Adding debug handlers to kubelet server"
Jun 25 18:47:14.455315 kubelet[2664]: E0625 18:47:14.454753    2664 cri_stats_provider.go:448] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs"
Jun 25 18:47:14.455315 kubelet[2664]: E0625 18:47:14.454787    2664 kubelet.go:1431] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem"
Jun 25 18:47:14.470424 kubelet[2664]: I0625 18:47:14.469496    2664 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Jun 25 18:47:14.475552 kubelet[2664]: I0625 18:47:14.475525    2664 volume_manager.go:291] "Starting Kubelet Volume Manager"
Jun 25 18:47:14.475665 kubelet[2664]: I0625 18:47:14.475651    2664 desired_state_of_world_populator.go:151] "Desired state populator starts to run"
Jun 25 18:47:14.475804 kubelet[2664]: I0625 18:47:14.475793    2664 reconciler_new.go:29] "Reconciler: start to sync state"
Jun 25 18:47:14.476064 sudo[2681]:     root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin
Jun 25 18:47:14.476375 sudo[2681]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0)
Jun 25 18:47:14.484552 kubelet[2664]: I0625 18:47:14.484375    2664 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4"
Jun 25 18:47:14.485268 kubelet[2664]: I0625 18:47:14.485247    2664 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6"
Jun 25 18:47:14.485268 kubelet[2664]: I0625 18:47:14.485267    2664 status_manager.go:217] "Starting to sync pod status with apiserver"
Jun 25 18:47:14.485591 kubelet[2664]: I0625 18:47:14.485286    2664 kubelet.go:2303] "Starting kubelet main sync loop"
Jun 25 18:47:14.485591 kubelet[2664]: E0625 18:47:14.485336    2664 kubelet.go:2327] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]"
Jun 25 18:47:14.538239 kubelet[2664]: I0625 18:47:14.538134    2664 cpu_manager.go:214] "Starting CPU manager" policy="none"
Jun 25 18:47:14.538239 kubelet[2664]: I0625 18:47:14.538158    2664 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s"
Jun 25 18:47:14.538239 kubelet[2664]: I0625 18:47:14.538176    2664 state_mem.go:36] "Initialized new in-memory state store"
Jun 25 18:47:14.539386 kubelet[2664]: I0625 18:47:14.539343    2664 state_mem.go:88] "Updated default CPUSet" cpuSet=""
Jun 25 18:47:14.539386 kubelet[2664]: I0625 18:47:14.539375    2664 state_mem.go:96] "Updated CPUSet assignments" assignments={}
Jun 25 18:47:14.539386 kubelet[2664]: I0625 18:47:14.539383    2664 policy_none.go:49] "None policy: Start"
Jun 25 18:47:14.540005 kubelet[2664]: I0625 18:47:14.539984    2664 memory_manager.go:169] "Starting memorymanager" policy="None"
Jun 25 18:47:14.540064 kubelet[2664]: I0625 18:47:14.540012    2664 state_mem.go:35] "Initializing new in-memory state store"
Jun 25 18:47:14.540522 kubelet[2664]: I0625 18:47:14.540502    2664 state_mem.go:75] "Updated machine memory state"
Jun 25 18:47:14.541917 kubelet[2664]: I0625 18:47:14.541568    2664 manager.go:471] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found"
Jun 25 18:47:14.543403 kubelet[2664]: I0625 18:47:14.542802    2664 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Jun 25 18:47:14.586382 kubelet[2664]: I0625 18:47:14.586347    2664 topology_manager.go:215] "Topology Admit Handler" podUID="cf2738f0339c495ab6929b1dd88bcc6c" podNamespace="kube-system" podName="kube-apiserver-localhost"
Jun 25 18:47:14.586523 kubelet[2664]: I0625 18:47:14.586506    2664 topology_manager.go:215] "Topology Admit Handler" podUID="d27baad490d2d4f748c86b318d7d74ef" podNamespace="kube-system" podName="kube-controller-manager-localhost"
Jun 25 18:47:14.586558 kubelet[2664]: I0625 18:47:14.586552    2664 topology_manager.go:215] "Topology Admit Handler" podUID="9c3207d669e00aa24ded52617c0d65d0" podNamespace="kube-system" podName="kube-scheduler-localhost"
Jun 25 18:47:14.647022 kubelet[2664]: I0625 18:47:14.646994    2664 kubelet_node_status.go:70] "Attempting to register node" node="localhost"
Jun 25 18:47:14.654254 kubelet[2664]: I0625 18:47:14.653489    2664 kubelet_node_status.go:108] "Node was previously registered" node="localhost"
Jun 25 18:47:14.654254 kubelet[2664]: I0625 18:47:14.653575    2664 kubelet_node_status.go:73] "Successfully registered node" node="localhost"
Jun 25 18:47:14.676101 kubelet[2664]: I0625 18:47:14.676076    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/cf2738f0339c495ab6929b1dd88bcc6c-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"cf2738f0339c495ab6929b1dd88bcc6c\") " pod="kube-system/kube-apiserver-localhost"
Jun 25 18:47:14.676198 kubelet[2664]: I0625 18:47:14.676113    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:14.676198 kubelet[2664]: I0625 18:47:14.676150    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:14.676198 kubelet[2664]: I0625 18:47:14.676167    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/cf2738f0339c495ab6929b1dd88bcc6c-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"cf2738f0339c495ab6929b1dd88bcc6c\") " pod="kube-system/kube-apiserver-localhost"
Jun 25 18:47:14.676198 kubelet[2664]: I0625 18:47:14.676186    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/cf2738f0339c495ab6929b1dd88bcc6c-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"cf2738f0339c495ab6929b1dd88bcc6c\") " pod="kube-system/kube-apiserver-localhost"
Jun 25 18:47:14.676306 kubelet[2664]: I0625 18:47:14.676203    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:14.676306 kubelet[2664]: I0625 18:47:14.676301    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:14.676878 kubelet[2664]: I0625 18:47:14.676354    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost"
Jun 25 18:47:14.676878 kubelet[2664]: I0625 18:47:14.676415    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/9c3207d669e00aa24ded52617c0d65d0-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"9c3207d669e00aa24ded52617c0d65d0\") " pod="kube-system/kube-scheduler-localhost"
Jun 25 18:47:14.895461 kubelet[2664]: E0625 18:47:14.895373    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:14.896941 kubelet[2664]: E0625 18:47:14.896902    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:14.897142 kubelet[2664]: E0625 18:47:14.897086    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:14.912591 sudo[2681]: pam_unix(sudo:session): session closed for user root
Jun 25 18:47:15.451417 kubelet[2664]: I0625 18:47:15.451377    2664 apiserver.go:52] "Watching apiserver"
Jun 25 18:47:15.476262 kubelet[2664]: I0625 18:47:15.476236    2664 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world"
Jun 25 18:47:15.501921 kubelet[2664]: E0625 18:47:15.501886    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:15.502519 kubelet[2664]: E0625 18:47:15.502499    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:15.508592 kubelet[2664]: E0625 18:47:15.508559    2664 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost"
Jun 25 18:47:15.509826 kubelet[2664]: E0625 18:47:15.509804    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:15.523021 kubelet[2664]: I0625 18:47:15.522987    2664 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.522941026 podCreationTimestamp="2024-06-25 18:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:47:15.522271941 +0000 UTC m=+1.122855337" watchObservedRunningTime="2024-06-25 18:47:15.522941026 +0000 UTC m=+1.123524422"
Jun 25 18:47:15.530062 kubelet[2664]: I0625 18:47:15.530025    2664 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.529998044 podCreationTimestamp="2024-06-25 18:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:47:15.52999144 +0000 UTC m=+1.130574876" watchObservedRunningTime="2024-06-25 18:47:15.529998044 +0000 UTC m=+1.130581440"
Jun 25 18:47:15.543884 kubelet[2664]: I0625 18:47:15.543857    2664 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.543829323 podCreationTimestamp="2024-06-25 18:47:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:47:15.536786154 +0000 UTC m=+1.137369550" watchObservedRunningTime="2024-06-25 18:47:15.543829323 +0000 UTC m=+1.144412720"
Jun 25 18:47:16.503624 kubelet[2664]: E0625 18:47:16.503578    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:17.326901 sudo[1769]: pam_unix(sudo:session): session closed for user root
Jun 25 18:47:17.328327 sshd[1762]: pam_unix(sshd:session): session closed for user core
Jun 25 18:47:17.332308 systemd[1]: sshd@6-10.0.0.141:22-10.0.0.1:43230.service: Deactivated successfully.
Jun 25 18:47:17.334569 systemd[1]: session-7.scope: Deactivated successfully.
Jun 25 18:47:17.334624 systemd-logind[1538]: Session 7 logged out. Waiting for processes to exit.
Jun 25 18:47:17.336169 systemd-logind[1538]: Removed session 7.
Jun 25 18:47:18.804253 kubelet[2664]: E0625 18:47:18.804203    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:19.674077 kubelet[2664]: E0625 18:47:19.674036    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:20.510224 kubelet[2664]: E0625 18:47:20.510176    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:23.255512 kubelet[2664]: E0625 18:47:23.255466    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:23.514520 kubelet[2664]: E0625 18:47:23.514396    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:27.935480 kubelet[2664]: I0625 18:47:27.935453    2664 kuberuntime_manager.go:1528] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24"
Jun 25 18:47:27.936175 kubelet[2664]: I0625 18:47:27.935987    2664 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24"
Jun 25 18:47:27.936240 containerd[1563]: time="2024-06-25T18:47:27.935818976Z" level=info msg="No cni config template is specified, wait for other system components to drop the config."
Jun 25 18:47:28.799691 kubelet[2664]: I0625 18:47:28.799621    2664 topology_manager.go:215] "Topology Admit Handler" podUID="60035cb9-2d59-4bd9-9256-adf0d1057f49" podNamespace="kube-system" podName="kube-proxy-jqzsn"
Jun 25 18:47:28.814279 kubelet[2664]: I0625 18:47:28.812496    2664 topology_manager.go:215] "Topology Admit Handler" podUID="71aed90d-6dae-4911-b0e6-4e9089070e4e" podNamespace="kube-system" podName="cilium-p8jq6"
Jun 25 18:47:28.825327 kubelet[2664]: E0625 18:47:28.825297    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:28.874928 kubelet[2664]: I0625 18:47:28.874898    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/71aed90d-6dae-4911-b0e6-4e9089070e4e-clustermesh-secrets\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875080 kubelet[2664]: I0625 18:47:28.874948    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/60035cb9-2d59-4bd9-9256-adf0d1057f49-xtables-lock\") pod \"kube-proxy-jqzsn\" (UID: \"60035cb9-2d59-4bd9-9256-adf0d1057f49\") " pod="kube-system/kube-proxy-jqzsn"
Jun 25 18:47:28.875080 kubelet[2664]: I0625 18:47:28.874972    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvptp\" (UniqueName: \"kubernetes.io/projected/60035cb9-2d59-4bd9-9256-adf0d1057f49-kube-api-access-jvptp\") pod \"kube-proxy-jqzsn\" (UID: \"60035cb9-2d59-4bd9-9256-adf0d1057f49\") " pod="kube-system/kube-proxy-jqzsn"
Jun 25 18:47:28.875080 kubelet[2664]: I0625 18:47:28.874994    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-config-path\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875080 kubelet[2664]: I0625 18:47:28.875012    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-host-proc-sys-net\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875080 kubelet[2664]: I0625 18:47:28.875031    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/60035cb9-2d59-4bd9-9256-adf0d1057f49-lib-modules\") pod \"kube-proxy-jqzsn\" (UID: \"60035cb9-2d59-4bd9-9256-adf0d1057f49\") " pod="kube-system/kube-proxy-jqzsn"
Jun 25 18:47:28.875339 kubelet[2664]: I0625 18:47:28.875050    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-xtables-lock\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875339 kubelet[2664]: I0625 18:47:28.875076    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s2g2\" (UniqueName: \"kubernetes.io/projected/71aed90d-6dae-4911-b0e6-4e9089070e4e-kube-api-access-4s2g2\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875339 kubelet[2664]: I0625 18:47:28.875115    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-hostproc\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875339 kubelet[2664]: I0625 18:47:28.875168    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-cgroup\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875592 kubelet[2664]: I0625 18:47:28.875439    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-host-proc-sys-kernel\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875592 kubelet[2664]: I0625 18:47:28.875485    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-run\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875592 kubelet[2664]: I0625 18:47:28.875527    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/60035cb9-2d59-4bd9-9256-adf0d1057f49-kube-proxy\") pod \"kube-proxy-jqzsn\" (UID: \"60035cb9-2d59-4bd9-9256-adf0d1057f49\") " pod="kube-system/kube-proxy-jqzsn"
Jun 25 18:47:28.875592 kubelet[2664]: I0625 18:47:28.875550    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cni-path\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875592 kubelet[2664]: I0625 18:47:28.875570    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-etc-cni-netd\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875849 kubelet[2664]: I0625 18:47:28.875751    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-lib-modules\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875849 kubelet[2664]: I0625 18:47:28.875778    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-bpf-maps\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.875849 kubelet[2664]: I0625 18:47:28.875811    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/71aed90d-6dae-4911-b0e6-4e9089070e4e-hubble-tls\") pod \"cilium-p8jq6\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") " pod="kube-system/cilium-p8jq6"
Jun 25 18:47:28.894448 kubelet[2664]: I0625 18:47:28.893232    2664 topology_manager.go:215] "Topology Admit Handler" podUID="fd86cd23-88a9-489b-ad27-9ea7130e570c" podNamespace="kube-system" podName="cilium-operator-6bc8ccdb58-42sxg"
Jun 25 18:47:28.977368 kubelet[2664]: I0625 18:47:28.977306    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/fd86cd23-88a9-489b-ad27-9ea7130e570c-cilium-config-path\") pod \"cilium-operator-6bc8ccdb58-42sxg\" (UID: \"fd86cd23-88a9-489b-ad27-9ea7130e570c\") " pod="kube-system/cilium-operator-6bc8ccdb58-42sxg"
Jun 25 18:47:28.977829 kubelet[2664]: I0625 18:47:28.977346    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbtwd\" (UniqueName: \"kubernetes.io/projected/fd86cd23-88a9-489b-ad27-9ea7130e570c-kube-api-access-xbtwd\") pod \"cilium-operator-6bc8ccdb58-42sxg\" (UID: \"fd86cd23-88a9-489b-ad27-9ea7130e570c\") " pod="kube-system/cilium-operator-6bc8ccdb58-42sxg"
Jun 25 18:47:29.116572 kubelet[2664]: E0625 18:47:29.116474    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:29.117250 containerd[1563]: time="2024-06-25T18:47:29.117125672Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jqzsn,Uid:60035cb9-2d59-4bd9-9256-adf0d1057f49,Namespace:kube-system,Attempt:0,}"
Jun 25 18:47:29.121311 kubelet[2664]: E0625 18:47:29.121287    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:29.122153 containerd[1563]: time="2024-06-25T18:47:29.121785664Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-p8jq6,Uid:71aed90d-6dae-4911-b0e6-4e9089070e4e,Namespace:kube-system,Attempt:0,}"
Jun 25 18:47:29.143131 containerd[1563]: time="2024-06-25T18:47:29.142662583Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 25 18:47:29.143131 containerd[1563]: time="2024-06-25T18:47:29.143107773Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:29.143131 containerd[1563]: time="2024-06-25T18:47:29.143125577Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 25 18:47:29.143131 containerd[1563]: time="2024-06-25T18:47:29.143140141Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:29.150616 containerd[1563]: time="2024-06-25T18:47:29.150399175Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 25 18:47:29.150616 containerd[1563]: time="2024-06-25T18:47:29.150448587Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:29.150616 containerd[1563]: time="2024-06-25T18:47:29.150462870Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 25 18:47:29.150616 containerd[1563]: time="2024-06-25T18:47:29.150472513Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:29.189285 containerd[1563]: time="2024-06-25T18:47:29.188505831Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-p8jq6,Uid:71aed90d-6dae-4911-b0e6-4e9089070e4e,Namespace:kube-system,Attempt:0,} returns sandbox id \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\""
Jun 25 18:47:29.189449 kubelet[2664]: E0625 18:47:29.189165    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:29.192040 containerd[1563]: time="2024-06-25T18:47:29.191846377Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\""
Jun 25 18:47:29.196679 containerd[1563]: time="2024-06-25T18:47:29.196638721Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-jqzsn,Uid:60035cb9-2d59-4bd9-9256-adf0d1057f49,Namespace:kube-system,Attempt:0,} returns sandbox id \"9e9599aa97fa8e17063cffd46def077fd6123590550858832a0478e0185dc807\""
Jun 25 18:47:29.197690 kubelet[2664]: E0625 18:47:29.197657    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:29.200546 containerd[1563]: time="2024-06-25T18:47:29.200461306Z" level=info msg="CreateContainer within sandbox \"9e9599aa97fa8e17063cffd46def077fd6123590550858832a0478e0185dc807\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}"
Jun 25 18:47:29.202429 kubelet[2664]: E0625 18:47:29.202401    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:29.204156 containerd[1563]: time="2024-06-25T18:47:29.203981015Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6bc8ccdb58-42sxg,Uid:fd86cd23-88a9-489b-ad27-9ea7130e570c,Namespace:kube-system,Attempt:0,}"
Jun 25 18:47:29.271391 containerd[1563]: time="2024-06-25T18:47:29.271346903Z" level=info msg="CreateContainer within sandbox \"9e9599aa97fa8e17063cffd46def077fd6123590550858832a0478e0185dc807\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"d7c259ee3d637a5ef6183a0e91d670cda52a18fdfcc19491c368a9ccb5d0437f\""
Jun 25 18:47:29.274193 containerd[1563]: time="2024-06-25T18:47:29.272719482Z" level=info msg="StartContainer for \"d7c259ee3d637a5ef6183a0e91d670cda52a18fdfcc19491c368a9ccb5d0437f\""
Jun 25 18:47:29.280351 containerd[1563]: time="2024-06-25T18:47:29.280155399Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 25 18:47:29.280351 containerd[1563]: time="2024-06-25T18:47:29.280275989Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:29.280351 containerd[1563]: time="2024-06-25T18:47:29.280290873Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 25 18:47:29.280351 containerd[1563]: time="2024-06-25T18:47:29.280300955Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:29.325157 containerd[1563]: time="2024-06-25T18:47:29.325113189Z" level=info msg="StartContainer for \"d7c259ee3d637a5ef6183a0e91d670cda52a18fdfcc19491c368a9ccb5d0437f\" returns successfully"
Jun 25 18:47:29.336639 containerd[1563]: time="2024-06-25T18:47:29.336601668Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-6bc8ccdb58-42sxg,Uid:fd86cd23-88a9-489b-ad27-9ea7130e570c,Namespace:kube-system,Attempt:0,} returns sandbox id \"136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be\""
Jun 25 18:47:29.337626 kubelet[2664]: E0625 18:47:29.337601    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:29.525602 kubelet[2664]: E0625 18:47:29.525283    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:29.535733 kubelet[2664]: I0625 18:47:29.535689    2664 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-jqzsn" podStartSLOduration=1.535655097 podCreationTimestamp="2024-06-25 18:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:47:29.533130633 +0000 UTC m=+15.133714029" watchObservedRunningTime="2024-06-25 18:47:29.535655097 +0000 UTC m=+15.136238493"
Jun 25 18:47:30.141914 update_engine[1544]: I0625 18:47:30.141813  1544 update_attempter.cc:509] Updating boot flags...
Jun 25 18:47:30.169097 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (3032)
Jun 25 18:47:30.216246 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 36 scanned by (udev-worker) (2979)
Jun 25 18:47:32.002327 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3895217572.mount: Deactivated successfully.
Jun 25 18:47:33.331655 containerd[1563]: time="2024-06-25T18:47:33.331598457Z" level=info msg="ImageCreate event name:\"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:33.332133 containerd[1563]: time="2024-06-25T18:47:33.332045268Z" level=info msg="stop pulling image quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5: active requests=0, bytes read=157651546"
Jun 25 18:47:33.332830 containerd[1563]: time="2024-06-25T18:47:33.332775616Z" level=info msg="ImageCreate event name:\"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:33.334970 containerd[1563]: time="2024-06-25T18:47:33.334349375Z" level=info msg="Pulled image \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" with image id \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\", repo tag \"\", repo digest \"quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\", size \"157636062\" in 4.142459308s"
Jun 25 18:47:33.334970 containerd[1563]: time="2024-06-25T18:47:33.334384703Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\""
Jun 25 18:47:33.335059 containerd[1563]: time="2024-06-25T18:47:33.334986185Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\""
Jun 25 18:47:33.341815 containerd[1563]: time="2024-06-25T18:47:33.341781203Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}"
Jun 25 18:47:33.353943 containerd[1563]: time="2024-06-25T18:47:33.353879498Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\""
Jun 25 18:47:33.355157 containerd[1563]: time="2024-06-25T18:47:33.354338031Z" level=info msg="StartContainer for \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\""
Jun 25 18:47:33.484294 containerd[1563]: time="2024-06-25T18:47:33.484137526Z" level=info msg="StartContainer for \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\" returns successfully"
Jun 25 18:47:33.502519 containerd[1563]: time="2024-06-25T18:47:33.502463405Z" level=info msg="shim disconnected" id=fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1 namespace=k8s.io
Jun 25 18:47:33.502519 containerd[1563]: time="2024-06-25T18:47:33.502515655Z" level=warning msg="cleaning up after shim disconnected" id=fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1 namespace=k8s.io
Jun 25 18:47:33.502693 containerd[1563]: time="2024-06-25T18:47:33.502565105Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:47:33.511692 containerd[1563]: time="2024-06-25T18:47:33.511642267Z" level=warning msg="cleanup warnings time=\"2024-06-25T18:47:33Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io
Jun 25 18:47:33.545224 kubelet[2664]: E0625 18:47:33.545161    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:33.547960 containerd[1563]: time="2024-06-25T18:47:33.547924789Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}"
Jun 25 18:47:33.558145 containerd[1563]: time="2024-06-25T18:47:33.558088531Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\""
Jun 25 18:47:33.558593 containerd[1563]: time="2024-06-25T18:47:33.558560066Z" level=info msg="StartContainer for \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\""
Jun 25 18:47:33.600278 containerd[1563]: time="2024-06-25T18:47:33.600163627Z" level=info msg="StartContainer for \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\" returns successfully"
Jun 25 18:47:33.632584 systemd[1]: systemd-sysctl.service: Deactivated successfully.
Jun 25 18:47:33.632865 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables.
Jun 25 18:47:33.632943 systemd[1]: Stopping systemd-sysctl.service - Apply Kernel Variables...
Jun 25 18:47:33.642526 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Jun 25 18:47:33.659666 containerd[1563]: time="2024-06-25T18:47:33.659607128Z" level=info msg="shim disconnected" id=70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14 namespace=k8s.io
Jun 25 18:47:33.659666 containerd[1563]: time="2024-06-25T18:47:33.659662259Z" level=warning msg="cleaning up after shim disconnected" id=70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14 namespace=k8s.io
Jun 25 18:47:33.659907 containerd[1563]: time="2024-06-25T18:47:33.659680023Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:47:33.661906 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Jun 25 18:47:34.351440 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1-rootfs.mount: Deactivated successfully.
Jun 25 18:47:34.492017 containerd[1563]: time="2024-06-25T18:47:34.491973885Z" level=info msg="ImageCreate event name:\"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:34.493219 containerd[1563]: time="2024-06-25T18:47:34.493177758Z" level=info msg="stop pulling image quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e: active requests=0, bytes read=17138386"
Jun 25 18:47:34.494014 containerd[1563]: time="2024-06-25T18:47:34.493991715Z" level=info msg="ImageCreate event name:\"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jun 25 18:47:34.495429 containerd[1563]: time="2024-06-25T18:47:34.495306330Z" level=info msg="Pulled image \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" with image id \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\", repo tag \"\", repo digest \"quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\", size \"17128551\" in 1.160289779s"
Jun 25 18:47:34.495429 containerd[1563]: time="2024-06-25T18:47:34.495342377Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\""
Jun 25 18:47:34.497390 containerd[1563]: time="2024-06-25T18:47:34.497345364Z" level=info msg="CreateContainer within sandbox \"136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}"
Jun 25 18:47:34.512488 containerd[1563]: time="2024-06-25T18:47:34.512443606Z" level=info msg="CreateContainer within sandbox \"136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\""
Jun 25 18:47:34.514248 containerd[1563]: time="2024-06-25T18:47:34.513518374Z" level=info msg="StartContainer for \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\""
Jun 25 18:47:34.553565 kubelet[2664]: E0625 18:47:34.553537    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:34.559674 containerd[1563]: time="2024-06-25T18:47:34.559381849Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}"
Jun 25 18:47:34.577126 containerd[1563]: time="2024-06-25T18:47:34.577089275Z" level=info msg="StartContainer for \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\" returns successfully"
Jun 25 18:47:34.578188 containerd[1563]: time="2024-06-25T18:47:34.578144000Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\""
Jun 25 18:47:34.579490 containerd[1563]: time="2024-06-25T18:47:34.579465975Z" level=info msg="StartContainer for \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\""
Jun 25 18:47:34.704609 containerd[1563]: time="2024-06-25T18:47:34.704429157Z" level=info msg="StartContainer for \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\" returns successfully"
Jun 25 18:47:34.727866 containerd[1563]: time="2024-06-25T18:47:34.727801320Z" level=info msg="shim disconnected" id=063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323 namespace=k8s.io
Jun 25 18:47:34.727866 containerd[1563]: time="2024-06-25T18:47:34.727861531Z" level=warning msg="cleaning up after shim disconnected" id=063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323 namespace=k8s.io
Jun 25 18:47:34.727866 containerd[1563]: time="2024-06-25T18:47:34.727870213Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:47:35.566558 kubelet[2664]: E0625 18:47:35.561513    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:35.573501 kubelet[2664]: E0625 18:47:35.573460    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:35.576194 kubelet[2664]: I0625 18:47:35.576170    2664 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-6bc8ccdb58-42sxg" podStartSLOduration=2.419171442 podCreationTimestamp="2024-06-25 18:47:28 +0000 UTC" firstStartedPulling="2024-06-25 18:47:29.338569794 +0000 UTC m=+14.939153190" lastFinishedPulling="2024-06-25 18:47:34.495534254 +0000 UTC m=+20.096117650" observedRunningTime="2024-06-25 18:47:35.575099871 +0000 UTC m=+21.175683307" watchObservedRunningTime="2024-06-25 18:47:35.576135902 +0000 UTC m=+21.176719298"
Jun 25 18:47:35.580316 containerd[1563]: time="2024-06-25T18:47:35.580157485Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}"
Jun 25 18:47:35.594431 containerd[1563]: time="2024-06-25T18:47:35.594381753Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\""
Jun 25 18:47:35.595304 containerd[1563]: time="2024-06-25T18:47:35.595016710Z" level=info msg="StartContainer for \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\""
Jun 25 18:47:35.649895 containerd[1563]: time="2024-06-25T18:47:35.649845077Z" level=info msg="StartContainer for \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\" returns successfully"
Jun 25 18:47:35.668606 containerd[1563]: time="2024-06-25T18:47:35.668540091Z" level=info msg="shim disconnected" id=3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b namespace=k8s.io
Jun 25 18:47:35.668606 containerd[1563]: time="2024-06-25T18:47:35.668595381Z" level=warning msg="cleaning up after shim disconnected" id=3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b namespace=k8s.io
Jun 25 18:47:35.668606 containerd[1563]: time="2024-06-25T18:47:35.668606223Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:47:36.349927 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b-rootfs.mount: Deactivated successfully.
Jun 25 18:47:36.577857 kubelet[2664]: E0625 18:47:36.577819    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:36.578258 kubelet[2664]: E0625 18:47:36.577879    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:36.580991 containerd[1563]: time="2024-06-25T18:47:36.580221553Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}"
Jun 25 18:47:36.592318 containerd[1563]: time="2024-06-25T18:47:36.591413248Z" level=info msg="CreateContainer within sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\""
Jun 25 18:47:36.592318 containerd[1563]: time="2024-06-25T18:47:36.591931420Z" level=info msg="StartContainer for \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\""
Jun 25 18:47:36.649875 containerd[1563]: time="2024-06-25T18:47:36.649774267Z" level=info msg="StartContainer for \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\" returns successfully"
Jun 25 18:47:36.802468 kubelet[2664]: I0625 18:47:36.802279    2664 kubelet_node_status.go:493] "Fast updating node status as it just became ready"
Jun 25 18:47:36.822951 kubelet[2664]: I0625 18:47:36.822920    2664 topology_manager.go:215] "Topology Admit Handler" podUID="c50f2a4c-9287-4bbb-adac-65bb4e0172e8" podNamespace="kube-system" podName="coredns-5dd5756b68-vhsn8"
Jun 25 18:47:36.826107 kubelet[2664]: I0625 18:47:36.823199    2664 topology_manager.go:215] "Topology Admit Handler" podUID="efbd4616-6b8c-424f-ae7a-1ec1ef2a80c4" podNamespace="kube-system" podName="coredns-5dd5756b68-zk649"
Jun 25 18:47:36.836141 kubelet[2664]: I0625 18:47:36.836113    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c50f2a4c-9287-4bbb-adac-65bb4e0172e8-config-volume\") pod \"coredns-5dd5756b68-vhsn8\" (UID: \"c50f2a4c-9287-4bbb-adac-65bb4e0172e8\") " pod="kube-system/coredns-5dd5756b68-vhsn8"
Jun 25 18:47:36.836354 kubelet[2664]: I0625 18:47:36.836342    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efbd4616-6b8c-424f-ae7a-1ec1ef2a80c4-config-volume\") pod \"coredns-5dd5756b68-zk649\" (UID: \"efbd4616-6b8c-424f-ae7a-1ec1ef2a80c4\") " pod="kube-system/coredns-5dd5756b68-zk649"
Jun 25 18:47:36.836600 kubelet[2664]: I0625 18:47:36.836581    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpd9l\" (UniqueName: \"kubernetes.io/projected/c50f2a4c-9287-4bbb-adac-65bb4e0172e8-kube-api-access-rpd9l\") pod \"coredns-5dd5756b68-vhsn8\" (UID: \"c50f2a4c-9287-4bbb-adac-65bb4e0172e8\") " pod="kube-system/coredns-5dd5756b68-vhsn8"
Jun 25 18:47:36.836694 kubelet[2664]: I0625 18:47:36.836684    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqt7f\" (UniqueName: \"kubernetes.io/projected/efbd4616-6b8c-424f-ae7a-1ec1ef2a80c4-kube-api-access-vqt7f\") pod \"coredns-5dd5756b68-zk649\" (UID: \"efbd4616-6b8c-424f-ae7a-1ec1ef2a80c4\") " pod="kube-system/coredns-5dd5756b68-zk649"
Jun 25 18:47:37.125568 kubelet[2664]: E0625 18:47:37.125463    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:37.127378 kubelet[2664]: E0625 18:47:37.127349    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:37.127575 containerd[1563]: time="2024-06-25T18:47:37.127261431Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-vhsn8,Uid:c50f2a4c-9287-4bbb-adac-65bb4e0172e8,Namespace:kube-system,Attempt:0,}"
Jun 25 18:47:37.127702 containerd[1563]: time="2024-06-25T18:47:37.127677541Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-zk649,Uid:efbd4616-6b8c-424f-ae7a-1ec1ef2a80c4,Namespace:kube-system,Attempt:0,}"
Jun 25 18:47:37.583865 kubelet[2664]: E0625 18:47:37.583124    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:37.597089 kubelet[2664]: I0625 18:47:37.597063    2664 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-p8jq6" podStartSLOduration=5.453653179 podCreationTimestamp="2024-06-25 18:47:28 +0000 UTC" firstStartedPulling="2024-06-25 18:47:29.191355536 +0000 UTC m=+14.791938932" lastFinishedPulling="2024-06-25 18:47:33.334731373 +0000 UTC m=+18.935314769" observedRunningTime="2024-06-25 18:47:37.595956995 +0000 UTC m=+23.196540471" watchObservedRunningTime="2024-06-25 18:47:37.597029016 +0000 UTC m=+23.197612412"
Jun 25 18:47:38.583954 kubelet[2664]: E0625 18:47:38.583878    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:38.869915 systemd-networkd[1237]: cilium_host: Link UP
Jun 25 18:47:38.870046 systemd-networkd[1237]: cilium_net: Link UP
Jun 25 18:47:38.870049 systemd-networkd[1237]: cilium_net: Gained carrier
Jun 25 18:47:38.870168 systemd-networkd[1237]: cilium_host: Gained carrier
Jun 25 18:47:38.870326 systemd-networkd[1237]: cilium_net: Gained IPv6LL
Jun 25 18:47:38.870472 systemd-networkd[1237]: cilium_host: Gained IPv6LL
Jun 25 18:47:38.938161 systemd-networkd[1237]: cilium_vxlan: Link UP
Jun 25 18:47:38.938168 systemd-networkd[1237]: cilium_vxlan: Gained carrier
Jun 25 18:47:39.238235 kernel: NET: Registered PF_ALG protocol family
Jun 25 18:47:39.585240 kubelet[2664]: E0625 18:47:39.585125    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:39.786094 systemd-networkd[1237]: lxc_health: Link UP
Jun 25 18:47:39.788825 systemd-networkd[1237]: lxc_health: Gained carrier
Jun 25 18:47:40.257786 systemd-networkd[1237]: lxc540b41549f6d: Link UP
Jun 25 18:47:40.267226 kernel: eth0: renamed from tmpcef98
Jun 25 18:47:40.274300 systemd-networkd[1237]: lxc820830bdec6f: Link UP
Jun 25 18:47:40.282266 kernel: eth0: renamed from tmp5f545
Jun 25 18:47:40.286070 systemd-networkd[1237]: lxc540b41549f6d: Gained carrier
Jun 25 18:47:40.287829 systemd-networkd[1237]: lxc820830bdec6f: Gained carrier
Jun 25 18:47:40.903349 systemd-networkd[1237]: cilium_vxlan: Gained IPv6LL
Jun 25 18:47:41.123895 kubelet[2664]: E0625 18:47:41.123855    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:41.479352 systemd-networkd[1237]: lxc540b41549f6d: Gained IPv6LL
Jun 25 18:47:41.543325 systemd-networkd[1237]: lxc820830bdec6f: Gained IPv6LL
Jun 25 18:47:41.735364 systemd-networkd[1237]: lxc_health: Gained IPv6LL
Jun 25 18:47:43.934623 containerd[1563]: time="2024-06-25T18:47:43.934503644Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 25 18:47:43.934623 containerd[1563]: time="2024-06-25T18:47:43.934564812Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:43.935132 containerd[1563]: time="2024-06-25T18:47:43.934655944Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 25 18:47:43.935132 containerd[1563]: time="2024-06-25T18:47:43.934709951Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:43.935132 containerd[1563]: time="2024-06-25T18:47:43.934481561Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 25 18:47:43.935503 containerd[1563]: time="2024-06-25T18:47:43.935436327Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:43.935503 containerd[1563]: time="2024-06-25T18:47:43.935471531Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 25 18:47:43.935591 containerd[1563]: time="2024-06-25T18:47:43.935482893Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:47:43.961792 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address
Jun 25 18:47:43.964120 systemd-resolved[1450]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address
Jun 25 18:47:43.983885 containerd[1563]: time="2024-06-25T18:47:43.983718715Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-vhsn8,Uid:c50f2a4c-9287-4bbb-adac-65bb4e0172e8,Namespace:kube-system,Attempt:0,} returns sandbox id \"cef98c4f97504897dfa8295a1d2046fd8ca254abfff1a25927646b964311091b\""
Jun 25 18:47:43.985778 kubelet[2664]: E0625 18:47:43.984410    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:43.986115 containerd[1563]: time="2024-06-25T18:47:43.985322406Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-zk649,Uid:efbd4616-6b8c-424f-ae7a-1ec1ef2a80c4,Namespace:kube-system,Attempt:0,} returns sandbox id \"5f545bd200ac0e55813d1a699ada0bb308726949b6dfba4be3b3e8770cb3a1dd\""
Jun 25 18:47:43.986964 kubelet[2664]: E0625 18:47:43.986795    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:43.988990 containerd[1563]: time="2024-06-25T18:47:43.988947403Z" level=info msg="CreateContainer within sandbox \"5f545bd200ac0e55813d1a699ada0bb308726949b6dfba4be3b3e8770cb3a1dd\" for container &ContainerMetadata{Name:coredns,Attempt:0,}"
Jun 25 18:47:44.003529 containerd[1563]: time="2024-06-25T18:47:44.003378690Z" level=info msg="CreateContainer within sandbox \"cef98c4f97504897dfa8295a1d2046fd8ca254abfff1a25927646b964311091b\" for container &ContainerMetadata{Name:coredns,Attempt:0,}"
Jun 25 18:47:44.006029 containerd[1563]: time="2024-06-25T18:47:44.005976539Z" level=info msg="CreateContainer within sandbox \"5f545bd200ac0e55813d1a699ada0bb308726949b6dfba4be3b3e8770cb3a1dd\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"5ad9136e93ca6739c364714fd85ee8191bd9435bc91dbcf7802880c7f12381a4\""
Jun 25 18:47:44.008192 containerd[1563]: time="2024-06-25T18:47:44.006759318Z" level=info msg="StartContainer for \"5ad9136e93ca6739c364714fd85ee8191bd9435bc91dbcf7802880c7f12381a4\""
Jun 25 18:47:44.013416 containerd[1563]: time="2024-06-25T18:47:44.013377315Z" level=info msg="CreateContainer within sandbox \"cef98c4f97504897dfa8295a1d2046fd8ca254abfff1a25927646b964311091b\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"0f1f4666eb85233ab0af30dedf82686e6b516a0835a47b344e7b6f5634c08d14\""
Jun 25 18:47:44.014294 containerd[1563]: time="2024-06-25T18:47:44.013834693Z" level=info msg="StartContainer for \"0f1f4666eb85233ab0af30dedf82686e6b516a0835a47b344e7b6f5634c08d14\""
Jun 25 18:47:44.065846 containerd[1563]: time="2024-06-25T18:47:44.062406361Z" level=info msg="StartContainer for \"5ad9136e93ca6739c364714fd85ee8191bd9435bc91dbcf7802880c7f12381a4\" returns successfully"
Jun 25 18:47:44.069463 containerd[1563]: time="2024-06-25T18:47:44.069400526Z" level=info msg="StartContainer for \"0f1f4666eb85233ab0af30dedf82686e6b516a0835a47b344e7b6f5634c08d14\" returns successfully"
Jun 25 18:47:44.594951 kubelet[2664]: E0625 18:47:44.594915    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:44.609345 kubelet[2664]: I0625 18:47:44.608006    2664 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5dd5756b68-vhsn8" podStartSLOduration=16.60795957 podCreationTimestamp="2024-06-25 18:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:47:44.605586709 +0000 UTC m=+30.206170065" watchObservedRunningTime="2024-06-25 18:47:44.60795957 +0000 UTC m=+30.208543006"
Jun 25 18:47:44.624848 kubelet[2664]: E0625 18:47:44.624164    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:44.638784 kubelet[2664]: I0625 18:47:44.636827    2664 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5dd5756b68-zk649" podStartSLOduration=16.636788738 podCreationTimestamp="2024-06-25 18:47:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:47:44.636647921 +0000 UTC m=+30.237231317" watchObservedRunningTime="2024-06-25 18:47:44.636788738 +0000 UTC m=+30.237372134"
Jun 25 18:47:45.626305 kubelet[2664]: E0625 18:47:45.625872    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:45.626305 kubelet[2664]: E0625 18:47:45.625942    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:46.425487 systemd[1]: Started sshd@7-10.0.0.141:22-10.0.0.1:60888.service - OpenSSH per-connection server daemon (10.0.0.1:60888).
Jun 25 18:47:46.467118 sshd[4058]: Accepted publickey for core from 10.0.0.1 port 60888 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:47:46.467826 sshd[4058]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:47:46.471601 systemd-logind[1538]: New session 8 of user core.
Jun 25 18:47:46.478542 systemd[1]: Started session-8.scope - Session 8 of User core.
Jun 25 18:47:46.628969 kubelet[2664]: E0625 18:47:46.628940    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:46.629835 kubelet[2664]: E0625 18:47:46.629795    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:46.632732 sshd[4058]: pam_unix(sshd:session): session closed for user core
Jun 25 18:47:46.635925 systemd[1]: sshd@7-10.0.0.141:22-10.0.0.1:60888.service: Deactivated successfully.
Jun 25 18:47:46.640649 systemd-logind[1538]: Session 8 logged out. Waiting for processes to exit.
Jun 25 18:47:46.640875 systemd[1]: session-8.scope: Deactivated successfully.
Jun 25 18:47:46.642995 systemd-logind[1538]: Removed session 8.
Jun 25 18:47:51.644503 systemd[1]: Started sshd@8-10.0.0.141:22-10.0.0.1:51294.service - OpenSSH per-connection server daemon (10.0.0.1:51294).
Jun 25 18:47:51.675712 sshd[4074]: Accepted publickey for core from 10.0.0.1 port 51294 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:47:51.677112 sshd[4074]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:47:51.681552 systemd-logind[1538]: New session 9 of user core.
Jun 25 18:47:51.692570 systemd[1]: Started session-9.scope - Session 9 of User core.
Jun 25 18:47:51.799438 sshd[4074]: pam_unix(sshd:session): session closed for user core
Jun 25 18:47:51.802965 systemd[1]: sshd@8-10.0.0.141:22-10.0.0.1:51294.service: Deactivated successfully.
Jun 25 18:47:51.805156 systemd-logind[1538]: Session 9 logged out. Waiting for processes to exit.
Jun 25 18:47:51.805543 systemd[1]: session-9.scope: Deactivated successfully.
Jun 25 18:47:51.806825 systemd-logind[1538]: Removed session 9.
Jun 25 18:47:56.817449 systemd[1]: Started sshd@9-10.0.0.141:22-10.0.0.1:51304.service - OpenSSH per-connection server daemon (10.0.0.1:51304).
Jun 25 18:47:56.848503 sshd[4090]: Accepted publickey for core from 10.0.0.1 port 51304 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:47:56.849824 sshd[4090]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:47:56.853263 systemd-logind[1538]: New session 10 of user core.
Jun 25 18:47:56.863422 systemd[1]: Started session-10.scope - Session 10 of User core.
Jun 25 18:47:56.967627 sshd[4090]: pam_unix(sshd:session): session closed for user core
Jun 25 18:47:56.970299 systemd[1]: sshd@9-10.0.0.141:22-10.0.0.1:51304.service: Deactivated successfully.
Jun 25 18:47:56.973409 systemd-logind[1538]: Session 10 logged out. Waiting for processes to exit.
Jun 25 18:47:56.973942 systemd[1]: session-10.scope: Deactivated successfully.
Jun 25 18:47:56.974924 systemd-logind[1538]: Removed session 10.
Jun 25 18:47:58.000530 kubelet[2664]: I0625 18:47:58.000462    2664 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness"
Jun 25 18:47:58.002225 kubelet[2664]: E0625 18:47:58.002186    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:47:58.655525 kubelet[2664]: E0625 18:47:58.655491    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:01.983425 systemd[1]: Started sshd@10-10.0.0.141:22-10.0.0.1:52864.service - OpenSSH per-connection server daemon (10.0.0.1:52864).
Jun 25 18:48:02.014147 sshd[4109]: Accepted publickey for core from 10.0.0.1 port 52864 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:02.015020 sshd[4109]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:02.018677 systemd-logind[1538]: New session 11 of user core.
Jun 25 18:48:02.023489 systemd[1]: Started session-11.scope - Session 11 of User core.
Jun 25 18:48:02.127334 sshd[4109]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:02.135429 systemd[1]: Started sshd@11-10.0.0.141:22-10.0.0.1:52876.service - OpenSSH per-connection server daemon (10.0.0.1:52876).
Jun 25 18:48:02.135781 systemd[1]: sshd@10-10.0.0.141:22-10.0.0.1:52864.service: Deactivated successfully.
Jun 25 18:48:02.138805 systemd[1]: session-11.scope: Deactivated successfully.
Jun 25 18:48:02.139183 systemd-logind[1538]: Session 11 logged out. Waiting for processes to exit.
Jun 25 18:48:02.141640 systemd-logind[1538]: Removed session 11.
Jun 25 18:48:02.167316 sshd[4123]: Accepted publickey for core from 10.0.0.1 port 52876 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:02.168450 sshd[4123]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:02.172396 systemd-logind[1538]: New session 12 of user core.
Jun 25 18:48:02.180423 systemd[1]: Started session-12.scope - Session 12 of User core.
Jun 25 18:48:02.834437 sshd[4123]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:02.841474 systemd[1]: Started sshd@12-10.0.0.141:22-10.0.0.1:52886.service - OpenSSH per-connection server daemon (10.0.0.1:52886).
Jun 25 18:48:02.850811 systemd[1]: sshd@11-10.0.0.141:22-10.0.0.1:52876.service: Deactivated successfully.
Jun 25 18:48:02.855024 systemd[1]: session-12.scope: Deactivated successfully.
Jun 25 18:48:02.856231 systemd-logind[1538]: Session 12 logged out. Waiting for processes to exit.
Jun 25 18:48:02.857357 systemd-logind[1538]: Removed session 12.
Jun 25 18:48:02.883768 sshd[4138]: Accepted publickey for core from 10.0.0.1 port 52886 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:02.884931 sshd[4138]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:02.888934 systemd-logind[1538]: New session 13 of user core.
Jun 25 18:48:02.899674 systemd[1]: Started session-13.scope - Session 13 of User core.
Jun 25 18:48:03.009469 sshd[4138]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:03.012997 systemd[1]: sshd@12-10.0.0.141:22-10.0.0.1:52886.service: Deactivated successfully.
Jun 25 18:48:03.014994 systemd-logind[1538]: Session 13 logged out. Waiting for processes to exit.
Jun 25 18:48:03.015456 systemd[1]: session-13.scope: Deactivated successfully.
Jun 25 18:48:03.018063 systemd-logind[1538]: Removed session 13.
Jun 25 18:48:08.022449 systemd[1]: Started sshd@13-10.0.0.141:22-10.0.0.1:52898.service - OpenSSH per-connection server daemon (10.0.0.1:52898).
Jun 25 18:48:08.053853 sshd[4156]: Accepted publickey for core from 10.0.0.1 port 52898 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:08.054544 sshd[4156]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:08.058030 systemd-logind[1538]: New session 14 of user core.
Jun 25 18:48:08.076513 systemd[1]: Started session-14.scope - Session 14 of User core.
Jun 25 18:48:08.181074 sshd[4156]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:08.188415 systemd[1]: Started sshd@14-10.0.0.141:22-10.0.0.1:52902.service - OpenSSH per-connection server daemon (10.0.0.1:52902).
Jun 25 18:48:08.188798 systemd[1]: sshd@13-10.0.0.141:22-10.0.0.1:52898.service: Deactivated successfully.
Jun 25 18:48:08.192351 systemd[1]: session-14.scope: Deactivated successfully.
Jun 25 18:48:08.193701 systemd-logind[1538]: Session 14 logged out. Waiting for processes to exit.
Jun 25 18:48:08.194692 systemd-logind[1538]: Removed session 14.
Jun 25 18:48:08.218874 sshd[4169]: Accepted publickey for core from 10.0.0.1 port 52902 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:08.220017 sshd[4169]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:08.223472 systemd-logind[1538]: New session 15 of user core.
Jun 25 18:48:08.243473 systemd[1]: Started session-15.scope - Session 15 of User core.
Jun 25 18:48:08.468128 sshd[4169]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:08.480586 systemd[1]: Started sshd@15-10.0.0.141:22-10.0.0.1:52910.service - OpenSSH per-connection server daemon (10.0.0.1:52910).
Jun 25 18:48:08.481484 systemd[1]: sshd@14-10.0.0.141:22-10.0.0.1:52902.service: Deactivated successfully.
Jun 25 18:48:08.483266 systemd[1]: session-15.scope: Deactivated successfully.
Jun 25 18:48:08.484716 systemd-logind[1538]: Session 15 logged out. Waiting for processes to exit.
Jun 25 18:48:08.485556 systemd-logind[1538]: Removed session 15.
Jun 25 18:48:08.515332 sshd[4182]: Accepted publickey for core from 10.0.0.1 port 52910 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:08.516580 sshd[4182]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:08.521460 systemd-logind[1538]: New session 16 of user core.
Jun 25 18:48:08.528504 systemd[1]: Started session-16.scope - Session 16 of User core.
Jun 25 18:48:09.263049 sshd[4182]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:09.273239 systemd[1]: Started sshd@16-10.0.0.141:22-10.0.0.1:46702.service - OpenSSH per-connection server daemon (10.0.0.1:46702).
Jun 25 18:48:09.273647 systemd[1]: sshd@15-10.0.0.141:22-10.0.0.1:52910.service: Deactivated successfully.
Jun 25 18:48:09.278872 systemd-logind[1538]: Session 16 logged out. Waiting for processes to exit.
Jun 25 18:48:09.279026 systemd[1]: session-16.scope: Deactivated successfully.
Jun 25 18:48:09.288927 systemd-logind[1538]: Removed session 16.
Jun 25 18:48:09.315785 sshd[4202]: Accepted publickey for core from 10.0.0.1 port 46702 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:09.317062 sshd[4202]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:09.321286 systemd-logind[1538]: New session 17 of user core.
Jun 25 18:48:09.325475 systemd[1]: Started session-17.scope - Session 17 of User core.
Jun 25 18:48:09.602442 sshd[4202]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:09.614466 systemd[1]: Started sshd@17-10.0.0.141:22-10.0.0.1:46704.service - OpenSSH per-connection server daemon (10.0.0.1:46704).
Jun 25 18:48:09.614846 systemd[1]: sshd@16-10.0.0.141:22-10.0.0.1:46702.service: Deactivated successfully.
Jun 25 18:48:09.617187 systemd-logind[1538]: Session 17 logged out. Waiting for processes to exit.
Jun 25 18:48:09.617627 systemd[1]: session-17.scope: Deactivated successfully.
Jun 25 18:48:09.619605 systemd-logind[1538]: Removed session 17.
Jun 25 18:48:09.650818 sshd[4217]: Accepted publickey for core from 10.0.0.1 port 46704 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:09.652002 sshd[4217]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:09.655993 systemd-logind[1538]: New session 18 of user core.
Jun 25 18:48:09.662508 systemd[1]: Started session-18.scope - Session 18 of User core.
Jun 25 18:48:09.773785 sshd[4217]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:09.777298 systemd[1]: sshd@17-10.0.0.141:22-10.0.0.1:46704.service: Deactivated successfully.
Jun 25 18:48:09.779340 systemd-logind[1538]: Session 18 logged out. Waiting for processes to exit.
Jun 25 18:48:09.779960 systemd[1]: session-18.scope: Deactivated successfully.
Jun 25 18:48:09.781784 systemd-logind[1538]: Removed session 18.
Jun 25 18:48:14.784442 systemd[1]: Started sshd@18-10.0.0.141:22-10.0.0.1:46720.service - OpenSSH per-connection server daemon (10.0.0.1:46720).
Jun 25 18:48:14.817876 sshd[4240]: Accepted publickey for core from 10.0.0.1 port 46720 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:14.819067 sshd[4240]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:14.822821 systemd-logind[1538]: New session 19 of user core.
Jun 25 18:48:14.832477 systemd[1]: Started session-19.scope - Session 19 of User core.
Jun 25 18:48:14.939724 sshd[4240]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:14.944432 systemd-logind[1538]: Session 19 logged out. Waiting for processes to exit.
Jun 25 18:48:14.945627 systemd[1]: sshd@18-10.0.0.141:22-10.0.0.1:46720.service: Deactivated successfully.
Jun 25 18:48:14.947204 systemd[1]: session-19.scope: Deactivated successfully.
Jun 25 18:48:14.948861 systemd-logind[1538]: Removed session 19.
Jun 25 18:48:19.951476 systemd[1]: Started sshd@19-10.0.0.141:22-10.0.0.1:52848.service - OpenSSH per-connection server daemon (10.0.0.1:52848).
Jun 25 18:48:19.991665 sshd[4256]: Accepted publickey for core from 10.0.0.1 port 52848 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:19.993437 sshd[4256]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:19.998157 systemd-logind[1538]: New session 20 of user core.
Jun 25 18:48:20.004506 systemd[1]: Started session-20.scope - Session 20 of User core.
Jun 25 18:48:20.110598 sshd[4256]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:20.113764 systemd[1]: sshd@19-10.0.0.141:22-10.0.0.1:52848.service: Deactivated successfully.
Jun 25 18:48:20.115952 systemd-logind[1538]: Session 20 logged out. Waiting for processes to exit.
Jun 25 18:48:20.116138 systemd[1]: session-20.scope: Deactivated successfully.
Jun 25 18:48:20.118001 systemd-logind[1538]: Removed session 20.
Jun 25 18:48:22.486775 kubelet[2664]: E0625 18:48:22.486529    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:25.120550 systemd[1]: Started sshd@20-10.0.0.141:22-10.0.0.1:52850.service - OpenSSH per-connection server daemon (10.0.0.1:52850).
Jun 25 18:48:25.153759 sshd[4271]: Accepted publickey for core from 10.0.0.1 port 52850 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:25.154164 sshd[4271]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:25.159017 systemd-logind[1538]: New session 21 of user core.
Jun 25 18:48:25.168528 systemd[1]: Started session-21.scope - Session 21 of User core.
Jun 25 18:48:25.294137 sshd[4271]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:25.305404 systemd[1]: Started sshd@21-10.0.0.141:22-10.0.0.1:52860.service - OpenSSH per-connection server daemon (10.0.0.1:52860).
Jun 25 18:48:25.305827 systemd[1]: sshd@20-10.0.0.141:22-10.0.0.1:52850.service: Deactivated successfully.
Jun 25 18:48:25.307388 systemd[1]: session-21.scope: Deactivated successfully.
Jun 25 18:48:25.309440 systemd-logind[1538]: Session 21 logged out. Waiting for processes to exit.
Jun 25 18:48:25.311656 systemd-logind[1538]: Removed session 21.
Jun 25 18:48:25.343590 sshd[4283]: Accepted publickey for core from 10.0.0.1 port 52860 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:25.344781 sshd[4283]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:25.349336 systemd-logind[1538]: New session 22 of user core.
Jun 25 18:48:25.360437 systemd[1]: Started session-22.scope - Session 22 of User core.
Jun 25 18:48:28.399585 containerd[1563]: time="2024-06-25T18:48:28.399540889Z" level=info msg="StopContainer for \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\" with timeout 30 (s)"
Jun 25 18:48:28.401347 containerd[1563]: time="2024-06-25T18:48:28.401319388Z" level=info msg="Stop container \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\" with signal terminated"
Jun 25 18:48:28.431425 containerd[1563]: time="2024-06-25T18:48:28.431381429Z" level=error msg="failed to reload cni configuration after receiving fs change event(REMOVE        \"/etc/cni/net.d/05-cilium.conf\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config"
Jun 25 18:48:28.437780 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403-rootfs.mount: Deactivated successfully.
Jun 25 18:48:28.438901 containerd[1563]: time="2024-06-25T18:48:28.438872330Z" level=info msg="StopContainer for \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\" with timeout 2 (s)"
Jun 25 18:48:28.439160 containerd[1563]: time="2024-06-25T18:48:28.439138361Z" level=info msg="Stop container \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\" with signal terminated"
Jun 25 18:48:28.442915 containerd[1563]: time="2024-06-25T18:48:28.442863792Z" level=info msg="shim disconnected" id=e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403 namespace=k8s.io
Jun 25 18:48:28.442915 containerd[1563]: time="2024-06-25T18:48:28.442913590Z" level=warning msg="cleaning up after shim disconnected" id=e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403 namespace=k8s.io
Jun 25 18:48:28.443115 containerd[1563]: time="2024-06-25T18:48:28.442924510Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:48:28.444993 systemd-networkd[1237]: lxc_health: Link DOWN
Jun 25 18:48:28.445084 systemd-networkd[1237]: lxc_health: Lost carrier
Jun 25 18:48:28.458875 containerd[1563]: time="2024-06-25T18:48:28.458832240Z" level=info msg="StopContainer for \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\" returns successfully"
Jun 25 18:48:28.460449 containerd[1563]: time="2024-06-25T18:48:28.460417625Z" level=info msg="StopPodSandbox for \"136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be\""
Jun 25 18:48:28.460514 containerd[1563]: time="2024-06-25T18:48:28.460464224Z" level=info msg="Container to stop \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jun 25 18:48:28.464324 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be-shm.mount: Deactivated successfully.
Jun 25 18:48:28.487179 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be-rootfs.mount: Deactivated successfully.
Jun 25 18:48:28.494015 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c-rootfs.mount: Deactivated successfully.
Jun 25 18:48:28.495993 containerd[1563]: time="2024-06-25T18:48:28.495177424Z" level=info msg="shim disconnected" id=136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be namespace=k8s.io
Jun 25 18:48:28.495993 containerd[1563]: time="2024-06-25T18:48:28.495995876Z" level=warning msg="cleaning up after shim disconnected" id=136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be namespace=k8s.io
Jun 25 18:48:28.496130 containerd[1563]: time="2024-06-25T18:48:28.496012595Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:48:28.496130 containerd[1563]: time="2024-06-25T18:48:28.495246022Z" level=info msg="shim disconnected" id=ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c namespace=k8s.io
Jun 25 18:48:28.496130 containerd[1563]: time="2024-06-25T18:48:28.496062833Z" level=warning msg="cleaning up after shim disconnected" id=ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c namespace=k8s.io
Jun 25 18:48:28.496130 containerd[1563]: time="2024-06-25T18:48:28.496074953Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:48:28.508835 containerd[1563]: time="2024-06-25T18:48:28.508764714Z" level=info msg="TearDown network for sandbox \"136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be\" successfully"
Jun 25 18:48:28.508835 containerd[1563]: time="2024-06-25T18:48:28.508833072Z" level=info msg="StopPodSandbox for \"136d46dc16eb3e16fa8fecb50bb111f7ee19a055aa3ab9811447c0b6a9b3b2be\" returns successfully"
Jun 25 18:48:28.509479 containerd[1563]: time="2024-06-25T18:48:28.509343614Z" level=info msg="StopContainer for \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\" returns successfully"
Jun 25 18:48:28.509860 containerd[1563]: time="2024-06-25T18:48:28.509791999Z" level=info msg="StopPodSandbox for \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\""
Jun 25 18:48:28.509860 containerd[1563]: time="2024-06-25T18:48:28.509827038Z" level=info msg="Container to stop \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jun 25 18:48:28.509940 containerd[1563]: time="2024-06-25T18:48:28.509863356Z" level=info msg="Container to stop \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jun 25 18:48:28.509940 containerd[1563]: time="2024-06-25T18:48:28.509873556Z" level=info msg="Container to stop \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jun 25 18:48:28.509940 containerd[1563]: time="2024-06-25T18:48:28.509883196Z" level=info msg="Container to stop \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jun 25 18:48:28.509940 containerd[1563]: time="2024-06-25T18:48:28.509892555Z" level=info msg="Container to stop \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\" must be in running or unknown state, current state \"CONTAINER_EXITED\""
Jun 25 18:48:28.541045 containerd[1563]: time="2024-06-25T18:48:28.540808447Z" level=info msg="shim disconnected" id=94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18 namespace=k8s.io
Jun 25 18:48:28.541045 containerd[1563]: time="2024-06-25T18:48:28.540870605Z" level=warning msg="cleaning up after shim disconnected" id=94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18 namespace=k8s.io
Jun 25 18:48:28.541045 containerd[1563]: time="2024-06-25T18:48:28.540881324Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:48:28.561325 containerd[1563]: time="2024-06-25T18:48:28.560355811Z" level=warning msg="cleanup warnings time=\"2024-06-25T18:48:28Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io
Jun 25 18:48:28.567252 containerd[1563]: time="2024-06-25T18:48:28.563399746Z" level=info msg="TearDown network for sandbox \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" successfully"
Jun 25 18:48:28.567252 containerd[1563]: time="2024-06-25T18:48:28.563429625Z" level=info msg="StopPodSandbox for \"94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18\" returns successfully"
Jun 25 18:48:28.592077 kubelet[2664]: I0625 18:48:28.591832    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-host-proc-sys-kernel\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.592077 kubelet[2664]: I0625 18:48:28.592026    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-bpf-maps\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.592538 kubelet[2664]: I0625 18:48:28.592048    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.593824 kubelet[2664]: I0625 18:48:28.592145    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s2g2\" (UniqueName: \"kubernetes.io/projected/71aed90d-6dae-4911-b0e6-4e9089070e4e-kube-api-access-4s2g2\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.593824 kubelet[2664]: I0625 18:48:28.592628    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/71aed90d-6dae-4911-b0e6-4e9089070e4e-hubble-tls\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.593824 kubelet[2664]: I0625 18:48:28.592661    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-hostproc\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.593824 kubelet[2664]: I0625 18:48:28.592687    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-config-path\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.593824 kubelet[2664]: I0625 18:48:28.592704    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cni-path\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.593824 kubelet[2664]: I0625 18:48:28.592723    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbtwd\" (UniqueName: \"kubernetes.io/projected/fd86cd23-88a9-489b-ad27-9ea7130e570c-kube-api-access-xbtwd\") pod \"fd86cd23-88a9-489b-ad27-9ea7130e570c\" (UID: \"fd86cd23-88a9-489b-ad27-9ea7130e570c\") "
Jun 25 18:48:28.594075 kubelet[2664]: I0625 18:48:28.592745    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-hostproc" (OuterVolumeSpecName: "hostproc") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.594075 kubelet[2664]: I0625 18:48:28.592755    2664 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-bpf-maps\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.594075 kubelet[2664]: I0625 18:48:28.592158    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.594075 kubelet[2664]: I0625 18:48:28.593161    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cni-path" (OuterVolumeSpecName: "cni-path") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.595104 kubelet[2664]: I0625 18:48:28.595070    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71aed90d-6dae-4911-b0e6-4e9089070e4e-kube-api-access-4s2g2" (OuterVolumeSpecName: "kube-api-access-4s2g2") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "kube-api-access-4s2g2". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jun 25 18:48:28.595339 kubelet[2664]: I0625 18:48:28.595302    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71aed90d-6dae-4911-b0e6-4e9089070e4e-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jun 25 18:48:28.595571 kubelet[2664]: I0625 18:48:28.595549    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jun 25 18:48:28.596276 kubelet[2664]: I0625 18:48:28.596254    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd86cd23-88a9-489b-ad27-9ea7130e570c-kube-api-access-xbtwd" (OuterVolumeSpecName: "kube-api-access-xbtwd") pod "fd86cd23-88a9-489b-ad27-9ea7130e570c" (UID: "fd86cd23-88a9-489b-ad27-9ea7130e570c"). InnerVolumeSpecName "kube-api-access-xbtwd". PluginName "kubernetes.io/projected", VolumeGidValue ""
Jun 25 18:48:28.693733 kubelet[2664]: I0625 18:48:28.693624    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/71aed90d-6dae-4911-b0e6-4e9089070e4e-clustermesh-secrets\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.693733 kubelet[2664]: I0625 18:48:28.693662    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-etc-cni-netd\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.693733 kubelet[2664]: I0625 18:48:28.693679    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-lib-modules\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.693733 kubelet[2664]: I0625 18:48:28.693701    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-cgroup\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.693733 kubelet[2664]: I0625 18:48:28.693720    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-run\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.693733 kubelet[2664]: I0625 18:48:28.693742    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/fd86cd23-88a9-489b-ad27-9ea7130e570c-cilium-config-path\") pod \"fd86cd23-88a9-489b-ad27-9ea7130e570c\" (UID: \"fd86cd23-88a9-489b-ad27-9ea7130e570c\") "
Jun 25 18:48:28.694422 kubelet[2664]: I0625 18:48:28.693762    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-host-proc-sys-net\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.694422 kubelet[2664]: I0625 18:48:28.693779    2664 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-xtables-lock\") pod \"71aed90d-6dae-4911-b0e6-4e9089070e4e\" (UID: \"71aed90d-6dae-4911-b0e6-4e9089070e4e\") "
Jun 25 18:48:28.694422 kubelet[2664]: I0625 18:48:28.693818    2664 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-host-proc-sys-kernel\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.694422 kubelet[2664]: I0625 18:48:28.693817    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.694422 kubelet[2664]: I0625 18:48:28.693831    2664 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-4s2g2\" (UniqueName: \"kubernetes.io/projected/71aed90d-6dae-4911-b0e6-4e9089070e4e-kube-api-access-4s2g2\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.694422 kubelet[2664]: I0625 18:48:28.693852    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.694597 kubelet[2664]: I0625 18:48:28.693867    2664 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/71aed90d-6dae-4911-b0e6-4e9089070e4e-hubble-tls\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.694597 kubelet[2664]: I0625 18:48:28.693873    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.694597 kubelet[2664]: I0625 18:48:28.693881    2664 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-hostproc\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.694597 kubelet[2664]: I0625 18:48:28.693892    2664 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cni-path\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.694597 kubelet[2664]: I0625 18:48:28.693904    2664 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-xbtwd\" (UniqueName: \"kubernetes.io/projected/fd86cd23-88a9-489b-ad27-9ea7130e570c-kube-api-access-xbtwd\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.694597 kubelet[2664]: I0625 18:48:28.693914    2664 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-config-path\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.694821 kubelet[2664]: I0625 18:48:28.694753    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.694821 kubelet[2664]: I0625 18:48:28.694789    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.694821 kubelet[2664]: I0625 18:48:28.694805    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue ""
Jun 25 18:48:28.695718 kubelet[2664]: I0625 18:48:28.695689    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd86cd23-88a9-489b-ad27-9ea7130e570c-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "fd86cd23-88a9-489b-ad27-9ea7130e570c" (UID: "fd86cd23-88a9-489b-ad27-9ea7130e570c"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue ""
Jun 25 18:48:28.696841 kubelet[2664]: I0625 18:48:28.696815    2664 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71aed90d-6dae-4911-b0e6-4e9089070e4e-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "71aed90d-6dae-4911-b0e6-4e9089070e4e" (UID: "71aed90d-6dae-4911-b0e6-4e9089070e4e"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue ""
Jun 25 18:48:28.707077 kubelet[2664]: I0625 18:48:28.707055    2664 scope.go:117] "RemoveContainer" containerID="e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403"
Jun 25 18:48:28.710154 containerd[1563]: time="2024-06-25T18:48:28.710108475Z" level=info msg="RemoveContainer for \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\""
Jun 25 18:48:28.718461 containerd[1563]: time="2024-06-25T18:48:28.718417828Z" level=info msg="RemoveContainer for \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\" returns successfully"
Jun 25 18:48:28.718690 kubelet[2664]: I0625 18:48:28.718652    2664 scope.go:117] "RemoveContainer" containerID="e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403"
Jun 25 18:48:28.723682 containerd[1563]: time="2024-06-25T18:48:28.718906091Z" level=error msg="ContainerStatus for \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\": not found"
Jun 25 18:48:28.729628 kubelet[2664]: E0625 18:48:28.729581    2664 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\": not found" containerID="e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403"
Jun 25 18:48:28.729691 kubelet[2664]: I0625 18:48:28.729673    2664 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403"} err="failed to get container status \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\": rpc error: code = NotFound desc = an error occurred when try to find container \"e3e2fbe55cc640b143df42eef2870ef9fcba5d92f6c04b2fff5f9cd6d21b0403\": not found"
Jun 25 18:48:28.729691 kubelet[2664]: I0625 18:48:28.729687    2664 scope.go:117] "RemoveContainer" containerID="ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c"
Jun 25 18:48:28.730611 containerd[1563]: time="2024-06-25T18:48:28.730586168Z" level=info msg="RemoveContainer for \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\""
Jun 25 18:48:28.732970 containerd[1563]: time="2024-06-25T18:48:28.732944206Z" level=info msg="RemoveContainer for \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\" returns successfully"
Jun 25 18:48:28.733095 kubelet[2664]: I0625 18:48:28.733078    2664 scope.go:117] "RemoveContainer" containerID="3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b"
Jun 25 18:48:28.733792 containerd[1563]: time="2024-06-25T18:48:28.733772138Z" level=info msg="RemoveContainer for \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\""
Jun 25 18:48:28.735856 containerd[1563]: time="2024-06-25T18:48:28.735827787Z" level=info msg="RemoveContainer for \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\" returns successfully"
Jun 25 18:48:28.736010 kubelet[2664]: I0625 18:48:28.735979    2664 scope.go:117] "RemoveContainer" containerID="063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323"
Jun 25 18:48:28.736920 containerd[1563]: time="2024-06-25T18:48:28.736879910Z" level=info msg="RemoveContainer for \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\""
Jun 25 18:48:28.738922 containerd[1563]: time="2024-06-25T18:48:28.738896960Z" level=info msg="RemoveContainer for \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\" returns successfully"
Jun 25 18:48:28.739038 kubelet[2664]: I0625 18:48:28.739022    2664 scope.go:117] "RemoveContainer" containerID="70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14"
Jun 25 18:48:28.739830 containerd[1563]: time="2024-06-25T18:48:28.739805809Z" level=info msg="RemoveContainer for \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\""
Jun 25 18:48:28.741796 containerd[1563]: time="2024-06-25T18:48:28.741766581Z" level=info msg="RemoveContainer for \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\" returns successfully"
Jun 25 18:48:28.741949 kubelet[2664]: I0625 18:48:28.741926    2664 scope.go:117] "RemoveContainer" containerID="fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1"
Jun 25 18:48:28.742741 containerd[1563]: time="2024-06-25T18:48:28.742719828Z" level=info msg="RemoveContainer for \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\""
Jun 25 18:48:28.744811 containerd[1563]: time="2024-06-25T18:48:28.744787797Z" level=info msg="RemoveContainer for \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\" returns successfully"
Jun 25 18:48:28.744924 kubelet[2664]: I0625 18:48:28.744905    2664 scope.go:117] "RemoveContainer" containerID="ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c"
Jun 25 18:48:28.745074 containerd[1563]: time="2024-06-25T18:48:28.745044668Z" level=error msg="ContainerStatus for \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\": not found"
Jun 25 18:48:28.745176 kubelet[2664]: E0625 18:48:28.745163    2664 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\": not found" containerID="ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c"
Jun 25 18:48:28.745212 kubelet[2664]: I0625 18:48:28.745190    2664 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c"} err="failed to get container status \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\": rpc error: code = NotFound desc = an error occurred when try to find container \"ba33ad9956a296c55fe1bcd92b757a3bec98bd4723d48595374f1a882114011c\": not found"
Jun 25 18:48:28.745212 kubelet[2664]: I0625 18:48:28.745202    2664 scope.go:117] "RemoveContainer" containerID="3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b"
Jun 25 18:48:28.745348 containerd[1563]: time="2024-06-25T18:48:28.745325738Z" level=error msg="ContainerStatus for \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\": not found"
Jun 25 18:48:28.745426 kubelet[2664]: E0625 18:48:28.745413    2664 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\": not found" containerID="3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b"
Jun 25 18:48:28.745461 kubelet[2664]: I0625 18:48:28.745434    2664 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b"} err="failed to get container status \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\": rpc error: code = NotFound desc = an error occurred when try to find container \"3e100300f293ff7dd7a27dfd76333f2985aa1ecdd209e93b2863464f21814a8b\": not found"
Jun 25 18:48:28.745461 kubelet[2664]: I0625 18:48:28.745458    2664 scope.go:117] "RemoveContainer" containerID="063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323"
Jun 25 18:48:28.745615 containerd[1563]: time="2024-06-25T18:48:28.745586809Z" level=error msg="ContainerStatus for \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\": not found"
Jun 25 18:48:28.745709 kubelet[2664]: E0625 18:48:28.745695    2664 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\": not found" containerID="063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323"
Jun 25 18:48:28.745741 kubelet[2664]: I0625 18:48:28.745721    2664 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323"} err="failed to get container status \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\": rpc error: code = NotFound desc = an error occurred when try to find container \"063a115df28b7ddfac18041f986c4bec2170e5bcbfc6bf15bf6ab58e55f20323\": not found"
Jun 25 18:48:28.745741 kubelet[2664]: I0625 18:48:28.745730    2664 scope.go:117] "RemoveContainer" containerID="70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14"
Jun 25 18:48:28.745856 containerd[1563]: time="2024-06-25T18:48:28.745834601Z" level=error msg="ContainerStatus for \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\": not found"
Jun 25 18:48:28.745937 kubelet[2664]: E0625 18:48:28.745923    2664 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\": not found" containerID="70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14"
Jun 25 18:48:28.745969 kubelet[2664]: I0625 18:48:28.745947    2664 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14"} err="failed to get container status \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\": rpc error: code = NotFound desc = an error occurred when try to find container \"70d53f0d5cc1b4b9e1ac5b2469c8039f3e3fa3abb382552fe006ee59e85acc14\": not found"
Jun 25 18:48:28.745969 kubelet[2664]: I0625 18:48:28.745957    2664 scope.go:117] "RemoveContainer" containerID="fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1"
Jun 25 18:48:28.746094 containerd[1563]: time="2024-06-25T18:48:28.746072632Z" level=error msg="ContainerStatus for \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\": not found"
Jun 25 18:48:28.746164 kubelet[2664]: E0625 18:48:28.746151    2664 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\": not found" containerID="fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1"
Jun 25 18:48:28.746197 kubelet[2664]: I0625 18:48:28.746175    2664 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1"} err="failed to get container status \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\": rpc error: code = NotFound desc = an error occurred when try to find container \"fbecf721f70605d0267b0857f162fce86bf419c388c98b287aec9e7e48ad2ad1\": not found"
Jun 25 18:48:28.794514 kubelet[2664]: I0625 18:48:28.794476    2664 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/71aed90d-6dae-4911-b0e6-4e9089070e4e-clustermesh-secrets\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.794514 kubelet[2664]: I0625 18:48:28.794507    2664 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-etc-cni-netd\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.794627 kubelet[2664]: I0625 18:48:28.794525    2664 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-lib-modules\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.794627 kubelet[2664]: I0625 18:48:28.794544    2664 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/fd86cd23-88a9-489b-ad27-9ea7130e570c-cilium-config-path\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.794627 kubelet[2664]: I0625 18:48:28.794562    2664 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-cgroup\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.794627 kubelet[2664]: I0625 18:48:28.794578    2664 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-cilium-run\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.794627 kubelet[2664]: I0625 18:48:28.794595    2664 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-host-proc-sys-net\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:28.794627 kubelet[2664]: I0625 18:48:28.794611    2664 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/71aed90d-6dae-4911-b0e6-4e9089070e4e-xtables-lock\") on node \"localhost\" DevicePath \"\""
Jun 25 18:48:29.411805 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18-rootfs.mount: Deactivated successfully.
Jun 25 18:48:29.411952 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-94171d483e8e1aa0d79caf351901a5b38f75a04594e186d2ea73e3acdc8ada18-shm.mount: Deactivated successfully.
Jun 25 18:48:29.412061 systemd[1]: var-lib-kubelet-pods-fd86cd23\x2d88a9\x2d489b\x2dad27\x2d9ea7130e570c-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dxbtwd.mount: Deactivated successfully.
Jun 25 18:48:29.412142 systemd[1]: var-lib-kubelet-pods-71aed90d\x2d6dae\x2d4911\x2db0e6\x2d4e9089070e4e-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d4s2g2.mount: Deactivated successfully.
Jun 25 18:48:29.412236 systemd[1]: var-lib-kubelet-pods-71aed90d\x2d6dae\x2d4911\x2db0e6\x2d4e9089070e4e-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully.
Jun 25 18:48:29.412322 systemd[1]: var-lib-kubelet-pods-71aed90d\x2d6dae\x2d4911\x2db0e6\x2d4e9089070e4e-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully.
Jun 25 18:48:29.562482 kubelet[2664]: E0625 18:48:29.562457    2664 kubelet.go:2855] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"
Jun 25 18:48:30.355924 sshd[4283]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:30.362447 systemd[1]: Started sshd@22-10.0.0.141:22-10.0.0.1:36324.service - OpenSSH per-connection server daemon (10.0.0.1:36324).
Jun 25 18:48:30.362804 systemd[1]: sshd@21-10.0.0.141:22-10.0.0.1:52860.service: Deactivated successfully.
Jun 25 18:48:30.364622 systemd-logind[1538]: Session 22 logged out. Waiting for processes to exit.
Jun 25 18:48:30.365262 systemd[1]: session-22.scope: Deactivated successfully.
Jun 25 18:48:30.366610 systemd-logind[1538]: Removed session 22.
Jun 25 18:48:30.395084 sshd[4450]: Accepted publickey for core from 10.0.0.1 port 36324 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:30.396256 sshd[4450]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:30.400074 systemd-logind[1538]: New session 23 of user core.
Jun 25 18:48:30.410432 systemd[1]: Started session-23.scope - Session 23 of User core.
Jun 25 18:48:30.488345 kubelet[2664]: I0625 18:48:30.488314    2664 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="71aed90d-6dae-4911-b0e6-4e9089070e4e" path="/var/lib/kubelet/pods/71aed90d-6dae-4911-b0e6-4e9089070e4e/volumes"
Jun 25 18:48:30.488903 kubelet[2664]: I0625 18:48:30.488883    2664 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="fd86cd23-88a9-489b-ad27-9ea7130e570c" path="/var/lib/kubelet/pods/fd86cd23-88a9-489b-ad27-9ea7130e570c/volumes"
Jun 25 18:48:31.367146 sshd[4450]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:31.378762 systemd[1]: Started sshd@23-10.0.0.141:22-10.0.0.1:36326.service - OpenSSH per-connection server daemon (10.0.0.1:36326).
Jun 25 18:48:31.383639 kubelet[2664]: I0625 18:48:31.383341    2664 topology_manager.go:215] "Topology Admit Handler" podUID="382bb7b1-3db7-4ca4-af80-e226104a2f75" podNamespace="kube-system" podName="cilium-b55b2"
Jun 25 18:48:31.383639 kubelet[2664]: E0625 18:48:31.383395    2664 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="fd86cd23-88a9-489b-ad27-9ea7130e570c" containerName="cilium-operator"
Jun 25 18:48:31.383639 kubelet[2664]: E0625 18:48:31.383404    2664 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="71aed90d-6dae-4911-b0e6-4e9089070e4e" containerName="mount-bpf-fs"
Jun 25 18:48:31.383639 kubelet[2664]: E0625 18:48:31.383410    2664 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="71aed90d-6dae-4911-b0e6-4e9089070e4e" containerName="clean-cilium-state"
Jun 25 18:48:31.383639 kubelet[2664]: E0625 18:48:31.383417    2664 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="71aed90d-6dae-4911-b0e6-4e9089070e4e" containerName="cilium-agent"
Jun 25 18:48:31.383639 kubelet[2664]: E0625 18:48:31.383423    2664 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="71aed90d-6dae-4911-b0e6-4e9089070e4e" containerName="apply-sysctl-overwrites"
Jun 25 18:48:31.383639 kubelet[2664]: E0625 18:48:31.383432    2664 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="71aed90d-6dae-4911-b0e6-4e9089070e4e" containerName="mount-cgroup"
Jun 25 18:48:31.383639 kubelet[2664]: I0625 18:48:31.383454    2664 memory_manager.go:346] "RemoveStaleState removing state" podUID="fd86cd23-88a9-489b-ad27-9ea7130e570c" containerName="cilium-operator"
Jun 25 18:48:31.383639 kubelet[2664]: I0625 18:48:31.383460    2664 memory_manager.go:346] "RemoveStaleState removing state" podUID="71aed90d-6dae-4911-b0e6-4e9089070e4e" containerName="cilium-agent"
Jun 25 18:48:31.379227 systemd[1]: sshd@22-10.0.0.141:22-10.0.0.1:36324.service: Deactivated successfully.
Jun 25 18:48:31.389133 systemd[1]: session-23.scope: Deactivated successfully.
Jun 25 18:48:31.393094 systemd-logind[1538]: Session 23 logged out. Waiting for processes to exit.
Jun 25 18:48:31.395809 systemd-logind[1538]: Removed session 23.
Jun 25 18:48:31.405963 kubelet[2664]: I0625 18:48:31.405918    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-cni-path\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.405963 kubelet[2664]: I0625 18:48:31.405962    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-etc-cni-netd\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406096 kubelet[2664]: I0625 18:48:31.406024    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-host-proc-sys-kernel\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406096 kubelet[2664]: I0625 18:48:31.406075    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-hostproc\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406155 kubelet[2664]: I0625 18:48:31.406098    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/382bb7b1-3db7-4ca4-af80-e226104a2f75-cilium-config-path\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406155 kubelet[2664]: I0625 18:48:31.406123    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/382bb7b1-3db7-4ca4-af80-e226104a2f75-cilium-ipsec-secrets\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406155 kubelet[2664]: I0625 18:48:31.406144    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/382bb7b1-3db7-4ca4-af80-e226104a2f75-hubble-tls\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406243 kubelet[2664]: I0625 18:48:31.406174    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-cilium-run\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406243 kubelet[2664]: I0625 18:48:31.406195    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-bpf-maps\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406243 kubelet[2664]: I0625 18:48:31.406235    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/382bb7b1-3db7-4ca4-af80-e226104a2f75-clustermesh-secrets\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406311 kubelet[2664]: I0625 18:48:31.406257    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wg82z\" (UniqueName: \"kubernetes.io/projected/382bb7b1-3db7-4ca4-af80-e226104a2f75-kube-api-access-wg82z\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406311 kubelet[2664]: I0625 18:48:31.406276    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-xtables-lock\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406311 kubelet[2664]: I0625 18:48:31.406295    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-lib-modules\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406375 kubelet[2664]: I0625 18:48:31.406316    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-cilium-cgroup\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.406375 kubelet[2664]: I0625 18:48:31.406334    2664 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/382bb7b1-3db7-4ca4-af80-e226104a2f75-host-proc-sys-net\") pod \"cilium-b55b2\" (UID: \"382bb7b1-3db7-4ca4-af80-e226104a2f75\") " pod="kube-system/cilium-b55b2"
Jun 25 18:48:31.410183 sshd[4464]: Accepted publickey for core from 10.0.0.1 port 36326 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:31.411642 sshd[4464]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:31.420288 systemd-logind[1538]: New session 24 of user core.
Jun 25 18:48:31.426500 systemd[1]: Started session-24.scope - Session 24 of User core.
Jun 25 18:48:31.476157 sshd[4464]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:31.489489 systemd[1]: Started sshd@24-10.0.0.141:22-10.0.0.1:36338.service - OpenSSH per-connection server daemon (10.0.0.1:36338).
Jun 25 18:48:31.489858 systemd[1]: sshd@23-10.0.0.141:22-10.0.0.1:36326.service: Deactivated successfully.
Jun 25 18:48:31.493017 systemd-logind[1538]: Session 24 logged out. Waiting for processes to exit.
Jun 25 18:48:31.493708 systemd[1]: session-24.scope: Deactivated successfully.
Jun 25 18:48:31.495715 systemd-logind[1538]: Removed session 24.
Jun 25 18:48:31.521278 sshd[4473]: Accepted publickey for core from 10.0.0.1 port 36338 ssh2: RSA SHA256:PTHQXr0iRYYg3MbKKJZ6aC6iEkqmHU1AdffEoJcWF3A
Jun 25 18:48:31.522169 sshd[4473]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0)
Jun 25 18:48:31.530267 systemd-logind[1538]: New session 25 of user core.
Jun 25 18:48:31.535475 systemd[1]: Started session-25.scope - Session 25 of User core.
Jun 25 18:48:31.693762 kubelet[2664]: E0625 18:48:31.693636    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:31.694131 containerd[1563]: time="2024-06-25T18:48:31.694080901Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-b55b2,Uid:382bb7b1-3db7-4ca4-af80-e226104a2f75,Namespace:kube-system,Attempt:0,}"
Jun 25 18:48:31.712840 containerd[1563]: time="2024-06-25T18:48:31.712751704Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jun 25 18:48:31.712840 containerd[1563]: time="2024-06-25T18:48:31.712807142Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:48:31.712840 containerd[1563]: time="2024-06-25T18:48:31.712831342Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jun 25 18:48:31.712991 containerd[1563]: time="2024-06-25T18:48:31.712847781Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jun 25 18:48:31.751509 containerd[1563]: time="2024-06-25T18:48:31.751401673Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-b55b2,Uid:382bb7b1-3db7-4ca4-af80-e226104a2f75,Namespace:kube-system,Attempt:0,} returns sandbox id \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\""
Jun 25 18:48:31.752109 kubelet[2664]: E0625 18:48:31.752088    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:31.756872 containerd[1563]: time="2024-06-25T18:48:31.756816443Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}"
Jun 25 18:48:31.776055 containerd[1563]: time="2024-06-25T18:48:31.776004312Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"16c1f9077e81f9e5c4c7b20ef7d26314b82c9973ebd11999702d4c81f0eab2f9\""
Jun 25 18:48:31.777355 containerd[1563]: time="2024-06-25T18:48:31.776526577Z" level=info msg="StartContainer for \"16c1f9077e81f9e5c4c7b20ef7d26314b82c9973ebd11999702d4c81f0eab2f9\""
Jun 25 18:48:31.825598 containerd[1563]: time="2024-06-25T18:48:31.825557379Z" level=info msg="StartContainer for \"16c1f9077e81f9e5c4c7b20ef7d26314b82c9973ebd11999702d4c81f0eab2f9\" returns successfully"
Jun 25 18:48:31.870173 containerd[1563]: time="2024-06-25T18:48:31.870117624Z" level=info msg="shim disconnected" id=16c1f9077e81f9e5c4c7b20ef7d26314b82c9973ebd11999702d4c81f0eab2f9 namespace=k8s.io
Jun 25 18:48:31.870389 containerd[1563]: time="2024-06-25T18:48:31.870171263Z" level=warning msg="cleaning up after shim disconnected" id=16c1f9077e81f9e5c4c7b20ef7d26314b82c9973ebd11999702d4c81f0eab2f9 namespace=k8s.io
Jun 25 18:48:31.870389 containerd[1563]: time="2024-06-25T18:48:31.870193102Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:48:32.722179 kubelet[2664]: E0625 18:48:32.722131    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:32.725021 containerd[1563]: time="2024-06-25T18:48:32.724985895Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}"
Jun 25 18:48:32.743496 containerd[1563]: time="2024-06-25T18:48:32.743349746Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"bdf5288c124ddcfed65c0aeaff6a5f93a43e75f13ea1ca09e0355be537b30f56\""
Jun 25 18:48:32.744033 containerd[1563]: time="2024-06-25T18:48:32.743840254Z" level=info msg="StartContainer for \"bdf5288c124ddcfed65c0aeaff6a5f93a43e75f13ea1ca09e0355be537b30f56\""
Jun 25 18:48:32.765104 systemd[1]: run-containerd-runc-k8s.io-bdf5288c124ddcfed65c0aeaff6a5f93a43e75f13ea1ca09e0355be537b30f56-runc.A64Mlw.mount: Deactivated successfully.
Jun 25 18:48:32.787472 containerd[1563]: time="2024-06-25T18:48:32.787325302Z" level=info msg="StartContainer for \"bdf5288c124ddcfed65c0aeaff6a5f93a43e75f13ea1ca09e0355be537b30f56\" returns successfully"
Jun 25 18:48:32.819691 containerd[1563]: time="2024-06-25T18:48:32.819512000Z" level=info msg="shim disconnected" id=bdf5288c124ddcfed65c0aeaff6a5f93a43e75f13ea1ca09e0355be537b30f56 namespace=k8s.io
Jun 25 18:48:32.819691 containerd[1563]: time="2024-06-25T18:48:32.819560558Z" level=warning msg="cleaning up after shim disconnected" id=bdf5288c124ddcfed65c0aeaff6a5f93a43e75f13ea1ca09e0355be537b30f56 namespace=k8s.io
Jun 25 18:48:32.819691 containerd[1563]: time="2024-06-25T18:48:32.819569038Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:48:33.511588 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-bdf5288c124ddcfed65c0aeaff6a5f93a43e75f13ea1ca09e0355be537b30f56-rootfs.mount: Deactivated successfully.
Jun 25 18:48:33.725301 kubelet[2664]: E0625 18:48:33.725269    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:33.727483 containerd[1563]: time="2024-06-25T18:48:33.727445585Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}"
Jun 25 18:48:33.759914 containerd[1563]: time="2024-06-25T18:48:33.759869984Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"151f2dac6115d7d0961a1c862b02a00152717795ee5132c889f6c55defcf9145\""
Jun 25 18:48:33.760796 containerd[1563]: time="2024-06-25T18:48:33.760760523Z" level=info msg="StartContainer for \"151f2dac6115d7d0961a1c862b02a00152717795ee5132c889f6c55defcf9145\""
Jun 25 18:48:33.804135 containerd[1563]: time="2024-06-25T18:48:33.804042707Z" level=info msg="StartContainer for \"151f2dac6115d7d0961a1c862b02a00152717795ee5132c889f6c55defcf9145\" returns successfully"
Jun 25 18:48:33.825985 containerd[1563]: time="2024-06-25T18:48:33.825928633Z" level=info msg="shim disconnected" id=151f2dac6115d7d0961a1c862b02a00152717795ee5132c889f6c55defcf9145 namespace=k8s.io
Jun 25 18:48:33.825985 containerd[1563]: time="2024-06-25T18:48:33.825981111Z" level=warning msg="cleaning up after shim disconnected" id=151f2dac6115d7d0961a1c862b02a00152717795ee5132c889f6c55defcf9145 namespace=k8s.io
Jun 25 18:48:33.825985 containerd[1563]: time="2024-06-25T18:48:33.825989671Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:48:34.511634 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-151f2dac6115d7d0961a1c862b02a00152717795ee5132c889f6c55defcf9145-rootfs.mount: Deactivated successfully.
Jun 25 18:48:34.563076 kubelet[2664]: E0625 18:48:34.563035    2664 kubelet.go:2855] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"
Jun 25 18:48:34.728863 kubelet[2664]: E0625 18:48:34.728834    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:34.731974 containerd[1563]: time="2024-06-25T18:48:34.731584840Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}"
Jun 25 18:48:34.743775 containerd[1563]: time="2024-06-25T18:48:34.743736579Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"285532a78f10b43ca836f5c8cb577f99d56735b9e6412576fbf03c39e61e8cad\""
Jun 25 18:48:34.745720 containerd[1563]: time="2024-06-25T18:48:34.745689017Z" level=info msg="StartContainer for \"285532a78f10b43ca836f5c8cb577f99d56735b9e6412576fbf03c39e61e8cad\""
Jun 25 18:48:34.786188 containerd[1563]: time="2024-06-25T18:48:34.786022831Z" level=info msg="StartContainer for \"285532a78f10b43ca836f5c8cb577f99d56735b9e6412576fbf03c39e61e8cad\" returns successfully"
Jun 25 18:48:34.803754 containerd[1563]: time="2024-06-25T18:48:34.803659932Z" level=error msg="collecting metrics for 285532a78f10b43ca836f5c8cb577f99d56735b9e6412576fbf03c39e61e8cad" error="cgroups: cgroup deleted: unknown"
Jun 25 18:48:34.804428 containerd[1563]: time="2024-06-25T18:48:34.804342558Z" level=info msg="shim disconnected" id=285532a78f10b43ca836f5c8cb577f99d56735b9e6412576fbf03c39e61e8cad namespace=k8s.io
Jun 25 18:48:34.804428 containerd[1563]: time="2024-06-25T18:48:34.804386717Z" level=warning msg="cleaning up after shim disconnected" id=285532a78f10b43ca836f5c8cb577f99d56735b9e6412576fbf03c39e61e8cad namespace=k8s.io
Jun 25 18:48:34.804428 containerd[1563]: time="2024-06-25T18:48:34.804395557Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jun 25 18:48:35.511745 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-285532a78f10b43ca836f5c8cb577f99d56735b9e6412576fbf03c39e61e8cad-rootfs.mount: Deactivated successfully.
Jun 25 18:48:35.732420 kubelet[2664]: E0625 18:48:35.732389    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:35.735853 containerd[1563]: time="2024-06-25T18:48:35.735561321Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}"
Jun 25 18:48:35.755841 containerd[1563]: time="2024-06-25T18:48:35.753223616Z" level=info msg="CreateContainer within sandbox \"80da64e5d005a821248f9c87360e669b2a22c036ab152a22e8a396a5db3684bc\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"77a1cb1f2b02fd2f9f3c4ba46627555ce1ed1eff79131ad93dc44bc729501ed6\""
Jun 25 18:48:35.755841 containerd[1563]: time="2024-06-25T18:48:35.753998401Z" level=info msg="StartContainer for \"77a1cb1f2b02fd2f9f3c4ba46627555ce1ed1eff79131ad93dc44bc729501ed6\""
Jun 25 18:48:35.810035 containerd[1563]: time="2024-06-25T18:48:35.809925910Z" level=info msg="StartContainer for \"77a1cb1f2b02fd2f9f3c4ba46627555ce1ed1eff79131ad93dc44bc729501ed6\" returns successfully"
Jun 25 18:48:36.100696 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aes-ce))
Jun 25 18:48:36.737811 kubelet[2664]: E0625 18:48:36.737783    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:36.751193 kubelet[2664]: I0625 18:48:36.750978    2664 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-b55b2" podStartSLOduration=5.750941642 podCreationTimestamp="2024-06-25 18:48:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:48:36.750624008 +0000 UTC m=+82.351207404" watchObservedRunningTime="2024-06-25 18:48:36.750941642 +0000 UTC m=+82.351525038"
Jun 25 18:48:36.925123 kubelet[2664]: I0625 18:48:36.925097    2664 setters.go:552] "Node became not ready" node="localhost" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-06-25T18:48:36Z","lastTransitionTime":"2024-06-25T18:48:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"}
Jun 25 18:48:37.740912 kubelet[2664]: E0625 18:48:37.740563    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:37.804544 systemd[1]: run-containerd-runc-k8s.io-77a1cb1f2b02fd2f9f3c4ba46627555ce1ed1eff79131ad93dc44bc729501ed6-runc.JoQ8nY.mount: Deactivated successfully.
Jun 25 18:48:38.936313 systemd-networkd[1237]: lxc_health: Link UP
Jun 25 18:48:38.947339 systemd-networkd[1237]: lxc_health: Gained carrier
Jun 25 18:48:39.695610 kubelet[2664]: E0625 18:48:39.695557    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:39.746134 kubelet[2664]: E0625 18:48:39.745100    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jun 25 18:48:40.103383 systemd-networkd[1237]: lxc_health: Gained IPv6LL
Jun 25 18:48:44.255405 sshd[4473]: pam_unix(sshd:session): session closed for user core
Jun 25 18:48:44.259056 systemd[1]: sshd@24-10.0.0.141:22-10.0.0.1:36338.service: Deactivated successfully.
Jun 25 18:48:44.261448 systemd-logind[1538]: Session 25 logged out. Waiting for processes to exit.
Jun 25 18:48:44.261888 systemd[1]: session-25.scope: Deactivated successfully.
Jun 25 18:48:44.264192 systemd-logind[1538]: Removed session 25.
Jun 25 18:48:44.487763 kubelet[2664]: E0625 18:48:44.486582    2664 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"