Jan 17 11:59:49.875511 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1]
Jan 17 11:59:49.875532 kernel: Linux version 6.6.71-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 13.3.1_p20240614 p17) 13.3.1 20240614, GNU ld (Gentoo 2.42 p3) 2.42.0) #1 SMP PREEMPT Fri Jan 17 10:42:25 -00 2025
Jan 17 11:59:49.875542 kernel: KASLR enabled
Jan 17 11:59:49.875548 kernel: efi: EFI v2.7 by EDK II
Jan 17 11:59:49.875553 kernel: efi: SMBIOS 3.0=0xdced0000 MEMATTR=0xdba86018 ACPI 2.0=0xd9710018 RNG=0xd971e498 MEMRESERVE=0xd9b43d18 
Jan 17 11:59:49.875559 kernel: random: crng init done
Jan 17 11:59:49.875566 kernel: ACPI: Early table checksum verification disabled
Jan 17 11:59:49.875572 kernel: ACPI: RSDP 0x00000000D9710018 000024 (v02 BOCHS )
Jan 17 11:59:49.875578 kernel: ACPI: XSDT 0x00000000D971FE98 000064 (v01 BOCHS  BXPC     00000001      01000013)
Jan 17 11:59:49.875585 kernel: ACPI: FACP 0x00000000D971FA98 000114 (v06 BOCHS  BXPC     00000001 BXPC 00000001)
Jan 17 11:59:49.875591 kernel: ACPI: DSDT 0x00000000D9717518 0014A2 (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jan 17 11:59:49.875597 kernel: ACPI: APIC 0x00000000D971FC18 0001A8 (v04 BOCHS  BXPC     00000001 BXPC 00000001)
Jan 17 11:59:49.875603 kernel: ACPI: PPTT 0x00000000D971D898 00009C (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jan 17 11:59:49.875609 kernel: ACPI: GTDT 0x00000000D971E818 000060 (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jan 17 11:59:49.875617 kernel: ACPI: MCFG 0x00000000D971E918 00003C (v01 BOCHS  BXPC     00000001 BXPC 00000001)
Jan 17 11:59:49.875625 kernel: ACPI: SPCR 0x00000000D971FF98 000050 (v02 BOCHS  BXPC     00000001 BXPC 00000001)
Jan 17 11:59:49.875631 kernel: ACPI: DBG2 0x00000000D971E418 000057 (v00 BOCHS  BXPC     00000001 BXPC 00000001)
Jan 17 11:59:49.875638 kernel: ACPI: IORT 0x00000000D971E718 000080 (v03 BOCHS  BXPC     00000001 BXPC 00000001)
Jan 17 11:59:49.875644 kernel: ACPI: SPCR: console: pl011,mmio,0x9000000,9600
Jan 17 11:59:49.875651 kernel: NUMA: Failed to initialise from firmware
Jan 17 11:59:49.875658 kernel: NUMA: Faking a node at [mem 0x0000000040000000-0x00000000dcffffff]
Jan 17 11:59:49.875665 kernel: NUMA: NODE_DATA [mem 0xdc958800-0xdc95dfff]
Jan 17 11:59:49.875671 kernel: Zone ranges:
Jan 17 11:59:49.875677 kernel:   DMA      [mem 0x0000000040000000-0x00000000dcffffff]
Jan 17 11:59:49.875684 kernel:   DMA32    empty
Jan 17 11:59:49.875692 kernel:   Normal   empty
Jan 17 11:59:49.875698 kernel: Movable zone start for each node
Jan 17 11:59:49.875704 kernel: Early memory node ranges
Jan 17 11:59:49.875711 kernel:   node   0: [mem 0x0000000040000000-0x00000000d976ffff]
Jan 17 11:59:49.875717 kernel:   node   0: [mem 0x00000000d9770000-0x00000000d9b3ffff]
Jan 17 11:59:49.875724 kernel:   node   0: [mem 0x00000000d9b40000-0x00000000dce1ffff]
Jan 17 11:59:49.875730 kernel:   node   0: [mem 0x00000000dce20000-0x00000000dceaffff]
Jan 17 11:59:49.875736 kernel:   node   0: [mem 0x00000000dceb0000-0x00000000dcebffff]
Jan 17 11:59:49.875743 kernel:   node   0: [mem 0x00000000dcec0000-0x00000000dcfdffff]
Jan 17 11:59:49.875749 kernel:   node   0: [mem 0x00000000dcfe0000-0x00000000dcffffff]
Jan 17 11:59:49.875755 kernel: Initmem setup node 0 [mem 0x0000000040000000-0x00000000dcffffff]
Jan 17 11:59:49.875762 kernel: On node 0, zone DMA: 12288 pages in unavailable ranges
Jan 17 11:59:49.875770 kernel: psci: probing for conduit method from ACPI.
Jan 17 11:59:49.875776 kernel: psci: PSCIv1.1 detected in firmware.
Jan 17 11:59:49.875783 kernel: psci: Using standard PSCI v0.2 function IDs
Jan 17 11:59:49.875792 kernel: psci: Trusted OS migration not required
Jan 17 11:59:49.875799 kernel: psci: SMC Calling Convention v1.1
Jan 17 11:59:49.875806 kernel: smccc: KVM: hypervisor services detected (0x00000000 0x00000000 0x00000000 0x00000003)
Jan 17 11:59:49.875823 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976
Jan 17 11:59:49.875830 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096
Jan 17 11:59:49.875836 kernel: pcpu-alloc: [0] 0 [0] 1 [0] 2 [0] 3 
Jan 17 11:59:49.875843 kernel: Detected PIPT I-cache on CPU0
Jan 17 11:59:49.875849 kernel: CPU features: detected: GIC system register CPU interface
Jan 17 11:59:49.875856 kernel: CPU features: detected: Hardware dirty bit management
Jan 17 11:59:49.875863 kernel: CPU features: detected: Spectre-v4
Jan 17 11:59:49.875869 kernel: CPU features: detected: Spectre-BHB
Jan 17 11:59:49.875876 kernel: CPU features: kernel page table isolation forced ON by KASLR
Jan 17 11:59:49.875883 kernel: CPU features: detected: Kernel page table isolation (KPTI)
Jan 17 11:59:49.875892 kernel: CPU features: detected: ARM erratum 1418040
Jan 17 11:59:49.875899 kernel: CPU features: detected: SSBS not fully self-synchronizing
Jan 17 11:59:49.875906 kernel: alternatives: applying boot alternatives
Jan 17 11:59:49.875914 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=1dec90e7382e4708d8bb0385f9465c79a53a2c2baf70ef34aed11855f47d17b3
Jan 17 11:59:49.875921 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space.
Jan 17 11:59:49.875928 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear)
Jan 17 11:59:49.875935 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear)
Jan 17 11:59:49.875942 kernel: Fallback order for Node 0: 0 
Jan 17 11:59:49.875949 kernel: Built 1 zonelists, mobility grouping on.  Total pages: 633024
Jan 17 11:59:49.875956 kernel: Policy zone: DMA
Jan 17 11:59:49.875962 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off
Jan 17 11:59:49.875970 kernel: software IO TLB: area num 4.
Jan 17 11:59:49.875987 kernel: software IO TLB: mapped [mem 0x00000000d2e00000-0x00000000d6e00000] (64MB)
Jan 17 11:59:49.875995 kernel: Memory: 2386532K/2572288K available (10240K kernel code, 2184K rwdata, 8096K rodata, 39360K init, 897K bss, 185756K reserved, 0K cma-reserved)
Jan 17 11:59:49.876002 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1
Jan 17 11:59:49.876009 kernel: rcu: Preemptible hierarchical RCU implementation.
Jan 17 11:59:49.876016 kernel: rcu:         RCU event tracing is enabled.
Jan 17 11:59:49.876023 kernel: rcu:         RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4.
Jan 17 11:59:49.876031 kernel:         Trampoline variant of Tasks RCU enabled.
Jan 17 11:59:49.876037 kernel:         Tracing variant of Tasks RCU enabled.
Jan 17 11:59:49.876044 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies.
Jan 17 11:59:49.876051 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4
Jan 17 11:59:49.876058 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0
Jan 17 11:59:49.876067 kernel: GICv3: 256 SPIs implemented
Jan 17 11:59:49.876074 kernel: GICv3: 0 Extended SPIs implemented
Jan 17 11:59:49.876081 kernel: Root IRQ handler: gic_handle_irq
Jan 17 11:59:49.876087 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI
Jan 17 11:59:49.876094 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000080a0000
Jan 17 11:59:49.876101 kernel: ITS [mem 0x08080000-0x0809ffff]
Jan 17 11:59:49.876108 kernel: ITS@0x0000000008080000: allocated 8192 Devices @400c0000 (indirect, esz 8, psz 64K, shr 1)
Jan 17 11:59:49.876114 kernel: ITS@0x0000000008080000: allocated 8192 Interrupt Collections @400d0000 (flat, esz 8, psz 64K, shr 1)
Jan 17 11:59:49.876121 kernel: GICv3: using LPI property table @0x00000000400f0000
Jan 17 11:59:49.876128 kernel: GICv3: CPU0: using allocated LPI pending table @0x0000000040100000
Jan 17 11:59:49.876135 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention.
Jan 17 11:59:49.876143 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jan 17 11:59:49.876149 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt).
Jan 17 11:59:49.876156 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns
Jan 17 11:59:49.876163 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns
Jan 17 11:59:49.876170 kernel: arm-pv: using stolen time PV
Jan 17 11:59:49.876177 kernel: Console: colour dummy device 80x25
Jan 17 11:59:49.876184 kernel: ACPI: Core revision 20230628
Jan 17 11:59:49.876191 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000)
Jan 17 11:59:49.876198 kernel: pid_max: default: 32768 minimum: 301
Jan 17 11:59:49.876205 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity
Jan 17 11:59:49.876214 kernel: landlock: Up and running.
Jan 17 11:59:49.876220 kernel: SELinux:  Initializing.
Jan 17 11:59:49.876227 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear)
Jan 17 11:59:49.876234 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear)
Jan 17 11:59:49.876241 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4.
Jan 17 11:59:49.876248 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=4.
Jan 17 11:59:49.876255 kernel: rcu: Hierarchical SRCU implementation.
Jan 17 11:59:49.876262 kernel: rcu:         Max phase no-delay instances is 400.
Jan 17 11:59:49.876269 kernel: Platform MSI: ITS@0x8080000 domain created
Jan 17 11:59:49.876277 kernel: PCI/MSI: ITS@0x8080000 domain created
Jan 17 11:59:49.876284 kernel: Remapping and enabling EFI services.
Jan 17 11:59:49.876291 kernel: smp: Bringing up secondary CPUs ...
Jan 17 11:59:49.876298 kernel: Detected PIPT I-cache on CPU1
Jan 17 11:59:49.876305 kernel: GICv3: CPU1: found redistributor 1 region 0:0x00000000080c0000
Jan 17 11:59:49.876312 kernel: GICv3: CPU1: using allocated LPI pending table @0x0000000040110000
Jan 17 11:59:49.876319 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jan 17 11:59:49.876325 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1]
Jan 17 11:59:49.876332 kernel: Detected PIPT I-cache on CPU2
Jan 17 11:59:49.876339 kernel: GICv3: CPU2: found redistributor 2 region 0:0x00000000080e0000
Jan 17 11:59:49.876348 kernel: GICv3: CPU2: using allocated LPI pending table @0x0000000040120000
Jan 17 11:59:49.876354 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jan 17 11:59:49.876366 kernel: CPU2: Booted secondary processor 0x0000000002 [0x413fd0c1]
Jan 17 11:59:49.876375 kernel: Detected PIPT I-cache on CPU3
Jan 17 11:59:49.876382 kernel: GICv3: CPU3: found redistributor 3 region 0:0x0000000008100000
Jan 17 11:59:49.876389 kernel: GICv3: CPU3: using allocated LPI pending table @0x0000000040130000
Jan 17 11:59:49.876396 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040
Jan 17 11:59:49.876403 kernel: CPU3: Booted secondary processor 0x0000000003 [0x413fd0c1]
Jan 17 11:59:49.876411 kernel: smp: Brought up 1 node, 4 CPUs
Jan 17 11:59:49.876419 kernel: SMP: Total of 4 processors activated.
Jan 17 11:59:49.876427 kernel: CPU features: detected: 32-bit EL0 Support
Jan 17 11:59:49.876434 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence
Jan 17 11:59:49.876441 kernel: CPU features: detected: Common not Private translations
Jan 17 11:59:49.876449 kernel: CPU features: detected: CRC32 instructions
Jan 17 11:59:49.876456 kernel: CPU features: detected: Enhanced Virtualization Traps
Jan 17 11:59:49.876463 kernel: CPU features: detected: RCpc load-acquire (LDAPR)
Jan 17 11:59:49.876470 kernel: CPU features: detected: LSE atomic instructions
Jan 17 11:59:49.876478 kernel: CPU features: detected: Privileged Access Never
Jan 17 11:59:49.876485 kernel: CPU features: detected: RAS Extension Support
Jan 17 11:59:49.876493 kernel: CPU features: detected: Speculative Store Bypassing Safe (SSBS)
Jan 17 11:59:49.876500 kernel: CPU: All CPU(s) started at EL1
Jan 17 11:59:49.876507 kernel: alternatives: applying system-wide alternatives
Jan 17 11:59:49.876514 kernel: devtmpfs: initialized
Jan 17 11:59:49.876522 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns
Jan 17 11:59:49.876530 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear)
Jan 17 11:59:49.876537 kernel: pinctrl core: initialized pinctrl subsystem
Jan 17 11:59:49.876546 kernel: SMBIOS 3.0.0 present.
Jan 17 11:59:49.876553 kernel: DMI: QEMU KVM Virtual Machine, BIOS edk2-20230524-3.fc38 05/24/2023
Jan 17 11:59:49.876560 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family
Jan 17 11:59:49.876568 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations
Jan 17 11:59:49.876575 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations
Jan 17 11:59:49.876582 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations
Jan 17 11:59:49.876590 kernel: audit: initializing netlink subsys (disabled)
Jan 17 11:59:49.876597 kernel: audit: type=2000 audit(0.024:1): state=initialized audit_enabled=0 res=1
Jan 17 11:59:49.876605 kernel: thermal_sys: Registered thermal governor 'step_wise'
Jan 17 11:59:49.876614 kernel: cpuidle: using governor menu
Jan 17 11:59:49.876621 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers.
Jan 17 11:59:49.876629 kernel: ASID allocator initialised with 32768 entries
Jan 17 11:59:49.876636 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5
Jan 17 11:59:49.876643 kernel: Serial: AMBA PL011 UART driver
Jan 17 11:59:49.876650 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL
Jan 17 11:59:49.876657 kernel: Modules: 0 pages in range for non-PLT usage
Jan 17 11:59:49.876664 kernel: Modules: 509040 pages in range for PLT usage
Jan 17 11:59:49.876672 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages
Jan 17 11:59:49.876681 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page
Jan 17 11:59:49.876688 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages
Jan 17 11:59:49.876695 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page
Jan 17 11:59:49.876702 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages
Jan 17 11:59:49.876710 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page
Jan 17 11:59:49.876717 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages
Jan 17 11:59:49.876724 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page
Jan 17 11:59:49.876731 kernel: ACPI: Added _OSI(Module Device)
Jan 17 11:59:49.876738 kernel: ACPI: Added _OSI(Processor Device)
Jan 17 11:59:49.876747 kernel: ACPI: Added _OSI(3.0 _SCP Extensions)
Jan 17 11:59:49.876754 kernel: ACPI: Added _OSI(Processor Aggregator Device)
Jan 17 11:59:49.876761 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded
Jan 17 11:59:49.876768 kernel: ACPI: Interpreter enabled
Jan 17 11:59:49.876776 kernel: ACPI: Using GIC for interrupt routing
Jan 17 11:59:49.876783 kernel: ACPI: MCFG table detected, 1 entries
Jan 17 11:59:49.876790 kernel: ARMH0011:00: ttyAMA0 at MMIO 0x9000000 (irq = 12, base_baud = 0) is a SBSA
Jan 17 11:59:49.876797 kernel: printk: console [ttyAMA0] enabled
Jan 17 11:59:49.876804 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff])
Jan 17 11:59:49.876953 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3]
Jan 17 11:59:49.877058 kernel: acpi PNP0A08:00: _OSC: platform does not support [LTR]
Jan 17 11:59:49.877124 kernel: acpi PNP0A08:00: _OSC: OS now controls [PCIeHotplug PME AER PCIeCapability]
Jan 17 11:59:49.877187 kernel: acpi PNP0A08:00: ECAM area [mem 0x4010000000-0x401fffffff] reserved by PNP0C02:00
Jan 17 11:59:49.877248 kernel: acpi PNP0A08:00: ECAM at [mem 0x4010000000-0x401fffffff] for [bus 00-ff]
Jan 17 11:59:49.877258 kernel: ACPI: Remapped I/O 0x000000003eff0000 to [io  0x0000-0xffff window]
Jan 17 11:59:49.877265 kernel: PCI host bridge to bus 0000:00
Jan 17 11:59:49.877336 kernel: pci_bus 0000:00: root bus resource [mem 0x10000000-0x3efeffff window]
Jan 17 11:59:49.877394 kernel: pci_bus 0000:00: root bus resource [io  0x0000-0xffff window]
Jan 17 11:59:49.877451 kernel: pci_bus 0000:00: root bus resource [mem 0x8000000000-0xffffffffff window]
Jan 17 11:59:49.877507 kernel: pci_bus 0000:00: root bus resource [bus 00-ff]
Jan 17 11:59:49.877586 kernel: pci 0000:00:00.0: [1b36:0008] type 00 class 0x060000
Jan 17 11:59:49.877660 kernel: pci 0000:00:01.0: [1af4:1005] type 00 class 0x00ff00
Jan 17 11:59:49.877729 kernel: pci 0000:00:01.0: reg 0x10: [io  0x0000-0x001f]
Jan 17 11:59:49.877822 kernel: pci 0000:00:01.0: reg 0x14: [mem 0x10000000-0x10000fff]
Jan 17 11:59:49.877894 kernel: pci 0000:00:01.0: reg 0x20: [mem 0x8000000000-0x8000003fff 64bit pref]
Jan 17 11:59:49.877961 kernel: pci 0000:00:01.0: BAR 4: assigned [mem 0x8000000000-0x8000003fff 64bit pref]
Jan 17 11:59:49.878047 kernel: pci 0000:00:01.0: BAR 1: assigned [mem 0x10000000-0x10000fff]
Jan 17 11:59:49.878112 kernel: pci 0000:00:01.0: BAR 0: assigned [io  0x1000-0x101f]
Jan 17 11:59:49.878172 kernel: pci_bus 0000:00: resource 4 [mem 0x10000000-0x3efeffff window]
Jan 17 11:59:49.878234 kernel: pci_bus 0000:00: resource 5 [io  0x0000-0xffff window]
Jan 17 11:59:49.878293 kernel: pci_bus 0000:00: resource 6 [mem 0x8000000000-0xffffffffff window]
Jan 17 11:59:49.878303 kernel: ACPI: PCI: Interrupt link GSI0 configured for IRQ 35
Jan 17 11:59:49.878312 kernel: ACPI: PCI: Interrupt link GSI1 configured for IRQ 36
Jan 17 11:59:49.878319 kernel: ACPI: PCI: Interrupt link GSI2 configured for IRQ 37
Jan 17 11:59:49.878327 kernel: ACPI: PCI: Interrupt link GSI3 configured for IRQ 38
Jan 17 11:59:49.878335 kernel: iommu: Default domain type: Translated
Jan 17 11:59:49.878343 kernel: iommu: DMA domain TLB invalidation policy: strict mode
Jan 17 11:59:49.878353 kernel: efivars: Registered efivars operations
Jan 17 11:59:49.878361 kernel: vgaarb: loaded
Jan 17 11:59:49.878368 kernel: clocksource: Switched to clocksource arch_sys_counter
Jan 17 11:59:49.878376 kernel: VFS: Disk quotas dquot_6.6.0
Jan 17 11:59:49.878384 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes)
Jan 17 11:59:49.878391 kernel: pnp: PnP ACPI init
Jan 17 11:59:49.878466 kernel: system 00:00: [mem 0x4010000000-0x401fffffff window] could not be reserved
Jan 17 11:59:49.878477 kernel: pnp: PnP ACPI: found 1 devices
Jan 17 11:59:49.878484 kernel: NET: Registered PF_INET protocol family
Jan 17 11:59:49.878494 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear)
Jan 17 11:59:49.878502 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear)
Jan 17 11:59:49.878509 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear)
Jan 17 11:59:49.878517 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear)
Jan 17 11:59:49.878524 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear)
Jan 17 11:59:49.878532 kernel: TCP: Hash tables configured (established 32768 bind 32768)
Jan 17 11:59:49.878539 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear)
Jan 17 11:59:49.878547 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear)
Jan 17 11:59:49.878556 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family
Jan 17 11:59:49.878563 kernel: PCI: CLS 0 bytes, default 64
Jan 17 11:59:49.878571 kernel: kvm [1]: HYP mode not available
Jan 17 11:59:49.878578 kernel: Initialise system trusted keyrings
Jan 17 11:59:49.878585 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0
Jan 17 11:59:49.878592 kernel: Key type asymmetric registered
Jan 17 11:59:49.878599 kernel: Asymmetric key parser 'x509' registered
Jan 17 11:59:49.878607 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250)
Jan 17 11:59:49.878614 kernel: io scheduler mq-deadline registered
Jan 17 11:59:49.878621 kernel: io scheduler kyber registered
Jan 17 11:59:49.878630 kernel: io scheduler bfq registered
Jan 17 11:59:49.878638 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXSYBUS:00/PNP0C0C:00/input/input0
Jan 17 11:59:49.878645 kernel: ACPI: button: Power Button [PWRB]
Jan 17 11:59:49.878653 kernel: ACPI: \_SB_.PCI0.GSI1: Enabled at IRQ 36
Jan 17 11:59:49.878719 kernel: virtio-pci 0000:00:01.0: enabling device (0005 -> 0007)
Jan 17 11:59:49.878729 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled
Jan 17 11:59:49.878737 kernel: thunder_xcv, ver 1.0
Jan 17 11:59:49.878744 kernel: thunder_bgx, ver 1.0
Jan 17 11:59:49.878751 kernel: nicpf, ver 1.0
Jan 17 11:59:49.878760 kernel: nicvf, ver 1.0
Jan 17 11:59:49.878852 kernel: rtc-efi rtc-efi.0: registered as rtc0
Jan 17 11:59:49.878918 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-01-17T11:59:49 UTC (1737115189)
Jan 17 11:59:49.878928 kernel: hid: raw HID events driver (C) Jiri Kosina
Jan 17 11:59:49.878935 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 7 counters available
Jan 17 11:59:49.878943 kernel: watchdog: Delayed init of the lockup detector failed: -19
Jan 17 11:59:49.878950 kernel: watchdog: Hard watchdog permanently disabled
Jan 17 11:59:49.878958 kernel: NET: Registered PF_INET6 protocol family
Jan 17 11:59:49.878967 kernel: Segment Routing with IPv6
Jan 17 11:59:49.878975 kernel: In-situ OAM (IOAM) with IPv6
Jan 17 11:59:49.878993 kernel: NET: Registered PF_PACKET protocol family
Jan 17 11:59:49.879021 kernel: Key type dns_resolver registered
Jan 17 11:59:49.879030 kernel: registered taskstats version 1
Jan 17 11:59:49.879037 kernel: Loading compiled-in X.509 certificates
Jan 17 11:59:49.879044 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.71-flatcar: e5b890cba32c3e1c766d9a9b821ee4d2154ffee7'
Jan 17 11:59:49.879052 kernel: Key type .fscrypt registered
Jan 17 11:59:49.879059 kernel: Key type fscrypt-provisioning registered
Jan 17 11:59:49.879069 kernel: ima: No TPM chip found, activating TPM-bypass!
Jan 17 11:59:49.879076 kernel: ima: Allocated hash algorithm: sha1
Jan 17 11:59:49.879084 kernel: ima: No architecture policies found
Jan 17 11:59:49.879091 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng)
Jan 17 11:59:49.879098 kernel: clk: Disabling unused clocks
Jan 17 11:59:49.879106 kernel: Freeing unused kernel memory: 39360K
Jan 17 11:59:49.879113 kernel: Run /init as init process
Jan 17 11:59:49.879120 kernel:   with arguments:
Jan 17 11:59:49.879127 kernel:     /init
Jan 17 11:59:49.879137 kernel:   with environment:
Jan 17 11:59:49.879144 kernel:     HOME=/
Jan 17 11:59:49.879151 kernel:     TERM=linux
Jan 17 11:59:49.879158 kernel:     BOOT_IMAGE=/flatcar/vmlinuz-a
Jan 17 11:59:49.879168 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified)
Jan 17 11:59:49.879177 systemd[1]: Detected virtualization kvm.
Jan 17 11:59:49.879185 systemd[1]: Detected architecture arm64.
Jan 17 11:59:49.879195 systemd[1]: Running in initrd.
Jan 17 11:59:49.879203 systemd[1]: No hostname configured, using default hostname.
Jan 17 11:59:49.879210 systemd[1]: Hostname set to <localhost>.
Jan 17 11:59:49.879218 systemd[1]: Initializing machine ID from VM UUID.
Jan 17 11:59:49.879226 systemd[1]: Queued start job for default target initrd.target.
Jan 17 11:59:49.879234 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Jan 17 11:59:49.879242 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Jan 17 11:59:49.879250 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM...
Jan 17 11:59:49.879259 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM...
Jan 17 11:59:49.879273 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT...
Jan 17 11:59:49.879281 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A...
Jan 17 11:59:49.879314 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132...
Jan 17 11:59:49.879322 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr...
Jan 17 11:59:49.879330 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Jan 17 11:59:49.879338 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes.
Jan 17 11:59:49.879347 systemd[1]: Reached target paths.target - Path Units.
Jan 17 11:59:49.879356 systemd[1]: Reached target slices.target - Slice Units.
Jan 17 11:59:49.879364 systemd[1]: Reached target swap.target - Swaps.
Jan 17 11:59:49.879372 systemd[1]: Reached target timers.target - Timer Units.
Jan 17 11:59:49.879380 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket.
Jan 17 11:59:49.879388 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket.
Jan 17 11:59:49.879396 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log).
Jan 17 11:59:49.879404 systemd[1]: Listening on systemd-journald.socket - Journal Socket.
Jan 17 11:59:49.879412 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket.
Jan 17 11:59:49.879422 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket.
Jan 17 11:59:49.879430 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket.
Jan 17 11:59:49.879437 systemd[1]: Reached target sockets.target - Socket Units.
Jan 17 11:59:49.879445 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup...
Jan 17 11:59:49.879453 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes...
Jan 17 11:59:49.879461 systemd[1]: Finished network-cleanup.service - Network Cleanup.
Jan 17 11:59:49.879469 systemd[1]: Starting systemd-fsck-usr.service...
Jan 17 11:59:49.879477 systemd[1]: Starting systemd-journald.service - Journal Service...
Jan 17 11:59:49.879486 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules...
Jan 17 11:59:49.879494 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Jan 17 11:59:49.879502 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup.
Jan 17 11:59:49.879510 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes.
Jan 17 11:59:49.879518 systemd[1]: Finished systemd-fsck-usr.service.
Jan 17 11:59:49.879526 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully...
Jan 17 11:59:49.879557 systemd-journald[236]: Collecting audit messages is disabled.
Jan 17 11:59:49.879577 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Jan 17 11:59:49.879585 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully.
Jan 17 11:59:49.879596 systemd-journald[236]: Journal started
Jan 17 11:59:49.879614 systemd-journald[236]: Runtime Journal (/run/log/journal/21d97b2ded8f4e43985b852446ef9d83) is 5.9M, max 47.3M, 41.4M free.
Jan 17 11:59:49.879653 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Jan 17 11:59:49.871430 systemd-modules-load[237]: Inserted module 'overlay'
Jan 17 11:59:49.884024 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this.
Jan 17 11:59:49.884071 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev...
Jan 17 11:59:49.886082 kernel: Bridge firewalling registered
Jan 17 11:59:49.886114 systemd[1]: Started systemd-journald.service - Journal Service.
Jan 17 11:59:49.885626 systemd-modules-load[237]: Inserted module 'br_netfilter'
Jan 17 11:59:49.887346 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules.
Jan 17 11:59:49.891273 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Jan 17 11:59:49.893210 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories...
Jan 17 11:59:49.896030 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Jan 17 11:59:49.904404 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories.
Jan 17 11:59:49.906101 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Jan 17 11:59:49.908397 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jan 17 11:59:49.919203 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook...
Jan 17 11:59:49.921216 systemd[1]: Starting systemd-resolved.service - Network Name Resolution...
Jan 17 11:59:49.929344 dracut-cmdline[274]: dracut-dracut-053
Jan 17 11:59:49.931956 dracut-cmdline[274]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected acpi=force verity.usrhash=1dec90e7382e4708d8bb0385f9465c79a53a2c2baf70ef34aed11855f47d17b3
Jan 17 11:59:49.948567 systemd-resolved[276]: Positive Trust Anchors:
Jan 17 11:59:49.948585 systemd-resolved[276]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Jan 17 11:59:49.948617 systemd-resolved[276]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test
Jan 17 11:59:49.953189 systemd-resolved[276]: Defaulting to hostname 'linux'.
Jan 17 11:59:49.954092 systemd[1]: Started systemd-resolved.service - Network Name Resolution.
Jan 17 11:59:49.956072 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups.
Jan 17 11:59:50.004009 kernel: SCSI subsystem initialized
Jan 17 11:59:50.009002 kernel: Loading iSCSI transport class v2.0-870.
Jan 17 11:59:50.018019 kernel: iscsi: registered transport (tcp)
Jan 17 11:59:50.031007 kernel: iscsi: registered transport (qla4xxx)
Jan 17 11:59:50.031032 kernel: QLogic iSCSI HBA Driver
Jan 17 11:59:50.074060 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook.
Jan 17 11:59:50.082228 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook...
Jan 17 11:59:50.100005 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log.
Jan 17 11:59:50.100057 kernel: device-mapper: uevent: version 1.0.3
Jan 17 11:59:50.101219 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com
Jan 17 11:59:50.151014 kernel: raid6: neonx8   gen() 15792 MB/s
Jan 17 11:59:50.167994 kernel: raid6: neonx4   gen() 15654 MB/s
Jan 17 11:59:50.184995 kernel: raid6: neonx2   gen() 13290 MB/s
Jan 17 11:59:50.201994 kernel: raid6: neonx1   gen() 10478 MB/s
Jan 17 11:59:50.218995 kernel: raid6: int64x8  gen()  6949 MB/s
Jan 17 11:59:50.235991 kernel: raid6: int64x4  gen()  7344 MB/s
Jan 17 11:59:50.252996 kernel: raid6: int64x2  gen()  6128 MB/s
Jan 17 11:59:50.269993 kernel: raid6: int64x1  gen()  5058 MB/s
Jan 17 11:59:50.270009 kernel: raid6: using algorithm neonx8 gen() 15792 MB/s
Jan 17 11:59:50.286998 kernel: raid6: .... xor() 11921 MB/s, rmw enabled
Jan 17 11:59:50.287013 kernel: raid6: using neon recovery algorithm
Jan 17 11:59:50.294259 kernel: xor: measuring software checksum speed
Jan 17 11:59:50.294278 kernel:    8regs           : 19816 MB/sec
Jan 17 11:59:50.294288 kernel:    32regs          : 19413 MB/sec
Jan 17 11:59:50.295213 kernel:    arm64_neon      : 27070 MB/sec
Jan 17 11:59:50.295225 kernel: xor: using function: arm64_neon (27070 MB/sec)
Jan 17 11:59:50.349004 kernel: Btrfs loaded, zoned=no, fsverity=no
Jan 17 11:59:50.362681 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook.
Jan 17 11:59:50.370151 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files...
Jan 17 11:59:50.381025 systemd-udevd[458]: Using default interface naming scheme 'v255'.
Jan 17 11:59:50.384200 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files.
Jan 17 11:59:50.391164 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook...
Jan 17 11:59:50.402596 dracut-pre-trigger[465]: rd.md=0: removing MD RAID activation
Jan 17 11:59:50.428390 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook.
Jan 17 11:59:50.439167 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices...
Jan 17 11:59:50.476044 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices.
Jan 17 11:59:50.485148 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook...
Jan 17 11:59:50.496599 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook.
Jan 17 11:59:50.498969 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems.
Jan 17 11:59:50.501130 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes.
Jan 17 11:59:50.502907 systemd[1]: Reached target remote-fs.target - Remote File Systems.
Jan 17 11:59:50.516163 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook...
Jan 17 11:59:50.522484 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully.
Jan 17 11:59:50.524989 kernel: virtio_blk virtio1: 1/0/0 default/read/poll queues
Jan 17 11:59:50.529604 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB)
Jan 17 11:59:50.529709 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk.
Jan 17 11:59:50.529721 kernel: GPT:9289727 != 19775487
Jan 17 11:59:50.529731 kernel: GPT:Alternate GPT header not at the end of the disk.
Jan 17 11:59:50.529740 kernel: GPT:9289727 != 19775487
Jan 17 11:59:50.529756 kernel: GPT: Use GNU Parted to correct GPT errors.
Jan 17 11:59:50.529766 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Jan 17 11:59:50.523185 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jan 17 11:59:50.526323 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Jan 17 11:59:50.527328 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully.
Jan 17 11:59:50.527489 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup.
Jan 17 11:59:50.533519 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup...
Jan 17 11:59:50.542252 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Jan 17 11:59:50.545004 kernel: BTRFS: device fsid 8c8354db-e4b6-4022-87e4-d06cc74d2d9f devid 1 transid 40 /dev/vda3 scanned by (udev-worker) (503)
Jan 17 11:59:50.547007 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (517)
Jan 17 11:59:50.547252 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook.
Jan 17 11:59:50.555854 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Jan 17 11:59:50.563726 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM.
Jan 17 11:59:50.570596 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT.
Jan 17 11:59:50.574227 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132.
Jan 17 11:59:50.575115 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A.
Jan 17 11:59:50.580522 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM.
Jan 17 11:59:50.597158 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary...
Jan 17 11:59:50.598731 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters...
Jan 17 11:59:50.604316 disk-uuid[550]: Primary Header is updated.
Jan 17 11:59:50.604316 disk-uuid[550]: Secondary Entries is updated.
Jan 17 11:59:50.604316 disk-uuid[550]: Secondary Header is updated.
Jan 17 11:59:50.612004 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Jan 17 11:59:50.615922 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jan 17 11:59:51.622000 kernel:  vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9
Jan 17 11:59:51.622386 disk-uuid[551]: The operation has completed successfully.
Jan 17 11:59:51.651512 systemd[1]: disk-uuid.service: Deactivated successfully.
Jan 17 11:59:51.651606 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary.
Jan 17 11:59:51.672190 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr...
Jan 17 11:59:51.675064 sh[574]: Success
Jan 17 11:59:51.690026 kernel: device-mapper: verity: sha256 using implementation "sha256-ce"
Jan 17 11:59:51.727479 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr.
Jan 17 11:59:51.729097 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr...
Jan 17 11:59:51.729832 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr.
Jan 17 11:59:51.745005 kernel: BTRFS info (device dm-0): first mount of filesystem 8c8354db-e4b6-4022-87e4-d06cc74d2d9f
Jan 17 11:59:51.745056 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm
Jan 17 11:59:51.745067 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead
Jan 17 11:59:51.745212 kernel: BTRFS info (device dm-0): disabling log replay at mount time
Jan 17 11:59:51.746240 kernel: BTRFS info (device dm-0): using free space tree
Jan 17 11:59:51.758081 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr.
Jan 17 11:59:51.759179 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met.
Jan 17 11:59:51.767126 systemd[1]: Starting ignition-setup.service - Ignition (setup)...
Jan 17 11:59:51.768426 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline...
Jan 17 11:59:51.778330 kernel: BTRFS info (device vda6): first mount of filesystem 5a5108d6-bc75-4f85-aab0-f326070fd0b5
Jan 17 11:59:51.778381 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm
Jan 17 11:59:51.778392 kernel: BTRFS info (device vda6): using free space tree
Jan 17 11:59:51.783118 kernel: BTRFS info (device vda6): auto enabling async discard
Jan 17 11:59:51.791343 systemd[1]: mnt-oem.mount: Deactivated successfully.
Jan 17 11:59:51.792834 kernel: BTRFS info (device vda6): last unmount of filesystem 5a5108d6-bc75-4f85-aab0-f326070fd0b5
Jan 17 11:59:51.798723 systemd[1]: Finished ignition-setup.service - Ignition (setup).
Jan 17 11:59:51.808165 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)...
Jan 17 11:59:51.886812 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline.
Jan 17 11:59:51.899184 systemd[1]: Starting systemd-networkd.service - Network Configuration...
Jan 17 11:59:51.923830 systemd-networkd[764]: lo: Link UP
Jan 17 11:59:51.923841 systemd-networkd[764]: lo: Gained carrier
Jan 17 11:59:51.924840 systemd-networkd[764]: Enumeration completed
Jan 17 11:59:51.924921 systemd[1]: Started systemd-networkd.service - Network Configuration.
Jan 17 11:59:51.925844 systemd[1]: Reached target network.target - Network.
Jan 17 11:59:51.927707 systemd-networkd[764]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jan 17 11:59:51.927710 systemd-networkd[764]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network.
Jan 17 11:59:51.932474 systemd-networkd[764]: eth0: Link UP
Jan 17 11:59:51.932489 systemd-networkd[764]: eth0: Gained carrier
Jan 17 11:59:51.932498 systemd-networkd[764]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jan 17 11:59:51.935217 ignition[666]: Ignition 2.19.0
Jan 17 11:59:51.935223 ignition[666]: Stage: fetch-offline
Jan 17 11:59:51.935260 ignition[666]: no configs at "/usr/lib/ignition/base.d"
Jan 17 11:59:51.935268 ignition[666]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jan 17 11:59:51.935422 ignition[666]: parsed url from cmdline: ""
Jan 17 11:59:51.935425 ignition[666]: no config URL provided
Jan 17 11:59:51.935430 ignition[666]: reading system config file "/usr/lib/ignition/user.ign"
Jan 17 11:59:51.935436 ignition[666]: no config at "/usr/lib/ignition/user.ign"
Jan 17 11:59:51.935459 ignition[666]: op(1): [started]  loading QEMU firmware config module
Jan 17 11:59:51.935465 ignition[666]: op(1): executing: "modprobe" "qemu_fw_cfg"
Jan 17 11:59:51.944387 ignition[666]: op(1): [finished] loading QEMU firmware config module
Jan 17 11:59:51.944995 ignition[666]: QEMU firmware config was not found. Ignoring...
Jan 17 11:59:51.959054 systemd-networkd[764]: eth0: DHCPv4 address 10.0.0.35/16, gateway 10.0.0.1 acquired from 10.0.0.1
Jan 17 11:59:51.968195 ignition[666]: parsing config with SHA512: 065d04589b759ccf511ea7ff7303bf7bcba29e99612a9b1e0d501902bb016dc8d9090060f17b5c98e92d5c12c959e61c92d416232a6a3fc0e88b70403c055e18
Jan 17 11:59:51.973366 unknown[666]: fetched base config from "system"
Jan 17 11:59:51.973383 unknown[666]: fetched user config from "qemu"
Jan 17 11:59:51.975444 ignition[666]: fetch-offline: fetch-offline passed
Jan 17 11:59:51.976178 ignition[666]: Ignition finished successfully
Jan 17 11:59:51.978317 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline).
Jan 17 11:59:51.979409 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json).
Jan 17 11:59:51.991125 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)...
Jan 17 11:59:52.002305 ignition[773]: Ignition 2.19.0
Jan 17 11:59:52.002316 ignition[773]: Stage: kargs
Jan 17 11:59:52.002484 ignition[773]: no configs at "/usr/lib/ignition/base.d"
Jan 17 11:59:52.002494 ignition[773]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jan 17 11:59:52.003376 ignition[773]: kargs: kargs passed
Jan 17 11:59:52.003420 ignition[773]: Ignition finished successfully
Jan 17 11:59:52.006660 systemd[1]: Finished ignition-kargs.service - Ignition (kargs).
Jan 17 11:59:52.025566 systemd[1]: Starting ignition-disks.service - Ignition (disks)...
Jan 17 11:59:52.037573 ignition[781]: Ignition 2.19.0
Jan 17 11:59:52.037582 ignition[781]: Stage: disks
Jan 17 11:59:52.037756 ignition[781]: no configs at "/usr/lib/ignition/base.d"
Jan 17 11:59:52.037766 ignition[781]: no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jan 17 11:59:52.038751 ignition[781]: disks: disks passed
Jan 17 11:59:52.040767 systemd[1]: Finished ignition-disks.service - Ignition (disks).
Jan 17 11:59:52.038835 ignition[781]: Ignition finished successfully
Jan 17 11:59:52.041848 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device.
Jan 17 11:59:52.042811 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems.
Jan 17 11:59:52.044284 systemd[1]: Reached target local-fs.target - Local File Systems.
Jan 17 11:59:52.045408 systemd[1]: Reached target sysinit.target - System Initialization.
Jan 17 11:59:52.046759 systemd[1]: Reached target basic.target - Basic System.
Jan 17 11:59:52.052185 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT...
Jan 17 11:59:52.064319 systemd-fsck[791]: ROOT: clean, 14/553520 files, 52654/553472 blocks
Jan 17 11:59:52.070131 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT.
Jan 17 11:59:52.081369 systemd[1]: Mounting sysroot.mount - /sysroot...
Jan 17 11:59:52.126997 kernel: EXT4-fs (vda9): mounted filesystem 5d516319-3144-49e6-9760-d0f29faba535 r/w with ordered data mode. Quota mode: none.
Jan 17 11:59:52.127755 systemd[1]: Mounted sysroot.mount - /sysroot.
Jan 17 11:59:52.128854 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System.
Jan 17 11:59:52.138127 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem...
Jan 17 11:59:52.139626 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr...
Jan 17 11:59:52.140375 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met.
Jan 17 11:59:52.140412 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot).
Jan 17 11:59:52.140432 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup.
Jan 17 11:59:52.146422 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr.
Jan 17 11:59:52.148237 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup...
Jan 17 11:59:52.151767 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (799)
Jan 17 11:59:52.151796 kernel: BTRFS info (device vda6): first mount of filesystem 5a5108d6-bc75-4f85-aab0-f326070fd0b5
Jan 17 11:59:52.151807 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm
Jan 17 11:59:52.151823 kernel: BTRFS info (device vda6): using free space tree
Jan 17 11:59:52.154991 kernel: BTRFS info (device vda6): auto enabling async discard
Jan 17 11:59:52.156468 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem.
Jan 17 11:59:52.198354 initrd-setup-root[823]: cut: /sysroot/etc/passwd: No such file or directory
Jan 17 11:59:52.202020 initrd-setup-root[830]: cut: /sysroot/etc/group: No such file or directory
Jan 17 11:59:52.205875 initrd-setup-root[837]: cut: /sysroot/etc/shadow: No such file or directory
Jan 17 11:59:52.208666 initrd-setup-root[844]: cut: /sysroot/etc/gshadow: No such file or directory
Jan 17 11:59:52.280166 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup.
Jan 17 11:59:52.289091 systemd[1]: Starting ignition-mount.service - Ignition (mount)...
Jan 17 11:59:52.290378 systemd[1]: Starting sysroot-boot.service - /sysroot/boot...
Jan 17 11:59:52.295007 kernel: BTRFS info (device vda6): last unmount of filesystem 5a5108d6-bc75-4f85-aab0-f326070fd0b5
Jan 17 11:59:52.316832 systemd[1]: Finished sysroot-boot.service - /sysroot/boot.
Jan 17 11:59:52.321875 ignition[912]: INFO     : Ignition 2.19.0
Jan 17 11:59:52.321875 ignition[912]: INFO     : Stage: mount
Jan 17 11:59:52.323087 ignition[912]: INFO     : no configs at "/usr/lib/ignition/base.d"
Jan 17 11:59:52.323087 ignition[912]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jan 17 11:59:52.323087 ignition[912]: INFO     : mount: mount passed
Jan 17 11:59:52.323087 ignition[912]: INFO     : Ignition finished successfully
Jan 17 11:59:52.325043 systemd[1]: Finished ignition-mount.service - Ignition (mount).
Jan 17 11:59:52.337076 systemd[1]: Starting ignition-files.service - Ignition (files)...
Jan 17 11:59:52.743449 systemd[1]: sysroot-oem.mount: Deactivated successfully.
Jan 17 11:59:52.759173 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem...
Jan 17 11:59:52.763998 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (925)
Jan 17 11:59:52.766405 kernel: BTRFS info (device vda6): first mount of filesystem 5a5108d6-bc75-4f85-aab0-f326070fd0b5
Jan 17 11:59:52.766437 kernel: BTRFS info (device vda6): using crc32c (crc32c-generic) checksum algorithm
Jan 17 11:59:52.766448 kernel: BTRFS info (device vda6): using free space tree
Jan 17 11:59:52.769001 kernel: BTRFS info (device vda6): auto enabling async discard
Jan 17 11:59:52.769590 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem.
Jan 17 11:59:52.785032 ignition[942]: INFO     : Ignition 2.19.0
Jan 17 11:59:52.785032 ignition[942]: INFO     : Stage: files
Jan 17 11:59:52.786246 ignition[942]: INFO     : no configs at "/usr/lib/ignition/base.d"
Jan 17 11:59:52.786246 ignition[942]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jan 17 11:59:52.786246 ignition[942]: DEBUG    : files: compiled without relabeling support, skipping
Jan 17 11:59:52.789132 ignition[942]: INFO     : files: ensureUsers: op(1): [started]  creating or modifying user "core"
Jan 17 11:59:52.789132 ignition[942]: DEBUG    : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core"
Jan 17 11:59:52.792167 ignition[942]: INFO     : files: ensureUsers: op(1): [finished] creating or modifying user "core"
Jan 17 11:59:52.793083 ignition[942]: INFO     : files: ensureUsers: op(2): [started]  adding ssh keys to user "core"
Jan 17 11:59:52.793083 ignition[942]: INFO     : files: ensureUsers: op(2): [finished] adding ssh keys to user "core"
Jan 17 11:59:52.792580 unknown[942]: wrote ssh authorized keys file for user: core
Jan 17 11:59:52.796503 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(3): [started]  writing file "/sysroot/etc/flatcar-cgroupv1"
Jan 17 11:59:52.797739 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1"
Jan 17 11:59:52.797739 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(4): [started]  writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz"
Jan 17 11:59:52.797739 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1
Jan 17 11:59:53.049003 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(4): GET result: OK
Jan 17 11:59:53.112932 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(5): [started]  writing file "/sysroot/home/core/install.sh"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(6): [started]  writing file "/sysroot/home/core/nginx.yaml"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(7): [started]  writing file "/sysroot/home/core/nfs-pod.yaml"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(8): [started]  writing file "/sysroot/home/core/nfs-pvc.yaml"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(9): [started]  writing file "/sysroot/etc/flatcar/update.conf"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf"
Jan 17 11:59:53.114360 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(a): [started]  writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-arm64.raw"
Jan 17 11:59:53.127770 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-arm64.raw"
Jan 17 11:59:53.127770 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(b): [started]  writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-arm64.raw"
Jan 17 11:59:53.127770 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.29.2-arm64.raw: attempt #1
Jan 17 11:59:53.441862 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(b): GET result: OK
Jan 17 11:59:53.626274 ignition[942]: INFO     : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-arm64.raw"
Jan 17 11:59:53.626274 ignition[942]: INFO     : files: op(c): [started]  processing unit "containerd.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(c): op(d): [started]  writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(c): [finished] processing unit "containerd.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(e): [started]  processing unit "prepare-helm.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(e): op(f): [started]  writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(e): [finished] processing unit "prepare-helm.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(10): [started]  processing unit "coreos-metadata.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(10): op(11): [started]  writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(10): op(11): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(10): [finished] processing unit "coreos-metadata.service"
Jan 17 11:59:53.628908 ignition[942]: INFO     : files: op(12): [started]  setting preset to disabled for "coreos-metadata.service"
Jan 17 11:59:53.651201 ignition[942]: INFO     : files: op(12): op(13): [started]  removing enablement symlink(s) for "coreos-metadata.service"
Jan 17 11:59:53.651132 systemd-networkd[764]: eth0: Gained IPv6LL
Jan 17 11:59:53.655209 ignition[942]: INFO     : files: op(12): op(13): [finished] removing enablement symlink(s) for "coreos-metadata.service"
Jan 17 11:59:53.656352 ignition[942]: INFO     : files: op(12): [finished] setting preset to disabled for "coreos-metadata.service"
Jan 17 11:59:53.656352 ignition[942]: INFO     : files: op(14): [started]  setting preset to enabled for "prepare-helm.service"
Jan 17 11:59:53.656352 ignition[942]: INFO     : files: op(14): [finished] setting preset to enabled for "prepare-helm.service"
Jan 17 11:59:53.656352 ignition[942]: INFO     : files: createResultFile: createFiles: op(15): [started]  writing file "/sysroot/etc/.ignition-result.json"
Jan 17 11:59:53.656352 ignition[942]: INFO     : files: createResultFile: createFiles: op(15): [finished] writing file "/sysroot/etc/.ignition-result.json"
Jan 17 11:59:53.656352 ignition[942]: INFO     : files: files passed
Jan 17 11:59:53.656352 ignition[942]: INFO     : Ignition finished successfully
Jan 17 11:59:53.658420 systemd[1]: Finished ignition-files.service - Ignition (files).
Jan 17 11:59:53.670115 systemd[1]: Starting ignition-quench.service - Ignition (record completion)...
Jan 17 11:59:53.671553 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion...
Jan 17 11:59:53.673564 systemd[1]: ignition-quench.service: Deactivated successfully.
Jan 17 11:59:53.673666 systemd[1]: Finished ignition-quench.service - Ignition (record completion).
Jan 17 11:59:53.679436 initrd-setup-root-after-ignition[970]: grep: /sysroot/oem/oem-release: No such file or directory
Jan 17 11:59:53.682557 initrd-setup-root-after-ignition[972]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory
Jan 17 11:59:53.682557 initrd-setup-root-after-ignition[972]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory
Jan 17 11:59:53.684917 initrd-setup-root-after-ignition[976]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory
Jan 17 11:59:53.684602 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion.
Jan 17 11:59:53.686157 systemd[1]: Reached target ignition-complete.target - Ignition Complete.
Jan 17 11:59:53.702214 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root...
Jan 17 11:59:53.721391 systemd[1]: initrd-parse-etc.service: Deactivated successfully.
Jan 17 11:59:53.721499 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root.
Jan 17 11:59:53.723199 systemd[1]: Reached target initrd-fs.target - Initrd File Systems.
Jan 17 11:59:53.724526 systemd[1]: Reached target initrd.target - Initrd Default Target.
Jan 17 11:59:53.725827 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met.
Jan 17 11:59:53.726589 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook...
Jan 17 11:59:53.742180 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook.
Jan 17 11:59:53.754164 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons...
Jan 17 11:59:53.762092 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups.
Jan 17 11:59:53.762973 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes.
Jan 17 11:59:53.764476 systemd[1]: Stopped target timers.target - Timer Units.
Jan 17 11:59:53.765764 systemd[1]: dracut-pre-pivot.service: Deactivated successfully.
Jan 17 11:59:53.765889 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook.
Jan 17 11:59:53.767666 systemd[1]: Stopped target initrd.target - Initrd Default Target.
Jan 17 11:59:53.769097 systemd[1]: Stopped target basic.target - Basic System.
Jan 17 11:59:53.770257 systemd[1]: Stopped target ignition-complete.target - Ignition Complete.
Jan 17 11:59:53.771486 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup.
Jan 17 11:59:53.772859 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device.
Jan 17 11:59:53.774284 systemd[1]: Stopped target remote-fs.target - Remote File Systems.
Jan 17 11:59:53.775611 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems.
Jan 17 11:59:53.776993 systemd[1]: Stopped target sysinit.target - System Initialization.
Jan 17 11:59:53.778416 systemd[1]: Stopped target local-fs.target - Local File Systems.
Jan 17 11:59:53.779677 systemd[1]: Stopped target swap.target - Swaps.
Jan 17 11:59:53.780811 systemd[1]: dracut-pre-mount.service: Deactivated successfully.
Jan 17 11:59:53.780932 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook.
Jan 17 11:59:53.782570 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes.
Jan 17 11:59:53.784055 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Jan 17 11:59:53.785524 systemd[1]: clevis-luks-askpass.path: Deactivated successfully.
Jan 17 11:59:53.786897 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Jan 17 11:59:53.787842 systemd[1]: dracut-initqueue.service: Deactivated successfully.
Jan 17 11:59:53.787948 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook.
Jan 17 11:59:53.790008 systemd[1]: ignition-fetch-offline.service: Deactivated successfully.
Jan 17 11:59:53.790120 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline).
Jan 17 11:59:53.791552 systemd[1]: Stopped target paths.target - Path Units.
Jan 17 11:59:53.792633 systemd[1]: systemd-ask-password-console.path: Deactivated successfully.
Jan 17 11:59:53.792722 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Jan 17 11:59:53.794147 systemd[1]: Stopped target slices.target - Slice Units.
Jan 17 11:59:53.795260 systemd[1]: Stopped target sockets.target - Socket Units.
Jan 17 11:59:53.796520 systemd[1]: iscsid.socket: Deactivated successfully.
Jan 17 11:59:53.796610 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket.
Jan 17 11:59:53.798082 systemd[1]: iscsiuio.socket: Deactivated successfully.
Jan 17 11:59:53.798164 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket.
Jan 17 11:59:53.799282 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully.
Jan 17 11:59:53.799386 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion.
Jan 17 11:59:53.800619 systemd[1]: ignition-files.service: Deactivated successfully.
Jan 17 11:59:53.800727 systemd[1]: Stopped ignition-files.service - Ignition (files).
Jan 17 11:59:53.813194 systemd[1]: Stopping ignition-mount.service - Ignition (mount)...
Jan 17 11:59:53.813851 systemd[1]: kmod-static-nodes.service: Deactivated successfully.
Jan 17 11:59:53.813995 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes.
Jan 17 11:59:53.816534 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot...
Jan 17 11:59:53.817340 systemd[1]: systemd-udev-trigger.service: Deactivated successfully.
Jan 17 11:59:53.817462 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices.
Jan 17 11:59:53.818779 systemd[1]: dracut-pre-trigger.service: Deactivated successfully.
Jan 17 11:59:53.818879 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook.
Jan 17 11:59:53.824608 ignition[996]: INFO     : Ignition 2.19.0
Jan 17 11:59:53.824608 ignition[996]: INFO     : Stage: umount
Jan 17 11:59:53.826075 ignition[996]: INFO     : no configs at "/usr/lib/ignition/base.d"
Jan 17 11:59:53.826075 ignition[996]: INFO     : no config dir at "/usr/lib/ignition/base.platform.d/qemu"
Jan 17 11:59:53.826075 ignition[996]: INFO     : umount: umount passed
Jan 17 11:59:53.826075 ignition[996]: INFO     : Ignition finished successfully
Jan 17 11:59:53.827380 systemd[1]: ignition-mount.service: Deactivated successfully.
Jan 17 11:59:53.828255 systemd[1]: Stopped ignition-mount.service - Ignition (mount).
Jan 17 11:59:53.829921 systemd[1]: sysroot-boot.mount: Deactivated successfully.
Jan 17 11:59:53.830448 systemd[1]: initrd-cleanup.service: Deactivated successfully.
Jan 17 11:59:53.830540 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons.
Jan 17 11:59:53.832223 systemd[1]: Stopped target network.target - Network.
Jan 17 11:59:53.832933 systemd[1]: ignition-disks.service: Deactivated successfully.
Jan 17 11:59:53.833021 systemd[1]: Stopped ignition-disks.service - Ignition (disks).
Jan 17 11:59:53.834295 systemd[1]: ignition-kargs.service: Deactivated successfully.
Jan 17 11:59:53.834338 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs).
Jan 17 11:59:53.835656 systemd[1]: ignition-setup.service: Deactivated successfully.
Jan 17 11:59:53.835697 systemd[1]: Stopped ignition-setup.service - Ignition (setup).
Jan 17 11:59:53.837021 systemd[1]: ignition-setup-pre.service: Deactivated successfully.
Jan 17 11:59:53.837061 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup.
Jan 17 11:59:53.838469 systemd[1]: Stopping systemd-networkd.service - Network Configuration...
Jan 17 11:59:53.839678 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution...
Jan 17 11:59:53.847011 systemd-networkd[764]: eth0: DHCPv6 lease lost
Jan 17 11:59:53.848456 systemd[1]: systemd-networkd.service: Deactivated successfully.
Jan 17 11:59:53.848571 systemd[1]: Stopped systemd-networkd.service - Network Configuration.
Jan 17 11:59:53.851149 systemd[1]: systemd-resolved.service: Deactivated successfully.
Jan 17 11:59:53.851975 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution.
Jan 17 11:59:53.853965 systemd[1]: systemd-networkd.socket: Deactivated successfully.
Jan 17 11:59:53.854415 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket.
Jan 17 11:59:53.862154 systemd[1]: Stopping network-cleanup.service - Network Cleanup...
Jan 17 11:59:53.862833 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully.
Jan 17 11:59:53.862896 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline.
Jan 17 11:59:53.864421 systemd[1]: systemd-sysctl.service: Deactivated successfully.
Jan 17 11:59:53.864465 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables.
Jan 17 11:59:53.865822 systemd[1]: systemd-modules-load.service: Deactivated successfully.
Jan 17 11:59:53.865866 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules.
Jan 17 11:59:53.866722 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully.
Jan 17 11:59:53.866763 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories.
Jan 17 11:59:53.868710 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files...
Jan 17 11:59:53.885334 systemd[1]: systemd-udevd.service: Deactivated successfully.
Jan 17 11:59:53.885496 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files.
Jan 17 11:59:53.887747 systemd[1]: sysroot-boot.service: Deactivated successfully.
Jan 17 11:59:53.887864 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot.
Jan 17 11:59:53.888807 systemd[1]: network-cleanup.service: Deactivated successfully.
Jan 17 11:59:53.888888 systemd[1]: Stopped network-cleanup.service - Network Cleanup.
Jan 17 11:59:53.891083 systemd[1]: systemd-udevd-control.socket: Deactivated successfully.
Jan 17 11:59:53.891142 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket.
Jan 17 11:59:53.892302 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully.
Jan 17 11:59:53.892336 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket.
Jan 17 11:59:53.893596 systemd[1]: dracut-pre-udev.service: Deactivated successfully.
Jan 17 11:59:53.893637 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook.
Jan 17 11:59:53.895668 systemd[1]: dracut-cmdline.service: Deactivated successfully.
Jan 17 11:59:53.895712 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook.
Jan 17 11:59:53.897684 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully.
Jan 17 11:59:53.897733 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters.
Jan 17 11:59:53.899804 systemd[1]: initrd-setup-root.service: Deactivated successfully.
Jan 17 11:59:53.899851 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup.
Jan 17 11:59:53.909102 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database...
Jan 17 11:59:53.909874 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully.
Jan 17 11:59:53.909926 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Jan 17 11:59:53.911534 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully.
Jan 17 11:59:53.911573 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup.
Jan 17 11:59:53.916370 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully.
Jan 17 11:59:53.916463 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database.
Jan 17 11:59:53.918115 systemd[1]: Reached target initrd-switch-root.target - Switch Root.
Jan 17 11:59:53.920151 systemd[1]: Starting initrd-switch-root.service - Switch Root...
Jan 17 11:59:53.929356 systemd[1]: Switching root.
Jan 17 11:59:53.948699 systemd-journald[236]: Journal stopped
Jan 17 11:59:54.733154 systemd-journald[236]: Received SIGTERM from PID 1 (systemd).
Jan 17 11:59:54.733223 kernel: SELinux:  policy capability network_peer_controls=1
Jan 17 11:59:54.733237 kernel: SELinux:  policy capability open_perms=1
Jan 17 11:59:54.733247 kernel: SELinux:  policy capability extended_socket_class=1
Jan 17 11:59:54.733263 kernel: SELinux:  policy capability always_check_network=0
Jan 17 11:59:54.733273 kernel: SELinux:  policy capability cgroup_seclabel=1
Jan 17 11:59:54.733283 kernel: SELinux:  policy capability nnp_nosuid_transition=1
Jan 17 11:59:54.733293 kernel: SELinux:  policy capability genfs_seclabel_symlinks=0
Jan 17 11:59:54.733303 kernel: SELinux:  policy capability ioctl_skip_cloexec=0
Jan 17 11:59:54.733316 kernel: audit: type=1403 audit(1737115194.121:2): auid=4294967295 ses=4294967295 lsm=selinux res=1
Jan 17 11:59:54.733331 systemd[1]: Successfully loaded SELinux policy in 30.835ms.
Jan 17 11:59:54.733347 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 9.075ms.
Jan 17 11:59:54.733359 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified)
Jan 17 11:59:54.733372 systemd[1]: Detected virtualization kvm.
Jan 17 11:59:54.733383 systemd[1]: Detected architecture arm64.
Jan 17 11:59:54.733393 systemd[1]: Detected first boot.
Jan 17 11:59:54.733404 systemd[1]: Initializing machine ID from VM UUID.
Jan 17 11:59:54.733414 zram_generator::config[1062]: No configuration found.
Jan 17 11:59:54.733425 systemd[1]: Populated /etc with preset unit settings.
Jan 17 11:59:54.733435 systemd[1]: Queued start job for default target multi-user.target.
Jan 17 11:59:54.733446 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6.
Jan 17 11:59:54.733459 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config.
Jan 17 11:59:54.733470 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run.
Jan 17 11:59:54.733481 systemd[1]: Created slice system-getty.slice - Slice /system/getty.
Jan 17 11:59:54.733491 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe.
Jan 17 11:59:54.733503 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty.
Jan 17 11:59:54.733513 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit.
Jan 17 11:59:54.733524 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck.
Jan 17 11:59:54.733538 systemd[1]: Created slice user.slice - User and Session Slice.
Jan 17 11:59:54.733548 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch.
Jan 17 11:59:54.733562 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch.
Jan 17 11:59:54.733573 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch.
Jan 17 11:59:54.733584 systemd[1]: Set up automount boot.automount - Boot partition Automount Point.
Jan 17 11:59:54.733595 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point.
Jan 17 11:59:54.733606 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM...
Jan 17 11:59:54.733617 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0...
Jan 17 11:59:54.733627 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre).
Jan 17 11:59:54.733638 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes.
Jan 17 11:59:54.733648 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes.
Jan 17 11:59:54.733660 systemd[1]: Reached target remote-fs.target - Remote File Systems.
Jan 17 11:59:54.733671 systemd[1]: Reached target slices.target - Slice Units.
Jan 17 11:59:54.733682 systemd[1]: Reached target swap.target - Swaps.
Jan 17 11:59:54.733693 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes.
Jan 17 11:59:54.733704 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket.
Jan 17 11:59:54.733715 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log).
Jan 17 11:59:54.733725 systemd[1]: Listening on systemd-journald.socket - Journal Socket.
Jan 17 11:59:54.733737 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket.
Jan 17 11:59:54.733749 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket.
Jan 17 11:59:54.733767 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket.
Jan 17 11:59:54.733781 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket.
Jan 17 11:59:54.733794 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System...
Jan 17 11:59:54.733805 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System...
Jan 17 11:59:54.733815 systemd[1]: Mounting media.mount - External Media Directory...
Jan 17 11:59:54.733826 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System...
Jan 17 11:59:54.733836 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System...
Jan 17 11:59:54.733847 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp...
Jan 17 11:59:54.733859 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files...
Jan 17 11:59:54.733870 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jan 17 11:59:54.734672 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes...
Jan 17 11:59:54.734694 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs...
Jan 17 11:59:54.734705 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jan 17 11:59:54.734716 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm...
Jan 17 11:59:54.734727 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jan 17 11:59:54.734751 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse...
Jan 17 11:59:54.734778 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jan 17 11:59:54.734790 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf).
Jan 17 11:59:54.734803 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling.
Jan 17 11:59:54.734816 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.)
Jan 17 11:59:54.734827 kernel: fuse: init (API version 7.39)
Jan 17 11:59:54.734837 systemd[1]: Starting systemd-journald.service - Journal Service...
Jan 17 11:59:54.734848 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules...
Jan 17 11:59:54.734858 kernel: loop: module loaded
Jan 17 11:59:54.734868 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line...
Jan 17 11:59:54.734879 kernel: ACPI: bus type drm_connector registered
Jan 17 11:59:54.734890 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems...
Jan 17 11:59:54.734933 systemd-journald[1140]: Collecting audit messages is disabled.
Jan 17 11:59:54.734957 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices...
Jan 17 11:59:54.734967 systemd-journald[1140]: Journal started
Jan 17 11:59:54.735025 systemd-journald[1140]: Runtime Journal (/run/log/journal/21d97b2ded8f4e43985b852446ef9d83) is 5.9M, max 47.3M, 41.4M free.
Jan 17 11:59:54.739587 systemd[1]: Started systemd-journald.service - Journal Service.
Jan 17 11:59:54.740789 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System.
Jan 17 11:59:54.741670 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System.
Jan 17 11:59:54.742589 systemd[1]: Mounted media.mount - External Media Directory.
Jan 17 11:59:54.743409 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System.
Jan 17 11:59:54.744412 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System.
Jan 17 11:59:54.745459 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp.
Jan 17 11:59:54.746519 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files.
Jan 17 11:59:54.747865 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes.
Jan 17 11:59:54.749054 systemd[1]: modprobe@configfs.service: Deactivated successfully.
Jan 17 11:59:54.749209 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs.
Jan 17 11:59:54.750271 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jan 17 11:59:54.750429 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jan 17 11:59:54.751537 systemd[1]: modprobe@drm.service: Deactivated successfully.
Jan 17 11:59:54.751691 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm.
Jan 17 11:59:54.752729 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jan 17 11:59:54.752892 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jan 17 11:59:54.754101 systemd[1]: modprobe@fuse.service: Deactivated successfully.
Jan 17 11:59:54.754254 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse.
Jan 17 11:59:54.755556 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jan 17 11:59:54.755776 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jan 17 11:59:54.756904 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules.
Jan 17 11:59:54.758291 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line.
Jan 17 11:59:54.759862 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems.
Jan 17 11:59:54.770918 systemd[1]: Reached target network-pre.target - Preparation for Network.
Jan 17 11:59:54.780158 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System...
Jan 17 11:59:54.781919 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System...
Jan 17 11:59:54.782772 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/).
Jan 17 11:59:54.786704 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database...
Jan 17 11:59:54.790687 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage...
Jan 17 11:59:54.791615 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore).
Jan 17 11:59:54.792612 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed...
Jan 17 11:59:54.793626 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met.
Jan 17 11:59:54.796815 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables...
Jan 17 11:59:54.797614 systemd-journald[1140]: Time spent on flushing to /var/log/journal/21d97b2ded8f4e43985b852446ef9d83 is 13.828ms for 843 entries.
Jan 17 11:59:54.797614 systemd-journald[1140]: System Journal (/var/log/journal/21d97b2ded8f4e43985b852446ef9d83) is 8.0M, max 195.6M, 187.6M free.
Jan 17 11:59:54.832999 systemd-journald[1140]: Received client request to flush runtime journal.
Jan 17 11:59:54.799745 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully...
Jan 17 11:59:54.802218 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices.
Jan 17 11:59:54.803238 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System.
Jan 17 11:59:54.804162 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System.
Jan 17 11:59:54.812217 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization...
Jan 17 11:59:54.813483 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed.
Jan 17 11:59:54.814962 systemd[1]: Reached target first-boot-complete.target - First Boot Complete.
Jan 17 11:59:54.824168 udevadm[1199]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in.
Jan 17 11:59:54.826106 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables.
Jan 17 11:59:54.835385 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage.
Jan 17 11:59:54.837054 systemd-tmpfiles[1193]: ACLs are not supported, ignoring.
Jan 17 11:59:54.837067 systemd-tmpfiles[1193]: ACLs are not supported, ignoring.
Jan 17 11:59:54.841255 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully.
Jan 17 11:59:54.850221 systemd[1]: Starting systemd-sysusers.service - Create System Users...
Jan 17 11:59:54.870098 systemd[1]: Finished systemd-sysusers.service - Create System Users.
Jan 17 11:59:54.885166 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev...
Jan 17 11:59:54.896453 systemd-tmpfiles[1215]: ACLs are not supported, ignoring.
Jan 17 11:59:54.896473 systemd-tmpfiles[1215]: ACLs are not supported, ignoring.
Jan 17 11:59:54.901524 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev.
Jan 17 11:59:55.229491 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database.
Jan 17 11:59:55.245236 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files...
Jan 17 11:59:55.267609 systemd-udevd[1221]: Using default interface naming scheme 'v255'.
Jan 17 11:59:55.280426 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files.
Jan 17 11:59:55.313274 systemd[1]: Starting systemd-networkd.service - Network Configuration...
Jan 17 11:59:55.322957 systemd[1]: Found device dev-ttyAMA0.device - /dev/ttyAMA0.
Jan 17 11:59:55.331154 systemd[1]: Starting systemd-userdbd.service - User Database Manager...
Jan 17 11:59:55.339999 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 40 scanned by (udev-worker) (1232)
Jan 17 11:59:55.372568 systemd[1]: Started systemd-userdbd.service - User Database Manager.
Jan 17 11:59:55.386533 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM.
Jan 17 11:59:55.417312 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup...
Jan 17 11:59:55.423871 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization.
Jan 17 11:59:55.426819 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes...
Jan 17 11:59:55.438890 systemd-networkd[1228]: lo: Link UP
Jan 17 11:59:55.438899 systemd-networkd[1228]: lo: Gained carrier
Jan 17 11:59:55.439581 systemd-networkd[1228]: Enumeration completed
Jan 17 11:59:55.439705 systemd[1]: Started systemd-networkd.service - Network Configuration.
Jan 17 11:59:55.440103 systemd-networkd[1228]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jan 17 11:59:55.440113 systemd-networkd[1228]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network.
Jan 17 11:59:55.440732 systemd-networkd[1228]: eth0: Link UP
Jan 17 11:59:55.440741 systemd-networkd[1228]: eth0: Gained carrier
Jan 17 11:59:55.440760 systemd-networkd[1228]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name.
Jan 17 11:59:55.449129 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured...
Jan 17 11:59:55.449539 lvm[1257]:   WARNING: Failed to connect to lvmetad. Falling back to device scanning.
Jan 17 11:59:55.465044 systemd-networkd[1228]: eth0: DHCPv4 address 10.0.0.35/16, gateway 10.0.0.1 acquired from 10.0.0.1
Jan 17 11:59:55.465306 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup.
Jan 17 11:59:55.479453 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes.
Jan 17 11:59:55.480566 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes.
Jan 17 11:59:55.493302 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes...
Jan 17 11:59:55.496833 lvm[1267]:   WARNING: Failed to connect to lvmetad. Falling back to device scanning.
Jan 17 11:59:55.525579 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes.
Jan 17 11:59:55.526701 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems.
Jan 17 11:59:55.527637 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw).
Jan 17 11:59:55.527666 systemd[1]: Reached target local-fs.target - Local File Systems.
Jan 17 11:59:55.528430 systemd[1]: Reached target machines.target - Containers.
Jan 17 11:59:55.530143 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink).
Jan 17 11:59:55.542164 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown...
Jan 17 11:59:55.544258 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache...
Jan 17 11:59:55.545118 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jan 17 11:59:55.546012 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM...
Jan 17 11:59:55.547906 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk...
Jan 17 11:59:55.553193 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/...
Jan 17 11:59:55.554816 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown.
Jan 17 11:59:55.561799 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM.
Jan 17 11:59:55.565294 kernel: loop0: detected capacity change from 0 to 114328
Jan 17 11:59:55.573873 systemd[1]: etc-machine\x2did.mount: Deactivated successfully.
Jan 17 11:59:55.574621 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk.
Jan 17 11:59:55.577011 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher
Jan 17 11:59:55.608118 kernel: loop1: detected capacity change from 0 to 114432
Jan 17 11:59:55.640012 kernel: loop2: detected capacity change from 0 to 194512
Jan 17 11:59:55.676015 kernel: loop3: detected capacity change from 0 to 114328
Jan 17 11:59:55.688007 kernel: loop4: detected capacity change from 0 to 114432
Jan 17 11:59:55.702284 kernel: loop5: detected capacity change from 0 to 194512
Jan 17 11:59:55.716304 (sd-merge)[1288]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'.
Jan 17 11:59:55.716727 (sd-merge)[1288]: Merged extensions into '/usr'.
Jan 17 11:59:55.721833 systemd[1]: Reloading requested from client PID 1275 ('systemd-sysext') (unit systemd-sysext.service)...
Jan 17 11:59:55.721854 systemd[1]: Reloading...
Jan 17 11:59:55.764020 zram_generator::config[1313]: No configuration found.
Jan 17 11:59:55.846037 ldconfig[1271]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start.
Jan 17 11:59:55.864602 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jan 17 11:59:55.912872 systemd[1]: Reloading finished in 190 ms.
Jan 17 11:59:55.932197 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache.
Jan 17 11:59:55.933366 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/.
Jan 17 11:59:55.953221 systemd[1]: Starting ensure-sysext.service...
Jan 17 11:59:55.955221 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories...
Jan 17 11:59:55.960023 systemd[1]: Reloading requested from client PID 1357 ('systemctl') (unit ensure-sysext.service)...
Jan 17 11:59:55.960041 systemd[1]: Reloading...
Jan 17 11:59:55.973118 systemd-tmpfiles[1358]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring.
Jan 17 11:59:55.973382 systemd-tmpfiles[1358]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring.
Jan 17 11:59:55.974043 systemd-tmpfiles[1358]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring.
Jan 17 11:59:55.974268 systemd-tmpfiles[1358]: ACLs are not supported, ignoring.
Jan 17 11:59:55.974319 systemd-tmpfiles[1358]: ACLs are not supported, ignoring.
Jan 17 11:59:55.976452 systemd-tmpfiles[1358]: Detected autofs mount point /boot during canonicalization of boot.
Jan 17 11:59:55.976465 systemd-tmpfiles[1358]: Skipping /boot
Jan 17 11:59:55.986144 systemd-tmpfiles[1358]: Detected autofs mount point /boot during canonicalization of boot.
Jan 17 11:59:55.986159 systemd-tmpfiles[1358]: Skipping /boot
Jan 17 11:59:55.997018 zram_generator::config[1384]: No configuration found.
Jan 17 11:59:56.093055 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jan 17 11:59:56.135942 systemd[1]: Reloading finished in 175 ms.
Jan 17 11:59:56.149681 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories.
Jan 17 11:59:56.166994 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules...
Jan 17 11:59:56.172134 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs...
Jan 17 11:59:56.174135 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog...
Jan 17 11:59:56.179147 systemd[1]: Starting systemd-resolved.service - Network Name Resolution...
Jan 17 11:59:56.181199 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP...
Jan 17 11:59:56.188176 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jan 17 11:59:56.192142 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jan 17 11:59:56.197483 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jan 17 11:59:56.200382 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jan 17 11:59:56.202211 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jan 17 11:59:56.203113 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jan 17 11:59:56.203275 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jan 17 11:59:56.204634 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jan 17 11:59:56.204785 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jan 17 11:59:56.208707 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog.
Jan 17 11:59:56.210466 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jan 17 11:59:56.210637 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jan 17 11:59:56.216742 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jan 17 11:59:56.224245 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jan 17 11:59:56.226762 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jan 17 11:59:56.230365 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jan 17 11:59:56.231163 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jan 17 11:59:56.233567 systemd[1]: Starting systemd-update-done.service - Update is Completed...
Jan 17 11:59:56.242641 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP.
Jan 17 11:59:56.244394 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jan 17 11:59:56.244542 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jan 17 11:59:56.246016 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jan 17 11:59:56.246162 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jan 17 11:59:56.249532 augenrules[1470]: No rules
Jan 17 11:59:56.247798 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jan 17 11:59:56.247943 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jan 17 11:59:56.252216 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules.
Jan 17 11:59:56.254893 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs.
Jan 17 11:59:56.257691 systemd[1]: Finished systemd-update-done.service - Update is Completed.
Jan 17 11:59:56.264266 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met.
Jan 17 11:59:56.264551 systemd-resolved[1434]: Positive Trust Anchors:
Jan 17 11:59:56.264814 systemd-resolved[1434]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d
Jan 17 11:59:56.264899 systemd-resolved[1434]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test
Jan 17 11:59:56.272256 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod...
Jan 17 11:59:56.273179 systemd-resolved[1434]: Defaulting to hostname 'linux'.
Jan 17 11:59:56.274251 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm...
Jan 17 11:59:56.277296 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore...
Jan 17 11:59:56.281166 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop...
Jan 17 11:59:56.282353 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met.
Jan 17 11:59:56.282632 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt).
Jan 17 11:59:56.283900 systemd[1]: Started systemd-resolved.service - Network Name Resolution.
Jan 17 11:59:56.285614 systemd[1]: modprobe@dm_mod.service: Deactivated successfully.
Jan 17 11:59:56.285892 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod.
Jan 17 11:59:56.287272 systemd[1]: modprobe@drm.service: Deactivated successfully.
Jan 17 11:59:56.287417 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm.
Jan 17 11:59:56.288671 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully.
Jan 17 11:59:56.288831 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore.
Jan 17 11:59:56.290153 systemd[1]: modprobe@loop.service: Deactivated successfully.
Jan 17 11:59:56.290369 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop.
Jan 17 11:59:56.292782 systemd[1]: Finished ensure-sysext.service.
Jan 17 11:59:56.297844 systemd[1]: Reached target network.target - Network.
Jan 17 11:59:56.298554 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups.
Jan 17 11:59:56.299559 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore).
Jan 17 11:59:56.299614 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met.
Jan 17 11:59:56.317196 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization...
Jan 17 11:59:56.358724 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization.
Jan 17 11:59:56.359760 systemd[1]: Reached target sysinit.target - System Initialization.
Jan 17 11:59:56.360588 systemd[1]: Started motdgen.path - Watch for update engine configuration changes.
Jan 17 11:59:56.361476 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data.
Jan 17 11:59:56.362817 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories.
Jan 17 11:59:56.362855 systemd-timesyncd[1501]: Contacted time server 10.0.0.1:123 (10.0.0.1).
Jan 17 11:59:56.362895 systemd-timesyncd[1501]: Initial clock synchronization to Fri 2025-01-17 11:59:56.639813 UTC.
Jan 17 11:59:56.363774 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate).
Jan 17 11:59:56.363802 systemd[1]: Reached target paths.target - Path Units.
Jan 17 11:59:56.364474 systemd[1]: Reached target time-set.target - System Time Set.
Jan 17 11:59:56.365514 systemd[1]: Started logrotate.timer - Daily rotation of log files.
Jan 17 11:59:56.366522 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information..
Jan 17 11:59:56.367770 systemd[1]: Reached target timers.target - Timer Units.
Jan 17 11:59:56.371814 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket.
Jan 17 11:59:56.375139 systemd[1]: Starting docker.socket - Docker Socket for the API...
Jan 17 11:59:56.376772 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket.
Jan 17 11:59:56.385206 systemd[1]: Listening on docker.socket - Docker Socket for the API.
Jan 17 11:59:56.386011 systemd[1]: Reached target sockets.target - Socket Units.
Jan 17 11:59:56.386691 systemd[1]: Reached target basic.target - Basic System.
Jan 17 11:59:56.387628 systemd[1]: System is tainted: cgroupsv1
Jan 17 11:59:56.387661 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met.
Jan 17 11:59:56.387680 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met.
Jan 17 11:59:56.388956 systemd[1]: Starting containerd.service - containerd container runtime...
Jan 17 11:59:56.392145 systemd[1]: Starting dbus.service - D-Bus System Message Bus...
Jan 17 11:59:56.394583 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit...
Jan 17 11:59:56.396568 systemd[1]: Starting extend-filesystems.service - Extend Filesystems...
Jan 17 11:59:56.397532 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment).
Jan 17 11:59:56.399207 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd...
Jan 17 11:59:56.403410 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin...
Jan 17 11:59:56.408213 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline...
Jan 17 11:59:56.413629 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys...
Jan 17 11:59:56.417631 jq[1507]: false
Jan 17 11:59:56.419388 systemd[1]: Starting systemd-logind.service - User Login Management...
Jan 17 11:59:56.423784 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0).
Jan 17 11:59:56.428154 systemd[1]: Starting update-engine.service - Update Engine...
Jan 17 11:59:56.429917 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition...
Jan 17 11:59:56.432584 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'.
Jan 17 11:59:56.436169 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped.
Jan 17 11:59:56.437845 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully.
Jan 17 11:59:56.438104 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline.
Jan 17 11:59:56.442018 extend-filesystems[1509]: Found loop3
Jan 17 11:59:56.442018 extend-filesystems[1509]: Found loop4
Jan 17 11:59:56.442018 extend-filesystems[1509]: Found loop5
Jan 17 11:59:56.442018 extend-filesystems[1509]: Found vda
Jan 17 11:59:56.442018 extend-filesystems[1509]: Found vda1
Jan 17 11:59:56.442018 extend-filesystems[1509]: Found vda2
Jan 17 11:59:56.442018 extend-filesystems[1509]: Found vda3
Jan 17 11:59:56.452682 systemd[1]: Started dbus.service - D-Bus System Message Bus.
Jan 17 11:59:56.449532 dbus-daemon[1506]: [system] SELinux support is enabled
Jan 17 11:59:56.457347 extend-filesystems[1509]: Found usr
Jan 17 11:59:56.457347 extend-filesystems[1509]: Found vda4
Jan 17 11:59:56.457347 extend-filesystems[1509]: Found vda6
Jan 17 11:59:56.457347 extend-filesystems[1509]: Found vda7
Jan 17 11:59:56.457347 extend-filesystems[1509]: Found vda9
Jan 17 11:59:56.457347 extend-filesystems[1509]: Checking size of /dev/vda9
Jan 17 11:59:56.466371 jq[1528]: true
Jan 17 11:59:56.456632 (ntainerd)[1534]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR
Jan 17 11:59:56.466499 systemd[1]: motdgen.service: Deactivated successfully.
Jan 17 11:59:56.466800 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd.
Jan 17 11:59:56.477197 extend-filesystems[1509]: Resized partition /dev/vda9
Jan 17 11:59:56.496544 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 40 scanned by (udev-worker) (1232)
Jan 17 11:59:56.496708 extend-filesystems[1546]: resize2fs 1.47.1 (20-May-2024)
Jan 17 11:59:56.511173 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks
Jan 17 11:59:56.493049 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml).
Jan 17 11:59:56.511430 tar[1530]: linux-arm64/helm
Jan 17 11:59:56.493133 systemd[1]: Reached target system-config.target - Load system-provided cloud configs.
Jan 17 11:59:56.498242 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url).
Jan 17 11:59:56.498262 systemd[1]: Reached target user-config.target - Load user-provided cloud configs.
Jan 17 11:59:56.515619 jq[1543]: true
Jan 17 11:59:56.582892 kernel: EXT4-fs (vda9): resized filesystem to 1864699
Jan 17 11:59:56.597311 update_engine[1522]: I20250117 11:59:56.577950  1522 main.cc:92] Flatcar Update Engine starting
Jan 17 11:59:56.597311 update_engine[1522]: I20250117 11:59:56.595777  1522 update_check_scheduler.cc:74] Next update check in 8m33s
Jan 17 11:59:56.597552 systemd-logind[1520]: Watching system buttons on /dev/input/event0 (Power Button)
Jan 17 11:59:56.599370 extend-filesystems[1546]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required
Jan 17 11:59:56.599370 extend-filesystems[1546]: old_desc_blocks = 1, new_desc_blocks = 1
Jan 17 11:59:56.599370 extend-filesystems[1546]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long.
Jan 17 11:59:56.608077 extend-filesystems[1509]: Resized filesystem in /dev/vda9
Jan 17 11:59:56.599599 systemd-logind[1520]: New seat seat0.
Jan 17 11:59:56.602320 systemd[1]: extend-filesystems.service: Deactivated successfully.
Jan 17 11:59:56.602569 systemd[1]: Finished extend-filesystems.service - Extend Filesystems.
Jan 17 11:59:56.616109 systemd[1]: Started systemd-logind.service - User Login Management.
Jan 17 11:59:56.618658 systemd[1]: Started update-engine.service - Update Engine.
Jan 17 11:59:56.621381 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details.
Jan 17 11:59:56.632263 systemd[1]: Started locksmithd.service - Cluster reboot manager.
Jan 17 11:59:56.635079 bash[1566]: Updated "/home/core/.ssh/authorized_keys"
Jan 17 11:59:56.644162 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition.
Jan 17 11:59:56.649475 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met.
Jan 17 11:59:56.771441 locksmithd[1575]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot"
Jan 17 11:59:56.934047 tar[1530]: linux-arm64/LICENSE
Jan 17 11:59:56.937005 tar[1530]: linux-arm64/README.md
Jan 17 11:59:56.937082 containerd[1534]: time="2025-01-17T11:59:56.935486120Z" level=info msg="starting containerd" revision=174e0d1785eeda18dc2beba45e1d5a188771636b version=v1.7.21
Jan 17 11:59:56.950589 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin.
Jan 17 11:59:56.961415 containerd[1534]: time="2025-01-17T11:59:56.961351880Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.962734960Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.71-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.962781280Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.962799760Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.962970400Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.962999160Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.963059240Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.963072200Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.963265480Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.963280600Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.963293760Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1
Jan 17 11:59:56.963938 containerd[1534]: time="2025-01-17T11:59:56.963303640Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1
Jan 17 11:59:56.964246 containerd[1534]: time="2025-01-17T11:59:56.963367840Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1
Jan 17 11:59:56.964246 containerd[1534]: time="2025-01-17T11:59:56.963545840Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1
Jan 17 11:59:56.964246 containerd[1534]: time="2025-01-17T11:59:56.963662520Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1
Jan 17 11:59:56.964246 containerd[1534]: time="2025-01-17T11:59:56.963677360Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1
Jan 17 11:59:56.964246 containerd[1534]: time="2025-01-17T11:59:56.963760720Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1
Jan 17 11:59:56.964246 containerd[1534]: time="2025-01-17T11:59:56.963804080Z" level=info msg="metadata content store policy set" policy=shared
Jan 17 11:59:56.967320 containerd[1534]: time="2025-01-17T11:59:56.967288880Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1
Jan 17 11:59:56.967447 containerd[1534]: time="2025-01-17T11:59:56.967430680Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1
Jan 17 11:59:56.967504 containerd[1534]: time="2025-01-17T11:59:56.967491800Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1
Jan 17 11:59:56.967607 containerd[1534]: time="2025-01-17T11:59:56.967592040Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1
Jan 17 11:59:56.967676 containerd[1534]: time="2025-01-17T11:59:56.967661320Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1
Jan 17 11:59:56.967880 containerd[1534]: time="2025-01-17T11:59:56.967857880Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1
Jan 17 11:59:56.968949 containerd[1534]: time="2025-01-17T11:59:56.968902520Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2
Jan 17 11:59:56.969098 containerd[1534]: time="2025-01-17T11:59:56.969078880Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2
Jan 17 11:59:56.969121 containerd[1534]: time="2025-01-17T11:59:56.969102680Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1
Jan 17 11:59:56.969159 containerd[1534]: time="2025-01-17T11:59:56.969117040Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1
Jan 17 11:59:56.969159 containerd[1534]: time="2025-01-17T11:59:56.969132480Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1
Jan 17 11:59:56.969159 containerd[1534]: time="2025-01-17T11:59:56.969145560Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1
Jan 17 11:59:56.969159 containerd[1534]: time="2025-01-17T11:59:56.969157880Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1
Jan 17 11:59:56.969224 containerd[1534]: time="2025-01-17T11:59:56.969171560Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1
Jan 17 11:59:56.969224 containerd[1534]: time="2025-01-17T11:59:56.969185960Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1
Jan 17 11:59:56.969224 containerd[1534]: time="2025-01-17T11:59:56.969199480Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1
Jan 17 11:59:56.969224 containerd[1534]: time="2025-01-17T11:59:56.969211600Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1
Jan 17 11:59:56.969224 containerd[1534]: time="2025-01-17T11:59:56.969223160Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1
Jan 17 11:59:56.969307 containerd[1534]: time="2025-01-17T11:59:56.969243520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969307 containerd[1534]: time="2025-01-17T11:59:56.969259480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969307 containerd[1534]: time="2025-01-17T11:59:56.969271880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969307 containerd[1534]: time="2025-01-17T11:59:56.969284080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969307 containerd[1534]: time="2025-01-17T11:59:56.969296840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969393 containerd[1534]: time="2025-01-17T11:59:56.969309920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969393 containerd[1534]: time="2025-01-17T11:59:56.969323080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969393 containerd[1534]: time="2025-01-17T11:59:56.969336560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969393 containerd[1534]: time="2025-01-17T11:59:56.969349080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969393 containerd[1534]: time="2025-01-17T11:59:56.969363040Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969393 containerd[1534]: time="2025-01-17T11:59:56.969375080Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969393 containerd[1534]: time="2025-01-17T11:59:56.969387120Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969511 containerd[1534]: time="2025-01-17T11:59:56.969399360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969511 containerd[1534]: time="2025-01-17T11:59:56.969415480Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1
Jan 17 11:59:56.969511 containerd[1534]: time="2025-01-17T11:59:56.969437280Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969511 containerd[1534]: time="2025-01-17T11:59:56.969449840Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969511 containerd[1534]: time="2025-01-17T11:59:56.969461160Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1
Jan 17 11:59:56.969597 containerd[1534]: time="2025-01-17T11:59:56.969584040Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1
Jan 17 11:59:56.969616 containerd[1534]: time="2025-01-17T11:59:56.969602840Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1
Jan 17 11:59:56.969636 containerd[1534]: time="2025-01-17T11:59:56.969614680Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1
Jan 17 11:59:56.969636 containerd[1534]: time="2025-01-17T11:59:56.969627480Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1
Jan 17 11:59:56.969686 containerd[1534]: time="2025-01-17T11:59:56.969637560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.969686 containerd[1534]: time="2025-01-17T11:59:56.969651480Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1
Jan 17 11:59:56.969686 containerd[1534]: time="2025-01-17T11:59:56.969661560Z" level=info msg="NRI interface is disabled by configuration."
Jan 17 11:59:56.969686 containerd[1534]: time="2025-01-17T11:59:56.969671880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1
Jan 17 11:59:56.970720 containerd[1534]: time="2025-01-17T11:59:56.970360640Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}"
Jan 17 11:59:56.970720 containerd[1534]: time="2025-01-17T11:59:56.970448680Z" level=info msg="Connect containerd service"
Jan 17 11:59:56.970720 containerd[1534]: time="2025-01-17T11:59:56.970500400Z" level=info msg="using legacy CRI server"
Jan 17 11:59:56.970720 containerd[1534]: time="2025-01-17T11:59:56.970509440Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this"
Jan 17 11:59:56.970720 containerd[1534]: time="2025-01-17T11:59:56.970620280Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\""
Jan 17 11:59:56.971874 containerd[1534]: time="2025-01-17T11:59:56.971832560Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config"
Jan 17 11:59:56.972414 containerd[1534]: time="2025-01-17T11:59:56.972320160Z" level=info msg="Start subscribing containerd event"
Jan 17 11:59:56.972414 containerd[1534]: time="2025-01-17T11:59:56.972374000Z" level=info msg="Start recovering state"
Jan 17 11:59:56.972487 containerd[1534]: time="2025-01-17T11:59:56.972440880Z" level=info msg="Start event monitor"
Jan 17 11:59:56.972487 containerd[1534]: time="2025-01-17T11:59:56.972457800Z" level=info msg="Start snapshots syncer"
Jan 17 11:59:56.972487 containerd[1534]: time="2025-01-17T11:59:56.972468800Z" level=info msg="Start cni network conf syncer for default"
Jan 17 11:59:56.972487 containerd[1534]: time="2025-01-17T11:59:56.972476400Z" level=info msg="Start streaming server"
Jan 17 11:59:56.972883 containerd[1534]: time="2025-01-17T11:59:56.972858840Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc
Jan 17 11:59:56.973045 containerd[1534]: time="2025-01-17T11:59:56.973030160Z" level=info msg=serving... address=/run/containerd/containerd.sock
Jan 17 11:59:56.973212 containerd[1534]: time="2025-01-17T11:59:56.973196040Z" level=info msg="containerd successfully booted in 0.038689s"
Jan 17 11:59:56.973320 systemd[1]: Started containerd.service - containerd container runtime.
Jan 17 11:59:56.979190 systemd-networkd[1228]: eth0: Gained IPv6LL
Jan 17 11:59:56.982394 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured.
Jan 17 11:59:56.984021 systemd[1]: Reached target network-online.target - Network is Online.
Jan 17 11:59:56.990216 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent...
Jan 17 11:59:56.992544 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jan 17 11:59:56.996100 systemd[1]: Starting nvidia.service - NVIDIA Configure Service...
Jan 17 11:59:57.024458 systemd[1]: Finished nvidia.service - NVIDIA Configure Service.
Jan 17 11:59:57.029345 systemd[1]: coreos-metadata.service: Deactivated successfully.
Jan 17 11:59:57.029587 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent.
Jan 17 11:59:57.031134 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met.
Jan 17 11:59:57.109952 sshd_keygen[1527]: ssh-keygen: generating new host keys: RSA ECDSA ED25519
Jan 17 11:59:57.131364 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys.
Jan 17 11:59:57.144258 systemd[1]: Starting issuegen.service - Generate /run/issue...
Jan 17 11:59:57.150232 systemd[1]: issuegen.service: Deactivated successfully.
Jan 17 11:59:57.150463 systemd[1]: Finished issuegen.service - Generate /run/issue.
Jan 17 11:59:57.153154 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions...
Jan 17 11:59:57.167393 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions.
Jan 17 11:59:57.170008 systemd[1]: Started getty@tty1.service - Getty on tty1.
Jan 17 11:59:57.171856 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0.
Jan 17 11:59:57.173176 systemd[1]: Reached target getty.target - Login Prompts.
Jan 17 11:59:57.495453 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jan 17 11:59:57.496695 systemd[1]: Reached target multi-user.target - Multi-User System.
Jan 17 11:59:57.499110 systemd[1]: Startup finished in 4.950s (kernel) + 3.408s (userspace) = 8.359s.
Jan 17 11:59:57.499181 (kubelet)[1643]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Jan 17 11:59:58.027656 kubelet[1643]: E0117 11:59:58.027508    1643 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Jan 17 11:59:58.030222 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Jan 17 11:59:58.030407 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Jan 17 12:00:02.962832 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd.
Jan 17 12:00:02.977423 systemd[1]: Started sshd@0-10.0.0.35:22-10.0.0.1:54554.service - OpenSSH per-connection server daemon (10.0.0.1:54554).
Jan 17 12:00:03.026385 sshd[1657]: Accepted publickey for core from 10.0.0.1 port 54554 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:00:03.028261 sshd[1657]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:00:03.040067 systemd-logind[1520]: New session 1 of user core.
Jan 17 12:00:03.040838 systemd[1]: Created slice user-500.slice - User Slice of UID 500.
Jan 17 12:00:03.050195 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500...
Jan 17 12:00:03.059276 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500.
Jan 17 12:00:03.062454 systemd[1]: Starting user@500.service - User Manager for UID 500...
Jan 17 12:00:03.068795 (systemd)[1663]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0)
Jan 17 12:00:03.139327 systemd[1663]: Queued start job for default target default.target.
Jan 17 12:00:03.140161 systemd[1663]: Created slice app.slice - User Application Slice.
Jan 17 12:00:03.140196 systemd[1663]: Reached target paths.target - Paths.
Jan 17 12:00:03.140208 systemd[1663]: Reached target timers.target - Timers.
Jan 17 12:00:03.151106 systemd[1663]: Starting dbus.socket - D-Bus User Message Bus Socket...
Jan 17 12:00:03.157495 systemd[1663]: Listening on dbus.socket - D-Bus User Message Bus Socket.
Jan 17 12:00:03.158106 systemd[1663]: Reached target sockets.target - Sockets.
Jan 17 12:00:03.158121 systemd[1663]: Reached target basic.target - Basic System.
Jan 17 12:00:03.158178 systemd[1663]: Reached target default.target - Main User Target.
Jan 17 12:00:03.158205 systemd[1663]: Startup finished in 84ms.
Jan 17 12:00:03.158390 systemd[1]: Started user@500.service - User Manager for UID 500.
Jan 17 12:00:03.160145 systemd[1]: Started session-1.scope - Session 1 of User core.
Jan 17 12:00:03.223276 systemd[1]: Started sshd@1-10.0.0.35:22-10.0.0.1:54570.service - OpenSSH per-connection server daemon (10.0.0.1:54570).
Jan 17 12:00:03.260690 sshd[1675]: Accepted publickey for core from 10.0.0.1 port 54570 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:00:03.261963 sshd[1675]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:00:03.266050 systemd-logind[1520]: New session 2 of user core.
Jan 17 12:00:03.274322 systemd[1]: Started session-2.scope - Session 2 of User core.
Jan 17 12:00:03.325736 sshd[1675]: pam_unix(sshd:session): session closed for user core
Jan 17 12:00:03.340279 systemd[1]: Started sshd@2-10.0.0.35:22-10.0.0.1:54574.service - OpenSSH per-connection server daemon (10.0.0.1:54574).
Jan 17 12:00:03.340647 systemd[1]: sshd@1-10.0.0.35:22-10.0.0.1:54570.service: Deactivated successfully.
Jan 17 12:00:03.342989 systemd[1]: session-2.scope: Deactivated successfully.
Jan 17 12:00:03.343566 systemd-logind[1520]: Session 2 logged out. Waiting for processes to exit.
Jan 17 12:00:03.344363 systemd-logind[1520]: Removed session 2.
Jan 17 12:00:03.371079 sshd[1680]: Accepted publickey for core from 10.0.0.1 port 54574 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:00:03.372213 sshd[1680]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:00:03.375703 systemd-logind[1520]: New session 3 of user core.
Jan 17 12:00:03.386195 systemd[1]: Started session-3.scope - Session 3 of User core.
Jan 17 12:00:03.433825 sshd[1680]: pam_unix(sshd:session): session closed for user core
Jan 17 12:00:03.444272 systemd[1]: Started sshd@3-10.0.0.35:22-10.0.0.1:54586.service - OpenSSH per-connection server daemon (10.0.0.1:54586).
Jan 17 12:00:03.444634 systemd[1]: sshd@2-10.0.0.35:22-10.0.0.1:54574.service: Deactivated successfully.
Jan 17 12:00:03.446417 systemd-logind[1520]: Session 3 logged out. Waiting for processes to exit.
Jan 17 12:00:03.447304 systemd[1]: session-3.scope: Deactivated successfully.
Jan 17 12:00:03.448438 systemd-logind[1520]: Removed session 3.
Jan 17 12:00:03.475053 sshd[1688]: Accepted publickey for core from 10.0.0.1 port 54586 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:00:03.476428 sshd[1688]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:00:03.480072 systemd-logind[1520]: New session 4 of user core.
Jan 17 12:00:03.489223 systemd[1]: Started session-4.scope - Session 4 of User core.
Jan 17 12:00:03.542327 sshd[1688]: pam_unix(sshd:session): session closed for user core
Jan 17 12:00:03.553285 systemd[1]: Started sshd@4-10.0.0.35:22-10.0.0.1:54592.service - OpenSSH per-connection server daemon (10.0.0.1:54592).
Jan 17 12:00:03.553641 systemd[1]: sshd@3-10.0.0.35:22-10.0.0.1:54586.service: Deactivated successfully.
Jan 17 12:00:03.555134 systemd-logind[1520]: Session 4 logged out. Waiting for processes to exit.
Jan 17 12:00:03.555798 systemd[1]: session-4.scope: Deactivated successfully.
Jan 17 12:00:03.557035 systemd-logind[1520]: Removed session 4.
Jan 17 12:00:03.585522 sshd[1696]: Accepted publickey for core from 10.0.0.1 port 54592 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:00:03.587243 sshd[1696]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:00:03.591986 systemd-logind[1520]: New session 5 of user core.
Jan 17 12:00:03.601303 systemd[1]: Started session-5.scope - Session 5 of User core.
Jan 17 12:00:03.663210 sudo[1703]:     core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh
Jan 17 12:00:03.663497 sudo[1703]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500)
Jan 17 12:00:03.965252 systemd[1]: Starting docker.service - Docker Application Container Engine...
Jan 17 12:00:03.967269 (dockerd)[1721]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU
Jan 17 12:00:04.241111 dockerd[1721]: time="2025-01-17T12:00:04.240894180Z" level=info msg="Starting up"
Jan 17 12:00:04.316697 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport2780480830-merged.mount: Deactivated successfully.
Jan 17 12:00:04.488920 dockerd[1721]: time="2025-01-17T12:00:04.488664301Z" level=info msg="Loading containers: start."
Jan 17 12:00:04.584024 kernel: Initializing XFRM netlink socket
Jan 17 12:00:04.655433 systemd-networkd[1228]: docker0: Link UP
Jan 17 12:00:04.673405 dockerd[1721]: time="2025-01-17T12:00:04.673369626Z" level=info msg="Loading containers: done."
Jan 17 12:00:04.688935 dockerd[1721]: time="2025-01-17T12:00:04.688875898Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2
Jan 17 12:00:04.689087 dockerd[1721]: time="2025-01-17T12:00:04.688979853Z" level=info msg="Docker daemon" commit=061aa95809be396a6b5542618d8a34b02a21ff77 containerd-snapshotter=false storage-driver=overlay2 version=26.1.0
Jan 17 12:00:04.689128 dockerd[1721]: time="2025-01-17T12:00:04.689107769Z" level=info msg="Daemon has completed initialization"
Jan 17 12:00:04.720358 dockerd[1721]: time="2025-01-17T12:00:04.720191914Z" level=info msg="API listen on /run/docker.sock"
Jan 17 12:00:04.720444 systemd[1]: Started docker.service - Docker Application Container Engine.
Jan 17 12:00:05.371937 containerd[1534]: time="2025-01-17T12:00:05.371904504Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.29.13\""
Jan 17 12:00:06.224007 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4260803225.mount: Deactivated successfully.
Jan 17 12:00:07.107610 containerd[1534]: time="2025-01-17T12:00:07.107466928Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.29.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:07.108450 containerd[1534]: time="2025-01-17T12:00:07.108165635Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.29.13: active requests=0, bytes read=32202459"
Jan 17 12:00:07.109144 containerd[1534]: time="2025-01-17T12:00:07.109097003Z" level=info msg="ImageCreate event name:\"sha256:5c8d3b261565d9e15723d572fb33e6ec92ceb342312c9418457857eb57d1ae9a\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:07.111973 containerd[1534]: time="2025-01-17T12:00:07.111946810Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:e5c42861045d0615769fad8a4e32e476fc5e59020157b60ced1bb7a69d4a5ce9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:07.113466 containerd[1534]: time="2025-01-17T12:00:07.113220468Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.29.13\" with image id \"sha256:5c8d3b261565d9e15723d572fb33e6ec92ceb342312c9418457857eb57d1ae9a\", repo tag \"registry.k8s.io/kube-apiserver:v1.29.13\", repo digest \"registry.k8s.io/kube-apiserver@sha256:e5c42861045d0615769fad8a4e32e476fc5e59020157b60ced1bb7a69d4a5ce9\", size \"32199257\" in 1.741279076s"
Jan 17 12:00:07.113466 containerd[1534]: time="2025-01-17T12:00:07.113259138Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.29.13\" returns image reference \"sha256:5c8d3b261565d9e15723d572fb33e6ec92ceb342312c9418457857eb57d1ae9a\""
Jan 17 12:00:07.131382 containerd[1534]: time="2025-01-17T12:00:07.131351348Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.29.13\""
Jan 17 12:00:08.280666 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1.
Jan 17 12:00:08.294242 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jan 17 12:00:08.379539 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jan 17 12:00:08.383561 (kubelet)[1953]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS
Jan 17 12:00:08.480130 containerd[1534]: time="2025-01-17T12:00:08.479454393Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.29.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:08.480704 containerd[1534]: time="2025-01-17T12:00:08.480678352Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.29.13: active requests=0, bytes read=29381104"
Jan 17 12:00:08.481837 containerd[1534]: time="2025-01-17T12:00:08.481803893Z" level=info msg="ImageCreate event name:\"sha256:bcc4e3c2095eb1aad9487d6679a8871f05390959aaf5091f391510033742cf7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:08.486007 containerd[1534]: time="2025-01-17T12:00:08.484581055Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:fc2838399752740bdd36c7e9287d4406feff6bef2baff393174b34ccd447b780\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:08.486242 containerd[1534]: time="2025-01-17T12:00:08.486185584Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.29.13\" with image id \"sha256:bcc4e3c2095eb1aad9487d6679a8871f05390959aaf5091f391510033742cf7c\", repo tag \"registry.k8s.io/kube-controller-manager:v1.29.13\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:fc2838399752740bdd36c7e9287d4406feff6bef2baff393174b34ccd447b780\", size \"30784892\" in 1.35479715s"
Jan 17 12:00:08.486242 containerd[1534]: time="2025-01-17T12:00:08.486237313Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.29.13\" returns image reference \"sha256:bcc4e3c2095eb1aad9487d6679a8871f05390959aaf5091f391510033742cf7c\""
Jan 17 12:00:08.505852 containerd[1534]: time="2025-01-17T12:00:08.505810822Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.29.13\""
Jan 17 12:00:08.515951 kubelet[1953]: E0117 12:00:08.515903    1953 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory"
Jan 17 12:00:08.520242 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE
Jan 17 12:00:08.520422 systemd[1]: kubelet.service: Failed with result 'exit-code'.
Jan 17 12:00:09.465900 containerd[1534]: time="2025-01-17T12:00:09.465848467Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.29.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:09.466766 containerd[1534]: time="2025-01-17T12:00:09.466700408Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.29.13: active requests=0, bytes read=15765674"
Jan 17 12:00:09.467369 containerd[1534]: time="2025-01-17T12:00:09.467304190Z" level=info msg="ImageCreate event name:\"sha256:09e2786faf24867b706964cc8c35c296f197dc7a57806a75388efa13868bf50c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:09.470252 containerd[1534]: time="2025-01-17T12:00:09.470216402Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:a4f1649a5249c0784963d85644b1e614548f032da9b4fb00a760bac02818ce4f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:09.471508 containerd[1534]: time="2025-01-17T12:00:09.471457679Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.29.13\" with image id \"sha256:09e2786faf24867b706964cc8c35c296f197dc7a57806a75388efa13868bf50c\", repo tag \"registry.k8s.io/kube-scheduler:v1.29.13\", repo digest \"registry.k8s.io/kube-scheduler@sha256:a4f1649a5249c0784963d85644b1e614548f032da9b4fb00a760bac02818ce4f\", size \"17169480\" in 965.603604ms"
Jan 17 12:00:09.471508 containerd[1534]: time="2025-01-17T12:00:09.471495219Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.29.13\" returns image reference \"sha256:09e2786faf24867b706964cc8c35c296f197dc7a57806a75388efa13868bf50c\""
Jan 17 12:00:09.490355 containerd[1534]: time="2025-01-17T12:00:09.490310542Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.13\""
Jan 17 12:00:10.412172 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3105220296.mount: Deactivated successfully.
Jan 17 12:00:10.721793 containerd[1534]: time="2025-01-17T12:00:10.721660353Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.29.13\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:10.722641 containerd[1534]: time="2025-01-17T12:00:10.722516989Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.29.13: active requests=0, bytes read=25274684"
Jan 17 12:00:10.723581 containerd[1534]: time="2025-01-17T12:00:10.723536936Z" level=info msg="ImageCreate event name:\"sha256:e3bc26919d7c787204f912c4bc2584bac5686761ae4da96585475c68dcc57181\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:10.725225 containerd[1534]: time="2025-01-17T12:00:10.725164086Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:dd45846de733434501e436638a7a240f2d379bf0a6bb0404a7684e0cf52c4011\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:10.725996 containerd[1534]: time="2025-01-17T12:00:10.725952871Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.29.13\" with image id \"sha256:e3bc26919d7c787204f912c4bc2584bac5686761ae4da96585475c68dcc57181\", repo tag \"registry.k8s.io/kube-proxy:v1.29.13\", repo digest \"registry.k8s.io/kube-proxy@sha256:dd45846de733434501e436638a7a240f2d379bf0a6bb0404a7684e0cf52c4011\", size \"25273701\" in 1.23560254s"
Jan 17 12:00:10.726047 containerd[1534]: time="2025-01-17T12:00:10.726003095Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.13\" returns image reference \"sha256:e3bc26919d7c787204f912c4bc2584bac5686761ae4da96585475c68dcc57181\""
Jan 17 12:00:10.744840 containerd[1534]: time="2025-01-17T12:00:10.744713913Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\""
Jan 17 12:00:11.298240 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3808100510.mount: Deactivated successfully.
Jan 17 12:00:11.872928 containerd[1534]: time="2025-01-17T12:00:11.872860392Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:11.873634 containerd[1534]: time="2025-01-17T12:00:11.873597347Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485383"
Jan 17 12:00:11.874621 containerd[1534]: time="2025-01-17T12:00:11.874570958Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:11.877732 containerd[1534]: time="2025-01-17T12:00:11.877671749Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:11.879025 containerd[1534]: time="2025-01-17T12:00:11.878969884Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 1.134218229s"
Jan 17 12:00:11.879091 containerd[1534]: time="2025-01-17T12:00:11.879024333Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\""
Jan 17 12:00:11.898127 containerd[1534]: time="2025-01-17T12:00:11.898093141Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\""
Jan 17 12:00:12.308745 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2915051679.mount: Deactivated successfully.
Jan 17 12:00:12.312953 containerd[1534]: time="2025-01-17T12:00:12.312903108Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:12.313412 containerd[1534]: time="2025-01-17T12:00:12.313371687Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=268823"
Jan 17 12:00:12.314318 containerd[1534]: time="2025-01-17T12:00:12.314282764Z" level=info msg="ImageCreate event name:\"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:12.316460 containerd[1534]: time="2025-01-17T12:00:12.316425811Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:12.318067 containerd[1534]: time="2025-01-17T12:00:12.318025451Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"268051\" in 419.89343ms"
Jan 17 12:00:12.318110 containerd[1534]: time="2025-01-17T12:00:12.318066281Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\""
Jan 17 12:00:12.336060 containerd[1534]: time="2025-01-17T12:00:12.336019658Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\""
Jan 17 12:00:12.868615 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount871216942.mount: Deactivated successfully.
Jan 17 12:00:14.125527 containerd[1534]: time="2025-01-17T12:00:14.125476762Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.10-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:14.126797 containerd[1534]: time="2025-01-17T12:00:14.126754838Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.10-0: active requests=0, bytes read=65200788"
Jan 17 12:00:14.127550 containerd[1534]: time="2025-01-17T12:00:14.127494113Z" level=info msg="ImageCreate event name:\"sha256:79f8d13ae8b8839cadfb2f83416935f5184206d386028e2d1263577f0ab3620b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:14.130663 containerd[1534]: time="2025-01-17T12:00:14.130612188Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:14.131956 containerd[1534]: time="2025-01-17T12:00:14.131928601Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.10-0\" with image id \"sha256:79f8d13ae8b8839cadfb2f83416935f5184206d386028e2d1263577f0ab3620b\", repo tag \"registry.k8s.io/etcd:3.5.10-0\", repo digest \"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\", size \"65198393\" in 1.795869608s"
Jan 17 12:00:14.132024 containerd[1534]: time="2025-01-17T12:00:14.131957223Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\" returns image reference \"sha256:79f8d13ae8b8839cadfb2f83416935f5184206d386028e2d1263577f0ab3620b\""
Jan 17 12:00:17.793378 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Jan 17 12:00:17.804281 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jan 17 12:00:17.818418 systemd[1]: Reloading requested from client PID 2180 ('systemctl') (unit session-5.scope)...
Jan 17 12:00:17.818436 systemd[1]: Reloading...
Jan 17 12:00:17.885298 zram_generator::config[2220]: No configuration found.
Jan 17 12:00:18.009557 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jan 17 12:00:18.058119 systemd[1]: Reloading finished in 239 ms.
Jan 17 12:00:18.100492 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM
Jan 17 12:00:18.100565 systemd[1]: kubelet.service: Failed with result 'signal'.
Jan 17 12:00:18.100817 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Jan 17 12:00:18.102515 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jan 17 12:00:18.188685 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jan 17 12:00:18.192193 (kubelet)[2276]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS
Jan 17 12:00:18.233219 kubelet[2276]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jan 17 12:00:18.233219 kubelet[2276]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI.
Jan 17 12:00:18.233219 kubelet[2276]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jan 17 12:00:18.233590 kubelet[2276]: I0117 12:00:18.233259    2276 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Jan 17 12:00:18.755915 kubelet[2276]: I0117 12:00:18.754687    2276 server.go:487] "Kubelet version" kubeletVersion="v1.29.2"
Jan 17 12:00:18.755915 kubelet[2276]: I0117 12:00:18.754722    2276 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Jan 17 12:00:18.755915 kubelet[2276]: I0117 12:00:18.754922    2276 server.go:919] "Client rotation is on, will bootstrap in background"
Jan 17 12:00:18.785250 kubelet[2276]: I0117 12:00:18.785210    2276 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt"
Jan 17 12:00:18.786321 kubelet[2276]: E0117 12:00:18.786294    2276 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.35:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:18.795990 kubelet[2276]: I0117 12:00:18.795962    2276 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Jan 17 12:00:18.796854 kubelet[2276]: I0117 12:00:18.796819    2276 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Jan 17 12:00:18.797046 kubelet[2276]: I0117 12:00:18.797023    2276 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null}
Jan 17 12:00:18.797046 kubelet[2276]: I0117 12:00:18.797047    2276 topology_manager.go:138] "Creating topology manager with none policy"
Jan 17 12:00:18.797159 kubelet[2276]: I0117 12:00:18.797056    2276 container_manager_linux.go:301] "Creating device plugin manager"
Jan 17 12:00:18.798314 kubelet[2276]: I0117 12:00:18.798274    2276 state_mem.go:36] "Initialized new in-memory state store"
Jan 17 12:00:18.802449 kubelet[2276]: I0117 12:00:18.802424    2276 kubelet.go:396] "Attempting to sync node with API server"
Jan 17 12:00:18.802449 kubelet[2276]: I0117 12:00:18.802449    2276 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests"
Jan 17 12:00:18.802514 kubelet[2276]: I0117 12:00:18.802469    2276 kubelet.go:312] "Adding apiserver pod source"
Jan 17 12:00:18.802514 kubelet[2276]: I0117 12:00:18.802482    2276 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Jan 17 12:00:18.803105 kubelet[2276]: W0117 12:00:18.803045    2276 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: Get "https://10.0.0.35:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:18.803105 kubelet[2276]: E0117 12:00:18.803088    2276 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.35:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:18.803546 kubelet[2276]: W0117 12:00:18.803482    2276 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: Get "https://10.0.0.35:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:18.803546 kubelet[2276]: E0117 12:00:18.803517    2276 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.35:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:18.806846 kubelet[2276]: I0117 12:00:18.806805    2276 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1"
Jan 17 12:00:18.807354 kubelet[2276]: I0117 12:00:18.807336    2276 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode"
Jan 17 12:00:18.807460 kubelet[2276]: W0117 12:00:18.807449    2276 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating.
Jan 17 12:00:18.808199 kubelet[2276]: I0117 12:00:18.808185    2276 server.go:1256] "Started kubelet"
Jan 17 12:00:18.809024 kubelet[2276]: I0117 12:00:18.808307    2276 server.go:162] "Starting to listen" address="0.0.0.0" port=10250
Jan 17 12:00:18.809084 kubelet[2276]: I0117 12:00:18.809055    2276 server.go:461] "Adding debug handlers to kubelet server"
Jan 17 12:00:18.811795 kubelet[2276]: I0117 12:00:18.808314    2276 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10
Jan 17 12:00:18.811795 kubelet[2276]: I0117 12:00:18.810045    2276 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock"
Jan 17 12:00:18.813696 kubelet[2276]: I0117 12:00:18.812049    2276 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Jan 17 12:00:18.820354 kubelet[2276]: I0117 12:00:18.819585    2276 volume_manager.go:291] "Starting Kubelet Volume Manager"
Jan 17 12:00:18.820354 kubelet[2276]: I0117 12:00:18.819663    2276 desired_state_of_world_populator.go:151] "Desired state populator starts to run"
Jan 17 12:00:18.820354 kubelet[2276]: I0117 12:00:18.819714    2276 reconciler_new.go:29] "Reconciler: start to sync state"
Jan 17 12:00:18.820354 kubelet[2276]: W0117 12:00:18.819969    2276 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: Get "https://10.0.0.35:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:18.820354 kubelet[2276]: E0117 12:00:18.820023    2276 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.35:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:18.820354 kubelet[2276]: E0117 12:00:18.820307    2276 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.35:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.35:6443: connect: connection refused" interval="200ms"
Jan 17 12:00:18.821014 kubelet[2276]: I0117 12:00:18.820872    2276 factory.go:221] Registration of the systemd container factory successfully
Jan 17 12:00:18.821014 kubelet[2276]: I0117 12:00:18.820968    2276 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory
Jan 17 12:00:18.821462 kubelet[2276]: E0117 12:00:18.821441    2276 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem"
Jan 17 12:00:18.822155 kubelet[2276]: I0117 12:00:18.822138    2276 factory.go:221] Registration of the containerd container factory successfully
Jan 17 12:00:18.826729 kubelet[2276]: E0117 12:00:18.826008    2276 event.go:355] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.35:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.35:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.181b79167efc42c8  default    0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-01-17 12:00:18.808160968 +0000 UTC m=+0.612730537,LastTimestamp:2025-01-17 12:00:18.808160968 +0000 UTC m=+0.612730537,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}"
Jan 17 12:00:18.837100 kubelet[2276]: I0117 12:00:18.836916    2276 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4"
Jan 17 12:00:18.837947 kubelet[2276]: I0117 12:00:18.837919    2276 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6"
Jan 17 12:00:18.837947 kubelet[2276]: I0117 12:00:18.837947    2276 status_manager.go:217] "Starting to sync pod status with apiserver"
Jan 17 12:00:18.838057 kubelet[2276]: I0117 12:00:18.837965    2276 kubelet.go:2329] "Starting kubelet main sync loop"
Jan 17 12:00:18.838079 kubelet[2276]: E0117 12:00:18.838063    2276 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]"
Jan 17 12:00:18.839571 kubelet[2276]: W0117 12:00:18.839172    2276 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.RuntimeClass: Get "https://10.0.0.35:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:18.839571 kubelet[2276]: E0117 12:00:18.839222    2276 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.35:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:18.840553 kubelet[2276]: I0117 12:00:18.840528    2276 cpu_manager.go:214] "Starting CPU manager" policy="none"
Jan 17 12:00:18.840553 kubelet[2276]: I0117 12:00:18.840546    2276 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s"
Jan 17 12:00:18.840659 kubelet[2276]: I0117 12:00:18.840562    2276 state_mem.go:36] "Initialized new in-memory state store"
Jan 17 12:00:18.913905 kubelet[2276]: I0117 12:00:18.913870    2276 policy_none.go:49] "None policy: Start"
Jan 17 12:00:18.914637 kubelet[2276]: I0117 12:00:18.914612    2276 memory_manager.go:170] "Starting memorymanager" policy="None"
Jan 17 12:00:18.914700 kubelet[2276]: I0117 12:00:18.914689    2276 state_mem.go:35] "Initializing new in-memory state store"
Jan 17 12:00:18.921414 kubelet[2276]: I0117 12:00:18.920823    2276 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found"
Jan 17 12:00:18.921414 kubelet[2276]: I0117 12:00:18.921078    2276 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Jan 17 12:00:18.921515 kubelet[2276]: I0117 12:00:18.921493    2276 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jan 17 12:00:18.922808 kubelet[2276]: E0117 12:00:18.922788    2276 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found"
Jan 17 12:00:18.923718 kubelet[2276]: E0117 12:00:18.923680    2276 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.35:6443/api/v1/nodes\": dial tcp 10.0.0.35:6443: connect: connection refused" node="localhost"
Jan 17 12:00:18.938995 kubelet[2276]: I0117 12:00:18.938952    2276 topology_manager.go:215] "Topology Admit Handler" podUID="3e2527a0b135d137a07d61e8f70971ec" podNamespace="kube-system" podName="kube-apiserver-localhost"
Jan 17 12:00:18.940447 kubelet[2276]: I0117 12:00:18.940403    2276 topology_manager.go:215] "Topology Admit Handler" podUID="dd466de870bdf0e573d7965dbd759acf" podNamespace="kube-system" podName="kube-controller-manager-localhost"
Jan 17 12:00:18.941226 kubelet[2276]: I0117 12:00:18.941210    2276 topology_manager.go:215] "Topology Admit Handler" podUID="605dd245551545e29d4e79fb03fd341e" podNamespace="kube-system" podName="kube-scheduler-localhost"
Jan 17 12:00:19.022102 kubelet[2276]: I0117 12:00:19.021201    2276 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:19.022102 kubelet[2276]: I0117 12:00:19.021275    2276 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:19.022102 kubelet[2276]: I0117 12:00:19.021300    2276 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:19.022102 kubelet[2276]: I0117 12:00:19.021329    2276 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/605dd245551545e29d4e79fb03fd341e-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"605dd245551545e29d4e79fb03fd341e\") " pod="kube-system/kube-scheduler-localhost"
Jan 17 12:00:19.022102 kubelet[2276]: I0117 12:00:19.021375    2276 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3e2527a0b135d137a07d61e8f70971ec-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"3e2527a0b135d137a07d61e8f70971ec\") " pod="kube-system/kube-apiserver-localhost"
Jan 17 12:00:19.022294 kubelet[2276]: E0117 12:00:19.021396    2276 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.35:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.35:6443: connect: connection refused" interval="400ms"
Jan 17 12:00:19.022294 kubelet[2276]: I0117 12:00:19.021414    2276 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3e2527a0b135d137a07d61e8f70971ec-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"3e2527a0b135d137a07d61e8f70971ec\") " pod="kube-system/kube-apiserver-localhost"
Jan 17 12:00:19.022294 kubelet[2276]: I0117 12:00:19.021455    2276 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3e2527a0b135d137a07d61e8f70971ec-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"3e2527a0b135d137a07d61e8f70971ec\") " pod="kube-system/kube-apiserver-localhost"
Jan 17 12:00:19.022294 kubelet[2276]: I0117 12:00:19.021493    2276 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:19.022294 kubelet[2276]: I0117 12:00:19.021526    2276 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:19.125891 kubelet[2276]: I0117 12:00:19.125865    2276 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jan 17 12:00:19.126281 kubelet[2276]: E0117 12:00:19.126235    2276 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.35:6443/api/v1/nodes\": dial tcp 10.0.0.35:6443: connect: connection refused" node="localhost"
Jan 17 12:00:19.249409 kubelet[2276]: E0117 12:00:19.249297    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:19.249409 kubelet[2276]: E0117 12:00:19.249346    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:19.249791 kubelet[2276]: E0117 12:00:19.249754    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:19.250464 containerd[1534]: time="2025-01-17T12:00:19.250077737Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:605dd245551545e29d4e79fb03fd341e,Namespace:kube-system,Attempt:0,}"
Jan 17 12:00:19.250464 containerd[1534]: time="2025-01-17T12:00:19.250164576Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:dd466de870bdf0e573d7965dbd759acf,Namespace:kube-system,Attempt:0,}"
Jan 17 12:00:19.250464 containerd[1534]: time="2025-01-17T12:00:19.250312607Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:3e2527a0b135d137a07d61e8f70971ec,Namespace:kube-system,Attempt:0,}"
Jan 17 12:00:19.422020 kubelet[2276]: E0117 12:00:19.421946    2276 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.35:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.35:6443: connect: connection refused" interval="800ms"
Jan 17 12:00:19.527988 kubelet[2276]: I0117 12:00:19.527931    2276 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jan 17 12:00:19.528676 kubelet[2276]: E0117 12:00:19.528652    2276 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.0.0.35:6443/api/v1/nodes\": dial tcp 10.0.0.35:6443: connect: connection refused" node="localhost"
Jan 17 12:00:19.660895 kubelet[2276]: W0117 12:00:19.660831    2276 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: Get "https://10.0.0.35:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:19.660895 kubelet[2276]: E0117 12:00:19.660891    2276 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.35:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:19.692896 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2738618302.mount: Deactivated successfully.
Jan 17 12:00:19.699594 containerd[1534]: time="2025-01-17T12:00:19.699232456Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jan 17 12:00:19.701196 containerd[1534]: time="2025-01-17T12:00:19.701138467Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0"
Jan 17 12:00:19.701922 containerd[1534]: time="2025-01-17T12:00:19.701867201Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jan 17 12:00:19.702818 containerd[1534]: time="2025-01-17T12:00:19.702788449Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jan 17 12:00:19.703930 containerd[1534]: time="2025-01-17T12:00:19.703865421Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jan 17 12:00:19.704542 containerd[1534]: time="2025-01-17T12:00:19.704471251Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0"
Jan 17 12:00:19.705070 containerd[1534]: time="2025-01-17T12:00:19.705014405Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269175"
Jan 17 12:00:19.707725 containerd[1534]: time="2025-01-17T12:00:19.707692310Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}"
Jan 17 12:00:19.710014 containerd[1534]: time="2025-01-17T12:00:19.709761500Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 459.538698ms"
Jan 17 12:00:19.710400 containerd[1534]: time="2025-01-17T12:00:19.710370455Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 460.026351ms"
Jan 17 12:00:19.711091 containerd[1534]: time="2025-01-17T12:00:19.711066650Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 460.905561ms"
Jan 17 12:00:19.822219 kubelet[2276]: W0117 12:00:19.822155    2276 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: Get "https://10.0.0.35:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:19.822219 kubelet[2276]: E0117 12:00:19.822214    2276 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.35:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:19.850383 containerd[1534]: time="2025-01-17T12:00:19.850237786Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jan 17 12:00:19.850383 containerd[1534]: time="2025-01-17T12:00:19.850282949Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jan 17 12:00:19.850383 containerd[1534]: time="2025-01-17T12:00:19.850297856Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:19.850383 containerd[1534]: time="2025-01-17T12:00:19.850110833Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jan 17 12:00:19.850383 containerd[1534]: time="2025-01-17T12:00:19.850190660Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jan 17 12:00:19.850383 containerd[1534]: time="2025-01-17T12:00:19.850202041Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:19.850383 containerd[1534]: time="2025-01-17T12:00:19.850290282Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:19.851182 containerd[1534]: time="2025-01-17T12:00:19.851017735Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:19.851956 containerd[1534]: time="2025-01-17T12:00:19.851858875Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jan 17 12:00:19.851956 containerd[1534]: time="2025-01-17T12:00:19.851913976Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jan 17 12:00:19.852233 containerd[1534]: time="2025-01-17T12:00:19.851939342Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:19.852683 containerd[1534]: time="2025-01-17T12:00:19.852610251Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:19.902830 containerd[1534]: time="2025-01-17T12:00:19.902709890Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:3e2527a0b135d137a07d61e8f70971ec,Namespace:kube-system,Attempt:0,} returns sandbox id \"d239cf6999589b02c949093ef0e43b8536ff712fd11b957569c16f60ae044bfe\""
Jan 17 12:00:19.908065 kubelet[2276]: E0117 12:00:19.905728    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:19.908122 containerd[1534]: time="2025-01-17T12:00:19.907900397Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:dd466de870bdf0e573d7965dbd759acf,Namespace:kube-system,Attempt:0,} returns sandbox id \"1ddbfa6459b9cb2581799b31297f9fc7df45a6b25b00368c848e109afdebda84\""
Jan 17 12:00:19.913347 kubelet[2276]: E0117 12:00:19.908573    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:19.913410 containerd[1534]: time="2025-01-17T12:00:19.910299951Z" level=info msg="CreateContainer within sandbox \"1ddbfa6459b9cb2581799b31297f9fc7df45a6b25b00368c848e109afdebda84\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}"
Jan 17 12:00:19.913410 containerd[1534]: time="2025-01-17T12:00:19.910449505Z" level=info msg="CreateContainer within sandbox \"d239cf6999589b02c949093ef0e43b8536ff712fd11b957569c16f60ae044bfe\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}"
Jan 17 12:00:19.913604 containerd[1534]: time="2025-01-17T12:00:19.913571223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:605dd245551545e29d4e79fb03fd341e,Namespace:kube-system,Attempt:0,} returns sandbox id \"63d2748ffbe69ab79315479ee179cf1e072d3f4f9248c43714d54ae32547f692\""
Jan 17 12:00:19.914333 kubelet[2276]: E0117 12:00:19.914313    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:19.915707 containerd[1534]: time="2025-01-17T12:00:19.915680927Z" level=info msg="CreateContainer within sandbox \"63d2748ffbe69ab79315479ee179cf1e072d3f4f9248c43714d54ae32547f692\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}"
Jan 17 12:00:19.929844 containerd[1534]: time="2025-01-17T12:00:19.929800026Z" level=info msg="CreateContainer within sandbox \"1ddbfa6459b9cb2581799b31297f9fc7df45a6b25b00368c848e109afdebda84\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"e2725b766c76b2c433db7804b55b5396476f4bc50005f28b4a303946f6aa86c3\""
Jan 17 12:00:19.930795 containerd[1534]: time="2025-01-17T12:00:19.930766917Z" level=info msg="StartContainer for \"e2725b766c76b2c433db7804b55b5396476f4bc50005f28b4a303946f6aa86c3\""
Jan 17 12:00:19.934233 containerd[1534]: time="2025-01-17T12:00:19.934193673Z" level=info msg="CreateContainer within sandbox \"d239cf6999589b02c949093ef0e43b8536ff712fd11b957569c16f60ae044bfe\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"c03be60ed13873dbdddeaa765e212bbe6c749335a782a34ca44c1892fef4a885\""
Jan 17 12:00:19.935230 containerd[1534]: time="2025-01-17T12:00:19.935200998Z" level=info msg="StartContainer for \"c03be60ed13873dbdddeaa765e212bbe6c749335a782a34ca44c1892fef4a885\""
Jan 17 12:00:19.938965 containerd[1534]: time="2025-01-17T12:00:19.938927183Z" level=info msg="CreateContainer within sandbox \"63d2748ffbe69ab79315479ee179cf1e072d3f4f9248c43714d54ae32547f692\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"5cd68a650bbe3567094a974fb98b7b51a1a5dfbf5692570cd925d637cb4ffdde\""
Jan 17 12:00:19.939400 containerd[1534]: time="2025-01-17T12:00:19.939382096Z" level=info msg="StartContainer for \"5cd68a650bbe3567094a974fb98b7b51a1a5dfbf5692570cd925d637cb4ffdde\""
Jan 17 12:00:19.994963 kubelet[2276]: W0117 12:00:19.994294    2276 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.RuntimeClass: Get "https://10.0.0.35:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:19.994963 kubelet[2276]: E0117 12:00:19.994448    2276 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.35:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:20.056273 containerd[1534]: time="2025-01-17T12:00:20.053615559Z" level=info msg="StartContainer for \"e2725b766c76b2c433db7804b55b5396476f4bc50005f28b4a303946f6aa86c3\" returns successfully"
Jan 17 12:00:20.056273 containerd[1534]: time="2025-01-17T12:00:20.053759350Z" level=info msg="StartContainer for \"c03be60ed13873dbdddeaa765e212bbe6c749335a782a34ca44c1892fef4a885\" returns successfully"
Jan 17 12:00:20.056273 containerd[1534]: time="2025-01-17T12:00:20.053785752Z" level=info msg="StartContainer for \"5cd68a650bbe3567094a974fb98b7b51a1a5dfbf5692570cd925d637cb4ffdde\" returns successfully"
Jan 17 12:00:20.099082 kubelet[2276]: W0117 12:00:20.098923    2276 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: Get "https://10.0.0.35:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:20.099082 kubelet[2276]: E0117 12:00:20.099041    2276 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.35:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.35:6443: connect: connection refused
Jan 17 12:00:20.100000 kubelet[2276]: E0117 12:00:20.099420    2276 event.go:355] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.35:6443/api/v1/namespaces/default/events\": dial tcp 10.0.0.35:6443: connect: connection refused" event="&Event{ObjectMeta:{localhost.181b79167efc42c8  default    0 0001-01-01 00:00:00 +0000 UTC <nil> <nil> map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:localhost,UID:localhost,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:localhost,},FirstTimestamp:2025-01-17 12:00:18.808160968 +0000 UTC m=+0.612730537,LastTimestamp:2025-01-17 12:00:18.808160968 +0000 UTC m=+0.612730537,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:localhost,}"
Jan 17 12:00:20.330338 kubelet[2276]: I0117 12:00:20.330004    2276 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jan 17 12:00:20.846322 kubelet[2276]: E0117 12:00:20.846293    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:20.850252 kubelet[2276]: E0117 12:00:20.850152    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:20.851266 kubelet[2276]: E0117 12:00:20.851166    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:21.584169 kubelet[2276]: I0117 12:00:21.584125    2276 kubelet_node_status.go:76] "Successfully registered node" node="localhost"
Jan 17 12:00:21.611114 kubelet[2276]: E0117 12:00:21.611075    2276 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found"
Jan 17 12:00:21.711775 kubelet[2276]: E0117 12:00:21.711732    2276 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found"
Jan 17 12:00:21.812598 kubelet[2276]: E0117 12:00:21.812544    2276 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found"
Jan 17 12:00:21.853698 kubelet[2276]: E0117 12:00:21.853590    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:21.913165 kubelet[2276]: E0117 12:00:21.913119    2276 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found"
Jan 17 12:00:22.013884 kubelet[2276]: E0117 12:00:22.013847    2276 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found"
Jan 17 12:00:22.114479 kubelet[2276]: E0117 12:00:22.114366    2276 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found"
Jan 17 12:00:22.214950 kubelet[2276]: E0117 12:00:22.214887    2276 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found"
Jan 17 12:00:22.315889 kubelet[2276]: E0117 12:00:22.315843    2276 kubelet_node_status.go:462] "Error getting the current node from lister" err="node \"localhost\" not found"
Jan 17 12:00:22.805326 kubelet[2276]: I0117 12:00:22.805289    2276 apiserver.go:52] "Watching apiserver"
Jan 17 12:00:22.820638 kubelet[2276]: I0117 12:00:22.820591    2276 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world"
Jan 17 12:00:24.146414 kubelet[2276]: E0117 12:00:24.146385    2276 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:24.318795 systemd[1]: Reloading requested from client PID 2554 ('systemctl') (unit session-5.scope)...
Jan 17 12:00:24.318809 systemd[1]: Reloading...
Jan 17 12:00:24.377036 zram_generator::config[2593]: No configuration found.
Jan 17 12:00:24.544069 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly.
Jan 17 12:00:24.600073 systemd[1]: Reloading finished in 280 ms.
Jan 17 12:00:24.624798 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent...
Jan 17 12:00:24.637188 systemd[1]: kubelet.service: Deactivated successfully.
Jan 17 12:00:24.637447 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent.
Jan 17 12:00:24.648312 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent...
Jan 17 12:00:24.737643 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent.
Jan 17 12:00:24.741353 (kubelet)[2645]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS
Jan 17 12:00:24.780437 kubelet[2645]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jan 17 12:00:24.780437 kubelet[2645]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI.
Jan 17 12:00:24.780437 kubelet[2645]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information.
Jan 17 12:00:24.780781 kubelet[2645]: I0117 12:00:24.780476    2645 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime"
Jan 17 12:00:24.784489 kubelet[2645]: I0117 12:00:24.784461    2645 server.go:487] "Kubelet version" kubeletVersion="v1.29.2"
Jan 17 12:00:24.784573 kubelet[2645]: I0117 12:00:24.784525    2645 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK=""
Jan 17 12:00:24.784777 kubelet[2645]: I0117 12:00:24.784763    2645 server.go:919] "Client rotation is on, will bootstrap in background"
Jan 17 12:00:24.786614 kubelet[2645]: I0117 12:00:24.786524    2645 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem".
Jan 17 12:00:24.788958 kubelet[2645]: I0117 12:00:24.788923    2645 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt"
Jan 17 12:00:24.796708 kubelet[2645]: I0117 12:00:24.796594    2645 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified.  defaulting to /"
Jan 17 12:00:24.797515 kubelet[2645]: I0117 12:00:24.797275    2645 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[]
Jan 17 12:00:24.797661 kubelet[2645]: I0117 12:00:24.797616    2645 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null}
Jan 17 12:00:24.797661 kubelet[2645]: I0117 12:00:24.797652    2645 topology_manager.go:138] "Creating topology manager with none policy"
Jan 17 12:00:24.797661 kubelet[2645]: I0117 12:00:24.797661    2645 container_manager_linux.go:301] "Creating device plugin manager"
Jan 17 12:00:24.797800 kubelet[2645]: I0117 12:00:24.797698    2645 state_mem.go:36] "Initialized new in-memory state store"
Jan 17 12:00:24.798477 kubelet[2645]: I0117 12:00:24.798442    2645 kubelet.go:396] "Attempting to sync node with API server"
Jan 17 12:00:24.798477 kubelet[2645]: I0117 12:00:24.798480    2645 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests"
Jan 17 12:00:24.798546 kubelet[2645]: I0117 12:00:24.798513    2645 kubelet.go:312] "Adding apiserver pod source"
Jan 17 12:00:24.798546 kubelet[2645]: I0117 12:00:24.798528    2645 apiserver.go:42] "Waiting for node sync before watching apiserver pods"
Jan 17 12:00:24.800161 kubelet[2645]: I0117 12:00:24.800131    2645 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="v1.7.21" apiVersion="v1"
Jan 17 12:00:24.800342 kubelet[2645]: I0117 12:00:24.800309    2645 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode"
Jan 17 12:00:24.801583 kubelet[2645]: I0117 12:00:24.800657    2645 server.go:1256] "Started kubelet"
Jan 17 12:00:24.801583 kubelet[2645]: I0117 12:00:24.801290    2645 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10
Jan 17 12:00:24.801583 kubelet[2645]: I0117 12:00:24.801491    2645 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock"
Jan 17 12:00:24.801583 kubelet[2645]: I0117 12:00:24.801537    2645 server.go:162] "Starting to listen" address="0.0.0.0" port=10250
Jan 17 12:00:24.802096 kubelet[2645]: I0117 12:00:24.802064    2645 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer"
Jan 17 12:00:24.802636 kubelet[2645]: I0117 12:00:24.802616    2645 server.go:461] "Adding debug handlers to kubelet server"
Jan 17 12:00:24.815241 kubelet[2645]: I0117 12:00:24.815157    2645 volume_manager.go:291] "Starting Kubelet Volume Manager"
Jan 17 12:00:24.815241 kubelet[2645]: I0117 12:00:24.815246    2645 desired_state_of_world_populator.go:151] "Desired state populator starts to run"
Jan 17 12:00:24.815375 kubelet[2645]: I0117 12:00:24.815361    2645 reconciler_new.go:29] "Reconciler: start to sync state"
Jan 17 12:00:24.817782 kubelet[2645]: E0117 12:00:24.817612    2645 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem"
Jan 17 12:00:24.820289 kubelet[2645]: I0117 12:00:24.819741    2645 factory.go:221] Registration of the systemd container factory successfully
Jan 17 12:00:24.820289 kubelet[2645]: I0117 12:00:24.819831    2645 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory
Jan 17 12:00:24.821754 kubelet[2645]: I0117 12:00:24.821734    2645 factory.go:221] Registration of the containerd container factory successfully
Jan 17 12:00:24.831580 kubelet[2645]: I0117 12:00:24.829386    2645 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4"
Jan 17 12:00:24.832630 kubelet[2645]: I0117 12:00:24.832583    2645 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6"
Jan 17 12:00:24.832630 kubelet[2645]: I0117 12:00:24.832611    2645 status_manager.go:217] "Starting to sync pod status with apiserver"
Jan 17 12:00:24.832630 kubelet[2645]: I0117 12:00:24.832629    2645 kubelet.go:2329] "Starting kubelet main sync loop"
Jan 17 12:00:24.832723 kubelet[2645]: E0117 12:00:24.832679    2645 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]"
Jan 17 12:00:24.866891 kubelet[2645]: I0117 12:00:24.866642    2645 cpu_manager.go:214] "Starting CPU manager" policy="none"
Jan 17 12:00:24.866891 kubelet[2645]: I0117 12:00:24.866665    2645 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s"
Jan 17 12:00:24.866891 kubelet[2645]: I0117 12:00:24.866684    2645 state_mem.go:36] "Initialized new in-memory state store"
Jan 17 12:00:24.866891 kubelet[2645]: I0117 12:00:24.866827    2645 state_mem.go:88] "Updated default CPUSet" cpuSet=""
Jan 17 12:00:24.866891 kubelet[2645]: I0117 12:00:24.866846    2645 state_mem.go:96] "Updated CPUSet assignments" assignments={}
Jan 17 12:00:24.866891 kubelet[2645]: I0117 12:00:24.866852    2645 policy_none.go:49] "None policy: Start"
Jan 17 12:00:24.867993 kubelet[2645]: I0117 12:00:24.867533    2645 memory_manager.go:170] "Starting memorymanager" policy="None"
Jan 17 12:00:24.867993 kubelet[2645]: I0117 12:00:24.867560    2645 state_mem.go:35] "Initializing new in-memory state store"
Jan 17 12:00:24.867993 kubelet[2645]: I0117 12:00:24.867720    2645 state_mem.go:75] "Updated machine memory state"
Jan 17 12:00:24.868794 kubelet[2645]: I0117 12:00:24.868770    2645 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found"
Jan 17 12:00:24.869030 kubelet[2645]: I0117 12:00:24.869006    2645 plugin_manager.go:118] "Starting Kubelet Plugin Manager"
Jan 17 12:00:24.919114 kubelet[2645]: I0117 12:00:24.918923    2645 kubelet_node_status.go:73] "Attempting to register node" node="localhost"
Jan 17 12:00:24.924330 kubelet[2645]: I0117 12:00:24.924195    2645 kubelet_node_status.go:112] "Node was previously registered" node="localhost"
Jan 17 12:00:24.924330 kubelet[2645]: I0117 12:00:24.924296    2645 kubelet_node_status.go:76] "Successfully registered node" node="localhost"
Jan 17 12:00:24.932895 kubelet[2645]: I0117 12:00:24.932794    2645 topology_manager.go:215] "Topology Admit Handler" podUID="3e2527a0b135d137a07d61e8f70971ec" podNamespace="kube-system" podName="kube-apiserver-localhost"
Jan 17 12:00:24.932895 kubelet[2645]: I0117 12:00:24.932877    2645 topology_manager.go:215] "Topology Admit Handler" podUID="dd466de870bdf0e573d7965dbd759acf" podNamespace="kube-system" podName="kube-controller-manager-localhost"
Jan 17 12:00:24.933003 kubelet[2645]: I0117 12:00:24.932932    2645 topology_manager.go:215] "Topology Admit Handler" podUID="605dd245551545e29d4e79fb03fd341e" podNamespace="kube-system" podName="kube-scheduler-localhost"
Jan 17 12:00:24.939734 kubelet[2645]: E0117 12:00:24.939708    2645 kubelet.go:1921] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:25.117134 kubelet[2645]: I0117 12:00:25.117090    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/3e2527a0b135d137a07d61e8f70971ec-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"3e2527a0b135d137a07d61e8f70971ec\") " pod="kube-system/kube-apiserver-localhost"
Jan 17 12:00:25.117473 kubelet[2645]: I0117 12:00:25.117348    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/3e2527a0b135d137a07d61e8f70971ec-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"3e2527a0b135d137a07d61e8f70971ec\") " pod="kube-system/kube-apiserver-localhost"
Jan 17 12:00:25.117473 kubelet[2645]: I0117 12:00:25.117378    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:25.117473 kubelet[2645]: I0117 12:00:25.117428    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:25.117473 kubelet[2645]: I0117 12:00:25.117453    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/605dd245551545e29d4e79fb03fd341e-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"605dd245551545e29d4e79fb03fd341e\") " pod="kube-system/kube-scheduler-localhost"
Jan 17 12:00:25.117698 kubelet[2645]: I0117 12:00:25.117558    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/3e2527a0b135d137a07d61e8f70971ec-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"3e2527a0b135d137a07d61e8f70971ec\") " pod="kube-system/kube-apiserver-localhost"
Jan 17 12:00:25.117698 kubelet[2645]: I0117 12:00:25.117599    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:25.117698 kubelet[2645]: I0117 12:00:25.117620    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:25.117698 kubelet[2645]: I0117 12:00:25.117640    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/dd466de870bdf0e573d7965dbd759acf-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"dd466de870bdf0e573d7965dbd759acf\") " pod="kube-system/kube-controller-manager-localhost"
Jan 17 12:00:25.238905 kubelet[2645]: E0117 12:00:25.238657    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:25.239168 kubelet[2645]: E0117 12:00:25.239088    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:25.240616 kubelet[2645]: E0117 12:00:25.240598    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:25.566295 sudo[1703]: pam_unix(sudo:session): session closed for user root
Jan 17 12:00:25.569707 sshd[1696]: pam_unix(sshd:session): session closed for user core
Jan 17 12:00:25.572197 systemd[1]: sshd@4-10.0.0.35:22-10.0.0.1:54592.service: Deactivated successfully.
Jan 17 12:00:25.574806 systemd-logind[1520]: Session 5 logged out. Waiting for processes to exit.
Jan 17 12:00:25.574922 systemd[1]: session-5.scope: Deactivated successfully.
Jan 17 12:00:25.576017 systemd-logind[1520]: Removed session 5.
Jan 17 12:00:25.799349 kubelet[2645]: I0117 12:00:25.799294    2645 apiserver.go:52] "Watching apiserver"
Jan 17 12:00:25.816036 kubelet[2645]: I0117 12:00:25.816003    2645 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world"
Jan 17 12:00:25.846657 kubelet[2645]: E0117 12:00:25.845737    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:25.846657 kubelet[2645]: E0117 12:00:25.846130    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:25.848694 kubelet[2645]: E0117 12:00:25.847198    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:25.865349 kubelet[2645]: I0117 12:00:25.865264    2645 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=1.86521098 podStartE2EDuration="1.86521098s" podCreationTimestamp="2025-01-17 12:00:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:00:25.86510153 +0000 UTC m=+1.120485295" watchObservedRunningTime="2025-01-17 12:00:25.86521098 +0000 UTC m=+1.120594705"
Jan 17 12:00:25.872595 kubelet[2645]: I0117 12:00:25.872432    2645 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=1.872046925 podStartE2EDuration="1.872046925s" podCreationTimestamp="2025-01-17 12:00:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:00:25.87200533 +0000 UTC m=+1.127389096" watchObservedRunningTime="2025-01-17 12:00:25.872046925 +0000 UTC m=+1.127430690"
Jan 17 12:00:25.879011 kubelet[2645]: I0117 12:00:25.878963    2645 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.878933311 podStartE2EDuration="1.878933311s" podCreationTimestamp="2025-01-17 12:00:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:00:25.878866816 +0000 UTC m=+1.134250581" watchObservedRunningTime="2025-01-17 12:00:25.878933311 +0000 UTC m=+1.134317076"
Jan 17 12:00:26.847724 kubelet[2645]: E0117 12:00:26.847680    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:26.949973 kubelet[2645]: E0117 12:00:26.949904    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:27.849886 kubelet[2645]: E0117 12:00:27.849822    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:34.970336 kubelet[2645]: E0117 12:00:34.970275    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:35.859503 kubelet[2645]: E0117 12:00:35.859449    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:36.365449 kubelet[2645]: E0117 12:00:36.365413    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:36.861454 kubelet[2645]: E0117 12:00:36.861161    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:36.960884 kubelet[2645]: E0117 12:00:36.960853    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:40.379001 kubelet[2645]: I0117 12:00:40.376973    2645 kuberuntime_manager.go:1529] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24"
Jan 17 12:00:40.380281 containerd[1534]: time="2025-01-17T12:00:40.380205576Z" level=info msg="No cni config template is specified, wait for other system components to drop the config."
Jan 17 12:00:40.381347 kubelet[2645]: I0117 12:00:40.380556    2645 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24"
Jan 17 12:00:40.506651 kubelet[2645]: I0117 12:00:40.506613    2645 topology_manager.go:215] "Topology Admit Handler" podUID="53755f40-caaa-4301-9577-a3bed28f4c0e" podNamespace="kube-system" podName="kube-proxy-zb52b"
Jan 17 12:00:40.510230 kubelet[2645]: I0117 12:00:40.508330    2645 topology_manager.go:215] "Topology Admit Handler" podUID="03a617fb-a44c-4257-aeb6-e951ac607dd0" podNamespace="kube-flannel" podName="kube-flannel-ds-qtljh"
Jan 17 12:00:40.521523 kubelet[2645]: I0117 12:00:40.521480    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-plugin\" (UniqueName: \"kubernetes.io/host-path/03a617fb-a44c-4257-aeb6-e951ac607dd0-cni-plugin\") pod \"kube-flannel-ds-qtljh\" (UID: \"03a617fb-a44c-4257-aeb6-e951ac607dd0\") " pod="kube-flannel/kube-flannel-ds-qtljh"
Jan 17 12:00:40.521523 kubelet[2645]: I0117 12:00:40.521527    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/53755f40-caaa-4301-9577-a3bed28f4c0e-kube-proxy\") pod \"kube-proxy-zb52b\" (UID: \"53755f40-caaa-4301-9577-a3bed28f4c0e\") " pod="kube-system/kube-proxy-zb52b"
Jan 17 12:00:40.521673 kubelet[2645]: I0117 12:00:40.521550    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/53755f40-caaa-4301-9577-a3bed28f4c0e-lib-modules\") pod \"kube-proxy-zb52b\" (UID: \"53755f40-caaa-4301-9577-a3bed28f4c0e\") " pod="kube-system/kube-proxy-zb52b"
Jan 17 12:00:40.521673 kubelet[2645]: I0117 12:00:40.521569    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni\" (UniqueName: \"kubernetes.io/host-path/03a617fb-a44c-4257-aeb6-e951ac607dd0-cni\") pod \"kube-flannel-ds-qtljh\" (UID: \"03a617fb-a44c-4257-aeb6-e951ac607dd0\") " pod="kube-flannel/kube-flannel-ds-qtljh"
Jan 17 12:00:40.521673 kubelet[2645]: I0117 12:00:40.521589    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/03a617fb-a44c-4257-aeb6-e951ac607dd0-xtables-lock\") pod \"kube-flannel-ds-qtljh\" (UID: \"03a617fb-a44c-4257-aeb6-e951ac607dd0\") " pod="kube-flannel/kube-flannel-ds-qtljh"
Jan 17 12:00:40.521673 kubelet[2645]: I0117 12:00:40.521608    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/03a617fb-a44c-4257-aeb6-e951ac607dd0-run\") pod \"kube-flannel-ds-qtljh\" (UID: \"03a617fb-a44c-4257-aeb6-e951ac607dd0\") " pod="kube-flannel/kube-flannel-ds-qtljh"
Jan 17 12:00:40.521673 kubelet[2645]: I0117 12:00:40.521626    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/53755f40-caaa-4301-9577-a3bed28f4c0e-xtables-lock\") pod \"kube-proxy-zb52b\" (UID: \"53755f40-caaa-4301-9577-a3bed28f4c0e\") " pod="kube-system/kube-proxy-zb52b"
Jan 17 12:00:40.521780 kubelet[2645]: I0117 12:00:40.521651    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2ztd\" (UniqueName: \"kubernetes.io/projected/53755f40-caaa-4301-9577-a3bed28f4c0e-kube-api-access-w2ztd\") pod \"kube-proxy-zb52b\" (UID: \"53755f40-caaa-4301-9577-a3bed28f4c0e\") " pod="kube-system/kube-proxy-zb52b"
Jan 17 12:00:40.521780 kubelet[2645]: I0117 12:00:40.521670    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flannel-cfg\" (UniqueName: \"kubernetes.io/configmap/03a617fb-a44c-4257-aeb6-e951ac607dd0-flannel-cfg\") pod \"kube-flannel-ds-qtljh\" (UID: \"03a617fb-a44c-4257-aeb6-e951ac607dd0\") " pod="kube-flannel/kube-flannel-ds-qtljh"
Jan 17 12:00:40.521780 kubelet[2645]: I0117 12:00:40.521689    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45hmf\" (UniqueName: \"kubernetes.io/projected/03a617fb-a44c-4257-aeb6-e951ac607dd0-kube-api-access-45hmf\") pod \"kube-flannel-ds-qtljh\" (UID: \"03a617fb-a44c-4257-aeb6-e951ac607dd0\") " pod="kube-flannel/kube-flannel-ds-qtljh"
Jan 17 12:00:40.810156 kubelet[2645]: E0117 12:00:40.809971    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:40.810482 kubelet[2645]: E0117 12:00:40.810291    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:40.810522 containerd[1534]: time="2025-01-17T12:00:40.810415728Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-zb52b,Uid:53755f40-caaa-4301-9577-a3bed28f4c0e,Namespace:kube-system,Attempt:0,}"
Jan 17 12:00:40.811316 containerd[1534]: time="2025-01-17T12:00:40.811285219Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-qtljh,Uid:03a617fb-a44c-4257-aeb6-e951ac607dd0,Namespace:kube-flannel,Attempt:0,}"
Jan 17 12:00:40.839269 containerd[1534]: time="2025-01-17T12:00:40.839066573Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jan 17 12:00:40.839269 containerd[1534]: time="2025-01-17T12:00:40.839135757Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jan 17 12:00:40.839392 containerd[1534]: time="2025-01-17T12:00:40.839221385Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:40.839392 containerd[1534]: time="2025-01-17T12:00:40.839369955Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:40.844373 containerd[1534]: time="2025-01-17T12:00:40.841481103Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jan 17 12:00:40.844373 containerd[1534]: time="2025-01-17T12:00:40.841539803Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jan 17 12:00:40.844373 containerd[1534]: time="2025-01-17T12:00:40.841556288Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:40.844373 containerd[1534]: time="2025-01-17T12:00:40.841637075Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:40.881115 containerd[1534]: time="2025-01-17T12:00:40.881060852Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-zb52b,Uid:53755f40-caaa-4301-9577-a3bed28f4c0e,Namespace:kube-system,Attempt:0,} returns sandbox id \"cee23b3fae12936c9a9500070ea169e791369c9c5fbd68274c8979abb71f159e\""
Jan 17 12:00:40.882367 kubelet[2645]: E0117 12:00:40.881642    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:40.886515 containerd[1534]: time="2025-01-17T12:00:40.886477428Z" level=info msg="CreateContainer within sandbox \"cee23b3fae12936c9a9500070ea169e791369c9c5fbd68274c8979abb71f159e\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}"
Jan 17 12:00:40.886906 containerd[1534]: time="2025-01-17T12:00:40.886883965Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-flannel-ds-qtljh,Uid:03a617fb-a44c-4257-aeb6-e951ac607dd0,Namespace:kube-flannel,Attempt:0,} returns sandbox id \"706448d5438a12b9b8e06b0ceb550751bc10b3cd3dedf19403b9379b708ece3e\""
Jan 17 12:00:40.887712 kubelet[2645]: E0117 12:00:40.887694    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:40.898451 containerd[1534]: time="2025-01-17T12:00:40.898360492Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\""
Jan 17 12:00:40.904253 containerd[1534]: time="2025-01-17T12:00:40.904213334Z" level=info msg="CreateContainer within sandbox \"cee23b3fae12936c9a9500070ea169e791369c9c5fbd68274c8979abb71f159e\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"a5461fd73331a0c604f7a5a22467a1309d500a38f7881cea6416362f22dbd36a\""
Jan 17 12:00:40.904740 containerd[1534]: time="2025-01-17T12:00:40.904688694Z" level=info msg="StartContainer for \"a5461fd73331a0c604f7a5a22467a1309d500a38f7881cea6416362f22dbd36a\""
Jan 17 12:00:40.956146 containerd[1534]: time="2025-01-17T12:00:40.956099890Z" level=info msg="StartContainer for \"a5461fd73331a0c604f7a5a22467a1309d500a38f7881cea6416362f22dbd36a\" returns successfully"
Jan 17 12:00:41.873412 kubelet[2645]: E0117 12:00:41.873046    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:41.886217 kubelet[2645]: I0117 12:00:41.886078    2645 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-zb52b" podStartSLOduration=1.886037959 podStartE2EDuration="1.886037959s" podCreationTimestamp="2025-01-17 12:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:00:41.885971578 +0000 UTC m=+17.141355343" watchObservedRunningTime="2025-01-17 12:00:41.886037959 +0000 UTC m=+17.141421724"
Jan 17 12:00:41.938844 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4256573791.mount: Deactivated successfully.
Jan 17 12:00:41.973662 containerd[1534]: time="2025-01-17T12:00:41.973482095Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin:v1.1.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:41.974467 containerd[1534]: time="2025-01-17T12:00:41.974356053Z" level=info msg="stop pulling image docker.io/flannel/flannel-cni-plugin:v1.1.2: active requests=0, bytes read=3673531"
Jan 17 12:00:41.975163 containerd[1534]: time="2025-01-17T12:00:41.975135262Z" level=info msg="ImageCreate event name:\"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:41.977477 containerd[1534]: time="2025-01-17T12:00:41.977447318Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:41.979016 containerd[1534]: time="2025-01-17T12:00:41.978739370Z" level=info msg="Pulled image \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" with image id \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\", repo tag \"docker.io/flannel/flannel-cni-plugin:v1.1.2\", repo digest \"docker.io/flannel/flannel-cni-plugin@sha256:bf4b62b131666d040f35a327d906ee5a3418280b68a88d9b9c7e828057210443\", size \"3662650\" in 1.080344106s"
Jan 17 12:00:41.979016 containerd[1534]: time="2025-01-17T12:00:41.978774701Z" level=info msg="PullImage \"docker.io/flannel/flannel-cni-plugin:v1.1.2\" returns image reference \"sha256:b45062ceea496fc421523388cb91166abc7715a15c2e2cbab4e6f8c9d5dc0ab8\""
Jan 17 12:00:41.980427 containerd[1534]: time="2025-01-17T12:00:41.980393137Z" level=info msg="CreateContainer within sandbox \"706448d5438a12b9b8e06b0ceb550751bc10b3cd3dedf19403b9379b708ece3e\" for container &ContainerMetadata{Name:install-cni-plugin,Attempt:0,}"
Jan 17 12:00:41.999939 containerd[1534]: time="2025-01-17T12:00:41.999897150Z" level=info msg="CreateContainer within sandbox \"706448d5438a12b9b8e06b0ceb550751bc10b3cd3dedf19403b9379b708ece3e\" for &ContainerMetadata{Name:install-cni-plugin,Attempt:0,} returns container id \"c7f21da4a44797a80ada31914ce2f37c53d9809a4150258280f3d7040b1c49ed\""
Jan 17 12:00:42.000372 containerd[1534]: time="2025-01-17T12:00:42.000346613Z" level=info msg="StartContainer for \"c7f21da4a44797a80ada31914ce2f37c53d9809a4150258280f3d7040b1c49ed\""
Jan 17 12:00:42.057331 containerd[1534]: time="2025-01-17T12:00:42.057278708Z" level=info msg="StartContainer for \"c7f21da4a44797a80ada31914ce2f37c53d9809a4150258280f3d7040b1c49ed\" returns successfully"
Jan 17 12:00:42.082125 containerd[1534]: time="2025-01-17T12:00:42.077627831Z" level=info msg="shim disconnected" id=c7f21da4a44797a80ada31914ce2f37c53d9809a4150258280f3d7040b1c49ed namespace=k8s.io
Jan 17 12:00:42.082125 containerd[1534]: time="2025-01-17T12:00:42.082118832Z" level=warning msg="cleaning up after shim disconnected" id=c7f21da4a44797a80ada31914ce2f37c53d9809a4150258280f3d7040b1c49ed namespace=k8s.io
Jan 17 12:00:42.082125 containerd[1534]: time="2025-01-17T12:00:42.082133436Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jan 17 12:00:42.154232 update_engine[1522]: I20250117 12:00:42.154086  1522 update_attempter.cc:509] Updating boot flags...
Jan 17 12:00:42.173014 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 40 scanned by (udev-worker) (3023)
Jan 17 12:00:42.196041 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 40 scanned by (udev-worker) (2907)
Jan 17 12:00:42.633367 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount638520581.mount: Deactivated successfully.
Jan 17 12:00:42.875790 kubelet[2645]: E0117 12:00:42.875717    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:42.877959 containerd[1534]: time="2025-01-17T12:00:42.877077425Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\""
Jan 17 12:00:43.883442 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2329895872.mount: Deactivated successfully.
Jan 17 12:00:44.452863 containerd[1534]: time="2025-01-17T12:00:44.452641858Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel:v0.22.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:44.454033 containerd[1534]: time="2025-01-17T12:00:44.453922770Z" level=info msg="stop pulling image docker.io/flannel/flannel:v0.22.0: active requests=0, bytes read=26874261"
Jan 17 12:00:44.454831 containerd[1534]: time="2025-01-17T12:00:44.454734072Z" level=info msg="ImageCreate event name:\"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:44.458454 containerd[1534]: time="2025-01-17T12:00:44.458393197Z" level=info msg="ImageCreate event name:\"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}"
Jan 17 12:00:44.460435 containerd[1534]: time="2025-01-17T12:00:44.460028966Z" level=info msg="Pulled image \"docker.io/flannel/flannel:v0.22.0\" with image id \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\", repo tag \"docker.io/flannel/flannel:v0.22.0\", repo digest \"docker.io/flannel/flannel@sha256:5f83f1243057458e27249157394e3859cf31cc075354af150d497f2ebc8b54db\", size \"26863435\" in 1.582911809s"
Jan 17 12:00:44.460435 containerd[1534]: time="2025-01-17T12:00:44.460062775Z" level=info msg="PullImage \"docker.io/flannel/flannel:v0.22.0\" returns image reference \"sha256:b3d1319ea6da12d4a1dd21a923f6a71f942a7b6e2c4763b8a3cca0725fb8aadf\""
Jan 17 12:00:44.480180 containerd[1534]: time="2025-01-17T12:00:44.480146687Z" level=info msg="CreateContainer within sandbox \"706448d5438a12b9b8e06b0ceb550751bc10b3cd3dedf19403b9379b708ece3e\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}"
Jan 17 12:00:44.493650 containerd[1534]: time="2025-01-17T12:00:44.493598619Z" level=info msg="CreateContainer within sandbox \"706448d5438a12b9b8e06b0ceb550751bc10b3cd3dedf19403b9379b708ece3e\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"821e9c1cffa1065c9952cc916fbeb81ce939997ab418b8359237a10040583399\""
Jan 17 12:00:44.494199 containerd[1534]: time="2025-01-17T12:00:44.494166175Z" level=info msg="StartContainer for \"821e9c1cffa1065c9952cc916fbeb81ce939997ab418b8359237a10040583399\""
Jan 17 12:00:44.551293 containerd[1534]: time="2025-01-17T12:00:44.551159777Z" level=info msg="StartContainer for \"821e9c1cffa1065c9952cc916fbeb81ce939997ab418b8359237a10040583399\" returns successfully"
Jan 17 12:00:44.567953 containerd[1534]: time="2025-01-17T12:00:44.567886328Z" level=info msg="shim disconnected" id=821e9c1cffa1065c9952cc916fbeb81ce939997ab418b8359237a10040583399 namespace=k8s.io
Jan 17 12:00:44.567953 containerd[1534]: time="2025-01-17T12:00:44.567940663Z" level=warning msg="cleaning up after shim disconnected" id=821e9c1cffa1065c9952cc916fbeb81ce939997ab418b8359237a10040583399 namespace=k8s.io
Jan 17 12:00:44.567953 containerd[1534]: time="2025-01-17T12:00:44.567949025Z" level=info msg="cleaning up dead shim" namespace=k8s.io
Jan 17 12:00:44.584637 kubelet[2645]: I0117 12:00:44.583867    2645 kubelet_node_status.go:497] "Fast updating node status as it just became ready"
Jan 17 12:00:44.613612 kubelet[2645]: I0117 12:00:44.613569    2645 topology_manager.go:215] "Topology Admit Handler" podUID="1556d61a-088e-4331-a92a-beaee4a257b5" podNamespace="kube-system" podName="coredns-76f75df574-ghmwl"
Jan 17 12:00:44.613768 kubelet[2645]: I0117 12:00:44.613751    2645 topology_manager.go:215] "Topology Admit Handler" podUID="38cf02b6-fc2c-420e-879c-599727e10909" podNamespace="kube-system" podName="coredns-76f75df574-vpz5m"
Jan 17 12:00:44.751161 kubelet[2645]: I0117 12:00:44.751053    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwdw\" (UniqueName: \"kubernetes.io/projected/1556d61a-088e-4331-a92a-beaee4a257b5-kube-api-access-djwdw\") pod \"coredns-76f75df574-ghmwl\" (UID: \"1556d61a-088e-4331-a92a-beaee4a257b5\") " pod="kube-system/coredns-76f75df574-ghmwl"
Jan 17 12:00:44.751161 kubelet[2645]: I0117 12:00:44.751104    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1556d61a-088e-4331-a92a-beaee4a257b5-config-volume\") pod \"coredns-76f75df574-ghmwl\" (UID: \"1556d61a-088e-4331-a92a-beaee4a257b5\") " pod="kube-system/coredns-76f75df574-ghmwl"
Jan 17 12:00:44.751161 kubelet[2645]: I0117 12:00:44.751127    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/38cf02b6-fc2c-420e-879c-599727e10909-config-volume\") pod \"coredns-76f75df574-vpz5m\" (UID: \"38cf02b6-fc2c-420e-879c-599727e10909\") " pod="kube-system/coredns-76f75df574-vpz5m"
Jan 17 12:00:44.751161 kubelet[2645]: I0117 12:00:44.751157    2645 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwrkg\" (UniqueName: \"kubernetes.io/projected/38cf02b6-fc2c-420e-879c-599727e10909-kube-api-access-fwrkg\") pod \"coredns-76f75df574-vpz5m\" (UID: \"38cf02b6-fc2c-420e-879c-599727e10909\") " pod="kube-system/coredns-76f75df574-vpz5m"
Jan 17 12:00:44.806691 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-821e9c1cffa1065c9952cc916fbeb81ce939997ab418b8359237a10040583399-rootfs.mount: Deactivated successfully.
Jan 17 12:00:44.885119 kubelet[2645]: E0117 12:00:44.885042    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:44.889146 containerd[1534]: time="2025-01-17T12:00:44.889095647Z" level=info msg="CreateContainer within sandbox \"706448d5438a12b9b8e06b0ceb550751bc10b3cd3dedf19403b9379b708ece3e\" for container &ContainerMetadata{Name:kube-flannel,Attempt:0,}"
Jan 17 12:00:44.902922 containerd[1534]: time="2025-01-17T12:00:44.902873948Z" level=info msg="CreateContainer within sandbox \"706448d5438a12b9b8e06b0ceb550751bc10b3cd3dedf19403b9379b708ece3e\" for &ContainerMetadata{Name:kube-flannel,Attempt:0,} returns container id \"7ce0871fb43837af9bca750eae5401b7892130d26cd4b3a7d728ba2157a0c9cc\""
Jan 17 12:00:44.903340 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2003448312.mount: Deactivated successfully.
Jan 17 12:00:44.904498 containerd[1534]: time="2025-01-17T12:00:44.903715779Z" level=info msg="StartContainer for \"7ce0871fb43837af9bca750eae5401b7892130d26cd4b3a7d728ba2157a0c9cc\""
Jan 17 12:00:44.916885 kubelet[2645]: E0117 12:00:44.916846    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:44.917084 kubelet[2645]: E0117 12:00:44.917035    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:44.917828 containerd[1534]: time="2025-01-17T12:00:44.917782840Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-vpz5m,Uid:38cf02b6-fc2c-420e-879c-599727e10909,Namespace:kube-system,Attempt:0,}"
Jan 17 12:00:44.918563 containerd[1534]: time="2025-01-17T12:00:44.918366600Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-ghmwl,Uid:1556d61a-088e-4331-a92a-beaee4a257b5,Namespace:kube-system,Attempt:0,}"
Jan 17 12:00:44.983690 containerd[1534]: time="2025-01-17T12:00:44.980439397Z" level=info msg="StartContainer for \"7ce0871fb43837af9bca750eae5401b7892130d26cd4b3a7d728ba2157a0c9cc\" returns successfully"
Jan 17 12:00:45.008734 containerd[1534]: time="2025-01-17T12:00:45.008161481Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-vpz5m,Uid:38cf02b6-fc2c-420e-879c-599727e10909,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"3ef3c18e8bd751415b0e97ee7daaed3dec701776b4c55037469bc82396e335cf\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory"
Jan 17 12:00:45.008838 kubelet[2645]: E0117 12:00:45.008414    2645 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3ef3c18e8bd751415b0e97ee7daaed3dec701776b4c55037469bc82396e335cf\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory"
Jan 17 12:00:45.008838 kubelet[2645]: E0117 12:00:45.008469    2645 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3ef3c18e8bd751415b0e97ee7daaed3dec701776b4c55037469bc82396e335cf\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-76f75df574-vpz5m"
Jan 17 12:00:45.008838 kubelet[2645]: E0117 12:00:45.008489    2645 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3ef3c18e8bd751415b0e97ee7daaed3dec701776b4c55037469bc82396e335cf\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-76f75df574-vpz5m"
Jan 17 12:00:45.008838 kubelet[2645]: E0117 12:00:45.008533    2645 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-76f75df574-vpz5m_kube-system(38cf02b6-fc2c-420e-879c-599727e10909)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-76f75df574-vpz5m_kube-system(38cf02b6-fc2c-420e-879c-599727e10909)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3ef3c18e8bd751415b0e97ee7daaed3dec701776b4c55037469bc82396e335cf\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-76f75df574-vpz5m" podUID="38cf02b6-fc2c-420e-879c-599727e10909"
Jan 17 12:00:45.019617 containerd[1534]: time="2025-01-17T12:00:45.018864160Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-ghmwl,Uid:1556d61a-088e-4331-a92a-beaee4a257b5,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"59dbeb7b70caafa7637e087c5293877aa1bac5472757ec5393bf56aa6578a2ed\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory"
Jan 17 12:00:45.019730 kubelet[2645]: E0117 12:00:45.019088    2645 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59dbeb7b70caafa7637e087c5293877aa1bac5472757ec5393bf56aa6578a2ed\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory"
Jan 17 12:00:45.019730 kubelet[2645]: E0117 12:00:45.019143    2645 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59dbeb7b70caafa7637e087c5293877aa1bac5472757ec5393bf56aa6578a2ed\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-76f75df574-ghmwl"
Jan 17 12:00:45.019730 kubelet[2645]: E0117 12:00:45.019164    2645 kuberuntime_manager.go:1172] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"59dbeb7b70caafa7637e087c5293877aa1bac5472757ec5393bf56aa6578a2ed\": plugin type=\"flannel\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory" pod="kube-system/coredns-76f75df574-ghmwl"
Jan 17 12:00:45.019730 kubelet[2645]: E0117 12:00:45.019207    2645 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-76f75df574-ghmwl_kube-system(1556d61a-088e-4331-a92a-beaee4a257b5)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-76f75df574-ghmwl_kube-system(1556d61a-088e-4331-a92a-beaee4a257b5)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"59dbeb7b70caafa7637e087c5293877aa1bac5472757ec5393bf56aa6578a2ed\\\": plugin type=\\\"flannel\\\" failed (add): loadFlannelSubnetEnv failed: open /run/flannel/subnet.env: no such file or directory\"" pod="kube-system/coredns-76f75df574-ghmwl" podUID="1556d61a-088e-4331-a92a-beaee4a257b5"
Jan 17 12:00:45.885245 kubelet[2645]: E0117 12:00:45.885204    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:45.896175 kubelet[2645]: I0117 12:00:45.896118    2645 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-flannel/kube-flannel-ds-qtljh" podStartSLOduration=2.324093598 podStartE2EDuration="5.895966811s" podCreationTimestamp="2025-01-17 12:00:40 +0000 UTC" firstStartedPulling="2025-01-17 12:00:40.888835699 +0000 UTC m=+16.144219464" lastFinishedPulling="2025-01-17 12:00:44.460708912 +0000 UTC m=+19.716092677" observedRunningTime="2025-01-17 12:00:45.89596269 +0000 UTC m=+21.151346455" watchObservedRunningTime="2025-01-17 12:00:45.895966811 +0000 UTC m=+21.151350576"
Jan 17 12:00:46.069595 systemd-networkd[1228]: flannel.1: Link UP
Jan 17 12:00:46.069604 systemd-networkd[1228]: flannel.1: Gained carrier
Jan 17 12:00:46.886902 kubelet[2645]: E0117 12:00:46.886862    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:47.411179 systemd-networkd[1228]: flannel.1: Gained IPv6LL
Jan 17 12:00:53.829279 systemd[1]: Started sshd@5-10.0.0.35:22-10.0.0.1:44328.service - OpenSSH per-connection server daemon (10.0.0.1:44328).
Jan 17 12:00:53.860916 sshd[3313]: Accepted publickey for core from 10.0.0.1 port 44328 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:00:53.862749 sshd[3313]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:00:53.866678 systemd-logind[1520]: New session 6 of user core.
Jan 17 12:00:53.873252 systemd[1]: Started session-6.scope - Session 6 of User core.
Jan 17 12:00:53.989771 sshd[3313]: pam_unix(sshd:session): session closed for user core
Jan 17 12:00:53.993065 systemd[1]: sshd@5-10.0.0.35:22-10.0.0.1:44328.service: Deactivated successfully.
Jan 17 12:00:53.995194 systemd-logind[1520]: Session 6 logged out. Waiting for processes to exit.
Jan 17 12:00:53.995225 systemd[1]: session-6.scope: Deactivated successfully.
Jan 17 12:00:53.996843 systemd-logind[1520]: Removed session 6.
Jan 17 12:00:56.833866 kubelet[2645]: E0117 12:00:56.833688    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:56.834254 containerd[1534]: time="2025-01-17T12:00:56.834133798Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-ghmwl,Uid:1556d61a-088e-4331-a92a-beaee4a257b5,Namespace:kube-system,Attempt:0,}"
Jan 17 12:00:56.872380 systemd-networkd[1228]: cni0: Link UP
Jan 17 12:00:56.872387 systemd-networkd[1228]: cni0: Gained carrier
Jan 17 12:00:56.874609 systemd-networkd[1228]: cni0: Lost carrier
Jan 17 12:00:56.878660 systemd-networkd[1228]: vethed680c80: Link UP
Jan 17 12:00:56.881289 kernel: cni0: port 1(vethed680c80) entered blocking state
Jan 17 12:00:56.881348 kernel: cni0: port 1(vethed680c80) entered disabled state
Jan 17 12:00:56.881366 kernel: vethed680c80: entered allmulticast mode
Jan 17 12:00:56.882043 kernel: vethed680c80: entered promiscuous mode
Jan 17 12:00:56.883059 kernel: cni0: port 1(vethed680c80) entered blocking state
Jan 17 12:00:56.883110 kernel: cni0: port 1(vethed680c80) entered forwarding state
Jan 17 12:00:56.884000 kernel: cni0: port 1(vethed680c80) entered disabled state
Jan 17 12:00:56.896338 kernel: cni0: port 1(vethed680c80) entered blocking state
Jan 17 12:00:56.896398 kernel: cni0: port 1(vethed680c80) entered forwarding state
Jan 17 12:00:56.896776 systemd-networkd[1228]: vethed680c80: Gained carrier
Jan 17 12:00:56.897232 systemd-networkd[1228]: cni0: Gained carrier
Jan 17 12:00:56.898616 containerd[1534]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x400001a938), "name":"cbr0", "type":"bridge"}
Jan 17 12:00:56.898616 containerd[1534]: delegateAdd: netconf sent to delegate plugin:
Jan 17 12:00:56.914335 containerd[1534]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2025-01-17T12:00:56.914108567Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jan 17 12:00:56.914335 containerd[1534]: time="2025-01-17T12:00:56.914164736Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jan 17 12:00:56.914335 containerd[1534]: time="2025-01-17T12:00:56.914180018Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:56.914335 containerd[1534]: time="2025-01-17T12:00:56.914271473Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:56.937773 systemd-resolved[1434]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address
Jan 17 12:00:56.955045 containerd[1534]: time="2025-01-17T12:00:56.954960792Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-ghmwl,Uid:1556d61a-088e-4331-a92a-beaee4a257b5,Namespace:kube-system,Attempt:0,} returns sandbox id \"9776acb9bf63fd7d92ddce9d1aa7b0c94a403fd8b28ef631061272d105975dbd\""
Jan 17 12:00:56.956006 kubelet[2645]: E0117 12:00:56.955884    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:56.971871 containerd[1534]: time="2025-01-17T12:00:56.971814782Z" level=info msg="CreateContainer within sandbox \"9776acb9bf63fd7d92ddce9d1aa7b0c94a403fd8b28ef631061272d105975dbd\" for container &ContainerMetadata{Name:coredns,Attempt:0,}"
Jan 17 12:00:56.982385 containerd[1534]: time="2025-01-17T12:00:56.982334659Z" level=info msg="CreateContainer within sandbox \"9776acb9bf63fd7d92ddce9d1aa7b0c94a403fd8b28ef631061272d105975dbd\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"76b3328a8a198d7ee367be7318d98264b4cffbd39c377bfa7d697742ed33d8a8\""
Jan 17 12:00:56.983011 containerd[1534]: time="2025-01-17T12:00:56.982963041Z" level=info msg="StartContainer for \"76b3328a8a198d7ee367be7318d98264b4cffbd39c377bfa7d697742ed33d8a8\""
Jan 17 12:00:57.034448 containerd[1534]: time="2025-01-17T12:00:57.034337665Z" level=info msg="StartContainer for \"76b3328a8a198d7ee367be7318d98264b4cffbd39c377bfa7d697742ed33d8a8\" returns successfully"
Jan 17 12:00:57.906329 kubelet[2645]: E0117 12:00:57.906292    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:57.915365 kubelet[2645]: I0117 12:00:57.915226    2645 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-76f75df574-ghmwl" podStartSLOduration=17.915185588 podStartE2EDuration="17.915185588s" podCreationTimestamp="2025-01-17 12:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:00:57.91481337 +0000 UTC m=+33.170197135" watchObservedRunningTime="2025-01-17 12:00:57.915185588 +0000 UTC m=+33.170569353"
Jan 17 12:00:57.971211 systemd-networkd[1228]: cni0: Gained IPv6LL
Jan 17 12:00:58.355114 systemd-networkd[1228]: vethed680c80: Gained IPv6LL
Jan 17 12:00:59.000384 systemd[1]: Started sshd@6-10.0.0.35:22-10.0.0.1:44342.service - OpenSSH per-connection server daemon (10.0.0.1:44342).
Jan 17 12:00:59.032092 sshd[3470]: Accepted publickey for core from 10.0.0.1 port 44342 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:00:59.033417 sshd[3470]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:00:59.037283 systemd-logind[1520]: New session 7 of user core.
Jan 17 12:00:59.050272 systemd[1]: Started session-7.scope - Session 7 of User core.
Jan 17 12:00:59.166165 sshd[3470]: pam_unix(sshd:session): session closed for user core
Jan 17 12:00:59.169415 systemd[1]: sshd@6-10.0.0.35:22-10.0.0.1:44342.service: Deactivated successfully.
Jan 17 12:00:59.172125 systemd-logind[1520]: Session 7 logged out. Waiting for processes to exit.
Jan 17 12:00:59.172509 systemd[1]: session-7.scope: Deactivated successfully.
Jan 17 12:00:59.173861 systemd-logind[1520]: Removed session 7.
Jan 17 12:00:59.834219 kubelet[2645]: E0117 12:00:59.834050    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:59.834637 containerd[1534]: time="2025-01-17T12:00:59.834448102Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-vpz5m,Uid:38cf02b6-fc2c-420e-879c-599727e10909,Namespace:kube-system,Attempt:0,}"
Jan 17 12:00:59.860558 systemd-networkd[1228]: veth64409df9: Link UP
Jan 17 12:00:59.864332 kernel: cni0: port 2(veth64409df9) entered blocking state
Jan 17 12:00:59.864417 kernel: cni0: port 2(veth64409df9) entered disabled state
Jan 17 12:00:59.864434 kernel: veth64409df9: entered allmulticast mode
Jan 17 12:00:59.864449 kernel: veth64409df9: entered promiscuous mode
Jan 17 12:00:59.871229 kernel: cni0: port 2(veth64409df9) entered blocking state
Jan 17 12:00:59.871343 kernel: cni0: port 2(veth64409df9) entered forwarding state
Jan 17 12:00:59.871459 systemd-networkd[1228]: veth64409df9: Gained carrier
Jan 17 12:00:59.873510 containerd[1534]: map[string]interface {}{"cniVersion":"0.3.1", "hairpinMode":true, "ipMasq":false, "ipam":map[string]interface {}{"ranges":[][]map[string]interface {}{[]map[string]interface {}{map[string]interface {}{"subnet":"192.168.0.0/24"}}}, "routes":[]types.Route{types.Route{Dst:net.IPNet{IP:net.IP{0xc0, 0xa8, 0x0, 0x0}, Mask:net.IPMask{0xff, 0xff, 0x80, 0x0}}, GW:net.IP(nil)}}, "type":"host-local"}, "isDefaultGateway":true, "isGateway":true, "mtu":(*uint)(0x4000014938), "name":"cbr0", "type":"bridge"}
Jan 17 12:00:59.873510 containerd[1534]: delegateAdd: netconf sent to delegate plugin:
Jan 17 12:00:59.888223 containerd[1534]: {"cniVersion":"0.3.1","hairpinMode":true,"ipMasq":false,"ipam":{"ranges":[[{"subnet":"192.168.0.0/24"}]],"routes":[{"dst":"192.168.0.0/17"}],"type":"host-local"},"isDefaultGateway":true,"isGateway":true,"mtu":1450,"name":"cbr0","type":"bridge"}time="2025-01-17T12:00:59.888128322Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1
Jan 17 12:00:59.888223 containerd[1534]: time="2025-01-17T12:00:59.888194851Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1
Jan 17 12:00:59.888223 containerd[1534]: time="2025-01-17T12:00:59.888215975Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:59.888685 containerd[1534]: time="2025-01-17T12:00:59.888584309Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1
Jan 17 12:00:59.916183 systemd-resolved[1434]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address
Jan 17 12:00:59.932945 containerd[1534]: time="2025-01-17T12:00:59.932878114Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-76f75df574-vpz5m,Uid:38cf02b6-fc2c-420e-879c-599727e10909,Namespace:kube-system,Attempt:0,} returns sandbox id \"d125afbd8fdaed47a6386aadc7cae01ee3c88440d645a1bcc46818e3b8dafbc3\""
Jan 17 12:00:59.933545 kubelet[2645]: E0117 12:00:59.933527    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:00:59.936862 containerd[1534]: time="2025-01-17T12:00:59.936829892Z" level=info msg="CreateContainer within sandbox \"d125afbd8fdaed47a6386aadc7cae01ee3c88440d645a1bcc46818e3b8dafbc3\" for container &ContainerMetadata{Name:coredns,Attempt:0,}"
Jan 17 12:00:59.950709 containerd[1534]: time="2025-01-17T12:00:59.950659997Z" level=info msg="CreateContainer within sandbox \"d125afbd8fdaed47a6386aadc7cae01ee3c88440d645a1bcc46818e3b8dafbc3\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"1da5dd64b58f4fd5ae84cfc85b1c1556815afe55aee793a4f199213a185a828f\""
Jan 17 12:00:59.951422 containerd[1534]: time="2025-01-17T12:00:59.951195115Z" level=info msg="StartContainer for \"1da5dd64b58f4fd5ae84cfc85b1c1556815afe55aee793a4f199213a185a828f\""
Jan 17 12:01:00.002317 containerd[1534]: time="2025-01-17T12:01:00.002266746Z" level=info msg="StartContainer for \"1da5dd64b58f4fd5ae84cfc85b1c1556815afe55aee793a4f199213a185a828f\" returns successfully"
Jan 17 12:01:00.914103 kubelet[2645]: E0117 12:01:00.913356    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:01:01.107141 systemd-networkd[1228]: veth64409df9: Gained IPv6LL
Jan 17 12:01:01.914993 kubelet[2645]: E0117 12:01:01.914930    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:01:04.179199 systemd[1]: Started sshd@7-10.0.0.35:22-10.0.0.1:56596.service - OpenSSH per-connection server daemon (10.0.0.1:56596).
Jan 17 12:01:04.211244 sshd[3615]: Accepted publickey for core from 10.0.0.1 port 56596 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:04.214550 sshd[3615]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:04.225892 systemd-logind[1520]: New session 8 of user core.
Jan 17 12:01:04.237248 systemd[1]: Started session-8.scope - Session 8 of User core.
Jan 17 12:01:04.350218 sshd[3615]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:04.356317 systemd[1]: Started sshd@8-10.0.0.35:22-10.0.0.1:56612.service - OpenSSH per-connection server daemon (10.0.0.1:56612).
Jan 17 12:01:04.356697 systemd[1]: sshd@7-10.0.0.35:22-10.0.0.1:56596.service: Deactivated successfully.
Jan 17 12:01:04.361175 systemd-logind[1520]: Session 8 logged out. Waiting for processes to exit.
Jan 17 12:01:04.361314 systemd[1]: session-8.scope: Deactivated successfully.
Jan 17 12:01:04.362557 systemd-logind[1520]: Removed session 8.
Jan 17 12:01:04.390257 sshd[3629]: Accepted publickey for core from 10.0.0.1 port 56612 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:04.391585 sshd[3629]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:04.395760 systemd-logind[1520]: New session 9 of user core.
Jan 17 12:01:04.409237 systemd[1]: Started session-9.scope - Session 9 of User core.
Jan 17 12:01:04.547464 sshd[3629]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:04.555274 systemd[1]: Started sshd@9-10.0.0.35:22-10.0.0.1:56624.service - OpenSSH per-connection server daemon (10.0.0.1:56624).
Jan 17 12:01:04.555800 systemd[1]: sshd@8-10.0.0.35:22-10.0.0.1:56612.service: Deactivated successfully.
Jan 17 12:01:04.559414 systemd-logind[1520]: Session 9 logged out. Waiting for processes to exit.
Jan 17 12:01:04.564274 systemd[1]: session-9.scope: Deactivated successfully.
Jan 17 12:01:04.571476 systemd-logind[1520]: Removed session 9.
Jan 17 12:01:04.591469 sshd[3642]: Accepted publickey for core from 10.0.0.1 port 56624 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:04.592874 sshd[3642]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:04.596491 systemd-logind[1520]: New session 10 of user core.
Jan 17 12:01:04.604343 systemd[1]: Started session-10.scope - Session 10 of User core.
Jan 17 12:01:04.710201 sshd[3642]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:04.713921 systemd[1]: sshd@9-10.0.0.35:22-10.0.0.1:56624.service: Deactivated successfully.
Jan 17 12:01:04.714247 systemd-logind[1520]: Session 10 logged out. Waiting for processes to exit.
Jan 17 12:01:04.716575 systemd[1]: session-10.scope: Deactivated successfully.
Jan 17 12:01:04.717561 systemd-logind[1520]: Removed session 10.
Jan 17 12:01:04.918207 kubelet[2645]: E0117 12:01:04.918141    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:01:04.931847 kubelet[2645]: I0117 12:01:04.931596    2645 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-76f75df574-vpz5m" podStartSLOduration=24.931544072 podStartE2EDuration="24.931544072s" podCreationTimestamp="2025-01-17 12:00:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-17 12:01:00.924031503 +0000 UTC m=+36.179415268" watchObservedRunningTime="2025-01-17 12:01:04.931544072 +0000 UTC m=+40.186927797"
Jan 17 12:01:05.921954 kubelet[2645]: E0117 12:01:05.921912    2645 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8"
Jan 17 12:01:09.724212 systemd[1]: Started sshd@10-10.0.0.35:22-10.0.0.1:56628.service - OpenSSH per-connection server daemon (10.0.0.1:56628).
Jan 17 12:01:09.754756 sshd[3686]: Accepted publickey for core from 10.0.0.1 port 56628 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:09.755927 sshd[3686]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:09.759169 systemd-logind[1520]: New session 11 of user core.
Jan 17 12:01:09.773219 systemd[1]: Started session-11.scope - Session 11 of User core.
Jan 17 12:01:09.879539 sshd[3686]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:09.885196 systemd[1]: Started sshd@11-10.0.0.35:22-10.0.0.1:56636.service - OpenSSH per-connection server daemon (10.0.0.1:56636).
Jan 17 12:01:09.885575 systemd[1]: sshd@10-10.0.0.35:22-10.0.0.1:56628.service: Deactivated successfully.
Jan 17 12:01:09.888662 systemd[1]: session-11.scope: Deactivated successfully.
Jan 17 12:01:09.888994 systemd-logind[1520]: Session 11 logged out. Waiting for processes to exit.
Jan 17 12:01:09.891362 systemd-logind[1520]: Removed session 11.
Jan 17 12:01:09.919285 sshd[3699]: Accepted publickey for core from 10.0.0.1 port 56636 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:09.920540 sshd[3699]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:09.924841 systemd-logind[1520]: New session 12 of user core.
Jan 17 12:01:09.931339 systemd[1]: Started session-12.scope - Session 12 of User core.
Jan 17 12:01:10.127965 sshd[3699]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:10.134203 systemd[1]: Started sshd@12-10.0.0.35:22-10.0.0.1:56642.service - OpenSSH per-connection server daemon (10.0.0.1:56642).
Jan 17 12:01:10.134588 systemd[1]: sshd@11-10.0.0.35:22-10.0.0.1:56636.service: Deactivated successfully.
Jan 17 12:01:10.137736 systemd[1]: session-12.scope: Deactivated successfully.
Jan 17 12:01:10.138551 systemd-logind[1520]: Session 12 logged out. Waiting for processes to exit.
Jan 17 12:01:10.139520 systemd-logind[1520]: Removed session 12.
Jan 17 12:01:10.166110 sshd[3712]: Accepted publickey for core from 10.0.0.1 port 56642 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:10.167287 sshd[3712]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:10.171550 systemd-logind[1520]: New session 13 of user core.
Jan 17 12:01:10.184219 systemd[1]: Started session-13.scope - Session 13 of User core.
Jan 17 12:01:11.239901 sshd[3712]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:11.248267 systemd[1]: Started sshd@13-10.0.0.35:22-10.0.0.1:56648.service - OpenSSH per-connection server daemon (10.0.0.1:56648).
Jan 17 12:01:11.248669 systemd[1]: sshd@12-10.0.0.35:22-10.0.0.1:56642.service: Deactivated successfully.
Jan 17 12:01:11.257272 systemd[1]: session-13.scope: Deactivated successfully.
Jan 17 12:01:11.260359 systemd-logind[1520]: Session 13 logged out. Waiting for processes to exit.
Jan 17 12:01:11.265188 systemd-logind[1520]: Removed session 13.
Jan 17 12:01:11.286859 sshd[3750]: Accepted publickey for core from 10.0.0.1 port 56648 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:11.288182 sshd[3750]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:11.292042 systemd-logind[1520]: New session 14 of user core.
Jan 17 12:01:11.302266 systemd[1]: Started session-14.scope - Session 14 of User core.
Jan 17 12:01:11.509851 sshd[3750]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:11.521610 systemd[1]: Started sshd@14-10.0.0.35:22-10.0.0.1:56652.service - OpenSSH per-connection server daemon (10.0.0.1:56652).
Jan 17 12:01:11.522624 systemd[1]: sshd@13-10.0.0.35:22-10.0.0.1:56648.service: Deactivated successfully.
Jan 17 12:01:11.525522 systemd[1]: session-14.scope: Deactivated successfully.
Jan 17 12:01:11.526900 systemd-logind[1520]: Session 14 logged out. Waiting for processes to exit.
Jan 17 12:01:11.528043 systemd-logind[1520]: Removed session 14.
Jan 17 12:01:11.552849 sshd[3769]: Accepted publickey for core from 10.0.0.1 port 56652 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:11.554188 sshd[3769]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:11.558253 systemd-logind[1520]: New session 15 of user core.
Jan 17 12:01:11.565266 systemd[1]: Started session-15.scope - Session 15 of User core.
Jan 17 12:01:11.669309 sshd[3769]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:11.672053 systemd[1]: sshd@14-10.0.0.35:22-10.0.0.1:56652.service: Deactivated successfully.
Jan 17 12:01:11.675359 systemd[1]: session-15.scope: Deactivated successfully.
Jan 17 12:01:11.676669 systemd-logind[1520]: Session 15 logged out. Waiting for processes to exit.
Jan 17 12:01:11.677842 systemd-logind[1520]: Removed session 15.
Jan 17 12:01:16.680272 systemd[1]: Started sshd@15-10.0.0.35:22-10.0.0.1:44476.service - OpenSSH per-connection server daemon (10.0.0.1:44476).
Jan 17 12:01:16.710850 sshd[3812]: Accepted publickey for core from 10.0.0.1 port 44476 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:16.712077 sshd[3812]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:16.715398 systemd-logind[1520]: New session 16 of user core.
Jan 17 12:01:16.728201 systemd[1]: Started session-16.scope - Session 16 of User core.
Jan 17 12:01:16.831720 sshd[3812]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:16.835045 systemd[1]: sshd@15-10.0.0.35:22-10.0.0.1:44476.service: Deactivated successfully.
Jan 17 12:01:16.837471 systemd-logind[1520]: Session 16 logged out. Waiting for processes to exit.
Jan 17 12:01:16.837792 systemd[1]: session-16.scope: Deactivated successfully.
Jan 17 12:01:16.839131 systemd-logind[1520]: Removed session 16.
Jan 17 12:01:21.847257 systemd[1]: Started sshd@16-10.0.0.35:22-10.0.0.1:44490.service - OpenSSH per-connection server daemon (10.0.0.1:44490).
Jan 17 12:01:21.882949 sshd[3848]: Accepted publickey for core from 10.0.0.1 port 44490 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:21.884647 sshd[3848]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:21.893703 systemd-logind[1520]: New session 17 of user core.
Jan 17 12:01:21.904275 systemd[1]: Started session-17.scope - Session 17 of User core.
Jan 17 12:01:22.010427 sshd[3848]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:22.013603 systemd[1]: sshd@16-10.0.0.35:22-10.0.0.1:44490.service: Deactivated successfully.
Jan 17 12:01:22.015892 systemd[1]: session-17.scope: Deactivated successfully.
Jan 17 12:01:22.016635 systemd-logind[1520]: Session 17 logged out. Waiting for processes to exit.
Jan 17 12:01:22.017471 systemd-logind[1520]: Removed session 17.
Jan 17 12:01:27.021220 systemd[1]: Started sshd@17-10.0.0.35:22-10.0.0.1:41126.service - OpenSSH per-connection server daemon (10.0.0.1:41126).
Jan 17 12:01:27.085541 sshd[3886]: Accepted publickey for core from 10.0.0.1 port 41126 ssh2: RSA SHA256:xsFjL0Ru499iNfhRyIcjP6wTIWZ5oE8f5Pm6hYv+KHo
Jan 17 12:01:27.086877 sshd[3886]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0)
Jan 17 12:01:27.090887 systemd-logind[1520]: New session 18 of user core.
Jan 17 12:01:27.102266 systemd[1]: Started session-18.scope - Session 18 of User core.
Jan 17 12:01:27.212674 sshd[3886]: pam_unix(sshd:session): session closed for user core
Jan 17 12:01:27.217300 systemd[1]: sshd@17-10.0.0.35:22-10.0.0.1:41126.service: Deactivated successfully.
Jan 17 12:01:27.219038 systemd[1]: session-18.scope: Deactivated successfully.
Jan 17 12:01:27.220485 systemd-logind[1520]: Session 18 logged out. Waiting for processes to exit.
Jan 17 12:01:27.222316 systemd-logind[1520]: Removed session 18.