Jun 25 18:47:25.889536 kernel: Linux version 6.6.35-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 13.2.1_p20240210 p14) 13.2.1 20240210, GNU ld (Gentoo 2.41 p5) 2.41.0) #1 SMP PREEMPT_DYNAMIC Tue Jun 25 17:21:28 -00 2024 Jun 25 18:47:25.889556 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=4483672da8ac4c95f5ee13a489103440a13110ce1f63977ab5a6a33d0c137bf8 Jun 25 18:47:25.889567 kernel: BIOS-provided physical RAM map: Jun 25 18:47:25.889573 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009ffff] usable Jun 25 18:47:25.889579 kernel: BIOS-e820: [mem 0x0000000000100000-0x00000000007fffff] usable Jun 25 18:47:25.889585 kernel: BIOS-e820: [mem 0x0000000000800000-0x0000000000807fff] ACPI NVS Jun 25 18:47:25.889593 kernel: BIOS-e820: [mem 0x0000000000808000-0x000000000080afff] usable Jun 25 18:47:25.889599 kernel: BIOS-e820: [mem 0x000000000080b000-0x000000000080bfff] ACPI NVS Jun 25 18:47:25.889605 kernel: BIOS-e820: [mem 0x000000000080c000-0x000000000080ffff] usable Jun 25 18:47:25.889611 kernel: BIOS-e820: [mem 0x0000000000810000-0x00000000008fffff] ACPI NVS Jun 25 18:47:25.889620 kernel: BIOS-e820: [mem 0x0000000000900000-0x000000009c8eefff] usable Jun 25 18:47:25.889626 kernel: BIOS-e820: [mem 0x000000009c8ef000-0x000000009c9eefff] reserved Jun 25 18:47:25.889632 kernel: BIOS-e820: [mem 0x000000009c9ef000-0x000000009caeefff] type 20 Jun 25 18:47:25.889639 kernel: BIOS-e820: [mem 0x000000009caef000-0x000000009cb6efff] reserved Jun 25 18:47:25.889646 kernel: BIOS-e820: [mem 0x000000009cb6f000-0x000000009cb7efff] ACPI data Jun 25 18:47:25.889655 kernel: BIOS-e820: [mem 0x000000009cb7f000-0x000000009cbfefff] ACPI NVS Jun 25 18:47:25.889662 kernel: BIOS-e820: [mem 0x000000009cbff000-0x000000009cf3ffff] usable Jun 25 18:47:25.889669 kernel: BIOS-e820: [mem 0x000000009cf40000-0x000000009cf5ffff] reserved Jun 25 18:47:25.889676 kernel: BIOS-e820: [mem 0x000000009cf60000-0x000000009cffffff] ACPI NVS Jun 25 18:47:25.889682 kernel: NX (Execute Disable) protection: active Jun 25 18:47:25.889689 kernel: APIC: Static calls initialized Jun 25 18:47:25.889696 kernel: efi: EFI v2.7 by EDK II Jun 25 18:47:25.889702 kernel: efi: SMBIOS=0x9c9ab000 ACPI=0x9cb7e000 ACPI 2.0=0x9cb7e014 MEMATTR=0x9b4f9018 Jun 25 18:47:25.889709 kernel: SMBIOS 2.8 present. Jun 25 18:47:25.889716 kernel: DMI: QEMU Standard PC (i440FX + PIIX, 1996), BIOS 0.0.0 02/06/2015 Jun 25 18:47:25.889723 kernel: Hypervisor detected: KVM Jun 25 18:47:25.889730 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Jun 25 18:47:25.889739 kernel: kvm-clock: using sched offset of 4221946074 cycles Jun 25 18:47:25.889746 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Jun 25 18:47:25.889753 kernel: tsc: Detected 2794.750 MHz processor Jun 25 18:47:25.889760 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Jun 25 18:47:25.889767 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Jun 25 18:47:25.889774 kernel: last_pfn = 0x9cf40 max_arch_pfn = 0x400000000 Jun 25 18:47:25.889781 kernel: MTRR map: 4 entries (2 fixed + 2 variable; max 18), built from 8 variable MTRRs Jun 25 18:47:25.889788 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Jun 25 18:47:25.889795 kernel: Using GB pages for direct mapping Jun 25 18:47:25.889812 kernel: Secure boot disabled Jun 25 18:47:25.889819 kernel: ACPI: Early table checksum verification disabled Jun 25 18:47:25.889826 kernel: ACPI: RSDP 0x000000009CB7E014 000024 (v02 BOCHS ) Jun 25 18:47:25.889834 kernel: ACPI: XSDT 0x000000009CB7D0E8 00004C (v01 BOCHS BXPC 00000001 01000013) Jun 25 18:47:25.889926 kernel: ACPI: FACP 0x000000009CB7A000 000074 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jun 25 18:47:25.889934 kernel: ACPI: DSDT 0x000000009CB7B000 001A39 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jun 25 18:47:25.889943 kernel: ACPI: FACS 0x000000009CBDD000 000040 Jun 25 18:47:25.889951 kernel: ACPI: APIC 0x000000009CB79000 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jun 25 18:47:25.889958 kernel: ACPI: HPET 0x000000009CB78000 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jun 25 18:47:25.889965 kernel: ACPI: WAET 0x000000009CB77000 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Jun 25 18:47:25.889972 kernel: ACPI: BGRT 0x000000009CB76000 000038 (v01 INTEL EDK2 00000002 01000013) Jun 25 18:47:25.889979 kernel: ACPI: Reserving FACP table memory at [mem 0x9cb7a000-0x9cb7a073] Jun 25 18:47:25.889986 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cb7b000-0x9cb7ca38] Jun 25 18:47:25.889993 kernel: ACPI: Reserving FACS table memory at [mem 0x9cbdd000-0x9cbdd03f] Jun 25 18:47:25.890003 kernel: ACPI: Reserving APIC table memory at [mem 0x9cb79000-0x9cb7908f] Jun 25 18:47:25.890010 kernel: ACPI: Reserving HPET table memory at [mem 0x9cb78000-0x9cb78037] Jun 25 18:47:25.890017 kernel: ACPI: Reserving WAET table memory at [mem 0x9cb77000-0x9cb77027] Jun 25 18:47:25.890024 kernel: ACPI: Reserving BGRT table memory at [mem 0x9cb76000-0x9cb76037] Jun 25 18:47:25.890031 kernel: No NUMA configuration found Jun 25 18:47:25.890038 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cf3ffff] Jun 25 18:47:25.890046 kernel: NODE_DATA(0) allocated [mem 0x9cea6000-0x9ceabfff] Jun 25 18:47:25.890053 kernel: Zone ranges: Jun 25 18:47:25.890060 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Jun 25 18:47:25.890070 kernel: DMA32 [mem 0x0000000001000000-0x000000009cf3ffff] Jun 25 18:47:25.890076 kernel: Normal empty Jun 25 18:47:25.890084 kernel: Movable zone start for each node Jun 25 18:47:25.890091 kernel: Early memory node ranges Jun 25 18:47:25.890098 kernel: node 0: [mem 0x0000000000001000-0x000000000009ffff] Jun 25 18:47:25.890105 kernel: node 0: [mem 0x0000000000100000-0x00000000007fffff] Jun 25 18:47:25.890112 kernel: node 0: [mem 0x0000000000808000-0x000000000080afff] Jun 25 18:47:25.890119 kernel: node 0: [mem 0x000000000080c000-0x000000000080ffff] Jun 25 18:47:25.890126 kernel: node 0: [mem 0x0000000000900000-0x000000009c8eefff] Jun 25 18:47:25.890133 kernel: node 0: [mem 0x000000009cbff000-0x000000009cf3ffff] Jun 25 18:47:25.890143 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cf3ffff] Jun 25 18:47:25.890150 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jun 25 18:47:25.890157 kernel: On node 0, zone DMA: 96 pages in unavailable ranges Jun 25 18:47:25.890164 kernel: On node 0, zone DMA: 8 pages in unavailable ranges Jun 25 18:47:25.890171 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Jun 25 18:47:25.890178 kernel: On node 0, zone DMA: 240 pages in unavailable ranges Jun 25 18:47:25.890185 kernel: On node 0, zone DMA32: 784 pages in unavailable ranges Jun 25 18:47:25.890192 kernel: On node 0, zone DMA32: 12480 pages in unavailable ranges Jun 25 18:47:25.890200 kernel: ACPI: PM-Timer IO Port: 0xb008 Jun 25 18:47:25.890209 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Jun 25 18:47:25.890216 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Jun 25 18:47:25.890223 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Jun 25 18:47:25.890230 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Jun 25 18:47:25.890237 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Jun 25 18:47:25.890244 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Jun 25 18:47:25.890251 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Jun 25 18:47:25.890258 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Jun 25 18:47:25.890265 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Jun 25 18:47:25.890275 kernel: TSC deadline timer available Jun 25 18:47:25.890282 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Jun 25 18:47:25.890289 kernel: kvm-guest: APIC: eoi() replaced with kvm_guest_apic_eoi_write() Jun 25 18:47:25.890296 kernel: kvm-guest: KVM setup pv remote TLB flush Jun 25 18:47:25.890303 kernel: kvm-guest: setup PV sched yield Jun 25 18:47:25.890310 kernel: [mem 0x9d000000-0xffffffff] available for PCI devices Jun 25 18:47:25.890317 kernel: Booting paravirtualized kernel on KVM Jun 25 18:47:25.890325 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Jun 25 18:47:25.890332 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:4 nr_cpu_ids:4 nr_node_ids:1 Jun 25 18:47:25.890342 kernel: percpu: Embedded 58 pages/cpu s196904 r8192 d32472 u524288 Jun 25 18:47:25.890349 kernel: pcpu-alloc: s196904 r8192 d32472 u524288 alloc=1*2097152 Jun 25 18:47:25.890356 kernel: pcpu-alloc: [0] 0 1 2 3 Jun 25 18:47:25.890363 kernel: kvm-guest: PV spinlocks enabled Jun 25 18:47:25.890370 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Jun 25 18:47:25.890379 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=4483672da8ac4c95f5ee13a489103440a13110ce1f63977ab5a6a33d0c137bf8 Jun 25 18:47:25.890387 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jun 25 18:47:25.890394 kernel: random: crng init done Jun 25 18:47:25.890404 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jun 25 18:47:25.890411 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jun 25 18:47:25.890418 kernel: Fallback order for Node 0: 0 Jun 25 18:47:25.890425 kernel: Built 1 zonelists, mobility grouping on. Total pages: 629759 Jun 25 18:47:25.890432 kernel: Policy zone: DMA32 Jun 25 18:47:25.890439 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jun 25 18:47:25.890447 kernel: Memory: 2388204K/2567000K available (12288K kernel code, 2302K rwdata, 22636K rodata, 49384K init, 1964K bss, 178536K reserved, 0K cma-reserved) Jun 25 18:47:25.890455 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Jun 25 18:47:25.890462 kernel: ftrace: allocating 37650 entries in 148 pages Jun 25 18:47:25.890471 kernel: ftrace: allocated 148 pages with 3 groups Jun 25 18:47:25.890478 kernel: Dynamic Preempt: voluntary Jun 25 18:47:25.890485 kernel: rcu: Preemptible hierarchical RCU implementation. Jun 25 18:47:25.890493 kernel: rcu: RCU event tracing is enabled. Jun 25 18:47:25.890501 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Jun 25 18:47:25.890517 kernel: Trampoline variant of Tasks RCU enabled. Jun 25 18:47:25.890525 kernel: Rude variant of Tasks RCU enabled. Jun 25 18:47:25.890532 kernel: Tracing variant of Tasks RCU enabled. Jun 25 18:47:25.890540 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jun 25 18:47:25.890548 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Jun 25 18:47:25.890555 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Jun 25 18:47:25.890562 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jun 25 18:47:25.890572 kernel: Console: colour dummy device 80x25 Jun 25 18:47:25.890579 kernel: printk: console [ttyS0] enabled Jun 25 18:47:25.890587 kernel: ACPI: Core revision 20230628 Jun 25 18:47:25.890595 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Jun 25 18:47:25.890602 kernel: APIC: Switch to symmetric I/O mode setup Jun 25 18:47:25.890611 kernel: x2apic enabled Jun 25 18:47:25.890619 kernel: APIC: Switched APIC routing to: physical x2apic Jun 25 18:47:25.890626 kernel: kvm-guest: APIC: send_IPI_mask() replaced with kvm_send_ipi_mask() Jun 25 18:47:25.890634 kernel: kvm-guest: APIC: send_IPI_mask_allbutself() replaced with kvm_send_ipi_mask_allbutself() Jun 25 18:47:25.890641 kernel: kvm-guest: setup PV IPIs Jun 25 18:47:25.890649 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Jun 25 18:47:25.890657 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Jun 25 18:47:25.890664 kernel: Calibrating delay loop (skipped) preset value.. 5589.50 BogoMIPS (lpj=2794750) Jun 25 18:47:25.890672 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Jun 25 18:47:25.890681 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Jun 25 18:47:25.890689 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Jun 25 18:47:25.890696 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Jun 25 18:47:25.890704 kernel: Spectre V2 : Mitigation: Retpolines Jun 25 18:47:25.890711 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Jun 25 18:47:25.890718 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Jun 25 18:47:25.890726 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Jun 25 18:47:25.890733 kernel: RETBleed: Mitigation: untrained return thunk Jun 25 18:47:25.890741 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Jun 25 18:47:25.890751 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl Jun 25 18:47:25.890758 kernel: Speculative Return Stack Overflow: IBPB-extending microcode not applied! Jun 25 18:47:25.890766 kernel: Speculative Return Stack Overflow: WARNING: See https://kernel.org/doc/html/latest/admin-guide/hw-vuln/srso.html for mitigation options. Jun 25 18:47:25.890774 kernel: Speculative Return Stack Overflow: Vulnerable: Safe RET, no microcode Jun 25 18:47:25.890781 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Jun 25 18:47:25.890789 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Jun 25 18:47:25.890796 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Jun 25 18:47:25.890810 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Jun 25 18:47:25.890818 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'compacted' format. Jun 25 18:47:25.890828 kernel: Freeing SMP alternatives memory: 32K Jun 25 18:47:25.890835 kernel: pid_max: default: 32768 minimum: 301 Jun 25 18:47:25.890852 kernel: LSM: initializing lsm=lockdown,capability,selinux,integrity Jun 25 18:47:25.890860 kernel: SELinux: Initializing. Jun 25 18:47:25.890868 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jun 25 18:47:25.890878 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jun 25 18:47:25.890888 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Jun 25 18:47:25.890899 kernel: RCU Tasks: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1. Jun 25 18:47:25.890913 kernel: RCU Tasks Rude: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1. Jun 25 18:47:25.890922 kernel: RCU Tasks Trace: Setting shift to 2 and lim to 1 rcu_task_cb_adjust=1. Jun 25 18:47:25.890930 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Jun 25 18:47:25.890937 kernel: ... version: 0 Jun 25 18:47:25.890945 kernel: ... bit width: 48 Jun 25 18:47:25.890952 kernel: ... generic registers: 6 Jun 25 18:47:25.890960 kernel: ... value mask: 0000ffffffffffff Jun 25 18:47:25.890967 kernel: ... max period: 00007fffffffffff Jun 25 18:47:25.890974 kernel: ... fixed-purpose events: 0 Jun 25 18:47:25.890984 kernel: ... event mask: 000000000000003f Jun 25 18:47:25.890991 kernel: signal: max sigframe size: 1776 Jun 25 18:47:25.890999 kernel: rcu: Hierarchical SRCU implementation. Jun 25 18:47:25.891006 kernel: rcu: Max phase no-delay instances is 400. Jun 25 18:47:25.891014 kernel: smp: Bringing up secondary CPUs ... Jun 25 18:47:25.891021 kernel: smpboot: x86: Booting SMP configuration: Jun 25 18:47:25.891029 kernel: .... node #0, CPUs: #1 #2 #3 Jun 25 18:47:25.891036 kernel: smp: Brought up 1 node, 4 CPUs Jun 25 18:47:25.891044 kernel: smpboot: Max logical packages: 1 Jun 25 18:47:25.891051 kernel: smpboot: Total of 4 processors activated (22358.00 BogoMIPS) Jun 25 18:47:25.891061 kernel: devtmpfs: initialized Jun 25 18:47:25.891068 kernel: x86/mm: Memory block size: 128MB Jun 25 18:47:25.891076 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00800000-0x00807fff] (32768 bytes) Jun 25 18:47:25.891083 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x0080b000-0x0080bfff] (4096 bytes) Jun 25 18:47:25.891091 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x00810000-0x008fffff] (983040 bytes) Jun 25 18:47:25.891099 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cb7f000-0x9cbfefff] (524288 bytes) Jun 25 18:47:25.891106 kernel: ACPI: PM: Registering ACPI NVS region [mem 0x9cf60000-0x9cffffff] (655360 bytes) Jun 25 18:47:25.891114 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jun 25 18:47:25.891121 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Jun 25 18:47:25.891131 kernel: pinctrl core: initialized pinctrl subsystem Jun 25 18:47:25.891139 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jun 25 18:47:25.891146 kernel: audit: initializing netlink subsys (disabled) Jun 25 18:47:25.891154 kernel: audit: type=2000 audit(1719341245.601:1): state=initialized audit_enabled=0 res=1 Jun 25 18:47:25.891161 kernel: thermal_sys: Registered thermal governor 'step_wise' Jun 25 18:47:25.891169 kernel: thermal_sys: Registered thermal governor 'user_space' Jun 25 18:47:25.891176 kernel: cpuidle: using governor menu Jun 25 18:47:25.891184 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jun 25 18:47:25.891191 kernel: dca service started, version 1.12.1 Jun 25 18:47:25.891201 kernel: PCI: Using configuration type 1 for base access Jun 25 18:47:25.891209 kernel: PCI: Using configuration type 1 for extended access Jun 25 18:47:25.891216 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Jun 25 18:47:25.891224 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jun 25 18:47:25.891231 kernel: HugeTLB: 16380 KiB vmemmap can be freed for a 1.00 GiB page Jun 25 18:47:25.891239 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jun 25 18:47:25.891246 kernel: HugeTLB: 28 KiB vmemmap can be freed for a 2.00 MiB page Jun 25 18:47:25.891254 kernel: ACPI: Added _OSI(Module Device) Jun 25 18:47:25.891263 kernel: ACPI: Added _OSI(Processor Device) Jun 25 18:47:25.891271 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jun 25 18:47:25.891278 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jun 25 18:47:25.891286 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jun 25 18:47:25.891293 kernel: ACPI: _OSC evaluation for CPUs failed, trying _PDC Jun 25 18:47:25.891301 kernel: ACPI: Interpreter enabled Jun 25 18:47:25.891308 kernel: ACPI: PM: (supports S0 S3 S5) Jun 25 18:47:25.891316 kernel: ACPI: Using IOAPIC for interrupt routing Jun 25 18:47:25.891323 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Jun 25 18:47:25.891331 kernel: PCI: Using E820 reservations for host bridge windows Jun 25 18:47:25.891341 kernel: ACPI: Enabled 2 GPEs in block 00 to 0F Jun 25 18:47:25.891348 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Jun 25 18:47:25.891523 kernel: acpi PNP0A03:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Jun 25 18:47:25.891535 kernel: acpiphp: Slot [3] registered Jun 25 18:47:25.891542 kernel: acpiphp: Slot [4] registered Jun 25 18:47:25.891550 kernel: acpiphp: Slot [5] registered Jun 25 18:47:25.891557 kernel: acpiphp: Slot [6] registered Jun 25 18:47:25.891564 kernel: acpiphp: Slot [7] registered Jun 25 18:47:25.891575 kernel: acpiphp: Slot [8] registered Jun 25 18:47:25.891582 kernel: acpiphp: Slot [9] registered Jun 25 18:47:25.891590 kernel: acpiphp: Slot [10] registered Jun 25 18:47:25.891597 kernel: acpiphp: Slot [11] registered Jun 25 18:47:25.891604 kernel: acpiphp: Slot [12] registered Jun 25 18:47:25.891612 kernel: acpiphp: Slot [13] registered Jun 25 18:47:25.891619 kernel: acpiphp: Slot [14] registered Jun 25 18:47:25.891626 kernel: acpiphp: Slot [15] registered Jun 25 18:47:25.891634 kernel: acpiphp: Slot [16] registered Jun 25 18:47:25.891643 kernel: acpiphp: Slot [17] registered Jun 25 18:47:25.891650 kernel: acpiphp: Slot [18] registered Jun 25 18:47:25.891658 kernel: acpiphp: Slot [19] registered Jun 25 18:47:25.891665 kernel: acpiphp: Slot [20] registered Jun 25 18:47:25.891672 kernel: acpiphp: Slot [21] registered Jun 25 18:47:25.891680 kernel: acpiphp: Slot [22] registered Jun 25 18:47:25.891687 kernel: acpiphp: Slot [23] registered Jun 25 18:47:25.891694 kernel: acpiphp: Slot [24] registered Jun 25 18:47:25.891702 kernel: acpiphp: Slot [25] registered Jun 25 18:47:25.891709 kernel: acpiphp: Slot [26] registered Jun 25 18:47:25.891718 kernel: acpiphp: Slot [27] registered Jun 25 18:47:25.891726 kernel: acpiphp: Slot [28] registered Jun 25 18:47:25.891733 kernel: acpiphp: Slot [29] registered Jun 25 18:47:25.891740 kernel: acpiphp: Slot [30] registered Jun 25 18:47:25.891748 kernel: acpiphp: Slot [31] registered Jun 25 18:47:25.891755 kernel: PCI host bridge to bus 0000:00 Jun 25 18:47:25.891907 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Jun 25 18:47:25.892021 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Jun 25 18:47:25.892134 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Jun 25 18:47:25.892242 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xfebfffff window] Jun 25 18:47:25.892350 kernel: pci_bus 0000:00: root bus resource [mem 0x800000000-0x87fffffff window] Jun 25 18:47:25.892457 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Jun 25 18:47:25.892596 kernel: pci 0000:00:00.0: [8086:1237] type 00 class 0x060000 Jun 25 18:47:25.892726 kernel: pci 0000:00:01.0: [8086:7000] type 00 class 0x060100 Jun 25 18:47:25.892885 kernel: pci 0000:00:01.1: [8086:7010] type 00 class 0x010180 Jun 25 18:47:25.893007 kernel: pci 0000:00:01.1: reg 0x20: [io 0xc0c0-0xc0cf] Jun 25 18:47:25.893126 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x10: [io 0x01f0-0x01f7] Jun 25 18:47:25.893246 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x14: [io 0x03f6] Jun 25 18:47:25.893367 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x18: [io 0x0170-0x0177] Jun 25 18:47:25.893486 kernel: pci 0000:00:01.1: legacy IDE quirk: reg 0x1c: [io 0x0376] Jun 25 18:47:25.893618 kernel: pci 0000:00:01.3: [8086:7113] type 00 class 0x068000 Jun 25 18:47:25.893829 kernel: pci 0000:00:01.3: quirk: [io 0xb000-0xb03f] claimed by PIIX4 ACPI Jun 25 18:47:25.893967 kernel: pci 0000:00:01.3: quirk: [io 0xb100-0xb10f] claimed by PIIX4 SMB Jun 25 18:47:25.894096 kernel: pci 0000:00:02.0: [1234:1111] type 00 class 0x030000 Jun 25 18:47:25.894216 kernel: pci 0000:00:02.0: reg 0x10: [mem 0xc0000000-0xc0ffffff pref] Jun 25 18:47:25.894334 kernel: pci 0000:00:02.0: reg 0x18: [mem 0xc1043000-0xc1043fff] Jun 25 18:47:25.894453 kernel: pci 0000:00:02.0: reg 0x30: [mem 0xffff0000-0xffffffff pref] Jun 25 18:47:25.894574 kernel: pci 0000:00:02.0: BAR 0: assigned to efifb Jun 25 18:47:25.894700 kernel: pci 0000:00:02.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Jun 25 18:47:25.894838 kernel: pci 0000:00:03.0: [1af4:1005] type 00 class 0x00ff00 Jun 25 18:47:25.895018 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc0a0-0xc0bf] Jun 25 18:47:25.895192 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xc1042000-0xc1042fff] Jun 25 18:47:25.895352 kernel: pci 0000:00:03.0: reg 0x20: [mem 0x800000000-0x800003fff 64bit pref] Jun 25 18:47:25.895484 kernel: pci 0000:00:04.0: [1af4:1001] type 00 class 0x010000 Jun 25 18:47:25.895610 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc000-0xc07f] Jun 25 18:47:25.895729 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xc1041000-0xc1041fff] Jun 25 18:47:25.895875 kernel: pci 0000:00:04.0: reg 0x20: [mem 0x800004000-0x800007fff 64bit pref] Jun 25 18:47:25.896005 kernel: pci 0000:00:05.0: [1af4:1000] type 00 class 0x020000 Jun 25 18:47:25.896138 kernel: pci 0000:00:05.0: reg 0x10: [io 0xc080-0xc09f] Jun 25 18:47:25.896259 kernel: pci 0000:00:05.0: reg 0x14: [mem 0xc1040000-0xc1040fff] Jun 25 18:47:25.896379 kernel: pci 0000:00:05.0: reg 0x20: [mem 0x800008000-0x80000bfff 64bit pref] Jun 25 18:47:25.896498 kernel: pci 0000:00:05.0: reg 0x30: [mem 0xfffc0000-0xffffffff pref] Jun 25 18:47:25.896512 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Jun 25 18:47:25.896520 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Jun 25 18:47:25.896528 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Jun 25 18:47:25.896535 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Jun 25 18:47:25.896542 kernel: ACPI: PCI: Interrupt link LNKS configured for IRQ 9 Jun 25 18:47:25.896550 kernel: iommu: Default domain type: Translated Jun 25 18:47:25.896558 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Jun 25 18:47:25.896565 kernel: efivars: Registered efivars operations Jun 25 18:47:25.896572 kernel: PCI: Using ACPI for IRQ routing Jun 25 18:47:25.896582 kernel: PCI: pci_cache_line_size set to 64 bytes Jun 25 18:47:25.896590 kernel: e820: reserve RAM buffer [mem 0x0080b000-0x008fffff] Jun 25 18:47:25.896597 kernel: e820: reserve RAM buffer [mem 0x00810000-0x008fffff] Jun 25 18:47:25.896604 kernel: e820: reserve RAM buffer [mem 0x9c8ef000-0x9fffffff] Jun 25 18:47:25.896611 kernel: e820: reserve RAM buffer [mem 0x9cf40000-0x9fffffff] Jun 25 18:47:25.896731 kernel: pci 0000:00:02.0: vgaarb: setting as boot VGA device Jun 25 18:47:25.896873 kernel: pci 0000:00:02.0: vgaarb: bridge control possible Jun 25 18:47:25.896996 kernel: pci 0000:00:02.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Jun 25 18:47:25.897010 kernel: vgaarb: loaded Jun 25 18:47:25.897018 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Jun 25 18:47:25.897025 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Jun 25 18:47:25.897033 kernel: clocksource: Switched to clocksource kvm-clock Jun 25 18:47:25.897040 kernel: VFS: Disk quotas dquot_6.6.0 Jun 25 18:47:25.897048 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jun 25 18:47:25.897055 kernel: pnp: PnP ACPI init Jun 25 18:47:25.897181 kernel: pnp 00:02: [dma 2] Jun 25 18:47:25.897195 kernel: pnp: PnP ACPI: found 6 devices Jun 25 18:47:25.897203 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Jun 25 18:47:25.897210 kernel: NET: Registered PF_INET protocol family Jun 25 18:47:25.897218 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jun 25 18:47:25.897225 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jun 25 18:47:25.897233 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jun 25 18:47:25.897240 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jun 25 18:47:25.897248 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jun 25 18:47:25.897256 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jun 25 18:47:25.897266 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jun 25 18:47:25.897274 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jun 25 18:47:25.897281 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jun 25 18:47:25.897288 kernel: NET: Registered PF_XDP protocol family Jun 25 18:47:25.897451 kernel: pci 0000:00:05.0: can't claim BAR 6 [mem 0xfffc0000-0xffffffff pref]: no compatible bridge window Jun 25 18:47:25.897603 kernel: pci 0000:00:05.0: BAR 6: assigned [mem 0x9d000000-0x9d03ffff pref] Jun 25 18:47:25.897719 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Jun 25 18:47:25.897855 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Jun 25 18:47:25.897990 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Jun 25 18:47:25.898101 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xfebfffff window] Jun 25 18:47:25.898212 kernel: pci_bus 0000:00: resource 8 [mem 0x800000000-0x87fffffff window] Jun 25 18:47:25.898334 kernel: pci 0000:00:01.0: PIIX3: Enabling Passive Release Jun 25 18:47:25.898514 kernel: pci 0000:00:00.0: Limiting direct PCI/PCI transfers Jun 25 18:47:25.898526 kernel: PCI: CLS 0 bytes, default 64 Jun 25 18:47:25.898534 kernel: Initialise system trusted keyrings Jun 25 18:47:25.898541 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jun 25 18:47:25.898552 kernel: Key type asymmetric registered Jun 25 18:47:25.898560 kernel: Asymmetric key parser 'x509' registered Jun 25 18:47:25.898567 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 251) Jun 25 18:47:25.898575 kernel: io scheduler mq-deadline registered Jun 25 18:47:25.898582 kernel: io scheduler kyber registered Jun 25 18:47:25.898589 kernel: io scheduler bfq registered Jun 25 18:47:25.898597 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Jun 25 18:47:25.898605 kernel: ACPI: \_SB_.LNKC: Enabled at IRQ 11 Jun 25 18:47:25.898613 kernel: ACPI: \_SB_.LNKD: Enabled at IRQ 10 Jun 25 18:47:25.898623 kernel: ACPI: \_SB_.LNKA: Enabled at IRQ 10 Jun 25 18:47:25.898630 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jun 25 18:47:25.898638 kernel: 00:04: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Jun 25 18:47:25.898645 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Jun 25 18:47:25.898669 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Jun 25 18:47:25.898679 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Jun 25 18:47:25.898818 kernel: rtc_cmos 00:05: RTC can wake from S4 Jun 25 18:47:25.898830 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Jun 25 18:47:25.899031 kernel: rtc_cmos 00:05: registered as rtc0 Jun 25 18:47:25.899148 kernel: rtc_cmos 00:05: setting system clock to 2024-06-25T18:47:25 UTC (1719341245) Jun 25 18:47:25.899258 kernel: rtc_cmos 00:05: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Jun 25 18:47:25.899269 kernel: amd_pstate: the _CPC object is not present in SBIOS or ACPI disabled Jun 25 18:47:25.899277 kernel: efifb: probing for efifb Jun 25 18:47:25.899284 kernel: efifb: framebuffer at 0xc0000000, using 1408k, total 1408k Jun 25 18:47:25.899292 kernel: efifb: mode is 800x600x24, linelength=2400, pages=1 Jun 25 18:47:25.899300 kernel: efifb: scrolling: redraw Jun 25 18:47:25.899308 kernel: efifb: Truecolor: size=0:8:8:8, shift=0:16:8:0 Jun 25 18:47:25.899320 kernel: Console: switching to colour frame buffer device 100x37 Jun 25 18:47:25.899330 kernel: fb0: EFI VGA frame buffer device Jun 25 18:47:25.899338 kernel: pstore: Using crash dump compression: deflate Jun 25 18:47:25.899350 kernel: pstore: Registered efi_pstore as persistent store backend Jun 25 18:47:25.899358 kernel: NET: Registered PF_INET6 protocol family Jun 25 18:47:25.899366 kernel: Segment Routing with IPv6 Jun 25 18:47:25.899373 kernel: In-situ OAM (IOAM) with IPv6 Jun 25 18:47:25.899381 kernel: NET: Registered PF_PACKET protocol family Jun 25 18:47:25.899389 kernel: Key type dns_resolver registered Jun 25 18:47:25.899399 kernel: IPI shorthand broadcast: enabled Jun 25 18:47:25.899407 kernel: sched_clock: Marking stable (690002730, 111960906)->(814978875, -13015239) Jun 25 18:47:25.899416 kernel: registered taskstats version 1 Jun 25 18:47:25.899424 kernel: Loading compiled-in X.509 certificates Jun 25 18:47:25.899432 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.35-flatcar: 60204e9db5f484c670a1c92aec37e9a0c4d3ae90' Jun 25 18:47:25.899440 kernel: Key type .fscrypt registered Jun 25 18:47:25.899450 kernel: Key type fscrypt-provisioning registered Jun 25 18:47:25.899458 kernel: ima: No TPM chip found, activating TPM-bypass! Jun 25 18:47:25.899465 kernel: ima: Allocated hash algorithm: sha1 Jun 25 18:47:25.899473 kernel: ima: No architecture policies found Jun 25 18:47:25.899481 kernel: clk: Disabling unused clocks Jun 25 18:47:25.899489 kernel: Freeing unused kernel image (initmem) memory: 49384K Jun 25 18:47:25.899496 kernel: Write protecting the kernel read-only data: 36864k Jun 25 18:47:25.899504 kernel: Freeing unused kernel image (rodata/data gap) memory: 1940K Jun 25 18:47:25.899512 kernel: Run /init as init process Jun 25 18:47:25.899522 kernel: with arguments: Jun 25 18:47:25.899530 kernel: /init Jun 25 18:47:25.899537 kernel: with environment: Jun 25 18:47:25.899545 kernel: HOME=/ Jun 25 18:47:25.899552 kernel: TERM=linux Jun 25 18:47:25.899560 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jun 25 18:47:25.899569 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jun 25 18:47:25.899581 systemd[1]: Detected virtualization kvm. Jun 25 18:47:25.899590 systemd[1]: Detected architecture x86-64. Jun 25 18:47:25.899598 systemd[1]: Running in initrd. Jun 25 18:47:25.899606 systemd[1]: No hostname configured, using default hostname. Jun 25 18:47:25.899614 systemd[1]: Hostname set to . Jun 25 18:47:25.899622 systemd[1]: Initializing machine ID from VM UUID. Jun 25 18:47:25.899631 systemd[1]: Queued start job for default target initrd.target. Jun 25 18:47:25.899639 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jun 25 18:47:25.899650 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jun 25 18:47:25.899659 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jun 25 18:47:25.899667 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jun 25 18:47:25.899675 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jun 25 18:47:25.899684 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jun 25 18:47:25.899694 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jun 25 18:47:25.899703 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jun 25 18:47:25.899713 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jun 25 18:47:25.899722 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jun 25 18:47:25.899730 systemd[1]: Reached target paths.target - Path Units. Jun 25 18:47:25.899738 systemd[1]: Reached target slices.target - Slice Units. Jun 25 18:47:25.899746 systemd[1]: Reached target swap.target - Swaps. Jun 25 18:47:25.899754 systemd[1]: Reached target timers.target - Timer Units. Jun 25 18:47:25.899763 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jun 25 18:47:25.899771 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jun 25 18:47:25.899781 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jun 25 18:47:25.899789 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jun 25 18:47:25.899798 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jun 25 18:47:25.899815 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jun 25 18:47:25.899823 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jun 25 18:47:25.899832 systemd[1]: Reached target sockets.target - Socket Units. Jun 25 18:47:25.899850 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jun 25 18:47:25.899859 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jun 25 18:47:25.899867 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jun 25 18:47:25.899886 systemd[1]: Starting systemd-fsck-usr.service... Jun 25 18:47:25.899901 systemd[1]: Starting systemd-journald.service - Journal Service... Jun 25 18:47:25.899909 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jun 25 18:47:25.899918 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jun 25 18:47:25.899926 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jun 25 18:47:25.899934 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jun 25 18:47:25.899943 systemd[1]: Finished systemd-fsck-usr.service. Jun 25 18:47:25.899974 systemd-journald[193]: Collecting audit messages is disabled. Jun 25 18:47:25.899995 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jun 25 18:47:25.900004 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jun 25 18:47:25.900012 systemd-journald[193]: Journal started Jun 25 18:47:25.900029 systemd-journald[193]: Runtime Journal (/run/log/journal/e59d93c2bcf6416384c1fd403fc9f628) is 6.0M, max 48.3M, 42.3M free. Jun 25 18:47:25.900065 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jun 25 18:47:25.896112 systemd-modules-load[194]: Inserted module 'overlay' Jun 25 18:47:25.905002 systemd[1]: Started systemd-journald.service - Journal Service. Jun 25 18:47:25.905428 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jun 25 18:47:25.911069 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jun 25 18:47:25.914128 systemd[1]: Starting systemd-tmpfiles-setup.service - Create Volatile Files and Directories... Jun 25 18:47:25.924482 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jun 25 18:47:25.927648 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jun 25 18:47:25.930578 systemd[1]: Finished systemd-tmpfiles-setup.service - Create Volatile Files and Directories. Jun 25 18:47:25.934544 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jun 25 18:47:25.936061 systemd-modules-load[194]: Inserted module 'br_netfilter' Jun 25 18:47:25.937941 kernel: Bridge firewalling registered Jun 25 18:47:25.936110 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jun 25 18:47:25.937118 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jun 25 18:47:25.940206 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jun 25 18:47:25.952167 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jun 25 18:47:25.954030 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jun 25 18:47:25.962505 dracut-cmdline[222]: dracut-dracut-053 Jun 25 18:47:25.965311 dracut-cmdline[222]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=4483672da8ac4c95f5ee13a489103440a13110ce1f63977ab5a6a33d0c137bf8 Jun 25 18:47:25.986861 systemd-resolved[231]: Positive Trust Anchors: Jun 25 18:47:25.986875 systemd-resolved[231]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jun 25 18:47:25.986906 systemd-resolved[231]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa corp home internal intranet lan local private test Jun 25 18:47:25.989415 systemd-resolved[231]: Defaulting to hostname 'linux'. Jun 25 18:47:25.990396 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jun 25 18:47:25.996416 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jun 25 18:47:26.068882 kernel: SCSI subsystem initialized Jun 25 18:47:26.079875 kernel: Loading iSCSI transport class v2.0-870. Jun 25 18:47:26.092891 kernel: iscsi: registered transport (tcp) Jun 25 18:47:26.117881 kernel: iscsi: registered transport (qla4xxx) Jun 25 18:47:26.117966 kernel: QLogic iSCSI HBA Driver Jun 25 18:47:26.176279 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jun 25 18:47:26.187977 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jun 25 18:47:26.214306 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jun 25 18:47:26.214337 kernel: device-mapper: uevent: version 1.0.3 Jun 25 18:47:26.215346 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jun 25 18:47:26.260870 kernel: raid6: avx2x4 gen() 30421 MB/s Jun 25 18:47:26.277866 kernel: raid6: avx2x2 gen() 24979 MB/s Jun 25 18:47:26.295087 kernel: raid6: avx2x1 gen() 23860 MB/s Jun 25 18:47:26.295108 kernel: raid6: using algorithm avx2x4 gen() 30421 MB/s Jun 25 18:47:26.312966 kernel: raid6: .... xor() 7813 MB/s, rmw enabled Jun 25 18:47:26.312980 kernel: raid6: using avx2x2 recovery algorithm Jun 25 18:47:26.339868 kernel: xor: automatically using best checksumming function avx Jun 25 18:47:26.539874 kernel: Btrfs loaded, zoned=no, fsverity=no Jun 25 18:47:26.554114 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jun 25 18:47:26.564008 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jun 25 18:47:26.577018 systemd-udevd[412]: Using default interface naming scheme 'v255'. Jun 25 18:47:26.581232 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jun 25 18:47:26.589201 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jun 25 18:47:26.604805 dracut-pre-trigger[416]: rd.md=0: removing MD RAID activation Jun 25 18:47:26.641774 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jun 25 18:47:26.659973 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jun 25 18:47:26.725621 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jun 25 18:47:26.739326 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jun 25 18:47:26.752149 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jun 25 18:47:26.754966 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jun 25 18:47:26.756655 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jun 25 18:47:26.757911 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jun 25 18:47:26.766860 kernel: virtio_blk virtio1: 4/0/0 default/read/poll queues Jun 25 18:47:26.800030 kernel: cryptd: max_cpu_qlen set to 1000 Jun 25 18:47:26.800060 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Jun 25 18:47:26.800242 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Jun 25 18:47:26.800257 kernel: GPT:9289727 != 19775487 Jun 25 18:47:26.800271 kernel: GPT:Alternate GPT header not at the end of the disk. Jun 25 18:47:26.800285 kernel: GPT:9289727 != 19775487 Jun 25 18:47:26.800298 kernel: GPT: Use GNU Parted to correct GPT errors. Jun 25 18:47:26.800311 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jun 25 18:47:26.800324 kernel: AVX2 version of gcm_enc/dec engaged. Jun 25 18:47:26.800338 kernel: AES CTR mode by8 optimization enabled Jun 25 18:47:26.772105 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jun 25 18:47:26.786861 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jun 25 18:47:26.786920 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jun 25 18:47:26.788403 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jun 25 18:47:26.789555 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jun 25 18:47:26.789605 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jun 25 18:47:26.790813 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jun 25 18:47:26.816313 kernel: libata version 3.00 loaded. Jun 25 18:47:26.803439 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jun 25 18:47:26.806158 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jun 25 18:47:26.819722 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jun 25 18:47:26.819872 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jun 25 18:47:26.827862 kernel: ata_piix 0000:00:01.1: version 2.13 Jun 25 18:47:26.835536 kernel: BTRFS: device fsid 329ce27e-ea89-47b5-8f8b-f762c8412eb0 devid 1 transid 31 /dev/vda3 scanned by (udev-worker) (464) Jun 25 18:47:26.835549 kernel: scsi host0: ata_piix Jun 25 18:47:26.835711 kernel: scsi host1: ata_piix Jun 25 18:47:26.835914 kernel: ata1: PATA max MWDMA2 cmd 0x1f0 ctl 0x3f6 bmdma 0xc0c0 irq 14 Jun 25 18:47:26.835925 kernel: ata2: PATA max MWDMA2 cmd 0x170 ctl 0x376 bmdma 0xc0c8 irq 15 Jun 25 18:47:26.835935 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (473) Jun 25 18:47:26.848247 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT. Jun 25 18:47:26.855485 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM. Jun 25 18:47:26.859360 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A. Jun 25 18:47:26.859431 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132. Jun 25 18:47:26.866542 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jun 25 18:47:26.881071 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jun 25 18:47:26.884423 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jun 25 18:47:26.889165 disk-uuid[541]: Primary Header is updated. Jun 25 18:47:26.889165 disk-uuid[541]: Secondary Entries is updated. Jun 25 18:47:26.889165 disk-uuid[541]: Secondary Header is updated. Jun 25 18:47:26.892539 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jun 25 18:47:26.896867 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jun 25 18:47:26.901868 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jun 25 18:47:26.903869 kernel: block device autoloading is deprecated and will be removed. Jun 25 18:47:26.908941 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jun 25 18:47:26.921024 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jun 25 18:47:26.947270 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jun 25 18:47:26.988069 kernel: ata2: found unknown device (class 0) Jun 25 18:47:26.989077 kernel: ata2.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Jun 25 18:47:26.992936 kernel: scsi 1:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Jun 25 18:47:27.064267 kernel: sr 1:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Jun 25 18:47:27.076950 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jun 25 18:47:27.076970 kernel: sr 1:0:0:0: Attached scsi CD-ROM sr0 Jun 25 18:47:27.901726 disk-uuid[542]: The operation has completed successfully. Jun 25 18:47:27.903408 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Jun 25 18:47:27.927822 systemd[1]: disk-uuid.service: Deactivated successfully. Jun 25 18:47:27.927963 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jun 25 18:47:27.956083 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jun 25 18:47:27.959090 sh[586]: Success Jun 25 18:47:27.971881 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Jun 25 18:47:28.005346 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jun 25 18:47:28.019400 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jun 25 18:47:28.022190 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jun 25 18:47:28.040201 kernel: BTRFS info (device dm-0): first mount of filesystem 329ce27e-ea89-47b5-8f8b-f762c8412eb0 Jun 25 18:47:28.040229 kernel: BTRFS info (device dm-0): using crc32c (crc32c-intel) checksum algorithm Jun 25 18:47:28.040241 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jun 25 18:47:28.042013 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jun 25 18:47:28.042038 kernel: BTRFS info (device dm-0): using free space tree Jun 25 18:47:28.047116 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jun 25 18:47:28.047415 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jun 25 18:47:28.057013 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jun 25 18:47:28.059360 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jun 25 18:47:28.083029 kernel: BTRFS info (device vda6): first mount of filesystem e6704e83-f8c1-4f1f-ad66-682b94c5899a Jun 25 18:47:28.083075 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jun 25 18:47:28.083091 kernel: BTRFS info (device vda6): using free space tree Jun 25 18:47:28.086284 kernel: BTRFS info (device vda6): auto enabling async discard Jun 25 18:47:28.095869 systemd[1]: mnt-oem.mount: Deactivated successfully. Jun 25 18:47:28.097857 kernel: BTRFS info (device vda6): last unmount of filesystem e6704e83-f8c1-4f1f-ad66-682b94c5899a Jun 25 18:47:28.177150 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jun 25 18:47:28.189985 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jun 25 18:47:28.219591 systemd-networkd[764]: lo: Link UP Jun 25 18:47:28.219603 systemd-networkd[764]: lo: Gained carrier Jun 25 18:47:28.221176 systemd-networkd[764]: Enumeration completed Jun 25 18:47:28.221250 systemd[1]: Started systemd-networkd.service - Network Configuration. Jun 25 18:47:28.221552 systemd-networkd[764]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jun 25 18:47:28.221556 systemd-networkd[764]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jun 25 18:47:28.222770 systemd[1]: Reached target network.target - Network. Jun 25 18:47:28.223092 systemd-networkd[764]: eth0: Link UP Jun 25 18:47:28.223096 systemd-networkd[764]: eth0: Gained carrier Jun 25 18:47:28.223103 systemd-networkd[764]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jun 25 18:47:28.241887 systemd-networkd[764]: eth0: DHCPv4 address 10.0.0.150/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jun 25 18:47:28.267864 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jun 25 18:47:28.280984 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jun 25 18:47:28.331909 ignition[769]: Ignition 2.19.0 Jun 25 18:47:28.331922 ignition[769]: Stage: fetch-offline Jun 25 18:47:28.331962 ignition[769]: no configs at "/usr/lib/ignition/base.d" Jun 25 18:47:28.331974 ignition[769]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jun 25 18:47:28.332086 ignition[769]: parsed url from cmdline: "" Jun 25 18:47:28.332091 ignition[769]: no config URL provided Jun 25 18:47:28.332099 ignition[769]: reading system config file "/usr/lib/ignition/user.ign" Jun 25 18:47:28.332112 ignition[769]: no config at "/usr/lib/ignition/user.ign" Jun 25 18:47:28.332143 ignition[769]: op(1): [started] loading QEMU firmware config module Jun 25 18:47:28.332149 ignition[769]: op(1): executing: "modprobe" "qemu_fw_cfg" Jun 25 18:47:28.339207 ignition[769]: op(1): [finished] loading QEMU firmware config module Jun 25 18:47:28.339231 ignition[769]: QEMU firmware config was not found. Ignoring... Jun 25 18:47:28.381080 ignition[769]: parsing config with SHA512: 2f4514e02c5226d75e17c3f0ff252446b505b993d3a168814edcfe83a1a7907423e90d0714b99dd0b6c564237bcb2997e98109dc9ee6959c74c2d80b75cdb514 Jun 25 18:47:28.384786 unknown[769]: fetched base config from "system" Jun 25 18:47:28.384804 unknown[769]: fetched user config from "qemu" Jun 25 18:47:28.385217 ignition[769]: fetch-offline: fetch-offline passed Jun 25 18:47:28.385381 systemd-resolved[231]: Detected conflict on linux IN A 10.0.0.150 Jun 25 18:47:28.385271 ignition[769]: Ignition finished successfully Jun 25 18:47:28.385392 systemd-resolved[231]: Hostname conflict, changing published hostname from 'linux' to 'linux10'. Jun 25 18:47:28.387663 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jun 25 18:47:28.389122 systemd[1]: ignition-fetch.service - Ignition (fetch) was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Jun 25 18:47:28.393981 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jun 25 18:47:28.409676 ignition[781]: Ignition 2.19.0 Jun 25 18:47:28.409687 ignition[781]: Stage: kargs Jun 25 18:47:28.409895 ignition[781]: no configs at "/usr/lib/ignition/base.d" Jun 25 18:47:28.419615 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jun 25 18:47:28.409907 ignition[781]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jun 25 18:47:28.410768 ignition[781]: kargs: kargs passed Jun 25 18:47:28.410811 ignition[781]: Ignition finished successfully Jun 25 18:47:28.428989 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jun 25 18:47:28.442537 ignition[789]: Ignition 2.19.0 Jun 25 18:47:28.442548 ignition[789]: Stage: disks Jun 25 18:47:28.442732 ignition[789]: no configs at "/usr/lib/ignition/base.d" Jun 25 18:47:28.442750 ignition[789]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jun 25 18:47:28.443644 ignition[789]: disks: disks passed Jun 25 18:47:28.445924 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jun 25 18:47:28.443689 ignition[789]: Ignition finished successfully Jun 25 18:47:28.447188 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jun 25 18:47:28.448757 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jun 25 18:47:28.450967 systemd[1]: Reached target local-fs.target - Local File Systems. Jun 25 18:47:28.452028 systemd[1]: Reached target sysinit.target - System Initialization. Jun 25 18:47:28.452082 systemd[1]: Reached target basic.target - Basic System. Jun 25 18:47:28.464987 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jun 25 18:47:28.489971 systemd-fsck[800]: ROOT: clean, 14/553520 files, 52654/553472 blocks Jun 25 18:47:28.619815 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jun 25 18:47:28.629937 systemd[1]: Mounting sysroot.mount - /sysroot... Jun 25 18:47:28.725903 kernel: EXT4-fs (vda9): mounted filesystem ed685e11-963b-427a-9b96-a4691c40e909 r/w with ordered data mode. Quota mode: none. Jun 25 18:47:28.726479 systemd[1]: Mounted sysroot.mount - /sysroot. Jun 25 18:47:28.728009 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jun 25 18:47:28.738934 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jun 25 18:47:28.741036 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jun 25 18:47:28.743500 systemd[1]: flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent was skipped because no trigger condition checks were met. Jun 25 18:47:28.748014 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (808) Jun 25 18:47:28.743568 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jun 25 18:47:28.743602 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jun 25 18:47:28.757179 kernel: BTRFS info (device vda6): first mount of filesystem e6704e83-f8c1-4f1f-ad66-682b94c5899a Jun 25 18:47:28.757197 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jun 25 18:47:28.757209 kernel: BTRFS info (device vda6): using free space tree Jun 25 18:47:28.757224 kernel: BTRFS info (device vda6): auto enabling async discard Jun 25 18:47:28.749135 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jun 25 18:47:28.756055 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jun 25 18:47:28.758363 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jun 25 18:47:28.793611 initrd-setup-root[832]: cut: /sysroot/etc/passwd: No such file or directory Jun 25 18:47:28.797612 initrd-setup-root[839]: cut: /sysroot/etc/group: No such file or directory Jun 25 18:47:28.801555 initrd-setup-root[846]: cut: /sysroot/etc/shadow: No such file or directory Jun 25 18:47:28.805366 initrd-setup-root[853]: cut: /sysroot/etc/gshadow: No such file or directory Jun 25 18:47:28.885526 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jun 25 18:47:28.896960 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jun 25 18:47:28.898665 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jun 25 18:47:28.906871 kernel: BTRFS info (device vda6): last unmount of filesystem e6704e83-f8c1-4f1f-ad66-682b94c5899a Jun 25 18:47:28.926688 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jun 25 18:47:28.965433 ignition[926]: INFO : Ignition 2.19.0 Jun 25 18:47:28.965433 ignition[926]: INFO : Stage: mount Jun 25 18:47:28.967191 ignition[926]: INFO : no configs at "/usr/lib/ignition/base.d" Jun 25 18:47:28.967191 ignition[926]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jun 25 18:47:28.969873 ignition[926]: INFO : mount: mount passed Jun 25 18:47:28.970655 ignition[926]: INFO : Ignition finished successfully Jun 25 18:47:28.973448 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jun 25 18:47:28.993951 systemd[1]: Starting ignition-files.service - Ignition (files)... Jun 25 18:47:29.037681 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jun 25 18:47:29.047052 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jun 25 18:47:29.056318 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/vda6 scanned by mount (936) Jun 25 18:47:29.056346 kernel: BTRFS info (device vda6): first mount of filesystem e6704e83-f8c1-4f1f-ad66-682b94c5899a Jun 25 18:47:29.056364 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Jun 25 18:47:29.057885 kernel: BTRFS info (device vda6): using free space tree Jun 25 18:47:29.060867 kernel: BTRFS info (device vda6): auto enabling async discard Jun 25 18:47:29.061643 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jun 25 18:47:29.087963 ignition[953]: INFO : Ignition 2.19.0 Jun 25 18:47:29.087963 ignition[953]: INFO : Stage: files Jun 25 18:47:29.089787 ignition[953]: INFO : no configs at "/usr/lib/ignition/base.d" Jun 25 18:47:29.089787 ignition[953]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jun 25 18:47:29.089787 ignition[953]: DEBUG : files: compiled without relabeling support, skipping Jun 25 18:47:29.089787 ignition[953]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jun 25 18:47:29.089787 ignition[953]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jun 25 18:47:29.096642 ignition[953]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jun 25 18:47:29.096642 ignition[953]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jun 25 18:47:29.096642 ignition[953]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jun 25 18:47:29.096642 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Jun 25 18:47:29.096642 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Jun 25 18:47:29.096642 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jun 25 18:47:29.096642 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://get.helm.sh/helm-v3.13.2-linux-amd64.tar.gz: attempt #1 Jun 25 18:47:29.092402 unknown[953]: wrote ssh authorized keys file for user: core Jun 25 18:47:29.120657 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Jun 25 18:47:29.192104 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-amd64.tar.gz" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/install.sh" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/install.sh" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nginx.yaml" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nginx.yaml" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/etc/flatcar/update.conf" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.28.7-x86-64.raw" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.28.7-x86-64.raw" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.28.7-x86-64.raw" Jun 25 18:47:29.194501 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.28.7-x86-64.raw: attempt #1 Jun 25 18:47:29.551895 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Jun 25 18:47:29.871573 ignition[953]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.28.7-x86-64.raw" Jun 25 18:47:29.871573 ignition[953]: INFO : files: op(c): [started] processing unit "containerd.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(c): op(d): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(c): op(d): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(c): [finished] processing unit "containerd.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(e): [started] processing unit "prepare-helm.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(e): op(f): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(e): op(f): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(e): [finished] processing unit "prepare-helm.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(10): [started] processing unit "coreos-metadata.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(10): op(11): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(10): op(11): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(10): [finished] processing unit "coreos-metadata.service" Jun 25 18:47:29.876033 ignition[953]: INFO : files: op(12): [started] setting preset to disabled for "coreos-metadata.service" Jun 25 18:47:29.898767 ignition[953]: INFO : files: op(12): op(13): [started] removing enablement symlink(s) for "coreos-metadata.service" Jun 25 18:47:29.902618 ignition[953]: INFO : files: op(12): op(13): [finished] removing enablement symlink(s) for "coreos-metadata.service" Jun 25 18:47:29.904215 ignition[953]: INFO : files: op(12): [finished] setting preset to disabled for "coreos-metadata.service" Jun 25 18:47:29.904215 ignition[953]: INFO : files: op(14): [started] setting preset to enabled for "prepare-helm.service" Jun 25 18:47:29.904215 ignition[953]: INFO : files: op(14): [finished] setting preset to enabled for "prepare-helm.service" Jun 25 18:47:29.904215 ignition[953]: INFO : files: createResultFile: createFiles: op(15): [started] writing file "/sysroot/etc/.ignition-result.json" Jun 25 18:47:29.904215 ignition[953]: INFO : files: createResultFile: createFiles: op(15): [finished] writing file "/sysroot/etc/.ignition-result.json" Jun 25 18:47:29.904215 ignition[953]: INFO : files: files passed Jun 25 18:47:29.904215 ignition[953]: INFO : Ignition finished successfully Jun 25 18:47:29.905458 systemd[1]: Finished ignition-files.service - Ignition (files). Jun 25 18:47:29.918031 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jun 25 18:47:29.919778 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jun 25 18:47:29.921791 systemd[1]: ignition-quench.service: Deactivated successfully. Jun 25 18:47:29.921912 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jun 25 18:47:29.929348 initrd-setup-root-after-ignition[982]: grep: /sysroot/oem/oem-release: No such file or directory Jun 25 18:47:29.931446 initrd-setup-root-after-ignition[984]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jun 25 18:47:29.933137 initrd-setup-root-after-ignition[984]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jun 25 18:47:29.934766 initrd-setup-root-after-ignition[988]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jun 25 18:47:29.934317 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jun 25 18:47:29.936543 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jun 25 18:47:29.948002 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jun 25 18:47:29.970655 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jun 25 18:47:29.970796 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jun 25 18:47:29.973117 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jun 25 18:47:29.975229 systemd[1]: Reached target initrd.target - Initrd Default Target. Jun 25 18:47:29.976322 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jun 25 18:47:29.989968 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jun 25 18:47:30.003458 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jun 25 18:47:30.004903 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jun 25 18:47:30.018263 systemd[1]: Stopped target network.target - Network. Jun 25 18:47:30.018415 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jun 25 18:47:30.018815 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jun 25 18:47:30.071271 ignition[1008]: INFO : Ignition 2.19.0 Jun 25 18:47:30.071271 ignition[1008]: INFO : Stage: umount Jun 25 18:47:30.071271 ignition[1008]: INFO : no configs at "/usr/lib/ignition/base.d" Jun 25 18:47:30.071271 ignition[1008]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Jun 25 18:47:30.071271 ignition[1008]: INFO : umount: umount passed Jun 25 18:47:30.071271 ignition[1008]: INFO : Ignition finished successfully Jun 25 18:47:30.019130 systemd[1]: Stopped target timers.target - Timer Units. Jun 25 18:47:30.019447 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jun 25 18:47:30.019553 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jun 25 18:47:30.020625 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jun 25 18:47:30.021146 systemd[1]: Stopped target basic.target - Basic System. Jun 25 18:47:30.021468 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jun 25 18:47:30.021815 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jun 25 18:47:30.022184 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jun 25 18:47:30.022491 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jun 25 18:47:30.022825 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jun 25 18:47:30.023167 systemd[1]: Stopped target sysinit.target - System Initialization. Jun 25 18:47:30.023545 systemd[1]: Stopped target local-fs.target - Local File Systems. Jun 25 18:47:30.023895 systemd[1]: Stopped target swap.target - Swaps. Jun 25 18:47:30.024145 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jun 25 18:47:30.024250 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jun 25 18:47:30.024694 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jun 25 18:47:30.025220 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jun 25 18:47:30.025504 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jun 25 18:47:30.025631 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jun 25 18:47:30.026221 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jun 25 18:47:30.026359 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jun 25 18:47:30.026885 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jun 25 18:47:30.026990 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jun 25 18:47:30.027449 systemd[1]: Stopped target paths.target - Path Units. Jun 25 18:47:30.027736 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jun 25 18:47:30.027880 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jun 25 18:47:30.028264 systemd[1]: Stopped target slices.target - Slice Units. Jun 25 18:47:30.028584 systemd[1]: Stopped target sockets.target - Socket Units. Jun 25 18:47:30.029119 systemd[1]: iscsid.socket: Deactivated successfully. Jun 25 18:47:30.029222 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jun 25 18:47:30.029686 systemd[1]: iscsiuio.socket: Deactivated successfully. Jun 25 18:47:30.029808 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jun 25 18:47:30.030163 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jun 25 18:47:30.030271 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jun 25 18:47:30.030702 systemd[1]: ignition-files.service: Deactivated successfully. Jun 25 18:47:30.030813 systemd[1]: Stopped ignition-files.service - Ignition (files). Jun 25 18:47:30.031959 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jun 25 18:47:30.032321 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jun 25 18:47:30.032435 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jun 25 18:47:30.033386 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jun 25 18:47:30.034300 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jun 25 18:47:30.034998 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jun 25 18:47:30.035332 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jun 25 18:47:30.035469 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jun 25 18:47:30.035898 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jun 25 18:47:30.036036 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jun 25 18:47:30.040454 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jun 25 18:47:30.040586 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jun 25 18:47:30.055243 systemd[1]: ignition-mount.service: Deactivated successfully. Jun 25 18:47:30.055382 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jun 25 18:47:30.055807 systemd[1]: ignition-disks.service: Deactivated successfully. Jun 25 18:47:30.055877 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jun 25 18:47:30.056151 systemd[1]: ignition-kargs.service: Deactivated successfully. Jun 25 18:47:30.056199 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jun 25 18:47:30.056586 systemd[1]: ignition-setup.service: Deactivated successfully. Jun 25 18:47:30.056627 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jun 25 18:47:30.056804 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jun 25 18:47:30.056861 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jun 25 18:47:30.062048 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jun 25 18:47:30.065170 systemd[1]: systemd-resolved.service: Deactivated successfully. Jun 25 18:47:30.065312 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jun 25 18:47:30.069040 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jun 25 18:47:30.069109 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create Volatile Files and Directories. Jun 25 18:47:30.070115 systemd-networkd[764]: eth0: DHCPv6 lease lost Jun 25 18:47:30.071539 systemd[1]: systemd-networkd.service: Deactivated successfully. Jun 25 18:47:30.071686 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jun 25 18:47:30.074632 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jun 25 18:47:30.074719 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jun 25 18:47:30.080949 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jun 25 18:47:30.082487 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jun 25 18:47:30.082578 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jun 25 18:47:30.085245 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jun 25 18:47:30.085293 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jun 25 18:47:30.086765 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jun 25 18:47:30.086812 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jun 25 18:47:30.088244 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jun 25 18:47:30.101406 systemd[1]: network-cleanup.service: Deactivated successfully. Jun 25 18:47:30.101539 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jun 25 18:47:30.112215 systemd[1]: systemd-udevd.service: Deactivated successfully. Jun 25 18:47:30.112425 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jun 25 18:47:30.114465 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jun 25 18:47:30.114522 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jun 25 18:47:30.117191 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jun 25 18:47:30.117232 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jun 25 18:47:30.119663 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jun 25 18:47:30.119742 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jun 25 18:47:30.122183 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jun 25 18:47:30.122231 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jun 25 18:47:30.124571 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jun 25 18:47:30.124619 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jun 25 18:47:30.132999 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jun 25 18:47:30.134803 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jun 25 18:47:30.134877 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jun 25 18:47:30.137274 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jun 25 18:47:30.137328 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jun 25 18:47:30.140971 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jun 25 18:47:30.141083 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jun 25 18:47:30.306340 systemd[1]: sysroot-boot.service: Deactivated successfully. Jun 25 18:47:30.306524 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jun 25 18:47:30.308284 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jun 25 18:47:30.309464 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jun 25 18:47:30.309556 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jun 25 18:47:30.317023 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jun 25 18:47:30.324326 systemd[1]: Switching root. Jun 25 18:47:30.356036 systemd-journald[193]: Journal stopped Jun 25 18:47:31.570896 systemd-journald[193]: Received SIGTERM from PID 1 (systemd). Jun 25 18:47:31.570977 kernel: SELinux: policy capability network_peer_controls=1 Jun 25 18:47:31.570996 kernel: SELinux: policy capability open_perms=1 Jun 25 18:47:31.571024 kernel: SELinux: policy capability extended_socket_class=1 Jun 25 18:47:31.571043 kernel: SELinux: policy capability always_check_network=0 Jun 25 18:47:31.571065 kernel: SELinux: policy capability cgroup_seclabel=1 Jun 25 18:47:31.571175 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jun 25 18:47:31.571206 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jun 25 18:47:31.571221 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jun 25 18:47:31.571245 kernel: audit: type=1403 audit(1719341250.778:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jun 25 18:47:31.571269 systemd[1]: Successfully loaded SELinux policy in 44.187ms. Jun 25 18:47:31.571301 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 12.986ms. Jun 25 18:47:31.571321 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jun 25 18:47:31.571338 systemd[1]: Detected virtualization kvm. Jun 25 18:47:31.571355 systemd[1]: Detected architecture x86-64. Jun 25 18:47:31.571376 systemd[1]: Detected first boot. Jun 25 18:47:31.571393 systemd[1]: Initializing machine ID from VM UUID. Jun 25 18:47:31.571410 zram_generator::config[1070]: No configuration found. Jun 25 18:47:31.571435 systemd[1]: Populated /etc with preset unit settings. Jun 25 18:47:31.571452 systemd[1]: Queued start job for default target multi-user.target. Jun 25 18:47:31.571470 systemd[1]: Unnecessary job was removed for dev-vda6.device - /dev/vda6. Jun 25 18:47:31.571488 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jun 25 18:47:31.571506 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jun 25 18:47:31.571526 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jun 25 18:47:31.571550 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jun 25 18:47:31.571569 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jun 25 18:47:31.571587 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jun 25 18:47:31.571604 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jun 25 18:47:31.571621 systemd[1]: Created slice user.slice - User and Session Slice. Jun 25 18:47:31.571644 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jun 25 18:47:31.571664 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jun 25 18:47:31.571692 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jun 25 18:47:31.571713 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jun 25 18:47:31.571731 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jun 25 18:47:31.571748 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jun 25 18:47:31.571766 systemd[1]: Expecting device dev-ttyS0.device - /dev/ttyS0... Jun 25 18:47:31.571783 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jun 25 18:47:31.571800 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jun 25 18:47:31.571818 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jun 25 18:47:31.571836 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jun 25 18:47:31.575891 systemd[1]: Reached target slices.target - Slice Units. Jun 25 18:47:31.575918 systemd[1]: Reached target swap.target - Swaps. Jun 25 18:47:31.575936 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jun 25 18:47:31.575954 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jun 25 18:47:31.575971 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jun 25 18:47:31.575988 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jun 25 18:47:31.576005 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jun 25 18:47:31.576023 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jun 25 18:47:31.576040 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jun 25 18:47:31.576060 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jun 25 18:47:31.576077 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jun 25 18:47:31.576095 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jun 25 18:47:31.576113 systemd[1]: Mounting media.mount - External Media Directory... Jun 25 18:47:31.576131 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jun 25 18:47:31.576148 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jun 25 18:47:31.576165 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jun 25 18:47:31.576183 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jun 25 18:47:31.576203 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jun 25 18:47:31.576221 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jun 25 18:47:31.576238 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jun 25 18:47:31.576255 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jun 25 18:47:31.576272 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jun 25 18:47:31.576291 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jun 25 18:47:31.576308 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jun 25 18:47:31.576325 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jun 25 18:47:31.576343 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jun 25 18:47:31.576363 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jun 25 18:47:31.576381 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Jun 25 18:47:31.576399 systemd[1]: systemd-journald.service: (This warning is only shown for the first unit using IP firewalling.) Jun 25 18:47:31.576416 kernel: fuse: init (API version 7.39) Jun 25 18:47:31.576433 kernel: loop: module loaded Jun 25 18:47:31.576450 systemd[1]: Starting systemd-journald.service - Journal Service... Jun 25 18:47:31.576467 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jun 25 18:47:31.576484 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jun 25 18:47:31.576504 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jun 25 18:47:31.576550 systemd-journald[1157]: Collecting audit messages is disabled. Jun 25 18:47:31.576580 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jun 25 18:47:31.576598 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jun 25 18:47:31.576615 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jun 25 18:47:31.576632 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jun 25 18:47:31.576649 systemd-journald[1157]: Journal started Jun 25 18:47:31.576692 systemd-journald[1157]: Runtime Journal (/run/log/journal/e59d93c2bcf6416384c1fd403fc9f628) is 6.0M, max 48.3M, 42.3M free. Jun 25 18:47:31.580817 systemd[1]: Started systemd-journald.service - Journal Service. Jun 25 18:47:31.582400 systemd[1]: Mounted media.mount - External Media Directory. Jun 25 18:47:31.583789 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jun 25 18:47:31.586425 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jun 25 18:47:31.587933 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jun 25 18:47:31.589542 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jun 25 18:47:31.592266 kernel: ACPI: bus type drm_connector registered Jun 25 18:47:31.592343 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jun 25 18:47:31.592562 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jun 25 18:47:31.594081 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jun 25 18:47:31.594294 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jun 25 18:47:31.595827 systemd[1]: modprobe@drm.service: Deactivated successfully. Jun 25 18:47:31.596169 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jun 25 18:47:31.597764 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jun 25 18:47:31.597988 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jun 25 18:47:31.599527 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jun 25 18:47:31.599746 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jun 25 18:47:31.601142 systemd[1]: modprobe@loop.service: Deactivated successfully. Jun 25 18:47:31.601362 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jun 25 18:47:31.603117 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jun 25 18:47:31.604873 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jun 25 18:47:31.606576 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jun 25 18:47:31.612261 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jun 25 18:47:31.624381 systemd[1]: Reached target network-pre.target - Preparation for Network. Jun 25 18:47:31.641931 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jun 25 18:47:31.644475 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jun 25 18:47:31.645818 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jun 25 18:47:31.649969 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jun 25 18:47:31.653655 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jun 25 18:47:31.656934 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jun 25 18:47:31.658375 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jun 25 18:47:31.659886 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jun 25 18:47:31.665064 systemd-journald[1157]: Time spent on flushing to /var/log/journal/e59d93c2bcf6416384c1fd403fc9f628 is 18.658ms for 980 entries. Jun 25 18:47:31.665064 systemd-journald[1157]: System Journal (/var/log/journal/e59d93c2bcf6416384c1fd403fc9f628) is 8.0M, max 195.6M, 187.6M free. Jun 25 18:47:31.882615 systemd-journald[1157]: Received client request to flush runtime journal. Jun 25 18:47:31.665091 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jun 25 18:47:31.668706 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jun 25 18:47:31.671535 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jun 25 18:47:31.673114 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jun 25 18:47:31.690604 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jun 25 18:47:31.700153 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jun 25 18:47:31.707396 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jun 25 18:47:31.711623 udevadm[1212]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jun 25 18:47:31.713943 systemd-tmpfiles[1205]: ACLs are not supported, ignoring. Jun 25 18:47:31.713958 systemd-tmpfiles[1205]: ACLs are not supported, ignoring. Jun 25 18:47:31.719797 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jun 25 18:47:31.726054 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jun 25 18:47:31.789997 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jun 25 18:47:31.799123 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jun 25 18:47:31.821300 systemd-tmpfiles[1225]: ACLs are not supported, ignoring. Jun 25 18:47:31.821319 systemd-tmpfiles[1225]: ACLs are not supported, ignoring. Jun 25 18:47:31.828470 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jun 25 18:47:31.845317 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jun 25 18:47:31.846892 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jun 25 18:47:31.885646 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jun 25 18:47:32.323364 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jun 25 18:47:32.336035 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jun 25 18:47:32.361511 systemd-udevd[1240]: Using default interface naming scheme 'v255'. Jun 25 18:47:32.375223 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jun 25 18:47:32.386372 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jun 25 18:47:32.400069 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jun 25 18:47:32.410380 systemd[1]: Found device dev-ttyS0.device - /dev/ttyS0. Jun 25 18:47:32.417869 kernel: BTRFS info: devid 1 device path /dev/mapper/usr changed to /dev/dm-0 scanned by (udev-worker) (1253) Jun 25 18:47:32.444880 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (1248) Jun 25 18:47:32.454918 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jun 25 18:47:32.499682 kernel: piix4_smbus 0000:00:01.3: SMBus Host Controller at 0xb100, revision 0 Jun 25 18:47:32.497130 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM. Jun 25 18:47:32.502868 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Jun 25 18:47:32.515864 kernel: ACPI: button: Power Button [PWRF] Jun 25 18:47:32.517875 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Jun 25 18:47:32.530402 systemd-networkd[1249]: lo: Link UP Jun 25 18:47:32.530413 systemd-networkd[1249]: lo: Gained carrier Jun 25 18:47:32.532529 systemd-networkd[1249]: Enumeration completed Jun 25 18:47:32.532660 systemd[1]: Started systemd-networkd.service - Network Configuration. Jun 25 18:47:32.533325 systemd-networkd[1249]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jun 25 18:47:32.533329 systemd-networkd[1249]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jun 25 18:47:32.534563 systemd-networkd[1249]: eth0: Link UP Jun 25 18:47:32.534574 systemd-networkd[1249]: eth0: Gained carrier Jun 25 18:47:32.534586 systemd-networkd[1249]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jun 25 18:47:32.546989 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jun 25 18:47:32.549165 kernel: mousedev: PS/2 mouse device common for all mice Jun 25 18:47:32.549108 systemd-networkd[1249]: eth0: DHCPv4 address 10.0.0.150/16, gateway 10.0.0.1 acquired from 10.0.0.1 Jun 25 18:47:32.571173 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jun 25 18:47:32.619410 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jun 25 18:47:32.672977 kernel: kvm_amd: TSC scaling supported Jun 25 18:47:32.673030 kernel: kvm_amd: Nested Virtualization enabled Jun 25 18:47:32.673047 kernel: kvm_amd: Nested Paging enabled Jun 25 18:47:32.673955 kernel: kvm_amd: LBR virtualization supported Jun 25 18:47:32.673976 kernel: kvm_amd: Virtual VMLOAD VMSAVE supported Jun 25 18:47:32.676922 kernel: kvm_amd: Virtual GIF supported Jun 25 18:47:32.695868 kernel: EDAC MC: Ver: 3.0.0 Jun 25 18:47:32.730613 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jun 25 18:47:32.743060 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jun 25 18:47:32.751429 lvm[1293]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jun 25 18:47:32.780432 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jun 25 18:47:32.783117 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jun 25 18:47:32.792971 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jun 25 18:47:32.797714 lvm[1299]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jun 25 18:47:32.827145 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jun 25 18:47:32.828739 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jun 25 18:47:32.830142 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jun 25 18:47:32.830170 systemd[1]: Reached target local-fs.target - Local File Systems. Jun 25 18:47:32.831382 systemd[1]: Reached target machines.target - Containers. Jun 25 18:47:32.833954 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jun 25 18:47:32.842969 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jun 25 18:47:32.848562 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jun 25 18:47:32.850009 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jun 25 18:47:32.851226 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jun 25 18:47:32.856207 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jun 25 18:47:32.860867 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jun 25 18:47:32.862379 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jun 25 18:47:32.873492 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jun 25 18:47:32.910142 kernel: loop0: detected capacity change from 0 to 209816 Jun 25 18:47:32.910220 kernel: block loop0: the capability attribute has been deprecated. Jun 25 18:47:32.941872 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jun 25 18:47:32.977871 kernel: loop1: detected capacity change from 0 to 139760 Jun 25 18:47:33.008472 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jun 25 18:47:33.009377 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jun 25 18:47:33.016873 kernel: loop2: detected capacity change from 0 to 80568 Jun 25 18:47:33.051870 kernel: loop3: detected capacity change from 0 to 209816 Jun 25 18:47:33.058859 kernel: loop4: detected capacity change from 0 to 139760 Jun 25 18:47:33.067859 kernel: loop5: detected capacity change from 0 to 80568 Jun 25 18:47:33.073010 (sd-merge)[1320]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes'. Jun 25 18:47:33.073642 (sd-merge)[1320]: Merged extensions into '/usr'. Jun 25 18:47:33.077668 systemd[1]: Reloading requested from client PID 1307 ('systemd-sysext') (unit systemd-sysext.service)... Jun 25 18:47:33.077682 systemd[1]: Reloading... Jun 25 18:47:33.122879 zram_generator::config[1343]: No configuration found. Jun 25 18:47:33.166806 ldconfig[1304]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jun 25 18:47:33.256217 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jun 25 18:47:33.321298 systemd[1]: Reloading finished in 243 ms. Jun 25 18:47:33.339718 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jun 25 18:47:33.341440 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jun 25 18:47:33.366134 systemd[1]: Starting ensure-sysext.service... Jun 25 18:47:33.377818 systemd[1]: Starting systemd-tmpfiles-setup.service - Create Volatile Files and Directories... Jun 25 18:47:33.380689 systemd[1]: Reloading requested from client PID 1390 ('systemctl') (unit ensure-sysext.service)... Jun 25 18:47:33.380702 systemd[1]: Reloading... Jun 25 18:47:33.402991 systemd-tmpfiles[1396]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jun 25 18:47:33.403379 systemd-tmpfiles[1396]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jun 25 18:47:33.404637 systemd-tmpfiles[1396]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jun 25 18:47:33.405089 systemd-tmpfiles[1396]: ACLs are not supported, ignoring. Jun 25 18:47:33.405192 systemd-tmpfiles[1396]: ACLs are not supported, ignoring. Jun 25 18:47:33.408806 systemd-tmpfiles[1396]: Detected autofs mount point /boot during canonicalization of boot. Jun 25 18:47:33.408821 systemd-tmpfiles[1396]: Skipping /boot Jun 25 18:47:33.422700 systemd-tmpfiles[1396]: Detected autofs mount point /boot during canonicalization of boot. Jun 25 18:47:33.422721 systemd-tmpfiles[1396]: Skipping /boot Jun 25 18:47:33.452912 zram_generator::config[1422]: No configuration found. Jun 25 18:47:33.551990 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jun 25 18:47:33.616401 systemd[1]: Reloading finished in 235 ms. Jun 25 18:47:33.640852 systemd[1]: Finished systemd-tmpfiles-setup.service - Create Volatile Files and Directories. Jun 25 18:47:33.659267 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jun 25 18:47:33.662142 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jun 25 18:47:33.664659 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jun 25 18:47:33.667991 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jun 25 18:47:33.674601 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jun 25 18:47:33.680828 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jun 25 18:47:33.681336 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jun 25 18:47:33.690306 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jun 25 18:47:33.697195 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jun 25 18:47:33.708287 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jun 25 18:47:33.709692 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jun 25 18:47:33.709906 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jun 25 18:47:33.711251 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jun 25 18:47:33.714785 augenrules[1494]: No rules Jun 25 18:47:33.719235 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jun 25 18:47:33.721500 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jun 25 18:47:33.721792 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jun 25 18:47:33.724024 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jun 25 18:47:33.724288 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jun 25 18:47:33.726672 systemd[1]: modprobe@loop.service: Deactivated successfully. Jun 25 18:47:33.726972 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jun 25 18:47:33.739854 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jun 25 18:47:33.744341 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jun 25 18:47:33.750545 systemd[1]: proc-xen.mount - /proc/xen was skipped because of an unmet condition check (ConditionVirtualization=xen). Jun 25 18:47:33.750812 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jun 25 18:47:33.763234 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jun 25 18:47:33.766263 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jun 25 18:47:33.770131 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jun 25 18:47:33.773084 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jun 25 18:47:33.774584 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jun 25 18:47:33.776130 systemd-resolved[1471]: Positive Trust Anchors: Jun 25 18:47:33.776153 systemd-resolved[1471]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jun 25 18:47:33.776195 systemd-resolved[1471]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa corp home internal intranet lan local private test Jun 25 18:47:33.780000 systemd-resolved[1471]: Defaulting to hostname 'linux'. Jun 25 18:47:33.780142 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jun 25 18:47:33.781562 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jun 25 18:47:33.781763 systemd[1]: xenserver-pv-version.service - Set fake PV driver version for XenServer was skipped because of an unmet condition check (ConditionVirtualization=xen). Jun 25 18:47:33.783183 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jun 25 18:47:33.785282 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jun 25 18:47:33.785574 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jun 25 18:47:33.787716 systemd[1]: modprobe@drm.service: Deactivated successfully. Jun 25 18:47:33.787997 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jun 25 18:47:33.789937 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jun 25 18:47:33.790193 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jun 25 18:47:33.792387 systemd[1]: modprobe@loop.service: Deactivated successfully. Jun 25 18:47:33.792685 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jun 25 18:47:33.795193 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jun 25 18:47:33.798767 systemd[1]: Finished ensure-sysext.service. Jun 25 18:47:33.807084 systemd[1]: Reached target network.target - Network. Jun 25 18:47:33.828642 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jun 25 18:47:33.829974 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jun 25 18:47:33.830048 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jun 25 18:47:33.844977 systemd[1]: Starting systemd-timesyncd.service - Network Time Synchronization... Jun 25 18:47:33.920457 systemd[1]: Started systemd-timesyncd.service - Network Time Synchronization. Jun 25 18:47:33.946569 systemd[1]: Reached target sysinit.target - System Initialization. Jun 25 18:47:33.947816 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jun 25 18:47:33.949199 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jun 25 18:47:34.730955 systemd-resolved[1471]: Clock change detected. Flushing caches. Jun 25 18:47:34.730990 systemd-timesyncd[1529]: Contacted time server 10.0.0.1:123 (10.0.0.1). Jun 25 18:47:34.731029 systemd-timesyncd[1529]: Initial clock synchronization to Tue 2024-06-25 18:47:34.730897 UTC. Jun 25 18:47:34.731074 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jun 25 18:47:34.732444 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jun 25 18:47:34.732476 systemd[1]: Reached target paths.target - Path Units. Jun 25 18:47:34.733463 systemd[1]: Reached target time-set.target - System Time Set. Jun 25 18:47:34.734706 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jun 25 18:47:34.736060 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jun 25 18:47:34.737434 systemd[1]: Reached target timers.target - Timer Units. Jun 25 18:47:34.758946 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jun 25 18:47:34.762066 systemd[1]: Starting docker.socket - Docker Socket for the API... Jun 25 18:47:34.764901 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jun 25 18:47:34.772528 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jun 25 18:47:34.773816 systemd[1]: Reached target sockets.target - Socket Units. Jun 25 18:47:34.774878 systemd[1]: Reached target basic.target - Basic System. Jun 25 18:47:34.776072 systemd[1]: System is tainted: cgroupsv1 Jun 25 18:47:34.776117 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jun 25 18:47:34.776144 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jun 25 18:47:34.777503 systemd[1]: Starting containerd.service - containerd container runtime... Jun 25 18:47:34.786965 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jun 25 18:47:34.789163 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jun 25 18:47:34.794294 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jun 25 18:47:34.798657 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jun 25 18:47:34.802542 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jun 25 18:47:34.809047 jq[1536]: false Jun 25 18:47:34.808826 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jun 25 18:47:34.812375 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jun 25 18:47:34.818163 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jun 25 18:47:34.820938 extend-filesystems[1537]: Found loop3 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found loop4 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found loop5 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found sr0 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found vda Jun 25 18:47:34.830422 extend-filesystems[1537]: Found vda1 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found vda2 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found vda3 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found usr Jun 25 18:47:34.830422 extend-filesystems[1537]: Found vda4 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found vda6 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found vda7 Jun 25 18:47:34.830422 extend-filesystems[1537]: Found vda9 Jun 25 18:47:34.830422 extend-filesystems[1537]: Checking size of /dev/vda9 Jun 25 18:47:34.833996 systemd[1]: Starting systemd-logind.service - User Login Management... Jun 25 18:47:34.836550 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jun 25 18:47:34.838988 systemd[1]: Starting update-engine.service - Update Engine... Jun 25 18:47:34.847012 dbus-daemon[1534]: [system] SELinux support is enabled Jun 25 18:47:34.847987 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jun 25 18:47:34.850328 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jun 25 18:47:34.853841 extend-filesystems[1537]: Resized partition /dev/vda9 Jun 25 18:47:34.861504 jq[1558]: true Jun 25 18:47:34.862001 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jun 25 18:47:34.866581 extend-filesystems[1561]: resize2fs 1.47.0 (5-Feb-2023) Jun 25 18:47:34.862376 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jun 25 18:47:34.862707 systemd[1]: motdgen.service: Deactivated successfully. Jun 25 18:47:34.863007 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jun 25 18:47:34.870718 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jun 25 18:47:34.871657 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jun 25 18:47:34.888221 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (1258) Jun 25 18:47:34.888282 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Jun 25 18:47:34.888341 update_engine[1556]: I0625 18:47:34.879933 1556 main.cc:92] Flatcar Update Engine starting Jun 25 18:47:34.891242 update_engine[1556]: I0625 18:47:34.890084 1556 update_check_scheduler.cc:74] Next update check in 5m59s Jun 25 18:47:34.905080 jq[1568]: true Jun 25 18:47:34.908812 (ntainerd)[1570]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jun 25 18:47:34.923661 tar[1565]: linux-amd64/helm Jun 25 18:47:34.927071 systemd[1]: Started update-engine.service - Update Engine. Jun 25 18:47:34.934692 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jun 25 18:47:34.934723 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jun 25 18:47:34.937068 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jun 25 18:47:34.937087 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jun 25 18:47:34.939858 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jun 25 18:47:34.949383 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jun 25 18:47:34.952338 systemd-logind[1553]: Watching system buttons on /dev/input/event1 (Power Button) Jun 25 18:47:34.952372 systemd-logind[1553]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jun 25 18:47:34.954836 systemd-logind[1553]: New seat seat0. Jun 25 18:47:34.956043 systemd[1]: Started systemd-logind.service - User Login Management. Jun 25 18:47:34.959250 systemd-networkd[1249]: eth0: Gained IPv6LL Jun 25 18:47:34.963910 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jun 25 18:47:34.967503 systemd[1]: Reached target network-online.target - Network is Online. Jun 25 18:47:34.980665 systemd[1]: Starting coreos-metadata.service - QEMU metadata agent... Jun 25 18:47:35.000464 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jun 25 18:47:35.002045 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Jun 25 18:47:35.004478 locksmithd[1593]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jun 25 18:47:35.004982 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jun 25 18:47:35.029252 systemd[1]: coreos-metadata.service: Deactivated successfully. Jun 25 18:47:35.029705 systemd[1]: Finished coreos-metadata.service - QEMU metadata agent. Jun 25 18:47:35.031617 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jun 25 18:47:35.183348 sshd_keygen[1564]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jun 25 18:47:35.183502 extend-filesystems[1561]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Jun 25 18:47:35.183502 extend-filesystems[1561]: old_desc_blocks = 1, new_desc_blocks = 1 Jun 25 18:47:35.183502 extend-filesystems[1561]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Jun 25 18:47:35.199367 containerd[1570]: time="2024-06-25T18:47:35.183292375Z" level=info msg="starting containerd" revision=cd7148ac666309abf41fd4a49a8a5895b905e7f3 version=v1.7.18 Jun 25 18:47:35.186964 systemd[1]: extend-filesystems.service: Deactivated successfully. Jun 25 18:47:35.199692 extend-filesystems[1537]: Resized filesystem in /dev/vda9 Jun 25 18:47:35.206725 bash[1597]: Updated "/home/core/.ssh/authorized_keys" Jun 25 18:47:35.187380 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jun 25 18:47:35.201353 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jun 25 18:47:35.203358 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jun 25 18:47:35.217317 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jun 25 18:47:35.229178 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jun 25 18:47:35.229483 containerd[1570]: time="2024-06-25T18:47:35.227764310Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jun 25 18:47:35.229483 containerd[1570]: time="2024-06-25T18:47:35.227851674Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jun 25 18:47:35.232212 containerd[1570]: time="2024-06-25T18:47:35.231973643Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.35-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jun 25 18:47:35.232212 containerd[1570]: time="2024-06-25T18:47:35.232019509Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jun 25 18:47:35.233430 containerd[1570]: time="2024-06-25T18:47:35.232518985Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jun 25 18:47:35.233430 containerd[1570]: time="2024-06-25T18:47:35.232544623Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jun 25 18:47:35.233430 containerd[1570]: time="2024-06-25T18:47:35.232664228Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jun 25 18:47:35.233430 containerd[1570]: time="2024-06-25T18:47:35.232743767Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jun 25 18:47:35.233430 containerd[1570]: time="2024-06-25T18:47:35.232761760Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jun 25 18:47:35.233430 containerd[1570]: time="2024-06-25T18:47:35.232869432Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jun 25 18:47:35.233656 containerd[1570]: time="2024-06-25T18:47:35.233635218Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jun 25 18:47:35.233750 containerd[1570]: time="2024-06-25T18:47:35.233726860Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Jun 25 18:47:35.233827 containerd[1570]: time="2024-06-25T18:47:35.233808573Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jun 25 18:47:35.234136 containerd[1570]: time="2024-06-25T18:47:35.234112333Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jun 25 18:47:35.234222 containerd[1570]: time="2024-06-25T18:47:35.234205738Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jun 25 18:47:35.234360 containerd[1570]: time="2024-06-25T18:47:35.234339328Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Jun 25 18:47:35.234423 containerd[1570]: time="2024-06-25T18:47:35.234409069Z" level=info msg="metadata content store policy set" policy=shared Jun 25 18:47:35.244387 systemd[1]: Starting issuegen.service - Generate /run/issue... Jun 25 18:47:35.252716 systemd[1]: issuegen.service: Deactivated successfully. Jun 25 18:47:35.253219 systemd[1]: Finished issuegen.service - Generate /run/issue. Jun 25 18:47:35.272517 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jun 25 18:47:35.287195 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jun 25 18:47:35.298081 containerd[1570]: time="2024-06-25T18:47:35.298019368Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jun 25 18:47:35.298081 containerd[1570]: time="2024-06-25T18:47:35.298083649Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jun 25 18:47:35.298224 containerd[1570]: time="2024-06-25T18:47:35.298103276Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jun 25 18:47:35.298245 containerd[1570]: time="2024-06-25T18:47:35.298231135Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jun 25 18:47:35.298266 containerd[1570]: time="2024-06-25T18:47:35.298252796Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jun 25 18:47:35.298286 containerd[1570]: time="2024-06-25T18:47:35.298269307Z" level=info msg="NRI interface is disabled by configuration." Jun 25 18:47:35.298306 containerd[1570]: time="2024-06-25T18:47:35.298286409Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298452891Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298482367Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298501452Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298519576Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298536979Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298559702Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298583075Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298599286Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298615456Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298633620Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298648769Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jun 25 18:47:35.298675 containerd[1570]: time="2024-06-25T18:47:35.298664187Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jun 25 18:47:35.298900 containerd[1570]: time="2024-06-25T18:47:35.298799191Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jun 25 18:47:35.299602 systemd[1]: Started getty@tty1.service - Getty on tty1. Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.299630930Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.299664172Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.299682487Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.299709427Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.299779519Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.299797322Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.299948065Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.299969174Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.299988440Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.300005573Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.300021462Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.300035879Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.300069402Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jun 25 18:47:35.300829 containerd[1570]: time="2024-06-25T18:47:35.300274387Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.301127 containerd[1570]: time="2024-06-25T18:47:35.300304132Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.301127 containerd[1570]: time="2024-06-25T18:47:35.300321385Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.301127 containerd[1570]: time="2024-06-25T18:47:35.300339419Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.301127 containerd[1570]: time="2024-06-25T18:47:35.300355248Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.301127 containerd[1570]: time="2024-06-25T18:47:35.300372961Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.301127 containerd[1570]: time="2024-06-25T18:47:35.300388871Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.301127 containerd[1570]: time="2024-06-25T18:47:35.300403088Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jun 25 18:47:35.301313 containerd[1570]: time="2024-06-25T18:47:35.300719261Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jun 25 18:47:35.301313 containerd[1570]: time="2024-06-25T18:47:35.300788781Z" level=info msg="Connect containerd service" Jun 25 18:47:35.301313 containerd[1570]: time="2024-06-25T18:47:35.300816132Z" level=info msg="using legacy CRI server" Jun 25 18:47:35.301313 containerd[1570]: time="2024-06-25T18:47:35.300826933Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jun 25 18:47:35.301313 containerd[1570]: time="2024-06-25T18:47:35.300957568Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.301693127Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.301752017Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.301773477Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.301874617Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.301894695Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.301845492Z" level=info msg="Start subscribing containerd event" Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.302091193Z" level=info msg="Start recovering state" Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.302206509Z" level=info msg="Start event monitor" Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.302312688Z" level=info msg="Start snapshots syncer" Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.302435849Z" level=info msg="Start cni network conf syncer for default" Jun 25 18:47:35.302695 containerd[1570]: time="2024-06-25T18:47:35.302449505Z" level=info msg="Start streaming server" Jun 25 18:47:35.302930 containerd[1570]: time="2024-06-25T18:47:35.302794792Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jun 25 18:47:35.303978 containerd[1570]: time="2024-06-25T18:47:35.303003584Z" level=info msg=serving... address=/run/containerd/containerd.sock Jun 25 18:47:35.304420 containerd[1570]: time="2024-06-25T18:47:35.304399621Z" level=info msg="containerd successfully booted in 0.124912s" Jun 25 18:47:35.307115 systemd[1]: Started serial-getty@ttyS0.service - Serial Getty on ttyS0. Jun 25 18:47:35.309514 systemd[1]: Reached target getty.target - Login Prompts. Jun 25 18:47:35.312605 systemd[1]: Started containerd.service - containerd container runtime. Jun 25 18:47:35.547722 tar[1565]: linux-amd64/LICENSE Jun 25 18:47:35.547811 tar[1565]: linux-amd64/README.md Jun 25 18:47:35.563672 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jun 25 18:47:35.844710 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jun 25 18:47:35.846402 systemd[1]: Reached target multi-user.target - Multi-User System. Jun 25 18:47:35.847622 systemd[1]: Startup finished in 5.892s (kernel) + 4.331s (userspace) = 10.224s. Jun 25 18:47:35.853030 (kubelet)[1673]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jun 25 18:47:36.340798 kubelet[1673]: E0625 18:47:36.340650 1673 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jun 25 18:47:36.345137 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jun 25 18:47:36.345442 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jun 25 18:47:44.297979 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jun 25 18:47:44.311397 systemd[1]: Started sshd@0-10.0.0.150:22-10.0.0.1:60380.service - OpenSSH per-connection server daemon (10.0.0.1:60380). Jun 25 18:47:44.343915 sshd[1688]: Accepted publickey for core from 10.0.0.1 port 60380 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:47:44.345850 sshd[1688]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:47:44.355004 systemd-logind[1553]: New session 1 of user core. Jun 25 18:47:44.356109 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jun 25 18:47:44.365333 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jun 25 18:47:44.377022 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jun 25 18:47:44.392487 systemd[1]: Starting user@500.service - User Manager for UID 500... Jun 25 18:47:44.395419 (systemd)[1693]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:47:44.491256 systemd[1693]: Queued start job for default target default.target. Jun 25 18:47:44.491606 systemd[1693]: Created slice app.slice - User Application Slice. Jun 25 18:47:44.491628 systemd[1693]: Reached target paths.target - Paths. Jun 25 18:47:44.491641 systemd[1693]: Reached target timers.target - Timers. Jun 25 18:47:44.500231 systemd[1693]: Starting dbus.socket - D-Bus User Message Bus Socket... Jun 25 18:47:44.506531 systemd[1693]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jun 25 18:47:44.506591 systemd[1693]: Reached target sockets.target - Sockets. Jun 25 18:47:44.506603 systemd[1693]: Reached target basic.target - Basic System. Jun 25 18:47:44.506638 systemd[1693]: Reached target default.target - Main User Target. Jun 25 18:47:44.506668 systemd[1693]: Startup finished in 104ms. Jun 25 18:47:44.507496 systemd[1]: Started user@500.service - User Manager for UID 500. Jun 25 18:47:44.509407 systemd[1]: Started session-1.scope - Session 1 of User core. Jun 25 18:47:44.567432 systemd[1]: Started sshd@1-10.0.0.150:22-10.0.0.1:60388.service - OpenSSH per-connection server daemon (10.0.0.1:60388). Jun 25 18:47:44.596809 sshd[1706]: Accepted publickey for core from 10.0.0.1 port 60388 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:47:44.598336 sshd[1706]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:47:44.602256 systemd-logind[1553]: New session 2 of user core. Jun 25 18:47:44.608463 systemd[1]: Started session-2.scope - Session 2 of User core. Jun 25 18:47:44.660999 sshd[1706]: pam_unix(sshd:session): session closed for user core Jun 25 18:47:44.673397 systemd[1]: Started sshd@2-10.0.0.150:22-10.0.0.1:60390.service - OpenSSH per-connection server daemon (10.0.0.1:60390). Jun 25 18:47:44.673858 systemd[1]: sshd@1-10.0.0.150:22-10.0.0.1:60388.service: Deactivated successfully. Jun 25 18:47:44.676067 systemd-logind[1553]: Session 2 logged out. Waiting for processes to exit. Jun 25 18:47:44.677007 systemd[1]: session-2.scope: Deactivated successfully. Jun 25 18:47:44.677961 systemd-logind[1553]: Removed session 2. Jun 25 18:47:44.698352 sshd[1711]: Accepted publickey for core from 10.0.0.1 port 60390 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:47:44.699866 sshd[1711]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:47:44.703505 systemd-logind[1553]: New session 3 of user core. Jun 25 18:47:44.714380 systemd[1]: Started session-3.scope - Session 3 of User core. Jun 25 18:47:44.763234 sshd[1711]: pam_unix(sshd:session): session closed for user core Jun 25 18:47:44.772363 systemd[1]: Started sshd@3-10.0.0.150:22-10.0.0.1:60394.service - OpenSSH per-connection server daemon (10.0.0.1:60394). Jun 25 18:47:44.772838 systemd[1]: sshd@2-10.0.0.150:22-10.0.0.1:60390.service: Deactivated successfully. Jun 25 18:47:44.775212 systemd-logind[1553]: Session 3 logged out. Waiting for processes to exit. Jun 25 18:47:44.776166 systemd[1]: session-3.scope: Deactivated successfully. Jun 25 18:47:44.777368 systemd-logind[1553]: Removed session 3. Jun 25 18:47:44.797558 sshd[1719]: Accepted publickey for core from 10.0.0.1 port 60394 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:47:44.799091 sshd[1719]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:47:44.803213 systemd-logind[1553]: New session 4 of user core. Jun 25 18:47:44.814396 systemd[1]: Started session-4.scope - Session 4 of User core. Jun 25 18:47:44.868456 sshd[1719]: pam_unix(sshd:session): session closed for user core Jun 25 18:47:44.876390 systemd[1]: Started sshd@4-10.0.0.150:22-10.0.0.1:60410.service - OpenSSH per-connection server daemon (10.0.0.1:60410). Jun 25 18:47:44.876880 systemd[1]: sshd@3-10.0.0.150:22-10.0.0.1:60394.service: Deactivated successfully. Jun 25 18:47:44.880559 systemd-logind[1553]: Session 4 logged out. Waiting for processes to exit. Jun 25 18:47:44.881231 systemd[1]: session-4.scope: Deactivated successfully. Jun 25 18:47:44.881945 systemd-logind[1553]: Removed session 4. Jun 25 18:47:44.901813 sshd[1727]: Accepted publickey for core from 10.0.0.1 port 60410 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:47:44.903150 sshd[1727]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:47:44.906651 systemd-logind[1553]: New session 5 of user core. Jun 25 18:47:44.917402 systemd[1]: Started session-5.scope - Session 5 of User core. Jun 25 18:47:44.975061 sudo[1734]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jun 25 18:47:44.975375 sudo[1734]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Jun 25 18:47:44.993988 sudo[1734]: pam_unix(sudo:session): session closed for user root Jun 25 18:47:44.995944 sshd[1727]: pam_unix(sshd:session): session closed for user core Jun 25 18:47:45.005431 systemd[1]: Started sshd@5-10.0.0.150:22-10.0.0.1:60426.service - OpenSSH per-connection server daemon (10.0.0.1:60426). Jun 25 18:47:45.006127 systemd[1]: sshd@4-10.0.0.150:22-10.0.0.1:60410.service: Deactivated successfully. Jun 25 18:47:45.008174 systemd[1]: session-5.scope: Deactivated successfully. Jun 25 18:47:45.008962 systemd-logind[1553]: Session 5 logged out. Waiting for processes to exit. Jun 25 18:47:45.010553 systemd-logind[1553]: Removed session 5. Jun 25 18:47:45.031119 sshd[1737]: Accepted publickey for core from 10.0.0.1 port 60426 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:47:45.032536 sshd[1737]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:47:45.036249 systemd-logind[1553]: New session 6 of user core. Jun 25 18:47:45.051549 systemd[1]: Started session-6.scope - Session 6 of User core. Jun 25 18:47:45.106549 sudo[1744]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jun 25 18:47:45.106848 sudo[1744]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Jun 25 18:47:45.110664 sudo[1744]: pam_unix(sudo:session): session closed for user root Jun 25 18:47:45.116327 sudo[1743]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/systemctl restart audit-rules Jun 25 18:47:45.116619 sudo[1743]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Jun 25 18:47:45.136374 systemd[1]: Stopping audit-rules.service - Load Security Auditing Rules... Jun 25 18:47:45.138098 auditctl[1747]: No rules Jun 25 18:47:45.138530 systemd[1]: audit-rules.service: Deactivated successfully. Jun 25 18:47:45.138868 systemd[1]: Stopped audit-rules.service - Load Security Auditing Rules. Jun 25 18:47:45.141728 systemd[1]: Starting audit-rules.service - Load Security Auditing Rules... Jun 25 18:47:45.170565 augenrules[1766]: No rules Jun 25 18:47:45.171552 systemd[1]: Finished audit-rules.service - Load Security Auditing Rules. Jun 25 18:47:45.173116 sudo[1743]: pam_unix(sudo:session): session closed for user root Jun 25 18:47:45.174997 sshd[1737]: pam_unix(sshd:session): session closed for user core Jun 25 18:47:45.185382 systemd[1]: Started sshd@6-10.0.0.150:22-10.0.0.1:60428.service - OpenSSH per-connection server daemon (10.0.0.1:60428). Jun 25 18:47:45.185938 systemd[1]: sshd@5-10.0.0.150:22-10.0.0.1:60426.service: Deactivated successfully. Jun 25 18:47:45.189021 systemd-logind[1553]: Session 6 logged out. Waiting for processes to exit. Jun 25 18:47:45.189220 systemd[1]: session-6.scope: Deactivated successfully. Jun 25 18:47:45.190542 systemd-logind[1553]: Removed session 6. Jun 25 18:47:45.214369 sshd[1772]: Accepted publickey for core from 10.0.0.1 port 60428 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:47:45.215731 sshd[1772]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:47:45.219679 systemd-logind[1553]: New session 7 of user core. Jun 25 18:47:45.229541 systemd[1]: Started session-7.scope - Session 7 of User core. Jun 25 18:47:45.283474 sudo[1779]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jun 25 18:47:45.283771 sudo[1779]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Jun 25 18:47:45.395350 systemd[1]: Starting docker.service - Docker Application Container Engine... Jun 25 18:47:45.395726 (dockerd)[1789]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jun 25 18:47:45.632954 dockerd[1789]: time="2024-06-25T18:47:45.632889417Z" level=info msg="Starting up" Jun 25 18:47:46.180892 dockerd[1789]: time="2024-06-25T18:47:46.180837810Z" level=info msg="Loading containers: start." Jun 25 18:47:46.299189 kernel: Initializing XFRM netlink socket Jun 25 18:47:46.378455 systemd-networkd[1249]: docker0: Link UP Jun 25 18:47:46.391981 dockerd[1789]: time="2024-06-25T18:47:46.391940798Z" level=info msg="Loading containers: done." Jun 25 18:47:46.441529 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck468987759-merged.mount: Deactivated successfully. Jun 25 18:47:46.442540 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jun 25 18:47:46.443766 dockerd[1789]: time="2024-06-25T18:47:46.443721717Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jun 25 18:47:46.443969 dockerd[1789]: time="2024-06-25T18:47:46.443938865Z" level=info msg="Docker daemon" commit=fca702de7f71362c8d103073c7e4a1d0a467fadd graphdriver=overlay2 version=24.0.9 Jun 25 18:47:46.444108 dockerd[1789]: time="2024-06-25T18:47:46.444083716Z" level=info msg="Daemon has completed initialization" Jun 25 18:47:46.448285 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jun 25 18:47:46.476752 dockerd[1789]: time="2024-06-25T18:47:46.476698138Z" level=info msg="API listen on /run/docker.sock" Jun 25 18:47:46.476901 systemd[1]: Started docker.service - Docker Application Container Engine. Jun 25 18:47:46.594628 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jun 25 18:47:46.601820 (kubelet)[1933]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jun 25 18:47:46.668398 kubelet[1933]: E0625 18:47:46.668340 1933 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jun 25 18:47:46.676171 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jun 25 18:47:46.676465 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jun 25 18:47:47.228264 containerd[1570]: time="2024-06-25T18:47:47.228214571Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.28.11\"" Jun 25 18:47:48.122803 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1206187284.mount: Deactivated successfully. Jun 25 18:47:49.313925 containerd[1570]: time="2024-06-25T18:47:49.313852532Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.28.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:49.314624 containerd[1570]: time="2024-06-25T18:47:49.314546162Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.28.11: active requests=0, bytes read=34605178" Jun 25 18:47:49.315799 containerd[1570]: time="2024-06-25T18:47:49.315760099Z" level=info msg="ImageCreate event name:\"sha256:b2de212bf8c1b7b0d1b2703356ac7ddcfccaadfcdcd32c1ae914b6078d11e524\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:49.319192 containerd[1570]: time="2024-06-25T18:47:49.319115991Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:aec9d1701c304eee8607d728a39baaa511d65bef6dd9861010618f63fbadeb10\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:49.320498 containerd[1570]: time="2024-06-25T18:47:49.320451285Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.28.11\" with image id \"sha256:b2de212bf8c1b7b0d1b2703356ac7ddcfccaadfcdcd32c1ae914b6078d11e524\", repo tag \"registry.k8s.io/kube-apiserver:v1.28.11\", repo digest \"registry.k8s.io/kube-apiserver@sha256:aec9d1701c304eee8607d728a39baaa511d65bef6dd9861010618f63fbadeb10\", size \"34601978\" in 2.092200145s" Jun 25 18:47:49.320498 containerd[1570]: time="2024-06-25T18:47:49.320491450Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.28.11\" returns image reference \"sha256:b2de212bf8c1b7b0d1b2703356ac7ddcfccaadfcdcd32c1ae914b6078d11e524\"" Jun 25 18:47:49.347608 containerd[1570]: time="2024-06-25T18:47:49.347545496Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.28.11\"" Jun 25 18:47:51.005606 containerd[1570]: time="2024-06-25T18:47:51.005545476Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.28.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:51.006494 containerd[1570]: time="2024-06-25T18:47:51.006455532Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.28.11: active requests=0, bytes read=31719491" Jun 25 18:47:51.007660 containerd[1570]: time="2024-06-25T18:47:51.007610147Z" level=info msg="ImageCreate event name:\"sha256:20145ae80ad309fd0c963e2539f6ef0be795ace696539514894b290892c1884b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:51.010540 containerd[1570]: time="2024-06-25T18:47:51.010506498Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:6014c3572ec683841bbb16f87b94da28ee0254b95e2dba2d1850d62bd0111f09\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:51.011421 containerd[1570]: time="2024-06-25T18:47:51.011396627Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.28.11\" with image id \"sha256:20145ae80ad309fd0c963e2539f6ef0be795ace696539514894b290892c1884b\", repo tag \"registry.k8s.io/kube-controller-manager:v1.28.11\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:6014c3572ec683841bbb16f87b94da28ee0254b95e2dba2d1850d62bd0111f09\", size \"33315989\" in 1.663800927s" Jun 25 18:47:51.011468 containerd[1570]: time="2024-06-25T18:47:51.011423527Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.28.11\" returns image reference \"sha256:20145ae80ad309fd0c963e2539f6ef0be795ace696539514894b290892c1884b\"" Jun 25 18:47:51.034585 containerd[1570]: time="2024-06-25T18:47:51.034535472Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.28.11\"" Jun 25 18:47:52.212904 containerd[1570]: time="2024-06-25T18:47:52.212848178Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.28.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:52.213716 containerd[1570]: time="2024-06-25T18:47:52.213680529Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.28.11: active requests=0, bytes read=16925505" Jun 25 18:47:52.214973 containerd[1570]: time="2024-06-25T18:47:52.214941964Z" level=info msg="ImageCreate event name:\"sha256:12c62a5a0745d200eb8333ea6244f6d6328e64c5c3b645a4ade456cc645399b9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:52.217413 containerd[1570]: time="2024-06-25T18:47:52.217383792Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:46cf7475c8daffb743c856a1aea0ddea35e5acd2418be18b1e22cf98d9c9b445\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:52.218380 containerd[1570]: time="2024-06-25T18:47:52.218353210Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.28.11\" with image id \"sha256:12c62a5a0745d200eb8333ea6244f6d6328e64c5c3b645a4ade456cc645399b9\", repo tag \"registry.k8s.io/kube-scheduler:v1.28.11\", repo digest \"registry.k8s.io/kube-scheduler@sha256:46cf7475c8daffb743c856a1aea0ddea35e5acd2418be18b1e22cf98d9c9b445\", size \"18522021\" in 1.18377606s" Jun 25 18:47:52.218419 containerd[1570]: time="2024-06-25T18:47:52.218380080Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.28.11\" returns image reference \"sha256:12c62a5a0745d200eb8333ea6244f6d6328e64c5c3b645a4ade456cc645399b9\"" Jun 25 18:47:52.238448 containerd[1570]: time="2024-06-25T18:47:52.238408123Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.11\"" Jun 25 18:47:53.169819 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2568580664.mount: Deactivated successfully. Jun 25 18:47:53.855289 containerd[1570]: time="2024-06-25T18:47:53.855229863Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.28.11\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:53.877624 containerd[1570]: time="2024-06-25T18:47:53.877582204Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.28.11: active requests=0, bytes read=28118419" Jun 25 18:47:53.901656 containerd[1570]: time="2024-06-25T18:47:53.901602772Z" level=info msg="ImageCreate event name:\"sha256:a3eea76ce409e136fe98838847fda217ce169eb7d1ceef544671d75f68e5a29c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:53.924233 containerd[1570]: time="2024-06-25T18:47:53.924192157Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:ae4b671d4cfc23dd75030bb4490207cd939b3b11a799bcb4119698cd712eb5b4\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:53.924894 containerd[1570]: time="2024-06-25T18:47:53.924859088Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.28.11\" with image id \"sha256:a3eea76ce409e136fe98838847fda217ce169eb7d1ceef544671d75f68e5a29c\", repo tag \"registry.k8s.io/kube-proxy:v1.28.11\", repo digest \"registry.k8s.io/kube-proxy@sha256:ae4b671d4cfc23dd75030bb4490207cd939b3b11a799bcb4119698cd712eb5b4\", size \"28117438\" in 1.686415128s" Jun 25 18:47:53.924894 containerd[1570]: time="2024-06-25T18:47:53.924887982Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.28.11\" returns image reference \"sha256:a3eea76ce409e136fe98838847fda217ce169eb7d1ceef544671d75f68e5a29c\"" Jun 25 18:47:53.947975 containerd[1570]: time="2024-06-25T18:47:53.947934474Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Jun 25 18:47:54.558431 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount438143480.mount: Deactivated successfully. Jun 25 18:47:54.564450 containerd[1570]: time="2024-06-25T18:47:54.564398715Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:54.565189 containerd[1570]: time="2024-06-25T18:47:54.565127551Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=322290" Jun 25 18:47:54.566435 containerd[1570]: time="2024-06-25T18:47:54.566390219Z" level=info msg="ImageCreate event name:\"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:54.568584 containerd[1570]: time="2024-06-25T18:47:54.568526033Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:54.569323 containerd[1570]: time="2024-06-25T18:47:54.569263466Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"321520\" in 621.122915ms" Jun 25 18:47:54.569323 containerd[1570]: time="2024-06-25T18:47:54.569310083Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:e6f1816883972d4be47bd48879a08919b96afcd344132622e4d444987919323c\"" Jun 25 18:47:54.590471 containerd[1570]: time="2024-06-25T18:47:54.590437357Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\"" Jun 25 18:47:55.130014 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount896371776.mount: Deactivated successfully. Jun 25 18:47:56.798064 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jun 25 18:47:56.811422 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jun 25 18:47:56.950700 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jun 25 18:47:56.955797 (kubelet)[2114]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jun 25 18:47:57.013397 kubelet[2114]: E0625 18:47:57.013325 2114 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jun 25 18:47:57.018239 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jun 25 18:47:57.018517 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jun 25 18:47:57.361788 containerd[1570]: time="2024-06-25T18:47:57.361728066Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.10-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:57.362640 containerd[1570]: time="2024-06-25T18:47:57.362592157Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.10-0: active requests=0, bytes read=56651625" Jun 25 18:47:57.363934 containerd[1570]: time="2024-06-25T18:47:57.363897714Z" level=info msg="ImageCreate event name:\"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:57.366913 containerd[1570]: time="2024-06-25T18:47:57.366878043Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:57.368123 containerd[1570]: time="2024-06-25T18:47:57.368085857Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.10-0\" with image id \"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\", repo tag \"registry.k8s.io/etcd:3.5.10-0\", repo digest \"registry.k8s.io/etcd@sha256:22f892d7672adc0b9c86df67792afdb8b2dc08880f49f669eaaa59c47d7908c2\", size \"56649232\" in 2.777519448s" Jun 25 18:47:57.368184 containerd[1570]: time="2024-06-25T18:47:57.368125682Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.10-0\" returns image reference \"sha256:a0eed15eed4498c145ef2f1883fcd300d7adbb759df73c901abd5383dda668e7\"" Jun 25 18:47:57.500061 containerd[1570]: time="2024-06-25T18:47:57.500009574Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\"" Jun 25 18:47:58.108532 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4231353051.mount: Deactivated successfully. Jun 25 18:47:58.627352 containerd[1570]: time="2024-06-25T18:47:58.627276399Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.10.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:58.628204 containerd[1570]: time="2024-06-25T18:47:58.628109271Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.10.1: active requests=0, bytes read=16191749" Jun 25 18:47:58.629286 containerd[1570]: time="2024-06-25T18:47:58.629253446Z" level=info msg="ImageCreate event name:\"sha256:ead0a4a53df89fd173874b46093b6e62d8c72967bbf606d672c9e8c9b601a4fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:58.632424 containerd[1570]: time="2024-06-25T18:47:58.632379868Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:47:58.633385 containerd[1570]: time="2024-06-25T18:47:58.633337694Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.10.1\" with image id \"sha256:ead0a4a53df89fd173874b46093b6e62d8c72967bbf606d672c9e8c9b601a4fc\", repo tag \"registry.k8s.io/coredns/coredns:v1.10.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e\", size \"16190758\" in 1.133278527s" Jun 25 18:47:58.633430 containerd[1570]: time="2024-06-25T18:47:58.633389211Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\" returns image reference \"sha256:ead0a4a53df89fd173874b46093b6e62d8c72967bbf606d672c9e8c9b601a4fc\"" Jun 25 18:48:00.872340 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jun 25 18:48:00.882515 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jun 25 18:48:00.912685 systemd[1]: Reloading requested from client PID 2219 ('systemctl') (unit session-7.scope)... Jun 25 18:48:00.912707 systemd[1]: Reloading... Jun 25 18:48:01.010187 zram_generator::config[2259]: No configuration found. Jun 25 18:48:01.164600 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jun 25 18:48:01.240604 systemd[1]: Reloading finished in 327 ms. Jun 25 18:48:01.291604 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jun 25 18:48:01.291710 systemd[1]: kubelet.service: Failed with result 'signal'. Jun 25 18:48:01.292061 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jun 25 18:48:01.300545 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jun 25 18:48:01.445757 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jun 25 18:48:01.450661 (kubelet)[2316]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jun 25 18:48:01.510067 kubelet[2316]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jun 25 18:48:01.510067 kubelet[2316]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jun 25 18:48:01.510067 kubelet[2316]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jun 25 18:48:01.510501 kubelet[2316]: I0625 18:48:01.510111 2316 server.go:203] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jun 25 18:48:02.094776 kubelet[2316]: I0625 18:48:02.094727 2316 server.go:467] "Kubelet version" kubeletVersion="v1.28.7" Jun 25 18:48:02.094776 kubelet[2316]: I0625 18:48:02.094761 2316 server.go:469] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jun 25 18:48:02.096609 kubelet[2316]: I0625 18:48:02.095452 2316 server.go:895] "Client rotation is on, will bootstrap in background" Jun 25 18:48:02.109931 kubelet[2316]: I0625 18:48:02.109888 2316 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jun 25 18:48:02.110304 kubelet[2316]: E0625 18:48:02.110250 2316 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.150:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:02.119809 kubelet[2316]: I0625 18:48:02.119789 2316 server.go:725] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jun 25 18:48:02.120835 kubelet[2316]: I0625 18:48:02.120811 2316 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jun 25 18:48:02.120977 kubelet[2316]: I0625 18:48:02.120960 2316 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jun 25 18:48:02.121342 kubelet[2316]: I0625 18:48:02.121326 2316 topology_manager.go:138] "Creating topology manager with none policy" Jun 25 18:48:02.121342 kubelet[2316]: I0625 18:48:02.121341 2316 container_manager_linux.go:301] "Creating device plugin manager" Jun 25 18:48:02.121879 kubelet[2316]: I0625 18:48:02.121859 2316 state_mem.go:36] "Initialized new in-memory state store" Jun 25 18:48:02.122984 kubelet[2316]: I0625 18:48:02.122960 2316 kubelet.go:393] "Attempting to sync node with API server" Jun 25 18:48:02.122984 kubelet[2316]: I0625 18:48:02.122978 2316 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests" Jun 25 18:48:02.123042 kubelet[2316]: I0625 18:48:02.123001 2316 kubelet.go:309] "Adding apiserver pod source" Jun 25 18:48:02.123042 kubelet[2316]: I0625 18:48:02.123017 2316 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jun 25 18:48:02.124347 kubelet[2316]: I0625 18:48:02.124045 2316 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="v1.7.18" apiVersion="v1" Jun 25 18:48:02.125331 kubelet[2316]: W0625 18:48:02.125279 2316 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://10.0.0.150:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:02.125331 kubelet[2316]: E0625 18:48:02.125331 2316 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.150:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:02.126825 kubelet[2316]: W0625 18:48:02.126779 2316 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://10.0.0.150:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:02.126883 kubelet[2316]: E0625 18:48:02.126834 2316 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.150:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:02.127039 kubelet[2316]: W0625 18:48:02.127026 2316 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jun 25 18:48:02.130737 kubelet[2316]: I0625 18:48:02.127885 2316 server.go:1232] "Started kubelet" Jun 25 18:48:02.130737 kubelet[2316]: I0625 18:48:02.128420 2316 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Jun 25 18:48:02.130737 kubelet[2316]: I0625 18:48:02.128792 2316 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10 Jun 25 18:48:02.130737 kubelet[2316]: I0625 18:48:02.129051 2316 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jun 25 18:48:02.130737 kubelet[2316]: I0625 18:48:02.129424 2316 server.go:462] "Adding debug handlers to kubelet server" Jun 25 18:48:02.131749 kubelet[2316]: I0625 18:48:02.131729 2316 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jun 25 18:48:02.134088 kubelet[2316]: I0625 18:48:02.133900 2316 volume_manager.go:291] "Starting Kubelet Volume Manager" Jun 25 18:48:02.135750 kubelet[2316]: E0625 18:48:02.134555 2316 event.go:289] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"localhost.17dc53c960e52539", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"localhost", UID:"localhost", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"localhost"}, FirstTimestamp:time.Date(2024, time.June, 25, 18, 48, 2, 127848761, time.Local), LastTimestamp:time.Date(2024, time.June, 25, 18, 48, 2, 127848761, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"kubelet", ReportingInstance:"localhost"}': 'Post "https://10.0.0.150:6443/api/v1/namespaces/default/events": dial tcp 10.0.0.150:6443: connect: connection refused'(may retry after sleeping) Jun 25 18:48:02.136364 kubelet[2316]: I0625 18:48:02.136342 2316 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Jun 25 18:48:02.136440 kubelet[2316]: I0625 18:48:02.136427 2316 reconciler_new.go:29] "Reconciler: start to sync state" Jun 25 18:48:02.137023 kubelet[2316]: E0625 18:48:02.137000 2316 cri_stats_provider.go:448] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Jun 25 18:48:02.137049 kubelet[2316]: E0625 18:48:02.137031 2316 kubelet.go:1431] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jun 25 18:48:02.137577 kubelet[2316]: W0625 18:48:02.137529 2316 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://10.0.0.150:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:02.137577 kubelet[2316]: E0625 18:48:02.137573 2316 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.150:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:02.137649 kubelet[2316]: E0625 18:48:02.137639 2316 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.150:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.150:6443: connect: connection refused" interval="200ms" Jun 25 18:48:02.154336 kubelet[2316]: I0625 18:48:02.154299 2316 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jun 25 18:48:02.155810 kubelet[2316]: I0625 18:48:02.155517 2316 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jun 25 18:48:02.155810 kubelet[2316]: I0625 18:48:02.155534 2316 status_manager.go:217] "Starting to sync pod status with apiserver" Jun 25 18:48:02.155810 kubelet[2316]: I0625 18:48:02.155550 2316 kubelet.go:2303] "Starting kubelet main sync loop" Jun 25 18:48:02.155810 kubelet[2316]: E0625 18:48:02.155597 2316 kubelet.go:2327] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jun 25 18:48:02.156704 kubelet[2316]: W0625 18:48:02.156656 2316 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://10.0.0.150:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:02.156704 kubelet[2316]: E0625 18:48:02.156680 2316 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.150:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:02.190825 kubelet[2316]: I0625 18:48:02.190797 2316 cpu_manager.go:214] "Starting CPU manager" policy="none" Jun 25 18:48:02.190825 kubelet[2316]: I0625 18:48:02.190817 2316 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jun 25 18:48:02.190825 kubelet[2316]: I0625 18:48:02.190835 2316 state_mem.go:36] "Initialized new in-memory state store" Jun 25 18:48:02.235939 kubelet[2316]: I0625 18:48:02.235898 2316 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Jun 25 18:48:02.236235 kubelet[2316]: E0625 18:48:02.236202 2316 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.150:6443/api/v1/nodes\": dial tcp 10.0.0.150:6443: connect: connection refused" node="localhost" Jun 25 18:48:02.256487 kubelet[2316]: E0625 18:48:02.256438 2316 kubelet.go:2327] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jun 25 18:48:02.338235 kubelet[2316]: E0625 18:48:02.338184 2316 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.150:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.150:6443: connect: connection refused" interval="400ms" Jun 25 18:48:02.437571 kubelet[2316]: I0625 18:48:02.437402 2316 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Jun 25 18:48:02.437730 kubelet[2316]: E0625 18:48:02.437707 2316 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.150:6443/api/v1/nodes\": dial tcp 10.0.0.150:6443: connect: connection refused" node="localhost" Jun 25 18:48:02.456947 kubelet[2316]: E0625 18:48:02.456912 2316 kubelet.go:2327] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jun 25 18:48:02.649259 kubelet[2316]: I0625 18:48:02.649195 2316 policy_none.go:49] "None policy: Start" Jun 25 18:48:02.650054 kubelet[2316]: I0625 18:48:02.650037 2316 memory_manager.go:169] "Starting memorymanager" policy="None" Jun 25 18:48:02.650110 kubelet[2316]: I0625 18:48:02.650073 2316 state_mem.go:35] "Initializing new in-memory state store" Jun 25 18:48:02.659076 kubelet[2316]: I0625 18:48:02.657228 2316 manager.go:471] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jun 25 18:48:02.659076 kubelet[2316]: I0625 18:48:02.657533 2316 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jun 25 18:48:02.659374 kubelet[2316]: E0625 18:48:02.659348 2316 eviction_manager.go:258] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"localhost\" not found" Jun 25 18:48:02.739519 kubelet[2316]: E0625 18:48:02.739378 2316 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.150:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.150:6443: connect: connection refused" interval="800ms" Jun 25 18:48:02.839948 kubelet[2316]: I0625 18:48:02.839905 2316 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Jun 25 18:48:02.840387 kubelet[2316]: E0625 18:48:02.840367 2316 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.150:6443/api/v1/nodes\": dial tcp 10.0.0.150:6443: connect: connection refused" node="localhost" Jun 25 18:48:02.857658 kubelet[2316]: I0625 18:48:02.857635 2316 topology_manager.go:215] "Topology Admit Handler" podUID="90568ebd3309c6cbd50a67732ab2ba14" podNamespace="kube-system" podName="kube-apiserver-localhost" Jun 25 18:48:02.858652 kubelet[2316]: I0625 18:48:02.858626 2316 topology_manager.go:215] "Topology Admit Handler" podUID="d27baad490d2d4f748c86b318d7d74ef" podNamespace="kube-system" podName="kube-controller-manager-localhost" Jun 25 18:48:02.859546 kubelet[2316]: I0625 18:48:02.859501 2316 topology_manager.go:215] "Topology Admit Handler" podUID="9c3207d669e00aa24ded52617c0d65d0" podNamespace="kube-system" podName="kube-scheduler-localhost" Jun 25 18:48:02.940884 kubelet[2316]: I0625 18:48:02.940820 2316 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/90568ebd3309c6cbd50a67732ab2ba14-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"90568ebd3309c6cbd50a67732ab2ba14\") " pod="kube-system/kube-apiserver-localhost" Jun 25 18:48:02.940884 kubelet[2316]: I0625 18:48:02.940872 2316 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:02.940884 kubelet[2316]: I0625 18:48:02.940895 2316 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:02.941099 kubelet[2316]: I0625 18:48:02.940918 2316 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:02.941099 kubelet[2316]: I0625 18:48:02.940942 2316 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:02.941099 kubelet[2316]: I0625 18:48:02.940963 2316 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/90568ebd3309c6cbd50a67732ab2ba14-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"90568ebd3309c6cbd50a67732ab2ba14\") " pod="kube-system/kube-apiserver-localhost" Jun 25 18:48:02.941099 kubelet[2316]: I0625 18:48:02.940983 2316 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:02.941099 kubelet[2316]: I0625 18:48:02.941003 2316 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/9c3207d669e00aa24ded52617c0d65d0-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"9c3207d669e00aa24ded52617c0d65d0\") " pod="kube-system/kube-scheduler-localhost" Jun 25 18:48:02.941263 kubelet[2316]: I0625 18:48:02.941022 2316 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/90568ebd3309c6cbd50a67732ab2ba14-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"90568ebd3309c6cbd50a67732ab2ba14\") " pod="kube-system/kube-apiserver-localhost" Jun 25 18:48:03.012667 kubelet[2316]: W0625 18:48:03.012503 2316 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://10.0.0.150:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:03.012667 kubelet[2316]: E0625 18:48:03.012576 2316 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.150:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:03.018887 kubelet[2316]: W0625 18:48:03.018844 2316 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://10.0.0.150:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:03.018887 kubelet[2316]: E0625 18:48:03.018877 2316 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.0.0.150:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:03.163121 kubelet[2316]: E0625 18:48:03.163090 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:03.163782 containerd[1570]: time="2024-06-25T18:48:03.163736220Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:90568ebd3309c6cbd50a67732ab2ba14,Namespace:kube-system,Attempt:0,}" Jun 25 18:48:03.164948 kubelet[2316]: E0625 18:48:03.164912 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:03.165502 containerd[1570]: time="2024-06-25T18:48:03.165274214Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:d27baad490d2d4f748c86b318d7d74ef,Namespace:kube-system,Attempt:0,}" Jun 25 18:48:03.166004 kubelet[2316]: E0625 18:48:03.165965 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:03.166396 containerd[1570]: time="2024-06-25T18:48:03.166363266Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:9c3207d669e00aa24ded52617c0d65d0,Namespace:kube-system,Attempt:0,}" Jun 25 18:48:03.247604 kubelet[2316]: W0625 18:48:03.247529 2316 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://10.0.0.150:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:03.247604 kubelet[2316]: E0625 18:48:03.247601 2316 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.0.0.150:6443/api/v1/nodes?fieldSelector=metadata.name%3Dlocalhost&limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:03.540456 kubelet[2316]: E0625 18:48:03.540407 2316 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.0.0.150:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/localhost?timeout=10s\": dial tcp 10.0.0.150:6443: connect: connection refused" interval="1.6s" Jun 25 18:48:03.619315 kubelet[2316]: W0625 18:48:03.619249 2316 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://10.0.0.150:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:03.619315 kubelet[2316]: E0625 18:48:03.619310 2316 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.0.0.150:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:03.641671 kubelet[2316]: I0625 18:48:03.641621 2316 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Jun 25 18:48:03.642191 kubelet[2316]: E0625 18:48:03.642121 2316 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.0.0.150:6443/api/v1/nodes\": dial tcp 10.0.0.150:6443: connect: connection refused" node="localhost" Jun 25 18:48:03.874491 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3360435364.mount: Deactivated successfully. Jun 25 18:48:03.882176 containerd[1570]: time="2024-06-25T18:48:03.882101144Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jun 25 18:48:03.883212 containerd[1570]: time="2024-06-25T18:48:03.883186830Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jun 25 18:48:03.883963 containerd[1570]: time="2024-06-25T18:48:03.883920005Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jun 25 18:48:03.884845 containerd[1570]: time="2024-06-25T18:48:03.884800656Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jun 25 18:48:03.885769 containerd[1570]: time="2024-06-25T18:48:03.885728536Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jun 25 18:48:03.886746 containerd[1570]: time="2024-06-25T18:48:03.886712661Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=312056" Jun 25 18:48:03.887754 containerd[1570]: time="2024-06-25T18:48:03.887705613Z" level=info msg="ImageCreate event name:\"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jun 25 18:48:03.892361 containerd[1570]: time="2024-06-25T18:48:03.892329312Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jun 25 18:48:03.893403 containerd[1570]: time="2024-06-25T18:48:03.893362830Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 728.019125ms" Jun 25 18:48:03.895032 containerd[1570]: time="2024-06-25T18:48:03.895005620Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 731.17912ms" Jun 25 18:48:03.896613 containerd[1570]: time="2024-06-25T18:48:03.896569843Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4873874c08efc72e9729683a83ffbb7502ee729e9a5ac097723806ea7fa13517\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"311286\" in 730.142166ms" Jun 25 18:48:04.044924 containerd[1570]: time="2024-06-25T18:48:04.043416891Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:04.044924 containerd[1570]: time="2024-06-25T18:48:04.044146990Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:04.044924 containerd[1570]: time="2024-06-25T18:48:04.044192495Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:04.044924 containerd[1570]: time="2024-06-25T18:48:04.044220909Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:04.046126 containerd[1570]: time="2024-06-25T18:48:04.045995907Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:04.046126 containerd[1570]: time="2024-06-25T18:48:04.046070507Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:04.046126 containerd[1570]: time="2024-06-25T18:48:04.046097117Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:04.046555 containerd[1570]: time="2024-06-25T18:48:04.046365550Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:04.048080 containerd[1570]: time="2024-06-25T18:48:04.047749274Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:04.048381 containerd[1570]: time="2024-06-25T18:48:04.048327448Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:04.048556 containerd[1570]: time="2024-06-25T18:48:04.048393412Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:04.048556 containerd[1570]: time="2024-06-25T18:48:04.048414101Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:04.113267 containerd[1570]: time="2024-06-25T18:48:04.113128891Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-localhost,Uid:9c3207d669e00aa24ded52617c0d65d0,Namespace:kube-system,Attempt:0,} returns sandbox id \"d0894ed96a73d1c84511e69f62f889ba5e5b3d991363fe96224a587ce7055a2f\"" Jun 25 18:48:04.114383 kubelet[2316]: E0625 18:48:04.114310 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:04.115867 containerd[1570]: time="2024-06-25T18:48:04.115825517Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-localhost,Uid:d27baad490d2d4f748c86b318d7d74ef,Namespace:kube-system,Attempt:0,} returns sandbox id \"00bbba156d0c21336de6ac60bd2b163aa142cf4fab296a5226abea9b70711e1a\"" Jun 25 18:48:04.116967 containerd[1570]: time="2024-06-25T18:48:04.116943594Z" level=info msg="CreateContainer within sandbox \"d0894ed96a73d1c84511e69f62f889ba5e5b3d991363fe96224a587ce7055a2f\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jun 25 18:48:04.118535 kubelet[2316]: E0625 18:48:04.118508 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:04.118929 containerd[1570]: time="2024-06-25T18:48:04.118852793Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-localhost,Uid:90568ebd3309c6cbd50a67732ab2ba14,Namespace:kube-system,Attempt:0,} returns sandbox id \"fc717334f6c5879e2a6fcf6bc8d62321da70827ec3959068d68d2f6e5359b261\"" Jun 25 18:48:04.119572 kubelet[2316]: E0625 18:48:04.119545 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:04.120516 containerd[1570]: time="2024-06-25T18:48:04.120498038Z" level=info msg="CreateContainer within sandbox \"00bbba156d0c21336de6ac60bd2b163aa142cf4fab296a5226abea9b70711e1a\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jun 25 18:48:04.121751 containerd[1570]: time="2024-06-25T18:48:04.121729557Z" level=info msg="CreateContainer within sandbox \"fc717334f6c5879e2a6fcf6bc8d62321da70827ec3959068d68d2f6e5359b261\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jun 25 18:48:04.146306 kubelet[2316]: E0625 18:48:04.146192 2316 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.0.0.150:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:04.674179 containerd[1570]: time="2024-06-25T18:48:04.674102316Z" level=info msg="CreateContainer within sandbox \"00bbba156d0c21336de6ac60bd2b163aa142cf4fab296a5226abea9b70711e1a\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"274a494904bd01b9750aecde73304077534b79558e74eec25fa26cb47561d256\"" Jun 25 18:48:04.674791 containerd[1570]: time="2024-06-25T18:48:04.674765910Z" level=info msg="StartContainer for \"274a494904bd01b9750aecde73304077534b79558e74eec25fa26cb47561d256\"" Jun 25 18:48:04.679994 containerd[1570]: time="2024-06-25T18:48:04.679889417Z" level=info msg="CreateContainer within sandbox \"d0894ed96a73d1c84511e69f62f889ba5e5b3d991363fe96224a587ce7055a2f\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"eb8e9525bc5cf8880251190c5106670f39abee8f6298cf2fc82e1fc598f87cad\"" Jun 25 18:48:04.680352 containerd[1570]: time="2024-06-25T18:48:04.680321918Z" level=info msg="StartContainer for \"eb8e9525bc5cf8880251190c5106670f39abee8f6298cf2fc82e1fc598f87cad\"" Jun 25 18:48:04.682451 containerd[1570]: time="2024-06-25T18:48:04.682414882Z" level=info msg="CreateContainer within sandbox \"fc717334f6c5879e2a6fcf6bc8d62321da70827ec3959068d68d2f6e5359b261\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"7f412bdf7a18aa746750d9ae913012e8f0b6211eb6ea0b7d904cbfd6fb2365d5\"" Jun 25 18:48:04.683626 containerd[1570]: time="2024-06-25T18:48:04.683342031Z" level=info msg="StartContainer for \"7f412bdf7a18aa746750d9ae913012e8f0b6211eb6ea0b7d904cbfd6fb2365d5\"" Jun 25 18:48:04.762608 containerd[1570]: time="2024-06-25T18:48:04.762310637Z" level=info msg="StartContainer for \"eb8e9525bc5cf8880251190c5106670f39abee8f6298cf2fc82e1fc598f87cad\" returns successfully" Jun 25 18:48:04.762608 containerd[1570]: time="2024-06-25T18:48:04.762472019Z" level=info msg="StartContainer for \"7f412bdf7a18aa746750d9ae913012e8f0b6211eb6ea0b7d904cbfd6fb2365d5\" returns successfully" Jun 25 18:48:04.762770 containerd[1570]: time="2024-06-25T18:48:04.762757114Z" level=info msg="StartContainer for \"274a494904bd01b9750aecde73304077534b79558e74eec25fa26cb47561d256\" returns successfully" Jun 25 18:48:04.797325 kubelet[2316]: W0625 18:48:04.796098 2316 reflector.go:535] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://10.0.0.150:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:04.797325 kubelet[2316]: E0625 18:48:04.796138 2316 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.0.0.150:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.0.0.150:6443: connect: connection refused Jun 25 18:48:05.173977 kubelet[2316]: E0625 18:48:05.173937 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:05.179199 kubelet[2316]: E0625 18:48:05.178357 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:05.184020 kubelet[2316]: E0625 18:48:05.183992 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:05.255201 kubelet[2316]: I0625 18:48:05.250930 2316 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Jun 25 18:48:05.911846 kubelet[2316]: E0625 18:48:05.911809 2316 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"localhost\" not found" node="localhost" Jun 25 18:48:06.009658 kubelet[2316]: I0625 18:48:06.008650 2316 kubelet_node_status.go:73] "Successfully registered node" node="localhost" Jun 25 18:48:06.125494 kubelet[2316]: I0625 18:48:06.125451 2316 apiserver.go:52] "Watching apiserver" Jun 25 18:48:06.136774 kubelet[2316]: I0625 18:48:06.136746 2316 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Jun 25 18:48:06.189319 kubelet[2316]: E0625 18:48:06.189207 2316 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-localhost" Jun 25 18:48:06.189319 kubelet[2316]: E0625 18:48:06.189235 2316 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:06.189736 kubelet[2316]: E0625 18:48:06.189488 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:06.189736 kubelet[2316]: E0625 18:48:06.189207 2316 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-apiserver-localhost" Jun 25 18:48:06.189736 kubelet[2316]: E0625 18:48:06.189553 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:06.189983 kubelet[2316]: E0625 18:48:06.189970 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:07.191498 kubelet[2316]: E0625 18:48:07.191468 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:07.191498 kubelet[2316]: E0625 18:48:07.191467 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:08.187135 kubelet[2316]: E0625 18:48:08.187098 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:08.187359 kubelet[2316]: E0625 18:48:08.187333 2316 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:08.526064 systemd[1]: Reloading requested from client PID 2591 ('systemctl') (unit session-7.scope)... Jun 25 18:48:08.526082 systemd[1]: Reloading... Jun 25 18:48:08.598192 zram_generator::config[2628]: No configuration found. Jun 25 18:48:09.174975 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jun 25 18:48:09.262656 systemd[1]: Reloading finished in 736 ms. Jun 25 18:48:09.301822 kubelet[2316]: I0625 18:48:09.301745 2316 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jun 25 18:48:09.301818 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jun 25 18:48:09.323853 systemd[1]: kubelet.service: Deactivated successfully. Jun 25 18:48:09.324391 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jun 25 18:48:09.332376 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jun 25 18:48:09.470829 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jun 25 18:48:09.476585 (kubelet)[2683]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jun 25 18:48:09.523343 kubelet[2683]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jun 25 18:48:09.524274 kubelet[2683]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jun 25 18:48:09.524274 kubelet[2683]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jun 25 18:48:09.524274 kubelet[2683]: I0625 18:48:09.523750 2683 server.go:203] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jun 25 18:48:09.529294 kubelet[2683]: I0625 18:48:09.529261 2683 server.go:467] "Kubelet version" kubeletVersion="v1.28.7" Jun 25 18:48:09.529294 kubelet[2683]: I0625 18:48:09.529282 2683 server.go:469] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jun 25 18:48:09.529427 kubelet[2683]: I0625 18:48:09.529409 2683 server.go:895] "Client rotation is on, will bootstrap in background" Jun 25 18:48:09.530652 kubelet[2683]: I0625 18:48:09.530624 2683 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jun 25 18:48:09.531660 kubelet[2683]: I0625 18:48:09.531626 2683 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jun 25 18:48:09.542017 kubelet[2683]: I0625 18:48:09.541986 2683 server.go:725] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jun 25 18:48:09.542590 kubelet[2683]: I0625 18:48:09.542548 2683 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jun 25 18:48:09.542807 kubelet[2683]: I0625 18:48:09.542751 2683 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"cgroupfs","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jun 25 18:48:09.542807 kubelet[2683]: I0625 18:48:09.542783 2683 topology_manager.go:138] "Creating topology manager with none policy" Jun 25 18:48:09.542807 kubelet[2683]: I0625 18:48:09.542793 2683 container_manager_linux.go:301] "Creating device plugin manager" Jun 25 18:48:09.542995 kubelet[2683]: I0625 18:48:09.542834 2683 state_mem.go:36] "Initialized new in-memory state store" Jun 25 18:48:09.542995 kubelet[2683]: I0625 18:48:09.542935 2683 kubelet.go:393] "Attempting to sync node with API server" Jun 25 18:48:09.542995 kubelet[2683]: I0625 18:48:09.542947 2683 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests" Jun 25 18:48:09.542995 kubelet[2683]: I0625 18:48:09.542975 2683 kubelet.go:309] "Adding apiserver pod source" Jun 25 18:48:09.542995 kubelet[2683]: I0625 18:48:09.542993 2683 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jun 25 18:48:09.545028 kubelet[2683]: I0625 18:48:09.544955 2683 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="v1.7.18" apiVersion="v1" Jun 25 18:48:09.545917 kubelet[2683]: I0625 18:48:09.545899 2683 server.go:1232] "Started kubelet" Jun 25 18:48:09.549181 kubelet[2683]: I0625 18:48:09.548469 2683 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jun 25 18:48:09.549554 kubelet[2683]: E0625 18:48:09.549532 2683 cri_stats_provider.go:448] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Jun 25 18:48:09.549664 kubelet[2683]: E0625 18:48:09.549645 2683 kubelet.go:1431] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jun 25 18:48:09.555035 kubelet[2683]: I0625 18:48:09.554403 2683 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Jun 25 18:48:09.556901 kubelet[2683]: I0625 18:48:09.555471 2683 server.go:462] "Adding debug handlers to kubelet server" Jun 25 18:48:09.557782 kubelet[2683]: I0625 18:48:09.557259 2683 volume_manager.go:291] "Starting Kubelet Volume Manager" Jun 25 18:48:09.558265 kubelet[2683]: I0625 18:48:09.558248 2683 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Jun 25 18:48:09.558349 kubelet[2683]: I0625 18:48:09.557732 2683 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10 Jun 25 18:48:09.558656 kubelet[2683]: I0625 18:48:09.558637 2683 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jun 25 18:48:09.558926 kubelet[2683]: I0625 18:48:09.558896 2683 reconciler_new.go:29] "Reconciler: start to sync state" Jun 25 18:48:09.567352 kubelet[2683]: I0625 18:48:09.566806 2683 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jun 25 18:48:09.568711 kubelet[2683]: I0625 18:48:09.568256 2683 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jun 25 18:48:09.568711 kubelet[2683]: I0625 18:48:09.568275 2683 status_manager.go:217] "Starting to sync pod status with apiserver" Jun 25 18:48:09.568711 kubelet[2683]: I0625 18:48:09.568292 2683 kubelet.go:2303] "Starting kubelet main sync loop" Jun 25 18:48:09.568711 kubelet[2683]: E0625 18:48:09.568625 2683 kubelet.go:2327] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jun 25 18:48:09.646909 kubelet[2683]: I0625 18:48:09.646626 2683 cpu_manager.go:214] "Starting CPU manager" policy="none" Jun 25 18:48:09.646909 kubelet[2683]: I0625 18:48:09.646648 2683 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jun 25 18:48:09.646909 kubelet[2683]: I0625 18:48:09.646663 2683 state_mem.go:36] "Initialized new in-memory state store" Jun 25 18:48:09.646909 kubelet[2683]: I0625 18:48:09.646802 2683 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jun 25 18:48:09.646909 kubelet[2683]: I0625 18:48:09.646820 2683 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jun 25 18:48:09.646909 kubelet[2683]: I0625 18:48:09.646826 2683 policy_none.go:49] "None policy: Start" Jun 25 18:48:09.647466 kubelet[2683]: I0625 18:48:09.647442 2683 memory_manager.go:169] "Starting memorymanager" policy="None" Jun 25 18:48:09.647466 kubelet[2683]: I0625 18:48:09.647465 2683 state_mem.go:35] "Initializing new in-memory state store" Jun 25 18:48:09.647683 kubelet[2683]: I0625 18:48:09.647613 2683 state_mem.go:75] "Updated machine memory state" Jun 25 18:48:09.649141 kubelet[2683]: I0625 18:48:09.649117 2683 manager.go:471] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jun 25 18:48:09.651097 kubelet[2683]: I0625 18:48:09.650967 2683 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jun 25 18:48:09.661738 kubelet[2683]: I0625 18:48:09.661513 2683 kubelet_node_status.go:70] "Attempting to register node" node="localhost" Jun 25 18:48:09.668779 kubelet[2683]: I0625 18:48:09.668744 2683 topology_manager.go:215] "Topology Admit Handler" podUID="d27baad490d2d4f748c86b318d7d74ef" podNamespace="kube-system" podName="kube-controller-manager-localhost" Jun 25 18:48:09.669044 kubelet[2683]: I0625 18:48:09.668953 2683 topology_manager.go:215] "Topology Admit Handler" podUID="9c3207d669e00aa24ded52617c0d65d0" podNamespace="kube-system" podName="kube-scheduler-localhost" Jun 25 18:48:09.669110 kubelet[2683]: I0625 18:48:09.669091 2683 topology_manager.go:215] "Topology Admit Handler" podUID="90568ebd3309c6cbd50a67732ab2ba14" podNamespace="kube-system" podName="kube-apiserver-localhost" Jun 25 18:48:09.753668 kubelet[2683]: E0625 18:48:09.753541 2683 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jun 25 18:48:09.753668 kubelet[2683]: E0625 18:48:09.753646 2683 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-scheduler-localhost\" already exists" pod="kube-system/kube-scheduler-localhost" Jun 25 18:48:09.755908 kubelet[2683]: I0625 18:48:09.755703 2683 kubelet_node_status.go:108] "Node was previously registered" node="localhost" Jun 25 18:48:09.755908 kubelet[2683]: I0625 18:48:09.755807 2683 kubelet_node_status.go:73] "Successfully registered node" node="localhost" Jun 25 18:48:09.759266 kubelet[2683]: I0625 18:48:09.759229 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/90568ebd3309c6cbd50a67732ab2ba14-ca-certs\") pod \"kube-apiserver-localhost\" (UID: \"90568ebd3309c6cbd50a67732ab2ba14\") " pod="kube-system/kube-apiserver-localhost" Jun 25 18:48:09.759330 kubelet[2683]: I0625 18:48:09.759281 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/90568ebd3309c6cbd50a67732ab2ba14-k8s-certs\") pod \"kube-apiserver-localhost\" (UID: \"90568ebd3309c6cbd50a67732ab2ba14\") " pod="kube-system/kube-apiserver-localhost" Jun 25 18:48:09.759330 kubelet[2683]: I0625 18:48:09.759318 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/90568ebd3309c6cbd50a67732ab2ba14-usr-share-ca-certificates\") pod \"kube-apiserver-localhost\" (UID: \"90568ebd3309c6cbd50a67732ab2ba14\") " pod="kube-system/kube-apiserver-localhost" Jun 25 18:48:09.759374 kubelet[2683]: I0625 18:48:09.759345 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-ca-certs\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:09.759397 kubelet[2683]: I0625 18:48:09.759374 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/9c3207d669e00aa24ded52617c0d65d0-kubeconfig\") pod \"kube-scheduler-localhost\" (UID: \"9c3207d669e00aa24ded52617c0d65d0\") " pod="kube-system/kube-scheduler-localhost" Jun 25 18:48:09.759489 kubelet[2683]: I0625 18:48:09.759402 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-flexvolume-dir\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:09.759489 kubelet[2683]: I0625 18:48:09.759428 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-k8s-certs\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:09.759489 kubelet[2683]: I0625 18:48:09.759455 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-kubeconfig\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:09.759489 kubelet[2683]: I0625 18:48:09.759482 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/d27baad490d2d4f748c86b318d7d74ef-usr-share-ca-certificates\") pod \"kube-controller-manager-localhost\" (UID: \"d27baad490d2d4f748c86b318d7d74ef\") " pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:10.038787 kubelet[2683]: E0625 18:48:10.038667 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:10.054936 kubelet[2683]: E0625 18:48:10.054723 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:10.054936 kubelet[2683]: E0625 18:48:10.054870 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:10.544531 kubelet[2683]: I0625 18:48:10.544225 2683 apiserver.go:52] "Watching apiserver" Jun 25 18:48:10.558556 kubelet[2683]: I0625 18:48:10.558519 2683 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Jun 25 18:48:10.588519 kubelet[2683]: E0625 18:48:10.588469 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:10.612426 kubelet[2683]: E0625 18:48:10.612380 2683 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-apiserver-localhost\" already exists" pod="kube-system/kube-apiserver-localhost" Jun 25 18:48:10.612900 kubelet[2683]: E0625 18:48:10.612878 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:10.614037 kubelet[2683]: E0625 18:48:10.613218 2683 kubelet.go:1890] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-localhost\" already exists" pod="kube-system/kube-controller-manager-localhost" Jun 25 18:48:10.614037 kubelet[2683]: E0625 18:48:10.613580 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:10.646170 kubelet[2683]: I0625 18:48:10.644260 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-localhost" podStartSLOduration=3.641728887 podCreationTimestamp="2024-06-25 18:48:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:48:10.641481212 +0000 UTC m=+1.160178835" watchObservedRunningTime="2024-06-25 18:48:10.641728887 +0000 UTC m=+1.160426510" Jun 25 18:48:10.666636 kubelet[2683]: I0625 18:48:10.666454 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-localhost" podStartSLOduration=1.666416732 podCreationTimestamp="2024-06-25 18:48:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:48:10.656360855 +0000 UTC m=+1.175058478" watchObservedRunningTime="2024-06-25 18:48:10.666416732 +0000 UTC m=+1.185114355" Jun 25 18:48:10.666636 kubelet[2683]: I0625 18:48:10.666526 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-localhost" podStartSLOduration=3.666514159 podCreationTimestamp="2024-06-25 18:48:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:48:10.665595045 +0000 UTC m=+1.184292658" watchObservedRunningTime="2024-06-25 18:48:10.666514159 +0000 UTC m=+1.185211782" Jun 25 18:48:11.591347 kubelet[2683]: E0625 18:48:11.591302 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:11.594208 kubelet[2683]: E0625 18:48:11.594181 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:12.984635 kubelet[2683]: E0625 18:48:12.984597 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:14.622952 sudo[1779]: pam_unix(sudo:session): session closed for user root Jun 25 18:48:14.624774 sshd[1772]: pam_unix(sshd:session): session closed for user core Jun 25 18:48:14.628999 systemd[1]: sshd@6-10.0.0.150:22-10.0.0.1:60428.service: Deactivated successfully. Jun 25 18:48:14.631546 systemd-logind[1553]: Session 7 logged out. Waiting for processes to exit. Jun 25 18:48:14.631668 systemd[1]: session-7.scope: Deactivated successfully. Jun 25 18:48:14.632855 systemd-logind[1553]: Removed session 7. Jun 25 18:48:18.991954 kubelet[2683]: E0625 18:48:18.991909 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:19.407895 kubelet[2683]: E0625 18:48:19.407853 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:19.602725 kubelet[2683]: E0625 18:48:19.602689 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:19.602725 kubelet[2683]: E0625 18:48:19.602720 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:20.624286 update_engine[1556]: I0625 18:48:20.624207 1556 update_attempter.cc:509] Updating boot flags... Jun 25 18:48:20.653628 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (2779) Jun 25 18:48:20.687216 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (2781) Jun 25 18:48:20.722178 kernel: BTRFS warning: duplicate device /dev/vda3 devid 1 generation 31 scanned by (udev-worker) (2781) Jun 25 18:48:22.888717 kubelet[2683]: I0625 18:48:22.888689 2683 kuberuntime_manager.go:1528] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jun 25 18:48:22.889325 containerd[1570]: time="2024-06-25T18:48:22.889135291Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jun 25 18:48:22.889628 kubelet[2683]: I0625 18:48:22.889356 2683 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jun 25 18:48:22.990811 kubelet[2683]: E0625 18:48:22.990724 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:23.710742 kubelet[2683]: I0625 18:48:23.710706 2683 topology_manager.go:215] "Topology Admit Handler" podUID="ad8dc28a-5c29-4c7b-bb23-30c98333374d" podNamespace="kube-system" podName="kube-proxy-xzzz9" Jun 25 18:48:23.745213 kubelet[2683]: I0625 18:48:23.742216 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/ad8dc28a-5c29-4c7b-bb23-30c98333374d-kube-proxy\") pod \"kube-proxy-xzzz9\" (UID: \"ad8dc28a-5c29-4c7b-bb23-30c98333374d\") " pod="kube-system/kube-proxy-xzzz9" Jun 25 18:48:23.745213 kubelet[2683]: I0625 18:48:23.742269 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khj9w\" (UniqueName: \"kubernetes.io/projected/ad8dc28a-5c29-4c7b-bb23-30c98333374d-kube-api-access-khj9w\") pod \"kube-proxy-xzzz9\" (UID: \"ad8dc28a-5c29-4c7b-bb23-30c98333374d\") " pod="kube-system/kube-proxy-xzzz9" Jun 25 18:48:23.745213 kubelet[2683]: I0625 18:48:23.742289 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/ad8dc28a-5c29-4c7b-bb23-30c98333374d-xtables-lock\") pod \"kube-proxy-xzzz9\" (UID: \"ad8dc28a-5c29-4c7b-bb23-30c98333374d\") " pod="kube-system/kube-proxy-xzzz9" Jun 25 18:48:23.745213 kubelet[2683]: I0625 18:48:23.742315 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ad8dc28a-5c29-4c7b-bb23-30c98333374d-lib-modules\") pod \"kube-proxy-xzzz9\" (UID: \"ad8dc28a-5c29-4c7b-bb23-30c98333374d\") " pod="kube-system/kube-proxy-xzzz9" Jun 25 18:48:23.831174 kubelet[2683]: I0625 18:48:23.831105 2683 topology_manager.go:215] "Topology Admit Handler" podUID="bae3c849-5dca-4f3a-9e6a-e445e0c1bbe7" podNamespace="tigera-operator" podName="tigera-operator-76c4974c85-2wbjc" Jun 25 18:48:23.944106 kubelet[2683]: I0625 18:48:23.944047 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/bae3c849-5dca-4f3a-9e6a-e445e0c1bbe7-var-lib-calico\") pod \"tigera-operator-76c4974c85-2wbjc\" (UID: \"bae3c849-5dca-4f3a-9e6a-e445e0c1bbe7\") " pod="tigera-operator/tigera-operator-76c4974c85-2wbjc" Jun 25 18:48:23.944106 kubelet[2683]: I0625 18:48:23.944099 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb87p\" (UniqueName: \"kubernetes.io/projected/bae3c849-5dca-4f3a-9e6a-e445e0c1bbe7-kube-api-access-zb87p\") pod \"tigera-operator-76c4974c85-2wbjc\" (UID: \"bae3c849-5dca-4f3a-9e6a-e445e0c1bbe7\") " pod="tigera-operator/tigera-operator-76c4974c85-2wbjc" Jun 25 18:48:24.014507 kubelet[2683]: E0625 18:48:24.014380 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:24.015128 containerd[1570]: time="2024-06-25T18:48:24.015065985Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xzzz9,Uid:ad8dc28a-5c29-4c7b-bb23-30c98333374d,Namespace:kube-system,Attempt:0,}" Jun 25 18:48:24.065613 containerd[1570]: time="2024-06-25T18:48:24.065510457Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:24.065613 containerd[1570]: time="2024-06-25T18:48:24.065577173Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:24.065811 containerd[1570]: time="2024-06-25T18:48:24.065636125Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:24.065811 containerd[1570]: time="2024-06-25T18:48:24.065732738Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:24.102657 containerd[1570]: time="2024-06-25T18:48:24.102604209Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-xzzz9,Uid:ad8dc28a-5c29-4c7b-bb23-30c98333374d,Namespace:kube-system,Attempt:0,} returns sandbox id \"19c1a500b71f0d9f3b513d10e362fe75217c67385e416032c3f2fe9190e0b1e4\"" Jun 25 18:48:24.103413 kubelet[2683]: E0625 18:48:24.103385 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:24.105802 containerd[1570]: time="2024-06-25T18:48:24.105739273Z" level=info msg="CreateContainer within sandbox \"19c1a500b71f0d9f3b513d10e362fe75217c67385e416032c3f2fe9190e0b1e4\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jun 25 18:48:24.123575 containerd[1570]: time="2024-06-25T18:48:24.123513686Z" level=info msg="CreateContainer within sandbox \"19c1a500b71f0d9f3b513d10e362fe75217c67385e416032c3f2fe9190e0b1e4\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"03ea60d7d106de47027a02cc154c5142629948c1048e7fabacc0f3dcdc8514b2\"" Jun 25 18:48:24.123929 containerd[1570]: time="2024-06-25T18:48:24.123894166Z" level=info msg="StartContainer for \"03ea60d7d106de47027a02cc154c5142629948c1048e7fabacc0f3dcdc8514b2\"" Jun 25 18:48:24.136175 containerd[1570]: time="2024-06-25T18:48:24.135667842Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-76c4974c85-2wbjc,Uid:bae3c849-5dca-4f3a-9e6a-e445e0c1bbe7,Namespace:tigera-operator,Attempt:0,}" Jun 25 18:48:24.160797 containerd[1570]: time="2024-06-25T18:48:24.160650810Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:24.161117 containerd[1570]: time="2024-06-25T18:48:24.160845869Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:24.161117 containerd[1570]: time="2024-06-25T18:48:24.160905241Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:24.161117 containerd[1570]: time="2024-06-25T18:48:24.160938695Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:24.187030 containerd[1570]: time="2024-06-25T18:48:24.186971299Z" level=info msg="StartContainer for \"03ea60d7d106de47027a02cc154c5142629948c1048e7fabacc0f3dcdc8514b2\" returns successfully" Jun 25 18:48:24.215525 containerd[1570]: time="2024-06-25T18:48:24.215478959Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-76c4974c85-2wbjc,Uid:bae3c849-5dca-4f3a-9e6a-e445e0c1bbe7,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"bb1fe274ca7c9d3101c4ffe04840b24e75978861f1e899ab522f4080528b42de\"" Jun 25 18:48:24.221380 containerd[1570]: time="2024-06-25T18:48:24.220409401Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.34.0\"" Jun 25 18:48:24.610950 kubelet[2683]: E0625 18:48:24.610922 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:25.737872 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2470868895.mount: Deactivated successfully. Jun 25 18:48:26.192495 containerd[1570]: time="2024-06-25T18:48:26.192423242Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.34.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:26.193412 containerd[1570]: time="2024-06-25T18:48:26.193379580Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.34.0: active requests=0, bytes read=22076096" Jun 25 18:48:26.194864 containerd[1570]: time="2024-06-25T18:48:26.194750962Z" level=info msg="ImageCreate event name:\"sha256:01249e32d0f6f7d0ad79761d634d16738f1a5792b893f202f9a417c63034411d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:26.203202 containerd[1570]: time="2024-06-25T18:48:26.203123683Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:479ddc7ff9ab095058b96f6710bbf070abada86332e267d6e5dcc1df36ba2cc5\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:26.203914 containerd[1570]: time="2024-06-25T18:48:26.203871596Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.34.0\" with image id \"sha256:01249e32d0f6f7d0ad79761d634d16738f1a5792b893f202f9a417c63034411d\", repo tag \"quay.io/tigera/operator:v1.34.0\", repo digest \"quay.io/tigera/operator@sha256:479ddc7ff9ab095058b96f6710bbf070abada86332e267d6e5dcc1df36ba2cc5\", size \"22070263\" in 1.98342794s" Jun 25 18:48:26.203955 containerd[1570]: time="2024-06-25T18:48:26.203913666Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.34.0\" returns image reference \"sha256:01249e32d0f6f7d0ad79761d634d16738f1a5792b893f202f9a417c63034411d\"" Jun 25 18:48:26.205700 containerd[1570]: time="2024-06-25T18:48:26.205671359Z" level=info msg="CreateContainer within sandbox \"bb1fe274ca7c9d3101c4ffe04840b24e75978861f1e899ab522f4080528b42de\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jun 25 18:48:26.242411 containerd[1570]: time="2024-06-25T18:48:26.242340628Z" level=info msg="CreateContainer within sandbox \"bb1fe274ca7c9d3101c4ffe04840b24e75978861f1e899ab522f4080528b42de\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"ad9fdff8668e28c1db2158af6ea707f8cc9e07ab01086c222df6f38f36ce1dad\"" Jun 25 18:48:26.243301 containerd[1570]: time="2024-06-25T18:48:26.243249647Z" level=info msg="StartContainer for \"ad9fdff8668e28c1db2158af6ea707f8cc9e07ab01086c222df6f38f36ce1dad\"" Jun 25 18:48:26.412636 containerd[1570]: time="2024-06-25T18:48:26.412560765Z" level=info msg="StartContainer for \"ad9fdff8668e28c1db2158af6ea707f8cc9e07ab01086c222df6f38f36ce1dad\" returns successfully" Jun 25 18:48:26.621547 kubelet[2683]: I0625 18:48:26.621466 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="tigera-operator/tigera-operator-76c4974c85-2wbjc" podStartSLOduration=1.634106902 podCreationTimestamp="2024-06-25 18:48:23 +0000 UTC" firstStartedPulling="2024-06-25 18:48:24.216904628 +0000 UTC m=+14.735602241" lastFinishedPulling="2024-06-25 18:48:26.204229653 +0000 UTC m=+16.722927276" observedRunningTime="2024-06-25 18:48:26.621339793 +0000 UTC m=+17.140037416" watchObservedRunningTime="2024-06-25 18:48:26.621431937 +0000 UTC m=+17.140129550" Jun 25 18:48:26.622341 kubelet[2683]: I0625 18:48:26.621660 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-xzzz9" podStartSLOduration=3.621643318 podCreationTimestamp="2024-06-25 18:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:48:24.634326618 +0000 UTC m=+15.153024251" watchObservedRunningTime="2024-06-25 18:48:26.621643318 +0000 UTC m=+17.140340941" Jun 25 18:48:30.545929 kubelet[2683]: I0625 18:48:30.542983 2683 topology_manager.go:215] "Topology Admit Handler" podUID="884b67b7-6945-4354-b19d-c5c37e2e8b68" podNamespace="calico-system" podName="calico-typha-7b5c7c7579-44nrd" Jun 25 18:48:30.597513 kubelet[2683]: I0625 18:48:30.597471 2683 topology_manager.go:215] "Topology Admit Handler" podUID="04f13a90-a54c-4fda-8fb1-db3288d0c52e" podNamespace="calico-system" podName="calico-node-hbcjh" Jun 25 18:48:30.687363 kubelet[2683]: I0625 18:48:30.687308 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/884b67b7-6945-4354-b19d-c5c37e2e8b68-typha-certs\") pod \"calico-typha-7b5c7c7579-44nrd\" (UID: \"884b67b7-6945-4354-b19d-c5c37e2e8b68\") " pod="calico-system/calico-typha-7b5c7c7579-44nrd" Jun 25 18:48:30.687363 kubelet[2683]: I0625 18:48:30.687360 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-lib-modules\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687580 kubelet[2683]: I0625 18:48:30.687387 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-xtables-lock\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687580 kubelet[2683]: I0625 18:48:30.687434 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04f13a90-a54c-4fda-8fb1-db3288d0c52e-tigera-ca-bundle\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687580 kubelet[2683]: I0625 18:48:30.687467 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-flexvol-driver-host\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687688 kubelet[2683]: I0625 18:48:30.687584 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-net-dir\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687688 kubelet[2683]: I0625 18:48:30.687635 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-var-run-calico\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687688 kubelet[2683]: I0625 18:48:30.687659 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/884b67b7-6945-4354-b19d-c5c37e2e8b68-tigera-ca-bundle\") pod \"calico-typha-7b5c7c7579-44nrd\" (UID: \"884b67b7-6945-4354-b19d-c5c37e2e8b68\") " pod="calico-system/calico-typha-7b5c7c7579-44nrd" Jun 25 18:48:30.687801 kubelet[2683]: I0625 18:48:30.687717 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdpgw\" (UniqueName: \"kubernetes.io/projected/884b67b7-6945-4354-b19d-c5c37e2e8b68-kube-api-access-xdpgw\") pod \"calico-typha-7b5c7c7579-44nrd\" (UID: \"884b67b7-6945-4354-b19d-c5c37e2e8b68\") " pod="calico-system/calico-typha-7b5c7c7579-44nrd" Jun 25 18:48:30.687801 kubelet[2683]: I0625 18:48:30.687757 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/04f13a90-a54c-4fda-8fb1-db3288d0c52e-node-certs\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687801 kubelet[2683]: I0625 18:48:30.687795 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-policysync\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687908 kubelet[2683]: I0625 18:48:30.687860 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-bin-dir\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687908 kubelet[2683]: I0625 18:48:30.687895 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw565\" (UniqueName: \"kubernetes.io/projected/04f13a90-a54c-4fda-8fb1-db3288d0c52e-kube-api-access-lw565\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687985 kubelet[2683]: I0625 18:48:30.687922 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-var-lib-calico\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.687985 kubelet[2683]: I0625 18:48:30.687955 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-log-dir\") pod \"calico-node-hbcjh\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " pod="calico-system/calico-node-hbcjh" Jun 25 18:48:30.702880 kubelet[2683]: I0625 18:48:30.702845 2683 topology_manager.go:215] "Topology Admit Handler" podUID="230b157d-8713-4407-885c-4ac300263b09" podNamespace="calico-system" podName="csi-node-driver-bbct9" Jun 25 18:48:30.703140 kubelet[2683]: E0625 18:48:30.703120 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-bbct9" podUID="230b157d-8713-4407-885c-4ac300263b09" Jun 25 18:48:30.790268 kubelet[2683]: I0625 18:48:30.789253 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/230b157d-8713-4407-885c-4ac300263b09-varrun\") pod \"csi-node-driver-bbct9\" (UID: \"230b157d-8713-4407-885c-4ac300263b09\") " pod="calico-system/csi-node-driver-bbct9" Jun 25 18:48:30.790268 kubelet[2683]: I0625 18:48:30.789308 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/230b157d-8713-4407-885c-4ac300263b09-socket-dir\") pod \"csi-node-driver-bbct9\" (UID: \"230b157d-8713-4407-885c-4ac300263b09\") " pod="calico-system/csi-node-driver-bbct9" Jun 25 18:48:30.790268 kubelet[2683]: I0625 18:48:30.789587 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/230b157d-8713-4407-885c-4ac300263b09-kubelet-dir\") pod \"csi-node-driver-bbct9\" (UID: \"230b157d-8713-4407-885c-4ac300263b09\") " pod="calico-system/csi-node-driver-bbct9" Jun 25 18:48:30.790268 kubelet[2683]: I0625 18:48:30.789703 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/230b157d-8713-4407-885c-4ac300263b09-registration-dir\") pod \"csi-node-driver-bbct9\" (UID: \"230b157d-8713-4407-885c-4ac300263b09\") " pod="calico-system/csi-node-driver-bbct9" Jun 25 18:48:30.790268 kubelet[2683]: I0625 18:48:30.789743 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dsfbs\" (UniqueName: \"kubernetes.io/projected/230b157d-8713-4407-885c-4ac300263b09-kube-api-access-dsfbs\") pod \"csi-node-driver-bbct9\" (UID: \"230b157d-8713-4407-885c-4ac300263b09\") " pod="calico-system/csi-node-driver-bbct9" Jun 25 18:48:30.806800 kubelet[2683]: E0625 18:48:30.806753 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.806800 kubelet[2683]: W0625 18:48:30.806790 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.807113 kubelet[2683]: E0625 18:48:30.806826 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.807221 kubelet[2683]: E0625 18:48:30.807131 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.807221 kubelet[2683]: W0625 18:48:30.807139 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.807221 kubelet[2683]: E0625 18:48:30.807201 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.853510 kubelet[2683]: E0625 18:48:30.853466 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:30.854140 containerd[1570]: time="2024-06-25T18:48:30.854101035Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7b5c7c7579-44nrd,Uid:884b67b7-6945-4354-b19d-c5c37e2e8b68,Namespace:calico-system,Attempt:0,}" Jun 25 18:48:30.891224 kubelet[2683]: E0625 18:48:30.891190 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.891224 kubelet[2683]: W0625 18:48:30.891212 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.891224 kubelet[2683]: E0625 18:48:30.891242 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.891550 kubelet[2683]: E0625 18:48:30.891535 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.891550 kubelet[2683]: W0625 18:48:30.891549 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.891631 kubelet[2683]: E0625 18:48:30.891568 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.891848 kubelet[2683]: E0625 18:48:30.891833 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.891848 kubelet[2683]: W0625 18:48:30.891845 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.891934 kubelet[2683]: E0625 18:48:30.891864 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.892163 kubelet[2683]: E0625 18:48:30.892120 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.892163 kubelet[2683]: W0625 18:48:30.892140 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.892253 kubelet[2683]: E0625 18:48:30.892203 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.892446 kubelet[2683]: E0625 18:48:30.892430 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.892446 kubelet[2683]: W0625 18:48:30.892440 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.892529 kubelet[2683]: E0625 18:48:30.892457 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.892655 kubelet[2683]: E0625 18:48:30.892641 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.892655 kubelet[2683]: W0625 18:48:30.892650 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.892720 kubelet[2683]: E0625 18:48:30.892666 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.892884 kubelet[2683]: E0625 18:48:30.892869 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.892884 kubelet[2683]: W0625 18:48:30.892881 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.892962 kubelet[2683]: E0625 18:48:30.892915 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.893116 kubelet[2683]: E0625 18:48:30.893100 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.893116 kubelet[2683]: W0625 18:48:30.893111 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.893218 kubelet[2683]: E0625 18:48:30.893136 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.893400 kubelet[2683]: E0625 18:48:30.893375 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.893400 kubelet[2683]: W0625 18:48:30.893391 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.893565 kubelet[2683]: E0625 18:48:30.893435 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.893654 kubelet[2683]: E0625 18:48:30.893632 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.893654 kubelet[2683]: W0625 18:48:30.893645 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.893729 kubelet[2683]: E0625 18:48:30.893677 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.893900 kubelet[2683]: E0625 18:48:30.893869 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.893900 kubelet[2683]: W0625 18:48:30.893886 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.893990 kubelet[2683]: E0625 18:48:30.893906 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.894133 kubelet[2683]: E0625 18:48:30.894114 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.894133 kubelet[2683]: W0625 18:48:30.894127 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.894225 kubelet[2683]: E0625 18:48:30.894145 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.894462 kubelet[2683]: E0625 18:48:30.894443 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.894462 kubelet[2683]: W0625 18:48:30.894455 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.894579 kubelet[2683]: E0625 18:48:30.894472 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.894699 kubelet[2683]: E0625 18:48:30.894680 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.894699 kubelet[2683]: W0625 18:48:30.894693 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.894775 kubelet[2683]: E0625 18:48:30.894711 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.894973 kubelet[2683]: E0625 18:48:30.894957 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.894973 kubelet[2683]: W0625 18:48:30.894969 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.895029 kubelet[2683]: E0625 18:48:30.894996 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.895189 kubelet[2683]: E0625 18:48:30.895174 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.895189 kubelet[2683]: W0625 18:48:30.895186 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.895250 kubelet[2683]: E0625 18:48:30.895213 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.895403 kubelet[2683]: E0625 18:48:30.895387 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.895403 kubelet[2683]: W0625 18:48:30.895400 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.895458 kubelet[2683]: E0625 18:48:30.895430 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.895616 kubelet[2683]: E0625 18:48:30.895601 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.895616 kubelet[2683]: W0625 18:48:30.895613 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.895696 kubelet[2683]: E0625 18:48:30.895631 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.895836 kubelet[2683]: E0625 18:48:30.895820 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.895836 kubelet[2683]: W0625 18:48:30.895832 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.895896 kubelet[2683]: E0625 18:48:30.895850 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.896046 kubelet[2683]: E0625 18:48:30.896029 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.896046 kubelet[2683]: W0625 18:48:30.896043 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.896094 kubelet[2683]: E0625 18:48:30.896068 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.896338 kubelet[2683]: E0625 18:48:30.896322 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.896338 kubelet[2683]: W0625 18:48:30.896337 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.896404 kubelet[2683]: E0625 18:48:30.896354 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.896579 kubelet[2683]: E0625 18:48:30.896564 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.896579 kubelet[2683]: W0625 18:48:30.896576 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.896624 kubelet[2683]: E0625 18:48:30.896591 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.896784 kubelet[2683]: E0625 18:48:30.896768 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.896784 kubelet[2683]: W0625 18:48:30.896779 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.896853 kubelet[2683]: E0625 18:48:30.896794 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.897070 kubelet[2683]: E0625 18:48:30.897053 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.897070 kubelet[2683]: W0625 18:48:30.897068 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.897124 kubelet[2683]: E0625 18:48:30.897086 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.897332 kubelet[2683]: E0625 18:48:30.897313 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.897332 kubelet[2683]: W0625 18:48:30.897330 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.897416 kubelet[2683]: E0625 18:48:30.897344 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:30.907495 kubelet[2683]: E0625 18:48:30.907459 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:30.908028 containerd[1570]: time="2024-06-25T18:48:30.907983847Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-hbcjh,Uid:04f13a90-a54c-4fda-8fb1-db3288d0c52e,Namespace:calico-system,Attempt:0,}" Jun 25 18:48:30.954054 kubelet[2683]: E0625 18:48:30.954026 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:30.954054 kubelet[2683]: W0625 18:48:30.954045 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:30.954054 kubelet[2683]: E0625 18:48:30.954065 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:31.152197 containerd[1570]: time="2024-06-25T18:48:31.151538984Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:31.152197 containerd[1570]: time="2024-06-25T18:48:31.151613775Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:31.152197 containerd[1570]: time="2024-06-25T18:48:31.151633883Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:31.152197 containerd[1570]: time="2024-06-25T18:48:31.151648851Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:31.159051 containerd[1570]: time="2024-06-25T18:48:31.158727975Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:31.159051 containerd[1570]: time="2024-06-25T18:48:31.158793157Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:31.159051 containerd[1570]: time="2024-06-25T18:48:31.158814659Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:31.159051 containerd[1570]: time="2024-06-25T18:48:31.158827914Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:31.256182 containerd[1570]: time="2024-06-25T18:48:31.256076216Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7b5c7c7579-44nrd,Uid:884b67b7-6945-4354-b19d-c5c37e2e8b68,Namespace:calico-system,Attempt:0,} returns sandbox id \"dd4e5e921fcc28929911f10b384a653e50aa034aeea8a3f6b191bdbbdb8b111c\"" Jun 25 18:48:31.257458 containerd[1570]: time="2024-06-25T18:48:31.257402418Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-hbcjh,Uid:04f13a90-a54c-4fda-8fb1-db3288d0c52e,Namespace:calico-system,Attempt:0,} returns sandbox id \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\"" Jun 25 18:48:31.257675 kubelet[2683]: E0625 18:48:31.257654 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:31.258501 kubelet[2683]: E0625 18:48:31.258478 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:31.260369 containerd[1570]: time="2024-06-25T18:48:31.260334349Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.28.0\"" Jun 25 18:48:32.569683 kubelet[2683]: E0625 18:48:32.569499 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-bbct9" podUID="230b157d-8713-4407-885c-4ac300263b09" Jun 25 18:48:33.083905 containerd[1570]: time="2024-06-25T18:48:33.083859076Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.28.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:33.084647 containerd[1570]: time="2024-06-25T18:48:33.084591877Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.28.0: active requests=0, bytes read=29458030" Jun 25 18:48:33.085803 containerd[1570]: time="2024-06-25T18:48:33.085773254Z" level=info msg="ImageCreate event name:\"sha256:a9372c0f51b54c589e5a16013ed3049b2a052dd6903d72603849fab2c4216fbc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:33.087762 containerd[1570]: time="2024-06-25T18:48:33.087735031Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:eff1501af12b7e27e2ef8f4e55d03d837bcb017aa5663e22e519059c452d51ed\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:33.088318 containerd[1570]: time="2024-06-25T18:48:33.088294867Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.28.0\" with image id \"sha256:a9372c0f51b54c589e5a16013ed3049b2a052dd6903d72603849fab2c4216fbc\", repo tag \"ghcr.io/flatcar/calico/typha:v3.28.0\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:eff1501af12b7e27e2ef8f4e55d03d837bcb017aa5663e22e519059c452d51ed\", size \"30905782\" in 1.82792442s" Jun 25 18:48:33.088360 containerd[1570]: time="2024-06-25T18:48:33.088324423Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.28.0\" returns image reference \"sha256:a9372c0f51b54c589e5a16013ed3049b2a052dd6903d72603849fab2c4216fbc\"" Jun 25 18:48:33.091825 containerd[1570]: time="2024-06-25T18:48:33.091784505Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.0\"" Jun 25 18:48:33.105872 containerd[1570]: time="2024-06-25T18:48:33.105804853Z" level=info msg="CreateContainer within sandbox \"dd4e5e921fcc28929911f10b384a653e50aa034aeea8a3f6b191bdbbdb8b111c\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jun 25 18:48:33.118869 containerd[1570]: time="2024-06-25T18:48:33.118821308Z" level=info msg="CreateContainer within sandbox \"dd4e5e921fcc28929911f10b384a653e50aa034aeea8a3f6b191bdbbdb8b111c\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"0a6234fc1fb7dc13c0d4eb48c4a2dc24d17ac3ebbbbb06c12580174016278e86\"" Jun 25 18:48:33.119381 containerd[1570]: time="2024-06-25T18:48:33.119354002Z" level=info msg="StartContainer for \"0a6234fc1fb7dc13c0d4eb48c4a2dc24d17ac3ebbbbb06c12580174016278e86\"" Jun 25 18:48:33.353413 containerd[1570]: time="2024-06-25T18:48:33.353258113Z" level=info msg="StartContainer for \"0a6234fc1fb7dc13c0d4eb48c4a2dc24d17ac3ebbbbb06c12580174016278e86\" returns successfully" Jun 25 18:48:33.653427 kubelet[2683]: E0625 18:48:33.653239 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:33.667376 kubelet[2683]: I0625 18:48:33.667291 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/calico-typha-7b5c7c7579-44nrd" podStartSLOduration=1.833756164 podCreationTimestamp="2024-06-25 18:48:30 +0000 UTC" firstStartedPulling="2024-06-25 18:48:31.258203739 +0000 UTC m=+21.776901372" lastFinishedPulling="2024-06-25 18:48:33.09154892 +0000 UTC m=+23.610246543" observedRunningTime="2024-06-25 18:48:33.667072692 +0000 UTC m=+24.185770315" watchObservedRunningTime="2024-06-25 18:48:33.667101335 +0000 UTC m=+24.185798958" Jun 25 18:48:33.740972 kubelet[2683]: E0625 18:48:33.740931 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.740972 kubelet[2683]: W0625 18:48:33.740956 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.740972 kubelet[2683]: E0625 18:48:33.740987 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.741307 kubelet[2683]: E0625 18:48:33.741227 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.741307 kubelet[2683]: W0625 18:48:33.741234 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.741307 kubelet[2683]: E0625 18:48:33.741247 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.741610 kubelet[2683]: E0625 18:48:33.741581 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.741670 kubelet[2683]: W0625 18:48:33.741615 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.741670 kubelet[2683]: E0625 18:48:33.741655 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.741955 kubelet[2683]: E0625 18:48:33.741942 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.741955 kubelet[2683]: W0625 18:48:33.741955 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.742022 kubelet[2683]: E0625 18:48:33.741971 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.742228 kubelet[2683]: E0625 18:48:33.742215 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.742280 kubelet[2683]: W0625 18:48:33.742228 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.742280 kubelet[2683]: E0625 18:48:33.742243 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.742483 kubelet[2683]: E0625 18:48:33.742465 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.742511 kubelet[2683]: W0625 18:48:33.742485 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.742511 kubelet[2683]: E0625 18:48:33.742497 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.742705 kubelet[2683]: E0625 18:48:33.742692 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.742737 kubelet[2683]: W0625 18:48:33.742705 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.742737 kubelet[2683]: E0625 18:48:33.742721 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.742921 kubelet[2683]: E0625 18:48:33.742909 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.742949 kubelet[2683]: W0625 18:48:33.742920 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.742949 kubelet[2683]: E0625 18:48:33.742934 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.743137 kubelet[2683]: E0625 18:48:33.743126 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.743209 kubelet[2683]: W0625 18:48:33.743137 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.743209 kubelet[2683]: E0625 18:48:33.743170 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.743376 kubelet[2683]: E0625 18:48:33.743364 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.743413 kubelet[2683]: W0625 18:48:33.743376 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.743413 kubelet[2683]: E0625 18:48:33.743390 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.743586 kubelet[2683]: E0625 18:48:33.743575 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.743625 kubelet[2683]: W0625 18:48:33.743586 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.743625 kubelet[2683]: E0625 18:48:33.743612 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.743846 kubelet[2683]: E0625 18:48:33.743820 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.743846 kubelet[2683]: W0625 18:48:33.743834 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.743912 kubelet[2683]: E0625 18:48:33.743851 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.744142 kubelet[2683]: E0625 18:48:33.744128 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.744142 kubelet[2683]: W0625 18:48:33.744141 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.744233 kubelet[2683]: E0625 18:48:33.744181 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.744427 kubelet[2683]: E0625 18:48:33.744415 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.744465 kubelet[2683]: W0625 18:48:33.744427 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.744465 kubelet[2683]: E0625 18:48:33.744441 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.744672 kubelet[2683]: E0625 18:48:33.744657 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.744672 kubelet[2683]: W0625 18:48:33.744671 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.744742 kubelet[2683]: E0625 18:48:33.744690 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.814089 kubelet[2683]: E0625 18:48:33.814046 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.814089 kubelet[2683]: W0625 18:48:33.814070 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.814089 kubelet[2683]: E0625 18:48:33.814095 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.814388 kubelet[2683]: E0625 18:48:33.814373 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.814388 kubelet[2683]: W0625 18:48:33.814382 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.814469 kubelet[2683]: E0625 18:48:33.814407 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.814654 kubelet[2683]: E0625 18:48:33.814639 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.814654 kubelet[2683]: W0625 18:48:33.814651 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.814735 kubelet[2683]: E0625 18:48:33.814670 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.815058 kubelet[2683]: E0625 18:48:33.815024 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.815058 kubelet[2683]: W0625 18:48:33.815051 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.815171 kubelet[2683]: E0625 18:48:33.815081 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.815337 kubelet[2683]: E0625 18:48:33.815316 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.815337 kubelet[2683]: W0625 18:48:33.815326 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.815410 kubelet[2683]: E0625 18:48:33.815344 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.815561 kubelet[2683]: E0625 18:48:33.815545 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.815561 kubelet[2683]: W0625 18:48:33.815555 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.815645 kubelet[2683]: E0625 18:48:33.815569 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.815816 kubelet[2683]: E0625 18:48:33.815800 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.815816 kubelet[2683]: W0625 18:48:33.815808 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.815891 kubelet[2683]: E0625 18:48:33.815849 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.816094 kubelet[2683]: E0625 18:48:33.816077 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.816094 kubelet[2683]: W0625 18:48:33.816091 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.816190 kubelet[2683]: E0625 18:48:33.816139 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.816391 kubelet[2683]: E0625 18:48:33.816376 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.816391 kubelet[2683]: W0625 18:48:33.816387 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.816471 kubelet[2683]: E0625 18:48:33.816429 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.816604 kubelet[2683]: E0625 18:48:33.816590 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.816604 kubelet[2683]: W0625 18:48:33.816601 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.816677 kubelet[2683]: E0625 18:48:33.816617 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.816851 kubelet[2683]: E0625 18:48:33.816835 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.816851 kubelet[2683]: W0625 18:48:33.816848 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.816928 kubelet[2683]: E0625 18:48:33.816868 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.817071 kubelet[2683]: E0625 18:48:33.817058 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.817071 kubelet[2683]: W0625 18:48:33.817068 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.817189 kubelet[2683]: E0625 18:48:33.817085 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.817338 kubelet[2683]: E0625 18:48:33.817324 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.817338 kubelet[2683]: W0625 18:48:33.817335 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.817406 kubelet[2683]: E0625 18:48:33.817353 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.817650 kubelet[2683]: E0625 18:48:33.817632 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.817650 kubelet[2683]: W0625 18:48:33.817644 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.817650 kubelet[2683]: E0625 18:48:33.817662 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.817882 kubelet[2683]: E0625 18:48:33.817866 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.817882 kubelet[2683]: W0625 18:48:33.817878 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.817962 kubelet[2683]: E0625 18:48:33.817895 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.818129 kubelet[2683]: E0625 18:48:33.818113 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.818129 kubelet[2683]: W0625 18:48:33.818127 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.818235 kubelet[2683]: E0625 18:48:33.818171 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.818413 kubelet[2683]: E0625 18:48:33.818397 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.818413 kubelet[2683]: W0625 18:48:33.818408 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.818479 kubelet[2683]: E0625 18:48:33.818419 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:33.818823 kubelet[2683]: E0625 18:48:33.818806 2683 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jun 25 18:48:33.818823 kubelet[2683]: W0625 18:48:33.818816 2683 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jun 25 18:48:33.818823 kubelet[2683]: E0625 18:48:33.818827 2683 plugins.go:723] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jun 25 18:48:34.393864 containerd[1570]: time="2024-06-25T18:48:34.393814248Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:34.394692 containerd[1570]: time="2024-06-25T18:48:34.394642038Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.0: active requests=0, bytes read=5140568" Jun 25 18:48:34.395897 containerd[1570]: time="2024-06-25T18:48:34.395803196Z" level=info msg="ImageCreate event name:\"sha256:587b28ecfc62e2a60919e6a39f9b25be37c77da99d8c84252716fa3a49a171b9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:34.398042 containerd[1570]: time="2024-06-25T18:48:34.398008321Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:e57c9db86f1cee1ae6f41257eed1ee2f363783177809217a2045502a09cf7cee\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:34.398786 containerd[1570]: time="2024-06-25T18:48:34.398737446Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.0\" with image id \"sha256:587b28ecfc62e2a60919e6a39f9b25be37c77da99d8c84252716fa3a49a171b9\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.0\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:e57c9db86f1cee1ae6f41257eed1ee2f363783177809217a2045502a09cf7cee\", size \"6588288\" in 1.306923424s" Jun 25 18:48:34.398786 containerd[1570]: time="2024-06-25T18:48:34.398778062Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.28.0\" returns image reference \"sha256:587b28ecfc62e2a60919e6a39f9b25be37c77da99d8c84252716fa3a49a171b9\"" Jun 25 18:48:34.400526 containerd[1570]: time="2024-06-25T18:48:34.400489336Z" level=info msg="CreateContainer within sandbox \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jun 25 18:48:34.504931 containerd[1570]: time="2024-06-25T18:48:34.504860157Z" level=info msg="CreateContainer within sandbox \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\"" Jun 25 18:48:34.505377 containerd[1570]: time="2024-06-25T18:48:34.505347074Z" level=info msg="StartContainer for \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\"" Jun 25 18:48:34.569681 kubelet[2683]: E0625 18:48:34.569644 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-bbct9" podUID="230b157d-8713-4407-885c-4ac300263b09" Jun 25 18:48:34.895252 containerd[1570]: time="2024-06-25T18:48:34.895112825Z" level=info msg="StartContainer for \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\" returns successfully" Jun 25 18:48:34.898673 containerd[1570]: time="2024-06-25T18:48:34.898630603Z" level=info msg="StopContainer for \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\" with timeout 5 (s)" Jun 25 18:48:34.899129 kubelet[2683]: I0625 18:48:34.899081 2683 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jun 25 18:48:34.900606 kubelet[2683]: E0625 18:48:34.900066 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:34.906734 containerd[1570]: time="2024-06-25T18:48:34.906529988Z" level=info msg="Stop container \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\" with signal terminated" Jun 25 18:48:34.934790 containerd[1570]: time="2024-06-25T18:48:34.934715881Z" level=info msg="shim disconnected" id=26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78 namespace=k8s.io Jun 25 18:48:34.934790 containerd[1570]: time="2024-06-25T18:48:34.934780713Z" level=warning msg="cleaning up after shim disconnected" id=26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78 namespace=k8s.io Jun 25 18:48:34.934790 containerd[1570]: time="2024-06-25T18:48:34.934789580Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jun 25 18:48:34.959543 containerd[1570]: time="2024-06-25T18:48:34.959491907Z" level=info msg="StopContainer for \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\" returns successfully" Jun 25 18:48:34.960115 containerd[1570]: time="2024-06-25T18:48:34.960092629Z" level=info msg="StopPodSandbox for \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\"" Jun 25 18:48:34.960226 containerd[1570]: time="2024-06-25T18:48:34.960128687Z" level=info msg="Container to stop \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jun 25 18:48:35.100821 containerd[1570]: time="2024-06-25T18:48:35.100570579Z" level=info msg="shim disconnected" id=f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395 namespace=k8s.io Jun 25 18:48:35.100821 containerd[1570]: time="2024-06-25T18:48:35.100635311Z" level=warning msg="cleaning up after shim disconnected" id=f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395 namespace=k8s.io Jun 25 18:48:35.100821 containerd[1570]: time="2024-06-25T18:48:35.100646032Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jun 25 18:48:35.102529 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78-rootfs.mount: Deactivated successfully. Jun 25 18:48:35.102735 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395-rootfs.mount: Deactivated successfully. Jun 25 18:48:35.102894 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395-shm.mount: Deactivated successfully. Jun 25 18:48:35.116015 containerd[1570]: time="2024-06-25T18:48:35.115967423Z" level=info msg="TearDown network for sandbox \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\" successfully" Jun 25 18:48:35.116015 containerd[1570]: time="2024-06-25T18:48:35.116001016Z" level=info msg="StopPodSandbox for \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\" returns successfully" Jun 25 18:48:35.222508 kubelet[2683]: I0625 18:48:35.222301 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-net-dir\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222508 kubelet[2683]: I0625 18:48:35.222371 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-flexvol-driver-host\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222508 kubelet[2683]: I0625 18:48:35.222404 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-policysync\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222508 kubelet[2683]: I0625 18:48:35.222434 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-log-dir\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222508 kubelet[2683]: I0625 18:48:35.222463 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-var-lib-calico\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222508 kubelet[2683]: I0625 18:48:35.222503 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lw565\" (UniqueName: \"kubernetes.io/projected/04f13a90-a54c-4fda-8fb1-db3288d0c52e-kube-api-access-lw565\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222851 kubelet[2683]: I0625 18:48:35.222538 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-lib-modules\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222851 kubelet[2683]: I0625 18:48:35.222570 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/04f13a90-a54c-4fda-8fb1-db3288d0c52e-node-certs\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222851 kubelet[2683]: I0625 18:48:35.222597 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-bin-dir\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222851 kubelet[2683]: I0625 18:48:35.222631 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-xtables-lock\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222851 kubelet[2683]: I0625 18:48:35.222662 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04f13a90-a54c-4fda-8fb1-db3288d0c52e-tigera-ca-bundle\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.222851 kubelet[2683]: I0625 18:48:35.222687 2683 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-var-run-calico\") pod \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\" (UID: \"04f13a90-a54c-4fda-8fb1-db3288d0c52e\") " Jun 25 18:48:35.223046 kubelet[2683]: I0625 18:48:35.222809 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-var-run-calico" (OuterVolumeSpecName: "var-run-calico") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "var-run-calico". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jun 25 18:48:35.223046 kubelet[2683]: I0625 18:48:35.222852 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-net-dir" (OuterVolumeSpecName: "cni-net-dir") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "cni-net-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jun 25 18:48:35.223046 kubelet[2683]: I0625 18:48:35.222873 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-flexvol-driver-host" (OuterVolumeSpecName: "flexvol-driver-host") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "flexvol-driver-host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jun 25 18:48:35.223046 kubelet[2683]: I0625 18:48:35.222893 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-policysync" (OuterVolumeSpecName: "policysync") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "policysync". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jun 25 18:48:35.223046 kubelet[2683]: I0625 18:48:35.222910 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-log-dir" (OuterVolumeSpecName: "cni-log-dir") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "cni-log-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jun 25 18:48:35.223257 kubelet[2683]: I0625 18:48:35.222927 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-var-lib-calico" (OuterVolumeSpecName: "var-lib-calico") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "var-lib-calico". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jun 25 18:48:35.223406 kubelet[2683]: I0625 18:48:35.223348 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-bin-dir" (OuterVolumeSpecName: "cni-bin-dir") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "cni-bin-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jun 25 18:48:35.223457 kubelet[2683]: I0625 18:48:35.223425 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jun 25 18:48:35.224061 kubelet[2683]: I0625 18:48:35.224023 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jun 25 18:48:35.225586 kubelet[2683]: I0625 18:48:35.224729 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04f13a90-a54c-4fda-8fb1-db3288d0c52e-tigera-ca-bundle" (OuterVolumeSpecName: "tigera-ca-bundle") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "tigera-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jun 25 18:48:35.227984 kubelet[2683]: I0625 18:48:35.227914 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04f13a90-a54c-4fda-8fb1-db3288d0c52e-kube-api-access-lw565" (OuterVolumeSpecName: "kube-api-access-lw565") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "kube-api-access-lw565". PluginName "kubernetes.io/projected", VolumeGidValue "" Jun 25 18:48:35.228508 kubelet[2683]: I0625 18:48:35.228424 2683 operation_generator.go:888] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04f13a90-a54c-4fda-8fb1-db3288d0c52e-node-certs" (OuterVolumeSpecName: "node-certs") pod "04f13a90-a54c-4fda-8fb1-db3288d0c52e" (UID: "04f13a90-a54c-4fda-8fb1-db3288d0c52e"). InnerVolumeSpecName "node-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jun 25 18:48:35.229425 systemd[1]: var-lib-kubelet-pods-04f13a90\x2da54c\x2d4fda\x2d8fb1\x2ddb3288d0c52e-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dlw565.mount: Deactivated successfully. Jun 25 18:48:35.233204 systemd[1]: var-lib-kubelet-pods-04f13a90\x2da54c\x2d4fda\x2d8fb1\x2ddb3288d0c52e-volumes-kubernetes.io\x7esecret-node\x2dcerts.mount: Deactivated successfully. Jun 25 18:48:35.323322 kubelet[2683]: I0625 18:48:35.323280 2683 reconciler_common.go:300] "Volume detached for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-var-run-calico\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323322 kubelet[2683]: I0625 18:48:35.323321 2683 reconciler_common.go:300] "Volume detached for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-net-dir\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323502 kubelet[2683]: I0625 18:48:35.323338 2683 reconciler_common.go:300] "Volume detached for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-log-dir\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323502 kubelet[2683]: I0625 18:48:35.323351 2683 reconciler_common.go:300] "Volume detached for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-flexvol-driver-host\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323502 kubelet[2683]: I0625 18:48:35.323360 2683 reconciler_common.go:300] "Volume detached for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-policysync\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323502 kubelet[2683]: I0625 18:48:35.323369 2683 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-lw565\" (UniqueName: \"kubernetes.io/projected/04f13a90-a54c-4fda-8fb1-db3288d0c52e-kube-api-access-lw565\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323502 kubelet[2683]: I0625 18:48:35.323378 2683 reconciler_common.go:300] "Volume detached for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-var-lib-calico\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323502 kubelet[2683]: I0625 18:48:35.323386 2683 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-lib-modules\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323502 kubelet[2683]: I0625 18:48:35.323399 2683 reconciler_common.go:300] "Volume detached for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/04f13a90-a54c-4fda-8fb1-db3288d0c52e-node-certs\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323502 kubelet[2683]: I0625 18:48:35.323408 2683 reconciler_common.go:300] "Volume detached for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-cni-bin-dir\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323701 kubelet[2683]: I0625 18:48:35.323416 2683 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/04f13a90-a54c-4fda-8fb1-db3288d0c52e-xtables-lock\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.323701 kubelet[2683]: I0625 18:48:35.323425 2683 reconciler_common.go:300] "Volume detached for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/04f13a90-a54c-4fda-8fb1-db3288d0c52e-tigera-ca-bundle\") on node \"localhost\" DevicePath \"\"" Jun 25 18:48:35.900764 kubelet[2683]: I0625 18:48:35.900725 2683 scope.go:117] "RemoveContainer" containerID="26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78" Jun 25 18:48:35.902206 containerd[1570]: time="2024-06-25T18:48:35.902146312Z" level=info msg="RemoveContainer for \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\"" Jun 25 18:48:35.906993 containerd[1570]: time="2024-06-25T18:48:35.906815849Z" level=info msg="RemoveContainer for \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\" returns successfully" Jun 25 18:48:35.907239 kubelet[2683]: I0625 18:48:35.907188 2683 scope.go:117] "RemoveContainer" containerID="26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78" Jun 25 18:48:35.907567 containerd[1570]: time="2024-06-25T18:48:35.907499486Z" level=error msg="ContainerStatus for \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\": not found" Jun 25 18:48:35.907741 kubelet[2683]: E0625 18:48:35.907715 2683 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\": not found" containerID="26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78" Jun 25 18:48:35.908082 kubelet[2683]: I0625 18:48:35.908049 2683 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78"} err="failed to get container status \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\": rpc error: code = NotFound desc = an error occurred when try to find container \"26a511dc45fe1cd6d2594f2b984620377ccfdd18cb3269e075013c09b3d7be78\": not found" Jun 25 18:48:35.930798 kubelet[2683]: I0625 18:48:35.930763 2683 topology_manager.go:215] "Topology Admit Handler" podUID="2f77eae3-5177-4a0b-b26b-7d47961d142e" podNamespace="calico-system" podName="calico-node-ckxph" Jun 25 18:48:35.930956 kubelet[2683]: E0625 18:48:35.930852 2683 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="04f13a90-a54c-4fda-8fb1-db3288d0c52e" containerName="flexvol-driver" Jun 25 18:48:35.930956 kubelet[2683]: I0625 18:48:35.930885 2683 memory_manager.go:346] "RemoveStaleState removing state" podUID="04f13a90-a54c-4fda-8fb1-db3288d0c52e" containerName="flexvol-driver" Jun 25 18:48:36.026512 kubelet[2683]: I0625 18:48:36.026470 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/2f77eae3-5177-4a0b-b26b-7d47961d142e-cni-bin-dir\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.026512 kubelet[2683]: I0625 18:48:36.026512 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/2f77eae3-5177-4a0b-b26b-7d47961d142e-cni-log-dir\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.026512 kubelet[2683]: I0625 18:48:36.026531 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/2f77eae3-5177-4a0b-b26b-7d47961d142e-xtables-lock\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.026786 kubelet[2683]: I0625 18:48:36.026585 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2f77eae3-5177-4a0b-b26b-7d47961d142e-tigera-ca-bundle\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.026786 kubelet[2683]: I0625 18:48:36.026704 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/2f77eae3-5177-4a0b-b26b-7d47961d142e-var-run-calico\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.026786 kubelet[2683]: I0625 18:48:36.026777 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/2f77eae3-5177-4a0b-b26b-7d47961d142e-var-lib-calico\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.026895 kubelet[2683]: I0625 18:48:36.026809 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/2f77eae3-5177-4a0b-b26b-7d47961d142e-lib-modules\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.026895 kubelet[2683]: I0625 18:48:36.026838 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/2f77eae3-5177-4a0b-b26b-7d47961d142e-policysync\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.026895 kubelet[2683]: I0625 18:48:36.026872 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/2f77eae3-5177-4a0b-b26b-7d47961d142e-node-certs\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.027000 kubelet[2683]: I0625 18:48:36.026911 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgtcp\" (UniqueName: \"kubernetes.io/projected/2f77eae3-5177-4a0b-b26b-7d47961d142e-kube-api-access-qgtcp\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.027000 kubelet[2683]: I0625 18:48:36.026933 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/2f77eae3-5177-4a0b-b26b-7d47961d142e-flexvol-driver-host\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.027000 kubelet[2683]: I0625 18:48:36.026997 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/2f77eae3-5177-4a0b-b26b-7d47961d142e-cni-net-dir\") pod \"calico-node-ckxph\" (UID: \"2f77eae3-5177-4a0b-b26b-7d47961d142e\") " pod="calico-system/calico-node-ckxph" Jun 25 18:48:36.235041 kubelet[2683]: E0625 18:48:36.234908 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:36.235488 containerd[1570]: time="2024-06-25T18:48:36.235445679Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ckxph,Uid:2f77eae3-5177-4a0b-b26b-7d47961d142e,Namespace:calico-system,Attempt:0,}" Jun 25 18:48:36.260324 containerd[1570]: time="2024-06-25T18:48:36.260141386Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:36.260324 containerd[1570]: time="2024-06-25T18:48:36.260270319Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:36.260324 containerd[1570]: time="2024-06-25T18:48:36.260289545Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:36.260324 containerd[1570]: time="2024-06-25T18:48:36.260300636Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:36.305053 containerd[1570]: time="2024-06-25T18:48:36.304984305Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-ckxph,Uid:2f77eae3-5177-4a0b-b26b-7d47961d142e,Namespace:calico-system,Attempt:0,} returns sandbox id \"7911e9e20a112e63d6f2166c8a4ac13cfb383161d32456277fc34cbcea9b487f\"" Jun 25 18:48:36.305650 kubelet[2683]: E0625 18:48:36.305612 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:36.307546 containerd[1570]: time="2024-06-25T18:48:36.307517064Z" level=info msg="CreateContainer within sandbox \"7911e9e20a112e63d6f2166c8a4ac13cfb383161d32456277fc34cbcea9b487f\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jun 25 18:48:36.325094 containerd[1570]: time="2024-06-25T18:48:36.325029093Z" level=info msg="CreateContainer within sandbox \"7911e9e20a112e63d6f2166c8a4ac13cfb383161d32456277fc34cbcea9b487f\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"92cc3d5dc8cb59d3729d538a168f5d4f391962819998accd43235bbd65bd3a2f\"" Jun 25 18:48:36.325636 containerd[1570]: time="2024-06-25T18:48:36.325592113Z" level=info msg="StartContainer for \"92cc3d5dc8cb59d3729d538a168f5d4f391962819998accd43235bbd65bd3a2f\"" Jun 25 18:48:36.392080 containerd[1570]: time="2024-06-25T18:48:36.392029038Z" level=info msg="StartContainer for \"92cc3d5dc8cb59d3729d538a168f5d4f391962819998accd43235bbd65bd3a2f\" returns successfully" Jun 25 18:48:36.453741 containerd[1570]: time="2024-06-25T18:48:36.453669860Z" level=info msg="shim disconnected" id=92cc3d5dc8cb59d3729d538a168f5d4f391962819998accd43235bbd65bd3a2f namespace=k8s.io Jun 25 18:48:36.453741 containerd[1570]: time="2024-06-25T18:48:36.453733641Z" level=warning msg="cleaning up after shim disconnected" id=92cc3d5dc8cb59d3729d538a168f5d4f391962819998accd43235bbd65bd3a2f namespace=k8s.io Jun 25 18:48:36.453741 containerd[1570]: time="2024-06-25T18:48:36.453742958Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jun 25 18:48:36.568881 kubelet[2683]: E0625 18:48:36.568834 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-bbct9" podUID="230b157d-8713-4407-885c-4ac300263b09" Jun 25 18:48:36.905707 kubelet[2683]: E0625 18:48:36.905477 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:36.906585 containerd[1570]: time="2024-06-25T18:48:36.906552653Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.28.0\"" Jun 25 18:48:37.572041 kubelet[2683]: I0625 18:48:37.571690 2683 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="04f13a90-a54c-4fda-8fb1-db3288d0c52e" path="/var/lib/kubelet/pods/04f13a90-a54c-4fda-8fb1-db3288d0c52e/volumes" Jun 25 18:48:38.569620 kubelet[2683]: E0625 18:48:38.569557 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-bbct9" podUID="230b157d-8713-4407-885c-4ac300263b09" Jun 25 18:48:40.213679 containerd[1570]: time="2024-06-25T18:48:40.213621686Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.28.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:40.214436 containerd[1570]: time="2024-06-25T18:48:40.214370404Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.28.0: active requests=0, bytes read=93087850" Jun 25 18:48:40.215629 containerd[1570]: time="2024-06-25T18:48:40.215598134Z" level=info msg="ImageCreate event name:\"sha256:107014d9f4c891a0235fa80b55df22451e8804ede5b891b632c5779ca3ab07a7\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:40.217972 containerd[1570]: time="2024-06-25T18:48:40.217948145Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:67fdc0954d3c96f9a7938fca4d5759c835b773dfb5cb513903e89d21462d886e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:40.218644 containerd[1570]: time="2024-06-25T18:48:40.218603047Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.28.0\" with image id \"sha256:107014d9f4c891a0235fa80b55df22451e8804ede5b891b632c5779ca3ab07a7\", repo tag \"ghcr.io/flatcar/calico/cni:v3.28.0\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:67fdc0954d3c96f9a7938fca4d5759c835b773dfb5cb513903e89d21462d886e\", size \"94535610\" in 3.312010659s" Jun 25 18:48:40.218691 containerd[1570]: time="2024-06-25T18:48:40.218645117Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.28.0\" returns image reference \"sha256:107014d9f4c891a0235fa80b55df22451e8804ede5b891b632c5779ca3ab07a7\"" Jun 25 18:48:40.220455 containerd[1570]: time="2024-06-25T18:48:40.220435435Z" level=info msg="CreateContainer within sandbox \"7911e9e20a112e63d6f2166c8a4ac13cfb383161d32456277fc34cbcea9b487f\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jun 25 18:48:40.234378 containerd[1570]: time="2024-06-25T18:48:40.234325066Z" level=info msg="CreateContainer within sandbox \"7911e9e20a112e63d6f2166c8a4ac13cfb383161d32456277fc34cbcea9b487f\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"2bae3b871fa1e7c5d7350ddecfd9c9598fd4ec1a12c21f44efcd69cf70bab177\"" Jun 25 18:48:40.234916 containerd[1570]: time="2024-06-25T18:48:40.234881854Z" level=info msg="StartContainer for \"2bae3b871fa1e7c5d7350ddecfd9c9598fd4ec1a12c21f44efcd69cf70bab177\"" Jun 25 18:48:40.569144 kubelet[2683]: E0625 18:48:40.569082 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-bbct9" podUID="230b157d-8713-4407-885c-4ac300263b09" Jun 25 18:48:41.104630 containerd[1570]: time="2024-06-25T18:48:41.104578509Z" level=info msg="StartContainer for \"2bae3b871fa1e7c5d7350ddecfd9c9598fd4ec1a12c21f44efcd69cf70bab177\" returns successfully" Jun 25 18:48:41.161117 systemd[1]: Started sshd@7-10.0.0.150:22-10.0.0.1:46920.service - OpenSSH per-connection server daemon (10.0.0.1:46920). Jun 25 18:48:41.215416 sshd[3555]: Accepted publickey for core from 10.0.0.1 port 46920 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:48:41.217476 sshd[3555]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:48:41.222898 systemd-logind[1553]: New session 8 of user core. Jun 25 18:48:41.238576 systemd[1]: Started session-8.scope - Session 8 of User core. Jun 25 18:48:41.526333 sshd[3555]: pam_unix(sshd:session): session closed for user core Jun 25 18:48:41.531163 systemd[1]: sshd@7-10.0.0.150:22-10.0.0.1:46920.service: Deactivated successfully. Jun 25 18:48:41.534093 systemd[1]: session-8.scope: Deactivated successfully. Jun 25 18:48:41.534920 systemd-logind[1553]: Session 8 logged out. Waiting for processes to exit. Jun 25 18:48:41.535935 systemd-logind[1553]: Removed session 8. Jun 25 18:48:42.110365 kubelet[2683]: E0625 18:48:42.110336 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:42.135302 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2bae3b871fa1e7c5d7350ddecfd9c9598fd4ec1a12c21f44efcd69cf70bab177-rootfs.mount: Deactivated successfully. Jun 25 18:48:42.142394 containerd[1570]: time="2024-06-25T18:48:42.142315633Z" level=info msg="shim disconnected" id=2bae3b871fa1e7c5d7350ddecfd9c9598fd4ec1a12c21f44efcd69cf70bab177 namespace=k8s.io Jun 25 18:48:42.142793 containerd[1570]: time="2024-06-25T18:48:42.142400342Z" level=warning msg="cleaning up after shim disconnected" id=2bae3b871fa1e7c5d7350ddecfd9c9598fd4ec1a12c21f44efcd69cf70bab177 namespace=k8s.io Jun 25 18:48:42.142793 containerd[1570]: time="2024-06-25T18:48:42.142415690Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jun 25 18:48:42.187135 kubelet[2683]: I0625 18:48:42.187099 2683 kubelet_node_status.go:493] "Fast updating node status as it just became ready" Jun 25 18:48:42.203765 kubelet[2683]: I0625 18:48:42.203725 2683 topology_manager.go:215] "Topology Admit Handler" podUID="dcdfe50f-5001-48fd-9832-129adec2bd17" podNamespace="calico-system" podName="calico-kube-controllers-84866796f8-tcv98" Jun 25 18:48:42.204146 kubelet[2683]: I0625 18:48:42.204131 2683 topology_manager.go:215] "Topology Admit Handler" podUID="f5ebc030-81ef-45b8-8422-d81d581072b4" podNamespace="kube-system" podName="coredns-5dd5756b68-bs4h4" Jun 25 18:48:42.204419 kubelet[2683]: I0625 18:48:42.204365 2683 topology_manager.go:215] "Topology Admit Handler" podUID="e0fe24fa-8aca-49e9-96c9-c8ecc910dd14" podNamespace="kube-system" podName="coredns-5dd5756b68-558t7" Jun 25 18:48:42.269488 kubelet[2683]: I0625 18:48:42.269438 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e0fe24fa-8aca-49e9-96c9-c8ecc910dd14-config-volume\") pod \"coredns-5dd5756b68-558t7\" (UID: \"e0fe24fa-8aca-49e9-96c9-c8ecc910dd14\") " pod="kube-system/coredns-5dd5756b68-558t7" Jun 25 18:48:42.269488 kubelet[2683]: I0625 18:48:42.269480 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkc8s\" (UniqueName: \"kubernetes.io/projected/e0fe24fa-8aca-49e9-96c9-c8ecc910dd14-kube-api-access-nkc8s\") pod \"coredns-5dd5756b68-558t7\" (UID: \"e0fe24fa-8aca-49e9-96c9-c8ecc910dd14\") " pod="kube-system/coredns-5dd5756b68-558t7" Jun 25 18:48:42.269679 kubelet[2683]: I0625 18:48:42.269512 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcdfe50f-5001-48fd-9832-129adec2bd17-tigera-ca-bundle\") pod \"calico-kube-controllers-84866796f8-tcv98\" (UID: \"dcdfe50f-5001-48fd-9832-129adec2bd17\") " pod="calico-system/calico-kube-controllers-84866796f8-tcv98" Jun 25 18:48:42.269679 kubelet[2683]: I0625 18:48:42.269534 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phf9z\" (UniqueName: \"kubernetes.io/projected/dcdfe50f-5001-48fd-9832-129adec2bd17-kube-api-access-phf9z\") pod \"calico-kube-controllers-84866796f8-tcv98\" (UID: \"dcdfe50f-5001-48fd-9832-129adec2bd17\") " pod="calico-system/calico-kube-controllers-84866796f8-tcv98" Jun 25 18:48:42.269748 kubelet[2683]: I0625 18:48:42.269678 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f5ebc030-81ef-45b8-8422-d81d581072b4-config-volume\") pod \"coredns-5dd5756b68-bs4h4\" (UID: \"f5ebc030-81ef-45b8-8422-d81d581072b4\") " pod="kube-system/coredns-5dd5756b68-bs4h4" Jun 25 18:48:42.269748 kubelet[2683]: I0625 18:48:42.269739 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj52m\" (UniqueName: \"kubernetes.io/projected/f5ebc030-81ef-45b8-8422-d81d581072b4-kube-api-access-tj52m\") pod \"coredns-5dd5756b68-bs4h4\" (UID: \"f5ebc030-81ef-45b8-8422-d81d581072b4\") " pod="kube-system/coredns-5dd5756b68-bs4h4" Jun 25 18:48:42.510998 kubelet[2683]: E0625 18:48:42.510877 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:42.511123 containerd[1570]: time="2024-06-25T18:48:42.510967810Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-84866796f8-tcv98,Uid:dcdfe50f-5001-48fd-9832-129adec2bd17,Namespace:calico-system,Attempt:0,}" Jun 25 18:48:42.511638 containerd[1570]: time="2024-06-25T18:48:42.511484311Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-bs4h4,Uid:f5ebc030-81ef-45b8-8422-d81d581072b4,Namespace:kube-system,Attempt:0,}" Jun 25 18:48:42.511916 kubelet[2683]: E0625 18:48:42.511855 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:42.512317 containerd[1570]: time="2024-06-25T18:48:42.512214905Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-558t7,Uid:e0fe24fa-8aca-49e9-96c9-c8ecc910dd14,Namespace:kube-system,Attempt:0,}" Jun 25 18:48:42.578593 containerd[1570]: time="2024-06-25T18:48:42.578542649Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-bbct9,Uid:230b157d-8713-4407-885c-4ac300263b09,Namespace:calico-system,Attempt:0,}" Jun 25 18:48:42.708834 containerd[1570]: time="2024-06-25T18:48:42.708734384Z" level=error msg="Failed to destroy network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.708834 containerd[1570]: time="2024-06-25T18:48:42.708793696Z" level=error msg="Failed to destroy network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.709067 containerd[1570]: time="2024-06-25T18:48:42.708755183Z" level=error msg="Failed to destroy network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.709067 containerd[1570]: time="2024-06-25T18:48:42.708787324Z" level=error msg="Failed to destroy network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.712783 containerd[1570]: time="2024-06-25T18:48:42.712733205Z" level=error msg="encountered an error cleaning up failed sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713172 containerd[1570]: time="2024-06-25T18:48:42.712799800Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-558t7,Uid:e0fe24fa-8aca-49e9-96c9-c8ecc910dd14,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713172 containerd[1570]: time="2024-06-25T18:48:42.712743464Z" level=error msg="encountered an error cleaning up failed sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713172 containerd[1570]: time="2024-06-25T18:48:42.712897995Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-bbct9,Uid:230b157d-8713-4407-885c-4ac300263b09,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713172 containerd[1570]: time="2024-06-25T18:48:42.712747822Z" level=error msg="encountered an error cleaning up failed sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713172 containerd[1570]: time="2024-06-25T18:48:42.712983024Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-84866796f8-tcv98,Uid:dcdfe50f-5001-48fd-9832-129adec2bd17,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713399 kubelet[2683]: E0625 18:48:42.713058 2683 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713399 kubelet[2683]: E0625 18:48:42.713120 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-5dd5756b68-558t7" Jun 25 18:48:42.713399 kubelet[2683]: E0625 18:48:42.713141 2683 kuberuntime_manager.go:1171] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-5dd5756b68-558t7" Jun 25 18:48:42.713399 kubelet[2683]: E0625 18:48:42.713058 2683 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713592 kubelet[2683]: E0625 18:48:42.713259 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-bbct9" Jun 25 18:48:42.713592 kubelet[2683]: E0625 18:48:42.713285 2683 kuberuntime_manager.go:1171] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-bbct9" Jun 25 18:48:42.713592 kubelet[2683]: E0625 18:48:42.713288 2683 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713592 kubelet[2683]: E0625 18:48:42.713338 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-84866796f8-tcv98" Jun 25 18:48:42.713726 kubelet[2683]: E0625 18:48:42.713345 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-bbct9_calico-system(230b157d-8713-4407-885c-4ac300263b09)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-bbct9_calico-system(230b157d-8713-4407-885c-4ac300263b09)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-bbct9" podUID="230b157d-8713-4407-885c-4ac300263b09" Jun 25 18:48:42.713726 kubelet[2683]: E0625 18:48:42.713354 2683 kuberuntime_manager.go:1171] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-84866796f8-tcv98" Jun 25 18:48:42.713726 kubelet[2683]: E0625 18:48:42.713426 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-84866796f8-tcv98_calico-system(dcdfe50f-5001-48fd-9832-129adec2bd17)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-84866796f8-tcv98_calico-system(dcdfe50f-5001-48fd-9832-129adec2bd17)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-84866796f8-tcv98" podUID="dcdfe50f-5001-48fd-9832-129adec2bd17" Jun 25 18:48:42.713878 kubelet[2683]: E0625 18:48:42.713479 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-5dd5756b68-558t7_kube-system(e0fe24fa-8aca-49e9-96c9-c8ecc910dd14)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-5dd5756b68-558t7_kube-system(e0fe24fa-8aca-49e9-96c9-c8ecc910dd14)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-5dd5756b68-558t7" podUID="e0fe24fa-8aca-49e9-96c9-c8ecc910dd14" Jun 25 18:48:42.713934 containerd[1570]: time="2024-06-25T18:48:42.713801945Z" level=error msg="encountered an error cleaning up failed sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.713934 containerd[1570]: time="2024-06-25T18:48:42.713877276Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-bs4h4,Uid:f5ebc030-81ef-45b8-8422-d81d581072b4,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.714080 kubelet[2683]: E0625 18:48:42.714034 2683 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:42.714080 kubelet[2683]: E0625 18:48:42.714072 2683 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-5dd5756b68-bs4h4" Jun 25 18:48:42.714214 kubelet[2683]: E0625 18:48:42.714087 2683 kuberuntime_manager.go:1171] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-5dd5756b68-bs4h4" Jun 25 18:48:42.714214 kubelet[2683]: E0625 18:48:42.714123 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-5dd5756b68-bs4h4_kube-system(f5ebc030-81ef-45b8-8422-d81d581072b4)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-5dd5756b68-bs4h4_kube-system(f5ebc030-81ef-45b8-8422-d81d581072b4)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-5dd5756b68-bs4h4" podUID="f5ebc030-81ef-45b8-8422-d81d581072b4" Jun 25 18:48:43.112352 kubelet[2683]: I0625 18:48:43.112293 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:48:43.112965 containerd[1570]: time="2024-06-25T18:48:43.112924674Z" level=info msg="StopPodSandbox for \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\"" Jun 25 18:48:43.113204 containerd[1570]: time="2024-06-25T18:48:43.113185165Z" level=info msg="Ensure that sandbox b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24 in task-service has been cleanup successfully" Jun 25 18:48:43.113595 kubelet[2683]: I0625 18:48:43.113571 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:48:43.114168 containerd[1570]: time="2024-06-25T18:48:43.114130973Z" level=info msg="StopPodSandbox for \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\"" Jun 25 18:48:43.114586 containerd[1570]: time="2024-06-25T18:48:43.114517640Z" level=info msg="Ensure that sandbox 1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7 in task-service has been cleanup successfully" Jun 25 18:48:43.119212 kubelet[2683]: I0625 18:48:43.117889 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:48:43.119313 containerd[1570]: time="2024-06-25T18:48:43.118515076Z" level=info msg="StopPodSandbox for \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\"" Jun 25 18:48:43.119313 containerd[1570]: time="2024-06-25T18:48:43.118713369Z" level=info msg="Ensure that sandbox 0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab in task-service has been cleanup successfully" Jun 25 18:48:43.122077 kubelet[2683]: E0625 18:48:43.122048 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:43.123136 containerd[1570]: time="2024-06-25T18:48:43.122755189Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.28.0\"" Jun 25 18:48:43.125514 kubelet[2683]: I0625 18:48:43.125136 2683 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:48:43.126807 containerd[1570]: time="2024-06-25T18:48:43.126758726Z" level=info msg="StopPodSandbox for \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\"" Jun 25 18:48:43.127001 containerd[1570]: time="2024-06-25T18:48:43.126983238Z" level=info msg="Ensure that sandbox a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe in task-service has been cleanup successfully" Jun 25 18:48:43.141348 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe-shm.mount: Deactivated successfully. Jun 25 18:48:43.153427 containerd[1570]: time="2024-06-25T18:48:43.153370562Z" level=error msg="StopPodSandbox for \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\" failed" error="failed to destroy network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:43.153824 kubelet[2683]: E0625 18:48:43.153631 2683 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:48:43.153824 kubelet[2683]: E0625 18:48:43.153679 2683 kuberuntime_manager.go:1380] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24"} Jun 25 18:48:43.153824 kubelet[2683]: E0625 18:48:43.153716 2683 kuberuntime_manager.go:1080] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"e0fe24fa-8aca-49e9-96c9-c8ecc910dd14\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jun 25 18:48:43.153824 kubelet[2683]: E0625 18:48:43.153744 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"e0fe24fa-8aca-49e9-96c9-c8ecc910dd14\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-5dd5756b68-558t7" podUID="e0fe24fa-8aca-49e9-96c9-c8ecc910dd14" Jun 25 18:48:43.163003 containerd[1570]: time="2024-06-25T18:48:43.162906081Z" level=error msg="StopPodSandbox for \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\" failed" error="failed to destroy network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:43.163228 kubelet[2683]: E0625 18:48:43.163194 2683 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:48:43.163277 kubelet[2683]: E0625 18:48:43.163252 2683 kuberuntime_manager.go:1380] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7"} Jun 25 18:48:43.163304 kubelet[2683]: E0625 18:48:43.163287 2683 kuberuntime_manager.go:1080] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"f5ebc030-81ef-45b8-8422-d81d581072b4\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jun 25 18:48:43.163388 kubelet[2683]: E0625 18:48:43.163316 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"f5ebc030-81ef-45b8-8422-d81d581072b4\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-5dd5756b68-bs4h4" podUID="f5ebc030-81ef-45b8-8422-d81d581072b4" Jun 25 18:48:43.165331 containerd[1570]: time="2024-06-25T18:48:43.165284252Z" level=error msg="StopPodSandbox for \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\" failed" error="failed to destroy network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:43.165540 kubelet[2683]: E0625 18:48:43.165519 2683 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:48:43.165579 kubelet[2683]: E0625 18:48:43.165549 2683 kuberuntime_manager.go:1380] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe"} Jun 25 18:48:43.165579 kubelet[2683]: E0625 18:48:43.165578 2683 kuberuntime_manager.go:1080] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"dcdfe50f-5001-48fd-9832-129adec2bd17\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jun 25 18:48:43.165655 kubelet[2683]: E0625 18:48:43.165606 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"dcdfe50f-5001-48fd-9832-129adec2bd17\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-84866796f8-tcv98" podUID="dcdfe50f-5001-48fd-9832-129adec2bd17" Jun 25 18:48:43.168257 containerd[1570]: time="2024-06-25T18:48:43.168218460Z" level=error msg="StopPodSandbox for \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\" failed" error="failed to destroy network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jun 25 18:48:43.168521 kubelet[2683]: E0625 18:48:43.168496 2683 remote_runtime.go:222] "StopPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to destroy network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\": plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" podSandboxID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:48:43.168562 kubelet[2683]: E0625 18:48:43.168557 2683 kuberuntime_manager.go:1380] "Failed to stop sandbox" podSandboxID={"Type":"containerd","ID":"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab"} Jun 25 18:48:43.168610 kubelet[2683]: E0625 18:48:43.168601 2683 kuberuntime_manager.go:1080] "killPodWithSyncResult failed" err="failed to \"KillPodSandbox\" for \"230b157d-8713-4407-885c-4ac300263b09\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" Jun 25 18:48:43.168660 kubelet[2683]: E0625 18:48:43.168643 2683 pod_workers.go:1300] "Error syncing pod, skipping" err="failed to \"KillPodSandbox\" for \"230b157d-8713-4407-885c-4ac300263b09\" with KillPodSandboxError: \"rpc error: code = Unknown desc = failed to destroy network for sandbox \\\"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\\\": plugin type=\\\"calico\\\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-bbct9" podUID="230b157d-8713-4407-885c-4ac300263b09" Jun 25 18:48:46.533357 systemd[1]: Started sshd@8-10.0.0.150:22-10.0.0.1:44030.service - OpenSSH per-connection server daemon (10.0.0.1:44030). Jun 25 18:48:46.564678 sshd[3841]: Accepted publickey for core from 10.0.0.1 port 44030 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:48:46.566529 sshd[3841]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:48:46.571680 systemd-logind[1553]: New session 9 of user core. Jun 25 18:48:46.577539 systemd[1]: Started session-9.scope - Session 9 of User core. Jun 25 18:48:46.620561 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount946201476.mount: Deactivated successfully. Jun 25 18:48:46.751487 sshd[3841]: pam_unix(sshd:session): session closed for user core Jun 25 18:48:46.755828 systemd[1]: sshd@8-10.0.0.150:22-10.0.0.1:44030.service: Deactivated successfully. Jun 25 18:48:46.758580 systemd-logind[1553]: Session 9 logged out. Waiting for processes to exit. Jun 25 18:48:46.758887 systemd[1]: session-9.scope: Deactivated successfully. Jun 25 18:48:46.759921 systemd-logind[1553]: Removed session 9. Jun 25 18:48:47.106099 containerd[1570]: time="2024-06-25T18:48:47.106040907Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.28.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:47.106713 containerd[1570]: time="2024-06-25T18:48:47.106675530Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.28.0: active requests=0, bytes read=115238750" Jun 25 18:48:47.107773 containerd[1570]: time="2024-06-25T18:48:47.107744969Z" level=info msg="ImageCreate event name:\"sha256:4e42b6f329bc1d197d97f6d2a1289b9e9f4a9560db3a36c8cffb5e95e64e4b49\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:47.109798 containerd[1570]: time="2024-06-25T18:48:47.109771607Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:95f8004836427050c9997ad0800819ced5636f6bda647b4158fc7c497910c8d0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:47.110358 containerd[1570]: time="2024-06-25T18:48:47.110311392Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.28.0\" with image id \"sha256:4e42b6f329bc1d197d97f6d2a1289b9e9f4a9560db3a36c8cffb5e95e64e4b49\", repo tag \"ghcr.io/flatcar/calico/node:v3.28.0\", repo digest \"ghcr.io/flatcar/calico/node@sha256:95f8004836427050c9997ad0800819ced5636f6bda647b4158fc7c497910c8d0\", size \"115238612\" in 3.987521769s" Jun 25 18:48:47.110388 containerd[1570]: time="2024-06-25T18:48:47.110356597Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.28.0\" returns image reference \"sha256:4e42b6f329bc1d197d97f6d2a1289b9e9f4a9560db3a36c8cffb5e95e64e4b49\"" Jun 25 18:48:47.117912 containerd[1570]: time="2024-06-25T18:48:47.117875004Z" level=info msg="CreateContainer within sandbox \"7911e9e20a112e63d6f2166c8a4ac13cfb383161d32456277fc34cbcea9b487f\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jun 25 18:48:47.135269 containerd[1570]: time="2024-06-25T18:48:47.135217571Z" level=info msg="CreateContainer within sandbox \"7911e9e20a112e63d6f2166c8a4ac13cfb383161d32456277fc34cbcea9b487f\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"7baf50e2637f2e9e0288bdb53a286730f8007d62d0bc792685f7e507ff32852d\"" Jun 25 18:48:47.135764 containerd[1570]: time="2024-06-25T18:48:47.135685019Z" level=info msg="StartContainer for \"7baf50e2637f2e9e0288bdb53a286730f8007d62d0bc792685f7e507ff32852d\"" Jun 25 18:48:47.228688 containerd[1570]: time="2024-06-25T18:48:47.228642490Z" level=info msg="StartContainer for \"7baf50e2637f2e9e0288bdb53a286730f8007d62d0bc792685f7e507ff32852d\" returns successfully" Jun 25 18:48:47.300625 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jun 25 18:48:47.300734 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jun 25 18:48:48.138318 kubelet[2683]: E0625 18:48:48.138047 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:48.146178 kubelet[2683]: I0625 18:48:48.146129 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/calico-node-ckxph" podStartSLOduration=2.941831602 podCreationTimestamp="2024-06-25 18:48:35 +0000 UTC" firstStartedPulling="2024-06-25 18:48:36.906347336 +0000 UTC m=+27.425044959" lastFinishedPulling="2024-06-25 18:48:47.110596488 +0000 UTC m=+37.629294111" observedRunningTime="2024-06-25 18:48:48.145276102 +0000 UTC m=+38.663973715" watchObservedRunningTime="2024-06-25 18:48:48.146080754 +0000 UTC m=+38.664778387" Jun 25 18:48:51.767539 systemd[1]: Started sshd@9-10.0.0.150:22-10.0.0.1:44038.service - OpenSSH per-connection server daemon (10.0.0.1:44038). Jun 25 18:48:51.797561 sshd[4091]: Accepted publickey for core from 10.0.0.1 port 44038 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:48:51.799544 sshd[4091]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:48:51.804041 systemd-logind[1553]: New session 10 of user core. Jun 25 18:48:51.816526 systemd[1]: Started session-10.scope - Session 10 of User core. Jun 25 18:48:51.945947 sshd[4091]: pam_unix(sshd:session): session closed for user core Jun 25 18:48:51.951791 systemd[1]: sshd@9-10.0.0.150:22-10.0.0.1:44038.service: Deactivated successfully. Jun 25 18:48:51.955340 systemd-logind[1553]: Session 10 logged out. Waiting for processes to exit. Jun 25 18:48:51.956085 systemd[1]: session-10.scope: Deactivated successfully. Jun 25 18:48:51.958345 systemd-logind[1553]: Removed session 10. Jun 25 18:48:53.575769 kubelet[2683]: I0625 18:48:53.575734 2683 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jun 25 18:48:53.576601 kubelet[2683]: E0625 18:48:53.576406 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:53.967278 systemd-networkd[1249]: vxlan.calico: Link UP Jun 25 18:48:53.967285 systemd-networkd[1249]: vxlan.calico: Gained carrier Jun 25 18:48:54.145721 kubelet[2683]: E0625 18:48:54.145680 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:55.022317 systemd-networkd[1249]: vxlan.calico: Gained IPv6LL Jun 25 18:48:55.569969 containerd[1570]: time="2024-06-25T18:48:55.569918288Z" level=info msg="StopPodSandbox for \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\"" Jun 25 18:48:55.570743 containerd[1570]: time="2024-06-25T18:48:55.569920472Z" level=info msg="StopPodSandbox for \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\"" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.630 [INFO][4315] k8s.go 608: Cleaning up netns ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.639 [INFO][4315] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" iface="eth0" netns="/var/run/netns/cni-448570a1-44d1-3aba-6bde-e097cbb74bfc" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.639 [INFO][4315] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" iface="eth0" netns="/var/run/netns/cni-448570a1-44d1-3aba-6bde-e097cbb74bfc" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.642 [INFO][4315] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" iface="eth0" netns="/var/run/netns/cni-448570a1-44d1-3aba-6bde-e097cbb74bfc" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.645 [INFO][4315] k8s.go 615: Releasing IP address(es) ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.645 [INFO][4315] utils.go 188: Calico CNI releasing IP address ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.709 [INFO][4324] ipam_plugin.go 411: Releasing address using handleID ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" HandleID="k8s-pod-network.a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.710 [INFO][4324] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.710 [INFO][4324] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.739 [WARNING][4324] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" HandleID="k8s-pod-network.a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.739 [INFO][4324] ipam_plugin.go 439: Releasing address using workloadID ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" HandleID="k8s-pod-network.a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.741 [INFO][4324] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:48:55.745511 containerd[1570]: 2024-06-25 18:48:55.743 [INFO][4315] k8s.go 621: Teardown processing complete. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:48:55.746248 containerd[1570]: time="2024-06-25T18:48:55.746209829Z" level=info msg="TearDown network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\" successfully" Jun 25 18:48:55.746248 containerd[1570]: time="2024-06-25T18:48:55.746247470Z" level=info msg="StopPodSandbox for \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\" returns successfully" Jun 25 18:48:55.747905 containerd[1570]: time="2024-06-25T18:48:55.747876899Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-84866796f8-tcv98,Uid:dcdfe50f-5001-48fd-9832-129adec2bd17,Namespace:calico-system,Attempt:1,}" Jun 25 18:48:55.748583 systemd[1]: run-netns-cni\x2d448570a1\x2d44d1\x2d3aba\x2d6bde\x2de097cbb74bfc.mount: Deactivated successfully. Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.641 [INFO][4305] k8s.go 608: Cleaning up netns ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.641 [INFO][4305] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" iface="eth0" netns="/var/run/netns/cni-cb0705cd-366b-8eb2-4393-0989f62cc41e" Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.647 [INFO][4305] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" iface="eth0" netns="/var/run/netns/cni-cb0705cd-366b-8eb2-4393-0989f62cc41e" Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.647 [INFO][4305] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" iface="eth0" netns="/var/run/netns/cni-cb0705cd-366b-8eb2-4393-0989f62cc41e" Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.647 [INFO][4305] k8s.go 615: Releasing IP address(es) ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.647 [INFO][4305] utils.go 188: Calico CNI releasing IP address ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.709 [INFO][4326] ipam_plugin.go 411: Releasing address using handleID ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" HandleID="k8s-pod-network.b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.710 [INFO][4326] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.741 [INFO][4326] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.745 [WARNING][4326] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" HandleID="k8s-pod-network.b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.745 [INFO][4326] ipam_plugin.go 439: Releasing address using workloadID ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" HandleID="k8s-pod-network.b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.747 [INFO][4326] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:48:55.753352 containerd[1570]: 2024-06-25 18:48:55.750 [INFO][4305] k8s.go 621: Teardown processing complete. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:48:55.753820 containerd[1570]: time="2024-06-25T18:48:55.753527017Z" level=info msg="TearDown network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\" successfully" Jun 25 18:48:55.753820 containerd[1570]: time="2024-06-25T18:48:55.753551102Z" level=info msg="StopPodSandbox for \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\" returns successfully" Jun 25 18:48:55.754042 kubelet[2683]: E0625 18:48:55.754021 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:55.754716 containerd[1570]: time="2024-06-25T18:48:55.754346836Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-558t7,Uid:e0fe24fa-8aca-49e9-96c9-c8ecc910dd14,Namespace:kube-system,Attempt:1,}" Jun 25 18:48:55.756473 systemd[1]: run-netns-cni\x2dcb0705cd\x2d366b\x2d8eb2\x2d4393\x2d0989f62cc41e.mount: Deactivated successfully. Jun 25 18:48:55.997991 systemd-networkd[1249]: cali931fb81213b: Link UP Jun 25 18:48:56.001438 systemd-networkd[1249]: cali931fb81213b: Gained carrier Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.919 [INFO][4341] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--5dd5756b68--558t7-eth0 coredns-5dd5756b68- kube-system e0fe24fa-8aca-49e9-96c9-c8ecc910dd14 805 0 2024-06-25 18:48:23 +0000 UTC map[k8s-app:kube-dns pod-template-hash:5dd5756b68 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-5dd5756b68-558t7 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali931fb81213b [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Namespace="kube-system" Pod="coredns-5dd5756b68-558t7" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--558t7-" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.919 [INFO][4341] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Namespace="kube-system" Pod="coredns-5dd5756b68-558t7" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.955 [INFO][4366] ipam_plugin.go 224: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" HandleID="k8s-pod-network.1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.964 [INFO][4366] ipam_plugin.go 264: Auto assigning IP ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" HandleID="k8s-pod-network.1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc0001fdad0), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-5dd5756b68-558t7", "timestamp":"2024-06-25 18:48:55.955474976 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.964 [INFO][4366] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.964 [INFO][4366] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.964 [INFO][4366] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.965 [INFO][4366] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" host="localhost" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.975 [INFO][4366] ipam.go 372: Looking up existing affinities for host host="localhost" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.979 [INFO][4366] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.980 [INFO][4366] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.982 [INFO][4366] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.982 [INFO][4366] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" host="localhost" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.984 [INFO][4366] ipam.go 1685: Creating new handle: k8s-pod-network.1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7 Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.987 [INFO][4366] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" host="localhost" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.991 [INFO][4366] ipam.go 1216: Successfully claimed IPs: [192.168.88.129/26] block=192.168.88.128/26 handle="k8s-pod-network.1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" host="localhost" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.991 [INFO][4366] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.129/26] handle="k8s-pod-network.1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" host="localhost" Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.991 [INFO][4366] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:48:56.010246 containerd[1570]: 2024-06-25 18:48:55.991 [INFO][4366] ipam_plugin.go 282: Calico CNI IPAM assigned addresses IPv4=[192.168.88.129/26] IPv6=[] ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" HandleID="k8s-pod-network.1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:56.012146 containerd[1570]: 2024-06-25 18:48:55.994 [INFO][4341] k8s.go 386: Populated endpoint ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Namespace="kube-system" Pod="coredns-5dd5756b68-558t7" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--558t7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--5dd5756b68--558t7-eth0", GenerateName:"coredns-5dd5756b68-", Namespace:"kube-system", SelfLink:"", UID:"e0fe24fa-8aca-49e9-96c9-c8ecc910dd14", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"5dd5756b68", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-5dd5756b68-558t7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali931fb81213b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:48:56.012146 containerd[1570]: 2024-06-25 18:48:55.995 [INFO][4341] k8s.go 387: Calico CNI using IPs: [192.168.88.129/32] ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Namespace="kube-system" Pod="coredns-5dd5756b68-558t7" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:56.012146 containerd[1570]: 2024-06-25 18:48:55.995 [INFO][4341] dataplane_linux.go 68: Setting the host side veth name to cali931fb81213b ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Namespace="kube-system" Pod="coredns-5dd5756b68-558t7" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:56.012146 containerd[1570]: 2024-06-25 18:48:55.998 [INFO][4341] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Namespace="kube-system" Pod="coredns-5dd5756b68-558t7" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:56.012146 containerd[1570]: 2024-06-25 18:48:55.998 [INFO][4341] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Namespace="kube-system" Pod="coredns-5dd5756b68-558t7" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--558t7-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--5dd5756b68--558t7-eth0", GenerateName:"coredns-5dd5756b68-", Namespace:"kube-system", SelfLink:"", UID:"e0fe24fa-8aca-49e9-96c9-c8ecc910dd14", ResourceVersion:"805", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"5dd5756b68", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7", Pod:"coredns-5dd5756b68-558t7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali931fb81213b", MAC:"02:62:f7:ec:30:a6", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:48:56.012146 containerd[1570]: 2024-06-25 18:48:56.006 [INFO][4341] k8s.go 500: Wrote updated endpoint to datastore ContainerID="1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7" Namespace="kube-system" Pod="coredns-5dd5756b68-558t7" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:48:56.027292 systemd-networkd[1249]: cali0344da49ff7: Link UP Jun 25 18:48:56.027506 systemd-networkd[1249]: cali0344da49ff7: Gained carrier Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:55.931 [INFO][4352] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0 calico-kube-controllers-84866796f8- calico-system dcdfe50f-5001-48fd-9832-129adec2bd17 804 0 2024-06-25 18:48:30 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:84866796f8 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s localhost calico-kube-controllers-84866796f8-tcv98 eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali0344da49ff7 [] []}} ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Namespace="calico-system" Pod="calico-kube-controllers-84866796f8-tcv98" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:55.931 [INFO][4352] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Namespace="calico-system" Pod="calico-kube-controllers-84866796f8-tcv98" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:55.965 [INFO][4370] ipam_plugin.go 224: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" HandleID="k8s-pod-network.a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:55.973 [INFO][4370] ipam_plugin.go 264: Auto assigning IP ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" HandleID="k8s-pod-network.a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000307300), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"calico-kube-controllers-84866796f8-tcv98", "timestamp":"2024-06-25 18:48:55.96505425 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:55.974 [INFO][4370] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:55.991 [INFO][4370] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:55.991 [INFO][4370] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:55.994 [INFO][4370] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" host="localhost" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:55.999 [INFO][4370] ipam.go 372: Looking up existing affinities for host host="localhost" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.004 [INFO][4370] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.008 [INFO][4370] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.010 [INFO][4370] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.010 [INFO][4370] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" host="localhost" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.011 [INFO][4370] ipam.go 1685: Creating new handle: k8s-pod-network.a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.015 [INFO][4370] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" host="localhost" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.019 [INFO][4370] ipam.go 1216: Successfully claimed IPs: [192.168.88.130/26] block=192.168.88.128/26 handle="k8s-pod-network.a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" host="localhost" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.020 [INFO][4370] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.130/26] handle="k8s-pod-network.a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" host="localhost" Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.020 [INFO][4370] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:48:56.040954 containerd[1570]: 2024-06-25 18:48:56.020 [INFO][4370] ipam_plugin.go 282: Calico CNI IPAM assigned addresses IPv4=[192.168.88.130/26] IPv6=[] ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" HandleID="k8s-pod-network.a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:56.041554 containerd[1570]: 2024-06-25 18:48:56.023 [INFO][4352] k8s.go 386: Populated endpoint ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Namespace="calico-system" Pod="calico-kube-controllers-84866796f8-tcv98" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0", GenerateName:"calico-kube-controllers-84866796f8-", Namespace:"calico-system", SelfLink:"", UID:"dcdfe50f-5001-48fd-9832-129adec2bd17", ResourceVersion:"804", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"84866796f8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-kube-controllers-84866796f8-tcv98", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali0344da49ff7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:48:56.041554 containerd[1570]: 2024-06-25 18:48:56.024 [INFO][4352] k8s.go 387: Calico CNI using IPs: [192.168.88.130/32] ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Namespace="calico-system" Pod="calico-kube-controllers-84866796f8-tcv98" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:56.041554 containerd[1570]: 2024-06-25 18:48:56.024 [INFO][4352] dataplane_linux.go 68: Setting the host side veth name to cali0344da49ff7 ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Namespace="calico-system" Pod="calico-kube-controllers-84866796f8-tcv98" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:56.041554 containerd[1570]: 2024-06-25 18:48:56.025 [INFO][4352] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Namespace="calico-system" Pod="calico-kube-controllers-84866796f8-tcv98" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:56.041554 containerd[1570]: 2024-06-25 18:48:56.026 [INFO][4352] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Namespace="calico-system" Pod="calico-kube-controllers-84866796f8-tcv98" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0", GenerateName:"calico-kube-controllers-84866796f8-", Namespace:"calico-system", SelfLink:"", UID:"dcdfe50f-5001-48fd-9832-129adec2bd17", ResourceVersion:"804", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"84866796f8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee", Pod:"calico-kube-controllers-84866796f8-tcv98", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali0344da49ff7", MAC:"0e:9e:14:0b:55:15", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:48:56.041554 containerd[1570]: 2024-06-25 18:48:56.034 [INFO][4352] k8s.go 500: Wrote updated endpoint to datastore ContainerID="a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee" Namespace="calico-system" Pod="calico-kube-controllers-84866796f8-tcv98" WorkloadEndpoint="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:48:56.047083 containerd[1570]: time="2024-06-25T18:48:56.046974741Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:56.047772 containerd[1570]: time="2024-06-25T18:48:56.047280455Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:56.047772 containerd[1570]: time="2024-06-25T18:48:56.047312174Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:56.047772 containerd[1570]: time="2024-06-25T18:48:56.047323966Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:56.065243 containerd[1570]: time="2024-06-25T18:48:56.065129361Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:56.065359 containerd[1570]: time="2024-06-25T18:48:56.065301435Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:56.065931 containerd[1570]: time="2024-06-25T18:48:56.065874781Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:56.066028 containerd[1570]: time="2024-06-25T18:48:56.065941005Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:56.076387 systemd-resolved[1471]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jun 25 18:48:56.096472 systemd-resolved[1471]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jun 25 18:48:56.104415 containerd[1570]: time="2024-06-25T18:48:56.104383011Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-558t7,Uid:e0fe24fa-8aca-49e9-96c9-c8ecc910dd14,Namespace:kube-system,Attempt:1,} returns sandbox id \"1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7\"" Jun 25 18:48:56.105049 kubelet[2683]: E0625 18:48:56.105026 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:56.108493 containerd[1570]: time="2024-06-25T18:48:56.108371839Z" level=info msg="CreateContainer within sandbox \"1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jun 25 18:48:56.124978 containerd[1570]: time="2024-06-25T18:48:56.124953427Z" level=info msg="CreateContainer within sandbox \"1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"3c2b4fc6340eb37bb73909ed416218b927f6eb37275f4d0b1d64375c7b614c18\"" Jun 25 18:48:56.125464 containerd[1570]: time="2024-06-25T18:48:56.125356334Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-84866796f8-tcv98,Uid:dcdfe50f-5001-48fd-9832-129adec2bd17,Namespace:calico-system,Attempt:1,} returns sandbox id \"a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee\"" Jun 25 18:48:56.125878 containerd[1570]: time="2024-06-25T18:48:56.125824633Z" level=info msg="StartContainer for \"3c2b4fc6340eb37bb73909ed416218b927f6eb37275f4d0b1d64375c7b614c18\"" Jun 25 18:48:56.156466 containerd[1570]: time="2024-06-25T18:48:56.156420198Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.28.0\"" Jun 25 18:48:56.189302 containerd[1570]: time="2024-06-25T18:48:56.189249367Z" level=info msg="StartContainer for \"3c2b4fc6340eb37bb73909ed416218b927f6eb37275f4d0b1d64375c7b614c18\" returns successfully" Jun 25 18:48:56.569511 containerd[1570]: time="2024-06-25T18:48:56.569446845Z" level=info msg="StopPodSandbox for \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\"" Jun 25 18:48:56.569657 containerd[1570]: time="2024-06-25T18:48:56.569455572Z" level=info msg="StopPodSandbox for \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\"" Jun 25 18:48:56.580081 kubelet[2683]: I0625 18:48:56.579887 2683 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jun 25 18:48:56.580877 kubelet[2683]: E0625 18:48:56.580863 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.626 [INFO][4568] k8s.go 608: Cleaning up netns ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.626 [INFO][4568] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" iface="eth0" netns="/var/run/netns/cni-1d8bc1f5-a347-1ac9-2493-5e7f313f37ea" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.627 [INFO][4568] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" iface="eth0" netns="/var/run/netns/cni-1d8bc1f5-a347-1ac9-2493-5e7f313f37ea" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.628 [INFO][4568] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" iface="eth0" netns="/var/run/netns/cni-1d8bc1f5-a347-1ac9-2493-5e7f313f37ea" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.629 [INFO][4568] k8s.go 615: Releasing IP address(es) ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.629 [INFO][4568] utils.go 188: Calico CNI releasing IP address ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.654 [INFO][4584] ipam_plugin.go 411: Releasing address using handleID ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" HandleID="k8s-pod-network.1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.655 [INFO][4584] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.655 [INFO][4584] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.660 [WARNING][4584] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" HandleID="k8s-pod-network.1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.660 [INFO][4584] ipam_plugin.go 439: Releasing address using workloadID ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" HandleID="k8s-pod-network.1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.661 [INFO][4584] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:48:56.666018 containerd[1570]: 2024-06-25 18:48:56.663 [INFO][4568] k8s.go 621: Teardown processing complete. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:48:56.666975 containerd[1570]: time="2024-06-25T18:48:56.666165071Z" level=info msg="TearDown network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\" successfully" Jun 25 18:48:56.666975 containerd[1570]: time="2024-06-25T18:48:56.666210426Z" level=info msg="StopPodSandbox for \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\" returns successfully" Jun 25 18:48:56.667041 kubelet[2683]: E0625 18:48:56.666887 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:56.669479 containerd[1570]: time="2024-06-25T18:48:56.669425702Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-bs4h4,Uid:f5ebc030-81ef-45b8-8422-d81d581072b4,Namespace:kube-system,Attempt:1,}" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.629 [INFO][4567] k8s.go 608: Cleaning up netns ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.629 [INFO][4567] dataplane_linux.go 530: Deleting workload's device in netns. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" iface="eth0" netns="/var/run/netns/cni-641d186a-128c-b533-1eac-f872d7fa6f6d" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.630 [INFO][4567] dataplane_linux.go 541: Entered netns, deleting veth. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" iface="eth0" netns="/var/run/netns/cni-641d186a-128c-b533-1eac-f872d7fa6f6d" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.630 [INFO][4567] dataplane_linux.go 568: Workload's veth was already gone. Nothing to do. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" iface="eth0" netns="/var/run/netns/cni-641d186a-128c-b533-1eac-f872d7fa6f6d" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.630 [INFO][4567] k8s.go 615: Releasing IP address(es) ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.630 [INFO][4567] utils.go 188: Calico CNI releasing IP address ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.655 [INFO][4585] ipam_plugin.go 411: Releasing address using handleID ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" HandleID="k8s-pod-network.0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.655 [INFO][4585] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.661 [INFO][4585] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.667 [WARNING][4585] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" HandleID="k8s-pod-network.0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.667 [INFO][4585] ipam_plugin.go 439: Releasing address using workloadID ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" HandleID="k8s-pod-network.0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.668 [INFO][4585] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:48:56.674994 containerd[1570]: 2024-06-25 18:48:56.671 [INFO][4567] k8s.go 621: Teardown processing complete. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:48:56.675497 containerd[1570]: time="2024-06-25T18:48:56.675132374Z" level=info msg="TearDown network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\" successfully" Jun 25 18:48:56.675497 containerd[1570]: time="2024-06-25T18:48:56.675167732Z" level=info msg="StopPodSandbox for \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\" returns successfully" Jun 25 18:48:56.675943 containerd[1570]: time="2024-06-25T18:48:56.675870370Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-bbct9,Uid:230b157d-8713-4407-885c-4ac300263b09,Namespace:calico-system,Attempt:1,}" Jun 25 18:48:56.761056 systemd[1]: run-netns-cni\x2d641d186a\x2d128c\x2db533\x2d1eac\x2df872d7fa6f6d.mount: Deactivated successfully. Jun 25 18:48:56.761299 systemd[1]: run-netns-cni\x2d1d8bc1f5\x2da347\x2d1ac9\x2d2493\x2d5e7f313f37ea.mount: Deactivated successfully. Jun 25 18:48:56.822251 systemd-networkd[1249]: calic2db4636d81: Link UP Jun 25 18:48:56.824868 systemd-networkd[1249]: calic2db4636d81: Gained carrier Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.731 [INFO][4620] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-coredns--5dd5756b68--bs4h4-eth0 coredns-5dd5756b68- kube-system f5ebc030-81ef-45b8-8422-d81d581072b4 823 0 2024-06-25 18:48:23 +0000 UTC map[k8s-app:kube-dns pod-template-hash:5dd5756b68 projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s localhost coredns-5dd5756b68-bs4h4 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calic2db4636d81 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Namespace="kube-system" Pod="coredns-5dd5756b68-bs4h4" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--bs4h4-" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.732 [INFO][4620] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Namespace="kube-system" Pod="coredns-5dd5756b68-bs4h4" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.782 [INFO][4647] ipam_plugin.go 224: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" HandleID="k8s-pod-network.58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.790 [INFO][4647] ipam_plugin.go 264: Auto assigning IP ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" HandleID="k8s-pod-network.58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc000680470), Attrs:map[string]string{"namespace":"kube-system", "node":"localhost", "pod":"coredns-5dd5756b68-bs4h4", "timestamp":"2024-06-25 18:48:56.782199798 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.790 [INFO][4647] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.790 [INFO][4647] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.790 [INFO][4647] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.793 [INFO][4647] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" host="localhost" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.797 [INFO][4647] ipam.go 372: Looking up existing affinities for host host="localhost" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.801 [INFO][4647] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.802 [INFO][4647] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.804 [INFO][4647] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.804 [INFO][4647] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" host="localhost" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.806 [INFO][4647] ipam.go 1685: Creating new handle: k8s-pod-network.58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256 Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.809 [INFO][4647] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" host="localhost" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.814 [INFO][4647] ipam.go 1216: Successfully claimed IPs: [192.168.88.131/26] block=192.168.88.128/26 handle="k8s-pod-network.58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" host="localhost" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.814 [INFO][4647] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.131/26] handle="k8s-pod-network.58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" host="localhost" Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.814 [INFO][4647] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:48:56.841844 containerd[1570]: 2024-06-25 18:48:56.814 [INFO][4647] ipam_plugin.go 282: Calico CNI IPAM assigned addresses IPv4=[192.168.88.131/26] IPv6=[] ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" HandleID="k8s-pod-network.58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.842396 containerd[1570]: 2024-06-25 18:48:56.818 [INFO][4620] k8s.go 386: Populated endpoint ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Namespace="kube-system" Pod="coredns-5dd5756b68-bs4h4" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--5dd5756b68--bs4h4-eth0", GenerateName:"coredns-5dd5756b68-", Namespace:"kube-system", SelfLink:"", UID:"f5ebc030-81ef-45b8-8422-d81d581072b4", ResourceVersion:"823", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"5dd5756b68", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"coredns-5dd5756b68-bs4h4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic2db4636d81", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:48:56.842396 containerd[1570]: 2024-06-25 18:48:56.818 [INFO][4620] k8s.go 387: Calico CNI using IPs: [192.168.88.131/32] ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Namespace="kube-system" Pod="coredns-5dd5756b68-bs4h4" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.842396 containerd[1570]: 2024-06-25 18:48:56.818 [INFO][4620] dataplane_linux.go 68: Setting the host side veth name to calic2db4636d81 ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Namespace="kube-system" Pod="coredns-5dd5756b68-bs4h4" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.842396 containerd[1570]: 2024-06-25 18:48:56.825 [INFO][4620] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Namespace="kube-system" Pod="coredns-5dd5756b68-bs4h4" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.842396 containerd[1570]: 2024-06-25 18:48:56.826 [INFO][4620] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Namespace="kube-system" Pod="coredns-5dd5756b68-bs4h4" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--5dd5756b68--bs4h4-eth0", GenerateName:"coredns-5dd5756b68-", Namespace:"kube-system", SelfLink:"", UID:"f5ebc030-81ef-45b8-8422-d81d581072b4", ResourceVersion:"823", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"5dd5756b68", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256", Pod:"coredns-5dd5756b68-bs4h4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic2db4636d81", MAC:"96:53:d7:f9:d9:b0", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:48:56.842396 containerd[1570]: 2024-06-25 18:48:56.836 [INFO][4620] k8s.go 500: Wrote updated endpoint to datastore ContainerID="58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256" Namespace="kube-system" Pod="coredns-5dd5756b68-bs4h4" WorkloadEndpoint="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:48:56.857249 systemd-networkd[1249]: cali8c83888f5a3: Link UP Jun 25 18:48:56.857957 systemd-networkd[1249]: cali8c83888f5a3: Gained carrier Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.729 [INFO][4630] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-csi--node--driver--bbct9-eth0 csi-node-driver- calico-system 230b157d-8713-4407-885c-4ac300263b09 824 0 2024-06-25 18:48:30 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:7d7f6c786c k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:default] map[] [] [] []} {k8s localhost csi-node-driver-bbct9 eth0 default [] [] [kns.calico-system ksa.calico-system.default] cali8c83888f5a3 [] []}} ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Namespace="calico-system" Pod="csi-node-driver-bbct9" WorkloadEndpoint="localhost-k8s-csi--node--driver--bbct9-" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.729 [INFO][4630] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Namespace="calico-system" Pod="csi-node-driver-bbct9" WorkloadEndpoint="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.782 [INFO][4646] ipam_plugin.go 224: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" HandleID="k8s-pod-network.bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.792 [INFO][4646] ipam_plugin.go 264: Auto assigning IP ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" HandleID="k8s-pod-network.bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00031a750), Attrs:map[string]string{"namespace":"calico-system", "node":"localhost", "pod":"csi-node-driver-bbct9", "timestamp":"2024-06-25 18:48:56.78261095 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.792 [INFO][4646] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.814 [INFO][4646] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.814 [INFO][4646] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.816 [INFO][4646] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" host="localhost" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.820 [INFO][4646] ipam.go 372: Looking up existing affinities for host host="localhost" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.825 [INFO][4646] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.827 [INFO][4646] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.832 [INFO][4646] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.832 [INFO][4646] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" host="localhost" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.835 [INFO][4646] ipam.go 1685: Creating new handle: k8s-pod-network.bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.838 [INFO][4646] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" host="localhost" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.844 [INFO][4646] ipam.go 1216: Successfully claimed IPs: [192.168.88.132/26] block=192.168.88.128/26 handle="k8s-pod-network.bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" host="localhost" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.844 [INFO][4646] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.132/26] handle="k8s-pod-network.bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" host="localhost" Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.844 [INFO][4646] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:48:56.870538 containerd[1570]: 2024-06-25 18:48:56.844 [INFO][4646] ipam_plugin.go 282: Calico CNI IPAM assigned addresses IPv4=[192.168.88.132/26] IPv6=[] ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" HandleID="k8s-pod-network.bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.871071 containerd[1570]: 2024-06-25 18:48:56.847 [INFO][4630] k8s.go 386: Populated endpoint ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Namespace="calico-system" Pod="csi-node-driver-bbct9" WorkloadEndpoint="localhost-k8s-csi--node--driver--bbct9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--bbct9-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"230b157d-8713-4407-885c-4ac300263b09", ResourceVersion:"824", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"7d7f6c786c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"csi-node-driver-bbct9", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"cali8c83888f5a3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:48:56.871071 containerd[1570]: 2024-06-25 18:48:56.848 [INFO][4630] k8s.go 387: Calico CNI using IPs: [192.168.88.132/32] ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Namespace="calico-system" Pod="csi-node-driver-bbct9" WorkloadEndpoint="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.871071 containerd[1570]: 2024-06-25 18:48:56.848 [INFO][4630] dataplane_linux.go 68: Setting the host side veth name to cali8c83888f5a3 ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Namespace="calico-system" Pod="csi-node-driver-bbct9" WorkloadEndpoint="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.871071 containerd[1570]: 2024-06-25 18:48:56.857 [INFO][4630] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Namespace="calico-system" Pod="csi-node-driver-bbct9" WorkloadEndpoint="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.871071 containerd[1570]: 2024-06-25 18:48:56.858 [INFO][4630] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Namespace="calico-system" Pod="csi-node-driver-bbct9" WorkloadEndpoint="localhost-k8s-csi--node--driver--bbct9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--bbct9-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"230b157d-8713-4407-885c-4ac300263b09", ResourceVersion:"824", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"7d7f6c786c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa", Pod:"csi-node-driver-bbct9", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"cali8c83888f5a3", MAC:"a6:00:bd:90:f2:35", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:48:56.871071 containerd[1570]: 2024-06-25 18:48:56.867 [INFO][4630] k8s.go 500: Wrote updated endpoint to datastore ContainerID="bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa" Namespace="calico-system" Pod="csi-node-driver-bbct9" WorkloadEndpoint="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:48:56.880320 containerd[1570]: time="2024-06-25T18:48:56.880067320Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:56.881717 containerd[1570]: time="2024-06-25T18:48:56.880260153Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:56.881717 containerd[1570]: time="2024-06-25T18:48:56.880853727Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:56.881717 containerd[1570]: time="2024-06-25T18:48:56.880880597Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:56.904925 containerd[1570]: time="2024-06-25T18:48:56.904658294Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:48:56.904925 containerd[1570]: time="2024-06-25T18:48:56.904726723Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:56.904925 containerd[1570]: time="2024-06-25T18:48:56.904741581Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:48:56.904925 containerd[1570]: time="2024-06-25T18:48:56.904752802Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:48:56.910816 systemd-resolved[1471]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jun 25 18:48:56.932203 systemd-resolved[1471]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jun 25 18:48:56.941480 containerd[1570]: time="2024-06-25T18:48:56.941403274Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5dd5756b68-bs4h4,Uid:f5ebc030-81ef-45b8-8422-d81d581072b4,Namespace:kube-system,Attempt:1,} returns sandbox id \"58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256\"" Jun 25 18:48:56.943276 kubelet[2683]: E0625 18:48:56.943142 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:56.948202 containerd[1570]: time="2024-06-25T18:48:56.948038951Z" level=info msg="CreateContainer within sandbox \"58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jun 25 18:48:56.954447 systemd[1]: Started sshd@10-10.0.0.150:22-10.0.0.1:38876.service - OpenSSH per-connection server daemon (10.0.0.1:38876). Jun 25 18:48:56.956590 containerd[1570]: time="2024-06-25T18:48:56.956318905Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-bbct9,Uid:230b157d-8713-4407-885c-4ac300263b09,Namespace:calico-system,Attempt:1,} returns sandbox id \"bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa\"" Jun 25 18:48:56.976468 containerd[1570]: time="2024-06-25T18:48:56.976377940Z" level=info msg="CreateContainer within sandbox \"58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"328bbb5accf87245a0d39932932bf98d4e3bc01f626134f92a147e2a1f2ea556\"" Jun 25 18:48:56.977320 containerd[1570]: time="2024-06-25T18:48:56.977175368Z" level=info msg="StartContainer for \"328bbb5accf87245a0d39932932bf98d4e3bc01f626134f92a147e2a1f2ea556\"" Jun 25 18:48:56.991321 sshd[4798]: Accepted publickey for core from 10.0.0.1 port 38876 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:48:56.993285 sshd[4798]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:48:56.999231 systemd-logind[1553]: New session 11 of user core. Jun 25 18:48:57.008774 systemd[1]: Started session-11.scope - Session 11 of User core. Jun 25 18:48:57.036995 containerd[1570]: time="2024-06-25T18:48:57.036952870Z" level=info msg="StartContainer for \"328bbb5accf87245a0d39932932bf98d4e3bc01f626134f92a147e2a1f2ea556\" returns successfully" Jun 25 18:48:57.130549 sshd[4798]: pam_unix(sshd:session): session closed for user core Jun 25 18:48:57.139403 systemd[1]: Started sshd@11-10.0.0.150:22-10.0.0.1:38890.service - OpenSSH per-connection server daemon (10.0.0.1:38890). Jun 25 18:48:57.140413 systemd[1]: sshd@10-10.0.0.150:22-10.0.0.1:38876.service: Deactivated successfully. Jun 25 18:48:57.144888 systemd[1]: session-11.scope: Deactivated successfully. Jun 25 18:48:57.146311 systemd-logind[1553]: Session 11 logged out. Waiting for processes to exit. Jun 25 18:48:57.147326 systemd-logind[1553]: Removed session 11. Jun 25 18:48:57.156286 kubelet[2683]: E0625 18:48:57.156250 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:57.162176 kubelet[2683]: E0625 18:48:57.162124 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:57.162408 kubelet[2683]: E0625 18:48:57.162311 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:57.168597 kubelet[2683]: I0625 18:48:57.168401 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5dd5756b68-bs4h4" podStartSLOduration=34.168355749 podCreationTimestamp="2024-06-25 18:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:48:57.166618949 +0000 UTC m=+47.685316582" watchObservedRunningTime="2024-06-25 18:48:57.168355749 +0000 UTC m=+47.687053372" Jun 25 18:48:57.169128 sshd[4848]: Accepted publickey for core from 10.0.0.1 port 38890 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:48:57.171352 sshd[4848]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:48:57.185102 systemd-logind[1553]: New session 12 of user core. Jun 25 18:48:57.193746 systemd[1]: Started session-12.scope - Session 12 of User core. Jun 25 18:48:57.198637 systemd-networkd[1249]: cali0344da49ff7: Gained IPv6LL Jun 25 18:48:57.583307 systemd-networkd[1249]: cali931fb81213b: Gained IPv6LL Jun 25 18:48:57.595532 sshd[4848]: pam_unix(sshd:session): session closed for user core Jun 25 18:48:57.601521 systemd[1]: Started sshd@12-10.0.0.150:22-10.0.0.1:38906.service - OpenSSH per-connection server daemon (10.0.0.1:38906). Jun 25 18:48:57.602087 systemd[1]: sshd@11-10.0.0.150:22-10.0.0.1:38890.service: Deactivated successfully. Jun 25 18:48:57.605733 systemd[1]: session-12.scope: Deactivated successfully. Jun 25 18:48:57.609066 systemd-logind[1553]: Session 12 logged out. Waiting for processes to exit. Jun 25 18:48:57.610660 systemd-logind[1553]: Removed session 12. Jun 25 18:48:57.630267 sshd[4865]: Accepted publickey for core from 10.0.0.1 port 38906 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:48:57.631798 sshd[4865]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:48:57.636296 systemd-logind[1553]: New session 13 of user core. Jun 25 18:48:57.641690 systemd[1]: Started session-13.scope - Session 13 of User core. Jun 25 18:48:57.820313 sshd[4865]: pam_unix(sshd:session): session closed for user core Jun 25 18:48:57.825769 systemd-logind[1553]: Session 13 logged out. Waiting for processes to exit. Jun 25 18:48:57.827466 systemd[1]: sshd@12-10.0.0.150:22-10.0.0.1:38906.service: Deactivated successfully. Jun 25 18:48:57.831074 systemd[1]: session-13.scope: Deactivated successfully. Jun 25 18:48:57.832589 systemd-logind[1553]: Removed session 13. Jun 25 18:48:58.115905 containerd[1570]: time="2024-06-25T18:48:58.115832147Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.28.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:58.117958 containerd[1570]: time="2024-06-25T18:48:58.117865544Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.28.0: active requests=0, bytes read=33505793" Jun 25 18:48:58.122247 containerd[1570]: time="2024-06-25T18:48:58.122192506Z" level=info msg="ImageCreate event name:\"sha256:428d92b02253980b402b9fb18f4cb58be36dc6bcf4893e07462732cb926ea783\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:58.124578 containerd[1570]: time="2024-06-25T18:48:58.124548338Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:c35e88abef622483409fff52313bf764a75095197be4c5a7c7830da342654de1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:58.125209 containerd[1570]: time="2024-06-25T18:48:58.125162020Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.28.0\" with image id \"sha256:428d92b02253980b402b9fb18f4cb58be36dc6bcf4893e07462732cb926ea783\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.28.0\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:c35e88abef622483409fff52313bf764a75095197be4c5a7c7830da342654de1\", size \"34953521\" in 1.968686558s" Jun 25 18:48:58.125255 containerd[1570]: time="2024-06-25T18:48:58.125207265Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.28.0\" returns image reference \"sha256:428d92b02253980b402b9fb18f4cb58be36dc6bcf4893e07462732cb926ea783\"" Jun 25 18:48:58.126541 containerd[1570]: time="2024-06-25T18:48:58.126262505Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.28.0\"" Jun 25 18:48:58.133919 containerd[1570]: time="2024-06-25T18:48:58.133860086Z" level=info msg="CreateContainer within sandbox \"a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Jun 25 18:48:58.156902 containerd[1570]: time="2024-06-25T18:48:58.156848094Z" level=info msg="CreateContainer within sandbox \"a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"b9ea66e36bedddca235cdb57013e3c5451943a9e203d495a43e6bf69af26d3ad\"" Jun 25 18:48:58.157585 containerd[1570]: time="2024-06-25T18:48:58.157537007Z" level=info msg="StartContainer for \"b9ea66e36bedddca235cdb57013e3c5451943a9e203d495a43e6bf69af26d3ad\"" Jun 25 18:48:58.166434 kubelet[2683]: E0625 18:48:58.166403 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:58.167023 kubelet[2683]: E0625 18:48:58.167003 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:58.176607 kubelet[2683]: I0625 18:48:58.176569 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5dd5756b68-558t7" podStartSLOduration=35.176530548 podCreationTimestamp="2024-06-25 18:48:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-06-25 18:48:57.180744113 +0000 UTC m=+47.699441746" watchObservedRunningTime="2024-06-25 18:48:58.176530548 +0000 UTC m=+48.695228171" Jun 25 18:48:58.243506 containerd[1570]: time="2024-06-25T18:48:58.243467677Z" level=info msg="StartContainer for \"b9ea66e36bedddca235cdb57013e3c5451943a9e203d495a43e6bf69af26d3ad\" returns successfully" Jun 25 18:48:58.542352 systemd-networkd[1249]: calic2db4636d81: Gained IPv6LL Jun 25 18:48:58.606431 systemd-networkd[1249]: cali8c83888f5a3: Gained IPv6LL Jun 25 18:48:59.169185 kubelet[2683]: E0625 18:48:59.169138 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:59.169673 kubelet[2683]: E0625 18:48:59.169443 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:48:59.227981 kubelet[2683]: I0625 18:48:59.225344 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-84866796f8-tcv98" podStartSLOduration=27.255424484 podCreationTimestamp="2024-06-25 18:48:30 +0000 UTC" firstStartedPulling="2024-06-25 18:48:56.15564906 +0000 UTC m=+46.674346684" lastFinishedPulling="2024-06-25 18:48:58.125516245 +0000 UTC m=+48.644213868" observedRunningTime="2024-06-25 18:48:59.177379373 +0000 UTC m=+49.696076996" watchObservedRunningTime="2024-06-25 18:48:59.225291668 +0000 UTC m=+49.743989291" Jun 25 18:48:59.726457 containerd[1570]: time="2024-06-25T18:48:59.726406340Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.28.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:59.727144 containerd[1570]: time="2024-06-25T18:48:59.727095934Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.28.0: active requests=0, bytes read=7641062" Jun 25 18:48:59.728323 containerd[1570]: time="2024-06-25T18:48:59.728287150Z" level=info msg="ImageCreate event name:\"sha256:1a094aeaf1521e225668c83cbf63c0ec63afbdb8c4dd7c3d2aab0ec917d103de\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:59.730765 containerd[1570]: time="2024-06-25T18:48:59.730733972Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:ac5f0089ad8eab325e5d16a59536f9292619adf16736b1554a439a66d543a63d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:48:59.731441 containerd[1570]: time="2024-06-25T18:48:59.731402356Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.28.0\" with image id \"sha256:1a094aeaf1521e225668c83cbf63c0ec63afbdb8c4dd7c3d2aab0ec917d103de\", repo tag \"ghcr.io/flatcar/calico/csi:v3.28.0\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:ac5f0089ad8eab325e5d16a59536f9292619adf16736b1554a439a66d543a63d\", size \"9088822\" in 1.605094677s" Jun 25 18:48:59.731441 containerd[1570]: time="2024-06-25T18:48:59.731437092Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.28.0\" returns image reference \"sha256:1a094aeaf1521e225668c83cbf63c0ec63afbdb8c4dd7c3d2aab0ec917d103de\"" Jun 25 18:48:59.732929 containerd[1570]: time="2024-06-25T18:48:59.732901320Z" level=info msg="CreateContainer within sandbox \"bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Jun 25 18:48:59.754076 containerd[1570]: time="2024-06-25T18:48:59.754029224Z" level=info msg="CreateContainer within sandbox \"bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"2927d841738095982fae33a3090bba2bf630fb98bf232a89aa0e3b2220c27c14\"" Jun 25 18:48:59.754929 containerd[1570]: time="2024-06-25T18:48:59.754886222Z" level=info msg="StartContainer for \"2927d841738095982fae33a3090bba2bf630fb98bf232a89aa0e3b2220c27c14\"" Jun 25 18:48:59.819954 containerd[1570]: time="2024-06-25T18:48:59.819916064Z" level=info msg="StartContainer for \"2927d841738095982fae33a3090bba2bf630fb98bf232a89aa0e3b2220c27c14\" returns successfully" Jun 25 18:48:59.822749 containerd[1570]: time="2024-06-25T18:48:59.822710900Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.0\"" Jun 25 18:49:00.172211 kubelet[2683]: E0625 18:49:00.172147 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:49:01.115289 containerd[1570]: time="2024-06-25T18:49:01.115236487Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:49:01.116323 containerd[1570]: time="2024-06-25T18:49:01.116273954Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.28.0: active requests=0, bytes read=10147655" Jun 25 18:49:01.121311 containerd[1570]: time="2024-06-25T18:49:01.119495800Z" level=info msg="ImageCreate event name:\"sha256:0f80feca743f4a84ddda4057266092db9134f9af9e20e12ea6fcfe51d7e3a020\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:49:01.124404 containerd[1570]: time="2024-06-25T18:49:01.124356312Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:b3caf3e7b3042b293728a5ab55d893798d60fec55993a9531e82997de0e534cc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:49:01.125704 containerd[1570]: time="2024-06-25T18:49:01.125489138Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.0\" with image id \"sha256:0f80feca743f4a84ddda4057266092db9134f9af9e20e12ea6fcfe51d7e3a020\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.0\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:b3caf3e7b3042b293728a5ab55d893798d60fec55993a9531e82997de0e534cc\", size \"11595367\" in 1.302738463s" Jun 25 18:49:01.125704 containerd[1570]: time="2024-06-25T18:49:01.125526628Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.28.0\" returns image reference \"sha256:0f80feca743f4a84ddda4057266092db9134f9af9e20e12ea6fcfe51d7e3a020\"" Jun 25 18:49:01.127752 containerd[1570]: time="2024-06-25T18:49:01.127650304Z" level=info msg="CreateContainer within sandbox \"bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Jun 25 18:49:01.149489 containerd[1570]: time="2024-06-25T18:49:01.149416180Z" level=info msg="CreateContainer within sandbox \"bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"e0a948ce3ab0cf7dde8d18e712f152c1448bd4bb3313a85712fbc7ff071057d6\"" Jun 25 18:49:01.150090 containerd[1570]: time="2024-06-25T18:49:01.150050791Z" level=info msg="StartContainer for \"e0a948ce3ab0cf7dde8d18e712f152c1448bd4bb3313a85712fbc7ff071057d6\"" Jun 25 18:49:01.240768 containerd[1570]: time="2024-06-25T18:49:01.240704185Z" level=info msg="StartContainer for \"e0a948ce3ab0cf7dde8d18e712f152c1448bd4bb3313a85712fbc7ff071057d6\" returns successfully" Jun 25 18:49:01.686487 kubelet[2683]: I0625 18:49:01.686450 2683 csi_plugin.go:99] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Jun 25 18:49:01.686487 kubelet[2683]: I0625 18:49:01.686484 2683 csi_plugin.go:112] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Jun 25 18:49:02.192618 kubelet[2683]: I0625 18:49:02.192460 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-system/csi-node-driver-bbct9" podStartSLOduration=28.024113857 podCreationTimestamp="2024-06-25 18:48:30 +0000 UTC" firstStartedPulling="2024-06-25 18:48:56.957635797 +0000 UTC m=+47.476333430" lastFinishedPulling="2024-06-25 18:49:01.125934384 +0000 UTC m=+51.644631997" observedRunningTime="2024-06-25 18:49:02.192076133 +0000 UTC m=+52.710773786" watchObservedRunningTime="2024-06-25 18:49:02.192412424 +0000 UTC m=+52.711110047" Jun 25 18:49:02.826529 systemd[1]: Started sshd@13-10.0.0.150:22-10.0.0.1:38916.service - OpenSSH per-connection server daemon (10.0.0.1:38916). Jun 25 18:49:02.855876 sshd[5045]: Accepted publickey for core from 10.0.0.1 port 38916 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:49:02.857495 sshd[5045]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:49:02.861744 systemd-logind[1553]: New session 14 of user core. Jun 25 18:49:02.870435 systemd[1]: Started session-14.scope - Session 14 of User core. Jun 25 18:49:02.985851 sshd[5045]: pam_unix(sshd:session): session closed for user core Jun 25 18:49:02.989998 systemd[1]: sshd@13-10.0.0.150:22-10.0.0.1:38916.service: Deactivated successfully. Jun 25 18:49:02.992137 systemd-logind[1553]: Session 14 logged out. Waiting for processes to exit. Jun 25 18:49:02.992234 systemd[1]: session-14.scope: Deactivated successfully. Jun 25 18:49:02.993178 systemd-logind[1553]: Removed session 14. Jun 25 18:49:07.997544 systemd[1]: Started sshd@14-10.0.0.150:22-10.0.0.1:48356.service - OpenSSH per-connection server daemon (10.0.0.1:48356). Jun 25 18:49:08.024029 sshd[5073]: Accepted publickey for core from 10.0.0.1 port 48356 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:49:08.025798 sshd[5073]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:49:08.030236 systemd-logind[1553]: New session 15 of user core. Jun 25 18:49:08.035501 systemd[1]: Started session-15.scope - Session 15 of User core. Jun 25 18:49:08.153782 sshd[5073]: pam_unix(sshd:session): session closed for user core Jun 25 18:49:08.162531 systemd[1]: Started sshd@15-10.0.0.150:22-10.0.0.1:48358.service - OpenSSH per-connection server daemon (10.0.0.1:48358). Jun 25 18:49:08.163227 systemd[1]: sshd@14-10.0.0.150:22-10.0.0.1:48356.service: Deactivated successfully. Jun 25 18:49:08.165896 systemd[1]: session-15.scope: Deactivated successfully. Jun 25 18:49:08.167944 systemd-logind[1553]: Session 15 logged out. Waiting for processes to exit. Jun 25 18:49:08.169588 systemd-logind[1553]: Removed session 15. Jun 25 18:49:08.191973 sshd[5087]: Accepted publickey for core from 10.0.0.1 port 48358 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:49:08.193870 sshd[5087]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:49:08.198090 systemd-logind[1553]: New session 16 of user core. Jun 25 18:49:08.205555 systemd[1]: Started session-16.scope - Session 16 of User core. Jun 25 18:49:08.475612 sshd[5087]: pam_unix(sshd:session): session closed for user core Jun 25 18:49:08.484354 systemd[1]: Started sshd@16-10.0.0.150:22-10.0.0.1:48364.service - OpenSSH per-connection server daemon (10.0.0.1:48364). Jun 25 18:49:08.484835 systemd[1]: sshd@15-10.0.0.150:22-10.0.0.1:48358.service: Deactivated successfully. Jun 25 18:49:08.488571 systemd-logind[1553]: Session 16 logged out. Waiting for processes to exit. Jun 25 18:49:08.489350 systemd[1]: session-16.scope: Deactivated successfully. Jun 25 18:49:08.490633 systemd-logind[1553]: Removed session 16. Jun 25 18:49:08.512857 sshd[5099]: Accepted publickey for core from 10.0.0.1 port 48364 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:49:08.514753 sshd[5099]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:49:08.520191 systemd-logind[1553]: New session 17 of user core. Jun 25 18:49:08.524905 systemd[1]: Started session-17.scope - Session 17 of User core. Jun 25 18:49:08.985237 kubelet[2683]: I0625 18:49:08.984067 2683 topology_manager.go:215] "Topology Admit Handler" podUID="34e811ce-72a9-4972-9793-56761df5702e" podNamespace="calico-apiserver" podName="calico-apiserver-7bfc7dcd54-5xf7n" Jun 25 18:49:09.126785 kubelet[2683]: I0625 18:49:09.126737 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzgwr\" (UniqueName: \"kubernetes.io/projected/34e811ce-72a9-4972-9793-56761df5702e-kube-api-access-xzgwr\") pod \"calico-apiserver-7bfc7dcd54-5xf7n\" (UID: \"34e811ce-72a9-4972-9793-56761df5702e\") " pod="calico-apiserver/calico-apiserver-7bfc7dcd54-5xf7n" Jun 25 18:49:09.126785 kubelet[2683]: I0625 18:49:09.126789 2683 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/34e811ce-72a9-4972-9793-56761df5702e-calico-apiserver-certs\") pod \"calico-apiserver-7bfc7dcd54-5xf7n\" (UID: \"34e811ce-72a9-4972-9793-56761df5702e\") " pod="calico-apiserver/calico-apiserver-7bfc7dcd54-5xf7n" Jun 25 18:49:09.228196 kubelet[2683]: E0625 18:49:09.228127 2683 secret.go:194] Couldn't get secret calico-apiserver/calico-apiserver-certs: secret "calico-apiserver-certs" not found Jun 25 18:49:09.229122 kubelet[2683]: E0625 18:49:09.229093 2683 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/34e811ce-72a9-4972-9793-56761df5702e-calico-apiserver-certs podName:34e811ce-72a9-4972-9793-56761df5702e nodeName:}" failed. No retries permitted until 2024-06-25 18:49:09.728220537 +0000 UTC m=+60.246918160 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "calico-apiserver-certs" (UniqueName: "kubernetes.io/secret/34e811ce-72a9-4972-9793-56761df5702e-calico-apiserver-certs") pod "calico-apiserver-7bfc7dcd54-5xf7n" (UID: "34e811ce-72a9-4972-9793-56761df5702e") : secret "calico-apiserver-certs" not found Jun 25 18:49:09.394101 sshd[5099]: pam_unix(sshd:session): session closed for user core Jun 25 18:49:09.403538 systemd[1]: Started sshd@17-10.0.0.150:22-10.0.0.1:48376.service - OpenSSH per-connection server daemon (10.0.0.1:48376). Jun 25 18:49:09.406121 systemd[1]: sshd@16-10.0.0.150:22-10.0.0.1:48364.service: Deactivated successfully. Jun 25 18:49:09.410809 systemd-logind[1553]: Session 17 logged out. Waiting for processes to exit. Jun 25 18:49:09.411694 systemd[1]: session-17.scope: Deactivated successfully. Jun 25 18:49:09.417199 systemd-logind[1553]: Removed session 17. Jun 25 18:49:09.438781 sshd[5122]: Accepted publickey for core from 10.0.0.1 port 48376 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:49:09.440346 sshd[5122]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:49:09.444327 systemd-logind[1553]: New session 18 of user core. Jun 25 18:49:09.456460 systemd[1]: Started session-18.scope - Session 18 of User core. Jun 25 18:49:09.555747 containerd[1570]: time="2024-06-25T18:49:09.555707605Z" level=info msg="StopPodSandbox for \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\"" Jun 25 18:49:09.556236 containerd[1570]: time="2024-06-25T18:49:09.555810628Z" level=info msg="TearDown network for sandbox \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\" successfully" Jun 25 18:49:09.556236 containerd[1570]: time="2024-06-25T18:49:09.555821849Z" level=info msg="StopPodSandbox for \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\" returns successfully" Jun 25 18:49:09.557322 containerd[1570]: time="2024-06-25T18:49:09.556399774Z" level=info msg="RemovePodSandbox for \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\"" Jun 25 18:49:09.561509 containerd[1570]: time="2024-06-25T18:49:09.561453884Z" level=info msg="Forcibly stopping sandbox \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\"" Jun 25 18:49:09.565828 containerd[1570]: time="2024-06-25T18:49:09.561579090Z" level=info msg="TearDown network for sandbox \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\" successfully" Jun 25 18:49:09.609271 containerd[1570]: time="2024-06-25T18:49:09.609222257Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jun 25 18:49:09.609407 containerd[1570]: time="2024-06-25T18:49:09.609297107Z" level=info msg="RemovePodSandbox \"f92a97ce4cb12ee1b7fc18b74e79bcee09dd1f630ead3ee82812d3f91d26b395\" returns successfully" Jun 25 18:49:09.610368 containerd[1570]: time="2024-06-25T18:49:09.609985769Z" level=info msg="StopPodSandbox for \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\"" Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.652 [WARNING][5150] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--5dd5756b68--bs4h4-eth0", GenerateName:"coredns-5dd5756b68-", Namespace:"kube-system", SelfLink:"", UID:"f5ebc030-81ef-45b8-8422-d81d581072b4", ResourceVersion:"875", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"5dd5756b68", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256", Pod:"coredns-5dd5756b68-bs4h4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic2db4636d81", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.653 [INFO][5150] k8s.go 608: Cleaning up netns ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.653 [INFO][5150] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" iface="eth0" netns="" Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.653 [INFO][5150] k8s.go 615: Releasing IP address(es) ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.653 [INFO][5150] utils.go 188: Calico CNI releasing IP address ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.680 [INFO][5158] ipam_plugin.go 411: Releasing address using handleID ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" HandleID="k8s-pod-network.1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.680 [INFO][5158] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.680 [INFO][5158] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.685 [WARNING][5158] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" HandleID="k8s-pod-network.1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.685 [INFO][5158] ipam_plugin.go 439: Releasing address using workloadID ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" HandleID="k8s-pod-network.1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.686 [INFO][5158] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:49:09.692810 containerd[1570]: 2024-06-25 18:49:09.689 [INFO][5150] k8s.go 621: Teardown processing complete. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:49:09.693772 containerd[1570]: time="2024-06-25T18:49:09.693621809Z" level=info msg="TearDown network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\" successfully" Jun 25 18:49:09.693772 containerd[1570]: time="2024-06-25T18:49:09.693649881Z" level=info msg="StopPodSandbox for \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\" returns successfully" Jun 25 18:49:09.694051 containerd[1570]: time="2024-06-25T18:49:09.694021268Z" level=info msg="RemovePodSandbox for \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\"" Jun 25 18:49:09.694051 containerd[1570]: time="2024-06-25T18:49:09.694049501Z" level=info msg="Forcibly stopping sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\"" Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.731 [WARNING][5180] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--5dd5756b68--bs4h4-eth0", GenerateName:"coredns-5dd5756b68-", Namespace:"kube-system", SelfLink:"", UID:"f5ebc030-81ef-45b8-8422-d81d581072b4", ResourceVersion:"875", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"5dd5756b68", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"58eb51ae4967f90ab5dba4454de37ce5826a59b4fee1d2286ed8badc5d230256", Pod:"coredns-5dd5756b68-bs4h4", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.131/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calic2db4636d81", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.731 [INFO][5180] k8s.go 608: Cleaning up netns ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.731 [INFO][5180] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" iface="eth0" netns="" Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.731 [INFO][5180] k8s.go 615: Releasing IP address(es) ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.731 [INFO][5180] utils.go 188: Calico CNI releasing IP address ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.759 [INFO][5187] ipam_plugin.go 411: Releasing address using handleID ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" HandleID="k8s-pod-network.1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.759 [INFO][5187] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.759 [INFO][5187] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.764 [WARNING][5187] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" HandleID="k8s-pod-network.1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.764 [INFO][5187] ipam_plugin.go 439: Releasing address using workloadID ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" HandleID="k8s-pod-network.1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Workload="localhost-k8s-coredns--5dd5756b68--bs4h4-eth0" Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.767 [INFO][5187] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:49:09.773979 containerd[1570]: 2024-06-25 18:49:09.770 [INFO][5180] k8s.go 621: Teardown processing complete. ContainerID="1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7" Jun 25 18:49:09.774459 containerd[1570]: time="2024-06-25T18:49:09.773979738Z" level=info msg="TearDown network for sandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\" successfully" Jun 25 18:49:09.793340 containerd[1570]: time="2024-06-25T18:49:09.793251681Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jun 25 18:49:09.793340 containerd[1570]: time="2024-06-25T18:49:09.793328205Z" level=info msg="RemovePodSandbox \"1f5af23dfe37a11b5a81ce88a10b9086e4bb2c79d75a946fbba9c9b7bfb754d7\" returns successfully" Jun 25 18:49:09.794493 sshd[5122]: pam_unix(sshd:session): session closed for user core Jun 25 18:49:09.795978 containerd[1570]: time="2024-06-25T18:49:09.795651613Z" level=info msg="StopPodSandbox for \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\"" Jun 25 18:49:09.803716 systemd[1]: Started sshd@18-10.0.0.150:22-10.0.0.1:48386.service - OpenSSH per-connection server daemon (10.0.0.1:48386). Jun 25 18:49:09.804847 systemd[1]: sshd@17-10.0.0.150:22-10.0.0.1:48376.service: Deactivated successfully. Jun 25 18:49:09.811173 systemd[1]: session-18.scope: Deactivated successfully. Jun 25 18:49:09.815651 systemd-logind[1553]: Session 18 logged out. Waiting for processes to exit. Jun 25 18:49:09.817475 systemd-logind[1553]: Removed session 18. Jun 25 18:49:09.849775 sshd[5201]: Accepted publickey for core from 10.0.0.1 port 48386 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:49:09.852865 sshd[5201]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:49:09.895823 systemd-logind[1553]: New session 19 of user core. Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.838 [WARNING][5214] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--5dd5756b68--558t7-eth0", GenerateName:"coredns-5dd5756b68-", Namespace:"kube-system", SelfLink:"", UID:"e0fe24fa-8aca-49e9-96c9-c8ecc910dd14", ResourceVersion:"872", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"5dd5756b68", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7", Pod:"coredns-5dd5756b68-558t7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali931fb81213b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.838 [INFO][5214] k8s.go 608: Cleaning up netns ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.838 [INFO][5214] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" iface="eth0" netns="" Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.838 [INFO][5214] k8s.go 615: Releasing IP address(es) ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.838 [INFO][5214] utils.go 188: Calico CNI releasing IP address ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.863 [INFO][5224] ipam_plugin.go 411: Releasing address using handleID ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" HandleID="k8s-pod-network.b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.863 [INFO][5224] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.865 [INFO][5224] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.887 [WARNING][5224] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" HandleID="k8s-pod-network.b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.887 [INFO][5224] ipam_plugin.go 439: Releasing address using workloadID ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" HandleID="k8s-pod-network.b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.891 [INFO][5224] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:49:09.899004 containerd[1570]: 2024-06-25 18:49:09.896 [INFO][5214] k8s.go 621: Teardown processing complete. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:49:09.899474 containerd[1570]: time="2024-06-25T18:49:09.899056559Z" level=info msg="TearDown network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\" successfully" Jun 25 18:49:09.899474 containerd[1570]: time="2024-06-25T18:49:09.899099570Z" level=info msg="StopPodSandbox for \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\" returns successfully" Jun 25 18:49:09.900350 containerd[1570]: time="2024-06-25T18:49:09.900324007Z" level=info msg="RemovePodSandbox for \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\"" Jun 25 18:49:09.900571 containerd[1570]: time="2024-06-25T18:49:09.900510917Z" level=info msg="Forcibly stopping sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\"" Jun 25 18:49:09.903466 containerd[1570]: time="2024-06-25T18:49:09.903437097Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bfc7dcd54-5xf7n,Uid:34e811ce-72a9-4972-9793-56761df5702e,Namespace:calico-apiserver,Attempt:0,}" Jun 25 18:49:09.903740 systemd[1]: Started session-19.scope - Session 19 of User core. Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.958 [WARNING][5251] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-coredns--5dd5756b68--558t7-eth0", GenerateName:"coredns-5dd5756b68-", Namespace:"kube-system", SelfLink:"", UID:"e0fe24fa-8aca-49e9-96c9-c8ecc910dd14", ResourceVersion:"872", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 23, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"5dd5756b68", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"1fd9c17426feebba7bb5ab5af20f0caeb161b15d1f5a99c263d502df7fd187a7", Pod:"coredns-5dd5756b68-558t7", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.88.129/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali931fb81213b", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.959 [INFO][5251] k8s.go 608: Cleaning up netns ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.959 [INFO][5251] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" iface="eth0" netns="" Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.959 [INFO][5251] k8s.go 615: Releasing IP address(es) ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.959 [INFO][5251] utils.go 188: Calico CNI releasing IP address ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.988 [INFO][5274] ipam_plugin.go 411: Releasing address using handleID ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" HandleID="k8s-pod-network.b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.988 [INFO][5274] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.988 [INFO][5274] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.994 [WARNING][5274] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" HandleID="k8s-pod-network.b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:09.994 [INFO][5274] ipam_plugin.go 439: Releasing address using workloadID ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" HandleID="k8s-pod-network.b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Workload="localhost-k8s-coredns--5dd5756b68--558t7-eth0" Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:10.000 [INFO][5274] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:49:10.008434 containerd[1570]: 2024-06-25 18:49:10.004 [INFO][5251] k8s.go 621: Teardown processing complete. ContainerID="b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24" Jun 25 18:49:10.008434 containerd[1570]: time="2024-06-25T18:49:10.006280127Z" level=info msg="TearDown network for sandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\" successfully" Jun 25 18:49:10.016550 containerd[1570]: time="2024-06-25T18:49:10.016507292Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jun 25 18:49:10.016624 containerd[1570]: time="2024-06-25T18:49:10.016575109Z" level=info msg="RemovePodSandbox \"b40e2944c0f58186bcb115f14f48855555e44a87bedc3243cd8fd7d72143ad24\" returns successfully" Jun 25 18:49:10.020259 containerd[1570]: time="2024-06-25T18:49:10.019416310Z" level=info msg="StopPodSandbox for \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\"" Jun 25 18:49:10.053766 sshd[5201]: pam_unix(sshd:session): session closed for user core Jun 25 18:49:10.061142 systemd[1]: sshd@18-10.0.0.150:22-10.0.0.1:48386.service: Deactivated successfully. Jun 25 18:49:10.068213 systemd-logind[1553]: Session 19 logged out. Waiting for processes to exit. Jun 25 18:49:10.068540 systemd[1]: session-19.scope: Deactivated successfully. Jun 25 18:49:10.073713 systemd-logind[1553]: Removed session 19. Jun 25 18:49:10.079322 systemd-networkd[1249]: cali94a7cae175f: Link UP Jun 25 18:49:10.079696 systemd-networkd[1249]: cali94a7cae175f: Gained carrier Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:09.968 [INFO][5257] plugin.go 326: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0 calico-apiserver-7bfc7dcd54- calico-apiserver 34e811ce-72a9-4972-9793-56761df5702e 1013 0 2024-06-25 18:49:08 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:7bfc7dcd54 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s localhost calico-apiserver-7bfc7dcd54-5xf7n eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali94a7cae175f [] []}} ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Namespace="calico-apiserver" Pod="calico-apiserver-7bfc7dcd54-5xf7n" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:09.968 [INFO][5257] k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Namespace="calico-apiserver" Pod="calico-apiserver-7bfc7dcd54-5xf7n" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.026 [INFO][5287] ipam_plugin.go 224: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" HandleID="k8s-pod-network.9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Workload="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.035 [INFO][5287] ipam_plugin.go 264: Auto assigning IP ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" HandleID="k8s-pod-network.9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Workload="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0xc00027f8e0), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"localhost", "pod":"calico-apiserver-7bfc7dcd54-5xf7n", "timestamp":"2024-06-25 18:49:10.026731392 +0000 UTC"}, Hostname:"localhost", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.035 [INFO][5287] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.035 [INFO][5287] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.035 [INFO][5287] ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'localhost' Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.036 [INFO][5287] ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" host="localhost" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.039 [INFO][5287] ipam.go 372: Looking up existing affinities for host host="localhost" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.046 [INFO][5287] ipam.go 489: Trying affinity for 192.168.88.128/26 host="localhost" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.048 [INFO][5287] ipam.go 155: Attempting to load block cidr=192.168.88.128/26 host="localhost" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.051 [INFO][5287] ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.88.128/26 host="localhost" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.051 [INFO][5287] ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.88.128/26 handle="k8s-pod-network.9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" host="localhost" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.053 [INFO][5287] ipam.go 1685: Creating new handle: k8s-pod-network.9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763 Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.057 [INFO][5287] ipam.go 1203: Writing block in order to claim IPs block=192.168.88.128/26 handle="k8s-pod-network.9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" host="localhost" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.070 [INFO][5287] ipam.go 1216: Successfully claimed IPs: [192.168.88.133/26] block=192.168.88.128/26 handle="k8s-pod-network.9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" host="localhost" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.070 [INFO][5287] ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.88.133/26] handle="k8s-pod-network.9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" host="localhost" Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.070 [INFO][5287] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:49:10.093177 containerd[1570]: 2024-06-25 18:49:10.070 [INFO][5287] ipam_plugin.go 282: Calico CNI IPAM assigned addresses IPv4=[192.168.88.133/26] IPv6=[] ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" HandleID="k8s-pod-network.9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Workload="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" Jun 25 18:49:10.093817 containerd[1570]: 2024-06-25 18:49:10.074 [INFO][5257] k8s.go 386: Populated endpoint ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Namespace="calico-apiserver" Pod="calico-apiserver-7bfc7dcd54-5xf7n" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0", GenerateName:"calico-apiserver-7bfc7dcd54-", Namespace:"calico-apiserver", SelfLink:"", UID:"34e811ce-72a9-4972-9793-56761df5702e", ResourceVersion:"1013", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 49, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7bfc7dcd54", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"", Pod:"calico-apiserver-7bfc7dcd54-5xf7n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali94a7cae175f", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:10.093817 containerd[1570]: 2024-06-25 18:49:10.074 [INFO][5257] k8s.go 387: Calico CNI using IPs: [192.168.88.133/32] ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Namespace="calico-apiserver" Pod="calico-apiserver-7bfc7dcd54-5xf7n" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" Jun 25 18:49:10.093817 containerd[1570]: 2024-06-25 18:49:10.074 [INFO][5257] dataplane_linux.go 68: Setting the host side veth name to cali94a7cae175f ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Namespace="calico-apiserver" Pod="calico-apiserver-7bfc7dcd54-5xf7n" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" Jun 25 18:49:10.093817 containerd[1570]: 2024-06-25 18:49:10.079 [INFO][5257] dataplane_linux.go 479: Disabling IPv4 forwarding ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Namespace="calico-apiserver" Pod="calico-apiserver-7bfc7dcd54-5xf7n" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" Jun 25 18:49:10.093817 containerd[1570]: 2024-06-25 18:49:10.079 [INFO][5257] k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Namespace="calico-apiserver" Pod="calico-apiserver-7bfc7dcd54-5xf7n" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0", GenerateName:"calico-apiserver-7bfc7dcd54-", Namespace:"calico-apiserver", SelfLink:"", UID:"34e811ce-72a9-4972-9793-56761df5702e", ResourceVersion:"1013", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 49, 8, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"7bfc7dcd54", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763", Pod:"calico-apiserver-7bfc7dcd54-5xf7n", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.88.133/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali94a7cae175f", MAC:"66:48:1b:69:0b:d8", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:10.093817 containerd[1570]: 2024-06-25 18:49:10.090 [INFO][5257] k8s.go 500: Wrote updated endpoint to datastore ContainerID="9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763" Namespace="calico-apiserver" Pod="calico-apiserver-7bfc7dcd54-5xf7n" WorkloadEndpoint="localhost-k8s-calico--apiserver--7bfc7dcd54--5xf7n-eth0" Jun 25 18:49:10.122259 containerd[1570]: time="2024-06-25T18:49:10.122139740Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jun 25 18:49:10.122259 containerd[1570]: time="2024-06-25T18:49:10.122231752Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:49:10.122401 containerd[1570]: time="2024-06-25T18:49:10.122253794Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jun 25 18:49:10.123510 containerd[1570]: time="2024-06-25T18:49:10.123312560Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.070 [WARNING][5311] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0", GenerateName:"calico-kube-controllers-84866796f8-", Namespace:"calico-system", SelfLink:"", UID:"dcdfe50f-5001-48fd-9832-129adec2bd17", ResourceVersion:"915", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"84866796f8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee", Pod:"calico-kube-controllers-84866796f8-tcv98", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali0344da49ff7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.072 [INFO][5311] k8s.go 608: Cleaning up netns ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.072 [INFO][5311] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" iface="eth0" netns="" Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.072 [INFO][5311] k8s.go 615: Releasing IP address(es) ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.072 [INFO][5311] utils.go 188: Calico CNI releasing IP address ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.110 [INFO][5326] ipam_plugin.go 411: Releasing address using handleID ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" HandleID="k8s-pod-network.a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.110 [INFO][5326] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.110 [INFO][5326] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.115 [WARNING][5326] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" HandleID="k8s-pod-network.a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.115 [INFO][5326] ipam_plugin.go 439: Releasing address using workloadID ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" HandleID="k8s-pod-network.a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.120 [INFO][5326] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:49:10.128346 containerd[1570]: 2024-06-25 18:49:10.124 [INFO][5311] k8s.go 621: Teardown processing complete. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:49:10.128346 containerd[1570]: time="2024-06-25T18:49:10.128333750Z" level=info msg="TearDown network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\" successfully" Jun 25 18:49:10.128912 containerd[1570]: time="2024-06-25T18:49:10.128358055Z" level=info msg="StopPodSandbox for \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\" returns successfully" Jun 25 18:49:10.128912 containerd[1570]: time="2024-06-25T18:49:10.128851480Z" level=info msg="RemovePodSandbox for \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\"" Jun 25 18:49:10.128912 containerd[1570]: time="2024-06-25T18:49:10.128874854Z" level=info msg="Forcibly stopping sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\"" Jun 25 18:49:10.156299 systemd-resolved[1471]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Jun 25 18:49:10.188087 containerd[1570]: time="2024-06-25T18:49:10.188001621Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-7bfc7dcd54-5xf7n,Uid:34e811ce-72a9-4972-9793-56761df5702e,Namespace:calico-apiserver,Attempt:0,} returns sandbox id \"9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763\"" Jun 25 18:49:10.189775 containerd[1570]: time="2024-06-25T18:49:10.189739712Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.28.0\"" Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.169 [WARNING][5387] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0", GenerateName:"calico-kube-controllers-84866796f8-", Namespace:"calico-system", SelfLink:"", UID:"dcdfe50f-5001-48fd-9832-129adec2bd17", ResourceVersion:"915", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"84866796f8", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"a359e2980a35b2337247983002822cecdedb6f100d70ce7cee2d1d4a013f73ee", Pod:"calico-kube-controllers-84866796f8-tcv98", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.88.130/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali0344da49ff7", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.169 [INFO][5387] k8s.go 608: Cleaning up netns ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.169 [INFO][5387] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" iface="eth0" netns="" Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.169 [INFO][5387] k8s.go 615: Releasing IP address(es) ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.169 [INFO][5387] utils.go 188: Calico CNI releasing IP address ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.188 [INFO][5404] ipam_plugin.go 411: Releasing address using handleID ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" HandleID="k8s-pod-network.a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.189 [INFO][5404] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.189 [INFO][5404] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.196 [WARNING][5404] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" HandleID="k8s-pod-network.a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.196 [INFO][5404] ipam_plugin.go 439: Releasing address using workloadID ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" HandleID="k8s-pod-network.a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Workload="localhost-k8s-calico--kube--controllers--84866796f8--tcv98-eth0" Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.198 [INFO][5404] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:49:10.204782 containerd[1570]: 2024-06-25 18:49:10.202 [INFO][5387] k8s.go 621: Teardown processing complete. ContainerID="a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe" Jun 25 18:49:10.205204 containerd[1570]: time="2024-06-25T18:49:10.204830892Z" level=info msg="TearDown network for sandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\" successfully" Jun 25 18:49:10.210899 containerd[1570]: time="2024-06-25T18:49:10.210840945Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jun 25 18:49:10.210942 containerd[1570]: time="2024-06-25T18:49:10.210907310Z" level=info msg="RemovePodSandbox \"a3c5a0c235bc2c103b8ddfee12e7ee03481cea17372e8193761667101ed544fe\" returns successfully" Jun 25 18:49:10.211367 containerd[1570]: time="2024-06-25T18:49:10.211335063Z" level=info msg="StopPodSandbox for \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\"" Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.244 [WARNING][5434] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--bbct9-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"230b157d-8713-4407-885c-4ac300263b09", ResourceVersion:"937", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"7d7f6c786c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa", Pod:"csi-node-driver-bbct9", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"cali8c83888f5a3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.245 [INFO][5434] k8s.go 608: Cleaning up netns ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.245 [INFO][5434] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" iface="eth0" netns="" Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.245 [INFO][5434] k8s.go 615: Releasing IP address(es) ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.245 [INFO][5434] utils.go 188: Calico CNI releasing IP address ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.263 [INFO][5442] ipam_plugin.go 411: Releasing address using handleID ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" HandleID="k8s-pod-network.0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.264 [INFO][5442] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.264 [INFO][5442] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.268 [WARNING][5442] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" HandleID="k8s-pod-network.0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.268 [INFO][5442] ipam_plugin.go 439: Releasing address using workloadID ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" HandleID="k8s-pod-network.0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.269 [INFO][5442] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:49:10.274662 containerd[1570]: 2024-06-25 18:49:10.272 [INFO][5434] k8s.go 621: Teardown processing complete. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:49:10.274662 containerd[1570]: time="2024-06-25T18:49:10.274623787Z" level=info msg="TearDown network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\" successfully" Jun 25 18:49:10.274662 containerd[1570]: time="2024-06-25T18:49:10.274657601Z" level=info msg="StopPodSandbox for \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\" returns successfully" Jun 25 18:49:10.275259 containerd[1570]: time="2024-06-25T18:49:10.275228601Z" level=info msg="RemovePodSandbox for \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\"" Jun 25 18:49:10.275297 containerd[1570]: time="2024-06-25T18:49:10.275259229Z" level=info msg="Forcibly stopping sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\"" Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.310 [WARNING][5465] k8s.go 572: CNI_CONTAINERID does not match WorkloadEndpoint ContainerID, don't delete WEP. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" WorkloadEndpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"localhost-k8s-csi--node--driver--bbct9-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"230b157d-8713-4407-885c-4ac300263b09", ResourceVersion:"937", Generation:0, CreationTimestamp:time.Date(2024, time.June, 25, 18, 48, 30, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"7d7f6c786c", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"default"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"localhost", ContainerID:"bfa9d8b1af5e6176ef362528fec7255479a6e982dc23926e55b8747f9c6709fa", Pod:"csi-node-driver-bbct9", Endpoint:"eth0", ServiceAccountName:"default", IPNetworks:[]string{"192.168.88.132/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.default"}, InterfaceName:"cali8c83888f5a3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.310 [INFO][5465] k8s.go 608: Cleaning up netns ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.310 [INFO][5465] dataplane_linux.go 526: CleanUpNamespace called with no netns name, ignoring. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" iface="eth0" netns="" Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.310 [INFO][5465] k8s.go 615: Releasing IP address(es) ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.310 [INFO][5465] utils.go 188: Calico CNI releasing IP address ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.333 [INFO][5473] ipam_plugin.go 411: Releasing address using handleID ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" HandleID="k8s-pod-network.0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.333 [INFO][5473] ipam_plugin.go 352: About to acquire host-wide IPAM lock. Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.333 [INFO][5473] ipam_plugin.go 367: Acquired host-wide IPAM lock. Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.338 [WARNING][5473] ipam_plugin.go 428: Asked to release address but it doesn't exist. Ignoring ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" HandleID="k8s-pod-network.0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.338 [INFO][5473] ipam_plugin.go 439: Releasing address using workloadID ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" HandleID="k8s-pod-network.0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Workload="localhost-k8s-csi--node--driver--bbct9-eth0" Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.339 [INFO][5473] ipam_plugin.go 373: Released host-wide IPAM lock. Jun 25 18:49:10.343396 containerd[1570]: 2024-06-25 18:49:10.341 [INFO][5465] k8s.go 621: Teardown processing complete. ContainerID="0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab" Jun 25 18:49:10.343939 containerd[1570]: time="2024-06-25T18:49:10.343433138Z" level=info msg="TearDown network for sandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\" successfully" Jun 25 18:49:10.347119 containerd[1570]: time="2024-06-25T18:49:10.347083595Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jun 25 18:49:10.347190 containerd[1570]: time="2024-06-25T18:49:10.347128089Z" level=info msg="RemovePodSandbox \"0ec177e7810f9d5245499c601bac5148f881edf8b3e71b03f36bc5710f4363ab\" returns successfully" Jun 25 18:49:11.534327 systemd-networkd[1249]: cali94a7cae175f: Gained IPv6LL Jun 25 18:49:12.080345 containerd[1570]: time="2024-06-25T18:49:12.080296686Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.28.0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:49:12.081029 containerd[1570]: time="2024-06-25T18:49:12.080967890Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.28.0: active requests=0, bytes read=40421260" Jun 25 18:49:12.085759 containerd[1570]: time="2024-06-25T18:49:12.085724156Z" level=info msg="ImageCreate event name:\"sha256:6c07591fd1cfafb48d575f75a6b9d8d3cc03bead5b684908ef5e7dd3132794d6\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:49:12.088048 containerd[1570]: time="2024-06-25T18:49:12.088010325Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:e8f124312a4c41451e51bfc00b6e98929e9eb0510905f3301542719a3e8d2fec\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jun 25 18:49:12.088667 containerd[1570]: time="2024-06-25T18:49:12.088634089Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.28.0\" with image id \"sha256:6c07591fd1cfafb48d575f75a6b9d8d3cc03bead5b684908ef5e7dd3132794d6\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.28.0\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:e8f124312a4c41451e51bfc00b6e98929e9eb0510905f3301542719a3e8d2fec\", size \"41869036\" in 1.898848571s" Jun 25 18:49:12.088703 containerd[1570]: time="2024-06-25T18:49:12.088663235Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.28.0\" returns image reference \"sha256:6c07591fd1cfafb48d575f75a6b9d8d3cc03bead5b684908ef5e7dd3132794d6\"" Jun 25 18:49:12.090261 containerd[1570]: time="2024-06-25T18:49:12.090217233Z" level=info msg="CreateContainer within sandbox \"9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jun 25 18:49:12.101178 containerd[1570]: time="2024-06-25T18:49:12.101133400Z" level=info msg="CreateContainer within sandbox \"9b05cbea9b0af0979d9fea7e032fc8300691c26efc6e28213dba37c8a4295763\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"da29d42d48dbcf5ef379b011e208f55948a8cabc8c24870f35a91c095d3c80cf\"" Jun 25 18:49:12.102273 containerd[1570]: time="2024-06-25T18:49:12.102231438Z" level=info msg="StartContainer for \"da29d42d48dbcf5ef379b011e208f55948a8cabc8c24870f35a91c095d3c80cf\"" Jun 25 18:49:12.169803 containerd[1570]: time="2024-06-25T18:49:12.169763701Z" level=info msg="StartContainer for \"da29d42d48dbcf5ef379b011e208f55948a8cabc8c24870f35a91c095d3c80cf\" returns successfully" Jun 25 18:49:12.216489 kubelet[2683]: I0625 18:49:12.216275 2683 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-7bfc7dcd54-5xf7n" podStartSLOduration=2.316638667 podCreationTimestamp="2024-06-25 18:49:08 +0000 UTC" firstStartedPulling="2024-06-25 18:49:10.189330474 +0000 UTC m=+60.708028097" lastFinishedPulling="2024-06-25 18:49:12.088927184 +0000 UTC m=+62.607624797" observedRunningTime="2024-06-25 18:49:12.21473895 +0000 UTC m=+62.733436573" watchObservedRunningTime="2024-06-25 18:49:12.216235367 +0000 UTC m=+62.734932990" Jun 25 18:49:15.060373 systemd[1]: Started sshd@19-10.0.0.150:22-10.0.0.1:48394.service - OpenSSH per-connection server daemon (10.0.0.1:48394). Jun 25 18:49:15.088817 sshd[5565]: Accepted publickey for core from 10.0.0.1 port 48394 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:49:15.090215 sshd[5565]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:49:15.094223 systemd-logind[1553]: New session 20 of user core. Jun 25 18:49:15.099391 systemd[1]: Started session-20.scope - Session 20 of User core. Jun 25 18:49:15.214897 sshd[5565]: pam_unix(sshd:session): session closed for user core Jun 25 18:49:15.219117 systemd[1]: sshd@19-10.0.0.150:22-10.0.0.1:48394.service: Deactivated successfully. Jun 25 18:49:15.221709 systemd[1]: session-20.scope: Deactivated successfully. Jun 25 18:49:15.222602 systemd-logind[1553]: Session 20 logged out. Waiting for processes to exit. Jun 25 18:49:15.223766 systemd-logind[1553]: Removed session 20. Jun 25 18:49:20.229356 systemd[1]: Started sshd@20-10.0.0.150:22-10.0.0.1:44964.service - OpenSSH per-connection server daemon (10.0.0.1:44964). Jun 25 18:49:20.255915 sshd[5586]: Accepted publickey for core from 10.0.0.1 port 44964 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:49:20.257367 sshd[5586]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:49:20.261076 systemd-logind[1553]: New session 21 of user core. Jun 25 18:49:20.266398 systemd[1]: Started session-21.scope - Session 21 of User core. Jun 25 18:49:20.365746 sshd[5586]: pam_unix(sshd:session): session closed for user core Jun 25 18:49:20.369491 systemd[1]: sshd@20-10.0.0.150:22-10.0.0.1:44964.service: Deactivated successfully. Jun 25 18:49:20.372138 systemd-logind[1553]: Session 21 logged out. Waiting for processes to exit. Jun 25 18:49:20.372174 systemd[1]: session-21.scope: Deactivated successfully. Jun 25 18:49:20.373213 systemd-logind[1553]: Removed session 21. Jun 25 18:49:22.569459 kubelet[2683]: E0625 18:49:22.569430 2683 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Jun 25 18:49:25.381394 systemd[1]: Started sshd@21-10.0.0.150:22-10.0.0.1:44966.service - OpenSSH per-connection server daemon (10.0.0.1:44966). Jun 25 18:49:25.407097 sshd[5609]: Accepted publickey for core from 10.0.0.1 port 44966 ssh2: RSA SHA256:aOL7xLJVSGgo2ACgb9Q96KiqqB5PNY5rPU/3iN9wkOM Jun 25 18:49:25.408474 sshd[5609]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Jun 25 18:49:25.412163 systemd-logind[1553]: New session 22 of user core. Jun 25 18:49:25.422570 systemd[1]: Started session-22.scope - Session 22 of User core. Jun 25 18:49:25.520467 sshd[5609]: pam_unix(sshd:session): session closed for user core Jun 25 18:49:25.523904 systemd[1]: sshd@21-10.0.0.150:22-10.0.0.1:44966.service: Deactivated successfully. Jun 25 18:49:25.526205 systemd-logind[1553]: Session 22 logged out. Waiting for processes to exit. Jun 25 18:49:25.526284 systemd[1]: session-22.scope: Deactivated successfully. Jun 25 18:49:25.527455 systemd-logind[1553]: Removed session 22.