Jan 30 13:22:30.375923 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Jan 30 13:22:30.375951 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241116 p3) 14.2.1 20241116, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT Wed Jan 29 09:30:22 -00 2025 Jan 30 13:22:30.375961 kernel: KASLR enabled Jan 30 13:22:30.375967 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Jan 30 13:22:30.375976 kernel: printk: bootconsole [pl11] enabled Jan 30 13:22:30.375982 kernel: efi: EFI v2.7 by EDK II Jan 30 13:22:30.375991 kernel: efi: ACPI 2.0=0x3fd5f018 SMBIOS=0x3e580000 SMBIOS 3.0=0x3e560000 MEMATTR=0x3ead5018 RNG=0x3fd5f998 MEMRESERVE=0x3e477598 Jan 30 13:22:30.375997 kernel: random: crng init done Jan 30 13:22:30.376003 kernel: secureboot: Secure boot disabled Jan 30 13:22:30.376009 kernel: ACPI: Early table checksum verification disabled Jan 30 13:22:30.376015 kernel: ACPI: RSDP 0x000000003FD5F018 000024 (v02 VRTUAL) Jan 30 13:22:30.376020 kernel: ACPI: XSDT 0x000000003FD5FF18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376026 kernel: ACPI: FACP 0x000000003FD5FC18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376034 kernel: ACPI: DSDT 0x000000003FD41018 01DFCD (v02 MSFTVM DSDT01 00000001 INTL 20230628) Jan 30 13:22:30.376041 kernel: ACPI: DBG2 0x000000003FD5FB18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376047 kernel: ACPI: GTDT 0x000000003FD5FD98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376053 kernel: ACPI: OEM0 0x000000003FD5F098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376062 kernel: ACPI: SPCR 0x000000003FD5FA98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376068 kernel: ACPI: APIC 0x000000003FD5F818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376074 kernel: ACPI: SRAT 0x000000003FD5F198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376081 kernel: ACPI: PPTT 0x000000003FD5F418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Jan 30 13:22:30.376087 kernel: ACPI: BGRT 0x000000003FD5FE98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376093 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Jan 30 13:22:30.376100 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] Jan 30 13:22:30.376106 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] Jan 30 13:22:30.376114 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] Jan 30 13:22:30.376120 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] Jan 30 13:22:30.376128 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] Jan 30 13:22:30.376136 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] Jan 30 13:22:30.376143 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] Jan 30 13:22:30.376149 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] Jan 30 13:22:30.376155 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] Jan 30 13:22:30.376162 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] Jan 30 13:22:30.376169 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] Jan 30 13:22:30.376175 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] Jan 30 13:22:30.378223 kernel: NUMA: NODE_DATA [mem 0x1bf7ef800-0x1bf7f4fff] Jan 30 13:22:30.378249 kernel: Zone ranges: Jan 30 13:22:30.378256 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Jan 30 13:22:30.378263 kernel: DMA32 empty Jan 30 13:22:30.378270 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Jan 30 13:22:30.378287 kernel: Movable zone start for each node Jan 30 13:22:30.378293 kernel: Early memory node ranges Jan 30 13:22:30.378300 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Jan 30 13:22:30.378307 kernel: node 0: [mem 0x0000000000824000-0x000000003e45ffff] Jan 30 13:22:30.378313 kernel: node 0: [mem 0x000000003e460000-0x000000003e46ffff] Jan 30 13:22:30.378322 kernel: node 0: [mem 0x000000003e470000-0x000000003e54ffff] Jan 30 13:22:30.378328 kernel: node 0: [mem 0x000000003e550000-0x000000003e87ffff] Jan 30 13:22:30.378335 kernel: node 0: [mem 0x000000003e880000-0x000000003fc7ffff] Jan 30 13:22:30.378341 kernel: node 0: [mem 0x000000003fc80000-0x000000003fcfffff] Jan 30 13:22:30.378348 kernel: node 0: [mem 0x000000003fd00000-0x000000003fffffff] Jan 30 13:22:30.378354 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Jan 30 13:22:30.378361 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Jan 30 13:22:30.378368 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Jan 30 13:22:30.378375 kernel: psci: probing for conduit method from ACPI. Jan 30 13:22:30.378381 kernel: psci: PSCIv1.1 detected in firmware. Jan 30 13:22:30.378388 kernel: psci: Using standard PSCI v0.2 function IDs Jan 30 13:22:30.378395 kernel: psci: MIGRATE_INFO_TYPE not supported. Jan 30 13:22:30.378403 kernel: psci: SMC Calling Convention v1.4 Jan 30 13:22:30.378410 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Jan 30 13:22:30.378416 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Jan 30 13:22:30.378423 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Jan 30 13:22:30.378429 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Jan 30 13:22:30.378436 kernel: pcpu-alloc: [0] 0 [0] 1 Jan 30 13:22:30.378443 kernel: Detected PIPT I-cache on CPU0 Jan 30 13:22:30.378449 kernel: CPU features: detected: GIC system register CPU interface Jan 30 13:22:30.378456 kernel: CPU features: detected: Hardware dirty bit management Jan 30 13:22:30.378462 kernel: CPU features: detected: Spectre-BHB Jan 30 13:22:30.378469 kernel: CPU features: kernel page table isolation forced ON by KASLR Jan 30 13:22:30.378477 kernel: CPU features: detected: Kernel page table isolation (KPTI) Jan 30 13:22:30.378484 kernel: CPU features: detected: ARM erratum 1418040 Jan 30 13:22:30.378490 kernel: CPU features: detected: ARM erratum 1542419 (kernel portion) Jan 30 13:22:30.378497 kernel: CPU features: detected: SSBS not fully self-synchronizing Jan 30 13:22:30.378503 kernel: alternatives: applying boot alternatives Jan 30 13:22:30.378511 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=e6957044c3256d96283265c263579aa4275d1d707b02496fcb081f5fc6356346 Jan 30 13:22:30.378518 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jan 30 13:22:30.378525 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 30 13:22:30.378532 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 30 13:22:30.378538 kernel: Fallback order for Node 0: 0 Jan 30 13:22:30.378545 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1032156 Jan 30 13:22:30.378553 kernel: Policy zone: Normal Jan 30 13:22:30.378559 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 30 13:22:30.378566 kernel: software IO TLB: area num 2. Jan 30 13:22:30.378572 kernel: software IO TLB: mapped [mem 0x000000003a460000-0x000000003e460000] (64MB) Jan 30 13:22:30.378579 kernel: Memory: 3982056K/4194160K available (10304K kernel code, 2186K rwdata, 8092K rodata, 39936K init, 897K bss, 212104K reserved, 0K cma-reserved) Jan 30 13:22:30.378586 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jan 30 13:22:30.378592 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 30 13:22:30.378600 kernel: rcu: RCU event tracing is enabled. Jan 30 13:22:30.378607 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jan 30 13:22:30.378614 kernel: Trampoline variant of Tasks RCU enabled. Jan 30 13:22:30.378620 kernel: Tracing variant of Tasks RCU enabled. Jan 30 13:22:30.378629 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 30 13:22:30.378635 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jan 30 13:22:30.378642 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Jan 30 13:22:30.378648 kernel: GICv3: 960 SPIs implemented Jan 30 13:22:30.378655 kernel: GICv3: 0 Extended SPIs implemented Jan 30 13:22:30.378661 kernel: Root IRQ handler: gic_handle_irq Jan 30 13:22:30.378668 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Jan 30 13:22:30.378674 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Jan 30 13:22:30.378680 kernel: ITS: No ITS available, not enabling LPIs Jan 30 13:22:30.378687 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 30 13:22:30.378694 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jan 30 13:22:30.378700 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Jan 30 13:22:30.378709 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Jan 30 13:22:30.378716 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Jan 30 13:22:30.378722 kernel: Console: colour dummy device 80x25 Jan 30 13:22:30.378730 kernel: printk: console [tty1] enabled Jan 30 13:22:30.378736 kernel: ACPI: Core revision 20230628 Jan 30 13:22:30.378743 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Jan 30 13:22:30.378750 kernel: pid_max: default: 32768 minimum: 301 Jan 30 13:22:30.378757 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 30 13:22:30.378764 kernel: landlock: Up and running. Jan 30 13:22:30.378772 kernel: SELinux: Initializing. Jan 30 13:22:30.378779 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 30 13:22:30.378786 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 30 13:22:30.378793 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 30 13:22:30.378800 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 30 13:22:30.378807 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3a8030, hints 0xe, misc 0x31e1 Jan 30 13:22:30.378814 kernel: Hyper-V: Host Build 10.0.22477.1594-1-0 Jan 30 13:22:30.378829 kernel: Hyper-V: enabling crash_kexec_post_notifiers Jan 30 13:22:30.378836 kernel: rcu: Hierarchical SRCU implementation. Jan 30 13:22:30.378843 kernel: rcu: Max phase no-delay instances is 400. Jan 30 13:22:30.378864 kernel: Remapping and enabling EFI services. Jan 30 13:22:30.378871 kernel: smp: Bringing up secondary CPUs ... Jan 30 13:22:30.378880 kernel: Detected PIPT I-cache on CPU1 Jan 30 13:22:30.378887 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Jan 30 13:22:30.378895 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jan 30 13:22:30.378902 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Jan 30 13:22:30.378909 kernel: smp: Brought up 1 node, 2 CPUs Jan 30 13:22:30.378918 kernel: SMP: Total of 2 processors activated. Jan 30 13:22:30.378925 kernel: CPU features: detected: 32-bit EL0 Support Jan 30 13:22:30.378933 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Jan 30 13:22:30.378940 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Jan 30 13:22:30.378947 kernel: CPU features: detected: CRC32 instructions Jan 30 13:22:30.378954 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Jan 30 13:22:30.378961 kernel: CPU features: detected: LSE atomic instructions Jan 30 13:22:30.378968 kernel: CPU features: detected: Privileged Access Never Jan 30 13:22:30.378975 kernel: CPU: All CPU(s) started at EL1 Jan 30 13:22:30.378984 kernel: alternatives: applying system-wide alternatives Jan 30 13:22:30.378991 kernel: devtmpfs: initialized Jan 30 13:22:30.378999 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 30 13:22:30.379006 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jan 30 13:22:30.379013 kernel: pinctrl core: initialized pinctrl subsystem Jan 30 13:22:30.379020 kernel: SMBIOS 3.1.0 present. Jan 30 13:22:30.379027 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 09/28/2024 Jan 30 13:22:30.379034 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 30 13:22:30.379042 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Jan 30 13:22:30.379051 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Jan 30 13:22:30.379058 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Jan 30 13:22:30.379065 kernel: audit: initializing netlink subsys (disabled) Jan 30 13:22:30.379073 kernel: audit: type=2000 audit(0.047:1): state=initialized audit_enabled=0 res=1 Jan 30 13:22:30.379080 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 30 13:22:30.379087 kernel: cpuidle: using governor menu Jan 30 13:22:30.379094 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Jan 30 13:22:30.379101 kernel: ASID allocator initialised with 32768 entries Jan 30 13:22:30.379108 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 30 13:22:30.379117 kernel: Serial: AMBA PL011 UART driver Jan 30 13:22:30.379124 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Jan 30 13:22:30.379131 kernel: Modules: 0 pages in range for non-PLT usage Jan 30 13:22:30.379139 kernel: Modules: 508880 pages in range for PLT usage Jan 30 13:22:30.379146 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 30 13:22:30.379153 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Jan 30 13:22:30.379161 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Jan 30 13:22:30.379168 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Jan 30 13:22:30.379175 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 30 13:22:30.379197 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Jan 30 13:22:30.379205 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Jan 30 13:22:30.379212 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Jan 30 13:22:30.379219 kernel: ACPI: Added _OSI(Module Device) Jan 30 13:22:30.379226 kernel: ACPI: Added _OSI(Processor Device) Jan 30 13:22:30.379234 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jan 30 13:22:30.379241 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 30 13:22:30.379248 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 30 13:22:30.379255 kernel: ACPI: Interpreter enabled Jan 30 13:22:30.379264 kernel: ACPI: Using GIC for interrupt routing Jan 30 13:22:30.379271 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Jan 30 13:22:30.379278 kernel: printk: console [ttyAMA0] enabled Jan 30 13:22:30.379285 kernel: printk: bootconsole [pl11] disabled Jan 30 13:22:30.379293 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Jan 30 13:22:30.379300 kernel: iommu: Default domain type: Translated Jan 30 13:22:30.379307 kernel: iommu: DMA domain TLB invalidation policy: strict mode Jan 30 13:22:30.379314 kernel: efivars: Registered efivars operations Jan 30 13:22:30.379321 kernel: vgaarb: loaded Jan 30 13:22:30.379330 kernel: clocksource: Switched to clocksource arch_sys_counter Jan 30 13:22:30.379337 kernel: VFS: Disk quotas dquot_6.6.0 Jan 30 13:22:30.379344 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 30 13:22:30.379351 kernel: pnp: PnP ACPI init Jan 30 13:22:30.379359 kernel: pnp: PnP ACPI: found 0 devices Jan 30 13:22:30.379366 kernel: NET: Registered PF_INET protocol family Jan 30 13:22:30.379373 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 30 13:22:30.379381 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 30 13:22:30.379388 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 30 13:22:30.379397 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 30 13:22:30.379404 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 30 13:22:30.379412 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 30 13:22:30.379419 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 30 13:22:30.379426 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 30 13:22:30.379433 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 30 13:22:30.379440 kernel: PCI: CLS 0 bytes, default 64 Jan 30 13:22:30.379447 kernel: kvm [1]: HYP mode not available Jan 30 13:22:30.379454 kernel: Initialise system trusted keyrings Jan 30 13:22:30.379463 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 30 13:22:30.379470 kernel: Key type asymmetric registered Jan 30 13:22:30.379477 kernel: Asymmetric key parser 'x509' registered Jan 30 13:22:30.379484 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jan 30 13:22:30.379491 kernel: io scheduler mq-deadline registered Jan 30 13:22:30.379498 kernel: io scheduler kyber registered Jan 30 13:22:30.379505 kernel: io scheduler bfq registered Jan 30 13:22:30.379513 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 30 13:22:30.379521 kernel: thunder_xcv, ver 1.0 Jan 30 13:22:30.379529 kernel: thunder_bgx, ver 1.0 Jan 30 13:22:30.379536 kernel: nicpf, ver 1.0 Jan 30 13:22:30.379543 kernel: nicvf, ver 1.0 Jan 30 13:22:30.379712 kernel: rtc-efi rtc-efi.0: registered as rtc0 Jan 30 13:22:30.379787 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-01-30T13:22:29 UTC (1738243349) Jan 30 13:22:30.379797 kernel: efifb: probing for efifb Jan 30 13:22:30.379804 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Jan 30 13:22:30.379811 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Jan 30 13:22:30.379821 kernel: efifb: scrolling: redraw Jan 30 13:22:30.379828 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Jan 30 13:22:30.379835 kernel: Console: switching to colour frame buffer device 128x48 Jan 30 13:22:30.379843 kernel: fb0: EFI VGA frame buffer device Jan 30 13:22:30.379850 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Jan 30 13:22:30.379857 kernel: hid: raw HID events driver (C) Jiri Kosina Jan 30 13:22:30.379864 kernel: No ACPI PMU IRQ for CPU0 Jan 30 13:22:30.379871 kernel: No ACPI PMU IRQ for CPU1 Jan 30 13:22:30.379878 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 1 counters available Jan 30 13:22:30.379888 kernel: watchdog: Delayed init of the lockup detector failed: -19 Jan 30 13:22:30.379895 kernel: watchdog: Hard watchdog permanently disabled Jan 30 13:22:30.379902 kernel: NET: Registered PF_INET6 protocol family Jan 30 13:22:30.379909 kernel: Segment Routing with IPv6 Jan 30 13:22:30.379916 kernel: In-situ OAM (IOAM) with IPv6 Jan 30 13:22:30.379923 kernel: NET: Registered PF_PACKET protocol family Jan 30 13:22:30.379930 kernel: Key type dns_resolver registered Jan 30 13:22:30.379937 kernel: registered taskstats version 1 Jan 30 13:22:30.379944 kernel: Loading compiled-in X.509 certificates Jan 30 13:22:30.379953 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: c31663d2c680b3b306c17f44b5295280d3a2e28a' Jan 30 13:22:30.379960 kernel: Key type .fscrypt registered Jan 30 13:22:30.379967 kernel: Key type fscrypt-provisioning registered Jan 30 13:22:30.379974 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 30 13:22:30.379982 kernel: ima: Allocated hash algorithm: sha1 Jan 30 13:22:30.379989 kernel: ima: No architecture policies found Jan 30 13:22:30.379996 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Jan 30 13:22:30.380003 kernel: clk: Disabling unused clocks Jan 30 13:22:30.380010 kernel: Freeing unused kernel memory: 39936K Jan 30 13:22:30.380019 kernel: Run /init as init process Jan 30 13:22:30.380026 kernel: with arguments: Jan 30 13:22:30.380033 kernel: /init Jan 30 13:22:30.380040 kernel: with environment: Jan 30 13:22:30.380047 kernel: HOME=/ Jan 30 13:22:30.380054 kernel: TERM=linux Jan 30 13:22:30.380061 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jan 30 13:22:30.380070 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 30 13:22:30.380081 systemd[1]: Detected virtualization microsoft. Jan 30 13:22:30.380089 systemd[1]: Detected architecture arm64. Jan 30 13:22:30.380096 systemd[1]: Running in initrd. Jan 30 13:22:30.380104 systemd[1]: No hostname configured, using default hostname. Jan 30 13:22:30.380112 systemd[1]: Hostname set to . Jan 30 13:22:30.380120 systemd[1]: Initializing machine ID from random generator. Jan 30 13:22:30.380128 systemd[1]: Queued start job for default target initrd.target. Jan 30 13:22:30.380135 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 13:22:30.380145 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 13:22:30.380153 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 30 13:22:30.380161 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 30 13:22:30.380169 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 30 13:22:30.380177 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 30 13:22:30.382239 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 30 13:22:30.382263 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 30 13:22:30.382271 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 13:22:30.382282 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 30 13:22:30.382290 systemd[1]: Reached target paths.target - Path Units. Jan 30 13:22:30.382298 systemd[1]: Reached target slices.target - Slice Units. Jan 30 13:22:30.382306 systemd[1]: Reached target swap.target - Swaps. Jan 30 13:22:30.382313 systemd[1]: Reached target timers.target - Timer Units. Jan 30 13:22:30.382321 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 30 13:22:30.382331 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 30 13:22:30.382341 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 30 13:22:30.382349 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 30 13:22:30.382356 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 30 13:22:30.382364 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 30 13:22:30.382375 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 13:22:30.382382 systemd[1]: Reached target sockets.target - Socket Units. Jan 30 13:22:30.382390 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 30 13:22:30.382398 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 30 13:22:30.382408 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 30 13:22:30.382415 systemd[1]: Starting systemd-fsck-usr.service... Jan 30 13:22:30.382423 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 30 13:22:30.382431 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 30 13:22:30.382474 systemd-journald[218]: Collecting audit messages is disabled. Jan 30 13:22:30.382498 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:22:30.382507 systemd-journald[218]: Journal started Jan 30 13:22:30.382530 systemd-journald[218]: Runtime Journal (/run/log/journal/39447d4c53d84c759bf241a727efd868) is 8.0M, max 78.5M, 70.5M free. Jan 30 13:22:30.388994 systemd-modules-load[219]: Inserted module 'overlay' Jan 30 13:22:30.408523 systemd[1]: Started systemd-journald.service - Journal Service. Jan 30 13:22:30.420202 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 30 13:22:30.430696 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 30 13:22:30.445147 kernel: Bridge firewalling registered Jan 30 13:22:30.431331 systemd-modules-load[219]: Inserted module 'br_netfilter' Jan 30 13:22:30.439554 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 13:22:30.452670 systemd[1]: Finished systemd-fsck-usr.service. Jan 30 13:22:30.462222 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 30 13:22:30.475652 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:22:30.501552 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 13:22:30.516421 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 30 13:22:30.532799 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 30 13:22:30.549368 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 30 13:22:30.562218 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:22:30.584425 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:22:30.591518 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 13:22:30.617484 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 30 13:22:30.625402 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 30 13:22:30.639217 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 13:22:30.668202 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 30 13:22:30.686533 dracut-cmdline[252]: dracut-dracut-053 Jan 30 13:22:30.686533 dracut-cmdline[252]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=e6957044c3256d96283265c263579aa4275d1d707b02496fcb081f5fc6356346 Jan 30 13:22:30.690026 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 13:22:30.759019 systemd-resolved[259]: Positive Trust Anchors: Jan 30 13:22:30.759035 systemd-resolved[259]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 30 13:22:30.759066 systemd-resolved[259]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 30 13:22:30.765802 systemd-resolved[259]: Defaulting to hostname 'linux'. Jan 30 13:22:30.766852 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 30 13:22:30.774249 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 30 13:22:30.882210 kernel: SCSI subsystem initialized Jan 30 13:22:30.890218 kernel: Loading iSCSI transport class v2.0-870. Jan 30 13:22:30.902223 kernel: iscsi: registered transport (tcp) Jan 30 13:22:30.920249 kernel: iscsi: registered transport (qla4xxx) Jan 30 13:22:30.920270 kernel: QLogic iSCSI HBA Driver Jan 30 13:22:30.957578 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 30 13:22:30.976327 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 30 13:22:31.011159 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 30 13:22:31.011247 kernel: device-mapper: uevent: version 1.0.3 Jan 30 13:22:31.020203 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 30 13:22:31.069210 kernel: raid6: neonx8 gen() 15783 MB/s Jan 30 13:22:31.091193 kernel: raid6: neonx4 gen() 15820 MB/s Jan 30 13:22:31.111193 kernel: raid6: neonx2 gen() 13337 MB/s Jan 30 13:22:31.133194 kernel: raid6: neonx1 gen() 10448 MB/s Jan 30 13:22:31.154192 kernel: raid6: int64x8 gen() 6771 MB/s Jan 30 13:22:31.174192 kernel: raid6: int64x4 gen() 7353 MB/s Jan 30 13:22:31.195193 kernel: raid6: int64x2 gen() 6073 MB/s Jan 30 13:22:31.219375 kernel: raid6: int64x1 gen() 5053 MB/s Jan 30 13:22:31.219386 kernel: raid6: using algorithm neonx4 gen() 15820 MB/s Jan 30 13:22:31.245625 kernel: raid6: .... xor() 12286 MB/s, rmw enabled Jan 30 13:22:31.245638 kernel: raid6: using neon recovery algorithm Jan 30 13:22:31.256957 kernel: xor: measuring software checksum speed Jan 30 13:22:31.256970 kernel: 8regs : 21636 MB/sec Jan 30 13:22:31.264398 kernel: 32regs : 20393 MB/sec Jan 30 13:22:31.264409 kernel: arm64_neon : 27870 MB/sec Jan 30 13:22:31.269436 kernel: xor: using function: arm64_neon (27870 MB/sec) Jan 30 13:22:31.323237 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 30 13:22:31.336975 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 30 13:22:31.353341 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 13:22:31.376210 systemd-udevd[439]: Using default interface naming scheme 'v255'. Jan 30 13:22:31.381889 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 13:22:31.402330 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 30 13:22:31.435779 dracut-pre-trigger[448]: rd.md=0: removing MD RAID activation Jan 30 13:22:31.466709 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 30 13:22:31.486664 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 30 13:22:31.532126 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 13:22:31.551571 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 30 13:22:31.589384 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 30 13:22:31.602259 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 30 13:22:31.626169 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 13:22:31.649763 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 30 13:22:31.658907 kernel: hv_vmbus: Vmbus version:5.3 Jan 30 13:22:31.675009 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 30 13:22:31.739142 kernel: pps_core: LinuxPPS API ver. 1 registered Jan 30 13:22:31.739171 kernel: hv_vmbus: registering driver hyperv_keyboard Jan 30 13:22:31.739193 kernel: hv_vmbus: registering driver hid_hyperv Jan 30 13:22:31.739204 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Jan 30 13:22:31.739214 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Jan 30 13:22:31.739223 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Jan 30 13:22:31.739232 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Jan 30 13:22:31.739366 kernel: PTP clock support registered Jan 30 13:22:31.707088 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 30 13:22:31.761977 kernel: hv_utils: Registering HyperV Utility Driver Jan 30 13:22:31.762005 kernel: hv_vmbus: registering driver hv_utils Jan 30 13:22:31.707267 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:22:31.675439 kernel: hv_utils: Heartbeat IC version 3.0 Jan 30 13:22:31.689158 kernel: hv_utils: Shutdown IC version 3.2 Jan 30 13:22:31.689176 kernel: hv_vmbus: registering driver hv_netvsc Jan 30 13:22:31.689184 kernel: hv_utils: TimeSync IC version 4.0 Jan 30 13:22:31.689194 kernel: hv_vmbus: registering driver hv_storvsc Jan 30 13:22:31.689201 systemd-journald[218]: Time jumped backwards, rotating. Jan 30 13:22:31.689238 kernel: scsi host0: storvsc_host_t Jan 30 13:22:31.757868 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 13:22:31.710569 kernel: scsi host1: storvsc_host_t Jan 30 13:22:31.710784 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Jan 30 13:22:31.656776 systemd-resolved[259]: Clock change detected. Flushing caches. Jan 30 13:22:31.661779 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 13:22:31.738813 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Jan 30 13:22:31.662004 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:22:31.703437 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:22:31.784873 kernel: hv_netvsc 002248ba-7efb-0022-48ba-7efb002248ba eth0: VF slot 1 added Jan 30 13:22:31.785081 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Jan 30 13:22:31.830016 kernel: hv_vmbus: registering driver hv_pci Jan 30 13:22:31.830040 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 30 13:22:31.830049 kernel: hv_pci 7637483c-19d3-4e2d-94c2-e66e36793b85: PCI VMBus probing: Using version 0x10004 Jan 30 13:22:31.925586 kernel: hv_pci 7637483c-19d3-4e2d-94c2-e66e36793b85: PCI host bridge to bus 19d3:00 Jan 30 13:22:31.925710 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Jan 30 13:22:31.925859 kernel: pci_bus 19d3:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Jan 30 13:22:31.925972 kernel: pci_bus 19d3:00: No busn resource found for root bus, will use [bus 00-ff] Jan 30 13:22:31.926083 kernel: pci 19d3:00:02.0: [15b3:1018] type 00 class 0x020000 Jan 30 13:22:31.926237 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Jan 30 13:22:31.926800 kernel: pci 19d3:00:02.0: reg 0x10: [mem 0xfc0000000-0xfc00fffff 64bit pref] Jan 30 13:22:31.926928 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Jan 30 13:22:31.927028 kernel: sd 0:0:0:0: [sda] Write Protect is off Jan 30 13:22:31.927121 kernel: pci 19d3:00:02.0: enabling Extended Tags Jan 30 13:22:31.927218 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Jan 30 13:22:31.927306 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Jan 30 13:22:31.927390 kernel: pci 19d3:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 19d3:00:02.0 (capable of 126.016 Gb/s with 8.0 GT/s PCIe x16 link) Jan 30 13:22:31.927481 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 13:22:31.927490 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Jan 30 13:22:31.927573 kernel: pci_bus 19d3:00: busn_res: [bus 00-ff] end is updated to 00 Jan 30 13:22:31.927657 kernel: pci 19d3:00:02.0: BAR 0: assigned [mem 0xfc0000000-0xfc00fffff 64bit pref] Jan 30 13:22:31.739178 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:22:31.755430 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 30 13:22:31.815878 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:22:31.831890 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 13:22:31.937583 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:22:31.989750 kernel: mlx5_core 19d3:00:02.0: enabling device (0000 -> 0002) Jan 30 13:22:32.221805 kernel: mlx5_core 19d3:00:02.0: firmware version: 16.30.1284 Jan 30 13:22:32.221961 kernel: hv_netvsc 002248ba-7efb-0022-48ba-7efb002248ba eth0: VF registering: eth1 Jan 30 13:22:32.222063 kernel: mlx5_core 19d3:00:02.0 eth1: joined to eth0 Jan 30 13:22:32.222158 kernel: mlx5_core 19d3:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Jan 30 13:22:32.230749 kernel: mlx5_core 19d3:00:02.0 enP6611s1: renamed from eth1 Jan 30 13:22:32.300557 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Jan 30 13:22:32.468762 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by (udev-worker) (502) Jan 30 13:22:32.484082 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Jan 30 13:22:32.505511 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Jan 30 13:22:32.590740 kernel: BTRFS: device fsid 1e2e5fa7-c757-4d5d-af66-73afe98fbaae devid 1 transid 39 /dev/sda3 scanned by (udev-worker) (482) Jan 30 13:22:32.605098 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Jan 30 13:22:32.612579 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Jan 30 13:22:32.645973 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 30 13:22:32.669803 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 13:22:33.682776 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 13:22:33.682833 disk-uuid[599]: The operation has completed successfully. Jan 30 13:22:33.752803 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 30 13:22:33.754751 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 30 13:22:33.800999 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 30 13:22:33.815537 sh[685]: Success Jan 30 13:22:33.835093 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Jan 30 13:22:34.154643 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 30 13:22:34.180908 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 30 13:22:34.192319 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 30 13:22:34.227091 kernel: BTRFS info (device dm-0): first mount of filesystem 1e2e5fa7-c757-4d5d-af66-73afe98fbaae Jan 30 13:22:34.227143 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Jan 30 13:22:34.234392 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 30 13:22:34.240215 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 30 13:22:34.245870 kernel: BTRFS info (device dm-0): using free space tree Jan 30 13:22:34.795772 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 30 13:22:34.801908 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 30 13:22:34.823025 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 30 13:22:34.831837 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 30 13:22:34.873428 kernel: BTRFS info (device sda6): first mount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:34.873497 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Jan 30 13:22:34.878118 kernel: BTRFS info (device sda6): using free space tree Jan 30 13:22:34.913459 kernel: BTRFS info (device sda6): auto enabling async discard Jan 30 13:22:34.930050 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 30 13:22:34.935396 kernel: BTRFS info (device sda6): last unmount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:34.943217 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 30 13:22:34.958029 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 30 13:22:34.965146 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 30 13:22:34.988548 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 30 13:22:35.022795 systemd-networkd[869]: lo: Link UP Jan 30 13:22:35.026277 systemd-networkd[869]: lo: Gained carrier Jan 30 13:22:35.028448 systemd-networkd[869]: Enumeration completed Jan 30 13:22:35.031003 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 30 13:22:35.031347 systemd-networkd[869]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:22:35.031350 systemd-networkd[869]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 13:22:35.037803 systemd[1]: Reached target network.target - Network. Jan 30 13:22:35.129743 kernel: mlx5_core 19d3:00:02.0 enP6611s1: Link up Jan 30 13:22:35.168743 kernel: hv_netvsc 002248ba-7efb-0022-48ba-7efb002248ba eth0: Data path switched to VF: enP6611s1 Jan 30 13:22:35.169244 systemd-networkd[869]: enP6611s1: Link UP Jan 30 13:22:35.169434 systemd-networkd[869]: eth0: Link UP Jan 30 13:22:35.169542 systemd-networkd[869]: eth0: Gained carrier Jan 30 13:22:35.169553 systemd-networkd[869]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:22:35.184269 systemd-networkd[869]: enP6611s1: Gained carrier Jan 30 13:22:35.205767 systemd-networkd[869]: eth0: DHCPv4 address 10.200.20.40/24, gateway 10.200.20.1 acquired from 168.63.129.16 Jan 30 13:22:36.325115 ignition[864]: Ignition 2.20.0 Jan 30 13:22:36.325127 ignition[864]: Stage: fetch-offline Jan 30 13:22:36.325170 ignition[864]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:36.335104 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 30 13:22:36.325178 ignition[864]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:36.325271 ignition[864]: parsed url from cmdline: "" Jan 30 13:22:36.325274 ignition[864]: no config URL provided Jan 30 13:22:36.325279 ignition[864]: reading system config file "/usr/lib/ignition/user.ign" Jan 30 13:22:36.325285 ignition[864]: no config at "/usr/lib/ignition/user.ign" Jan 30 13:22:36.363075 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jan 30 13:22:36.325291 ignition[864]: failed to fetch config: resource requires networking Jan 30 13:22:36.325486 ignition[864]: Ignition finished successfully Jan 30 13:22:36.386403 ignition[880]: Ignition 2.20.0 Jan 30 13:22:36.386410 ignition[880]: Stage: fetch Jan 30 13:22:36.386579 ignition[880]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:36.386588 ignition[880]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:36.386677 ignition[880]: parsed url from cmdline: "" Jan 30 13:22:36.386680 ignition[880]: no config URL provided Jan 30 13:22:36.386684 ignition[880]: reading system config file "/usr/lib/ignition/user.ign" Jan 30 13:22:36.386692 ignition[880]: no config at "/usr/lib/ignition/user.ign" Jan 30 13:22:36.386732 ignition[880]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Jan 30 13:22:36.494907 ignition[880]: GET result: OK Jan 30 13:22:36.494977 ignition[880]: config has been read from IMDS userdata Jan 30 13:22:36.495016 ignition[880]: parsing config with SHA512: 6f25ae24839f5dd6d1be3ca08131bb21cfa0c0c8a38bcc2ebc6f91db920acf07ad49d964bbf6f8cfeabc74dac631371b91a23873884712feff26df55f11a995a Jan 30 13:22:36.500032 unknown[880]: fetched base config from "system" Jan 30 13:22:36.500534 ignition[880]: fetch: fetch complete Jan 30 13:22:36.500039 unknown[880]: fetched base config from "system" Jan 30 13:22:36.500539 ignition[880]: fetch: fetch passed Jan 30 13:22:36.500047 unknown[880]: fetched user config from "azure" Jan 30 13:22:36.500603 ignition[880]: Ignition finished successfully Jan 30 13:22:36.503184 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jan 30 13:22:36.526895 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 30 13:22:36.556365 ignition[886]: Ignition 2.20.0 Jan 30 13:22:36.556393 ignition[886]: Stage: kargs Jan 30 13:22:36.561674 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 30 13:22:36.556615 ignition[886]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:36.556625 ignition[886]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:36.577033 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 30 13:22:36.557699 ignition[886]: kargs: kargs passed Jan 30 13:22:36.557767 ignition[886]: Ignition finished successfully Jan 30 13:22:36.604023 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 30 13:22:36.600626 ignition[892]: Ignition 2.20.0 Jan 30 13:22:36.610582 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 30 13:22:36.600633 ignition[892]: Stage: disks Jan 30 13:22:36.620837 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 30 13:22:36.600828 ignition[892]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:36.632763 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 30 13:22:36.600838 ignition[892]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:36.641809 systemd[1]: Reached target sysinit.target - System Initialization. Jan 30 13:22:36.601803 ignition[892]: disks: disks passed Jan 30 13:22:36.653631 systemd[1]: Reached target basic.target - Basic System. Jan 30 13:22:36.601857 ignition[892]: Ignition finished successfully Jan 30 13:22:36.669984 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 30 13:22:36.756400 systemd-fsck[901]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Jan 30 13:22:36.767423 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 30 13:22:36.787925 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 30 13:22:36.805921 systemd-networkd[869]: eth0: Gained IPv6LL Jan 30 13:22:36.851581 kernel: EXT4-fs (sda9): mounted filesystem 88903c49-366d-43ff-90b1-141790b6e85c r/w with ordered data mode. Quota mode: none. Jan 30 13:22:36.852127 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 30 13:22:36.861463 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 30 13:22:36.917823 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 30 13:22:36.929120 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 30 13:22:36.941759 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Jan 30 13:22:36.966218 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (912) Jan 30 13:22:36.958879 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 30 13:22:36.958923 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 30 13:22:37.024132 kernel: BTRFS info (device sda6): first mount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:37.024160 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Jan 30 13:22:37.024170 kernel: BTRFS info (device sda6): using free space tree Jan 30 13:22:37.024179 kernel: BTRFS info (device sda6): auto enabling async discard Jan 30 13:22:36.975005 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 30 13:22:37.015000 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 30 13:22:37.031452 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 30 13:22:37.061815 systemd-networkd[869]: enP6611s1: Gained IPv6LL Jan 30 13:22:37.738997 coreos-metadata[914]: Jan 30 13:22:37.738 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Jan 30 13:22:37.748362 coreos-metadata[914]: Jan 30 13:22:37.748 INFO Fetch successful Jan 30 13:22:37.748362 coreos-metadata[914]: Jan 30 13:22:37.748 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Jan 30 13:22:37.769015 coreos-metadata[914]: Jan 30 13:22:37.768 INFO Fetch successful Jan 30 13:22:37.775246 coreos-metadata[914]: Jan 30 13:22:37.774 INFO wrote hostname ci-4186.1.0-a-a7402f3063 to /sysroot/etc/hostname Jan 30 13:22:37.776803 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 30 13:22:38.124225 initrd-setup-root[943]: cut: /sysroot/etc/passwd: No such file or directory Jan 30 13:22:38.231683 initrd-setup-root[950]: cut: /sysroot/etc/group: No such file or directory Jan 30 13:22:38.241704 initrd-setup-root[957]: cut: /sysroot/etc/shadow: No such file or directory Jan 30 13:22:38.251500 initrd-setup-root[964]: cut: /sysroot/etc/gshadow: No such file or directory Jan 30 13:22:39.725379 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 30 13:22:39.744861 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 30 13:22:39.760966 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 30 13:22:39.780020 kernel: BTRFS info (device sda6): last unmount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:39.775970 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 30 13:22:39.804916 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 30 13:22:39.818761 ignition[1032]: INFO : Ignition 2.20.0 Jan 30 13:22:39.818761 ignition[1032]: INFO : Stage: mount Jan 30 13:22:39.818761 ignition[1032]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:39.818761 ignition[1032]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:39.848195 ignition[1032]: INFO : mount: mount passed Jan 30 13:22:39.848195 ignition[1032]: INFO : Ignition finished successfully Jan 30 13:22:39.833687 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 30 13:22:39.858996 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 30 13:22:39.878026 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 30 13:22:39.910732 kernel: BTRFS: device label OEM devid 1 transid 17 /dev/sda6 scanned by mount (1044) Jan 30 13:22:39.910785 kernel: BTRFS info (device sda6): first mount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:39.917032 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Jan 30 13:22:39.921965 kernel: BTRFS info (device sda6): using free space tree Jan 30 13:22:39.928944 kernel: BTRFS info (device sda6): auto enabling async discard Jan 30 13:22:39.930411 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 30 13:22:39.958024 ignition[1061]: INFO : Ignition 2.20.0 Jan 30 13:22:39.958024 ignition[1061]: INFO : Stage: files Jan 30 13:22:39.967498 ignition[1061]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:39.967498 ignition[1061]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:39.967498 ignition[1061]: DEBUG : files: compiled without relabeling support, skipping Jan 30 13:22:39.988825 ignition[1061]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 30 13:22:39.988825 ignition[1061]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 30 13:22:40.238396 ignition[1061]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 30 13:22:40.246866 ignition[1061]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 30 13:22:40.246866 ignition[1061]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 30 13:22:40.241298 unknown[1061]: wrote ssh authorized keys file for user: core Jan 30 13:22:40.317406 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jan 30 13:22:40.328599 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Jan 30 13:22:40.400812 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 30 13:22:40.527784 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jan 30 13:22:40.527784 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 Jan 30 13:22:40.993294 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 30 13:22:41.205081 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Jan 30 13:22:41.205081 ignition[1061]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 30 13:22:41.281174 ignition[1061]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: files passed Jan 30 13:22:41.294108 ignition[1061]: INFO : Ignition finished successfully Jan 30 13:22:41.294524 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 30 13:22:41.340024 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 30 13:22:41.359973 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 30 13:22:41.383415 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 30 13:22:41.383514 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 30 13:22:41.436126 initrd-setup-root-after-ignition[1090]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 30 13:22:41.436126 initrd-setup-root-after-ignition[1090]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 30 13:22:41.457572 initrd-setup-root-after-ignition[1094]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 30 13:22:41.449073 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 30 13:22:41.465673 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 30 13:22:41.500009 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 30 13:22:41.530082 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 30 13:22:41.530958 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 30 13:22:41.544377 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 30 13:22:41.557381 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 30 13:22:41.569507 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 30 13:22:41.572967 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 30 13:22:41.607956 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 30 13:22:41.625965 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 30 13:22:41.646177 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 30 13:22:41.653708 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 13:22:41.668791 systemd[1]: Stopped target timers.target - Timer Units. Jan 30 13:22:41.681030 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 30 13:22:41.681206 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 30 13:22:41.699876 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 30 13:22:41.714020 systemd[1]: Stopped target basic.target - Basic System. Jan 30 13:22:41.725611 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 30 13:22:41.738020 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 30 13:22:41.751600 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 30 13:22:41.765469 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 30 13:22:41.779552 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 30 13:22:41.793667 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 30 13:22:41.808137 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 30 13:22:41.821267 systemd[1]: Stopped target swap.target - Swaps. Jan 30 13:22:41.832374 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 30 13:22:41.832556 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 30 13:22:41.849838 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 30 13:22:41.862879 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 13:22:41.876325 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 30 13:22:41.876437 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 13:22:41.892274 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 30 13:22:41.892452 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 30 13:22:41.913132 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 30 13:22:41.913324 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 30 13:22:41.926898 systemd[1]: ignition-files.service: Deactivated successfully. Jan 30 13:22:41.927073 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 30 13:22:41.939606 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Jan 30 13:22:41.939805 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 30 13:22:41.975894 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 30 13:22:42.017088 ignition[1114]: INFO : Ignition 2.20.0 Jan 30 13:22:42.017088 ignition[1114]: INFO : Stage: umount Jan 30 13:22:42.017088 ignition[1114]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:42.017088 ignition[1114]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:42.017088 ignition[1114]: INFO : umount: umount passed Jan 30 13:22:42.017088 ignition[1114]: INFO : Ignition finished successfully Jan 30 13:22:41.994983 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 30 13:22:42.008276 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 30 13:22:42.008470 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 13:22:42.030298 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 30 13:22:42.030448 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 30 13:22:42.051280 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 30 13:22:42.051385 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 30 13:22:42.061955 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 30 13:22:42.062672 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 30 13:22:42.063823 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 30 13:22:42.082043 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 30 13:22:42.082123 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 30 13:22:42.093627 systemd[1]: ignition-fetch.service: Deactivated successfully. Jan 30 13:22:42.093687 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jan 30 13:22:42.105370 systemd[1]: Stopped target network.target - Network. Jan 30 13:22:42.111506 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 30 13:22:42.111589 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 30 13:22:42.125816 systemd[1]: Stopped target paths.target - Path Units. Jan 30 13:22:42.137775 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 30 13:22:42.141786 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 13:22:42.151402 systemd[1]: Stopped target slices.target - Slice Units. Jan 30 13:22:42.170066 systemd[1]: Stopped target sockets.target - Socket Units. Jan 30 13:22:42.182070 systemd[1]: iscsid.socket: Deactivated successfully. Jan 30 13:22:42.182174 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 30 13:22:42.193282 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 30 13:22:42.193335 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 30 13:22:42.204527 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 30 13:22:42.204582 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 30 13:22:42.215998 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 30 13:22:42.216045 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 30 13:22:42.228135 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 30 13:22:42.471100 kernel: hv_netvsc 002248ba-7efb-0022-48ba-7efb002248ba eth0: Data path switched from VF: enP6611s1 Jan 30 13:22:42.238959 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 30 13:22:42.251048 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 30 13:22:42.251153 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 30 13:22:42.255827 systemd-networkd[869]: eth0: DHCPv6 lease lost Jan 30 13:22:42.265942 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 30 13:22:42.266057 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 30 13:22:42.276013 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 30 13:22:42.276092 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 30 13:22:42.311275 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 30 13:22:42.324422 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 30 13:22:42.324516 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 30 13:22:42.340592 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 13:22:42.357259 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 30 13:22:42.357377 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 30 13:22:42.387075 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 30 13:22:42.387270 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:22:42.400598 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 30 13:22:42.400668 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 30 13:22:42.414293 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 30 13:22:42.414364 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 13:22:42.429312 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 30 13:22:42.429466 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 13:22:42.452196 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 30 13:22:42.452274 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 30 13:22:42.464585 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 30 13:22:42.464638 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 13:22:42.477663 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 30 13:22:42.477772 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 30 13:22:42.496678 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 30 13:22:42.496775 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 30 13:22:42.515386 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 30 13:22:42.515469 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:22:42.557015 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 30 13:22:42.573795 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 30 13:22:42.573876 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 13:22:42.590189 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jan 30 13:22:42.590249 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 13:22:42.605245 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 30 13:22:42.605298 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 13:22:42.618389 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 13:22:42.618442 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:22:42.631120 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 30 13:22:42.631243 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 30 13:22:42.643134 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 30 13:22:42.643227 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 30 13:22:44.664021 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 30 13:22:44.664145 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 30 13:22:44.670342 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 30 13:22:44.682215 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 30 13:22:44.682294 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 30 13:22:44.713058 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 30 13:22:44.727184 systemd[1]: Switching root. Jan 30 13:22:44.972288 systemd-journald[218]: Journal stopped Jan 30 13:22:30.375923 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Jan 30 13:22:30.375951 kernel: Linux version 6.6.74-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 14.2.1_p20241116 p3) 14.2.1 20241116, GNU ld (Gentoo 2.42 p6) 2.42.0) #1 SMP PREEMPT Wed Jan 29 09:30:22 -00 2025 Jan 30 13:22:30.375961 kernel: KASLR enabled Jan 30 13:22:30.375967 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Jan 30 13:22:30.375976 kernel: printk: bootconsole [pl11] enabled Jan 30 13:22:30.375982 kernel: efi: EFI v2.7 by EDK II Jan 30 13:22:30.375991 kernel: efi: ACPI 2.0=0x3fd5f018 SMBIOS=0x3e580000 SMBIOS 3.0=0x3e560000 MEMATTR=0x3ead5018 RNG=0x3fd5f998 MEMRESERVE=0x3e477598 Jan 30 13:22:30.375997 kernel: random: crng init done Jan 30 13:22:30.376003 kernel: secureboot: Secure boot disabled Jan 30 13:22:30.376009 kernel: ACPI: Early table checksum verification disabled Jan 30 13:22:30.376015 kernel: ACPI: RSDP 0x000000003FD5F018 000024 (v02 VRTUAL) Jan 30 13:22:30.376020 kernel: ACPI: XSDT 0x000000003FD5FF18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376026 kernel: ACPI: FACP 0x000000003FD5FC18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376034 kernel: ACPI: DSDT 0x000000003FD41018 01DFCD (v02 MSFTVM DSDT01 00000001 INTL 20230628) Jan 30 13:22:30.376041 kernel: ACPI: DBG2 0x000000003FD5FB18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376047 kernel: ACPI: GTDT 0x000000003FD5FD98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376053 kernel: ACPI: OEM0 0x000000003FD5F098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376062 kernel: ACPI: SPCR 0x000000003FD5FA98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376068 kernel: ACPI: APIC 0x000000003FD5F818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376074 kernel: ACPI: SRAT 0x000000003FD5F198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376081 kernel: ACPI: PPTT 0x000000003FD5F418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Jan 30 13:22:30.376087 kernel: ACPI: BGRT 0x000000003FD5FE98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Jan 30 13:22:30.376093 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Jan 30 13:22:30.376100 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x00000000-0x3fffffff] Jan 30 13:22:30.376106 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000-0x1bfffffff] Jan 30 13:22:30.376114 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1c0000000-0xfbfffffff] Jan 30 13:22:30.376120 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x1000000000-0xffffffffff] Jan 30 13:22:30.376128 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x10000000000-0x1ffffffffff] Jan 30 13:22:30.376136 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x20000000000-0x3ffffffffff] Jan 30 13:22:30.376143 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x40000000000-0x7ffffffffff] Jan 30 13:22:30.376149 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x80000000000-0xfffffffffff] Jan 30 13:22:30.376155 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x100000000000-0x1fffffffffff] Jan 30 13:22:30.376162 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x200000000000-0x3fffffffffff] Jan 30 13:22:30.376169 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x400000000000-0x7fffffffffff] Jan 30 13:22:30.376175 kernel: ACPI: SRAT: Node 0 PXM 0 [mem 0x800000000000-0xffffffffffff] Jan 30 13:22:30.378223 kernel: NUMA: NODE_DATA [mem 0x1bf7ef800-0x1bf7f4fff] Jan 30 13:22:30.378249 kernel: Zone ranges: Jan 30 13:22:30.378256 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Jan 30 13:22:30.378263 kernel: DMA32 empty Jan 30 13:22:30.378270 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Jan 30 13:22:30.378287 kernel: Movable zone start for each node Jan 30 13:22:30.378293 kernel: Early memory node ranges Jan 30 13:22:30.378300 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Jan 30 13:22:30.378307 kernel: node 0: [mem 0x0000000000824000-0x000000003e45ffff] Jan 30 13:22:30.378313 kernel: node 0: [mem 0x000000003e460000-0x000000003e46ffff] Jan 30 13:22:30.378322 kernel: node 0: [mem 0x000000003e470000-0x000000003e54ffff] Jan 30 13:22:30.378328 kernel: node 0: [mem 0x000000003e550000-0x000000003e87ffff] Jan 30 13:22:30.378335 kernel: node 0: [mem 0x000000003e880000-0x000000003fc7ffff] Jan 30 13:22:30.378341 kernel: node 0: [mem 0x000000003fc80000-0x000000003fcfffff] Jan 30 13:22:30.378348 kernel: node 0: [mem 0x000000003fd00000-0x000000003fffffff] Jan 30 13:22:30.378354 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Jan 30 13:22:30.378361 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Jan 30 13:22:30.378368 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Jan 30 13:22:30.378375 kernel: psci: probing for conduit method from ACPI. Jan 30 13:22:30.378381 kernel: psci: PSCIv1.1 detected in firmware. Jan 30 13:22:30.378388 kernel: psci: Using standard PSCI v0.2 function IDs Jan 30 13:22:30.378395 kernel: psci: MIGRATE_INFO_TYPE not supported. Jan 30 13:22:30.378403 kernel: psci: SMC Calling Convention v1.4 Jan 30 13:22:30.378410 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node 0 Jan 30 13:22:30.378416 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node 0 Jan 30 13:22:30.378423 kernel: percpu: Embedded 31 pages/cpu s86696 r8192 d32088 u126976 Jan 30 13:22:30.378429 kernel: pcpu-alloc: s86696 r8192 d32088 u126976 alloc=31*4096 Jan 30 13:22:30.378436 kernel: pcpu-alloc: [0] 0 [0] 1 Jan 30 13:22:30.378443 kernel: Detected PIPT I-cache on CPU0 Jan 30 13:22:30.378449 kernel: CPU features: detected: GIC system register CPU interface Jan 30 13:22:30.378456 kernel: CPU features: detected: Hardware dirty bit management Jan 30 13:22:30.378462 kernel: CPU features: detected: Spectre-BHB Jan 30 13:22:30.378469 kernel: CPU features: kernel page table isolation forced ON by KASLR Jan 30 13:22:30.378477 kernel: CPU features: detected: Kernel page table isolation (KPTI) Jan 30 13:22:30.378484 kernel: CPU features: detected: ARM erratum 1418040 Jan 30 13:22:30.378490 kernel: CPU features: detected: ARM erratum 1542419 (kernel portion) Jan 30 13:22:30.378497 kernel: CPU features: detected: SSBS not fully self-synchronizing Jan 30 13:22:30.378503 kernel: alternatives: applying boot alternatives Jan 30 13:22:30.378511 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=e6957044c3256d96283265c263579aa4275d1d707b02496fcb081f5fc6356346 Jan 30 13:22:30.378518 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Jan 30 13:22:30.378525 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Jan 30 13:22:30.378532 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Jan 30 13:22:30.378538 kernel: Fallback order for Node 0: 0 Jan 30 13:22:30.378545 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1032156 Jan 30 13:22:30.378553 kernel: Policy zone: Normal Jan 30 13:22:30.378559 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Jan 30 13:22:30.378566 kernel: software IO TLB: area num 2. Jan 30 13:22:30.378572 kernel: software IO TLB: mapped [mem 0x000000003a460000-0x000000003e460000] (64MB) Jan 30 13:22:30.378579 kernel: Memory: 3982056K/4194160K available (10304K kernel code, 2186K rwdata, 8092K rodata, 39936K init, 897K bss, 212104K reserved, 0K cma-reserved) Jan 30 13:22:30.378586 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Jan 30 13:22:30.378592 kernel: rcu: Preemptible hierarchical RCU implementation. Jan 30 13:22:30.378600 kernel: rcu: RCU event tracing is enabled. Jan 30 13:22:30.378607 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Jan 30 13:22:30.378614 kernel: Trampoline variant of Tasks RCU enabled. Jan 30 13:22:30.378620 kernel: Tracing variant of Tasks RCU enabled. Jan 30 13:22:30.378629 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Jan 30 13:22:30.378635 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Jan 30 13:22:30.378642 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Jan 30 13:22:30.378648 kernel: GICv3: 960 SPIs implemented Jan 30 13:22:30.378655 kernel: GICv3: 0 Extended SPIs implemented Jan 30 13:22:30.378661 kernel: Root IRQ handler: gic_handle_irq Jan 30 13:22:30.378668 kernel: GICv3: GICv3 features: 16 PPIs, DirectLPI Jan 30 13:22:30.378674 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Jan 30 13:22:30.378680 kernel: ITS: No ITS available, not enabling LPIs Jan 30 13:22:30.378687 kernel: rcu: srcu_init: Setting srcu_struct sizes based on contention. Jan 30 13:22:30.378694 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jan 30 13:22:30.378700 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Jan 30 13:22:30.378709 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Jan 30 13:22:30.378716 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Jan 30 13:22:30.378722 kernel: Console: colour dummy device 80x25 Jan 30 13:22:30.378730 kernel: printk: console [tty1] enabled Jan 30 13:22:30.378736 kernel: ACPI: Core revision 20230628 Jan 30 13:22:30.378743 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Jan 30 13:22:30.378750 kernel: pid_max: default: 32768 minimum: 301 Jan 30 13:22:30.378757 kernel: LSM: initializing lsm=lockdown,capability,landlock,selinux,integrity Jan 30 13:22:30.378764 kernel: landlock: Up and running. Jan 30 13:22:30.378772 kernel: SELinux: Initializing. Jan 30 13:22:30.378779 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 30 13:22:30.378786 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Jan 30 13:22:30.378793 kernel: RCU Tasks: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 30 13:22:30.378800 kernel: RCU Tasks Trace: Setting shift to 1 and lim to 1 rcu_task_cb_adjust=1 rcu_task_cpu_ids=2. Jan 30 13:22:30.378807 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3a8030, hints 0xe, misc 0x31e1 Jan 30 13:22:30.378814 kernel: Hyper-V: Host Build 10.0.22477.1594-1-0 Jan 30 13:22:30.378829 kernel: Hyper-V: enabling crash_kexec_post_notifiers Jan 30 13:22:30.378836 kernel: rcu: Hierarchical SRCU implementation. Jan 30 13:22:30.378843 kernel: rcu: Max phase no-delay instances is 400. Jan 30 13:22:30.378864 kernel: Remapping and enabling EFI services. Jan 30 13:22:30.378871 kernel: smp: Bringing up secondary CPUs ... Jan 30 13:22:30.378880 kernel: Detected PIPT I-cache on CPU1 Jan 30 13:22:30.378887 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Jan 30 13:22:30.378895 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Jan 30 13:22:30.378902 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Jan 30 13:22:30.378909 kernel: smp: Brought up 1 node, 2 CPUs Jan 30 13:22:30.378918 kernel: SMP: Total of 2 processors activated. Jan 30 13:22:30.378925 kernel: CPU features: detected: 32-bit EL0 Support Jan 30 13:22:30.378933 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Jan 30 13:22:30.378940 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Jan 30 13:22:30.378947 kernel: CPU features: detected: CRC32 instructions Jan 30 13:22:30.378954 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Jan 30 13:22:30.378961 kernel: CPU features: detected: LSE atomic instructions Jan 30 13:22:30.378968 kernel: CPU features: detected: Privileged Access Never Jan 30 13:22:30.378975 kernel: CPU: All CPU(s) started at EL1 Jan 30 13:22:30.378984 kernel: alternatives: applying system-wide alternatives Jan 30 13:22:30.378991 kernel: devtmpfs: initialized Jan 30 13:22:30.378999 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Jan 30 13:22:30.379006 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Jan 30 13:22:30.379013 kernel: pinctrl core: initialized pinctrl subsystem Jan 30 13:22:30.379020 kernel: SMBIOS 3.1.0 present. Jan 30 13:22:30.379027 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 09/28/2024 Jan 30 13:22:30.379034 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Jan 30 13:22:30.379042 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Jan 30 13:22:30.379051 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Jan 30 13:22:30.379058 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Jan 30 13:22:30.379065 kernel: audit: initializing netlink subsys (disabled) Jan 30 13:22:30.379073 kernel: audit: type=2000 audit(0.047:1): state=initialized audit_enabled=0 res=1 Jan 30 13:22:30.379080 kernel: thermal_sys: Registered thermal governor 'step_wise' Jan 30 13:22:30.379087 kernel: cpuidle: using governor menu Jan 30 13:22:30.379094 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Jan 30 13:22:30.379101 kernel: ASID allocator initialised with 32768 entries Jan 30 13:22:30.379108 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Jan 30 13:22:30.379117 kernel: Serial: AMBA PL011 UART driver Jan 30 13:22:30.379124 kernel: Modules: 2G module region forced by RANDOMIZE_MODULE_REGION_FULL Jan 30 13:22:30.379131 kernel: Modules: 0 pages in range for non-PLT usage Jan 30 13:22:30.379139 kernel: Modules: 508880 pages in range for PLT usage Jan 30 13:22:30.379146 kernel: HugeTLB: registered 1.00 GiB page size, pre-allocated 0 pages Jan 30 13:22:30.379153 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 1.00 GiB page Jan 30 13:22:30.379161 kernel: HugeTLB: registered 32.0 MiB page size, pre-allocated 0 pages Jan 30 13:22:30.379168 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 32.0 MiB page Jan 30 13:22:30.379175 kernel: HugeTLB: registered 2.00 MiB page size, pre-allocated 0 pages Jan 30 13:22:30.379197 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 2.00 MiB page Jan 30 13:22:30.379205 kernel: HugeTLB: registered 64.0 KiB page size, pre-allocated 0 pages Jan 30 13:22:30.379212 kernel: HugeTLB: 0 KiB vmemmap can be freed for a 64.0 KiB page Jan 30 13:22:30.379219 kernel: ACPI: Added _OSI(Module Device) Jan 30 13:22:30.379226 kernel: ACPI: Added _OSI(Processor Device) Jan 30 13:22:30.379234 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Jan 30 13:22:30.379241 kernel: ACPI: Added _OSI(Processor Aggregator Device) Jan 30 13:22:30.379248 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Jan 30 13:22:30.379255 kernel: ACPI: Interpreter enabled Jan 30 13:22:30.379264 kernel: ACPI: Using GIC for interrupt routing Jan 30 13:22:30.379271 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Jan 30 13:22:30.379278 kernel: printk: console [ttyAMA0] enabled Jan 30 13:22:30.379285 kernel: printk: bootconsole [pl11] disabled Jan 30 13:22:30.379293 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Jan 30 13:22:30.379300 kernel: iommu: Default domain type: Translated Jan 30 13:22:30.379307 kernel: iommu: DMA domain TLB invalidation policy: strict mode Jan 30 13:22:30.379314 kernel: efivars: Registered efivars operations Jan 30 13:22:30.379321 kernel: vgaarb: loaded Jan 30 13:22:30.379330 kernel: clocksource: Switched to clocksource arch_sys_counter Jan 30 13:22:30.379337 kernel: VFS: Disk quotas dquot_6.6.0 Jan 30 13:22:30.379344 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Jan 30 13:22:30.379351 kernel: pnp: PnP ACPI init Jan 30 13:22:30.379359 kernel: pnp: PnP ACPI: found 0 devices Jan 30 13:22:30.379366 kernel: NET: Registered PF_INET protocol family Jan 30 13:22:30.379373 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Jan 30 13:22:30.379381 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Jan 30 13:22:30.379388 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Jan 30 13:22:30.379397 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Jan 30 13:22:30.379404 kernel: TCP bind hash table entries: 32768 (order: 8, 1048576 bytes, linear) Jan 30 13:22:30.379412 kernel: TCP: Hash tables configured (established 32768 bind 32768) Jan 30 13:22:30.379419 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 30 13:22:30.379426 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Jan 30 13:22:30.379433 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Jan 30 13:22:30.379440 kernel: PCI: CLS 0 bytes, default 64 Jan 30 13:22:30.379447 kernel: kvm [1]: HYP mode not available Jan 30 13:22:30.379454 kernel: Initialise system trusted keyrings Jan 30 13:22:30.379463 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Jan 30 13:22:30.379470 kernel: Key type asymmetric registered Jan 30 13:22:30.379477 kernel: Asymmetric key parser 'x509' registered Jan 30 13:22:30.379484 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 250) Jan 30 13:22:30.379491 kernel: io scheduler mq-deadline registered Jan 30 13:22:30.379498 kernel: io scheduler kyber registered Jan 30 13:22:30.379505 kernel: io scheduler bfq registered Jan 30 13:22:30.379513 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Jan 30 13:22:30.379521 kernel: thunder_xcv, ver 1.0 Jan 30 13:22:30.379529 kernel: thunder_bgx, ver 1.0 Jan 30 13:22:30.379536 kernel: nicpf, ver 1.0 Jan 30 13:22:30.379543 kernel: nicvf, ver 1.0 Jan 30 13:22:30.379712 kernel: rtc-efi rtc-efi.0: registered as rtc0 Jan 30 13:22:30.379787 kernel: rtc-efi rtc-efi.0: setting system clock to 2025-01-30T13:22:29 UTC (1738243349) Jan 30 13:22:30.379797 kernel: efifb: probing for efifb Jan 30 13:22:30.379804 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Jan 30 13:22:30.379811 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Jan 30 13:22:30.379821 kernel: efifb: scrolling: redraw Jan 30 13:22:30.379828 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Jan 30 13:22:30.379835 kernel: Console: switching to colour frame buffer device 128x48 Jan 30 13:22:30.379843 kernel: fb0: EFI VGA frame buffer device Jan 30 13:22:30.379850 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Jan 30 13:22:30.379857 kernel: hid: raw HID events driver (C) Jiri Kosina Jan 30 13:22:30.379864 kernel: No ACPI PMU IRQ for CPU0 Jan 30 13:22:30.379871 kernel: No ACPI PMU IRQ for CPU1 Jan 30 13:22:30.379878 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 1 counters available Jan 30 13:22:30.379888 kernel: watchdog: Delayed init of the lockup detector failed: -19 Jan 30 13:22:30.379895 kernel: watchdog: Hard watchdog permanently disabled Jan 30 13:22:30.379902 kernel: NET: Registered PF_INET6 protocol family Jan 30 13:22:30.379909 kernel: Segment Routing with IPv6 Jan 30 13:22:30.379916 kernel: In-situ OAM (IOAM) with IPv6 Jan 30 13:22:30.379923 kernel: NET: Registered PF_PACKET protocol family Jan 30 13:22:30.379930 kernel: Key type dns_resolver registered Jan 30 13:22:30.379937 kernel: registered taskstats version 1 Jan 30 13:22:30.379944 kernel: Loading compiled-in X.509 certificates Jan 30 13:22:30.379953 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 6.6.74-flatcar: c31663d2c680b3b306c17f44b5295280d3a2e28a' Jan 30 13:22:30.379960 kernel: Key type .fscrypt registered Jan 30 13:22:30.379967 kernel: Key type fscrypt-provisioning registered Jan 30 13:22:30.379974 kernel: ima: No TPM chip found, activating TPM-bypass! Jan 30 13:22:30.379982 kernel: ima: Allocated hash algorithm: sha1 Jan 30 13:22:30.379989 kernel: ima: No architecture policies found Jan 30 13:22:30.379996 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Jan 30 13:22:30.380003 kernel: clk: Disabling unused clocks Jan 30 13:22:30.380010 kernel: Freeing unused kernel memory: 39936K Jan 30 13:22:30.380019 kernel: Run /init as init process Jan 30 13:22:30.380026 kernel: with arguments: Jan 30 13:22:30.380033 kernel: /init Jan 30 13:22:30.380040 kernel: with environment: Jan 30 13:22:30.380047 kernel: HOME=/ Jan 30 13:22:30.380054 kernel: TERM=linux Jan 30 13:22:30.380061 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Jan 30 13:22:30.380070 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 30 13:22:30.380081 systemd[1]: Detected virtualization microsoft. Jan 30 13:22:30.380089 systemd[1]: Detected architecture arm64. Jan 30 13:22:30.380096 systemd[1]: Running in initrd. Jan 30 13:22:30.380104 systemd[1]: No hostname configured, using default hostname. Jan 30 13:22:30.380112 systemd[1]: Hostname set to . Jan 30 13:22:30.380120 systemd[1]: Initializing machine ID from random generator. Jan 30 13:22:30.380128 systemd[1]: Queued start job for default target initrd.target. Jan 30 13:22:30.380135 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 13:22:30.380145 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 13:22:30.380153 systemd[1]: Expecting device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - /dev/disk/by-label/EFI-SYSTEM... Jan 30 13:22:30.380161 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 30 13:22:30.380169 systemd[1]: Expecting device dev-disk-by\x2dlabel-ROOT.device - /dev/disk/by-label/ROOT... Jan 30 13:22:30.380177 systemd[1]: Expecting device dev-disk-by\x2dpartlabel-USR\x2dA.device - /dev/disk/by-partlabel/USR-A... Jan 30 13:22:30.382239 systemd[1]: Expecting device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - /dev/disk/by-partuuid/7130c94a-213a-4e5a-8e26-6cce9662f132... Jan 30 13:22:30.382263 systemd[1]: Expecting device dev-mapper-usr.device - /dev/mapper/usr... Jan 30 13:22:30.382271 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 13:22:30.382282 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 30 13:22:30.382290 systemd[1]: Reached target paths.target - Path Units. Jan 30 13:22:30.382298 systemd[1]: Reached target slices.target - Slice Units. Jan 30 13:22:30.382306 systemd[1]: Reached target swap.target - Swaps. Jan 30 13:22:30.382313 systemd[1]: Reached target timers.target - Timer Units. Jan 30 13:22:30.382321 systemd[1]: Listening on iscsid.socket - Open-iSCSI iscsid Socket. Jan 30 13:22:30.382331 systemd[1]: Listening on iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 30 13:22:30.382341 systemd[1]: Listening on systemd-journald-dev-log.socket - Journal Socket (/dev/log). Jan 30 13:22:30.382349 systemd[1]: Listening on systemd-journald.socket - Journal Socket. Jan 30 13:22:30.382356 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 30 13:22:30.382364 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 30 13:22:30.382375 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 13:22:30.382382 systemd[1]: Reached target sockets.target - Socket Units. Jan 30 13:22:30.382390 systemd[1]: Starting ignition-setup-pre.service - Ignition env setup... Jan 30 13:22:30.382398 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 30 13:22:30.382408 systemd[1]: Finished network-cleanup.service - Network Cleanup. Jan 30 13:22:30.382415 systemd[1]: Starting systemd-fsck-usr.service... Jan 30 13:22:30.382423 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 30 13:22:30.382431 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 30 13:22:30.382474 systemd-journald[218]: Collecting audit messages is disabled. Jan 30 13:22:30.382498 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:22:30.382507 systemd-journald[218]: Journal started Jan 30 13:22:30.382530 systemd-journald[218]: Runtime Journal (/run/log/journal/39447d4c53d84c759bf241a727efd868) is 8.0M, max 78.5M, 70.5M free. Jan 30 13:22:30.388994 systemd-modules-load[219]: Inserted module 'overlay' Jan 30 13:22:30.408523 systemd[1]: Started systemd-journald.service - Journal Service. Jan 30 13:22:30.420202 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Jan 30 13:22:30.430696 systemd[1]: Finished ignition-setup-pre.service - Ignition env setup. Jan 30 13:22:30.445147 kernel: Bridge firewalling registered Jan 30 13:22:30.431331 systemd-modules-load[219]: Inserted module 'br_netfilter' Jan 30 13:22:30.439554 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 13:22:30.452670 systemd[1]: Finished systemd-fsck-usr.service. Jan 30 13:22:30.462222 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 30 13:22:30.475652 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:22:30.501552 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 13:22:30.516421 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 30 13:22:30.532799 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 30 13:22:30.549368 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 30 13:22:30.562218 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:22:30.584425 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:22:30.591518 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 13:22:30.617484 systemd[1]: Starting dracut-cmdline.service - dracut cmdline hook... Jan 30 13:22:30.625402 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 30 13:22:30.639217 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 13:22:30.668202 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 30 13:22:30.686533 dracut-cmdline[252]: dracut-dracut-053 Jan 30 13:22:30.686533 dracut-cmdline[252]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=e6957044c3256d96283265c263579aa4275d1d707b02496fcb081f5fc6356346 Jan 30 13:22:30.690026 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 13:22:30.759019 systemd-resolved[259]: Positive Trust Anchors: Jan 30 13:22:30.759035 systemd-resolved[259]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 30 13:22:30.759066 systemd-resolved[259]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 30 13:22:30.765802 systemd-resolved[259]: Defaulting to hostname 'linux'. Jan 30 13:22:30.766852 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 30 13:22:30.774249 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 30 13:22:30.882210 kernel: SCSI subsystem initialized Jan 30 13:22:30.890218 kernel: Loading iSCSI transport class v2.0-870. Jan 30 13:22:30.902223 kernel: iscsi: registered transport (tcp) Jan 30 13:22:30.920249 kernel: iscsi: registered transport (qla4xxx) Jan 30 13:22:30.920270 kernel: QLogic iSCSI HBA Driver Jan 30 13:22:30.957578 systemd[1]: Finished dracut-cmdline.service - dracut cmdline hook. Jan 30 13:22:30.976327 systemd[1]: Starting dracut-pre-udev.service - dracut pre-udev hook... Jan 30 13:22:31.011159 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Jan 30 13:22:31.011247 kernel: device-mapper: uevent: version 1.0.3 Jan 30 13:22:31.020203 kernel: device-mapper: ioctl: 4.48.0-ioctl (2023-03-01) initialised: dm-devel@redhat.com Jan 30 13:22:31.069210 kernel: raid6: neonx8 gen() 15783 MB/s Jan 30 13:22:31.091193 kernel: raid6: neonx4 gen() 15820 MB/s Jan 30 13:22:31.111193 kernel: raid6: neonx2 gen() 13337 MB/s Jan 30 13:22:31.133194 kernel: raid6: neonx1 gen() 10448 MB/s Jan 30 13:22:31.154192 kernel: raid6: int64x8 gen() 6771 MB/s Jan 30 13:22:31.174192 kernel: raid6: int64x4 gen() 7353 MB/s Jan 30 13:22:31.195193 kernel: raid6: int64x2 gen() 6073 MB/s Jan 30 13:22:31.219375 kernel: raid6: int64x1 gen() 5053 MB/s Jan 30 13:22:31.219386 kernel: raid6: using algorithm neonx4 gen() 15820 MB/s Jan 30 13:22:31.245625 kernel: raid6: .... xor() 12286 MB/s, rmw enabled Jan 30 13:22:31.245638 kernel: raid6: using neon recovery algorithm Jan 30 13:22:31.256957 kernel: xor: measuring software checksum speed Jan 30 13:22:31.256970 kernel: 8regs : 21636 MB/sec Jan 30 13:22:31.264398 kernel: 32regs : 20393 MB/sec Jan 30 13:22:31.264409 kernel: arm64_neon : 27870 MB/sec Jan 30 13:22:31.269436 kernel: xor: using function: arm64_neon (27870 MB/sec) Jan 30 13:22:31.323237 kernel: Btrfs loaded, zoned=no, fsverity=no Jan 30 13:22:31.336975 systemd[1]: Finished dracut-pre-udev.service - dracut pre-udev hook. Jan 30 13:22:31.353341 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 13:22:31.376210 systemd-udevd[439]: Using default interface naming scheme 'v255'. Jan 30 13:22:31.381889 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 13:22:31.402330 systemd[1]: Starting dracut-pre-trigger.service - dracut pre-trigger hook... Jan 30 13:22:31.435779 dracut-pre-trigger[448]: rd.md=0: removing MD RAID activation Jan 30 13:22:31.466709 systemd[1]: Finished dracut-pre-trigger.service - dracut pre-trigger hook. Jan 30 13:22:31.486664 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 30 13:22:31.532126 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 13:22:31.551571 systemd[1]: Starting dracut-initqueue.service - dracut initqueue hook... Jan 30 13:22:31.589384 systemd[1]: Finished dracut-initqueue.service - dracut initqueue hook. Jan 30 13:22:31.602259 systemd[1]: Reached target remote-fs-pre.target - Preparation for Remote File Systems. Jan 30 13:22:31.626169 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 13:22:31.649763 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 30 13:22:31.658907 kernel: hv_vmbus: Vmbus version:5.3 Jan 30 13:22:31.675009 systemd[1]: Starting dracut-pre-mount.service - dracut pre-mount hook... Jan 30 13:22:31.739142 kernel: pps_core: LinuxPPS API ver. 1 registered Jan 30 13:22:31.739171 kernel: hv_vmbus: registering driver hyperv_keyboard Jan 30 13:22:31.739193 kernel: hv_vmbus: registering driver hid_hyperv Jan 30 13:22:31.739204 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Jan 30 13:22:31.739214 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/MSFT1000:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Jan 30 13:22:31.739223 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Jan 30 13:22:31.739232 kernel: hid-hyperv 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Jan 30 13:22:31.739366 kernel: PTP clock support registered Jan 30 13:22:31.707088 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 30 13:22:31.761977 kernel: hv_utils: Registering HyperV Utility Driver Jan 30 13:22:31.762005 kernel: hv_vmbus: registering driver hv_utils Jan 30 13:22:31.707267 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:22:31.675439 kernel: hv_utils: Heartbeat IC version 3.0 Jan 30 13:22:31.689158 kernel: hv_utils: Shutdown IC version 3.2 Jan 30 13:22:31.689176 kernel: hv_vmbus: registering driver hv_netvsc Jan 30 13:22:31.689184 kernel: hv_utils: TimeSync IC version 4.0 Jan 30 13:22:31.689194 kernel: hv_vmbus: registering driver hv_storvsc Jan 30 13:22:31.689201 systemd-journald[218]: Time jumped backwards, rotating. Jan 30 13:22:31.689238 kernel: scsi host0: storvsc_host_t Jan 30 13:22:31.757868 systemd[1]: Stopping dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 13:22:31.710569 kernel: scsi host1: storvsc_host_t Jan 30 13:22:31.710784 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Jan 30 13:22:31.656776 systemd-resolved[259]: Clock change detected. Flushing caches. Jan 30 13:22:31.661779 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 13:22:31.738813 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Jan 30 13:22:31.662004 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:22:31.703437 systemd[1]: Stopping systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:22:31.784873 kernel: hv_netvsc 002248ba-7efb-0022-48ba-7efb002248ba eth0: VF slot 1 added Jan 30 13:22:31.785081 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Jan 30 13:22:31.830016 kernel: hv_vmbus: registering driver hv_pci Jan 30 13:22:31.830040 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Jan 30 13:22:31.830049 kernel: hv_pci 7637483c-19d3-4e2d-94c2-e66e36793b85: PCI VMBus probing: Using version 0x10004 Jan 30 13:22:31.925586 kernel: hv_pci 7637483c-19d3-4e2d-94c2-e66e36793b85: PCI host bridge to bus 19d3:00 Jan 30 13:22:31.925710 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Jan 30 13:22:31.925859 kernel: pci_bus 19d3:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Jan 30 13:22:31.925972 kernel: pci_bus 19d3:00: No busn resource found for root bus, will use [bus 00-ff] Jan 30 13:22:31.926083 kernel: pci 19d3:00:02.0: [15b3:1018] type 00 class 0x020000 Jan 30 13:22:31.926237 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Jan 30 13:22:31.926800 kernel: pci 19d3:00:02.0: reg 0x10: [mem 0xfc0000000-0xfc00fffff 64bit pref] Jan 30 13:22:31.926928 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Jan 30 13:22:31.927028 kernel: sd 0:0:0:0: [sda] Write Protect is off Jan 30 13:22:31.927121 kernel: pci 19d3:00:02.0: enabling Extended Tags Jan 30 13:22:31.927218 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Jan 30 13:22:31.927306 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Jan 30 13:22:31.927390 kernel: pci 19d3:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 19d3:00:02.0 (capable of 126.016 Gb/s with 8.0 GT/s PCIe x16 link) Jan 30 13:22:31.927481 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 13:22:31.927490 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Jan 30 13:22:31.927573 kernel: pci_bus 19d3:00: busn_res: [bus 00-ff] end is updated to 00 Jan 30 13:22:31.927657 kernel: pci 19d3:00:02.0: BAR 0: assigned [mem 0xfc0000000-0xfc00fffff 64bit pref] Jan 30 13:22:31.739178 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:22:31.755430 systemd[1]: Finished dracut-pre-mount.service - dracut pre-mount hook. Jan 30 13:22:31.815878 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:22:31.831890 systemd[1]: Starting dracut-cmdline-ask.service - dracut ask for additional cmdline parameters... Jan 30 13:22:31.937583 systemd[1]: Finished dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:22:31.989750 kernel: mlx5_core 19d3:00:02.0: enabling device (0000 -> 0002) Jan 30 13:22:32.221805 kernel: mlx5_core 19d3:00:02.0: firmware version: 16.30.1284 Jan 30 13:22:32.221961 kernel: hv_netvsc 002248ba-7efb-0022-48ba-7efb002248ba eth0: VF registering: eth1 Jan 30 13:22:32.222063 kernel: mlx5_core 19d3:00:02.0 eth1: joined to eth0 Jan 30 13:22:32.222158 kernel: mlx5_core 19d3:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0 basic) Jan 30 13:22:32.230749 kernel: mlx5_core 19d3:00:02.0 enP6611s1: renamed from eth1 Jan 30 13:22:32.300557 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device - Virtual_Disk EFI-SYSTEM. Jan 30 13:22:32.468762 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by (udev-worker) (502) Jan 30 13:22:32.484082 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Jan 30 13:22:32.505511 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device - Virtual_Disk ROOT. Jan 30 13:22:32.590740 kernel: BTRFS: device fsid 1e2e5fa7-c757-4d5d-af66-73afe98fbaae devid 1 transid 39 /dev/sda3 scanned by (udev-worker) (482) Jan 30 13:22:32.605098 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device - Virtual_Disk USR-A. Jan 30 13:22:32.612579 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device - Virtual_Disk USR-A. Jan 30 13:22:32.645973 systemd[1]: Starting disk-uuid.service - Generate new UUID for disk GPT if necessary... Jan 30 13:22:32.669803 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 13:22:33.682776 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Jan 30 13:22:33.682833 disk-uuid[599]: The operation has completed successfully. Jan 30 13:22:33.752803 systemd[1]: disk-uuid.service: Deactivated successfully. Jan 30 13:22:33.754751 systemd[1]: Finished disk-uuid.service - Generate new UUID for disk GPT if necessary. Jan 30 13:22:33.800999 systemd[1]: Starting verity-setup.service - Verity Setup for /dev/mapper/usr... Jan 30 13:22:33.815537 sh[685]: Success Jan 30 13:22:33.835093 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Jan 30 13:22:34.154643 systemd[1]: Found device dev-mapper-usr.device - /dev/mapper/usr. Jan 30 13:22:34.180908 systemd[1]: Mounting sysusr-usr.mount - /sysusr/usr... Jan 30 13:22:34.192319 systemd[1]: Finished verity-setup.service - Verity Setup for /dev/mapper/usr. Jan 30 13:22:34.227091 kernel: BTRFS info (device dm-0): first mount of filesystem 1e2e5fa7-c757-4d5d-af66-73afe98fbaae Jan 30 13:22:34.227143 kernel: BTRFS info (device dm-0): using crc32c (crc32c-generic) checksum algorithm Jan 30 13:22:34.234392 kernel: BTRFS warning (device dm-0): 'nologreplay' is deprecated, use 'rescue=nologreplay' instead Jan 30 13:22:34.240215 kernel: BTRFS info (device dm-0): disabling log replay at mount time Jan 30 13:22:34.245870 kernel: BTRFS info (device dm-0): using free space tree Jan 30 13:22:34.795772 systemd[1]: Mounted sysusr-usr.mount - /sysusr/usr. Jan 30 13:22:34.801908 systemd[1]: afterburn-network-kargs.service - Afterburn Initrd Setup Network Kernel Arguments was skipped because no trigger condition checks were met. Jan 30 13:22:34.823025 systemd[1]: Starting ignition-setup.service - Ignition (setup)... Jan 30 13:22:34.831837 systemd[1]: Starting parse-ip-for-networkd.service - Write systemd-networkd units from cmdline... Jan 30 13:22:34.873428 kernel: BTRFS info (device sda6): first mount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:34.873497 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Jan 30 13:22:34.878118 kernel: BTRFS info (device sda6): using free space tree Jan 30 13:22:34.913459 kernel: BTRFS info (device sda6): auto enabling async discard Jan 30 13:22:34.930050 systemd[1]: mnt-oem.mount: Deactivated successfully. Jan 30 13:22:34.935396 kernel: BTRFS info (device sda6): last unmount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:34.943217 systemd[1]: Finished ignition-setup.service - Ignition (setup). Jan 30 13:22:34.958029 systemd[1]: Starting ignition-fetch-offline.service - Ignition (fetch-offline)... Jan 30 13:22:34.965146 systemd[1]: Finished parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 30 13:22:34.988548 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 30 13:22:35.022795 systemd-networkd[869]: lo: Link UP Jan 30 13:22:35.026277 systemd-networkd[869]: lo: Gained carrier Jan 30 13:22:35.028448 systemd-networkd[869]: Enumeration completed Jan 30 13:22:35.031003 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 30 13:22:35.031347 systemd-networkd[869]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:22:35.031350 systemd-networkd[869]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 13:22:35.037803 systemd[1]: Reached target network.target - Network. Jan 30 13:22:35.129743 kernel: mlx5_core 19d3:00:02.0 enP6611s1: Link up Jan 30 13:22:35.168743 kernel: hv_netvsc 002248ba-7efb-0022-48ba-7efb002248ba eth0: Data path switched to VF: enP6611s1 Jan 30 13:22:35.169244 systemd-networkd[869]: enP6611s1: Link UP Jan 30 13:22:35.169434 systemd-networkd[869]: eth0: Link UP Jan 30 13:22:35.169542 systemd-networkd[869]: eth0: Gained carrier Jan 30 13:22:35.169553 systemd-networkd[869]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:22:35.184269 systemd-networkd[869]: enP6611s1: Gained carrier Jan 30 13:22:35.205767 systemd-networkd[869]: eth0: DHCPv4 address 10.200.20.40/24, gateway 10.200.20.1 acquired from 168.63.129.16 Jan 30 13:22:36.325115 ignition[864]: Ignition 2.20.0 Jan 30 13:22:36.325127 ignition[864]: Stage: fetch-offline Jan 30 13:22:36.325170 ignition[864]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:36.335104 systemd[1]: Finished ignition-fetch-offline.service - Ignition (fetch-offline). Jan 30 13:22:36.325178 ignition[864]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:36.325271 ignition[864]: parsed url from cmdline: "" Jan 30 13:22:36.325274 ignition[864]: no config URL provided Jan 30 13:22:36.325279 ignition[864]: reading system config file "/usr/lib/ignition/user.ign" Jan 30 13:22:36.325285 ignition[864]: no config at "/usr/lib/ignition/user.ign" Jan 30 13:22:36.363075 systemd[1]: Starting ignition-fetch.service - Ignition (fetch)... Jan 30 13:22:36.325291 ignition[864]: failed to fetch config: resource requires networking Jan 30 13:22:36.325486 ignition[864]: Ignition finished successfully Jan 30 13:22:36.386403 ignition[880]: Ignition 2.20.0 Jan 30 13:22:36.386410 ignition[880]: Stage: fetch Jan 30 13:22:36.386579 ignition[880]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:36.386588 ignition[880]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:36.386677 ignition[880]: parsed url from cmdline: "" Jan 30 13:22:36.386680 ignition[880]: no config URL provided Jan 30 13:22:36.386684 ignition[880]: reading system config file "/usr/lib/ignition/user.ign" Jan 30 13:22:36.386692 ignition[880]: no config at "/usr/lib/ignition/user.ign" Jan 30 13:22:36.386732 ignition[880]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Jan 30 13:22:36.494907 ignition[880]: GET result: OK Jan 30 13:22:36.494977 ignition[880]: config has been read from IMDS userdata Jan 30 13:22:36.495016 ignition[880]: parsing config with SHA512: 6f25ae24839f5dd6d1be3ca08131bb21cfa0c0c8a38bcc2ebc6f91db920acf07ad49d964bbf6f8cfeabc74dac631371b91a23873884712feff26df55f11a995a Jan 30 13:22:36.500032 unknown[880]: fetched base config from "system" Jan 30 13:22:36.500534 ignition[880]: fetch: fetch complete Jan 30 13:22:36.500039 unknown[880]: fetched base config from "system" Jan 30 13:22:36.500539 ignition[880]: fetch: fetch passed Jan 30 13:22:36.500047 unknown[880]: fetched user config from "azure" Jan 30 13:22:36.500603 ignition[880]: Ignition finished successfully Jan 30 13:22:36.503184 systemd[1]: Finished ignition-fetch.service - Ignition (fetch). Jan 30 13:22:36.526895 systemd[1]: Starting ignition-kargs.service - Ignition (kargs)... Jan 30 13:22:36.556365 ignition[886]: Ignition 2.20.0 Jan 30 13:22:36.556393 ignition[886]: Stage: kargs Jan 30 13:22:36.561674 systemd[1]: Finished ignition-kargs.service - Ignition (kargs). Jan 30 13:22:36.556615 ignition[886]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:36.556625 ignition[886]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:36.577033 systemd[1]: Starting ignition-disks.service - Ignition (disks)... Jan 30 13:22:36.557699 ignition[886]: kargs: kargs passed Jan 30 13:22:36.557767 ignition[886]: Ignition finished successfully Jan 30 13:22:36.604023 systemd[1]: Finished ignition-disks.service - Ignition (disks). Jan 30 13:22:36.600626 ignition[892]: Ignition 2.20.0 Jan 30 13:22:36.610582 systemd[1]: Reached target initrd-root-device.target - Initrd Root Device. Jan 30 13:22:36.600633 ignition[892]: Stage: disks Jan 30 13:22:36.620837 systemd[1]: Reached target local-fs-pre.target - Preparation for Local File Systems. Jan 30 13:22:36.600828 ignition[892]: no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:36.632763 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 30 13:22:36.600838 ignition[892]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:36.641809 systemd[1]: Reached target sysinit.target - System Initialization. Jan 30 13:22:36.601803 ignition[892]: disks: disks passed Jan 30 13:22:36.653631 systemd[1]: Reached target basic.target - Basic System. Jan 30 13:22:36.601857 ignition[892]: Ignition finished successfully Jan 30 13:22:36.669984 systemd[1]: Starting systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT... Jan 30 13:22:36.756400 systemd-fsck[901]: ROOT: clean, 14/7326000 files, 477710/7359488 blocks Jan 30 13:22:36.767423 systemd[1]: Finished systemd-fsck-root.service - File System Check on /dev/disk/by-label/ROOT. Jan 30 13:22:36.787925 systemd[1]: Mounting sysroot.mount - /sysroot... Jan 30 13:22:36.805921 systemd-networkd[869]: eth0: Gained IPv6LL Jan 30 13:22:36.851581 kernel: EXT4-fs (sda9): mounted filesystem 88903c49-366d-43ff-90b1-141790b6e85c r/w with ordered data mode. Quota mode: none. Jan 30 13:22:36.852127 systemd[1]: Mounted sysroot.mount - /sysroot. Jan 30 13:22:36.861463 systemd[1]: Reached target initrd-root-fs.target - Initrd Root File System. Jan 30 13:22:36.917823 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 30 13:22:36.929120 systemd[1]: Mounting sysroot-usr.mount - /sysroot/usr... Jan 30 13:22:36.941759 systemd[1]: Starting flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent... Jan 30 13:22:36.966218 kernel: BTRFS: device label OEM devid 1 transid 16 /dev/sda6 scanned by mount (912) Jan 30 13:22:36.958879 systemd[1]: ignition-remount-sysroot.service - Remount /sysroot read-write for Ignition was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Jan 30 13:22:36.958923 systemd[1]: Reached target ignition-diskful.target - Ignition Boot Disk Setup. Jan 30 13:22:37.024132 kernel: BTRFS info (device sda6): first mount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:37.024160 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Jan 30 13:22:37.024170 kernel: BTRFS info (device sda6): using free space tree Jan 30 13:22:37.024179 kernel: BTRFS info (device sda6): auto enabling async discard Jan 30 13:22:36.975005 systemd[1]: Mounted sysroot-usr.mount - /sysroot/usr. Jan 30 13:22:37.015000 systemd[1]: Starting initrd-setup-root.service - Root filesystem setup... Jan 30 13:22:37.031452 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 30 13:22:37.061815 systemd-networkd[869]: enP6611s1: Gained IPv6LL Jan 30 13:22:37.738997 coreos-metadata[914]: Jan 30 13:22:37.738 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Jan 30 13:22:37.748362 coreos-metadata[914]: Jan 30 13:22:37.748 INFO Fetch successful Jan 30 13:22:37.748362 coreos-metadata[914]: Jan 30 13:22:37.748 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Jan 30 13:22:37.769015 coreos-metadata[914]: Jan 30 13:22:37.768 INFO Fetch successful Jan 30 13:22:37.775246 coreos-metadata[914]: Jan 30 13:22:37.774 INFO wrote hostname ci-4186.1.0-a-a7402f3063 to /sysroot/etc/hostname Jan 30 13:22:37.776803 systemd[1]: Finished flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 30 13:22:38.124225 initrd-setup-root[943]: cut: /sysroot/etc/passwd: No such file or directory Jan 30 13:22:38.231683 initrd-setup-root[950]: cut: /sysroot/etc/group: No such file or directory Jan 30 13:22:38.241704 initrd-setup-root[957]: cut: /sysroot/etc/shadow: No such file or directory Jan 30 13:22:38.251500 initrd-setup-root[964]: cut: /sysroot/etc/gshadow: No such file or directory Jan 30 13:22:39.725379 systemd[1]: Finished initrd-setup-root.service - Root filesystem setup. Jan 30 13:22:39.744861 systemd[1]: Starting ignition-mount.service - Ignition (mount)... Jan 30 13:22:39.760966 systemd[1]: Starting sysroot-boot.service - /sysroot/boot... Jan 30 13:22:39.780020 kernel: BTRFS info (device sda6): last unmount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:39.775970 systemd[1]: sysroot-oem.mount: Deactivated successfully. Jan 30 13:22:39.804916 systemd[1]: Finished sysroot-boot.service - /sysroot/boot. Jan 30 13:22:39.818761 ignition[1032]: INFO : Ignition 2.20.0 Jan 30 13:22:39.818761 ignition[1032]: INFO : Stage: mount Jan 30 13:22:39.818761 ignition[1032]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:39.818761 ignition[1032]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:39.848195 ignition[1032]: INFO : mount: mount passed Jan 30 13:22:39.848195 ignition[1032]: INFO : Ignition finished successfully Jan 30 13:22:39.833687 systemd[1]: Finished ignition-mount.service - Ignition (mount). Jan 30 13:22:39.858996 systemd[1]: Starting ignition-files.service - Ignition (files)... Jan 30 13:22:39.878026 systemd[1]: Mounting sysroot-oem.mount - /sysroot/oem... Jan 30 13:22:39.910732 kernel: BTRFS: device label OEM devid 1 transid 17 /dev/sda6 scanned by mount (1044) Jan 30 13:22:39.910785 kernel: BTRFS info (device sda6): first mount of filesystem 5265f28b-8d78-4be2-8b05-2145d9ab7cfa Jan 30 13:22:39.917032 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Jan 30 13:22:39.921965 kernel: BTRFS info (device sda6): using free space tree Jan 30 13:22:39.928944 kernel: BTRFS info (device sda6): auto enabling async discard Jan 30 13:22:39.930411 systemd[1]: Mounted sysroot-oem.mount - /sysroot/oem. Jan 30 13:22:39.958024 ignition[1061]: INFO : Ignition 2.20.0 Jan 30 13:22:39.958024 ignition[1061]: INFO : Stage: files Jan 30 13:22:39.967498 ignition[1061]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:39.967498 ignition[1061]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:39.967498 ignition[1061]: DEBUG : files: compiled without relabeling support, skipping Jan 30 13:22:39.988825 ignition[1061]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Jan 30 13:22:39.988825 ignition[1061]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Jan 30 13:22:40.238396 ignition[1061]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Jan 30 13:22:40.246866 ignition[1061]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Jan 30 13:22:40.246866 ignition[1061]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Jan 30 13:22:40.241298 unknown[1061]: wrote ssh authorized keys file for user: core Jan 30 13:22:40.317406 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jan 30 13:22:40.328599 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Jan 30 13:22:40.400812 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Jan 30 13:22:40.527784 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Jan 30 13:22:40.527784 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/home/core/install.sh" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/home/core/install.sh" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/home/core/nginx.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/home/core/nginx.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/etc/flatcar/update.conf" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/etc/flatcar/update.conf" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Jan 30 13:22:40.551626 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.30.1-arm64.raw: attempt #1 Jan 30 13:22:40.993294 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(a): GET result: OK Jan 30 13:22:41.205081 ignition[1061]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.30.1-arm64.raw" Jan 30 13:22:41.205081 ignition[1061]: INFO : files: op(b): [started] processing unit "prepare-helm.service" Jan 30 13:22:41.281174 ignition[1061]: INFO : files: op(b): op(c): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: op(b): op(c): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: op(b): [finished] processing unit "prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: op(d): [started] setting preset to enabled for "prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: op(d): [finished] setting preset to enabled for "prepare-helm.service" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: createResultFile: createFiles: op(e): [started] writing file "/sysroot/etc/.ignition-result.json" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: createResultFile: createFiles: op(e): [finished] writing file "/sysroot/etc/.ignition-result.json" Jan 30 13:22:41.294108 ignition[1061]: INFO : files: files passed Jan 30 13:22:41.294108 ignition[1061]: INFO : Ignition finished successfully Jan 30 13:22:41.294524 systemd[1]: Finished ignition-files.service - Ignition (files). Jan 30 13:22:41.340024 systemd[1]: Starting ignition-quench.service - Ignition (record completion)... Jan 30 13:22:41.359973 systemd[1]: Starting initrd-setup-root-after-ignition.service - Root filesystem completion... Jan 30 13:22:41.383415 systemd[1]: ignition-quench.service: Deactivated successfully. Jan 30 13:22:41.383514 systemd[1]: Finished ignition-quench.service - Ignition (record completion). Jan 30 13:22:41.436126 initrd-setup-root-after-ignition[1090]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 30 13:22:41.436126 initrd-setup-root-after-ignition[1090]: grep: /sysroot/usr/share/flatcar/enabled-sysext.conf: No such file or directory Jan 30 13:22:41.457572 initrd-setup-root-after-ignition[1094]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Jan 30 13:22:41.449073 systemd[1]: Finished initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 30 13:22:41.465673 systemd[1]: Reached target ignition-complete.target - Ignition Complete. Jan 30 13:22:41.500009 systemd[1]: Starting initrd-parse-etc.service - Mountpoints Configured in the Real Root... Jan 30 13:22:41.530082 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Jan 30 13:22:41.530958 systemd[1]: Finished initrd-parse-etc.service - Mountpoints Configured in the Real Root. Jan 30 13:22:41.544377 systemd[1]: Reached target initrd-fs.target - Initrd File Systems. Jan 30 13:22:41.557381 systemd[1]: Reached target initrd.target - Initrd Default Target. Jan 30 13:22:41.569507 systemd[1]: dracut-mount.service - dracut mount hook was skipped because no trigger condition checks were met. Jan 30 13:22:41.572967 systemd[1]: Starting dracut-pre-pivot.service - dracut pre-pivot and cleanup hook... Jan 30 13:22:41.607956 systemd[1]: Finished dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 30 13:22:41.625965 systemd[1]: Starting initrd-cleanup.service - Cleaning Up and Shutting Down Daemons... Jan 30 13:22:41.646177 systemd[1]: Stopped target nss-lookup.target - Host and Network Name Lookups. Jan 30 13:22:41.653708 systemd[1]: Stopped target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 13:22:41.668791 systemd[1]: Stopped target timers.target - Timer Units. Jan 30 13:22:41.681030 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Jan 30 13:22:41.681206 systemd[1]: Stopped dracut-pre-pivot.service - dracut pre-pivot and cleanup hook. Jan 30 13:22:41.699876 systemd[1]: Stopped target initrd.target - Initrd Default Target. Jan 30 13:22:41.714020 systemd[1]: Stopped target basic.target - Basic System. Jan 30 13:22:41.725611 systemd[1]: Stopped target ignition-complete.target - Ignition Complete. Jan 30 13:22:41.738020 systemd[1]: Stopped target ignition-diskful.target - Ignition Boot Disk Setup. Jan 30 13:22:41.751600 systemd[1]: Stopped target initrd-root-device.target - Initrd Root Device. Jan 30 13:22:41.765469 systemd[1]: Stopped target remote-fs.target - Remote File Systems. Jan 30 13:22:41.779552 systemd[1]: Stopped target remote-fs-pre.target - Preparation for Remote File Systems. Jan 30 13:22:41.793667 systemd[1]: Stopped target sysinit.target - System Initialization. Jan 30 13:22:41.808137 systemd[1]: Stopped target local-fs.target - Local File Systems. Jan 30 13:22:41.821267 systemd[1]: Stopped target swap.target - Swaps. Jan 30 13:22:41.832374 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Jan 30 13:22:41.832556 systemd[1]: Stopped dracut-pre-mount.service - dracut pre-mount hook. Jan 30 13:22:41.849838 systemd[1]: Stopped target cryptsetup.target - Local Encrypted Volumes. Jan 30 13:22:41.862879 systemd[1]: Stopped target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 13:22:41.876325 systemd[1]: clevis-luks-askpass.path: Deactivated successfully. Jan 30 13:22:41.876437 systemd[1]: Stopped clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 13:22:41.892274 systemd[1]: dracut-initqueue.service: Deactivated successfully. Jan 30 13:22:41.892452 systemd[1]: Stopped dracut-initqueue.service - dracut initqueue hook. Jan 30 13:22:41.913132 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Jan 30 13:22:41.913324 systemd[1]: Stopped initrd-setup-root-after-ignition.service - Root filesystem completion. Jan 30 13:22:41.926898 systemd[1]: ignition-files.service: Deactivated successfully. Jan 30 13:22:41.927073 systemd[1]: Stopped ignition-files.service - Ignition (files). Jan 30 13:22:41.939606 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Jan 30 13:22:41.939805 systemd[1]: Stopped flatcar-metadata-hostname.service - Flatcar Metadata Hostname Agent. Jan 30 13:22:41.975894 systemd[1]: Stopping ignition-mount.service - Ignition (mount)... Jan 30 13:22:42.017088 ignition[1114]: INFO : Ignition 2.20.0 Jan 30 13:22:42.017088 ignition[1114]: INFO : Stage: umount Jan 30 13:22:42.017088 ignition[1114]: INFO : no configs at "/usr/lib/ignition/base.d" Jan 30 13:22:42.017088 ignition[1114]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Jan 30 13:22:42.017088 ignition[1114]: INFO : umount: umount passed Jan 30 13:22:42.017088 ignition[1114]: INFO : Ignition finished successfully Jan 30 13:22:41.994983 systemd[1]: Stopping sysroot-boot.service - /sysroot/boot... Jan 30 13:22:42.008276 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Jan 30 13:22:42.008470 systemd[1]: Stopped systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 13:22:42.030298 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Jan 30 13:22:42.030448 systemd[1]: Stopped dracut-pre-trigger.service - dracut pre-trigger hook. Jan 30 13:22:42.051280 systemd[1]: ignition-mount.service: Deactivated successfully. Jan 30 13:22:42.051385 systemd[1]: Stopped ignition-mount.service - Ignition (mount). Jan 30 13:22:42.061955 systemd[1]: sysroot-boot.mount: Deactivated successfully. Jan 30 13:22:42.062672 systemd[1]: ignition-disks.service: Deactivated successfully. Jan 30 13:22:42.063823 systemd[1]: Stopped ignition-disks.service - Ignition (disks). Jan 30 13:22:42.082043 systemd[1]: ignition-kargs.service: Deactivated successfully. Jan 30 13:22:42.082123 systemd[1]: Stopped ignition-kargs.service - Ignition (kargs). Jan 30 13:22:42.093627 systemd[1]: ignition-fetch.service: Deactivated successfully. Jan 30 13:22:42.093687 systemd[1]: Stopped ignition-fetch.service - Ignition (fetch). Jan 30 13:22:42.105370 systemd[1]: Stopped target network.target - Network. Jan 30 13:22:42.111506 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Jan 30 13:22:42.111589 systemd[1]: Stopped ignition-fetch-offline.service - Ignition (fetch-offline). Jan 30 13:22:42.125816 systemd[1]: Stopped target paths.target - Path Units. Jan 30 13:22:42.137775 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Jan 30 13:22:42.141786 systemd[1]: Stopped systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 13:22:42.151402 systemd[1]: Stopped target slices.target - Slice Units. Jan 30 13:22:42.170066 systemd[1]: Stopped target sockets.target - Socket Units. Jan 30 13:22:42.182070 systemd[1]: iscsid.socket: Deactivated successfully. Jan 30 13:22:42.182174 systemd[1]: Closed iscsid.socket - Open-iSCSI iscsid Socket. Jan 30 13:22:42.193282 systemd[1]: iscsiuio.socket: Deactivated successfully. Jan 30 13:22:42.193335 systemd[1]: Closed iscsiuio.socket - Open-iSCSI iscsiuio Socket. Jan 30 13:22:42.204527 systemd[1]: ignition-setup.service: Deactivated successfully. Jan 30 13:22:42.204582 systemd[1]: Stopped ignition-setup.service - Ignition (setup). Jan 30 13:22:42.215998 systemd[1]: ignition-setup-pre.service: Deactivated successfully. Jan 30 13:22:42.216045 systemd[1]: Stopped ignition-setup-pre.service - Ignition env setup. Jan 30 13:22:42.228135 systemd[1]: Stopping systemd-networkd.service - Network Configuration... Jan 30 13:22:42.471100 kernel: hv_netvsc 002248ba-7efb-0022-48ba-7efb002248ba eth0: Data path switched from VF: enP6611s1 Jan 30 13:22:42.238959 systemd[1]: Stopping systemd-resolved.service - Network Name Resolution... Jan 30 13:22:42.251048 systemd[1]: initrd-cleanup.service: Deactivated successfully. Jan 30 13:22:42.251153 systemd[1]: Finished initrd-cleanup.service - Cleaning Up and Shutting Down Daemons. Jan 30 13:22:42.255827 systemd-networkd[869]: eth0: DHCPv6 lease lost Jan 30 13:22:42.265942 systemd[1]: systemd-networkd.service: Deactivated successfully. Jan 30 13:22:42.266057 systemd[1]: Stopped systemd-networkd.service - Network Configuration. Jan 30 13:22:42.276013 systemd[1]: systemd-networkd.socket: Deactivated successfully. Jan 30 13:22:42.276092 systemd[1]: Closed systemd-networkd.socket - Network Service Netlink Socket. Jan 30 13:22:42.311275 systemd[1]: Stopping network-cleanup.service - Network Cleanup... Jan 30 13:22:42.324422 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Jan 30 13:22:42.324516 systemd[1]: Stopped parse-ip-for-networkd.service - Write systemd-networkd units from cmdline. Jan 30 13:22:42.340592 systemd[1]: Stopping systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 13:22:42.357259 systemd[1]: systemd-resolved.service: Deactivated successfully. Jan 30 13:22:42.357377 systemd[1]: Stopped systemd-resolved.service - Network Name Resolution. Jan 30 13:22:42.387075 systemd[1]: systemd-sysctl.service: Deactivated successfully. Jan 30 13:22:42.387270 systemd[1]: Stopped systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:22:42.400598 systemd[1]: systemd-modules-load.service: Deactivated successfully. Jan 30 13:22:42.400668 systemd[1]: Stopped systemd-modules-load.service - Load Kernel Modules. Jan 30 13:22:42.414293 systemd[1]: systemd-tmpfiles-setup.service: Deactivated successfully. Jan 30 13:22:42.414364 systemd[1]: Stopped systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 13:22:42.429312 systemd[1]: systemd-udevd.service: Deactivated successfully. Jan 30 13:22:42.429466 systemd[1]: Stopped systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 13:22:42.452196 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Jan 30 13:22:42.452274 systemd[1]: Closed systemd-udevd-control.socket - udev Control Socket. Jan 30 13:22:42.464585 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Jan 30 13:22:42.464638 systemd[1]: Closed systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 13:22:42.477663 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Jan 30 13:22:42.477772 systemd[1]: Stopped dracut-pre-udev.service - dracut pre-udev hook. Jan 30 13:22:42.496678 systemd[1]: dracut-cmdline.service: Deactivated successfully. Jan 30 13:22:42.496775 systemd[1]: Stopped dracut-cmdline.service - dracut cmdline hook. Jan 30 13:22:42.515386 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Jan 30 13:22:42.515469 systemd[1]: Stopped dracut-cmdline-ask.service - dracut ask for additional cmdline parameters. Jan 30 13:22:42.557015 systemd[1]: Starting initrd-udevadm-cleanup-db.service - Cleanup udev Database... Jan 30 13:22:42.573795 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Jan 30 13:22:42.573876 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 13:22:42.590189 systemd[1]: systemd-tmpfiles-setup-dev-early.service: Deactivated successfully. Jan 30 13:22:42.590249 systemd[1]: Stopped systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 13:22:42.605245 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Jan 30 13:22:42.605298 systemd[1]: Stopped kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 13:22:42.618389 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 13:22:42.618442 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:22:42.631120 systemd[1]: network-cleanup.service: Deactivated successfully. Jan 30 13:22:42.631243 systemd[1]: Stopped network-cleanup.service - Network Cleanup. Jan 30 13:22:42.643134 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Jan 30 13:22:42.643227 systemd[1]: Finished initrd-udevadm-cleanup-db.service - Cleanup udev Database. Jan 30 13:22:44.664021 systemd[1]: sysroot-boot.service: Deactivated successfully. Jan 30 13:22:44.664145 systemd[1]: Stopped sysroot-boot.service - /sysroot/boot. Jan 30 13:22:44.670342 systemd[1]: Reached target initrd-switch-root.target - Switch Root. Jan 30 13:22:44.682215 systemd[1]: initrd-setup-root.service: Deactivated successfully. Jan 30 13:22:44.682294 systemd[1]: Stopped initrd-setup-root.service - Root filesystem setup. Jan 30 13:22:44.713058 systemd[1]: Starting initrd-switch-root.service - Switch Root... Jan 30 13:22:44.727184 systemd[1]: Switching root. Jan 30 13:22:44.972288 systemd-journald[218]: Journal stopped Jan 30 13:22:57.732910 systemd-journald[218]: Received SIGTERM from PID 1 (systemd). Jan 30 13:22:57.732938 kernel: SELinux: policy capability network_peer_controls=1 Jan 30 13:22:57.732949 kernel: SELinux: policy capability open_perms=1 Jan 30 13:22:57.732960 kernel: SELinux: policy capability extended_socket_class=1 Jan 30 13:22:57.732968 kernel: SELinux: policy capability always_check_network=0 Jan 30 13:22:57.732976 kernel: SELinux: policy capability cgroup_seclabel=1 Jan 30 13:22:57.732984 kernel: SELinux: policy capability nnp_nosuid_transition=1 Jan 30 13:22:57.732992 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Jan 30 13:22:57.733000 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Jan 30 13:22:57.733008 kernel: audit: type=1403 audit(1738243369.986:2): auid=4294967295 ses=4294967295 lsm=selinux res=1 Jan 30 13:22:57.733018 systemd[1]: Successfully loaded SELinux policy in 455.282ms. Jan 30 13:22:57.733028 systemd[1]: Relabeled /dev, /dev/shm, /run, /sys/fs/cgroup in 10.745ms. Jan 30 13:22:57.733038 systemd[1]: systemd 255 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL +ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE +TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP -SYSVINIT default-hierarchy=unified) Jan 30 13:22:57.733047 systemd[1]: Detected virtualization microsoft. Jan 30 13:22:57.733056 systemd[1]: Detected architecture arm64. Jan 30 13:22:57.733067 systemd[1]: Detected first boot. Jan 30 13:22:57.733076 systemd[1]: Hostname set to . Jan 30 13:22:57.733087 systemd[1]: Initializing machine ID from random generator. Jan 30 13:22:57.733096 zram_generator::config[1156]: No configuration found. Jan 30 13:22:57.733105 systemd[1]: Populated /etc with preset unit settings. Jan 30 13:22:57.733115 systemd[1]: initrd-switch-root.service: Deactivated successfully. Jan 30 13:22:57.733125 systemd[1]: Stopped initrd-switch-root.service - Switch Root. Jan 30 13:22:57.733134 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Jan 30 13:22:57.733144 systemd[1]: Created slice system-addon\x2dconfig.slice - Slice /system/addon-config. Jan 30 13:22:57.733153 systemd[1]: Created slice system-addon\x2drun.slice - Slice /system/addon-run. Jan 30 13:22:57.733163 systemd[1]: Created slice system-getty.slice - Slice /system/getty. Jan 30 13:22:57.733172 systemd[1]: Created slice system-modprobe.slice - Slice /system/modprobe. Jan 30 13:22:57.733182 systemd[1]: Created slice system-serial\x2dgetty.slice - Slice /system/serial-getty. Jan 30 13:22:57.733193 systemd[1]: Created slice system-system\x2dcloudinit.slice - Slice /system/system-cloudinit. Jan 30 13:22:57.733202 systemd[1]: Created slice system-systemd\x2dfsck.slice - Slice /system/systemd-fsck. Jan 30 13:22:57.733212 systemd[1]: Created slice user.slice - User and Session Slice. Jan 30 13:22:57.733221 systemd[1]: Started clevis-luks-askpass.path - Forward Password Requests to Clevis Directory Watch. Jan 30 13:22:57.733230 systemd[1]: Started systemd-ask-password-console.path - Dispatch Password Requests to Console Directory Watch. Jan 30 13:22:57.733240 systemd[1]: Started systemd-ask-password-wall.path - Forward Password Requests to Wall Directory Watch. Jan 30 13:22:57.733249 systemd[1]: Set up automount boot.automount - Boot partition Automount Point. Jan 30 13:22:57.733258 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount - Arbitrary Executable File Formats File System Automount Point. Jan 30 13:22:57.733267 systemd[1]: Expecting device dev-disk-by\x2dlabel-OEM.device - /dev/disk/by-label/OEM... Jan 30 13:22:57.733278 systemd[1]: Expecting device dev-ttyAMA0.device - /dev/ttyAMA0... Jan 30 13:22:57.733288 systemd[1]: Reached target cryptsetup-pre.target - Local Encrypted Volumes (Pre). Jan 30 13:22:57.733297 systemd[1]: Stopped target initrd-switch-root.target - Switch Root. Jan 30 13:22:57.733309 systemd[1]: Stopped target initrd-fs.target - Initrd File Systems. Jan 30 13:22:57.733318 systemd[1]: Stopped target initrd-root-fs.target - Initrd Root File System. Jan 30 13:22:57.733328 systemd[1]: Reached target integritysetup.target - Local Integrity Protected Volumes. Jan 30 13:22:57.733337 systemd[1]: Reached target remote-cryptsetup.target - Remote Encrypted Volumes. Jan 30 13:22:57.733353 systemd[1]: Reached target remote-fs.target - Remote File Systems. Jan 30 13:22:57.733362 systemd[1]: Reached target slices.target - Slice Units. Jan 30 13:22:57.733371 systemd[1]: Reached target swap.target - Swaps. Jan 30 13:22:57.733381 systemd[1]: Reached target veritysetup.target - Local Verity Protected Volumes. Jan 30 13:22:57.733390 systemd[1]: Listening on systemd-coredump.socket - Process Core Dump Socket. Jan 30 13:22:57.733399 systemd[1]: Listening on systemd-networkd.socket - Network Service Netlink Socket. Jan 30 13:22:57.733408 systemd[1]: Listening on systemd-udevd-control.socket - udev Control Socket. Jan 30 13:22:57.733421 systemd[1]: Listening on systemd-udevd-kernel.socket - udev Kernel Socket. Jan 30 13:22:57.733431 systemd[1]: Listening on systemd-userdbd.socket - User Database Manager Socket. Jan 30 13:22:57.733440 systemd[1]: Mounting dev-hugepages.mount - Huge Pages File System... Jan 30 13:22:57.733449 systemd[1]: Mounting dev-mqueue.mount - POSIX Message Queue File System... Jan 30 13:22:57.733459 systemd[1]: Mounting media.mount - External Media Directory... Jan 30 13:22:57.733468 systemd[1]: Mounting sys-kernel-debug.mount - Kernel Debug File System... Jan 30 13:22:57.733478 systemd[1]: Mounting sys-kernel-tracing.mount - Kernel Trace File System... Jan 30 13:22:57.733488 systemd[1]: Mounting tmp.mount - Temporary Directory /tmp... Jan 30 13:22:57.733498 systemd[1]: var-lib-machines.mount - Virtual Machine and Container Storage (Compatibility) was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Jan 30 13:22:57.733508 systemd[1]: Reached target machines.target - Containers. Jan 30 13:22:57.733517 systemd[1]: Starting flatcar-tmpfiles.service - Create missing system files... Jan 30 13:22:57.733527 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 13:22:57.733536 systemd[1]: Starting kmod-static-nodes.service - Create List of Static Device Nodes... Jan 30 13:22:57.733545 systemd[1]: Starting modprobe@configfs.service - Load Kernel Module configfs... Jan 30 13:22:57.733556 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 30 13:22:57.733566 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 30 13:22:57.733575 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 30 13:22:57.733584 systemd[1]: Starting modprobe@fuse.service - Load Kernel Module fuse... Jan 30 13:22:57.733594 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 30 13:22:57.733604 systemd[1]: setup-nsswitch.service - Create /etc/nsswitch.conf was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Jan 30 13:22:57.733613 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Jan 30 13:22:57.733623 systemd[1]: Stopped systemd-fsck-root.service - File System Check on Root Device. Jan 30 13:22:57.733633 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Jan 30 13:22:57.733644 systemd[1]: Stopped systemd-fsck-usr.service. Jan 30 13:22:57.733653 systemd[1]: Starting systemd-journald.service - Journal Service... Jan 30 13:22:57.733681 systemd-journald[1235]: Collecting audit messages is disabled. Jan 30 13:22:57.733703 systemd-journald[1235]: Journal started Jan 30 13:22:57.733885 systemd-journald[1235]: Runtime Journal (/run/log/journal/3667f3bd8792463d89aaa05230dfef4c) is 8.0M, max 78.5M, 70.5M free. Jan 30 13:22:56.231609 systemd[1]: Queued start job for default target multi-user.target. Jan 30 13:22:56.539391 systemd[1]: Unnecessary job was removed for dev-sda6.device - /dev/sda6. Jan 30 13:22:56.539815 systemd[1]: systemd-journald.service: Deactivated successfully. Jan 30 13:22:56.540164 systemd[1]: systemd-journald.service: Consumed 3.510s CPU time. Jan 30 13:22:57.833771 kernel: fuse: init (API version 7.39) Jan 30 13:22:57.850543 systemd[1]: Starting systemd-modules-load.service - Load Kernel Modules... Jan 30 13:22:57.867956 systemd[1]: Starting systemd-network-generator.service - Generate network units from Kernel command line... Jan 30 13:22:57.886424 systemd[1]: Starting systemd-remount-fs.service - Remount Root and Kernel File Systems... Jan 30 13:22:57.894730 kernel: loop: module loaded Jan 30 13:22:57.905763 systemd[1]: Starting systemd-udev-trigger.service - Coldplug All udev Devices... Jan 30 13:22:57.924279 systemd[1]: verity-setup.service: Deactivated successfully. Jan 30 13:22:57.924353 systemd[1]: Stopped verity-setup.service. Jan 30 13:22:57.941946 systemd[1]: Started systemd-journald.service - Journal Service. Jan 30 13:22:57.942893 systemd[1]: Mounted dev-hugepages.mount - Huge Pages File System. Jan 30 13:22:57.950017 systemd[1]: Mounted dev-mqueue.mount - POSIX Message Queue File System. Jan 30 13:22:57.957231 systemd[1]: Mounted media.mount - External Media Directory. Jan 30 13:22:57.963833 systemd[1]: Mounted sys-kernel-debug.mount - Kernel Debug File System. Jan 30 13:22:57.971001 systemd[1]: Mounted sys-kernel-tracing.mount - Kernel Trace File System. Jan 30 13:22:57.978068 systemd[1]: Mounted tmp.mount - Temporary Directory /tmp. Jan 30 13:22:57.985760 systemd[1]: Finished kmod-static-nodes.service - Create List of Static Device Nodes. Jan 30 13:22:57.994038 systemd[1]: modprobe@configfs.service: Deactivated successfully. Jan 30 13:22:57.994193 systemd[1]: Finished modprobe@configfs.service - Load Kernel Module configfs. Jan 30 13:22:58.001401 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 30 13:22:58.001534 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 30 13:22:58.008980 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 30 13:22:58.009130 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 30 13:22:58.016858 systemd[1]: modprobe@fuse.service: Deactivated successfully. Jan 30 13:22:58.016988 systemd[1]: Finished modprobe@fuse.service - Load Kernel Module fuse. Jan 30 13:22:58.024069 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 30 13:22:58.024206 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 30 13:22:58.031459 systemd[1]: Finished systemd-modules-load.service - Load Kernel Modules. Jan 30 13:22:58.038635 systemd[1]: Finished systemd-network-generator.service - Generate network units from Kernel command line. Jan 30 13:22:58.047502 systemd[1]: Finished systemd-remount-fs.service - Remount Root and Kernel File Systems. Jan 30 13:22:58.056328 systemd[1]: Finished systemd-udev-trigger.service - Coldplug All udev Devices. Jan 30 13:22:58.073321 systemd[1]: Reached target network-pre.target - Preparation for Network. Jan 30 13:22:58.090734 kernel: ACPI: bus type drm_connector registered Jan 30 13:22:58.093857 systemd[1]: Mounting sys-fs-fuse-connections.mount - FUSE Control File System... Jan 30 13:22:58.101807 systemd[1]: Mounting sys-kernel-config.mount - Kernel Configuration File System... Jan 30 13:22:58.109140 systemd[1]: remount-root.service - Remount Root File System was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Jan 30 13:22:58.109189 systemd[1]: Reached target local-fs.target - Local File Systems. Jan 30 13:22:58.116624 systemd[1]: Listening on systemd-sysext.socket - System Extension Image Management (Varlink). Jan 30 13:22:58.125296 systemd[1]: Starting dracut-shutdown.service - Restore /run/initramfs on shutdown... Jan 30 13:22:58.133285 systemd[1]: Starting ldconfig.service - Rebuild Dynamic Linker Cache... Jan 30 13:22:58.139920 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 13:22:58.141457 systemd[1]: Starting systemd-hwdb-update.service - Rebuild Hardware Database... Jan 30 13:22:58.149253 systemd[1]: Starting systemd-journal-flush.service - Flush Journal to Persistent Storage... Jan 30 13:22:58.156463 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 30 13:22:58.157903 systemd[1]: Starting systemd-random-seed.service - Load/Save OS Random Seed... Jan 30 13:22:58.164355 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 30 13:22:58.165946 systemd[1]: Starting systemd-sysctl.service - Apply Kernel Variables... Jan 30 13:22:58.180357 systemd[1]: Starting systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/... Jan 30 13:22:58.190517 systemd[1]: Starting systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully... Jan 30 13:22:58.203075 systemd[1]: Starting systemd-udev-settle.service - Wait for udev To Complete Device Initialization... Jan 30 13:22:58.213109 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 30 13:22:58.213287 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 30 13:22:58.220585 systemd[1]: Mounted sys-fs-fuse-connections.mount - FUSE Control File System. Jan 30 13:22:58.228107 systemd[1]: Mounted sys-kernel-config.mount - Kernel Configuration File System. Jan 30 13:22:58.238869 udevadm[1270]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Jan 30 13:22:58.269241 systemd[1]: Finished dracut-shutdown.service - Restore /run/initramfs on shutdown. Jan 30 13:22:58.292747 kernel: loop0: detected capacity change from 0 to 194096 Jan 30 13:22:58.314244 systemd-journald[1235]: Time spent on flushing to /var/log/journal/3667f3bd8792463d89aaa05230dfef4c is 1.231984s for 902 entries. Jan 30 13:22:58.314244 systemd-journald[1235]: System Journal (/var/log/journal/3667f3bd8792463d89aaa05230dfef4c) is 11.8M, max 2.6G, 2.6G free. Jan 30 13:23:00.666966 systemd-journald[1235]: Received client request to flush runtime journal. Jan 30 13:23:00.667038 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Jan 30 13:23:00.667059 kernel: loop1: detected capacity change from 0 to 113552 Jan 30 13:23:00.667075 systemd-journald[1235]: /var/log/journal/3667f3bd8792463d89aaa05230dfef4c/system.journal: Realtime clock jumped backwards relative to last journal entry, rotating. Jan 30 13:23:00.667101 systemd-journald[1235]: Rotating system journal. Jan 30 13:23:00.667123 kernel: loop2: detected capacity change from 0 to 116784 Jan 30 13:22:58.480415 systemd[1]: Finished systemd-sysctl.service - Apply Kernel Variables. Jan 30 13:22:58.673804 systemd[1]: Finished systemd-random-seed.service - Load/Save OS Random Seed. Jan 30 13:22:58.677456 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Jan 30 13:22:58.677466 systemd-tmpfiles[1269]: ACLs are not supported, ignoring. Jan 30 13:22:58.682416 systemd[1]: Reached target first-boot-complete.target - First Boot Complete. Jan 30 13:22:58.697044 systemd[1]: Starting systemd-machine-id-commit.service - Commit a transient machine-id on disk... Jan 30 13:22:58.717557 systemd[1]: Finished systemd-tmpfiles-setup-dev-early.service - Create Static Device Nodes in /dev gracefully. Jan 30 13:23:00.621430 systemd[1]: Finished flatcar-tmpfiles.service - Create missing system files. Jan 30 13:23:00.647316 systemd[1]: Starting systemd-sysusers.service - Create System Users... Jan 30 13:23:00.671263 systemd[1]: Finished systemd-journal-flush.service - Flush Journal to Persistent Storage. Jan 30 13:23:01.015804 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Jan 30 13:23:01.016528 systemd[1]: Finished systemd-machine-id-commit.service - Commit a transient machine-id on disk. Jan 30 13:23:01.328785 systemd[1]: Finished systemd-sysusers.service - Create System Users. Jan 30 13:23:01.348766 kernel: loop3: detected capacity change from 0 to 28752 Jan 30 13:23:01.350014 systemd[1]: Starting systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev... Jan 30 13:23:01.370076 systemd-tmpfiles[1316]: ACLs are not supported, ignoring. Jan 30 13:23:01.370099 systemd-tmpfiles[1316]: ACLs are not supported, ignoring. Jan 30 13:23:01.375866 systemd[1]: Finished systemd-tmpfiles-setup-dev.service - Create Static Device Nodes in /dev. Jan 30 13:23:01.468742 kernel: loop4: detected capacity change from 0 to 194096 Jan 30 13:23:01.480739 kernel: loop5: detected capacity change from 0 to 113552 Jan 30 13:23:01.497823 kernel: loop6: detected capacity change from 0 to 116784 Jan 30 13:23:01.508761 kernel: loop7: detected capacity change from 0 to 28752 Jan 30 13:23:01.512946 (sd-merge)[1320]: Using extensions 'containerd-flatcar', 'docker-flatcar', 'kubernetes', 'oem-azure'. Jan 30 13:23:01.514284 (sd-merge)[1320]: Merged extensions into '/usr'. Jan 30 13:23:01.517822 systemd[1]: Reloading requested from client PID 1267 ('systemd-sysext') (unit systemd-sysext.service)... Jan 30 13:23:01.517840 systemd[1]: Reloading... Jan 30 13:23:01.622795 zram_generator::config[1346]: No configuration found. Jan 30 13:23:01.770770 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 13:23:01.828307 systemd[1]: Reloading finished in 310 ms. Jan 30 13:23:01.856530 systemd[1]: Finished systemd-hwdb-update.service - Rebuild Hardware Database. Jan 30 13:23:01.864731 systemd[1]: Finished systemd-sysext.service - Merge System Extension Images into /usr/ and /opt/. Jan 30 13:23:01.880977 systemd[1]: Starting ensure-sysext.service... Jan 30 13:23:01.889995 systemd[1]: Starting systemd-tmpfiles-setup.service - Create System Files and Directories... Jan 30 13:23:01.900853 systemd[1]: Starting systemd-udevd.service - Rule-based Manager for Device Events and Files... Jan 30 13:23:01.915849 systemd[1]: Reloading requested from client PID 1402 ('systemctl') (unit ensure-sysext.service)... Jan 30 13:23:01.915872 systemd[1]: Reloading... Jan 30 13:23:01.932222 systemd-udevd[1404]: Using default interface naming scheme 'v255'. Jan 30 13:23:01.945220 systemd-tmpfiles[1403]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Jan 30 13:23:01.946047 systemd-tmpfiles[1403]: /usr/lib/tmpfiles.d/systemd-flatcar.conf:6: Duplicate line for path "/var/log/journal", ignoring. Jan 30 13:23:01.946895 systemd-tmpfiles[1403]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Jan 30 13:23:01.947254 systemd-tmpfiles[1403]: ACLs are not supported, ignoring. Jan 30 13:23:01.947386 systemd-tmpfiles[1403]: ACLs are not supported, ignoring. Jan 30 13:23:01.953405 systemd-tmpfiles[1403]: Detected autofs mount point /boot during canonicalization of boot. Jan 30 13:23:01.953417 systemd-tmpfiles[1403]: Skipping /boot Jan 30 13:23:01.976145 systemd-tmpfiles[1403]: Detected autofs mount point /boot during canonicalization of boot. Jan 30 13:23:01.976159 systemd-tmpfiles[1403]: Skipping /boot Jan 30 13:23:01.993746 zram_generator::config[1428]: No configuration found. Jan 30 13:23:02.201049 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 13:23:02.262750 kernel: mousedev: PS/2 mouse device common for all mice Jan 30 13:23:02.274194 kernel: hv_vmbus: registering driver hyperv_fb Jan 30 13:23:02.296754 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Jan 30 13:23:02.296847 kernel: hv_vmbus: registering driver hv_balloon Jan 30 13:23:02.296864 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Jan 30 13:23:02.304230 kernel: Console: switching to colour dummy device 80x25 Jan 30 13:23:02.309385 systemd[1]: Condition check resulted in dev-ttyAMA0.device - /dev/ttyAMA0 being skipped. Jan 30 13:23:02.310415 systemd[1]: Reloading finished in 394 ms. Jan 30 13:23:02.313768 kernel: Console: switching to colour frame buffer device 128x48 Jan 30 13:23:02.337871 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Jan 30 13:23:02.337979 kernel: hv_balloon: Memory hot add disabled on ARM64 Jan 30 13:23:02.346277 systemd[1]: Started systemd-udevd.service - Rule-based Manager for Device Events and Files. Jan 30 13:23:02.367855 systemd[1]: Finished systemd-tmpfiles-setup.service - Create System Files and Directories. Jan 30 13:23:02.414218 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 30 13:23:02.424514 systemd[1]: Starting clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs... Jan 30 13:23:02.443080 systemd[1]: Starting systemd-journal-catalog-update.service - Rebuild Journal Catalog... Jan 30 13:23:02.446738 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (1467) Jan 30 13:23:02.459127 systemd[1]: Starting systemd-networkd.service - Network Configuration... Jan 30 13:23:02.481395 systemd[1]: Starting systemd-resolved.service - Network Name Resolution... Jan 30 13:23:02.494114 systemd[1]: Starting systemd-update-utmp.service - Record System Boot/Shutdown in UTMP... Jan 30 13:23:02.510989 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:23:02.537669 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 13:23:02.571759 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 30 13:23:02.581678 augenrules[1604]: No rules Jan 30 13:23:02.583743 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 30 13:23:02.596277 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 30 13:23:02.606702 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 13:23:02.612182 systemd[1]: Starting systemd-userdbd.service - User Database Manager... Jan 30 13:23:02.626804 systemd[1]: audit-rules.service: Deactivated successfully. Jan 30 13:23:02.627041 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 30 13:23:02.639063 systemd[1]: Finished systemd-journal-catalog-update.service - Rebuild Journal Catalog. Jan 30 13:23:02.652206 systemd[1]: Finished systemd-update-utmp.service - Record System Boot/Shutdown in UTMP. Jan 30 13:23:02.661176 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Jan 30 13:23:02.662791 systemd[1]: Stopped systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:23:02.670187 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 30 13:23:02.670344 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 30 13:23:02.678073 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 30 13:23:02.678209 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 30 13:23:02.686302 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 30 13:23:02.686442 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 30 13:23:02.715796 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device - Virtual_Disk OEM. Jan 30 13:23:02.723732 systemd[1]: Started systemd-userdbd.service - User Database Manager. Jan 30 13:23:02.731069 systemd[1]: Finished systemd-udev-settle.service - Wait for udev To Complete Device Initialization. Jan 30 13:23:02.747264 systemd[1]: Finished clean-ca-certificates.service - Clean up broken links in /etc/ssl/certs. Jan 30 13:23:02.778979 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 30 13:23:02.788583 systemd[1]: ignition-delete-config.service - Ignition (delete config) was skipped because no trigger condition checks were met. Jan 30 13:23:02.793952 systemd[1]: Starting lvm2-activation-early.service - Activation of LVM2 logical volumes... Jan 30 13:23:02.809852 systemd[1]: Starting modprobe@dm_mod.service - Load Kernel Module dm_mod... Jan 30 13:23:02.827749 systemd-resolved[1557]: Positive Trust Anchors: Jan 30 13:23:02.827772 systemd-resolved[1557]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Jan 30 13:23:02.827808 systemd-resolved[1557]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 170.0.0.192.in-addr.arpa 171.0.0.192.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa ipv4only.arpa resolver.arpa corp home internal intranet lan local private test Jan 30 13:23:02.836120 systemd[1]: Starting modprobe@drm.service - Load Kernel Module drm... Jan 30 13:23:02.843111 systemd-resolved[1557]: Using system hostname 'ci-4186.1.0-a-a7402f3063'. Jan 30 13:23:02.850247 lvm[1639]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 30 13:23:02.854847 systemd[1]: Starting modprobe@efi_pstore.service - Load Kernel Module efi_pstore... Jan 30 13:23:02.867354 systemd-networkd[1549]: lo: Link UP Jan 30 13:23:02.867679 systemd-networkd[1549]: lo: Gained carrier Jan 30 13:23:02.868929 augenrules[1638]: /sbin/augenrules: No change Jan 30 13:23:02.872227 systemd-networkd[1549]: Enumeration completed Jan 30 13:23:02.877147 systemd[1]: Starting modprobe@loop.service - Load Kernel Module loop... Jan 30 13:23:02.880734 systemd-networkd[1549]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:23:02.880743 systemd-networkd[1549]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 13:23:02.886778 systemd[1]: systemd-binfmt.service - Set Up Additional Binary Formats was skipped because no trigger condition checks were met. Jan 30 13:23:02.893996 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM... Jan 30 13:23:02.899188 augenrules[1662]: No rules Jan 30 13:23:02.900877 systemd[1]: Reached target time-set.target - System Time Set. Jan 30 13:23:02.908789 ldconfig[1262]: /sbin/ldconfig: /usr/lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Jan 30 13:23:02.912039 systemd[1]: Starting systemd-vconsole-setup.service - Virtual Console Setup... Jan 30 13:23:02.922149 systemd[1]: update-ca-certificates.service - Update CA bundle at /etc/ssl/certs/ca-certificates.crt was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Jan 30 13:23:02.922999 systemd[1]: Started systemd-resolved.service - Network Name Resolution. Jan 30 13:23:02.932558 systemd[1]: Started systemd-networkd.service - Network Configuration. Jan 30 13:23:02.943791 systemd[1]: Finished ensure-sysext.service. Jan 30 13:23:02.951771 systemd[1]: Finished ldconfig.service - Rebuild Dynamic Linker Cache. Jan 30 13:23:02.960084 systemd[1]: audit-rules.service: Deactivated successfully. Jan 30 13:23:02.961401 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 30 13:23:02.974773 kernel: mlx5_core 19d3:00:02.0 enP6611s1: Link up Jan 30 13:23:02.975108 systemd[1]: Finished lvm2-activation-early.service - Activation of LVM2 logical volumes. Jan 30 13:23:02.985085 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Jan 30 13:23:02.985250 systemd[1]: Finished modprobe@dm_mod.service - Load Kernel Module dm_mod. Jan 30 13:23:02.992974 systemd[1]: modprobe@drm.service: Deactivated successfully. Jan 30 13:23:02.993138 systemd[1]: Finished modprobe@drm.service - Load Kernel Module drm. Jan 30 13:23:03.008757 kernel: hv_netvsc 002248ba-7efb-0022-48ba-7efb002248ba eth0: Data path switched to VF: enP6611s1 Jan 30 13:23:03.008800 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Jan 30 13:23:03.009397 systemd-networkd[1549]: enP6611s1: Link UP Jan 30 13:23:03.009528 systemd-networkd[1549]: eth0: Link UP Jan 30 13:23:03.009531 systemd-networkd[1549]: eth0: Gained carrier Jan 30 13:23:03.009552 systemd-networkd[1549]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:23:03.010294 systemd[1]: Finished modprobe@efi_pstore.service - Load Kernel Module efi_pstore. Jan 30 13:23:03.019219 systemd[1]: modprobe@loop.service: Deactivated successfully. Jan 30 13:23:03.019370 systemd[1]: Finished modprobe@loop.service - Load Kernel Module loop. Jan 30 13:23:03.027924 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service - File System Check on /dev/disk/by-label/OEM. Jan 30 13:23:03.028100 systemd-networkd[1549]: enP6611s1: Gained carrier Jan 30 13:23:03.044509 systemd[1]: Reached target cryptsetup.target - Local Encrypted Volumes. Jan 30 13:23:03.044823 systemd-networkd[1549]: eth0: DHCPv4 address 10.200.20.40/24, gateway 10.200.20.1 acquired from 168.63.129.16 Jan 30 13:23:03.052103 systemd[1]: Reached target network.target - Network. Jan 30 13:23:03.058691 systemd[1]: Reached target nss-lookup.target - Host and Network Name Lookups. Jan 30 13:23:03.070911 systemd[1]: Starting lvm2-activation.service - Activation of LVM2 logical volumes... Jan 30 13:23:03.081992 systemd[1]: Starting systemd-networkd-wait-online.service - Wait for Network to be Configured... Jan 30 13:23:03.083249 lvm[1680]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Jan 30 13:23:03.092599 systemd[1]: systemd-pstore.service - Platform Persistent Storage Archival was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Jan 30 13:23:03.092688 systemd[1]: systemd-repart.service - Repartition Root Disk was skipped because no trigger condition checks were met. Jan 30 13:23:03.095976 systemd[1]: Starting systemd-update-done.service - Update is Completed... Jan 30 13:23:03.115855 systemd[1]: Finished lvm2-activation.service - Activation of LVM2 logical volumes. Jan 30 13:23:03.125140 systemd[1]: Finished systemd-update-done.service - Update is Completed. Jan 30 13:23:03.148175 systemd[1]: Finished systemd-vconsole-setup.service - Virtual Console Setup. Jan 30 13:23:03.155468 systemd[1]: Reached target sysinit.target - System Initialization. Jan 30 13:23:03.162112 systemd[1]: Started motdgen.path - Watch for update engine configuration changes. Jan 30 13:23:03.169241 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path - Watch for a cloud-config at /var/lib/flatcar-install/user_data. Jan 30 13:23:03.177092 systemd[1]: Started logrotate.timer - Daily rotation of log files. Jan 30 13:23:03.183436 systemd[1]: Started mdadm.timer - Weekly check for MD array's redundancy information.. Jan 30 13:23:03.190793 systemd[1]: Started systemd-tmpfiles-clean.timer - Daily Cleanup of Temporary Directories. Jan 30 13:23:03.198444 systemd[1]: update-engine-stub.timer - Update Engine Stub Timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Jan 30 13:23:03.198482 systemd[1]: Reached target paths.target - Path Units. Jan 30 13:23:03.203943 systemd[1]: Reached target timers.target - Timer Units. Jan 30 13:23:03.210766 systemd[1]: Listening on dbus.socket - D-Bus System Message Bus Socket. Jan 30 13:23:03.218891 systemd[1]: Starting docker.socket - Docker Socket for the API... Jan 30 13:23:03.229438 systemd[1]: Listening on sshd.socket - OpenSSH Server Socket. Jan 30 13:23:03.236519 systemd[1]: Listening on docker.socket - Docker Socket for the API. Jan 30 13:23:03.243386 systemd[1]: Reached target sockets.target - Socket Units. Jan 30 13:23:03.249287 systemd[1]: Reached target basic.target - Basic System. Jan 30 13:23:03.254978 systemd[1]: addon-config@oem.service - Configure Addon /oem was skipped because no trigger condition checks were met. Jan 30 13:23:03.255022 systemd[1]: addon-run@oem.service - Run Addon /oem was skipped because no trigger condition checks were met. Jan 30 13:23:03.260838 systemd[1]: Starting chronyd.service - NTP client/server... Jan 30 13:23:03.267912 systemd[1]: Starting containerd.service - containerd container runtime... Jan 30 13:23:03.285393 systemd[1]: Starting coreos-metadata.service - Flatcar Metadata Agent... Jan 30 13:23:03.292191 (chronyd)[1690]: chronyd.service: Referenced but unset environment variable evaluates to an empty string: OPTIONS Jan 30 13:23:03.294973 systemd[1]: Starting dbus.service - D-Bus System Message Bus... Jan 30 13:23:03.302840 systemd[1]: Starting enable-oem-cloudinit.service - Enable cloudinit... Jan 30 13:23:03.307238 chronyd[1699]: chronyd version 4.6.1 starting (+CMDMON +NTP +REFCLOCK +RTC +PRIVDROP +SCFILTER -SIGND +ASYNCDNS +NTS +SECHASH +IPV6 -DEBUG) Jan 30 13:23:03.313965 systemd[1]: Starting extend-filesystems.service - Extend Filesystems... Jan 30 13:23:03.319919 systemd[1]: flatcar-setup-environment.service - Modifies /etc/environment for CoreOS was skipped because of an unmet condition check (ConditionPathExists=/oem/bin/flatcar-setup-environment). Jan 30 13:23:03.319968 systemd[1]: hv_fcopy_daemon.service - Hyper-V FCOPY daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_fcopy). Jan 30 13:23:03.321638 chronyd[1699]: Timezone right/UTC failed leap second check, ignoring Jan 30 13:23:03.322006 chronyd[1699]: Loaded seccomp filter (level 2) Jan 30 13:23:03.322094 systemd[1]: Started hv_kvp_daemon.service - Hyper-V KVP daemon. Jan 30 13:23:03.331047 jq[1697]: false Jan 30 13:23:03.331575 systemd[1]: hv_vss_daemon.service - Hyper-V VSS daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/vmbus/hv_vss). Jan 30 13:23:03.332950 systemd[1]: Starting motdgen.service - Generate /run/flatcar/motd... Jan 30 13:23:03.342918 systemd[1]: Starting prepare-helm.service - Unpack helm to /opt/bin... Jan 30 13:23:03.344078 KVP[1701]: KVP starting; pid is:1701 Jan 30 13:23:03.356013 systemd[1]: Starting ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline... Jan 30 13:23:03.366751 extend-filesystems[1700]: Found loop4 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found loop5 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found loop6 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found loop7 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found sda Jan 30 13:23:03.378242 extend-filesystems[1700]: Found sda1 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found sda2 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found sda3 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found usr Jan 30 13:23:03.378242 extend-filesystems[1700]: Found sda4 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found sda6 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found sda7 Jan 30 13:23:03.378242 extend-filesystems[1700]: Found sda9 Jan 30 13:23:03.378242 extend-filesystems[1700]: Checking size of /dev/sda9 Jan 30 13:23:03.650118 kernel: hv_utils: KVP IC version 4.0 Jan 30 13:23:03.650156 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (1443) Jan 30 13:23:03.370772 systemd[1]: Starting sshd-keygen.service - Generate sshd host keys... Jan 30 13:23:03.377688 dbus-daemon[1693]: [system] SELinux support is enabled Jan 30 13:23:03.661950 extend-filesystems[1700]: Old size kept for /dev/sda9 Jan 30 13:23:03.661950 extend-filesystems[1700]: Found sr0 Jan 30 13:23:03.703981 coreos-metadata[1692]: Jan 30 13:23:03.419 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Jan 30 13:23:03.703981 coreos-metadata[1692]: Jan 30 13:23:03.442 INFO Fetch successful Jan 30 13:23:03.703981 coreos-metadata[1692]: Jan 30 13:23:03.442 INFO Fetching http://168.63.129.16/machine/?comp=goalstate: Attempt #1 Jan 30 13:23:03.703981 coreos-metadata[1692]: Jan 30 13:23:03.452 INFO Fetch successful Jan 30 13:23:03.703981 coreos-metadata[1692]: Jan 30 13:23:03.452 INFO Fetching http://168.63.129.16/machine/14c3077d-ba36-4136-b639-be7d96bd1129/8d7d602c%2D0e67%2D41fb%2D80b4%2D69e054c831dd.%5Fci%2D4186.1.0%2Da%2Da7402f3063?comp=config&type=sharedConfig&incarnation=1: Attempt #1 Jan 30 13:23:03.703981 coreos-metadata[1692]: Jan 30 13:23:03.454 INFO Fetch successful Jan 30 13:23:03.703981 coreos-metadata[1692]: Jan 30 13:23:03.454 INFO Fetching http://169.254.169.254/metadata/instance/compute/vmSize?api-version=2017-08-01&format=text: Attempt #1 Jan 30 13:23:03.703981 coreos-metadata[1692]: Jan 30 13:23:03.476 INFO Fetch successful Jan 30 13:23:03.408935 systemd[1]: Starting systemd-logind.service - User Login Management... Jan 30 13:23:03.415764 KVP[1701]: KVP LIC Version: 3.1 Jan 30 13:23:03.428090 systemd[1]: tcsd.service - TCG Core Services Daemon was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Jan 30 13:23:03.429109 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Jan 30 13:23:03.704662 update_engine[1724]: I20250130 13:23:03.500401 1724 main.cc:92] Flatcar Update Engine starting Jan 30 13:23:03.704662 update_engine[1724]: I20250130 13:23:03.504854 1724 update_check_scheduler.cc:74] Next update check in 7m48s Jan 30 13:23:03.435971 systemd[1]: Starting update-engine.service - Update Engine... Jan 30 13:23:03.704993 jq[1726]: true Jan 30 13:23:03.456040 systemd[1]: Starting update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition... Jan 30 13:23:03.484329 systemd[1]: Started dbus.service - D-Bus System Message Bus. Jan 30 13:23:03.509141 systemd[1]: Started chronyd.service - NTP client/server. Jan 30 13:23:03.523802 systemd-logind[1717]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Jan 30 13:23:03.527680 systemd-logind[1717]: New seat seat0. Jan 30 13:23:03.530165 systemd[1]: Started systemd-logind.service - User Login Management. Jan 30 13:23:03.547321 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Jan 30 13:23:03.547556 systemd[1]: Condition check resulted in enable-oem-cloudinit.service - Enable cloudinit being skipped. Jan 30 13:23:03.548043 systemd[1]: extend-filesystems.service: Deactivated successfully. Jan 30 13:23:03.548836 systemd[1]: Finished extend-filesystems.service - Extend Filesystems. Jan 30 13:23:03.590093 systemd[1]: motdgen.service: Deactivated successfully. Jan 30 13:23:03.590966 systemd[1]: Finished motdgen.service - Generate /run/flatcar/motd. Jan 30 13:23:03.671417 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Jan 30 13:23:03.671917 systemd[1]: Finished ssh-key-proc-cmdline.service - Install an ssh key from /proc/cmdline. Jan 30 13:23:03.714764 jq[1780]: true Jan 30 13:23:03.721787 (ntainerd)[1781]: containerd.service: Referenced but unset environment variable evaluates to an empty string: TORCX_IMAGEDIR, TORCX_UNPACKDIR Jan 30 13:23:03.726766 systemd[1]: Finished coreos-metadata.service - Flatcar Metadata Agent. Jan 30 13:23:03.758508 dbus-daemon[1693]: [system] Successfully activated service 'org.freedesktop.systemd1' Jan 30 13:23:03.759227 tar[1774]: linux-arm64/helm Jan 30 13:23:03.776311 systemd[1]: Started update-engine.service - Update Engine. Jan 30 13:23:03.789348 systemd[1]: packet-phone-home.service - Report Success to Packet was skipped because no trigger condition checks were met. Jan 30 13:23:03.790000 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service - Load cloud-config from /usr/share/oem/cloud-config.yml was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Jan 30 13:23:03.790124 systemd[1]: Reached target system-config.target - Load system-provided cloud configs. Jan 30 13:23:03.804246 systemd[1]: user-cloudinit-proc-cmdline.service - Load cloud-config from url defined in /proc/cmdline was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Jan 30 13:23:03.804364 systemd[1]: Reached target user-config.target - Load user-provided cloud configs. Jan 30 13:23:03.828032 systemd[1]: Started locksmithd.service - Cluster reboot manager. Jan 30 13:23:03.857752 bash[1810]: Updated "/home/core/.ssh/authorized_keys" Jan 30 13:23:03.861255 systemd[1]: Finished update-ssh-keys-after-ignition.service - Run update-ssh-keys once after Ignition. Jan 30 13:23:03.878599 systemd[1]: sshkeys.service was skipped because no trigger condition checks were met. Jan 30 13:23:03.971237 locksmithd[1811]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Jan 30 13:23:04.070784 containerd[1781]: time="2025-01-30T13:23:04.070270280Z" level=info msg="starting containerd" revision=9b2ad7760328148397346d10c7b2004271249db4 version=v1.7.23 Jan 30 13:23:04.099429 containerd[1781]: time="2025-01-30T13:23:04.099092120Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Jan 30 13:23:04.100683 containerd[1781]: time="2025-01-30T13:23:04.100634040Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/6.6.74-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:23:04.100683 containerd[1781]: time="2025-01-30T13:23:04.100680960Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Jan 30 13:23:04.100786 containerd[1781]: time="2025-01-30T13:23:04.100708520Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Jan 30 13:23:04.100919 containerd[1781]: time="2025-01-30T13:23:04.100895040Z" level=info msg="loading plugin \"io.containerd.warning.v1.deprecations\"..." type=io.containerd.warning.v1 Jan 30 13:23:04.100947 containerd[1781]: time="2025-01-30T13:23:04.100921440Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101011 containerd[1781]: time="2025-01-30T13:23:04.100989080Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.blockfile\"..." error="no scratch file generator: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101032 containerd[1781]: time="2025-01-30T13:23:04.101008680Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101232 containerd[1781]: time="2025-01-30T13:23:04.101208560Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101252 containerd[1781]: time="2025-01-30T13:23:04.101229680Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101252 containerd[1781]: time="2025-01-30T13:23:04.101242880Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." error="devmapper not configured: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101287 containerd[1781]: time="2025-01-30T13:23:04.101252240Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101342 containerd[1781]: time="2025-01-30T13:23:04.101323920Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101544 containerd[1781]: time="2025-01-30T13:23:04.101523720Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101659 containerd[1781]: time="2025-01-30T13:23:04.101636960Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Jan 30 13:23:04.101679 containerd[1781]: time="2025-01-30T13:23:04.101656480Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Jan 30 13:23:04.108867 containerd[1781]: time="2025-01-30T13:23:04.108817440Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Jan 30 13:23:04.108993 containerd[1781]: time="2025-01-30T13:23:04.108936160Z" level=info msg="metadata content store policy set" policy=shared Jan 30 13:23:04.123774 containerd[1781]: time="2025-01-30T13:23:04.123708080Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Jan 30 13:23:04.123924 containerd[1781]: time="2025-01-30T13:23:04.123798680Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Jan 30 13:23:04.123924 containerd[1781]: time="2025-01-30T13:23:04.123816600Z" level=info msg="loading plugin \"io.containerd.lease.v1.manager\"..." type=io.containerd.lease.v1 Jan 30 13:23:04.123924 containerd[1781]: time="2025-01-30T13:23:04.123836960Z" level=info msg="loading plugin \"io.containerd.streaming.v1.manager\"..." type=io.containerd.streaming.v1 Jan 30 13:23:04.123924 containerd[1781]: time="2025-01-30T13:23:04.123852840Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Jan 30 13:23:04.124119 containerd[1781]: time="2025-01-30T13:23:04.124051280Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Jan 30 13:23:04.124344 containerd[1781]: time="2025-01-30T13:23:04.124317200Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Jan 30 13:23:04.124456 containerd[1781]: time="2025-01-30T13:23:04.124436360Z" level=info msg="loading plugin \"io.containerd.runtime.v2.shim\"..." type=io.containerd.runtime.v2 Jan 30 13:23:04.124479 containerd[1781]: time="2025-01-30T13:23:04.124458160Z" level=info msg="loading plugin \"io.containerd.sandbox.store.v1.local\"..." type=io.containerd.sandbox.store.v1 Jan 30 13:23:04.124508 containerd[1781]: time="2025-01-30T13:23:04.124478840Z" level=info msg="loading plugin \"io.containerd.sandbox.controller.v1.local\"..." type=io.containerd.sandbox.controller.v1 Jan 30 13:23:04.124508 containerd[1781]: time="2025-01-30T13:23:04.124493840Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Jan 30 13:23:04.124543 containerd[1781]: time="2025-01-30T13:23:04.124508240Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Jan 30 13:23:04.124543 containerd[1781]: time="2025-01-30T13:23:04.124520920Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Jan 30 13:23:04.124543 containerd[1781]: time="2025-01-30T13:23:04.124534840Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Jan 30 13:23:04.124591 containerd[1781]: time="2025-01-30T13:23:04.124550280Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Jan 30 13:23:04.124591 containerd[1781]: time="2025-01-30T13:23:04.124563320Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Jan 30 13:23:04.124591 containerd[1781]: time="2025-01-30T13:23:04.124574480Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Jan 30 13:23:04.124591 containerd[1781]: time="2025-01-30T13:23:04.124585520Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Jan 30 13:23:04.124660 containerd[1781]: time="2025-01-30T13:23:04.124611520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124660 containerd[1781]: time="2025-01-30T13:23:04.124626920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124660 containerd[1781]: time="2025-01-30T13:23:04.124639360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124660 containerd[1781]: time="2025-01-30T13:23:04.124651920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124763 containerd[1781]: time="2025-01-30T13:23:04.124663600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124763 containerd[1781]: time="2025-01-30T13:23:04.124677320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124763 containerd[1781]: time="2025-01-30T13:23:04.124688480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124763 containerd[1781]: time="2025-01-30T13:23:04.124702520Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124763 containerd[1781]: time="2025-01-30T13:23:04.124734360Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandbox-controllers\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124763 containerd[1781]: time="2025-01-30T13:23:04.124750560Z" level=info msg="loading plugin \"io.containerd.grpc.v1.sandboxes\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124763 containerd[1781]: time="2025-01-30T13:23:04.124762600Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124882 containerd[1781]: time="2025-01-30T13:23:04.124775640Z" level=info msg="loading plugin \"io.containerd.grpc.v1.streaming\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124882 containerd[1781]: time="2025-01-30T13:23:04.124787320Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124882 containerd[1781]: time="2025-01-30T13:23:04.124802040Z" level=info msg="loading plugin \"io.containerd.transfer.v1.local\"..." type=io.containerd.transfer.v1 Jan 30 13:23:04.124882 containerd[1781]: time="2025-01-30T13:23:04.124822480Z" level=info msg="loading plugin \"io.containerd.grpc.v1.transfer\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124882 containerd[1781]: time="2025-01-30T13:23:04.124836920Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.124882 containerd[1781]: time="2025-01-30T13:23:04.124847800Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Jan 30 13:23:04.125027 containerd[1781]: time="2025-01-30T13:23:04.124906000Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Jan 30 13:23:04.125027 containerd[1781]: time="2025-01-30T13:23:04.124926240Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.tracing.processor.v1 Jan 30 13:23:04.125027 containerd[1781]: time="2025-01-30T13:23:04.124936600Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Jan 30 13:23:04.125027 containerd[1781]: time="2025-01-30T13:23:04.124948600Z" level=info msg="skip loading plugin \"io.containerd.internal.v1.tracing\"..." error="skip plugin: tracing endpoint not configured" type=io.containerd.internal.v1 Jan 30 13:23:04.125027 containerd[1781]: time="2025-01-30T13:23:04.124957880Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.125027 containerd[1781]: time="2025-01-30T13:23:04.124969840Z" level=info msg="loading plugin \"io.containerd.nri.v1.nri\"..." type=io.containerd.nri.v1 Jan 30 13:23:04.125027 containerd[1781]: time="2025-01-30T13:23:04.124979680Z" level=info msg="NRI interface is disabled by configuration." Jan 30 13:23:04.125027 containerd[1781]: time="2025-01-30T13:23:04.124990800Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Jan 30 13:23:04.125329 containerd[1781]: time="2025-01-30T13:23:04.125270800Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false PrivilegedWithoutHostDevicesAllDevicesAllowed:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0 Snapshotter: SandboxMode:podsandbox}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreBlockIONotEnabledErrors:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginSetupSerially:false NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.8 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false EnableCDI:false CDISpecDirs:[/etc/cdi /var/run/cdi] ImagePullProgressTimeout:5m0s DrainExecSyncIOTimeout:0s ImagePullWithSyncFs:false IgnoreDeprecationWarnings:[]} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Jan 30 13:23:04.125329 containerd[1781]: time="2025-01-30T13:23:04.125332560Z" level=info msg="Connect containerd service" Jan 30 13:23:04.125467 containerd[1781]: time="2025-01-30T13:23:04.125368840Z" level=info msg="using legacy CRI server" Jan 30 13:23:04.125467 containerd[1781]: time="2025-01-30T13:23:04.125375280Z" level=info msg="using experimental NRI integration - disable nri plugin to prevent this" Jan 30 13:23:04.125509 containerd[1781]: time="2025-01-30T13:23:04.125495680Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Jan 30 13:23:04.127015 containerd[1781]: time="2025-01-30T13:23:04.126959360Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 30 13:23:04.127594 containerd[1781]: time="2025-01-30T13:23:04.127187360Z" level=info msg="Start subscribing containerd event" Jan 30 13:23:04.127594 containerd[1781]: time="2025-01-30T13:23:04.127235560Z" level=info msg="Start recovering state" Jan 30 13:23:04.127594 containerd[1781]: time="2025-01-30T13:23:04.127306760Z" level=info msg="Start event monitor" Jan 30 13:23:04.127594 containerd[1781]: time="2025-01-30T13:23:04.127318520Z" level=info msg="Start snapshots syncer" Jan 30 13:23:04.127594 containerd[1781]: time="2025-01-30T13:23:04.127327840Z" level=info msg="Start cni network conf syncer for default" Jan 30 13:23:04.127594 containerd[1781]: time="2025-01-30T13:23:04.127335200Z" level=info msg="Start streaming server" Jan 30 13:23:04.127942 containerd[1781]: time="2025-01-30T13:23:04.127921720Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Jan 30 13:23:04.128041 containerd[1781]: time="2025-01-30T13:23:04.128028240Z" level=info msg=serving... address=/run/containerd/containerd.sock Jan 30 13:23:04.128223 systemd[1]: Started containerd.service - containerd container runtime. Jan 30 13:23:04.135208 containerd[1781]: time="2025-01-30T13:23:04.135086400Z" level=info msg="containerd successfully booted in 0.067237s" Jan 30 13:23:04.267440 tar[1774]: linux-arm64/LICENSE Jan 30 13:23:04.268566 tar[1774]: linux-arm64/README.md Jan 30 13:23:04.282790 systemd[1]: Finished prepare-helm.service - Unpack helm to /opt/bin. Jan 30 13:23:04.453899 systemd-networkd[1549]: enP6611s1: Gained IPv6LL Jan 30 13:23:04.519492 sshd_keygen[1723]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Jan 30 13:23:04.539143 systemd[1]: Finished sshd-keygen.service - Generate sshd host keys. Jan 30 13:23:04.552004 systemd[1]: Starting issuegen.service - Generate /run/issue... Jan 30 13:23:04.566083 systemd[1]: issuegen.service: Deactivated successfully. Jan 30 13:23:04.566258 systemd[1]: Finished issuegen.service - Generate /run/issue. Jan 30 13:23:04.580036 systemd[1]: Starting systemd-user-sessions.service - Permit User Sessions... Jan 30 13:23:04.596491 systemd[1]: Finished systemd-user-sessions.service - Permit User Sessions. Jan 30 13:23:04.608113 systemd[1]: Started getty@tty1.service - Getty on tty1. Jan 30 13:23:04.615051 systemd[1]: Started serial-getty@ttyAMA0.service - Serial Getty on ttyAMA0. Jan 30 13:23:04.622197 systemd[1]: Reached target getty.target - Login Prompts. Jan 30 13:23:04.773844 systemd-networkd[1549]: eth0: Gained IPv6LL Jan 30 13:23:04.776427 systemd[1]: Finished systemd-networkd-wait-online.service - Wait for Network to be Configured. Jan 30 13:23:04.784877 systemd[1]: Reached target network-online.target - Network is Online. Jan 30 13:23:04.796947 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:23:04.807077 systemd[1]: Starting nvidia.service - NVIDIA Configure Service... Jan 30 13:23:04.816992 systemd[1]: Starting waagent.service - Microsoft Azure Linux Agent... Jan 30 13:23:04.855379 systemd[1]: Started waagent.service - Microsoft Azure Linux Agent. Jan 30 13:23:04.863970 systemd[1]: Finished nvidia.service - NVIDIA Configure Service. Jan 30 13:23:05.456950 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:23:05.468888 (kubelet)[1869]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 13:23:05.471840 systemd[1]: Reached target multi-user.target - Multi-User System. Jan 30 13:23:05.481893 systemd[1]: Startup finished in 728ms (kernel) + 19.902s (initrd) + 15.948s (userspace) = 36.579s. Jan 30 13:23:05.522282 agetty[1843]: failed to open credentials directory Jan 30 13:23:05.523768 agetty[1844]: failed to open credentials directory Jan 30 13:23:05.548349 waagent[1860]: 2025-01-30T13:23:05.548245Z INFO Daemon Daemon Azure Linux Agent Version: 2.9.1.1 Jan 30 13:23:05.566350 waagent[1860]: 2025-01-30T13:23:05.566246Z INFO Daemon Daemon OS: flatcar 4186.1.0 Jan 30 13:23:05.572978 waagent[1860]: 2025-01-30T13:23:05.572888Z INFO Daemon Daemon Python: 3.11.10 Jan 30 13:23:05.581401 waagent[1860]: 2025-01-30T13:23:05.581048Z INFO Daemon Daemon Run daemon Jan 30 13:23:05.587872 waagent[1860]: 2025-01-30T13:23:05.586454Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='4186.1.0' Jan 30 13:23:05.597622 waagent[1860]: 2025-01-30T13:23:05.597316Z INFO Daemon Daemon Using waagent for provisioning Jan 30 13:23:05.603416 waagent[1860]: 2025-01-30T13:23:05.603334Z INFO Daemon Daemon Activate resource disk Jan 30 13:23:05.608816 waagent[1860]: 2025-01-30T13:23:05.608400Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Jan 30 13:23:05.622878 waagent[1860]: 2025-01-30T13:23:05.622413Z INFO Daemon Daemon Found device: None Jan 30 13:23:05.627313 waagent[1860]: 2025-01-30T13:23:05.627233Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Jan 30 13:23:05.637734 waagent[1860]: 2025-01-30T13:23:05.637457Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Jan 30 13:23:05.643631 login[1843]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:23:05.653694 waagent[1860]: 2025-01-30T13:23:05.650763Z INFO Daemon Daemon Clean protocol and wireserver endpoint Jan 30 13:23:05.656821 login[1844]: pam_unix(login:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:23:05.663853 waagent[1860]: 2025-01-30T13:23:05.663762Z INFO Daemon Daemon Running default provisioning handler Jan 30 13:23:05.673891 systemd-logind[1717]: New session 1 of user core. Jan 30 13:23:05.675305 systemd[1]: Created slice user-500.slice - User Slice of UID 500. Jan 30 13:23:05.681078 systemd[1]: Starting user-runtime-dir@500.service - User Runtime Directory /run/user/500... Jan 30 13:23:05.686034 waagent[1860]: 2025-01-30T13:23:05.685883Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 4. Jan 30 13:23:05.688294 systemd-logind[1717]: New session 2 of user core. Jan 30 13:23:05.705416 waagent[1860]: 2025-01-30T13:23:05.705317Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Jan 30 13:23:05.715888 systemd[1]: Finished user-runtime-dir@500.service - User Runtime Directory /run/user/500. Jan 30 13:23:05.720162 waagent[1860]: 2025-01-30T13:23:05.720046Z INFO Daemon Daemon cloud-init is enabled: False Jan 30 13:23:05.728872 waagent[1860]: 2025-01-30T13:23:05.728692Z INFO Daemon Daemon Copying ovf-env.xml Jan 30 13:23:05.737200 systemd[1]: Starting user@500.service - User Manager for UID 500... Jan 30 13:23:05.750425 (systemd)[1883]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Jan 30 13:23:05.793404 waagent[1860]: 2025-01-30T13:23:05.788118Z INFO Daemon Daemon Successfully mounted dvd Jan 30 13:23:05.811595 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Jan 30 13:23:05.814357 waagent[1860]: 2025-01-30T13:23:05.814251Z INFO Daemon Daemon Detect protocol endpoint Jan 30 13:23:05.820500 waagent[1860]: 2025-01-30T13:23:05.820399Z INFO Daemon Daemon Clean protocol and wireserver endpoint Jan 30 13:23:05.827877 waagent[1860]: 2025-01-30T13:23:05.827766Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Jan 30 13:23:05.835757 waagent[1860]: 2025-01-30T13:23:05.835644Z INFO Daemon Daemon Test for route to 168.63.129.16 Jan 30 13:23:05.846823 waagent[1860]: 2025-01-30T13:23:05.846278Z INFO Daemon Daemon Route to 168.63.129.16 exists Jan 30 13:23:05.863245 waagent[1860]: 2025-01-30T13:23:05.863135Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Jan 30 13:23:05.911600 waagent[1860]: 2025-01-30T13:23:05.911538Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Jan 30 13:23:05.935326 waagent[1860]: 2025-01-30T13:23:05.935272Z INFO Daemon Daemon Wire protocol version:2012-11-30 Jan 30 13:23:05.951272 waagent[1860]: 2025-01-30T13:23:05.951157Z INFO Daemon Daemon Server preferred version:2015-04-05 Jan 30 13:23:05.968123 systemd[1883]: Queued start job for default target default.target. Jan 30 13:23:05.978042 systemd[1883]: Created slice app.slice - User Application Slice. Jan 30 13:23:05.978075 systemd[1883]: Reached target paths.target - Paths. Jan 30 13:23:05.978088 systemd[1883]: Reached target timers.target - Timers. Jan 30 13:23:05.980933 systemd[1883]: Starting dbus.socket - D-Bus User Message Bus Socket... Jan 30 13:23:05.993394 systemd[1883]: Listening on dbus.socket - D-Bus User Message Bus Socket. Jan 30 13:23:05.993462 systemd[1883]: Reached target sockets.target - Sockets. Jan 30 13:23:05.993475 systemd[1883]: Reached target basic.target - Basic System. Jan 30 13:23:05.993537 systemd[1883]: Reached target default.target - Main User Target. Jan 30 13:23:05.993565 systemd[1883]: Startup finished in 227ms. Jan 30 13:23:05.993688 systemd[1]: Started user@500.service - User Manager for UID 500. Jan 30 13:23:05.997938 systemd[1]: Started session-1.scope - Session 1 of User core. Jan 30 13:23:05.999543 systemd[1]: Started session-2.scope - Session 2 of User core. Jan 30 13:23:06.137089 kubelet[1869]: E0130 13:23:06.137000 1869 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 13:23:06.139920 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 13:23:06.140075 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 13:23:06.219804 waagent[1860]: 2025-01-30T13:23:06.218960Z INFO Daemon Daemon Initializing goal state during protocol detection Jan 30 13:23:06.226011 waagent[1860]: 2025-01-30T13:23:06.225930Z INFO Daemon Daemon Forcing an update of the goal state. Jan 30 13:23:06.239788 waagent[1860]: 2025-01-30T13:23:06.239701Z INFO Daemon Fetched a new incarnation for the WireServer goal state [incarnation 1] Jan 30 13:23:06.304616 waagent[1860]: 2025-01-30T13:23:06.304556Z INFO Daemon Daemon HostGAPlugin version: 1.0.8.159 Jan 30 13:23:06.311934 waagent[1860]: 2025-01-30T13:23:06.311878Z INFO Daemon Jan 30 13:23:06.315415 waagent[1860]: 2025-01-30T13:23:06.315355Z INFO Daemon Fetched new vmSettings [HostGAPlugin correlation ID: d4da2e0b-c4d6-4716-bc52-7a396ef8e81f eTag: 15190768693255875840 source: Fabric] Jan 30 13:23:06.328278 waagent[1860]: 2025-01-30T13:23:06.328224Z INFO Daemon The vmSettings originated via Fabric; will ignore them. Jan 30 13:23:06.335683 waagent[1860]: 2025-01-30T13:23:06.335629Z INFO Daemon Jan 30 13:23:06.338656 waagent[1860]: 2025-01-30T13:23:06.338594Z INFO Daemon Fetching full goal state from the WireServer [incarnation 1] Jan 30 13:23:06.350969 waagent[1860]: 2025-01-30T13:23:06.350922Z INFO Daemon Daemon Downloading artifacts profile blob Jan 30 13:23:06.441748 waagent[1860]: 2025-01-30T13:23:06.437969Z INFO Daemon Downloaded certificate {'thumbprint': 'F3B1945C6DEB95EFA3F87AF728C1A410F3714569', 'hasPrivateKey': True} Jan 30 13:23:06.448782 waagent[1860]: 2025-01-30T13:23:06.448701Z INFO Daemon Downloaded certificate {'thumbprint': 'DB3A6A25068C54721FD2100CD3523D73AFD0670D', 'hasPrivateKey': False} Jan 30 13:23:06.460272 waagent[1860]: 2025-01-30T13:23:06.460214Z INFO Daemon Fetch goal state completed Jan 30 13:23:06.474060 waagent[1860]: 2025-01-30T13:23:06.473970Z INFO Daemon Daemon Starting provisioning Jan 30 13:23:06.479427 waagent[1860]: 2025-01-30T13:23:06.479356Z INFO Daemon Daemon Handle ovf-env.xml. Jan 30 13:23:06.484602 waagent[1860]: 2025-01-30T13:23:06.484542Z INFO Daemon Daemon Set hostname [ci-4186.1.0-a-a7402f3063] Jan 30 13:23:06.497070 waagent[1860]: 2025-01-30T13:23:06.496981Z INFO Daemon Daemon Publish hostname [ci-4186.1.0-a-a7402f3063] Jan 30 13:23:06.504046 waagent[1860]: 2025-01-30T13:23:06.503973Z INFO Daemon Daemon Examine /proc/net/route for primary interface Jan 30 13:23:06.511020 waagent[1860]: 2025-01-30T13:23:06.510943Z INFO Daemon Daemon Primary interface is [eth0] Jan 30 13:23:06.530175 systemd-networkd[1549]: eth0: found matching network '/usr/lib/systemd/network/zz-default.network', based on potentially unpredictable interface name. Jan 30 13:23:06.530183 systemd-networkd[1549]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Jan 30 13:23:06.530212 systemd-networkd[1549]: eth0: DHCP lease lost Jan 30 13:23:06.531710 waagent[1860]: 2025-01-30T13:23:06.531607Z INFO Daemon Daemon Create user account if not exists Jan 30 13:23:06.537849 waagent[1860]: 2025-01-30T13:23:06.537775Z INFO Daemon Daemon User core already exists, skip useradd Jan 30 13:23:06.543778 systemd-networkd[1549]: eth0: DHCPv6 lease lost Jan 30 13:23:06.544192 waagent[1860]: 2025-01-30T13:23:06.543954Z INFO Daemon Daemon Configure sudoer Jan 30 13:23:06.549284 waagent[1860]: 2025-01-30T13:23:06.549182Z INFO Daemon Daemon Configure sshd Jan 30 13:23:06.554256 waagent[1860]: 2025-01-30T13:23:06.554182Z INFO Daemon Daemon Added a configuration snippet disabling SSH password-based authentication methods. It also configures SSH client probing to keep connections alive. Jan 30 13:23:06.568579 waagent[1860]: 2025-01-30T13:23:06.568483Z INFO Daemon Daemon Deploy ssh public key. Jan 30 13:23:06.578295 systemd-networkd[1549]: eth0: DHCPv4 address 10.200.20.40/24, gateway 10.200.20.1 acquired from 168.63.129.16 Jan 30 13:23:07.825522 waagent[1860]: 2025-01-30T13:23:07.825449Z INFO Daemon Daemon Provisioning complete Jan 30 13:23:07.847894 waagent[1860]: 2025-01-30T13:23:07.847843Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Jan 30 13:23:07.855081 waagent[1860]: 2025-01-30T13:23:07.855014Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Jan 30 13:23:07.865874 waagent[1860]: 2025-01-30T13:23:07.865810Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.9.1.1 is the most current agent Jan 30 13:23:08.011464 waagent[1938]: 2025-01-30T13:23:08.010909Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.9.1.1) Jan 30 13:23:08.011464 waagent[1938]: 2025-01-30T13:23:08.011072Z INFO ExtHandler ExtHandler OS: flatcar 4186.1.0 Jan 30 13:23:08.011464 waagent[1938]: 2025-01-30T13:23:08.011125Z INFO ExtHandler ExtHandler Python: 3.11.10 Jan 30 13:23:08.023833 waagent[1938]: 2025-01-30T13:23:08.023745Z INFO ExtHandler ExtHandler Distro: flatcar-4186.1.0; OSUtil: FlatcarUtil; AgentService: waagent; Python: 3.11.10; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Jan 30 13:23:08.024166 waagent[1938]: 2025-01-30T13:23:08.024127Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Jan 30 13:23:08.024307 waagent[1938]: 2025-01-30T13:23:08.024274Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Jan 30 13:23:08.033537 waagent[1938]: 2025-01-30T13:23:08.033444Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Jan 30 13:23:08.041762 waagent[1938]: 2025-01-30T13:23:08.041695Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.159 Jan 30 13:23:08.042498 waagent[1938]: 2025-01-30T13:23:08.042453Z INFO ExtHandler Jan 30 13:23:08.043742 waagent[1938]: 2025-01-30T13:23:08.042635Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: b417f06b-76a3-4ac7-9fbd-11eef35a2614 eTag: 15190768693255875840 source: Fabric] Jan 30 13:23:08.043742 waagent[1938]: 2025-01-30T13:23:08.043023Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Jan 30 13:23:08.043742 waagent[1938]: 2025-01-30T13:23:08.043590Z INFO ExtHandler Jan 30 13:23:08.043742 waagent[1938]: 2025-01-30T13:23:08.043658Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Jan 30 13:23:08.048012 waagent[1938]: 2025-01-30T13:23:08.047969Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Jan 30 13:23:08.127554 waagent[1938]: 2025-01-30T13:23:08.127406Z INFO ExtHandler Downloaded certificate {'thumbprint': 'F3B1945C6DEB95EFA3F87AF728C1A410F3714569', 'hasPrivateKey': True} Jan 30 13:23:08.128012 waagent[1938]: 2025-01-30T13:23:08.127954Z INFO ExtHandler Downloaded certificate {'thumbprint': 'DB3A6A25068C54721FD2100CD3523D73AFD0670D', 'hasPrivateKey': False} Jan 30 13:23:08.128470 waagent[1938]: 2025-01-30T13:23:08.128424Z INFO ExtHandler Fetch goal state completed Jan 30 13:23:08.148428 waagent[1938]: 2025-01-30T13:23:08.148348Z INFO ExtHandler ExtHandler WALinuxAgent-2.9.1.1 running as process 1938 Jan 30 13:23:08.148603 waagent[1938]: 2025-01-30T13:23:08.148564Z INFO ExtHandler ExtHandler ******** AutoUpdate.Enabled is set to False, not processing the operation ******** Jan 30 13:23:08.150453 waagent[1938]: 2025-01-30T13:23:08.150394Z INFO ExtHandler ExtHandler Cgroup monitoring is not supported on ['flatcar', '4186.1.0', '', 'Flatcar Container Linux by Kinvolk'] Jan 30 13:23:08.150877 waagent[1938]: 2025-01-30T13:23:08.150836Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Jan 30 13:23:08.366113 waagent[1938]: 2025-01-30T13:23:08.366063Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Jan 30 13:23:08.366327 waagent[1938]: 2025-01-30T13:23:08.366287Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Jan 30 13:23:08.373964 waagent[1938]: 2025-01-30T13:23:08.373911Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Jan 30 13:23:08.381249 systemd[1]: Reloading requested from client PID 1953 ('systemctl') (unit waagent.service)... Jan 30 13:23:08.381555 systemd[1]: Reloading... Jan 30 13:23:08.474752 zram_generator::config[1987]: No configuration found. Jan 30 13:23:08.589100 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 13:23:08.671383 systemd[1]: Reloading finished in 289 ms. Jan 30 13:23:08.696451 waagent[1938]: 2025-01-30T13:23:08.696159Z INFO ExtHandler ExtHandler Executing systemctl daemon-reload for setting up waagent-network-setup.service Jan 30 13:23:08.703098 systemd[1]: Reloading requested from client PID 2041 ('systemctl') (unit waagent.service)... Jan 30 13:23:08.703120 systemd[1]: Reloading... Jan 30 13:23:08.782765 zram_generator::config[2075]: No configuration found. Jan 30 13:23:08.900504 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 13:23:08.991738 systemd[1]: Reloading finished in 288 ms. Jan 30 13:23:09.020505 waagent[1938]: 2025-01-30T13:23:09.019510Z INFO ExtHandler ExtHandler Successfully added and enabled the waagent-network-setup.service Jan 30 13:23:09.020505 waagent[1938]: 2025-01-30T13:23:09.019752Z INFO ExtHandler ExtHandler Persistent firewall rules setup successfully Jan 30 13:23:14.442775 waagent[1938]: 2025-01-30T13:23:14.442066Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Jan 30 13:23:14.443090 waagent[1938]: 2025-01-30T13:23:14.442768Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: configuration enabled [True], cgroups enabled [False], python supported: [True] Jan 30 13:23:14.443709 waagent[1938]: 2025-01-30T13:23:14.443583Z INFO ExtHandler ExtHandler Starting env monitor service. Jan 30 13:23:14.444149 waagent[1938]: 2025-01-30T13:23:14.444073Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Jan 30 13:23:14.444430 waagent[1938]: 2025-01-30T13:23:14.444355Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Jan 30 13:23:14.444663 waagent[1938]: 2025-01-30T13:23:14.444600Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Jan 30 13:23:14.444836 waagent[1938]: 2025-01-30T13:23:14.444764Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Jan 30 13:23:14.446192 waagent[1938]: 2025-01-30T13:23:14.445239Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Jan 30 13:23:14.446192 waagent[1938]: 2025-01-30T13:23:14.445342Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Jan 30 13:23:14.446192 waagent[1938]: 2025-01-30T13:23:14.445494Z INFO EnvHandler ExtHandler Configure routes Jan 30 13:23:14.446192 waagent[1938]: 2025-01-30T13:23:14.445558Z INFO EnvHandler ExtHandler Gateway:None Jan 30 13:23:14.446192 waagent[1938]: 2025-01-30T13:23:14.445604Z INFO EnvHandler ExtHandler Routes:None Jan 30 13:23:14.446535 waagent[1938]: 2025-01-30T13:23:14.446478Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Jan 30 13:23:14.446929 waagent[1938]: 2025-01-30T13:23:14.446878Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Jan 30 13:23:14.447246 waagent[1938]: 2025-01-30T13:23:14.447197Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Jan 30 13:23:14.447246 waagent[1938]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Jan 30 13:23:14.447246 waagent[1938]: eth0 00000000 0114C80A 0003 0 0 1024 00000000 0 0 0 Jan 30 13:23:14.447246 waagent[1938]: eth0 0014C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Jan 30 13:23:14.447246 waagent[1938]: eth0 0114C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Jan 30 13:23:14.447246 waagent[1938]: eth0 10813FA8 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Jan 30 13:23:14.447246 waagent[1938]: eth0 FEA9FEA9 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Jan 30 13:23:14.448076 waagent[1938]: 2025-01-30T13:23:14.448019Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Jan 30 13:23:14.448185 waagent[1938]: 2025-01-30T13:23:14.448108Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Jan 30 13:23:14.448620 waagent[1938]: 2025-01-30T13:23:14.448547Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Jan 30 13:23:14.460504 waagent[1938]: 2025-01-30T13:23:14.460434Z INFO ExtHandler ExtHandler Jan 30 13:23:14.460614 waagent[1938]: 2025-01-30T13:23:14.460565Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: 08960a8f-a706-422b-bfaa-075920f9d756 correlation 474297d2-6001-4496-96a4-6af82ea9acc9 created: 2025-01-30T13:21:24.558632Z] Jan 30 13:23:14.461137 waagent[1938]: 2025-01-30T13:23:14.461079Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Jan 30 13:23:14.462433 waagent[1938]: 2025-01-30T13:23:14.462361Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 1 ms] Jan 30 13:23:14.503510 waagent[1938]: 2025-01-30T13:23:14.503455Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.9.1.1 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: 0EE2B8D6-6FF5-4443-8023-CEB8B95250F4;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 0] Jan 30 13:23:14.723603 waagent[1938]: 2025-01-30T13:23:14.723096Z INFO MonitorHandler ExtHandler Network interfaces: Jan 30 13:23:14.723603 waagent[1938]: Executing ['ip', '-a', '-o', 'link']: Jan 30 13:23:14.723603 waagent[1938]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Jan 30 13:23:14.723603 waagent[1938]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 00:22:48:ba:7e:fb brd ff:ff:ff:ff:ff:ff Jan 30 13:23:14.723603 waagent[1938]: 3: enP6611s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 00:22:48:ba:7e:fb brd ff:ff:ff:ff:ff:ff\ altname enP6611p0s2 Jan 30 13:23:14.723603 waagent[1938]: Executing ['ip', '-4', '-a', '-o', 'address']: Jan 30 13:23:14.723603 waagent[1938]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Jan 30 13:23:14.723603 waagent[1938]: 2: eth0 inet 10.200.20.40/24 metric 1024 brd 10.200.20.255 scope global eth0\ valid_lft forever preferred_lft forever Jan 30 13:23:14.723603 waagent[1938]: Executing ['ip', '-6', '-a', '-o', 'address']: Jan 30 13:23:14.723603 waagent[1938]: 1: lo inet6 ::1/128 scope host noprefixroute \ valid_lft forever preferred_lft forever Jan 30 13:23:14.723603 waagent[1938]: 2: eth0 inet6 fe80::222:48ff:feba:7efb/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Jan 30 13:23:14.723603 waagent[1938]: 3: enP6611s1 inet6 fe80::222:48ff:feba:7efb/64 scope link proto kernel_ll \ valid_lft forever preferred_lft forever Jan 30 13:23:14.791942 waagent[1938]: 2025-01-30T13:23:14.791711Z INFO EnvHandler ExtHandler Successfully added Azure fabric firewall rules. Current Firewall rules: Jan 30 13:23:14.791942 waagent[1938]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Jan 30 13:23:14.791942 waagent[1938]: pkts bytes target prot opt in out source destination Jan 30 13:23:14.791942 waagent[1938]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Jan 30 13:23:14.791942 waagent[1938]: pkts bytes target prot opt in out source destination Jan 30 13:23:14.791942 waagent[1938]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Jan 30 13:23:14.791942 waagent[1938]: pkts bytes target prot opt in out source destination Jan 30 13:23:14.791942 waagent[1938]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Jan 30 13:23:14.791942 waagent[1938]: 1 52 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Jan 30 13:23:14.791942 waagent[1938]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Jan 30 13:23:14.796111 waagent[1938]: 2025-01-30T13:23:14.795644Z INFO EnvHandler ExtHandler Current Firewall rules: Jan 30 13:23:14.796111 waagent[1938]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Jan 30 13:23:14.796111 waagent[1938]: pkts bytes target prot opt in out source destination Jan 30 13:23:14.796111 waagent[1938]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Jan 30 13:23:14.796111 waagent[1938]: pkts bytes target prot opt in out source destination Jan 30 13:23:14.796111 waagent[1938]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Jan 30 13:23:14.796111 waagent[1938]: pkts bytes target prot opt in out source destination Jan 30 13:23:14.796111 waagent[1938]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Jan 30 13:23:14.796111 waagent[1938]: 1 52 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Jan 30 13:23:14.796111 waagent[1938]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Jan 30 13:23:14.796111 waagent[1938]: 2025-01-30T13:23:14.795989Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Jan 30 13:23:16.201217 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Jan 30 13:23:16.210934 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:23:16.657100 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:23:16.661960 (kubelet)[2172]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 13:23:16.732694 kubelet[2172]: E0130 13:23:16.732608 2172 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 13:23:16.736022 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 13:23:16.736188 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 13:23:26.951434 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Jan 30 13:23:26.960921 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:23:27.063769 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:23:27.074018 (kubelet)[2189]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 13:23:27.135969 chronyd[1699]: Selected source PHC0 Jan 30 13:23:27.143020 kubelet[2189]: E0130 13:23:27.142962 2189 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 13:23:27.145302 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 13:23:27.145432 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 13:23:31.728949 systemd[1]: Created slice system-sshd.slice - Slice /system/sshd. Jan 30 13:23:31.734997 systemd[1]: Started sshd@0-10.200.20.40:22-10.200.16.10:58980.service - OpenSSH per-connection server daemon (10.200.16.10:58980). Jan 30 13:23:32.366538 sshd[2198]: Accepted publickey for core from 10.200.16.10 port 58980 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:23:32.368041 sshd-session[2198]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:23:32.372232 systemd-logind[1717]: New session 3 of user core. Jan 30 13:23:32.377914 systemd[1]: Started session-3.scope - Session 3 of User core. Jan 30 13:23:32.755650 systemd[1]: Started sshd@1-10.200.20.40:22-10.200.16.10:58982.service - OpenSSH per-connection server daemon (10.200.16.10:58982). Jan 30 13:23:33.193315 sshd[2203]: Accepted publickey for core from 10.200.16.10 port 58982 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:23:33.194630 sshd-session[2203]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:23:33.199948 systemd-logind[1717]: New session 4 of user core. Jan 30 13:23:33.204919 systemd[1]: Started session-4.scope - Session 4 of User core. Jan 30 13:23:33.511844 sshd[2205]: Connection closed by 10.200.16.10 port 58982 Jan 30 13:23:33.512608 sshd-session[2203]: pam_unix(sshd:session): session closed for user core Jan 30 13:23:33.517359 systemd[1]: sshd@1-10.200.20.40:22-10.200.16.10:58982.service: Deactivated successfully. Jan 30 13:23:33.520672 systemd[1]: session-4.scope: Deactivated successfully. Jan 30 13:23:33.522000 systemd-logind[1717]: Session 4 logged out. Waiting for processes to exit. Jan 30 13:23:33.524317 systemd-logind[1717]: Removed session 4. Jan 30 13:23:33.591071 systemd[1]: Started sshd@2-10.200.20.40:22-10.200.16.10:58984.service - OpenSSH per-connection server daemon (10.200.16.10:58984). Jan 30 13:23:34.036411 sshd[2210]: Accepted publickey for core from 10.200.16.10 port 58984 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:23:34.037873 sshd-session[2210]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:23:34.042015 systemd-logind[1717]: New session 5 of user core. Jan 30 13:23:34.056898 systemd[1]: Started session-5.scope - Session 5 of User core. Jan 30 13:23:34.360905 sshd[2212]: Connection closed by 10.200.16.10 port 58984 Jan 30 13:23:34.359961 sshd-session[2210]: pam_unix(sshd:session): session closed for user core Jan 30 13:23:34.364088 systemd[1]: sshd@2-10.200.20.40:22-10.200.16.10:58984.service: Deactivated successfully. Jan 30 13:23:34.365833 systemd[1]: session-5.scope: Deactivated successfully. Jan 30 13:23:34.366523 systemd-logind[1717]: Session 5 logged out. Waiting for processes to exit. Jan 30 13:23:34.367421 systemd-logind[1717]: Removed session 5. Jan 30 13:23:34.439185 systemd[1]: Started sshd@3-10.200.20.40:22-10.200.16.10:58990.service - OpenSSH per-connection server daemon (10.200.16.10:58990). Jan 30 13:23:34.880867 sshd[2217]: Accepted publickey for core from 10.200.16.10 port 58990 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:23:34.882262 sshd-session[2217]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:23:34.887470 systemd-logind[1717]: New session 6 of user core. Jan 30 13:23:34.893929 systemd[1]: Started session-6.scope - Session 6 of User core. Jan 30 13:23:35.208561 sshd[2219]: Connection closed by 10.200.16.10 port 58990 Jan 30 13:23:35.209146 sshd-session[2217]: pam_unix(sshd:session): session closed for user core Jan 30 13:23:35.212974 systemd[1]: sshd@3-10.200.20.40:22-10.200.16.10:58990.service: Deactivated successfully. Jan 30 13:23:35.214640 systemd[1]: session-6.scope: Deactivated successfully. Jan 30 13:23:35.215400 systemd-logind[1717]: Session 6 logged out. Waiting for processes to exit. Jan 30 13:23:35.216540 systemd-logind[1717]: Removed session 6. Jan 30 13:23:35.292566 systemd[1]: Started sshd@4-10.200.20.40:22-10.200.16.10:59000.service - OpenSSH per-connection server daemon (10.200.16.10:59000). Jan 30 13:23:35.727638 sshd[2224]: Accepted publickey for core from 10.200.16.10 port 59000 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:23:35.729217 sshd-session[2224]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:23:35.733904 systemd-logind[1717]: New session 7 of user core. Jan 30 13:23:35.742134 systemd[1]: Started session-7.scope - Session 7 of User core. Jan 30 13:23:36.158110 sudo[2227]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/setenforce 1 Jan 30 13:23:36.158429 sudo[2227]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 13:23:36.202866 sudo[2227]: pam_unix(sudo:session): session closed for user root Jan 30 13:23:36.278348 sshd[2226]: Connection closed by 10.200.16.10 port 59000 Jan 30 13:23:36.279115 sshd-session[2224]: pam_unix(sshd:session): session closed for user core Jan 30 13:23:36.283259 systemd[1]: sshd@4-10.200.20.40:22-10.200.16.10:59000.service: Deactivated successfully. Jan 30 13:23:36.284877 systemd[1]: session-7.scope: Deactivated successfully. Jan 30 13:23:36.286064 systemd-logind[1717]: Session 7 logged out. Waiting for processes to exit. Jan 30 13:23:36.287061 systemd-logind[1717]: Removed session 7. Jan 30 13:23:36.368288 systemd[1]: Started sshd@5-10.200.20.40:22-10.200.16.10:44528.service - OpenSSH per-connection server daemon (10.200.16.10:44528). Jan 30 13:23:36.809644 sshd[2232]: Accepted publickey for core from 10.200.16.10 port 44528 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:23:36.811183 sshd-session[2232]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:23:36.815345 systemd-logind[1717]: New session 8 of user core. Jan 30 13:23:36.822901 systemd[1]: Started session-8.scope - Session 8 of User core. Jan 30 13:23:37.057599 sudo[2236]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/rm -rf /etc/audit/rules.d/80-selinux.rules /etc/audit/rules.d/99-default.rules Jan 30 13:23:37.058368 sudo[2236]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 13:23:37.062184 sudo[2236]: pam_unix(sudo:session): session closed for user root Jan 30 13:23:37.067921 sudo[2235]: core : PWD=/home/core ; USER=root ; COMMAND=/usr/sbin/systemctl restart audit-rules Jan 30 13:23:37.068221 sudo[2235]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 13:23:37.084085 systemd[1]: Starting audit-rules.service - Load Audit Rules... Jan 30 13:23:37.109001 augenrules[2258]: No rules Jan 30 13:23:37.110296 systemd[1]: audit-rules.service: Deactivated successfully. Jan 30 13:23:37.110504 systemd[1]: Finished audit-rules.service - Load Audit Rules. Jan 30 13:23:37.112394 sudo[2235]: pam_unix(sudo:session): session closed for user root Jan 30 13:23:37.191870 sshd[2234]: Connection closed by 10.200.16.10 port 44528 Jan 30 13:23:37.192305 sshd-session[2232]: pam_unix(sshd:session): session closed for user core Jan 30 13:23:37.196435 systemd[1]: sshd@5-10.200.20.40:22-10.200.16.10:44528.service: Deactivated successfully. Jan 30 13:23:37.199397 systemd[1]: session-8.scope: Deactivated successfully. Jan 30 13:23:37.200489 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 3. Jan 30 13:23:37.201063 systemd-logind[1717]: Session 8 logged out. Waiting for processes to exit. Jan 30 13:23:37.206962 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:23:37.207988 systemd-logind[1717]: Removed session 8. Jan 30 13:23:37.290033 systemd[1]: Started sshd@6-10.200.20.40:22-10.200.16.10:44536.service - OpenSSH per-connection server daemon (10.200.16.10:44536). Jan 30 13:23:37.578966 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:23:37.584357 (kubelet)[2276]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 13:23:37.628797 kubelet[2276]: E0130 13:23:37.628703 2276 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 13:23:37.630983 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 13:23:37.631117 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 13:23:37.723415 sshd[2269]: Accepted publickey for core from 10.200.16.10 port 44536 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:23:37.724970 sshd-session[2269]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:23:37.729223 systemd-logind[1717]: New session 9 of user core. Jan 30 13:23:37.737101 systemd[1]: Started session-9.scope - Session 9 of User core. Jan 30 13:23:37.971320 sudo[2285]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Jan 30 13:23:37.971621 sudo[2285]: pam_unix(sudo:session): session opened for user root(uid=0) by core(uid=500) Jan 30 13:23:38.769014 systemd[1]: Starting docker.service - Docker Application Container Engine... Jan 30 13:23:38.769198 (dockerd)[2303]: docker.service: Referenced but unset environment variable evaluates to an empty string: DOCKER_CGROUPS, DOCKER_OPTS, DOCKER_OPT_BIP, DOCKER_OPT_IPMASQ, DOCKER_OPT_MTU Jan 30 13:23:39.053754 dockerd[2303]: time="2025-01-30T13:23:39.051940416Z" level=info msg="Starting up" Jan 30 13:23:39.193708 dockerd[2303]: time="2025-01-30T13:23:39.193663721Z" level=info msg="Loading containers: start." Jan 30 13:23:39.334738 kernel: Initializing XFRM netlink socket Jan 30 13:23:39.392303 systemd-networkd[1549]: docker0: Link UP Jan 30 13:23:39.431124 dockerd[2303]: time="2025-01-30T13:23:39.431075122Z" level=info msg="Loading containers: done." Jan 30 13:23:39.451420 dockerd[2303]: time="2025-01-30T13:23:39.451364006Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Jan 30 13:23:39.451641 dockerd[2303]: time="2025-01-30T13:23:39.451488566Z" level=info msg="Docker daemon" commit=41ca978a0a5400cc24b274137efa9f25517fcc0b containerd-snapshotter=false storage-driver=overlay2 version=27.3.1 Jan 30 13:23:39.451668 dockerd[2303]: time="2025-01-30T13:23:39.451647806Z" level=info msg="Daemon has completed initialization" Jan 30 13:23:39.503747 dockerd[2303]: time="2025-01-30T13:23:39.503514175Z" level=info msg="API listen on /run/docker.sock" Jan 30 13:23:39.504277 systemd[1]: Started docker.service - Docker Application Container Engine. Jan 30 13:23:41.597215 containerd[1781]: time="2025-01-30T13:23:41.597172142Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.9\"" Jan 30 13:23:42.537857 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2898300046.mount: Deactivated successfully. Jan 30 13:23:46.911650 containerd[1781]: time="2025-01-30T13:23:46.911357798Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver:v1.30.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:23:46.914085 containerd[1781]: time="2025-01-30T13:23:46.914018198Z" level=info msg="stop pulling image registry.k8s.io/kube-apiserver:v1.30.9: active requests=0, bytes read=29864935" Jan 30 13:23:46.918433 containerd[1781]: time="2025-01-30T13:23:46.918381077Z" level=info msg="ImageCreate event name:\"sha256:5a490fe478de4f27039cf07d124901df2a58010e72f7afe3f65c70c05ada6715\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:23:46.922697 containerd[1781]: time="2025-01-30T13:23:46.922633516Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-apiserver@sha256:540de8f810ac963b8ed93f7393a8746d68e7e8a2c79ea58ff409ac5b9ca6a9fc\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:23:46.923846 containerd[1781]: time="2025-01-30T13:23:46.923628236Z" level=info msg="Pulled image \"registry.k8s.io/kube-apiserver:v1.30.9\" with image id \"sha256:5a490fe478de4f27039cf07d124901df2a58010e72f7afe3f65c70c05ada6715\", repo tag \"registry.k8s.io/kube-apiserver:v1.30.9\", repo digest \"registry.k8s.io/kube-apiserver@sha256:540de8f810ac963b8ed93f7393a8746d68e7e8a2c79ea58ff409ac5b9ca6a9fc\", size \"29861735\" in 5.326412014s" Jan 30 13:23:46.923846 containerd[1781]: time="2025-01-30T13:23:46.923674316Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.30.9\" returns image reference \"sha256:5a490fe478de4f27039cf07d124901df2a58010e72f7afe3f65c70c05ada6715\"" Jan 30 13:23:46.945495 containerd[1781]: time="2025-01-30T13:23:46.945218551Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.9\"" Jan 30 13:23:47.701224 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 4. Jan 30 13:23:47.708928 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:23:47.811728 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:23:47.824083 (kubelet)[2562]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 13:23:47.864694 kubelet[2562]: E0130 13:23:47.864633 2562 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 13:23:47.867421 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 13:23:47.867561 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 13:23:48.913456 update_engine[1724]: I20250130 13:23:48.912832 1724 update_attempter.cc:509] Updating boot flags... Jan 30 13:23:50.454025 kernel: hv_balloon: Max. dynamic memory size: 4096 MB Jan 30 13:23:50.660745 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (2584) Jan 30 13:23:50.779836 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (2584) Jan 30 13:23:50.905768 kernel: BTRFS warning: duplicate device /dev/sda3 devid 1 generation 39 scanned by (udev-worker) (2584) Jan 30 13:23:54.514895 containerd[1781]: time="2025-01-30T13:23:54.514838092Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager:v1.30.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:23:54.518823 containerd[1781]: time="2025-01-30T13:23:54.518770052Z" level=info msg="stop pulling image registry.k8s.io/kube-controller-manager:v1.30.9: active requests=0, bytes read=26901561" Jan 30 13:23:54.523736 containerd[1781]: time="2025-01-30T13:23:54.522734773Z" level=info msg="ImageCreate event name:\"sha256:cd43f1277f3b33fd1db15e7f98b093eb07e4d4530ff326356591daeb16369ca2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:23:54.529215 containerd[1781]: time="2025-01-30T13:23:54.529168773Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-controller-manager@sha256:6350693c04956b13db2519e01ca12a0bbe58466e9f12ef8617f1429da6081f43\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:23:54.529992 containerd[1781]: time="2025-01-30T13:23:54.529942253Z" level=info msg="Pulled image \"registry.k8s.io/kube-controller-manager:v1.30.9\" with image id \"sha256:cd43f1277f3b33fd1db15e7f98b093eb07e4d4530ff326356591daeb16369ca2\", repo tag \"registry.k8s.io/kube-controller-manager:v1.30.9\", repo digest \"registry.k8s.io/kube-controller-manager@sha256:6350693c04956b13db2519e01ca12a0bbe58466e9f12ef8617f1429da6081f43\", size \"28305351\" in 7.584666262s" Jan 30 13:23:54.529992 containerd[1781]: time="2025-01-30T13:23:54.529984613Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.30.9\" returns image reference \"sha256:cd43f1277f3b33fd1db15e7f98b093eb07e4d4530ff326356591daeb16369ca2\"" Jan 30 13:23:54.551599 containerd[1781]: time="2025-01-30T13:23:54.551563055Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.9\"" Jan 30 13:23:57.951242 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 5. Jan 30 13:23:57.959928 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:23:58.704775 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:23:58.709635 (kubelet)[2755]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 13:23:58.751832 kubelet[2755]: E0130 13:23:58.751783 2755 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 13:23:58.753952 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 13:23:58.754083 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 13:23:59.232750 containerd[1781]: time="2025-01-30T13:23:59.232680115Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler:v1.30.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:23:59.234753 containerd[1781]: time="2025-01-30T13:23:59.234474236Z" level=info msg="stop pulling image registry.k8s.io/kube-scheduler:v1.30.9: active requests=0, bytes read=16164338" Jan 30 13:23:59.237864 containerd[1781]: time="2025-01-30T13:23:59.237822396Z" level=info msg="ImageCreate event name:\"sha256:4ebb50f72fd1ba66a57f91b338174ab72034493ff261ebb9bbfd717d882178ce\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:23:59.243983 containerd[1781]: time="2025-01-30T13:23:59.243906996Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-scheduler@sha256:153efd6dc89e61a38ef273cf4c4cebd2bfee68082c2ee3d4fab5da94e4ae13d3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:23:59.245142 containerd[1781]: time="2025-01-30T13:23:59.244996796Z" level=info msg="Pulled image \"registry.k8s.io/kube-scheduler:v1.30.9\" with image id \"sha256:4ebb50f72fd1ba66a57f91b338174ab72034493ff261ebb9bbfd717d882178ce\", repo tag \"registry.k8s.io/kube-scheduler:v1.30.9\", repo digest \"registry.k8s.io/kube-scheduler@sha256:153efd6dc89e61a38ef273cf4c4cebd2bfee68082c2ee3d4fab5da94e4ae13d3\", size \"17568146\" in 4.693210141s" Jan 30 13:23:59.245142 containerd[1781]: time="2025-01-30T13:23:59.245040196Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.30.9\" returns image reference \"sha256:4ebb50f72fd1ba66a57f91b338174ab72034493ff261ebb9bbfd717d882178ce\"" Jan 30 13:23:59.267664 containerd[1781]: time="2025-01-30T13:23:59.267565638Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.9\"" Jan 30 13:24:00.333014 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3891867744.mount: Deactivated successfully. Jan 30 13:24:01.083125 containerd[1781]: time="2025-01-30T13:24:01.083062879Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy:v1.30.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:01.085606 containerd[1781]: time="2025-01-30T13:24:01.085417439Z" level=info msg="stop pulling image registry.k8s.io/kube-proxy:v1.30.9: active requests=0, bytes read=25662712" Jan 30 13:24:01.090752 containerd[1781]: time="2025-01-30T13:24:01.090681319Z" level=info msg="ImageCreate event name:\"sha256:d97113839930faa5ab88f70aff4bfb62f7381074a290dd5aadbec9b16b2567a2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:01.096488 containerd[1781]: time="2025-01-30T13:24:01.096424680Z" level=info msg="ImageCreate event name:\"registry.k8s.io/kube-proxy@sha256:d78dc40d97ff862fd8ddb47f80a5ba3feec17bc73e58a60e963885e33faa0083\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:01.097521 containerd[1781]: time="2025-01-30T13:24:01.097076440Z" level=info msg="Pulled image \"registry.k8s.io/kube-proxy:v1.30.9\" with image id \"sha256:d97113839930faa5ab88f70aff4bfb62f7381074a290dd5aadbec9b16b2567a2\", repo tag \"registry.k8s.io/kube-proxy:v1.30.9\", repo digest \"registry.k8s.io/kube-proxy@sha256:d78dc40d97ff862fd8ddb47f80a5ba3feec17bc73e58a60e963885e33faa0083\", size \"25661731\" in 1.829472082s" Jan 30 13:24:01.097521 containerd[1781]: time="2025-01-30T13:24:01.097114440Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.30.9\" returns image reference \"sha256:d97113839930faa5ab88f70aff4bfb62f7381074a290dd5aadbec9b16b2567a2\"" Jan 30 13:24:01.118058 containerd[1781]: time="2025-01-30T13:24:01.117989521Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\"" Jan 30 13:24:01.845885 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2100554896.mount: Deactivated successfully. Jan 30 13:24:03.138771 containerd[1781]: time="2025-01-30T13:24:03.138113496Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns:v1.11.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:03.140569 containerd[1781]: time="2025-01-30T13:24:03.140500936Z" level=info msg="stop pulling image registry.k8s.io/coredns/coredns:v1.11.1: active requests=0, bytes read=16485381" Jan 30 13:24:03.143890 containerd[1781]: time="2025-01-30T13:24:03.143826416Z" level=info msg="ImageCreate event name:\"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:03.149380 containerd[1781]: time="2025-01-30T13:24:03.148913336Z" level=info msg="ImageCreate event name:\"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:03.150209 containerd[1781]: time="2025-01-30T13:24:03.149645096Z" level=info msg="Pulled image \"registry.k8s.io/coredns/coredns:v1.11.1\" with image id \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\", repo tag \"registry.k8s.io/coredns/coredns:v1.11.1\", repo digest \"registry.k8s.io/coredns/coredns@sha256:1eeb4c7316bacb1d4c8ead65571cd92dd21e27359f0d4917f1a5822a73b75db1\", size \"16482581\" in 2.031585975s" Jan 30 13:24:03.150209 containerd[1781]: time="2025-01-30T13:24:03.149673336Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.11.1\" returns image reference \"sha256:2437cf762177702dec2dfe99a09c37427a15af6d9a57c456b65352667c223d93\"" Jan 30 13:24:03.173569 containerd[1781]: time="2025-01-30T13:24:03.173532298Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Jan 30 13:24:03.770884 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount170163167.mount: Deactivated successfully. Jan 30 13:24:03.792589 containerd[1781]: time="2025-01-30T13:24:03.792540459Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:03.795142 containerd[1781]: time="2025-01-30T13:24:03.795101659Z" level=info msg="stop pulling image registry.k8s.io/pause:3.9: active requests=0, bytes read=268821" Jan 30 13:24:03.799863 containerd[1781]: time="2025-01-30T13:24:03.799813420Z" level=info msg="ImageCreate event name:\"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:03.805614 containerd[1781]: time="2025-01-30T13:24:03.805556460Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:03.807026 containerd[1781]: time="2025-01-30T13:24:03.806570340Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.9\" with image id \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\", repo tag \"registry.k8s.io/pause:3.9\", repo digest \"registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097\", size \"268051\" in 632.785442ms" Jan 30 13:24:03.807026 containerd[1781]: time="2025-01-30T13:24:03.806606700Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" Jan 30 13:24:03.828898 containerd[1781]: time="2025-01-30T13:24:03.828846862Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\"" Jan 30 13:24:04.487372 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2186300291.mount: Deactivated successfully. Jan 30 13:24:08.951245 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 6. Jan 30 13:24:08.960095 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:24:09.075049 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:24:09.075675 (kubelet)[2890]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS, KUBELET_KUBEADM_ARGS Jan 30 13:24:09.125532 kubelet[2890]: E0130 13:24:09.125489 2890 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Jan 30 13:24:09.128996 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Jan 30 13:24:09.129142 systemd[1]: kubelet.service: Failed with result 'exit-code'. Jan 30 13:24:09.529776 containerd[1781]: time="2025-01-30T13:24:09.529201728Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd:3.5.12-0\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:09.531749 containerd[1781]: time="2025-01-30T13:24:09.531652929Z" level=info msg="stop pulling image registry.k8s.io/etcd:3.5.12-0: active requests=0, bytes read=66191472" Jan 30 13:24:09.538389 containerd[1781]: time="2025-01-30T13:24:09.538329889Z" level=info msg="ImageCreate event name:\"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:09.543669 containerd[1781]: time="2025-01-30T13:24:09.543606289Z" level=info msg="ImageCreate event name:\"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:09.544902 containerd[1781]: time="2025-01-30T13:24:09.544732329Z" level=info msg="Pulled image \"registry.k8s.io/etcd:3.5.12-0\" with image id \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\", repo tag \"registry.k8s.io/etcd:3.5.12-0\", repo digest \"registry.k8s.io/etcd@sha256:44a8e24dcbba3470ee1fee21d5e88d128c936e9b55d4bc51fbef8086f8ed123b\", size \"66189079\" in 5.715824227s" Jan 30 13:24:09.544902 containerd[1781]: time="2025-01-30T13:24:09.544777369Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.12-0\" returns image reference \"sha256:014faa467e29798aeef733fe6d1a3b5e382688217b053ad23410e6cccd5d22fd\"" Jan 30 13:24:13.804022 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:24:13.815969 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:24:13.836485 systemd[1]: Reloading requested from client PID 2964 ('systemctl') (unit session-9.scope)... Jan 30 13:24:13.836510 systemd[1]: Reloading... Jan 30 13:24:13.943758 zram_generator::config[3004]: No configuration found. Jan 30 13:24:14.062144 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 13:24:14.143268 systemd[1]: Reloading finished in 306 ms. Jan 30 13:24:14.190070 systemd[1]: kubelet.service: Control process exited, code=killed, status=15/TERM Jan 30 13:24:14.190162 systemd[1]: kubelet.service: Failed with result 'signal'. Jan 30 13:24:14.191786 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:24:14.197103 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:24:14.308771 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:24:14.321057 (kubelet)[3071]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 30 13:24:14.364056 kubelet[3071]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 13:24:14.364547 kubelet[3071]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 30 13:24:14.364547 kubelet[3071]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 13:24:14.365744 kubelet[3071]: I0130 13:24:14.365497 3071 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 30 13:24:14.740615 kubelet[3071]: I0130 13:24:14.740556 3071 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Jan 30 13:24:14.740615 kubelet[3071]: I0130 13:24:14.740590 3071 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 30 13:24:14.740905 kubelet[3071]: I0130 13:24:14.740862 3071 server.go:927] "Client rotation is on, will bootstrap in background" Jan 30 13:24:14.754444 kubelet[3071]: E0130 13:24:14.754390 3071 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.200.20.40:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:14.755465 kubelet[3071]: I0130 13:24:14.754622 3071 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 30 13:24:14.763340 kubelet[3071]: I0130 13:24:14.763248 3071 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 30 13:24:14.764240 kubelet[3071]: I0130 13:24:14.763679 3071 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 30 13:24:14.764240 kubelet[3071]: I0130 13:24:14.763736 3071 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4186.1.0-a-a7402f3063","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jan 30 13:24:14.764240 kubelet[3071]: I0130 13:24:14.763928 3071 topology_manager.go:138] "Creating topology manager with none policy" Jan 30 13:24:14.764240 kubelet[3071]: I0130 13:24:14.763939 3071 container_manager_linux.go:301] "Creating device plugin manager" Jan 30 13:24:14.764465 kubelet[3071]: I0130 13:24:14.764087 3071 state_mem.go:36] "Initialized new in-memory state store" Jan 30 13:24:14.764958 kubelet[3071]: I0130 13:24:14.764942 3071 kubelet.go:400] "Attempting to sync node with API server" Jan 30 13:24:14.765046 kubelet[3071]: I0130 13:24:14.765035 3071 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 30 13:24:14.765124 kubelet[3071]: I0130 13:24:14.765116 3071 kubelet.go:312] "Adding apiserver pod source" Jan 30 13:24:14.765188 kubelet[3071]: I0130 13:24:14.765179 3071 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 30 13:24:14.767047 kubelet[3071]: W0130 13:24:14.766839 3071 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.20.40:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4186.1.0-a-a7402f3063&limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:14.767047 kubelet[3071]: E0130 13:24:14.766903 3071 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.200.20.40:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4186.1.0-a-a7402f3063&limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:14.767315 kubelet[3071]: W0130 13:24:14.767267 3071 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.20.40:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:14.767315 kubelet[3071]: E0130 13:24:14.767314 3071 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.200.20.40:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:14.767429 kubelet[3071]: I0130 13:24:14.767409 3071 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Jan 30 13:24:14.767592 kubelet[3071]: I0130 13:24:14.767574 3071 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 30 13:24:14.767631 kubelet[3071]: W0130 13:24:14.767625 3071 probe.go:272] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Jan 30 13:24:14.769637 kubelet[3071]: I0130 13:24:14.769597 3071 server.go:1264] "Started kubelet" Jan 30 13:24:14.773544 kubelet[3071]: I0130 13:24:14.773348 3071 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 30 13:24:14.777439 kubelet[3071]: I0130 13:24:14.775689 3071 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 30 13:24:14.777439 kubelet[3071]: I0130 13:24:14.776089 3071 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 30 13:24:14.777439 kubelet[3071]: I0130 13:24:14.776113 3071 server.go:455] "Adding debug handlers to kubelet server" Jan 30 13:24:14.777439 kubelet[3071]: E0130 13:24:14.777164 3071 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.20.40:6443/api/v1/namespaces/default/events\": dial tcp 10.200.20.40:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4186.1.0-a-a7402f3063.181f7b3666b00fa3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4186.1.0-a-a7402f3063,UID:ci-4186.1.0-a-a7402f3063,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4186.1.0-a-a7402f3063,},FirstTimestamp:2025-01-30 13:24:14.769565603 +0000 UTC m=+0.445140904,LastTimestamp:2025-01-30 13:24:14.769565603 +0000 UTC m=+0.445140904,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4186.1.0-a-a7402f3063,}" Jan 30 13:24:14.779819 kubelet[3071]: I0130 13:24:14.779624 3071 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 30 13:24:14.783748 kubelet[3071]: I0130 13:24:14.781882 3071 volume_manager.go:291] "Starting Kubelet Volume Manager" Jan 30 13:24:14.783748 kubelet[3071]: I0130 13:24:14.782266 3071 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 30 13:24:14.783748 kubelet[3071]: I0130 13:24:14.782339 3071 reconciler.go:26] "Reconciler: start to sync state" Jan 30 13:24:14.783748 kubelet[3071]: E0130 13:24:14.782416 3071 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.40:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4186.1.0-a-a7402f3063?timeout=10s\": dial tcp 10.200.20.40:6443: connect: connection refused" interval="200ms" Jan 30 13:24:14.784021 kubelet[3071]: W0130 13:24:14.783694 3071 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.20.40:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:14.784102 kubelet[3071]: E0130 13:24:14.784091 3071 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.200.20.40:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:14.786276 kubelet[3071]: E0130 13:24:14.786241 3071 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 30 13:24:14.786621 kubelet[3071]: I0130 13:24:14.786604 3071 factory.go:221] Registration of the containerd container factory successfully Jan 30 13:24:14.786741 kubelet[3071]: I0130 13:24:14.786706 3071 factory.go:221] Registration of the systemd container factory successfully Jan 30 13:24:14.786891 kubelet[3071]: I0130 13:24:14.786872 3071 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 30 13:24:14.810117 kubelet[3071]: E0130 13:24:14.810000 3071 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://10.200.20.40:6443/api/v1/namespaces/default/events\": dial tcp 10.200.20.40:6443: connect: connection refused" event="&Event{ObjectMeta:{ci-4186.1.0-a-a7402f3063.181f7b3666b00fa3 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:ci-4186.1.0-a-a7402f3063,UID:ci-4186.1.0-a-a7402f3063,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:ci-4186.1.0-a-a7402f3063,},FirstTimestamp:2025-01-30 13:24:14.769565603 +0000 UTC m=+0.445140904,LastTimestamp:2025-01-30 13:24:14.769565603 +0000 UTC m=+0.445140904,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:ci-4186.1.0-a-a7402f3063,}" Jan 30 13:24:14.837134 kubelet[3071]: I0130 13:24:14.837091 3071 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 30 13:24:14.838884 kubelet[3071]: I0130 13:24:14.838832 3071 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 30 13:24:14.838884 kubelet[3071]: I0130 13:24:14.838893 3071 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 30 13:24:14.839042 kubelet[3071]: I0130 13:24:14.838918 3071 kubelet.go:2337] "Starting kubelet main sync loop" Jan 30 13:24:14.839042 kubelet[3071]: E0130 13:24:14.838974 3071 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 30 13:24:14.842186 kubelet[3071]: W0130 13:24:14.842022 3071 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.20.40:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:14.842339 kubelet[3071]: E0130 13:24:14.842183 3071 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.200.20.40:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:14.910960 kubelet[3071]: I0130 13:24:14.910923 3071 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:14.911489 kubelet[3071]: E0130 13:24:14.911403 3071 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.20.40:6443/api/v1/nodes\": dial tcp 10.200.20.40:6443: connect: connection refused" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:14.912169 kubelet[3071]: I0130 13:24:14.911893 3071 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 30 13:24:14.912169 kubelet[3071]: I0130 13:24:14.911909 3071 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 30 13:24:14.912169 kubelet[3071]: I0130 13:24:14.911929 3071 state_mem.go:36] "Initialized new in-memory state store" Jan 30 13:24:14.916935 kubelet[3071]: I0130 13:24:14.916915 3071 policy_none.go:49] "None policy: Start" Jan 30 13:24:14.917843 kubelet[3071]: I0130 13:24:14.917817 3071 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 30 13:24:14.917930 kubelet[3071]: I0130 13:24:14.917852 3071 state_mem.go:35] "Initializing new in-memory state store" Jan 30 13:24:14.926173 systemd[1]: Created slice kubepods.slice - libcontainer container kubepods.slice. Jan 30 13:24:14.938555 systemd[1]: Created slice kubepods-besteffort.slice - libcontainer container kubepods-besteffort.slice. Jan 30 13:24:14.939294 kubelet[3071]: E0130 13:24:14.939081 3071 kubelet.go:2361] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Jan 30 13:24:14.950263 systemd[1]: Created slice kubepods-burstable.slice - libcontainer container kubepods-burstable.slice. Jan 30 13:24:14.951684 kubelet[3071]: I0130 13:24:14.951636 3071 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 30 13:24:14.951953 kubelet[3071]: I0130 13:24:14.951906 3071 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 30 13:24:14.952168 kubelet[3071]: I0130 13:24:14.952030 3071 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 30 13:24:14.955946 kubelet[3071]: E0130 13:24:14.955913 3071 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-4186.1.0-a-a7402f3063\" not found" Jan 30 13:24:14.983440 kubelet[3071]: E0130 13:24:14.983391 3071 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.40:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4186.1.0-a-a7402f3063?timeout=10s\": dial tcp 10.200.20.40:6443: connect: connection refused" interval="400ms" Jan 30 13:24:15.113921 kubelet[3071]: I0130 13:24:15.113736 3071 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.114152 kubelet[3071]: E0130 13:24:15.114108 3071 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.20.40:6443/api/v1/nodes\": dial tcp 10.200.20.40:6443: connect: connection refused" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.139359 kubelet[3071]: I0130 13:24:15.139301 3071 topology_manager.go:215] "Topology Admit Handler" podUID="957c1958438ef3e58d469930af0ca4f8" podNamespace="kube-system" podName="kube-apiserver-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.141365 kubelet[3071]: I0130 13:24:15.141325 3071 topology_manager.go:215] "Topology Admit Handler" podUID="025418c384685ed6ca2074ea9ff6e827" podNamespace="kube-system" podName="kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.143268 kubelet[3071]: I0130 13:24:15.143009 3071 topology_manager.go:215] "Topology Admit Handler" podUID="bc4fb14880cd95e361140b684a985191" podNamespace="kube-system" podName="kube-scheduler-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.152815 systemd[1]: Created slice kubepods-burstable-pod957c1958438ef3e58d469930af0ca4f8.slice - libcontainer container kubepods-burstable-pod957c1958438ef3e58d469930af0ca4f8.slice. Jan 30 13:24:15.179251 systemd[1]: Created slice kubepods-burstable-pod025418c384685ed6ca2074ea9ff6e827.slice - libcontainer container kubepods-burstable-pod025418c384685ed6ca2074ea9ff6e827.slice. Jan 30 13:24:15.184843 kubelet[3071]: I0130 13:24:15.184800 3071 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/957c1958438ef3e58d469930af0ca4f8-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4186.1.0-a-a7402f3063\" (UID: \"957c1958438ef3e58d469930af0ca4f8\") " pod="kube-system/kube-apiserver-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.184843 kubelet[3071]: I0130 13:24:15.184842 3071 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-flexvolume-dir\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.185143 kubelet[3071]: I0130 13:24:15.184864 3071 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-k8s-certs\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.185143 kubelet[3071]: I0130 13:24:15.184881 3071 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.185143 kubelet[3071]: I0130 13:24:15.184899 3071 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/957c1958438ef3e58d469930af0ca4f8-ca-certs\") pod \"kube-apiserver-ci-4186.1.0-a-a7402f3063\" (UID: \"957c1958438ef3e58d469930af0ca4f8\") " pod="kube-system/kube-apiserver-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.185143 kubelet[3071]: I0130 13:24:15.184915 3071 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/957c1958438ef3e58d469930af0ca4f8-k8s-certs\") pod \"kube-apiserver-ci-4186.1.0-a-a7402f3063\" (UID: \"957c1958438ef3e58d469930af0ca4f8\") " pod="kube-system/kube-apiserver-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.185143 kubelet[3071]: I0130 13:24:15.184930 3071 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bc4fb14880cd95e361140b684a985191-kubeconfig\") pod \"kube-scheduler-ci-4186.1.0-a-a7402f3063\" (UID: \"bc4fb14880cd95e361140b684a985191\") " pod="kube-system/kube-scheduler-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.185272 kubelet[3071]: I0130 13:24:15.184958 3071 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-ca-certs\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.185272 kubelet[3071]: I0130 13:24:15.184983 3071 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-kubeconfig\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.190209 systemd[1]: Created slice kubepods-burstable-podbc4fb14880cd95e361140b684a985191.slice - libcontainer container kubepods-burstable-podbc4fb14880cd95e361140b684a985191.slice. Jan 30 13:24:15.384313 kubelet[3071]: E0130 13:24:15.384183 3071 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.40:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4186.1.0-a-a7402f3063?timeout=10s\": dial tcp 10.200.20.40:6443: connect: connection refused" interval="800ms" Jan 30 13:24:15.476539 containerd[1781]: time="2025-01-30T13:24:15.476475135Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4186.1.0-a-a7402f3063,Uid:957c1958438ef3e58d469930af0ca4f8,Namespace:kube-system,Attempt:0,}" Jan 30 13:24:15.483481 containerd[1781]: time="2025-01-30T13:24:15.483430338Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4186.1.0-a-a7402f3063,Uid:025418c384685ed6ca2074ea9ff6e827,Namespace:kube-system,Attempt:0,}" Jan 30 13:24:15.493504 containerd[1781]: time="2025-01-30T13:24:15.493460582Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4186.1.0-a-a7402f3063,Uid:bc4fb14880cd95e361140b684a985191,Namespace:kube-system,Attempt:0,}" Jan 30 13:24:15.517929 kubelet[3071]: I0130 13:24:15.517896 3071 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.518271 kubelet[3071]: E0130 13:24:15.518232 3071 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.20.40:6443/api/v1/nodes\": dial tcp 10.200.20.40:6443: connect: connection refused" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:15.829838 kubelet[3071]: W0130 13:24:15.829797 3071 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://10.200.20.40:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:15.829838 kubelet[3071]: E0130 13:24:15.829843 3071 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.200.20.40:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:15.986357 kubelet[3071]: W0130 13:24:15.986283 3071 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://10.200.20.40:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:15.986357 kubelet[3071]: E0130 13:24:15.986363 3071 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.200.20.40:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:16.039133 kubelet[3071]: W0130 13:24:16.039093 3071 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://10.200.20.40:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:16.039133 kubelet[3071]: E0130 13:24:16.039140 3071 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.200.20.40:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:16.142755 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1733261722.mount: Deactivated successfully. Jan 30 13:24:16.171709 containerd[1781]: time="2025-01-30T13:24:16.170880133Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 13:24:16.183881 containerd[1781]: time="2025-01-30T13:24:16.183820298Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=269173" Jan 30 13:24:16.185206 kubelet[3071]: E0130 13:24:16.185161 3071 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.40:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-4186.1.0-a-a7402f3063?timeout=10s\": dial tcp 10.200.20.40:6443: connect: connection refused" interval="1.6s" Jan 30 13:24:16.190129 containerd[1781]: time="2025-01-30T13:24:16.189514621Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 13:24:16.195770 containerd[1781]: time="2025-01-30T13:24:16.195560223Z" level=info msg="ImageCreate event name:\"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 13:24:16.200546 containerd[1781]: time="2025-01-30T13:24:16.200387425Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 30 13:24:16.204338 containerd[1781]: time="2025-01-30T13:24:16.203514666Z" level=info msg="ImageUpdate event name:\"registry.k8s.io/pause:3.8\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 13:24:16.206165 containerd[1781]: time="2025-01-30T13:24:16.206116507Z" level=info msg="ImageCreate event name:\"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"} labels:{key:\"io.cri-containerd.pinned\" value:\"pinned\"}" Jan 30 13:24:16.207115 containerd[1781]: time="2025-01-30T13:24:16.207076308Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 730.518133ms" Jan 30 13:24:16.208131 containerd[1781]: time="2025-01-30T13:24:16.208006828Z" level=info msg="stop pulling image registry.k8s.io/pause:3.8: active requests=0, bytes read=0" Jan 30 13:24:16.213173 containerd[1781]: time="2025-01-30T13:24:16.213122310Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 729.607572ms" Jan 30 13:24:16.242183 containerd[1781]: time="2025-01-30T13:24:16.241953202Z" level=info msg="Pulled image \"registry.k8s.io/pause:3.8\" with image id \"sha256:4e42fb3c9d90ed7895bc04a9d96fe3102a65b521f485cc5a4f3dd818afef9cef\", repo tag \"registry.k8s.io/pause:3.8\", repo digest \"registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d\", size \"268403\" in 748.4083ms" Jan 30 13:24:16.275520 kubelet[3071]: W0130 13:24:16.275458 3071 reflector.go:547] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://10.200.20.40:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4186.1.0-a-a7402f3063&limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:16.275782 kubelet[3071]: E0130 13:24:16.275753 3071 reflector.go:150] k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.200.20.40:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-4186.1.0-a-a7402f3063&limit=500&resourceVersion=0": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:16.321154 kubelet[3071]: I0130 13:24:16.321128 3071 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:16.322560 kubelet[3071]: E0130 13:24:16.322405 3071 kubelet_node_status.go:96] "Unable to register node with API server" err="Post \"https://10.200.20.40:6443/api/v1/nodes\": dial tcp 10.200.20.40:6443: connect: connection refused" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:16.406334 containerd[1781]: time="2025-01-30T13:24:16.406028627Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:24:16.406334 containerd[1781]: time="2025-01-30T13:24:16.406118107Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:24:16.406334 containerd[1781]: time="2025-01-30T13:24:16.406131387Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:16.406334 containerd[1781]: time="2025-01-30T13:24:16.406219627Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:16.412702 containerd[1781]: time="2025-01-30T13:24:16.410005109Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:24:16.412702 containerd[1781]: time="2025-01-30T13:24:16.411830590Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:24:16.412702 containerd[1781]: time="2025-01-30T13:24:16.411845910Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:16.412702 containerd[1781]: time="2025-01-30T13:24:16.411940030Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:16.413312 containerd[1781]: time="2025-01-30T13:24:16.413095350Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:24:16.413312 containerd[1781]: time="2025-01-30T13:24:16.413164630Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:24:16.413312 containerd[1781]: time="2025-01-30T13:24:16.413179430Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:16.413312 containerd[1781]: time="2025-01-30T13:24:16.413262710Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:16.437000 systemd[1]: Started cri-containerd-c373e6260a19f6d9a8d685157c186ad9e0cda264ff5cc7a7b3e9b85250688530.scope - libcontainer container c373e6260a19f6d9a8d685157c186ad9e0cda264ff5cc7a7b3e9b85250688530. Jan 30 13:24:16.439259 systemd[1]: Started cri-containerd-ef23064c1db537ab744004d406a50f85f84d6e9fe683a764eac73dac0cf2686c.scope - libcontainer container ef23064c1db537ab744004d406a50f85f84d6e9fe683a764eac73dac0cf2686c. Jan 30 13:24:16.455261 systemd[1]: Started cri-containerd-84302aeab2a24c01ed73ad16de1a40ca0f73be678cda6cb6c4c2548a7f8ea53b.scope - libcontainer container 84302aeab2a24c01ed73ad16de1a40ca0f73be678cda6cb6c4c2548a7f8ea53b. Jan 30 13:24:16.505260 containerd[1781]: time="2025-01-30T13:24:16.505220227Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-4186.1.0-a-a7402f3063,Uid:957c1958438ef3e58d469930af0ca4f8,Namespace:kube-system,Attempt:0,} returns sandbox id \"c373e6260a19f6d9a8d685157c186ad9e0cda264ff5cc7a7b3e9b85250688530\"" Jan 30 13:24:16.518484 containerd[1781]: time="2025-01-30T13:24:16.518118352Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-4186.1.0-a-a7402f3063,Uid:bc4fb14880cd95e361140b684a985191,Namespace:kube-system,Attempt:0,} returns sandbox id \"ef23064c1db537ab744004d406a50f85f84d6e9fe683a764eac73dac0cf2686c\"" Jan 30 13:24:16.521147 containerd[1781]: time="2025-01-30T13:24:16.521057873Z" level=info msg="CreateContainer within sandbox \"c373e6260a19f6d9a8d685157c186ad9e0cda264ff5cc7a7b3e9b85250688530\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Jan 30 13:24:16.523309 containerd[1781]: time="2025-01-30T13:24:16.523195274Z" level=info msg="CreateContainer within sandbox \"ef23064c1db537ab744004d406a50f85f84d6e9fe683a764eac73dac0cf2686c\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Jan 30 13:24:16.532594 containerd[1781]: time="2025-01-30T13:24:16.532326558Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-4186.1.0-a-a7402f3063,Uid:025418c384685ed6ca2074ea9ff6e827,Namespace:kube-system,Attempt:0,} returns sandbox id \"84302aeab2a24c01ed73ad16de1a40ca0f73be678cda6cb6c4c2548a7f8ea53b\"" Jan 30 13:24:16.538736 containerd[1781]: time="2025-01-30T13:24:16.538675320Z" level=info msg="CreateContainer within sandbox \"84302aeab2a24c01ed73ad16de1a40ca0f73be678cda6cb6c4c2548a7f8ea53b\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Jan 30 13:24:16.591753 containerd[1781]: time="2025-01-30T13:24:16.591658102Z" level=info msg="CreateContainer within sandbox \"c373e6260a19f6d9a8d685157c186ad9e0cda264ff5cc7a7b3e9b85250688530\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"23d0415c0401b13aaea5a61a49d17954f508862b06dac7d32e741b7e2f2a577d\"" Jan 30 13:24:16.595291 containerd[1781]: time="2025-01-30T13:24:16.595152583Z" level=info msg="CreateContainer within sandbox \"ef23064c1db537ab744004d406a50f85f84d6e9fe683a764eac73dac0cf2686c\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"46bd693ed4f47dc44be528c74a11ad9c128b3c08af5da03d662ddadce99c116d\"" Jan 30 13:24:16.595671 containerd[1781]: time="2025-01-30T13:24:16.595449983Z" level=info msg="StartContainer for \"23d0415c0401b13aaea5a61a49d17954f508862b06dac7d32e741b7e2f2a577d\"" Jan 30 13:24:16.603038 containerd[1781]: time="2025-01-30T13:24:16.602987426Z" level=info msg="CreateContainer within sandbox \"84302aeab2a24c01ed73ad16de1a40ca0f73be678cda6cb6c4c2548a7f8ea53b\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"84e034552dc81be72c90a75e70dc21954202aaf234ce015afc9727384c47c455\"" Jan 30 13:24:16.603667 containerd[1781]: time="2025-01-30T13:24:16.603226866Z" level=info msg="StartContainer for \"46bd693ed4f47dc44be528c74a11ad9c128b3c08af5da03d662ddadce99c116d\"" Jan 30 13:24:16.606044 containerd[1781]: time="2025-01-30T13:24:16.605993347Z" level=info msg="StartContainer for \"84e034552dc81be72c90a75e70dc21954202aaf234ce015afc9727384c47c455\"" Jan 30 13:24:16.626087 systemd[1]: Started cri-containerd-23d0415c0401b13aaea5a61a49d17954f508862b06dac7d32e741b7e2f2a577d.scope - libcontainer container 23d0415c0401b13aaea5a61a49d17954f508862b06dac7d32e741b7e2f2a577d. Jan 30 13:24:16.648980 systemd[1]: Started cri-containerd-46bd693ed4f47dc44be528c74a11ad9c128b3c08af5da03d662ddadce99c116d.scope - libcontainer container 46bd693ed4f47dc44be528c74a11ad9c128b3c08af5da03d662ddadce99c116d. Jan 30 13:24:16.657954 systemd[1]: Started cri-containerd-84e034552dc81be72c90a75e70dc21954202aaf234ce015afc9727384c47c455.scope - libcontainer container 84e034552dc81be72c90a75e70dc21954202aaf234ce015afc9727384c47c455. Jan 30 13:24:16.705770 containerd[1781]: time="2025-01-30T13:24:16.705643507Z" level=info msg="StartContainer for \"23d0415c0401b13aaea5a61a49d17954f508862b06dac7d32e741b7e2f2a577d\" returns successfully" Jan 30 13:24:16.731988 containerd[1781]: time="2025-01-30T13:24:16.731608478Z" level=info msg="StartContainer for \"84e034552dc81be72c90a75e70dc21954202aaf234ce015afc9727384c47c455\" returns successfully" Jan 30 13:24:16.731988 containerd[1781]: time="2025-01-30T13:24:16.731608478Z" level=info msg="StartContainer for \"46bd693ed4f47dc44be528c74a11ad9c128b3c08af5da03d662ddadce99c116d\" returns successfully" Jan 30 13:24:16.780945 kubelet[3071]: E0130 13:24:16.780898 3071 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.200.20.40:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.200.20.40:6443: connect: connection refused Jan 30 13:24:17.926738 kubelet[3071]: I0130 13:24:17.925205 3071 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:19.563137 kubelet[3071]: I0130 13:24:19.563089 3071 kubelet_node_status.go:76] "Successfully registered node" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:19.677648 kubelet[3071]: E0130 13:24:19.677592 3071 controller.go:145] "Failed to ensure lease exists, will retry" err="namespaces \"kube-node-lease\" not found" interval="3.2s" Jan 30 13:24:19.769497 kubelet[3071]: I0130 13:24:19.769226 3071 apiserver.go:52] "Watching apiserver" Jan 30 13:24:19.782871 kubelet[3071]: I0130 13:24:19.782842 3071 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Jan 30 13:24:22.253196 systemd[1]: Reloading requested from client PID 3351 ('systemctl') (unit session-9.scope)... Jan 30 13:24:22.253228 systemd[1]: Reloading... Jan 30 13:24:22.347783 zram_generator::config[3391]: No configuration found. Jan 30 13:24:22.476979 systemd[1]: /usr/lib/systemd/system/docker.socket:6: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Jan 30 13:24:22.573082 systemd[1]: Reloading finished in 319 ms. Jan 30 13:24:22.609833 systemd[1]: Stopping kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:24:22.610677 kubelet[3071]: I0130 13:24:22.610194 3071 dynamic_cafile_content.go:171] "Shutting down controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 30 13:24:22.625805 systemd[1]: kubelet.service: Deactivated successfully. Jan 30 13:24:22.626056 systemd[1]: Stopped kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:24:22.633010 systemd[1]: Starting kubelet.service - kubelet: The Kubernetes Node Agent... Jan 30 13:24:22.755902 systemd[1]: Started kubelet.service - kubelet: The Kubernetes Node Agent. Jan 30 13:24:22.765545 (kubelet)[3455]: kubelet.service: Referenced but unset environment variable evaluates to an empty string: KUBELET_EXTRA_ARGS Jan 30 13:24:22.823949 kubelet[3455]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 13:24:22.824401 kubelet[3455]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Jan 30 13:24:22.824453 kubelet[3455]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Jan 30 13:24:22.824582 kubelet[3455]: I0130 13:24:22.824550 3455 server.go:205] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Jan 30 13:24:22.828984 kubelet[3455]: I0130 13:24:22.828951 3455 server.go:484] "Kubelet version" kubeletVersion="v1.30.1" Jan 30 13:24:22.829170 kubelet[3455]: I0130 13:24:22.829160 3455 server.go:486] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Jan 30 13:24:22.829485 kubelet[3455]: I0130 13:24:22.829471 3455 server.go:927] "Client rotation is on, will bootstrap in background" Jan 30 13:24:22.830992 kubelet[3455]: I0130 13:24:22.830962 3455 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Jan 30 13:24:22.832675 kubelet[3455]: I0130 13:24:22.832649 3455 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Jan 30 13:24:22.843748 kubelet[3455]: I0130 13:24:22.843679 3455 server.go:742] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Jan 30 13:24:22.843937 kubelet[3455]: I0130 13:24:22.843898 3455 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Jan 30 13:24:22.844115 kubelet[3455]: I0130 13:24:22.843933 3455 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"ci-4186.1.0-a-a7402f3063","RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Jan 30 13:24:22.844191 kubelet[3455]: I0130 13:24:22.844120 3455 topology_manager.go:138] "Creating topology manager with none policy" Jan 30 13:24:22.844191 kubelet[3455]: I0130 13:24:22.844130 3455 container_manager_linux.go:301] "Creating device plugin manager" Jan 30 13:24:22.844191 kubelet[3455]: I0130 13:24:22.844164 3455 state_mem.go:36] "Initialized new in-memory state store" Jan 30 13:24:22.844301 kubelet[3455]: I0130 13:24:22.844278 3455 kubelet.go:400] "Attempting to sync node with API server" Jan 30 13:24:22.844301 kubelet[3455]: I0130 13:24:22.844293 3455 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Jan 30 13:24:22.844570 kubelet[3455]: I0130 13:24:22.844321 3455 kubelet.go:312] "Adding apiserver pod source" Jan 30 13:24:22.844570 kubelet[3455]: I0130 13:24:22.844337 3455 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Jan 30 13:24:22.847880 kubelet[3455]: I0130 13:24:22.847847 3455 kuberuntime_manager.go:261] "Container runtime initialized" containerRuntime="containerd" version="v1.7.23" apiVersion="v1" Jan 30 13:24:22.848099 kubelet[3455]: I0130 13:24:22.848079 3455 kubelet.go:815] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Jan 30 13:24:22.848512 kubelet[3455]: I0130 13:24:22.848491 3455 server.go:1264] "Started kubelet" Jan 30 13:24:22.852125 kubelet[3455]: I0130 13:24:22.852084 3455 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Jan 30 13:24:22.867757 kubelet[3455]: I0130 13:24:22.866801 3455 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Jan 30 13:24:22.867757 kubelet[3455]: I0130 13:24:22.867643 3455 server.go:455] "Adding debug handlers to kubelet server" Jan 30 13:24:22.868568 kubelet[3455]: I0130 13:24:22.868507 3455 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Jan 30 13:24:22.870756 kubelet[3455]: I0130 13:24:22.870732 3455 server.go:227] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Jan 30 13:24:22.872021 kubelet[3455]: I0130 13:24:22.871983 3455 volume_manager.go:291] "Starting Kubelet Volume Manager" Jan 30 13:24:22.876933 kubelet[3455]: I0130 13:24:22.876896 3455 desired_state_of_world_populator.go:149] "Desired state populator starts to run" Jan 30 13:24:22.877083 kubelet[3455]: I0130 13:24:22.877065 3455 reconciler.go:26] "Reconciler: start to sync state" Jan 30 13:24:22.881786 kubelet[3455]: I0130 13:24:22.880841 3455 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Jan 30 13:24:22.882698 kubelet[3455]: I0130 13:24:22.882656 3455 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Jan 30 13:24:22.882781 kubelet[3455]: I0130 13:24:22.882740 3455 status_manager.go:217] "Starting to sync pod status with apiserver" Jan 30 13:24:22.882781 kubelet[3455]: I0130 13:24:22.882763 3455 kubelet.go:2337] "Starting kubelet main sync loop" Jan 30 13:24:22.882835 kubelet[3455]: E0130 13:24:22.882809 3455 kubelet.go:2361] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Jan 30 13:24:22.893756 kubelet[3455]: E0130 13:24:22.893607 3455 kubelet.go:1467] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Jan 30 13:24:22.895985 kubelet[3455]: I0130 13:24:22.895942 3455 factory.go:221] Registration of the containerd container factory successfully Jan 30 13:24:22.895985 kubelet[3455]: I0130 13:24:22.895971 3455 factory.go:221] Registration of the systemd container factory successfully Jan 30 13:24:22.896146 kubelet[3455]: I0130 13:24:22.896084 3455 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Jan 30 13:24:22.942107 kubelet[3455]: I0130 13:24:22.942069 3455 cpu_manager.go:214] "Starting CPU manager" policy="none" Jan 30 13:24:22.942107 kubelet[3455]: I0130 13:24:22.942094 3455 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Jan 30 13:24:22.942107 kubelet[3455]: I0130 13:24:22.942123 3455 state_mem.go:36] "Initialized new in-memory state store" Jan 30 13:24:22.942333 kubelet[3455]: I0130 13:24:22.942310 3455 state_mem.go:88] "Updated default CPUSet" cpuSet="" Jan 30 13:24:22.942368 kubelet[3455]: I0130 13:24:22.942327 3455 state_mem.go:96] "Updated CPUSet assignments" assignments={} Jan 30 13:24:22.942368 kubelet[3455]: I0130 13:24:22.942349 3455 policy_none.go:49] "None policy: Start" Jan 30 13:24:22.943680 kubelet[3455]: I0130 13:24:22.943316 3455 memory_manager.go:170] "Starting memorymanager" policy="None" Jan 30 13:24:22.943680 kubelet[3455]: I0130 13:24:22.943352 3455 state_mem.go:35] "Initializing new in-memory state store" Jan 30 13:24:22.943680 kubelet[3455]: I0130 13:24:22.943533 3455 state_mem.go:75] "Updated machine memory state" Jan 30 13:24:22.948616 kubelet[3455]: I0130 13:24:22.948574 3455 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Jan 30 13:24:22.948835 kubelet[3455]: I0130 13:24:22.948789 3455 container_log_manager.go:186] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Jan 30 13:24:22.948923 kubelet[3455]: I0130 13:24:22.948906 3455 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Jan 30 13:24:22.975346 kubelet[3455]: I0130 13:24:22.975145 3455 kubelet_node_status.go:73] "Attempting to register node" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:22.983331 kubelet[3455]: I0130 13:24:22.983284 3455 topology_manager.go:215] "Topology Admit Handler" podUID="025418c384685ed6ca2074ea9ff6e827" podNamespace="kube-system" podName="kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:22.983999 kubelet[3455]: I0130 13:24:22.983647 3455 topology_manager.go:215] "Topology Admit Handler" podUID="bc4fb14880cd95e361140b684a985191" podNamespace="kube-system" podName="kube-scheduler-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:22.983999 kubelet[3455]: I0130 13:24:22.983695 3455 topology_manager.go:215] "Topology Admit Handler" podUID="957c1958438ef3e58d469930af0ca4f8" podNamespace="kube-system" podName="kube-apiserver-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:22.991502 kubelet[3455]: I0130 13:24:22.991455 3455 kubelet_node_status.go:112] "Node was previously registered" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:22.991705 kubelet[3455]: I0130 13:24:22.991548 3455 kubelet_node_status.go:76] "Successfully registered node" node="ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.003739 kubelet[3455]: W0130 13:24:23.003687 3455 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 30 13:24:23.003999 kubelet[3455]: W0130 13:24:23.003760 3455 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 30 13:24:23.003999 kubelet[3455]: W0130 13:24:23.003960 3455 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 30 13:24:23.178492 kubelet[3455]: I0130 13:24:23.178368 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-flexvolume-dir\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.178492 kubelet[3455]: I0130 13:24:23.178412 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-k8s-certs\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.178492 kubelet[3455]: I0130 13:24:23.178434 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/bc4fb14880cd95e361140b684a985191-kubeconfig\") pod \"kube-scheduler-ci-4186.1.0-a-a7402f3063\" (UID: \"bc4fb14880cd95e361140b684a985191\") " pod="kube-system/kube-scheduler-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.178492 kubelet[3455]: I0130 13:24:23.178453 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/957c1958438ef3e58d469930af0ca4f8-ca-certs\") pod \"kube-apiserver-ci-4186.1.0-a-a7402f3063\" (UID: \"957c1958438ef3e58d469930af0ca4f8\") " pod="kube-system/kube-apiserver-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.178492 kubelet[3455]: I0130 13:24:23.178471 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/957c1958438ef3e58d469930af0ca4f8-k8s-certs\") pod \"kube-apiserver-ci-4186.1.0-a-a7402f3063\" (UID: \"957c1958438ef3e58d469930af0ca4f8\") " pod="kube-system/kube-apiserver-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.178789 kubelet[3455]: I0130 13:24:23.178498 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/957c1958438ef3e58d469930af0ca4f8-usr-share-ca-certificates\") pod \"kube-apiserver-ci-4186.1.0-a-a7402f3063\" (UID: \"957c1958438ef3e58d469930af0ca4f8\") " pod="kube-system/kube-apiserver-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.178789 kubelet[3455]: I0130 13:24:23.178537 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-ca-certs\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.178789 kubelet[3455]: I0130 13:24:23.178554 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-kubeconfig\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.178789 kubelet[3455]: I0130 13:24:23.178572 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/025418c384685ed6ca2074ea9ff6e827-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" (UID: \"025418c384685ed6ca2074ea9ff6e827\") " pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.844758 kubelet[3455]: I0130 13:24:23.844702 3455 apiserver.go:52] "Watching apiserver" Jan 30 13:24:23.878066 kubelet[3455]: I0130 13:24:23.878032 3455 desired_state_of_world_populator.go:157] "Finished populating initial desired state of world" Jan 30 13:24:23.938522 kubelet[3455]: W0130 13:24:23.938409 3455 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 30 13:24:23.938877 kubelet[3455]: E0130 13:24:23.938494 3455 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-4186.1.0-a-a7402f3063\" already exists" pod="kube-system/kube-apiserver-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.939096 kubelet[3455]: W0130 13:24:23.938898 3455 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Jan 30 13:24:23.939096 kubelet[3455]: E0130 13:24:23.938953 3455 kubelet.go:1928] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-ci-4186.1.0-a-a7402f3063\" already exists" pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" Jan 30 13:24:23.945104 kubelet[3455]: I0130 13:24:23.945031 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-4186.1.0-a-a7402f3063" podStartSLOduration=1.945013675 podStartE2EDuration="1.945013675s" podCreationTimestamp="2025-01-30 13:24:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:24:23.944759955 +0000 UTC m=+1.174356942" watchObservedRunningTime="2025-01-30 13:24:23.945013675 +0000 UTC m=+1.174610662" Jan 30 13:24:23.956016 kubelet[3455]: I0130 13:24:23.955898 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-4186.1.0-a-a7402f3063" podStartSLOduration=1.955877837 podStartE2EDuration="1.955877837s" podCreationTimestamp="2025-01-30 13:24:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:24:23.955591517 +0000 UTC m=+1.185188464" watchObservedRunningTime="2025-01-30 13:24:23.955877837 +0000 UTC m=+1.185474824" Jan 30 13:24:23.967509 kubelet[3455]: I0130 13:24:23.967292 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-4186.1.0-a-a7402f3063" podStartSLOduration=1.967273598 podStartE2EDuration="1.967273598s" podCreationTimestamp="2025-01-30 13:24:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:24:23.967086518 +0000 UTC m=+1.196683505" watchObservedRunningTime="2025-01-30 13:24:23.967273598 +0000 UTC m=+1.196870585" Jan 30 13:24:31.047818 sudo[2285]: pam_unix(sudo:session): session closed for user root Jan 30 13:24:31.127926 sshd[2284]: Connection closed by 10.200.16.10 port 44536 Jan 30 13:24:31.129106 sshd-session[2269]: pam_unix(sshd:session): session closed for user core Jan 30 13:24:31.134115 systemd[1]: sshd@6-10.200.20.40:22-10.200.16.10:44536.service: Deactivated successfully. Jan 30 13:24:31.137798 systemd[1]: session-9.scope: Deactivated successfully. Jan 30 13:24:31.139448 systemd[1]: session-9.scope: Consumed 5.454s CPU time, 188.9M memory peak, 0B memory swap peak. Jan 30 13:24:31.140479 systemd-logind[1717]: Session 9 logged out. Waiting for processes to exit. Jan 30 13:24:31.141905 systemd-logind[1717]: Removed session 9. Jan 30 13:24:35.502704 kubelet[3455]: I0130 13:24:35.502666 3455 kuberuntime_manager.go:1523] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Jan 30 13:24:35.504404 kubelet[3455]: I0130 13:24:35.504379 3455 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Jan 30 13:24:35.504447 containerd[1781]: time="2025-01-30T13:24:35.504077078Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Jan 30 13:24:36.395811 kubelet[3455]: I0130 13:24:36.395328 3455 topology_manager.go:215] "Topology Admit Handler" podUID="43507cae-e1f8-4113-b10a-3ef3a3ebbf94" podNamespace="kube-system" podName="kube-proxy-ksvst" Jan 30 13:24:36.409985 systemd[1]: Created slice kubepods-besteffort-pod43507cae_e1f8_4113_b10a_3ef3a3ebbf94.slice - libcontainer container kubepods-besteffort-pod43507cae_e1f8_4113_b10a_3ef3a3ebbf94.slice. Jan 30 13:24:36.464661 kubelet[3455]: I0130 13:24:36.464621 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/43507cae-e1f8-4113-b10a-3ef3a3ebbf94-kube-proxy\") pod \"kube-proxy-ksvst\" (UID: \"43507cae-e1f8-4113-b10a-3ef3a3ebbf94\") " pod="kube-system/kube-proxy-ksvst" Jan 30 13:24:36.464661 kubelet[3455]: I0130 13:24:36.464666 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c686\" (UniqueName: \"kubernetes.io/projected/43507cae-e1f8-4113-b10a-3ef3a3ebbf94-kube-api-access-6c686\") pod \"kube-proxy-ksvst\" (UID: \"43507cae-e1f8-4113-b10a-3ef3a3ebbf94\") " pod="kube-system/kube-proxy-ksvst" Jan 30 13:24:36.464661 kubelet[3455]: I0130 13:24:36.464690 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/43507cae-e1f8-4113-b10a-3ef3a3ebbf94-lib-modules\") pod \"kube-proxy-ksvst\" (UID: \"43507cae-e1f8-4113-b10a-3ef3a3ebbf94\") " pod="kube-system/kube-proxy-ksvst" Jan 30 13:24:36.464661 kubelet[3455]: I0130 13:24:36.464721 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/43507cae-e1f8-4113-b10a-3ef3a3ebbf94-xtables-lock\") pod \"kube-proxy-ksvst\" (UID: \"43507cae-e1f8-4113-b10a-3ef3a3ebbf94\") " pod="kube-system/kube-proxy-ksvst" Jan 30 13:24:36.618986 kubelet[3455]: I0130 13:24:36.618530 3455 topology_manager.go:215] "Topology Admit Handler" podUID="7f888095-116a-44fe-bf32-ba587a63eabf" podNamespace="tigera-operator" podName="tigera-operator-7bc55997bb-dkgxw" Jan 30 13:24:36.634439 systemd[1]: Created slice kubepods-besteffort-pod7f888095_116a_44fe_bf32_ba587a63eabf.slice - libcontainer container kubepods-besteffort-pod7f888095_116a_44fe_bf32_ba587a63eabf.slice. Jan 30 13:24:36.666174 kubelet[3455]: I0130 13:24:36.666056 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf9m7\" (UniqueName: \"kubernetes.io/projected/7f888095-116a-44fe-bf32-ba587a63eabf-kube-api-access-hf9m7\") pod \"tigera-operator-7bc55997bb-dkgxw\" (UID: \"7f888095-116a-44fe-bf32-ba587a63eabf\") " pod="tigera-operator/tigera-operator-7bc55997bb-dkgxw" Jan 30 13:24:36.666174 kubelet[3455]: I0130 13:24:36.666107 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/7f888095-116a-44fe-bf32-ba587a63eabf-var-lib-calico\") pod \"tigera-operator-7bc55997bb-dkgxw\" (UID: \"7f888095-116a-44fe-bf32-ba587a63eabf\") " pod="tigera-operator/tigera-operator-7bc55997bb-dkgxw" Jan 30 13:24:36.720076 containerd[1781]: time="2025-01-30T13:24:36.719960316Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ksvst,Uid:43507cae-e1f8-4113-b10a-3ef3a3ebbf94,Namespace:kube-system,Attempt:0,}" Jan 30 13:24:36.766671 containerd[1781]: time="2025-01-30T13:24:36.766467225Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:24:36.766919 containerd[1781]: time="2025-01-30T13:24:36.766527665Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:24:36.766919 containerd[1781]: time="2025-01-30T13:24:36.766616145Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:36.766982 containerd[1781]: time="2025-01-30T13:24:36.766933225Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:36.793931 systemd[1]: Started cri-containerd-90b8457543f57864568ee8625e179cb2103dd89148cd25b379ebf5e8fa7bd5dd.scope - libcontainer container 90b8457543f57864568ee8625e179cb2103dd89148cd25b379ebf5e8fa7bd5dd. Jan 30 13:24:36.817360 containerd[1781]: time="2025-01-30T13:24:36.817313053Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-ksvst,Uid:43507cae-e1f8-4113-b10a-3ef3a3ebbf94,Namespace:kube-system,Attempt:0,} returns sandbox id \"90b8457543f57864568ee8625e179cb2103dd89148cd25b379ebf5e8fa7bd5dd\"" Jan 30 13:24:36.822498 containerd[1781]: time="2025-01-30T13:24:36.822450332Z" level=info msg="CreateContainer within sandbox \"90b8457543f57864568ee8625e179cb2103dd89148cd25b379ebf5e8fa7bd5dd\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Jan 30 13:24:36.860012 containerd[1781]: time="2025-01-30T13:24:36.859954563Z" level=info msg="CreateContainer within sandbox \"90b8457543f57864568ee8625e179cb2103dd89148cd25b379ebf5e8fa7bd5dd\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"016afb6cff3981e0f3d9f88b9680dcb2237e5805cee42d6e2496a4d31708e8ce\"" Jan 30 13:24:36.860765 containerd[1781]: time="2025-01-30T13:24:36.860708763Z" level=info msg="StartContainer for \"016afb6cff3981e0f3d9f88b9680dcb2237e5805cee42d6e2496a4d31708e8ce\"" Jan 30 13:24:36.887243 systemd[1]: Started cri-containerd-016afb6cff3981e0f3d9f88b9680dcb2237e5805cee42d6e2496a4d31708e8ce.scope - libcontainer container 016afb6cff3981e0f3d9f88b9680dcb2237e5805cee42d6e2496a4d31708e8ce. Jan 30 13:24:36.924843 containerd[1781]: time="2025-01-30T13:24:36.923801228Z" level=info msg="StartContainer for \"016afb6cff3981e0f3d9f88b9680dcb2237e5805cee42d6e2496a4d31708e8ce\" returns successfully" Jan 30 13:24:36.939331 containerd[1781]: time="2025-01-30T13:24:36.939234665Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7bc55997bb-dkgxw,Uid:7f888095-116a-44fe-bf32-ba587a63eabf,Namespace:tigera-operator,Attempt:0,}" Jan 30 13:24:36.973497 kubelet[3455]: I0130 13:24:36.973420 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/kube-proxy-ksvst" podStartSLOduration=0.973400457 podStartE2EDuration="973.400457ms" podCreationTimestamp="2025-01-30 13:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:24:36.973336217 +0000 UTC m=+14.202933204" watchObservedRunningTime="2025-01-30 13:24:36.973400457 +0000 UTC m=+14.202997404" Jan 30 13:24:36.993625 containerd[1781]: time="2025-01-30T13:24:36.993463572Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:24:36.993992 containerd[1781]: time="2025-01-30T13:24:36.993638372Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:24:36.993992 containerd[1781]: time="2025-01-30T13:24:36.993690172Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:36.993992 containerd[1781]: time="2025-01-30T13:24:36.993864052Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:37.014333 systemd[1]: Started cri-containerd-e7133782e8064279101770d46dfd856ac6da6e18d314d12c2707ec9cf7adfcef.scope - libcontainer container e7133782e8064279101770d46dfd856ac6da6e18d314d12c2707ec9cf7adfcef. Jan 30 13:24:37.050154 containerd[1781]: time="2025-01-30T13:24:37.049929399Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:tigera-operator-7bc55997bb-dkgxw,Uid:7f888095-116a-44fe-bf32-ba587a63eabf,Namespace:tigera-operator,Attempt:0,} returns sandbox id \"e7133782e8064279101770d46dfd856ac6da6e18d314d12c2707ec9cf7adfcef\"" Jan 30 13:24:37.053325 containerd[1781]: time="2025-01-30T13:24:37.052842358Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\"" Jan 30 13:24:39.041450 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3981197728.mount: Deactivated successfully. Jan 30 13:24:39.448928 containerd[1781]: time="2025-01-30T13:24:39.448803028Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator:v1.36.2\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:39.451549 containerd[1781]: time="2025-01-30T13:24:39.451473828Z" level=info msg="stop pulling image quay.io/tigera/operator:v1.36.2: active requests=0, bytes read=19124160" Jan 30 13:24:39.454886 containerd[1781]: time="2025-01-30T13:24:39.454821628Z" level=info msg="ImageCreate event name:\"sha256:30d521e4e84764b396aacbb2a373ca7a573f84571e3955b34329652acccfb73c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:39.459935 containerd[1781]: time="2025-01-30T13:24:39.459865948Z" level=info msg="ImageCreate event name:\"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:39.460705 containerd[1781]: time="2025-01-30T13:24:39.460557068Z" level=info msg="Pulled image \"quay.io/tigera/operator:v1.36.2\" with image id \"sha256:30d521e4e84764b396aacbb2a373ca7a573f84571e3955b34329652acccfb73c\", repo tag \"quay.io/tigera/operator:v1.36.2\", repo digest \"quay.io/tigera/operator@sha256:fc9ea45f2475fd99db1b36d2ff180a50017b1a5ea0e82a171c6b439b3a620764\", size \"19120155\" in 2.40767091s" Jan 30 13:24:39.460705 containerd[1781]: time="2025-01-30T13:24:39.460599868Z" level=info msg="PullImage \"quay.io/tigera/operator:v1.36.2\" returns image reference \"sha256:30d521e4e84764b396aacbb2a373ca7a573f84571e3955b34329652acccfb73c\"" Jan 30 13:24:39.463643 containerd[1781]: time="2025-01-30T13:24:39.463375628Z" level=info msg="CreateContainer within sandbox \"e7133782e8064279101770d46dfd856ac6da6e18d314d12c2707ec9cf7adfcef\" for container &ContainerMetadata{Name:tigera-operator,Attempt:0,}" Jan 30 13:24:39.506180 containerd[1781]: time="2025-01-30T13:24:39.506130709Z" level=info msg="CreateContainer within sandbox \"e7133782e8064279101770d46dfd856ac6da6e18d314d12c2707ec9cf7adfcef\" for &ContainerMetadata{Name:tigera-operator,Attempt:0,} returns container id \"dc648da9ac47d0bcfef1aa06c6e95e0035f7d30ae93b25c9f7bd7fb1b78f0feb\"" Jan 30 13:24:39.506960 containerd[1781]: time="2025-01-30T13:24:39.506919389Z" level=info msg="StartContainer for \"dc648da9ac47d0bcfef1aa06c6e95e0035f7d30ae93b25c9f7bd7fb1b78f0feb\"" Jan 30 13:24:39.531980 systemd[1]: Started cri-containerd-dc648da9ac47d0bcfef1aa06c6e95e0035f7d30ae93b25c9f7bd7fb1b78f0feb.scope - libcontainer container dc648da9ac47d0bcfef1aa06c6e95e0035f7d30ae93b25c9f7bd7fb1b78f0feb. Jan 30 13:24:39.559815 containerd[1781]: time="2025-01-30T13:24:39.559613710Z" level=info msg="StartContainer for \"dc648da9ac47d0bcfef1aa06c6e95e0035f7d30ae93b25c9f7bd7fb1b78f0feb\" returns successfully" Jan 30 13:24:39.977614 kubelet[3455]: I0130 13:24:39.977482 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="tigera-operator/tigera-operator-7bc55997bb-dkgxw" podStartSLOduration=1.567596731 podStartE2EDuration="3.97746212s" podCreationTimestamp="2025-01-30 13:24:36 +0000 UTC" firstStartedPulling="2025-01-30 13:24:37.051798679 +0000 UTC m=+14.281395666" lastFinishedPulling="2025-01-30 13:24:39.461664108 +0000 UTC m=+16.691261055" observedRunningTime="2025-01-30 13:24:39.97725752 +0000 UTC m=+17.206854547" watchObservedRunningTime="2025-01-30 13:24:39.97746212 +0000 UTC m=+17.207059107" Jan 30 13:24:44.658248 kubelet[3455]: I0130 13:24:44.658171 3455 topology_manager.go:215] "Topology Admit Handler" podUID="c6dbcf90-fed1-4eb3-9b55-393ac028ea12" podNamespace="calico-system" podName="calico-typha-7c75f4fd6-7422q" Jan 30 13:24:44.665649 kubelet[3455]: W0130 13:24:44.665331 3455 reflector.go:547] object-"calico-system"/"typha-certs": failed to list *v1.Secret: secrets "typha-certs" is forbidden: User "system:node:ci-4186.1.0-a-a7402f3063" cannot list resource "secrets" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4186.1.0-a-a7402f3063' and this object Jan 30 13:24:44.665649 kubelet[3455]: E0130 13:24:44.665376 3455 reflector.go:150] object-"calico-system"/"typha-certs": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "typha-certs" is forbidden: User "system:node:ci-4186.1.0-a-a7402f3063" cannot list resource "secrets" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4186.1.0-a-a7402f3063' and this object Jan 30 13:24:44.666180 kubelet[3455]: W0130 13:24:44.666114 3455 reflector.go:547] object-"calico-system"/"tigera-ca-bundle": failed to list *v1.ConfigMap: configmaps "tigera-ca-bundle" is forbidden: User "system:node:ci-4186.1.0-a-a7402f3063" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4186.1.0-a-a7402f3063' and this object Jan 30 13:24:44.666180 kubelet[3455]: E0130 13:24:44.666153 3455 reflector.go:150] object-"calico-system"/"tigera-ca-bundle": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "tigera-ca-bundle" is forbidden: User "system:node:ci-4186.1.0-a-a7402f3063" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4186.1.0-a-a7402f3063' and this object Jan 30 13:24:44.667963 kubelet[3455]: W0130 13:24:44.667925 3455 reflector.go:547] object-"calico-system"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4186.1.0-a-a7402f3063" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4186.1.0-a-a7402f3063' and this object Jan 30 13:24:44.667963 kubelet[3455]: E0130 13:24:44.667964 3455 reflector.go:150] object-"calico-system"/"kube-root-ca.crt": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:ci-4186.1.0-a-a7402f3063" cannot list resource "configmaps" in API group "" in the namespace "calico-system": no relationship found between node 'ci-4186.1.0-a-a7402f3063' and this object Jan 30 13:24:44.669814 systemd[1]: Created slice kubepods-besteffort-podc6dbcf90_fed1_4eb3_9b55_393ac028ea12.slice - libcontainer container kubepods-besteffort-podc6dbcf90_fed1_4eb3_9b55_393ac028ea12.slice. Jan 30 13:24:44.723589 kubelet[3455]: I0130 13:24:44.723316 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-typha-certs\") pod \"calico-typha-7c75f4fd6-7422q\" (UID: \"c6dbcf90-fed1-4eb3-9b55-393ac028ea12\") " pod="calico-system/calico-typha-7c75f4fd6-7422q" Jan 30 13:24:44.723589 kubelet[3455]: I0130 13:24:44.723365 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdxh2\" (UniqueName: \"kubernetes.io/projected/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-kube-api-access-wdxh2\") pod \"calico-typha-7c75f4fd6-7422q\" (UID: \"c6dbcf90-fed1-4eb3-9b55-393ac028ea12\") " pod="calico-system/calico-typha-7c75f4fd6-7422q" Jan 30 13:24:44.723589 kubelet[3455]: I0130 13:24:44.723409 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-tigera-ca-bundle\") pod \"calico-typha-7c75f4fd6-7422q\" (UID: \"c6dbcf90-fed1-4eb3-9b55-393ac028ea12\") " pod="calico-system/calico-typha-7c75f4fd6-7422q" Jan 30 13:24:44.766519 kubelet[3455]: I0130 13:24:44.766275 3455 topology_manager.go:215] "Topology Admit Handler" podUID="788d3ec6-3317-4475-a4e0-bb103af4d1c2" podNamespace="calico-system" podName="calico-node-f6k44" Jan 30 13:24:44.777483 systemd[1]: Created slice kubepods-besteffort-pod788d3ec6_3317_4475_a4e0_bb103af4d1c2.slice - libcontainer container kubepods-besteffort-pod788d3ec6_3317_4475_a4e0_bb103af4d1c2.slice. Jan 30 13:24:44.823928 kubelet[3455]: I0130 13:24:44.823690 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-var-run-calico\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.823928 kubelet[3455]: I0130 13:24:44.823756 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-net-dir\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.823928 kubelet[3455]: I0130 13:24:44.823804 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-var-lib-calico\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.824914 kubelet[3455]: I0130 13:24:44.824129 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-xtables-lock\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.824914 kubelet[3455]: I0130 13:24:44.824260 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-flexvol-driver-host\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.824914 kubelet[3455]: I0130 13:24:44.824286 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdjxc\" (UniqueName: \"kubernetes.io/projected/788d3ec6-3317-4475-a4e0-bb103af4d1c2-kube-api-access-kdjxc\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.824914 kubelet[3455]: I0130 13:24:44.824308 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/788d3ec6-3317-4475-a4e0-bb103af4d1c2-tigera-ca-bundle\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.824914 kubelet[3455]: I0130 13:24:44.824339 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/788d3ec6-3317-4475-a4e0-bb103af4d1c2-node-certs\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.825109 kubelet[3455]: I0130 13:24:44.824359 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-policysync\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.825109 kubelet[3455]: I0130 13:24:44.824378 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-log-dir\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.825109 kubelet[3455]: I0130 13:24:44.825028 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-lib-modules\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.825109 kubelet[3455]: I0130 13:24:44.825062 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-bin-dir\") pod \"calico-node-f6k44\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " pod="calico-system/calico-node-f6k44" Jan 30 13:24:44.913404 kubelet[3455]: I0130 13:24:44.911991 3455 topology_manager.go:215] "Topology Admit Handler" podUID="17a70524-69f8-4a61-9912-a962709a69cd" podNamespace="calico-system" podName="csi-node-driver-42f6k" Jan 30 13:24:44.915279 kubelet[3455]: E0130 13:24:44.915247 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:44.935045 kubelet[3455]: E0130 13:24:44.933485 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:44.935045 kubelet[3455]: W0130 13:24:44.933513 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:44.935045 kubelet[3455]: E0130 13:24:44.933544 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.008837 kubelet[3455]: E0130 13:24:45.008800 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.009254 kubelet[3455]: W0130 13:24:45.009082 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.009254 kubelet[3455]: E0130 13:24:45.009113 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.010506 kubelet[3455]: E0130 13:24:45.009967 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.010506 kubelet[3455]: W0130 13:24:45.009985 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.010506 kubelet[3455]: E0130 13:24:45.010003 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.011175 kubelet[3455]: E0130 13:24:45.010749 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.011175 kubelet[3455]: W0130 13:24:45.010765 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.011175 kubelet[3455]: E0130 13:24:45.010781 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.011834 kubelet[3455]: E0130 13:24:45.011621 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.011834 kubelet[3455]: W0130 13:24:45.011637 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.011834 kubelet[3455]: E0130 13:24:45.011658 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.012940 kubelet[3455]: E0130 13:24:45.012867 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.012940 kubelet[3455]: W0130 13:24:45.012882 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.012940 kubelet[3455]: E0130 13:24:45.012897 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.013806 kubelet[3455]: E0130 13:24:45.013613 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.013806 kubelet[3455]: W0130 13:24:45.013629 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.013806 kubelet[3455]: E0130 13:24:45.013642 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.014882 kubelet[3455]: E0130 13:24:45.014530 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.014882 kubelet[3455]: W0130 13:24:45.014546 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.014882 kubelet[3455]: E0130 13:24:45.014560 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.015591 kubelet[3455]: E0130 13:24:45.015498 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.015591 kubelet[3455]: W0130 13:24:45.015512 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.015591 kubelet[3455]: E0130 13:24:45.015525 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.016134 kubelet[3455]: E0130 13:24:45.016030 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.016134 kubelet[3455]: W0130 13:24:45.016044 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.016134 kubelet[3455]: E0130 13:24:45.016057 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.016364 kubelet[3455]: E0130 13:24:45.016307 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.016364 kubelet[3455]: W0130 13:24:45.016318 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.016364 kubelet[3455]: E0130 13:24:45.016330 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.016685 kubelet[3455]: E0130 13:24:45.016595 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.016685 kubelet[3455]: W0130 13:24:45.016607 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.016685 kubelet[3455]: E0130 13:24:45.016617 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.016955 kubelet[3455]: E0130 13:24:45.016881 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.016955 kubelet[3455]: W0130 13:24:45.016893 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.016955 kubelet[3455]: E0130 13:24:45.016904 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.017242 kubelet[3455]: E0130 13:24:45.017230 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.017381 kubelet[3455]: W0130 13:24:45.017291 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.017381 kubelet[3455]: E0130 13:24:45.017305 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.017621 kubelet[3455]: E0130 13:24:45.017552 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.017621 kubelet[3455]: W0130 13:24:45.017564 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.017621 kubelet[3455]: E0130 13:24:45.017574 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.018001 kubelet[3455]: E0130 13:24:45.017903 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.018001 kubelet[3455]: W0130 13:24:45.017915 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.018001 kubelet[3455]: E0130 13:24:45.017925 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.018299 kubelet[3455]: E0130 13:24:45.018145 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.018299 kubelet[3455]: W0130 13:24:45.018155 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.018299 kubelet[3455]: E0130 13:24:45.018164 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.018436 kubelet[3455]: E0130 13:24:45.018425 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.018491 kubelet[3455]: W0130 13:24:45.018481 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.018610 kubelet[3455]: E0130 13:24:45.018533 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.018988 kubelet[3455]: E0130 13:24:45.018925 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.018988 kubelet[3455]: W0130 13:24:45.018938 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.018988 kubelet[3455]: E0130 13:24:45.018949 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.019338 kubelet[3455]: E0130 13:24:45.019236 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.019338 kubelet[3455]: W0130 13:24:45.019248 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.019338 kubelet[3455]: E0130 13:24:45.019258 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.019600 kubelet[3455]: E0130 13:24:45.019492 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.019600 kubelet[3455]: W0130 13:24:45.019503 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.019600 kubelet[3455]: E0130 13:24:45.019512 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.026821 kubelet[3455]: E0130 13:24:45.026466 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.026821 kubelet[3455]: W0130 13:24:45.026492 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.026821 kubelet[3455]: E0130 13:24:45.026518 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.026821 kubelet[3455]: I0130 13:24:45.026549 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/17a70524-69f8-4a61-9912-a962709a69cd-registration-dir\") pod \"csi-node-driver-42f6k\" (UID: \"17a70524-69f8-4a61-9912-a962709a69cd\") " pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:45.027892 kubelet[3455]: E0130 13:24:45.027623 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.027892 kubelet[3455]: W0130 13:24:45.027650 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.027892 kubelet[3455]: E0130 13:24:45.027680 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.030371 kubelet[3455]: E0130 13:24:45.030225 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.030371 kubelet[3455]: W0130 13:24:45.030249 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.030371 kubelet[3455]: E0130 13:24:45.030279 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.030371 kubelet[3455]: I0130 13:24:45.030309 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"varrun\" (UniqueName: \"kubernetes.io/host-path/17a70524-69f8-4a61-9912-a962709a69cd-varrun\") pod \"csi-node-driver-42f6k\" (UID: \"17a70524-69f8-4a61-9912-a962709a69cd\") " pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:45.030579 kubelet[3455]: E0130 13:24:45.030540 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.030579 kubelet[3455]: W0130 13:24:45.030553 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.030579 kubelet[3455]: E0130 13:24:45.030572 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.030768 kubelet[3455]: E0130 13:24:45.030750 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.030768 kubelet[3455]: W0130 13:24:45.030765 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.030862 kubelet[3455]: E0130 13:24:45.030783 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.030961 kubelet[3455]: E0130 13:24:45.030946 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.030961 kubelet[3455]: W0130 13:24:45.030959 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.031051 kubelet[3455]: E0130 13:24:45.030974 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.031152 kubelet[3455]: E0130 13:24:45.031135 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.031152 kubelet[3455]: W0130 13:24:45.031148 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.031242 kubelet[3455]: E0130 13:24:45.031165 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.031320 kubelet[3455]: E0130 13:24:45.031305 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.031320 kubelet[3455]: W0130 13:24:45.031317 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.031375 kubelet[3455]: E0130 13:24:45.031330 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.031375 kubelet[3455]: I0130 13:24:45.031348 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/17a70524-69f8-4a61-9912-a962709a69cd-kubelet-dir\") pod \"csi-node-driver-42f6k\" (UID: \"17a70524-69f8-4a61-9912-a962709a69cd\") " pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:45.031528 kubelet[3455]: E0130 13:24:45.031498 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.031528 kubelet[3455]: W0130 13:24:45.031524 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.031652 kubelet[3455]: E0130 13:24:45.031604 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.031777 kubelet[3455]: E0130 13:24:45.031679 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.031777 kubelet[3455]: W0130 13:24:45.031692 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.031952 kubelet[3455]: E0130 13:24:45.031862 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.031952 kubelet[3455]: W0130 13:24:45.031876 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.031952 kubelet[3455]: E0130 13:24:45.031887 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.031952 kubelet[3455]: E0130 13:24:45.031752 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.032268 kubelet[3455]: E0130 13:24:45.032012 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.032268 kubelet[3455]: W0130 13:24:45.032019 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.032268 kubelet[3455]: E0130 13:24:45.032028 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.032268 kubelet[3455]: E0130 13:24:45.032170 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.032268 kubelet[3455]: W0130 13:24:45.032179 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.032268 kubelet[3455]: E0130 13:24:45.032194 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.032644 kubelet[3455]: E0130 13:24:45.032318 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.032644 kubelet[3455]: W0130 13:24:45.032325 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.032644 kubelet[3455]: E0130 13:24:45.032333 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.032644 kubelet[3455]: I0130 13:24:45.032359 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/17a70524-69f8-4a61-9912-a962709a69cd-socket-dir\") pod \"csi-node-driver-42f6k\" (UID: \"17a70524-69f8-4a61-9912-a962709a69cd\") " pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:45.032644 kubelet[3455]: E0130 13:24:45.032499 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.032644 kubelet[3455]: W0130 13:24:45.032509 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.032644 kubelet[3455]: E0130 13:24:45.032521 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.032644 kubelet[3455]: E0130 13:24:45.032656 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.032644 kubelet[3455]: W0130 13:24:45.032665 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.033232 kubelet[3455]: E0130 13:24:45.032679 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.033232 kubelet[3455]: I0130 13:24:45.032695 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4sht\" (UniqueName: \"kubernetes.io/projected/17a70524-69f8-4a61-9912-a962709a69cd-kube-api-access-c4sht\") pod \"csi-node-driver-42f6k\" (UID: \"17a70524-69f8-4a61-9912-a962709a69cd\") " pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:45.034034 kubelet[3455]: E0130 13:24:45.033769 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.034034 kubelet[3455]: W0130 13:24:45.033789 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.034034 kubelet[3455]: E0130 13:24:45.033815 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.034744 kubelet[3455]: E0130 13:24:45.034545 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.034744 kubelet[3455]: W0130 13:24:45.034563 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.034744 kubelet[3455]: E0130 13:24:45.034597 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.035538 kubelet[3455]: E0130 13:24:45.035345 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.035538 kubelet[3455]: W0130 13:24:45.035364 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.035538 kubelet[3455]: E0130 13:24:45.035381 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.036142 kubelet[3455]: E0130 13:24:45.036073 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.036142 kubelet[3455]: W0130 13:24:45.036091 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.036142 kubelet[3455]: E0130 13:24:45.036111 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.134227 kubelet[3455]: E0130 13:24:45.134183 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.134227 kubelet[3455]: W0130 13:24:45.134218 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.134399 kubelet[3455]: E0130 13:24:45.134241 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.134749 kubelet[3455]: E0130 13:24:45.134426 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.134749 kubelet[3455]: W0130 13:24:45.134441 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.134749 kubelet[3455]: E0130 13:24:45.134451 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.134749 kubelet[3455]: E0130 13:24:45.134656 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.134749 kubelet[3455]: W0130 13:24:45.134667 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.134749 kubelet[3455]: E0130 13:24:45.134679 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.134930 kubelet[3455]: E0130 13:24:45.134910 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.134930 kubelet[3455]: W0130 13:24:45.134925 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.134976 kubelet[3455]: E0130 13:24:45.134938 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.135741 kubelet[3455]: E0130 13:24:45.135105 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.135741 kubelet[3455]: W0130 13:24:45.135119 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.135741 kubelet[3455]: E0130 13:24:45.135133 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.135741 kubelet[3455]: E0130 13:24:45.135319 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.135741 kubelet[3455]: W0130 13:24:45.135327 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.135741 kubelet[3455]: E0130 13:24:45.135344 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.135741 kubelet[3455]: E0130 13:24:45.135514 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.135741 kubelet[3455]: W0130 13:24:45.135522 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.135741 kubelet[3455]: E0130 13:24:45.135536 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.135741 kubelet[3455]: E0130 13:24:45.135688 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.135953 kubelet[3455]: W0130 13:24:45.135695 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.135953 kubelet[3455]: E0130 13:24:45.135748 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.135953 kubelet[3455]: E0130 13:24:45.135926 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.135953 kubelet[3455]: W0130 13:24:45.135934 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.136033 kubelet[3455]: E0130 13:24:45.136012 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.136112 kubelet[3455]: E0130 13:24:45.136095 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.136112 kubelet[3455]: W0130 13:24:45.136108 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.136263 kubelet[3455]: E0130 13:24:45.136173 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.136299 kubelet[3455]: E0130 13:24:45.136281 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.136299 kubelet[3455]: W0130 13:24:45.136288 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.136353 kubelet[3455]: E0130 13:24:45.136298 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.136473 kubelet[3455]: E0130 13:24:45.136454 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.136473 kubelet[3455]: W0130 13:24:45.136466 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.136530 kubelet[3455]: E0130 13:24:45.136480 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.136648 kubelet[3455]: E0130 13:24:45.136631 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.136648 kubelet[3455]: W0130 13:24:45.136644 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.136725 kubelet[3455]: E0130 13:24:45.136659 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.136887 kubelet[3455]: E0130 13:24:45.136869 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.136887 kubelet[3455]: W0130 13:24:45.136884 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.136942 kubelet[3455]: E0130 13:24:45.136902 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.137388 kubelet[3455]: E0130 13:24:45.137130 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.137388 kubelet[3455]: W0130 13:24:45.137145 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.137388 kubelet[3455]: E0130 13:24:45.137163 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.137388 kubelet[3455]: E0130 13:24:45.137320 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.137388 kubelet[3455]: W0130 13:24:45.137328 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.137509 kubelet[3455]: E0130 13:24:45.137394 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.137601 kubelet[3455]: E0130 13:24:45.137582 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.137601 kubelet[3455]: W0130 13:24:45.137595 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.137674 kubelet[3455]: E0130 13:24:45.137657 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.138110 kubelet[3455]: E0130 13:24:45.137819 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.138110 kubelet[3455]: W0130 13:24:45.137833 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.138110 kubelet[3455]: E0130 13:24:45.137893 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.138110 kubelet[3455]: E0130 13:24:45.137985 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.138110 kubelet[3455]: W0130 13:24:45.137992 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.138110 kubelet[3455]: E0130 13:24:45.138068 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.138400 kubelet[3455]: E0130 13:24:45.138146 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.138400 kubelet[3455]: W0130 13:24:45.138153 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.138400 kubelet[3455]: E0130 13:24:45.138163 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.138400 kubelet[3455]: E0130 13:24:45.138349 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.138400 kubelet[3455]: W0130 13:24:45.138359 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.138400 kubelet[3455]: E0130 13:24:45.138373 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.138774 kubelet[3455]: E0130 13:24:45.138537 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.138774 kubelet[3455]: W0130 13:24:45.138553 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.138774 kubelet[3455]: E0130 13:24:45.138575 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.138904 kubelet[3455]: E0130 13:24:45.138883 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.138904 kubelet[3455]: W0130 13:24:45.138899 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.138958 kubelet[3455]: E0130 13:24:45.138912 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.139143 kubelet[3455]: E0130 13:24:45.139128 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.139143 kubelet[3455]: W0130 13:24:45.139141 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.139197 kubelet[3455]: E0130 13:24:45.139180 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.139802 kubelet[3455]: E0130 13:24:45.139352 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.139802 kubelet[3455]: W0130 13:24:45.139367 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.139802 kubelet[3455]: E0130 13:24:45.139407 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.139802 kubelet[3455]: E0130 13:24:45.139537 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.139802 kubelet[3455]: W0130 13:24:45.139544 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.139802 kubelet[3455]: E0130 13:24:45.139618 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.139952 kubelet[3455]: E0130 13:24:45.139934 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.139952 kubelet[3455]: W0130 13:24:45.139945 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.140008 kubelet[3455]: E0130 13:24:45.139958 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.140207 kubelet[3455]: E0130 13:24:45.140187 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.140207 kubelet[3455]: W0130 13:24:45.140202 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.140272 kubelet[3455]: E0130 13:24:45.140218 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.140734 kubelet[3455]: E0130 13:24:45.140392 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.140734 kubelet[3455]: W0130 13:24:45.140443 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.140734 kubelet[3455]: E0130 13:24:45.140461 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.141969 kubelet[3455]: E0130 13:24:45.141944 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.141969 kubelet[3455]: W0130 13:24:45.141968 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.142049 kubelet[3455]: E0130 13:24:45.141985 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.238024 kubelet[3455]: E0130 13:24:45.237989 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.238024 kubelet[3455]: W0130 13:24:45.238017 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.238204 kubelet[3455]: E0130 13:24:45.238040 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.238375 kubelet[3455]: E0130 13:24:45.238352 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.238375 kubelet[3455]: W0130 13:24:45.238368 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.238446 kubelet[3455]: E0130 13:24:45.238379 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.238599 kubelet[3455]: E0130 13:24:45.238580 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.238599 kubelet[3455]: W0130 13:24:45.238593 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.238673 kubelet[3455]: E0130 13:24:45.238602 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.238856 kubelet[3455]: E0130 13:24:45.238837 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.238856 kubelet[3455]: W0130 13:24:45.238852 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.238926 kubelet[3455]: E0130 13:24:45.238861 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.239073 kubelet[3455]: E0130 13:24:45.239052 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.239073 kubelet[3455]: W0130 13:24:45.239067 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.239149 kubelet[3455]: E0130 13:24:45.239076 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.239283 kubelet[3455]: E0130 13:24:45.239264 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.239283 kubelet[3455]: W0130 13:24:45.239278 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.239341 kubelet[3455]: E0130 13:24:45.239286 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.340689 kubelet[3455]: E0130 13:24:45.340549 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.340689 kubelet[3455]: W0130 13:24:45.340573 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.340689 kubelet[3455]: E0130 13:24:45.340593 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.341274 kubelet[3455]: E0130 13:24:45.341254 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.341274 kubelet[3455]: W0130 13:24:45.341268 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.341274 kubelet[3455]: E0130 13:24:45.341278 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.341551 kubelet[3455]: E0130 13:24:45.341532 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.341551 kubelet[3455]: W0130 13:24:45.341546 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.341727 kubelet[3455]: E0130 13:24:45.341555 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.341772 kubelet[3455]: E0130 13:24:45.341738 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.341772 kubelet[3455]: W0130 13:24:45.341751 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.341772 kubelet[3455]: E0130 13:24:45.341759 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.341937 kubelet[3455]: E0130 13:24:45.341920 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.341937 kubelet[3455]: W0130 13:24:45.341935 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.342010 kubelet[3455]: E0130 13:24:45.341944 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.342146 kubelet[3455]: E0130 13:24:45.342128 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.342146 kubelet[3455]: W0130 13:24:45.342137 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.342146 kubelet[3455]: E0130 13:24:45.342145 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.442814 kubelet[3455]: E0130 13:24:45.442782 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.442814 kubelet[3455]: W0130 13:24:45.442808 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.442968 kubelet[3455]: E0130 13:24:45.442830 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.442968 kubelet[3455]: E0130 13:24:45.442998 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.442968 kubelet[3455]: W0130 13:24:45.443009 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.442968 kubelet[3455]: E0130 13:24:45.443017 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.443364 kubelet[3455]: E0130 13:24:45.443174 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.443364 kubelet[3455]: W0130 13:24:45.443183 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.443364 kubelet[3455]: E0130 13:24:45.443192 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.443364 kubelet[3455]: E0130 13:24:45.443341 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.443364 kubelet[3455]: W0130 13:24:45.443350 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.443364 kubelet[3455]: E0130 13:24:45.443358 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.443789 kubelet[3455]: E0130 13:24:45.443496 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.443789 kubelet[3455]: W0130 13:24:45.443503 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.443789 kubelet[3455]: E0130 13:24:45.443511 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.443789 kubelet[3455]: E0130 13:24:45.443647 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.443789 kubelet[3455]: W0130 13:24:45.443655 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.443789 kubelet[3455]: E0130 13:24:45.443662 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.535873 kubelet[3455]: E0130 13:24:45.535653 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.535873 kubelet[3455]: W0130 13:24:45.535678 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.535873 kubelet[3455]: E0130 13:24:45.535698 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.542957 kubelet[3455]: E0130 13:24:45.542784 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.542957 kubelet[3455]: W0130 13:24:45.542809 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.542957 kubelet[3455]: E0130 13:24:45.542832 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.544829 kubelet[3455]: E0130 13:24:45.544625 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.544829 kubelet[3455]: W0130 13:24:45.544649 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.544829 kubelet[3455]: E0130 13:24:45.544671 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.545279 kubelet[3455]: E0130 13:24:45.545151 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.545279 kubelet[3455]: W0130 13:24:45.545165 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.545279 kubelet[3455]: E0130 13:24:45.545179 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.545790 kubelet[3455]: E0130 13:24:45.545670 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.545790 kubelet[3455]: W0130 13:24:45.545684 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.545790 kubelet[3455]: E0130 13:24:45.545697 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.546154 kubelet[3455]: E0130 13:24:45.546050 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.546154 kubelet[3455]: W0130 13:24:45.546100 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.546154 kubelet[3455]: E0130 13:24:45.546113 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.647319 kubelet[3455]: E0130 13:24:45.647170 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.647319 kubelet[3455]: W0130 13:24:45.647196 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.647319 kubelet[3455]: E0130 13:24:45.647218 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.647700 kubelet[3455]: E0130 13:24:45.647678 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.647700 kubelet[3455]: W0130 13:24:45.647698 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.647809 kubelet[3455]: E0130 13:24:45.647735 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.648078 kubelet[3455]: E0130 13:24:45.648053 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.648078 kubelet[3455]: W0130 13:24:45.648067 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.648164 kubelet[3455]: E0130 13:24:45.648080 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.648568 kubelet[3455]: E0130 13:24:45.648549 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.648568 kubelet[3455]: W0130 13:24:45.648566 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.648639 kubelet[3455]: E0130 13:24:45.648581 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.717279 kubelet[3455]: E0130 13:24:45.712018 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.717279 kubelet[3455]: W0130 13:24:45.712047 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.717279 kubelet[3455]: E0130 13:24:45.712072 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.717279 kubelet[3455]: E0130 13:24:45.716954 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.717279 kubelet[3455]: W0130 13:24:45.716975 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.717279 kubelet[3455]: E0130 13:24:45.717006 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.717279 kubelet[3455]: E0130 13:24:45.717222 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.717279 kubelet[3455]: W0130 13:24:45.717232 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.717279 kubelet[3455]: E0130 13:24:45.717241 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.750137 kubelet[3455]: E0130 13:24:45.749975 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.750137 kubelet[3455]: W0130 13:24:45.750002 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.750137 kubelet[3455]: E0130 13:24:45.750025 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.824518 kubelet[3455]: E0130 13:24:45.824355 3455 secret.go:194] Couldn't get secret calico-system/typha-certs: failed to sync secret cache: timed out waiting for the condition Jan 30 13:24:45.824518 kubelet[3455]: E0130 13:24:45.824467 3455 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-typha-certs podName:c6dbcf90-fed1-4eb3-9b55-393ac028ea12 nodeName:}" failed. No retries permitted until 2025-01-30 13:24:46.324444614 +0000 UTC m=+23.554041601 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "typha-certs" (UniqueName: "kubernetes.io/secret/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-typha-certs") pod "calico-typha-7c75f4fd6-7422q" (UID: "c6dbcf90-fed1-4eb3-9b55-393ac028ea12") : failed to sync secret cache: timed out waiting for the condition Jan 30 13:24:45.851448 kubelet[3455]: E0130 13:24:45.851410 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.851448 kubelet[3455]: W0130 13:24:45.851438 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.851448 kubelet[3455]: E0130 13:24:45.851459 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.952828 kubelet[3455]: E0130 13:24:45.952793 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:45.952828 kubelet[3455]: W0130 13:24:45.952819 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:45.952828 kubelet[3455]: E0130 13:24:45.952847 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:45.983929 containerd[1781]: time="2025-01-30T13:24:45.983876818Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-f6k44,Uid:788d3ec6-3317-4475-a4e0-bb103af4d1c2,Namespace:calico-system,Attempt:0,}" Jan 30 13:24:46.031695 containerd[1781]: time="2025-01-30T13:24:46.029505819Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:24:46.031695 containerd[1781]: time="2025-01-30T13:24:46.029570379Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:24:46.031695 containerd[1781]: time="2025-01-30T13:24:46.029585659Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:46.031695 containerd[1781]: time="2025-01-30T13:24:46.029673019Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:46.053972 systemd[1]: Started cri-containerd-8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618.scope - libcontainer container 8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618. Jan 30 13:24:46.054709 kubelet[3455]: E0130 13:24:46.054287 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:46.054709 kubelet[3455]: W0130 13:24:46.054395 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:46.055474 kubelet[3455]: E0130 13:24:46.055170 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:46.080848 containerd[1781]: time="2025-01-30T13:24:46.079752620Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-f6k44,Uid:788d3ec6-3317-4475-a4e0-bb103af4d1c2,Namespace:calico-system,Attempt:0,} returns sandbox id \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\"" Jan 30 13:24:46.083063 containerd[1781]: time="2025-01-30T13:24:46.082755780Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\"" Jan 30 13:24:46.156129 kubelet[3455]: E0130 13:24:46.156090 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:46.156129 kubelet[3455]: W0130 13:24:46.156118 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:46.156129 kubelet[3455]: E0130 13:24:46.156139 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:46.257709 kubelet[3455]: E0130 13:24:46.257602 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:46.257709 kubelet[3455]: W0130 13:24:46.257627 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:46.257709 kubelet[3455]: E0130 13:24:46.257647 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:46.358816 kubelet[3455]: E0130 13:24:46.358669 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:46.358816 kubelet[3455]: W0130 13:24:46.358696 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:46.358816 kubelet[3455]: E0130 13:24:46.358743 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:46.359340 kubelet[3455]: E0130 13:24:46.358980 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:46.359340 kubelet[3455]: W0130 13:24:46.358989 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:46.359340 kubelet[3455]: E0130 13:24:46.358999 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:46.359340 kubelet[3455]: E0130 13:24:46.359138 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:46.359340 kubelet[3455]: W0130 13:24:46.359146 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:46.359340 kubelet[3455]: E0130 13:24:46.359154 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:46.359340 kubelet[3455]: E0130 13:24:46.359283 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:46.359340 kubelet[3455]: W0130 13:24:46.359292 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:46.359340 kubelet[3455]: E0130 13:24:46.359300 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:46.359340 kubelet[3455]: E0130 13:24:46.359467 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:46.359898 kubelet[3455]: W0130 13:24:46.359475 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:46.359898 kubelet[3455]: E0130 13:24:46.359484 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:46.364094 kubelet[3455]: E0130 13:24:46.363995 3455 driver-call.go:262] Failed to unmarshal output for command: init, output: "", error: unexpected end of JSON input Jan 30 13:24:46.364094 kubelet[3455]: W0130 13:24:46.364022 3455 driver-call.go:149] FlexVolume: driver call failed: executable: /opt/libexec/kubernetes/kubelet-plugins/volume/exec/nodeagent~uds/uds, args: [init], error: executable file not found in $PATH, output: "" Jan 30 13:24:46.364094 kubelet[3455]: E0130 13:24:46.364044 3455 plugins.go:730] "Error dynamically probing plugins" err="error creating Flexvolume plugin from directory nodeagent~uds, skipping. Error: unexpected end of JSON input" Jan 30 13:24:46.475974 containerd[1781]: time="2025-01-30T13:24:46.475901229Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7c75f4fd6-7422q,Uid:c6dbcf90-fed1-4eb3-9b55-393ac028ea12,Namespace:calico-system,Attempt:0,}" Jan 30 13:24:46.532709 containerd[1781]: time="2025-01-30T13:24:46.532483431Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:24:46.532709 containerd[1781]: time="2025-01-30T13:24:46.532563431Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:24:46.532709 containerd[1781]: time="2025-01-30T13:24:46.532574991Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:46.533168 containerd[1781]: time="2025-01-30T13:24:46.532671511Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:24:46.551950 systemd[1]: Started cri-containerd-f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330.scope - libcontainer container f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330. Jan 30 13:24:46.587152 containerd[1781]: time="2025-01-30T13:24:46.587055152Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-7c75f4fd6-7422q,Uid:c6dbcf90-fed1-4eb3-9b55-393ac028ea12,Namespace:calico-system,Attempt:0,} returns sandbox id \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\"" Jan 30 13:24:46.886058 kubelet[3455]: E0130 13:24:46.885378 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:47.197957 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3122382902.mount: Deactivated successfully. Jan 30 13:24:47.332828 containerd[1781]: time="2025-01-30T13:24:47.332768916Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:47.335492 containerd[1781]: time="2025-01-30T13:24:47.335421314Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1: active requests=0, bytes read=6487603" Jan 30 13:24:47.338296 containerd[1781]: time="2025-01-30T13:24:47.338219113Z" level=info msg="ImageCreate event name:\"sha256:ece9bca32e64e726de8bbfc9e175a3ca91e0881cd40352bfcd1d107411f4f348\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:47.342126 containerd[1781]: time="2025-01-30T13:24:47.342061230Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:47.343156 containerd[1781]: time="2025-01-30T13:24:47.342756110Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" with image id \"sha256:ece9bca32e64e726de8bbfc9e175a3ca91e0881cd40352bfcd1d107411f4f348\", repo tag \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/pod2daemon-flexvol@sha256:a63f8b4ff531912d12d143664eb263fdbc6cd7b3ff4aa777dfb6e318a090462c\", size \"6487425\" in 1.25996001s" Jan 30 13:24:47.343156 containerd[1781]: time="2025-01-30T13:24:47.342794470Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/pod2daemon-flexvol:v3.29.1\" returns image reference \"sha256:ece9bca32e64e726de8bbfc9e175a3ca91e0881cd40352bfcd1d107411f4f348\"" Jan 30 13:24:47.344436 containerd[1781]: time="2025-01-30T13:24:47.344390869Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\"" Jan 30 13:24:47.346027 containerd[1781]: time="2025-01-30T13:24:47.345976028Z" level=info msg="CreateContainer within sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jan 30 13:24:47.381550 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3599331770.mount: Deactivated successfully. Jan 30 13:24:47.395152 containerd[1781]: time="2025-01-30T13:24:47.395057719Z" level=info msg="CreateContainer within sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\"" Jan 30 13:24:47.396395 containerd[1781]: time="2025-01-30T13:24:47.396291118Z" level=info msg="StartContainer for \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\"" Jan 30 13:24:47.426595 systemd[1]: Started cri-containerd-5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015.scope - libcontainer container 5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015. Jan 30 13:24:47.466045 containerd[1781]: time="2025-01-30T13:24:47.465984757Z" level=info msg="StartContainer for \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\" returns successfully" Jan 30 13:24:47.475748 systemd[1]: cri-containerd-5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015.scope: Deactivated successfully. Jan 30 13:24:48.001281 containerd[1781]: time="2025-01-30T13:24:48.001219961Z" level=info msg="shim disconnected" id=5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015 namespace=k8s.io Jan 30 13:24:48.001281 containerd[1781]: time="2025-01-30T13:24:48.001276601Z" level=warning msg="cleaning up after shim disconnected" id=5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015 namespace=k8s.io Jan 30 13:24:48.001281 containerd[1781]: time="2025-01-30T13:24:48.001288201Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:24:48.884775 kubelet[3455]: E0130 13:24:48.883491 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:49.459270 containerd[1781]: time="2025-01-30T13:24:49.459210100Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:49.461598 containerd[1781]: time="2025-01-30T13:24:49.461551579Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/typha:v3.29.1: active requests=0, bytes read=27861516" Jan 30 13:24:49.463757 containerd[1781]: time="2025-01-30T13:24:49.463644618Z" level=info msg="ImageCreate event name:\"sha256:1d1fc316829ae1650b0b1629b54232520f297e7c3b1444eecd290ae088902a28\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:49.468481 containerd[1781]: time="2025-01-30T13:24:49.468402975Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:49.469356 containerd[1781]: time="2025-01-30T13:24:49.469184094Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/typha:v3.29.1\" with image id \"sha256:1d1fc316829ae1650b0b1629b54232520f297e7c3b1444eecd290ae088902a28\", repo tag \"ghcr.io/flatcar/calico/typha:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/typha@sha256:768a194e1115c73bcbf35edb7afd18a63e16e08d940c79993565b6a3cca2da7c\", size \"29231162\" in 2.124749785s" Jan 30 13:24:49.469356 containerd[1781]: time="2025-01-30T13:24:49.469221294Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/typha:v3.29.1\" returns image reference \"sha256:1d1fc316829ae1650b0b1629b54232520f297e7c3b1444eecd290ae088902a28\"" Jan 30 13:24:49.471507 containerd[1781]: time="2025-01-30T13:24:49.471120573Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\"" Jan 30 13:24:49.487084 containerd[1781]: time="2025-01-30T13:24:49.484565885Z" level=info msg="CreateContainer within sandbox \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jan 30 13:24:49.524964 containerd[1781]: time="2025-01-30T13:24:49.524805541Z" level=info msg="CreateContainer within sandbox \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\"" Jan 30 13:24:49.525968 containerd[1781]: time="2025-01-30T13:24:49.525892341Z" level=info msg="StartContainer for \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\"" Jan 30 13:24:49.554956 systemd[1]: Started cri-containerd-ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a.scope - libcontainer container ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a. Jan 30 13:24:49.592923 containerd[1781]: time="2025-01-30T13:24:49.592842661Z" level=info msg="StartContainer for \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\" returns successfully" Jan 30 13:24:50.005096 kubelet[3455]: I0130 13:24:50.005005 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-7c75f4fd6-7422q" podStartSLOduration=3.124412156 podStartE2EDuration="6.004987898s" podCreationTimestamp="2025-01-30 13:24:44 +0000 UTC" firstStartedPulling="2025-01-30 13:24:46.589740712 +0000 UTC m=+23.819337699" lastFinishedPulling="2025-01-30 13:24:49.470316374 +0000 UTC m=+26.699913441" observedRunningTime="2025-01-30 13:24:50.004970458 +0000 UTC m=+27.234567445" watchObservedRunningTime="2025-01-30 13:24:50.004987898 +0000 UTC m=+27.234584885" Jan 30 13:24:50.883768 kubelet[3455]: E0130 13:24:50.883528 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:52.478948 containerd[1781]: time="2025-01-30T13:24:52.478879077Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:52.481368 containerd[1781]: time="2025-01-30T13:24:52.481200196Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/cni:v3.29.1: active requests=0, bytes read=89703123" Jan 30 13:24:52.484362 containerd[1781]: time="2025-01-30T13:24:52.484292554Z" level=info msg="ImageCreate event name:\"sha256:e5ca62af4ff61b88f55fe4e0d7723151103d3f6a470fd4ebb311a2de27a9597f\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:52.489679 containerd[1781]: time="2025-01-30T13:24:52.489600671Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:24:52.490587 containerd[1781]: time="2025-01-30T13:24:52.490422950Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/cni:v3.29.1\" with image id \"sha256:e5ca62af4ff61b88f55fe4e0d7723151103d3f6a470fd4ebb311a2de27a9597f\", repo tag \"ghcr.io/flatcar/calico/cni:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/cni@sha256:21e759d51c90dfb34fc1397dc180dd3a3fb564c2b0580d2f61ffe108f2a3c94b\", size \"91072777\" in 3.019258577s" Jan 30 13:24:52.490587 containerd[1781]: time="2025-01-30T13:24:52.490460350Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/cni:v3.29.1\" returns image reference \"sha256:e5ca62af4ff61b88f55fe4e0d7723151103d3f6a470fd4ebb311a2de27a9597f\"" Jan 30 13:24:52.494270 containerd[1781]: time="2025-01-30T13:24:52.494233868Z" level=info msg="CreateContainer within sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 30 13:24:52.533255 containerd[1781]: time="2025-01-30T13:24:52.533160005Z" level=info msg="CreateContainer within sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\"" Jan 30 13:24:52.535860 containerd[1781]: time="2025-01-30T13:24:52.533940524Z" level=info msg="StartContainer for \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\"" Jan 30 13:24:52.568941 systemd[1]: Started cri-containerd-8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5.scope - libcontainer container 8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5. Jan 30 13:24:52.603941 containerd[1781]: time="2025-01-30T13:24:52.603900883Z" level=info msg="StartContainer for \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\" returns successfully" Jan 30 13:24:52.884413 kubelet[3455]: E0130 13:24:52.884283 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:54.098853 containerd[1781]: time="2025-01-30T13:24:54.098788829Z" level=error msg="failed to reload cni configuration after receiving fs change event(WRITE \"/etc/cni/net.d/calico-kubeconfig\")" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Jan 30 13:24:54.101181 systemd[1]: cri-containerd-8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5.scope: Deactivated successfully. Jan 30 13:24:54.123203 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5-rootfs.mount: Deactivated successfully. Jan 30 13:24:54.197710 kubelet[3455]: I0130 13:24:54.197327 3455 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Jan 30 13:24:54.372404 kubelet[3455]: I0130 13:24:54.230092 3455 topology_manager.go:215] "Topology Admit Handler" podUID="efc446ea-7180-4d70-91d5-8919459a6efa" podNamespace="kube-system" podName="coredns-7db6d8ff4d-5djx9" Jan 30 13:24:54.372404 kubelet[3455]: I0130 13:24:54.238961 3455 topology_manager.go:215] "Topology Admit Handler" podUID="0b84efc9-a798-4f8e-8d4f-6bba2eb764e3" podNamespace="kube-system" podName="coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:54.372404 kubelet[3455]: I0130 13:24:54.240139 3455 topology_manager.go:215] "Topology Admit Handler" podUID="376f6955-c43d-49c5-a9d0-4891ee1dc8f6" podNamespace="calico-apiserver" podName="calico-apiserver-c978d777-x5x6w" Jan 30 13:24:54.372404 kubelet[3455]: I0130 13:24:54.244125 3455 topology_manager.go:215] "Topology Admit Handler" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" podNamespace="calico-system" podName="calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:54.372404 kubelet[3455]: I0130 13:24:54.244263 3455 topology_manager.go:215] "Topology Admit Handler" podUID="ce7cd8b7-6707-4478-a092-50a5a608f5e0" podNamespace="calico-apiserver" podName="calico-apiserver-c978d777-jmpf2" Jan 30 13:24:54.372404 kubelet[3455]: I0130 13:24:54.313343 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b84efc9-a798-4f8e-8d4f-6bba2eb764e3-config-volume\") pod \"coredns-7db6d8ff4d-5hs8w\" (UID: \"0b84efc9-a798-4f8e-8d4f-6bba2eb764e3\") " pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:54.372404 kubelet[3455]: I0130 13:24:54.313379 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mbdm\" (UniqueName: \"kubernetes.io/projected/aad42f51-7748-4be3-8b1d-97665b3346fc-kube-api-access-2mbdm\") pod \"calico-kube-controllers-6dd45f4c57-58wrn\" (UID: \"aad42f51-7748-4be3-8b1d-97665b3346fc\") " pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:54.239309 systemd[1]: Created slice kubepods-burstable-podefc446ea_7180_4d70_91d5_8919459a6efa.slice - libcontainer container kubepods-burstable-podefc446ea_7180_4d70_91d5_8919459a6efa.slice. Jan 30 13:24:54.372814 kubelet[3455]: I0130 13:24:54.313407 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aad42f51-7748-4be3-8b1d-97665b3346fc-tigera-ca-bundle\") pod \"calico-kube-controllers-6dd45f4c57-58wrn\" (UID: \"aad42f51-7748-4be3-8b1d-97665b3346fc\") " pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:54.372814 kubelet[3455]: I0130 13:24:54.313429 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kc2dr\" (UniqueName: \"kubernetes.io/projected/efc446ea-7180-4d70-91d5-8919459a6efa-kube-api-access-kc2dr\") pod \"coredns-7db6d8ff4d-5djx9\" (UID: \"efc446ea-7180-4d70-91d5-8919459a6efa\") " pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:54.372814 kubelet[3455]: I0130 13:24:54.313445 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/efc446ea-7180-4d70-91d5-8919459a6efa-config-volume\") pod \"coredns-7db6d8ff4d-5djx9\" (UID: \"efc446ea-7180-4d70-91d5-8919459a6efa\") " pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:54.372814 kubelet[3455]: I0130 13:24:54.313463 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/376f6955-c43d-49c5-a9d0-4891ee1dc8f6-calico-apiserver-certs\") pod \"calico-apiserver-c978d777-x5x6w\" (UID: \"376f6955-c43d-49c5-a9d0-4891ee1dc8f6\") " pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:54.372814 kubelet[3455]: I0130 13:24:54.313478 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x94kq\" (UniqueName: \"kubernetes.io/projected/376f6955-c43d-49c5-a9d0-4891ee1dc8f6-kube-api-access-x94kq\") pod \"calico-apiserver-c978d777-x5x6w\" (UID: \"376f6955-c43d-49c5-a9d0-4891ee1dc8f6\") " pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:54.256881 systemd[1]: Created slice kubepods-burstable-pod0b84efc9_a798_4f8e_8d4f_6bba2eb764e3.slice - libcontainer container kubepods-burstable-pod0b84efc9_a798_4f8e_8d4f_6bba2eb764e3.slice. Jan 30 13:24:54.372970 kubelet[3455]: I0130 13:24:54.313498 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kcdx\" (UniqueName: \"kubernetes.io/projected/0b84efc9-a798-4f8e-8d4f-6bba2eb764e3-kube-api-access-4kcdx\") pod \"coredns-7db6d8ff4d-5hs8w\" (UID: \"0b84efc9-a798-4f8e-8d4f-6bba2eb764e3\") " pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:54.372970 kubelet[3455]: I0130 13:24:54.313514 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"calico-apiserver-certs\" (UniqueName: \"kubernetes.io/secret/ce7cd8b7-6707-4478-a092-50a5a608f5e0-calico-apiserver-certs\") pod \"calico-apiserver-c978d777-jmpf2\" (UID: \"ce7cd8b7-6707-4478-a092-50a5a608f5e0\") " pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:54.372970 kubelet[3455]: I0130 13:24:54.313531 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fckw\" (UniqueName: \"kubernetes.io/projected/ce7cd8b7-6707-4478-a092-50a5a608f5e0-kube-api-access-5fckw\") pod \"calico-apiserver-c978d777-jmpf2\" (UID: \"ce7cd8b7-6707-4478-a092-50a5a608f5e0\") " pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:54.266059 systemd[1]: Created slice kubepods-besteffort-podce7cd8b7_6707_4478_a092_50a5a608f5e0.slice - libcontainer container kubepods-besteffort-podce7cd8b7_6707_4478_a092_50a5a608f5e0.slice. Jan 30 13:24:54.276030 systemd[1]: Created slice kubepods-besteffort-pod376f6955_c43d_49c5_a9d0_4891ee1dc8f6.slice - libcontainer container kubepods-besteffort-pod376f6955_c43d_49c5_a9d0_4891ee1dc8f6.slice. Jan 30 13:24:54.287030 systemd[1]: Created slice kubepods-besteffort-podaad42f51_7748_4be3_8b1d_97665b3346fc.slice - libcontainer container kubepods-besteffort-podaad42f51_7748_4be3_8b1d_97665b3346fc.slice. Jan 30 13:24:54.723350 containerd[1781]: time="2025-01-30T13:24:54.723296737Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:0,}" Jan 30 13:24:54.727401 containerd[1781]: time="2025-01-30T13:24:54.727113458Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:0,}" Jan 30 13:24:54.730224 containerd[1781]: time="2025-01-30T13:24:54.730171778Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:0,}" Jan 30 13:24:54.730510 containerd[1781]: time="2025-01-30T13:24:54.730461858Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:0,}" Jan 30 13:24:54.740178 containerd[1781]: time="2025-01-30T13:24:54.740123898Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:0,}" Jan 30 13:24:54.893926 systemd[1]: Created slice kubepods-besteffort-pod17a70524_69f8_4a61_9912_a962709a69cd.slice - libcontainer container kubepods-besteffort-pod17a70524_69f8_4a61_9912_a962709a69cd.slice. Jan 30 13:24:54.897408 containerd[1781]: time="2025-01-30T13:24:54.897132665Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:0,}" Jan 30 13:24:55.263586 containerd[1781]: time="2025-01-30T13:24:55.263487880Z" level=info msg="shim disconnected" id=8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5 namespace=k8s.io Jan 30 13:24:55.263586 containerd[1781]: time="2025-01-30T13:24:55.263570440Z" level=warning msg="cleaning up after shim disconnected" id=8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5 namespace=k8s.io Jan 30 13:24:55.263586 containerd[1781]: time="2025-01-30T13:24:55.263593600Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:24:55.548526 containerd[1781]: time="2025-01-30T13:24:55.548295189Z" level=error msg="Failed to destroy network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.549265 containerd[1781]: time="2025-01-30T13:24:55.549121109Z" level=error msg="encountered an error cleaning up failed sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.549265 containerd[1781]: time="2025-01-30T13:24:55.549194229Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.552151 kubelet[3455]: E0130 13:24:55.551034 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.552151 kubelet[3455]: E0130 13:24:55.551107 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:55.552151 kubelet[3455]: E0130 13:24:55.551126 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:55.552568 kubelet[3455]: E0130 13:24:55.551165 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" podUID="376f6955-c43d-49c5-a9d0-4891ee1dc8f6" Jan 30 13:24:55.575657 containerd[1781]: time="2025-01-30T13:24:55.575590264Z" level=error msg="Failed to destroy network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.576253 containerd[1781]: time="2025-01-30T13:24:55.576214624Z" level=error msg="encountered an error cleaning up failed sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.576775 containerd[1781]: time="2025-01-30T13:24:55.576739984Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:0,} failed, error" error="failed to setup network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.577072 kubelet[3455]: E0130 13:24:55.576967 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.577072 kubelet[3455]: E0130 13:24:55.577020 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:55.577072 kubelet[3455]: E0130 13:24:55.577038 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:55.577189 kubelet[3455]: E0130 13:24:55.577082 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" podUID="ce7cd8b7-6707-4478-a092-50a5a608f5e0" Jan 30 13:24:55.587232 containerd[1781]: time="2025-01-30T13:24:55.587080582Z" level=error msg="Failed to destroy network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.588093 containerd[1781]: time="2025-01-30T13:24:55.587889782Z" level=error msg="encountered an error cleaning up failed sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.588093 containerd[1781]: time="2025-01-30T13:24:55.587964262Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.589144 kubelet[3455]: E0130 13:24:55.588358 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.589144 kubelet[3455]: E0130 13:24:55.588413 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:55.589144 kubelet[3455]: E0130 13:24:55.588431 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:55.589337 containerd[1781]: time="2025-01-30T13:24:55.588586502Z" level=error msg="Failed to destroy network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.589337 containerd[1781]: time="2025-01-30T13:24:55.589104061Z" level=error msg="encountered an error cleaning up failed sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.589337 containerd[1781]: time="2025-01-30T13:24:55.589189861Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.589441 kubelet[3455]: E0130 13:24:55.588835 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5hs8w" podUID="0b84efc9-a798-4f8e-8d4f-6bba2eb764e3" Jan 30 13:24:55.589441 kubelet[3455]: E0130 13:24:55.589375 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.589441 kubelet[3455]: E0130 13:24:55.589414 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:55.589526 kubelet[3455]: E0130 13:24:55.589432 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:55.589526 kubelet[3455]: E0130 13:24:55.589463 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" Jan 30 13:24:55.590499 containerd[1781]: time="2025-01-30T13:24:55.589631341Z" level=error msg="Failed to destroy network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.590499 containerd[1781]: time="2025-01-30T13:24:55.590235461Z" level=error msg="encountered an error cleaning up failed sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.590499 containerd[1781]: time="2025-01-30T13:24:55.590287381Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.590624 kubelet[3455]: E0130 13:24:55.590463 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.590624 kubelet[3455]: E0130 13:24:55.590500 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:55.590624 kubelet[3455]: E0130 13:24:55.590514 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:55.590709 kubelet[3455]: E0130 13:24:55.590549 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:55.591800 containerd[1781]: time="2025-01-30T13:24:55.591599301Z" level=error msg="Failed to destroy network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.591997 containerd[1781]: time="2025-01-30T13:24:55.591941301Z" level=error msg="encountered an error cleaning up failed sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.592039 containerd[1781]: time="2025-01-30T13:24:55.592016821Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:0,} failed, error" error="failed to setup network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.592248 kubelet[3455]: E0130 13:24:55.592201 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:55.592300 kubelet[3455]: E0130 13:24:55.592257 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:55.592300 kubelet[3455]: E0130 13:24:55.592275 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:55.592351 kubelet[3455]: E0130 13:24:55.592312 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5djx9" podUID="efc446ea-7180-4d70-91d5-8919459a6efa" Jan 30 13:24:56.009384 containerd[1781]: time="2025-01-30T13:24:56.009333545Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\"" Jan 30 13:24:56.011759 kubelet[3455]: I0130 13:24:56.010859 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112" Jan 30 13:24:56.011997 containerd[1781]: time="2025-01-30T13:24:56.011672665Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:24:56.012512 containerd[1781]: time="2025-01-30T13:24:56.012430985Z" level=info msg="Ensure that sandbox 7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112 in task-service has been cleanup successfully" Jan 30 13:24:56.014630 containerd[1781]: time="2025-01-30T13:24:56.013500545Z" level=info msg="TearDown network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" successfully" Jan 30 13:24:56.014630 containerd[1781]: time="2025-01-30T13:24:56.014623864Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" returns successfully" Jan 30 13:24:56.014763 kubelet[3455]: I0130 13:24:56.014742 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611" Jan 30 13:24:56.016789 containerd[1781]: time="2025-01-30T13:24:56.016751704Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:1,}" Jan 30 13:24:56.017666 containerd[1781]: time="2025-01-30T13:24:56.017635664Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:24:56.017957 containerd[1781]: time="2025-01-30T13:24:56.017924624Z" level=info msg="Ensure that sandbox 618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611 in task-service has been cleanup successfully" Jan 30 13:24:56.019837 containerd[1781]: time="2025-01-30T13:24:56.019783783Z" level=info msg="TearDown network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" successfully" Jan 30 13:24:56.019837 containerd[1781]: time="2025-01-30T13:24:56.019807263Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" returns successfully" Jan 30 13:24:56.020497 containerd[1781]: time="2025-01-30T13:24:56.020313343Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:1,}" Jan 30 13:24:56.020741 kubelet[3455]: I0130 13:24:56.020666 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e" Jan 30 13:24:56.021651 containerd[1781]: time="2025-01-30T13:24:56.021357943Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:24:56.021651 containerd[1781]: time="2025-01-30T13:24:56.021554103Z" level=info msg="Ensure that sandbox 2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e in task-service has been cleanup successfully" Jan 30 13:24:56.023009 containerd[1781]: time="2025-01-30T13:24:56.022458223Z" level=info msg="TearDown network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" successfully" Jan 30 13:24:56.023009 containerd[1781]: time="2025-01-30T13:24:56.023005023Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" returns successfully" Jan 30 13:24:56.025097 containerd[1781]: time="2025-01-30T13:24:56.024422703Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:1,}" Jan 30 13:24:56.032540 kubelet[3455]: I0130 13:24:56.032128 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4" Jan 30 13:24:56.033647 containerd[1781]: time="2025-01-30T13:24:56.033314781Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:24:56.033778 containerd[1781]: time="2025-01-30T13:24:56.033567341Z" level=info msg="Ensure that sandbox 8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4 in task-service has been cleanup successfully" Jan 30 13:24:56.034005 containerd[1781]: time="2025-01-30T13:24:56.033986701Z" level=info msg="TearDown network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" successfully" Jan 30 13:24:56.034100 containerd[1781]: time="2025-01-30T13:24:56.034085301Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" returns successfully" Jan 30 13:24:56.038532 containerd[1781]: time="2025-01-30T13:24:56.038476620Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:1,}" Jan 30 13:24:56.043344 kubelet[3455]: I0130 13:24:56.042976 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14" Jan 30 13:24:56.044740 containerd[1781]: time="2025-01-30T13:24:56.044593739Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:24:56.045933 containerd[1781]: time="2025-01-30T13:24:56.045898299Z" level=info msg="Ensure that sandbox 7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14 in task-service has been cleanup successfully" Jan 30 13:24:56.047208 containerd[1781]: time="2025-01-30T13:24:56.047092539Z" level=info msg="TearDown network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" successfully" Jan 30 13:24:56.047208 containerd[1781]: time="2025-01-30T13:24:56.047134978Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" returns successfully" Jan 30 13:24:56.048184 kubelet[3455]: I0130 13:24:56.047751 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289" Jan 30 13:24:56.048271 containerd[1781]: time="2025-01-30T13:24:56.047970658Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:1,}" Jan 30 13:24:56.049237 containerd[1781]: time="2025-01-30T13:24:56.048856978Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:24:56.049237 containerd[1781]: time="2025-01-30T13:24:56.049039858Z" level=info msg="Ensure that sandbox c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289 in task-service has been cleanup successfully" Jan 30 13:24:56.049697 containerd[1781]: time="2025-01-30T13:24:56.049388218Z" level=info msg="TearDown network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" successfully" Jan 30 13:24:56.049697 containerd[1781]: time="2025-01-30T13:24:56.049451738Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" returns successfully" Jan 30 13:24:56.050640 containerd[1781]: time="2025-01-30T13:24:56.050606258Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:1,}" Jan 30 13:24:56.252433 containerd[1781]: time="2025-01-30T13:24:56.252289821Z" level=error msg="Failed to destroy network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.252899 containerd[1781]: time="2025-01-30T13:24:56.252846661Z" level=error msg="encountered an error cleaning up failed sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.253133 containerd[1781]: time="2025-01-30T13:24:56.253044941Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.253448 kubelet[3455]: E0130 13:24:56.253411 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.253530 kubelet[3455]: E0130 13:24:56.253467 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:56.253530 kubelet[3455]: E0130 13:24:56.253490 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:56.255141 kubelet[3455]: E0130 13:24:56.253527 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" podUID="376f6955-c43d-49c5-a9d0-4891ee1dc8f6" Jan 30 13:24:56.296946 containerd[1781]: time="2025-01-30T13:24:56.296744813Z" level=error msg="Failed to destroy network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.299990 containerd[1781]: time="2025-01-30T13:24:56.299678093Z" level=error msg="encountered an error cleaning up failed sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.300243 containerd[1781]: time="2025-01-30T13:24:56.299963413Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.300866 kubelet[3455]: E0130 13:24:56.300769 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.300866 kubelet[3455]: E0130 13:24:56.300821 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:56.300866 kubelet[3455]: E0130 13:24:56.300839 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:56.301683 kubelet[3455]: E0130 13:24:56.300877 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5hs8w" podUID="0b84efc9-a798-4f8e-8d4f-6bba2eb764e3" Jan 30 13:24:56.360934 containerd[1781]: time="2025-01-30T13:24:56.360840722Z" level=error msg="Failed to destroy network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.362611 containerd[1781]: time="2025-01-30T13:24:56.362140441Z" level=error msg="encountered an error cleaning up failed sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.362611 containerd[1781]: time="2025-01-30T13:24:56.362330481Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.363987 kubelet[3455]: E0130 13:24:56.362959 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.363987 kubelet[3455]: E0130 13:24:56.363015 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:56.363987 kubelet[3455]: E0130 13:24:56.363038 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:56.364157 kubelet[3455]: E0130 13:24:56.363079 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5djx9" podUID="efc446ea-7180-4d70-91d5-8919459a6efa" Jan 30 13:24:56.366503 containerd[1781]: time="2025-01-30T13:24:56.366456561Z" level=error msg="Failed to destroy network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.367765 containerd[1781]: time="2025-01-30T13:24:56.367649560Z" level=error msg="encountered an error cleaning up failed sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.367981 containerd[1781]: time="2025-01-30T13:24:56.367955680Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:1,} failed, error" error="failed to setup network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.369214 kubelet[3455]: E0130 13:24:56.368242 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.369214 kubelet[3455]: E0130 13:24:56.368295 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:56.369214 kubelet[3455]: E0130 13:24:56.368313 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:56.369375 kubelet[3455]: E0130 13:24:56.368351 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" podUID="ce7cd8b7-6707-4478-a092-50a5a608f5e0" Jan 30 13:24:56.377618 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e-shm.mount: Deactivated successfully. Jan 30 13:24:56.378437 systemd[1]: run-netns-cni\x2d7fbc0e98\x2dd6bf\x2d2790\x2da083\x2d4288f37db771.mount: Deactivated successfully. Jan 30 13:24:56.380228 containerd[1781]: time="2025-01-30T13:24:56.379978878Z" level=error msg="Failed to destroy network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.378527 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289-shm.mount: Deactivated successfully. Jan 30 13:24:56.378579 systemd[1]: run-netns-cni\x2ddf8ec9af\x2dbaf9\x2db5f2\x2d8c2c\x2da48194adab4d.mount: Deactivated successfully. Jan 30 13:24:56.378629 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611-shm.mount: Deactivated successfully. Jan 30 13:24:56.382859 containerd[1781]: time="2025-01-30T13:24:56.382100838Z" level=error msg="encountered an error cleaning up failed sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.382859 containerd[1781]: time="2025-01-30T13:24:56.382181438Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.384757 kubelet[3455]: E0130 13:24:56.383146 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.384757 kubelet[3455]: E0130 13:24:56.383197 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:56.384757 kubelet[3455]: E0130 13:24:56.383220 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:56.384427 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b-shm.mount: Deactivated successfully. Jan 30 13:24:56.385857 kubelet[3455]: E0130 13:24:56.383267 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:56.387130 containerd[1781]: time="2025-01-30T13:24:56.386711077Z" level=error msg="Failed to destroy network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.387445 containerd[1781]: time="2025-01-30T13:24:56.387407997Z" level=error msg="encountered an error cleaning up failed sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.389603 containerd[1781]: time="2025-01-30T13:24:56.387476597Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:1,} failed, error" error="failed to setup network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.389422 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203-shm.mount: Deactivated successfully. Jan 30 13:24:56.389854 kubelet[3455]: E0130 13:24:56.387679 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:56.389854 kubelet[3455]: E0130 13:24:56.389674 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:56.389854 kubelet[3455]: E0130 13:24:56.389708 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:56.390165 kubelet[3455]: E0130 13:24:56.389778 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" Jan 30 13:24:57.056221 kubelet[3455]: I0130 13:24:57.054936 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e" Jan 30 13:24:57.057479 containerd[1781]: time="2025-01-30T13:24:57.057369796Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" Jan 30 13:24:57.057858 containerd[1781]: time="2025-01-30T13:24:57.057834676Z" level=info msg="Ensure that sandbox d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e in task-service has been cleanup successfully" Jan 30 13:24:57.060973 containerd[1781]: time="2025-01-30T13:24:57.060936715Z" level=info msg="TearDown network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" successfully" Jan 30 13:24:57.061126 containerd[1781]: time="2025-01-30T13:24:57.061112555Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" returns successfully" Jan 30 13:24:57.062748 containerd[1781]: time="2025-01-30T13:24:57.062380595Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:24:57.062958 systemd[1]: run-netns-cni\x2d223db46f\x2de3f7\x2dccf8\x2d5b01\x2d5c7c4b51a810.mount: Deactivated successfully. Jan 30 13:24:57.065354 containerd[1781]: time="2025-01-30T13:24:57.065064914Z" level=info msg="TearDown network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" successfully" Jan 30 13:24:57.065354 containerd[1781]: time="2025-01-30T13:24:57.065089474Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" returns successfully" Jan 30 13:24:57.069736 kubelet[3455]: I0130 13:24:57.068798 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1" Jan 30 13:24:57.069950 containerd[1781]: time="2025-01-30T13:24:57.069351593Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:2,}" Jan 30 13:24:57.070431 containerd[1781]: time="2025-01-30T13:24:57.070379113Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" Jan 30 13:24:57.071779 containerd[1781]: time="2025-01-30T13:24:57.071123473Z" level=info msg="Ensure that sandbox a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1 in task-service has been cleanup successfully" Jan 30 13:24:57.074425 systemd[1]: run-netns-cni\x2dfaa495df\x2d93e0\x2dc4b2\x2dbca4\x2d6f3f6307db8a.mount: Deactivated successfully. Jan 30 13:24:57.075903 containerd[1781]: time="2025-01-30T13:24:57.074558512Z" level=info msg="TearDown network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" successfully" Jan 30 13:24:57.075903 containerd[1781]: time="2025-01-30T13:24:57.074582312Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" returns successfully" Jan 30 13:24:57.078711 containerd[1781]: time="2025-01-30T13:24:57.077299992Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:24:57.078978 containerd[1781]: time="2025-01-30T13:24:57.078953592Z" level=info msg="TearDown network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" successfully" Jan 30 13:24:57.079250 containerd[1781]: time="2025-01-30T13:24:57.079034552Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" returns successfully" Jan 30 13:24:57.079891 containerd[1781]: time="2025-01-30T13:24:57.079867752Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:2,}" Jan 30 13:24:57.080355 kubelet[3455]: I0130 13:24:57.080322 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051" Jan 30 13:24:57.083758 containerd[1781]: time="2025-01-30T13:24:57.083458871Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" Jan 30 13:24:57.083758 containerd[1781]: time="2025-01-30T13:24:57.083693871Z" level=info msg="Ensure that sandbox 2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051 in task-service has been cleanup successfully" Jan 30 13:24:57.084230 containerd[1781]: time="2025-01-30T13:24:57.084191151Z" level=info msg="TearDown network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" successfully" Jan 30 13:24:57.084364 containerd[1781]: time="2025-01-30T13:24:57.084345871Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" returns successfully" Jan 30 13:24:57.085856 containerd[1781]: time="2025-01-30T13:24:57.085777390Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:24:57.087290 containerd[1781]: time="2025-01-30T13:24:57.086989390Z" level=info msg="TearDown network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" successfully" Jan 30 13:24:57.087290 containerd[1781]: time="2025-01-30T13:24:57.087015470Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" returns successfully" Jan 30 13:24:57.088327 kubelet[3455]: I0130 13:24:57.088297 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b" Jan 30 13:24:57.089423 containerd[1781]: time="2025-01-30T13:24:57.089380750Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:2,}" Jan 30 13:24:57.091099 containerd[1781]: time="2025-01-30T13:24:57.091061310Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" Jan 30 13:24:57.092248 containerd[1781]: time="2025-01-30T13:24:57.091258469Z" level=info msg="Ensure that sandbox 77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b in task-service has been cleanup successfully" Jan 30 13:24:57.092248 containerd[1781]: time="2025-01-30T13:24:57.091787189Z" level=info msg="TearDown network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" successfully" Jan 30 13:24:57.092248 containerd[1781]: time="2025-01-30T13:24:57.092073509Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" returns successfully" Jan 30 13:24:57.093997 containerd[1781]: time="2025-01-30T13:24:57.093947829Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:24:57.094367 containerd[1781]: time="2025-01-30T13:24:57.094347909Z" level=info msg="TearDown network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" successfully" Jan 30 13:24:57.094668 containerd[1781]: time="2025-01-30T13:24:57.094508149Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" returns successfully" Jan 30 13:24:57.096130 containerd[1781]: time="2025-01-30T13:24:57.096102709Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:2,}" Jan 30 13:24:57.098180 kubelet[3455]: I0130 13:24:57.098036 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203" Jan 30 13:24:57.100254 containerd[1781]: time="2025-01-30T13:24:57.100065948Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" Jan 30 13:24:57.100452 containerd[1781]: time="2025-01-30T13:24:57.100255668Z" level=info msg="Ensure that sandbox 34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203 in task-service has been cleanup successfully" Jan 30 13:24:57.102358 containerd[1781]: time="2025-01-30T13:24:57.101649028Z" level=info msg="TearDown network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" successfully" Jan 30 13:24:57.102358 containerd[1781]: time="2025-01-30T13:24:57.101776308Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" returns successfully" Jan 30 13:24:57.103486 containerd[1781]: time="2025-01-30T13:24:57.103317147Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:24:57.103486 containerd[1781]: time="2025-01-30T13:24:57.103434507Z" level=info msg="TearDown network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" successfully" Jan 30 13:24:57.103486 containerd[1781]: time="2025-01-30T13:24:57.103459747Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" returns successfully" Jan 30 13:24:57.105280 containerd[1781]: time="2025-01-30T13:24:57.105151267Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:2,}" Jan 30 13:24:57.111682 kubelet[3455]: I0130 13:24:57.110894 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895" Jan 30 13:24:57.112082 containerd[1781]: time="2025-01-30T13:24:57.111934026Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" Jan 30 13:24:57.112680 containerd[1781]: time="2025-01-30T13:24:57.112320986Z" level=info msg="Ensure that sandbox 6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895 in task-service has been cleanup successfully" Jan 30 13:24:57.112680 containerd[1781]: time="2025-01-30T13:24:57.112557066Z" level=info msg="TearDown network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" successfully" Jan 30 13:24:57.112680 containerd[1781]: time="2025-01-30T13:24:57.112573586Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" returns successfully" Jan 30 13:24:57.119272 containerd[1781]: time="2025-01-30T13:24:57.117261545Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:24:57.119272 containerd[1781]: time="2025-01-30T13:24:57.117443105Z" level=info msg="TearDown network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" successfully" Jan 30 13:24:57.119272 containerd[1781]: time="2025-01-30T13:24:57.117470465Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" returns successfully" Jan 30 13:24:57.120756 containerd[1781]: time="2025-01-30T13:24:57.120695544Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:2,}" Jan 30 13:24:57.379418 systemd[1]: run-netns-cni\x2dda8d17fa\x2daeb3\x2d0e4a\x2d6679\x2d3f7935ec2c8d.mount: Deactivated successfully. Jan 30 13:24:57.379525 systemd[1]: run-netns-cni\x2dda5b194e\x2df35d\x2d28f7\x2da4cd\x2d0ee105e02c59.mount: Deactivated successfully. Jan 30 13:24:57.379575 systemd[1]: run-netns-cni\x2d99eaa06f\x2d9ce3\x2d3661\x2d49d9\x2d2a87d4d80a5a.mount: Deactivated successfully. Jan 30 13:24:57.379617 systemd[1]: run-netns-cni\x2d3fd51371\x2dc8c9\x2df2f6\x2d148c\x2daa26326613e3.mount: Deactivated successfully. Jan 30 13:24:57.394449 containerd[1781]: time="2025-01-30T13:24:57.394399375Z" level=error msg="Failed to destroy network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.400666 containerd[1781]: time="2025-01-30T13:24:57.400609933Z" level=error msg="encountered an error cleaning up failed sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.403349 containerd[1781]: time="2025-01-30T13:24:57.400694453Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.404098 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e-shm.mount: Deactivated successfully. Jan 30 13:24:57.406536 kubelet[3455]: E0130 13:24:57.406504 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.406691 kubelet[3455]: E0130 13:24:57.406667 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:57.406938 kubelet[3455]: E0130 13:24:57.406856 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:57.407278 kubelet[3455]: E0130 13:24:57.407015 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5hs8w" podUID="0b84efc9-a798-4f8e-8d4f-6bba2eb764e3" Jan 30 13:24:57.422906 containerd[1781]: time="2025-01-30T13:24:57.422854129Z" level=error msg="Failed to destroy network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.425204 containerd[1781]: time="2025-01-30T13:24:57.423361049Z" level=error msg="encountered an error cleaning up failed sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.425204 containerd[1781]: time="2025-01-30T13:24:57.423424129Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.425361 kubelet[3455]: E0130 13:24:57.424883 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.425361 kubelet[3455]: E0130 13:24:57.424941 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:57.425361 kubelet[3455]: E0130 13:24:57.424959 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:57.425933 kubelet[3455]: E0130 13:24:57.424995 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5djx9" podUID="efc446ea-7180-4d70-91d5-8919459a6efa" Jan 30 13:24:57.427397 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50-shm.mount: Deactivated successfully. Jan 30 13:24:57.436894 containerd[1781]: time="2025-01-30T13:24:57.436649767Z" level=error msg="Failed to destroy network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.439658 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39-shm.mount: Deactivated successfully. Jan 30 13:24:57.440406 containerd[1781]: time="2025-01-30T13:24:57.440122286Z" level=error msg="encountered an error cleaning up failed sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.442097 containerd[1781]: time="2025-01-30T13:24:57.442054566Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.444409 kubelet[3455]: E0130 13:24:57.443255 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.444409 kubelet[3455]: E0130 13:24:57.443313 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:57.444409 kubelet[3455]: E0130 13:24:57.443332 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:57.444610 kubelet[3455]: E0130 13:24:57.443375 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:57.494689 containerd[1781]: time="2025-01-30T13:24:57.494637836Z" level=error msg="Failed to destroy network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.497366 containerd[1781]: time="2025-01-30T13:24:57.495063916Z" level=error msg="encountered an error cleaning up failed sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.497366 containerd[1781]: time="2025-01-30T13:24:57.495180716Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.497544 kubelet[3455]: E0130 13:24:57.495871 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.497544 kubelet[3455]: E0130 13:24:57.495928 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:57.497544 kubelet[3455]: E0130 13:24:57.495966 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:57.497707 kubelet[3455]: E0130 13:24:57.497132 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" podUID="376f6955-c43d-49c5-a9d0-4891ee1dc8f6" Jan 30 13:24:57.500138 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9-shm.mount: Deactivated successfully. Jan 30 13:24:57.513455 containerd[1781]: time="2025-01-30T13:24:57.512237473Z" level=error msg="Failed to destroy network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.513561 containerd[1781]: time="2025-01-30T13:24:57.513474673Z" level=error msg="encountered an error cleaning up failed sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.514025 containerd[1781]: time="2025-01-30T13:24:57.513671153Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:2,} failed, error" error="failed to setup network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.514381 kubelet[3455]: E0130 13:24:57.514348 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.514591 kubelet[3455]: E0130 13:24:57.514567 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:57.514679 kubelet[3455]: E0130 13:24:57.514657 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:57.514822 kubelet[3455]: E0130 13:24:57.514797 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" Jan 30 13:24:57.523298 containerd[1781]: time="2025-01-30T13:24:57.523228871Z" level=error msg="Failed to destroy network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.523688 containerd[1781]: time="2025-01-30T13:24:57.523653871Z" level=error msg="encountered an error cleaning up failed sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.523814 containerd[1781]: time="2025-01-30T13:24:57.523736711Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:2,} failed, error" error="failed to setup network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.524070 kubelet[3455]: E0130 13:24:57.524027 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:57.524150 kubelet[3455]: E0130 13:24:57.524086 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:57.524150 kubelet[3455]: E0130 13:24:57.524105 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:57.524243 kubelet[3455]: E0130 13:24:57.524210 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" podUID="ce7cd8b7-6707-4478-a092-50a5a608f5e0" Jan 30 13:24:58.115211 kubelet[3455]: I0130 13:24:58.115141 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39" Jan 30 13:24:58.117053 containerd[1781]: time="2025-01-30T13:24:58.117004524Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\"" Jan 30 13:24:58.117575 containerd[1781]: time="2025-01-30T13:24:58.117482484Z" level=info msg="Ensure that sandbox 38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39 in task-service has been cleanup successfully" Jan 30 13:24:58.117736 containerd[1781]: time="2025-01-30T13:24:58.117697284Z" level=info msg="TearDown network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" successfully" Jan 30 13:24:58.117736 containerd[1781]: time="2025-01-30T13:24:58.117730484Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" returns successfully" Jan 30 13:24:58.119007 containerd[1781]: time="2025-01-30T13:24:58.118587123Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" Jan 30 13:24:58.119007 containerd[1781]: time="2025-01-30T13:24:58.118737963Z" level=info msg="TearDown network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" successfully" Jan 30 13:24:58.119007 containerd[1781]: time="2025-01-30T13:24:58.118775883Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" returns successfully" Jan 30 13:24:58.119399 containerd[1781]: time="2025-01-30T13:24:58.119361843Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:24:58.119950 containerd[1781]: time="2025-01-30T13:24:58.119569763Z" level=info msg="TearDown network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" successfully" Jan 30 13:24:58.120143 containerd[1781]: time="2025-01-30T13:24:58.120050043Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" returns successfully" Jan 30 13:24:58.122334 containerd[1781]: time="2025-01-30T13:24:58.122041483Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:3,}" Jan 30 13:24:58.123251 kubelet[3455]: I0130 13:24:58.122471 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf" Jan 30 13:24:58.124423 containerd[1781]: time="2025-01-30T13:24:58.123372363Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\"" Jan 30 13:24:58.124423 containerd[1781]: time="2025-01-30T13:24:58.123766963Z" level=info msg="Ensure that sandbox a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf in task-service has been cleanup successfully" Jan 30 13:24:58.124880 containerd[1781]: time="2025-01-30T13:24:58.124817962Z" level=info msg="TearDown network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" successfully" Jan 30 13:24:58.124948 containerd[1781]: time="2025-01-30T13:24:58.124876842Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" returns successfully" Jan 30 13:24:58.126104 containerd[1781]: time="2025-01-30T13:24:58.126064882Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" Jan 30 13:24:58.126338 containerd[1781]: time="2025-01-30T13:24:58.126292962Z" level=info msg="TearDown network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" successfully" Jan 30 13:24:58.126338 containerd[1781]: time="2025-01-30T13:24:58.126331602Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" returns successfully" Jan 30 13:24:58.127443 containerd[1781]: time="2025-01-30T13:24:58.127394602Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:24:58.128095 kubelet[3455]: I0130 13:24:58.128064 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956" Jan 30 13:24:58.128706 containerd[1781]: time="2025-01-30T13:24:58.128638482Z" level=info msg="TearDown network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" successfully" Jan 30 13:24:58.129354 containerd[1781]: time="2025-01-30T13:24:58.129317602Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" returns successfully" Jan 30 13:24:58.130073 containerd[1781]: time="2025-01-30T13:24:58.129605881Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\"" Jan 30 13:24:58.130073 containerd[1781]: time="2025-01-30T13:24:58.129977041Z" level=info msg="Ensure that sandbox d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956 in task-service has been cleanup successfully" Jan 30 13:24:58.130503 containerd[1781]: time="2025-01-30T13:24:58.130029321Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:3,}" Jan 30 13:24:58.130503 containerd[1781]: time="2025-01-30T13:24:58.130243161Z" level=info msg="TearDown network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" successfully" Jan 30 13:24:58.130503 containerd[1781]: time="2025-01-30T13:24:58.130258241Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" returns successfully" Jan 30 13:24:58.131239 containerd[1781]: time="2025-01-30T13:24:58.131212001Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" Jan 30 13:24:58.131520 containerd[1781]: time="2025-01-30T13:24:58.131471161Z" level=info msg="TearDown network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" successfully" Jan 30 13:24:58.131520 containerd[1781]: time="2025-01-30T13:24:58.131505001Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" returns successfully" Jan 30 13:24:58.131925 containerd[1781]: time="2025-01-30T13:24:58.131893241Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:24:58.133276 containerd[1781]: time="2025-01-30T13:24:58.132932841Z" level=info msg="TearDown network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" successfully" Jan 30 13:24:58.133276 containerd[1781]: time="2025-01-30T13:24:58.132962161Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" returns successfully" Jan 30 13:24:58.134769 containerd[1781]: time="2025-01-30T13:24:58.134701121Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:3,}" Jan 30 13:24:58.137364 kubelet[3455]: I0130 13:24:58.136278 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50" Jan 30 13:24:58.139047 containerd[1781]: time="2025-01-30T13:24:58.138997680Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\"" Jan 30 13:24:58.142508 containerd[1781]: time="2025-01-30T13:24:58.142031719Z" level=info msg="Ensure that sandbox 7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50 in task-service has been cleanup successfully" Jan 30 13:24:58.142508 containerd[1781]: time="2025-01-30T13:24:58.142416159Z" level=info msg="TearDown network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" successfully" Jan 30 13:24:58.142508 containerd[1781]: time="2025-01-30T13:24:58.142435479Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" returns successfully" Jan 30 13:24:58.145597 containerd[1781]: time="2025-01-30T13:24:58.145428319Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" Jan 30 13:24:58.145849 containerd[1781]: time="2025-01-30T13:24:58.145821279Z" level=info msg="TearDown network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" successfully" Jan 30 13:24:58.146184 containerd[1781]: time="2025-01-30T13:24:58.145963759Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" returns successfully" Jan 30 13:24:58.147253 containerd[1781]: time="2025-01-30T13:24:58.147086798Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:24:58.148326 containerd[1781]: time="2025-01-30T13:24:58.147615878Z" level=info msg="TearDown network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" successfully" Jan 30 13:24:58.148326 containerd[1781]: time="2025-01-30T13:24:58.147854598Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" returns successfully" Jan 30 13:24:58.148511 kubelet[3455]: I0130 13:24:58.148472 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e" Jan 30 13:24:58.149745 containerd[1781]: time="2025-01-30T13:24:58.149431838Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:3,}" Jan 30 13:24:58.151795 containerd[1781]: time="2025-01-30T13:24:58.151356638Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\"" Jan 30 13:24:58.151795 containerd[1781]: time="2025-01-30T13:24:58.151564238Z" level=info msg="Ensure that sandbox 5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e in task-service has been cleanup successfully" Jan 30 13:24:58.153201 containerd[1781]: time="2025-01-30T13:24:58.152772557Z" level=info msg="TearDown network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" successfully" Jan 30 13:24:58.153201 containerd[1781]: time="2025-01-30T13:24:58.152837837Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" returns successfully" Jan 30 13:24:58.154432 containerd[1781]: time="2025-01-30T13:24:58.154364917Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" Jan 30 13:24:58.154862 containerd[1781]: time="2025-01-30T13:24:58.154756677Z" level=info msg="TearDown network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" successfully" Jan 30 13:24:58.154862 containerd[1781]: time="2025-01-30T13:24:58.154779397Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" returns successfully" Jan 30 13:24:58.155189 containerd[1781]: time="2025-01-30T13:24:58.155154717Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:24:58.155315 containerd[1781]: time="2025-01-30T13:24:58.155243397Z" level=info msg="TearDown network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" successfully" Jan 30 13:24:58.155315 containerd[1781]: time="2025-01-30T13:24:58.155258797Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" returns successfully" Jan 30 13:24:58.156148 containerd[1781]: time="2025-01-30T13:24:58.156110717Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:3,}" Jan 30 13:24:58.156483 kubelet[3455]: I0130 13:24:58.156436 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9" Jan 30 13:24:58.157111 containerd[1781]: time="2025-01-30T13:24:58.157089837Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\"" Jan 30 13:24:58.157645 containerd[1781]: time="2025-01-30T13:24:58.157473956Z" level=info msg="Ensure that sandbox 3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9 in task-service has been cleanup successfully" Jan 30 13:24:58.157966 containerd[1781]: time="2025-01-30T13:24:58.157862236Z" level=info msg="TearDown network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" successfully" Jan 30 13:24:58.157966 containerd[1781]: time="2025-01-30T13:24:58.157882316Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" returns successfully" Jan 30 13:24:58.158277 containerd[1781]: time="2025-01-30T13:24:58.158250716Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" Jan 30 13:24:58.158389 containerd[1781]: time="2025-01-30T13:24:58.158370876Z" level=info msg="TearDown network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" successfully" Jan 30 13:24:58.158389 containerd[1781]: time="2025-01-30T13:24:58.158385796Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" returns successfully" Jan 30 13:24:58.159018 containerd[1781]: time="2025-01-30T13:24:58.158987556Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:24:58.159183 containerd[1781]: time="2025-01-30T13:24:58.159154996Z" level=info msg="TearDown network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" successfully" Jan 30 13:24:58.159183 containerd[1781]: time="2025-01-30T13:24:58.159171996Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" returns successfully" Jan 30 13:24:58.160046 containerd[1781]: time="2025-01-30T13:24:58.160002476Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:3,}" Jan 30 13:24:58.371264 systemd[1]: run-netns-cni\x2df3467a80\x2d9d56\x2d95ec\x2df5ef\x2d036538a0ca2b.mount: Deactivated successfully. Jan 30 13:24:58.372897 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956-shm.mount: Deactivated successfully. Jan 30 13:24:58.372955 systemd[1]: run-netns-cni\x2d0d96ee33\x2dbe9f\x2d93d3\x2d589d\x2d4252bb12423c.mount: Deactivated successfully. Jan 30 13:24:58.373002 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf-shm.mount: Deactivated successfully. Jan 30 13:24:58.373053 systemd[1]: run-netns-cni\x2d548ced0c\x2d4d6c\x2d1f69\x2d86f4\x2d625f12ee5aac.mount: Deactivated successfully. Jan 30 13:24:58.373096 systemd[1]: run-netns-cni\x2d9a74f558\x2deb22\x2d84c1\x2d4fa0\x2dd01754b15cc5.mount: Deactivated successfully. Jan 30 13:24:58.373544 systemd[1]: run-netns-cni\x2d15f26696\x2d022f\x2d0422\x2de062\x2da6634850b44c.mount: Deactivated successfully. Jan 30 13:24:58.373623 systemd[1]: run-netns-cni\x2dfe4c25eb\x2dc33b\x2d4a15\x2d63c2\x2d40816e7cd1f8.mount: Deactivated successfully. Jan 30 13:24:58.636386 containerd[1781]: time="2025-01-30T13:24:58.636033350Z" level=error msg="Failed to destroy network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.640821 containerd[1781]: time="2025-01-30T13:24:58.640764709Z" level=error msg="encountered an error cleaning up failed sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.641329 containerd[1781]: time="2025-01-30T13:24:58.641222469Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.641682 kubelet[3455]: E0130 13:24:58.641644 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.641830 kubelet[3455]: E0130 13:24:58.641739 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:58.641830 kubelet[3455]: E0130 13:24:58.641765 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:58.643803 kubelet[3455]: E0130 13:24:58.642005 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:58.689798 containerd[1781]: time="2025-01-30T13:24:58.689740900Z" level=error msg="Failed to destroy network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.690878 containerd[1781]: time="2025-01-30T13:24:58.690831420Z" level=error msg="encountered an error cleaning up failed sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.691895 containerd[1781]: time="2025-01-30T13:24:58.691839980Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.692632 kubelet[3455]: E0130 13:24:58.692157 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.692632 kubelet[3455]: E0130 13:24:58.692216 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:58.692632 kubelet[3455]: E0130 13:24:58.692239 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:58.692798 kubelet[3455]: E0130 13:24:58.692275 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5djx9" podUID="efc446ea-7180-4d70-91d5-8919459a6efa" Jan 30 13:24:58.699532 containerd[1781]: time="2025-01-30T13:24:58.699228338Z" level=error msg="Failed to destroy network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.701050 containerd[1781]: time="2025-01-30T13:24:58.700981338Z" level=error msg="encountered an error cleaning up failed sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.701160 containerd[1781]: time="2025-01-30T13:24:58.701079698Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.701387 kubelet[3455]: E0130 13:24:58.701267 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.701387 kubelet[3455]: E0130 13:24:58.701321 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:58.701387 kubelet[3455]: E0130 13:24:58.701342 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:58.701506 kubelet[3455]: E0130 13:24:58.701379 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" podUID="376f6955-c43d-49c5-a9d0-4891ee1dc8f6" Jan 30 13:24:58.725967 containerd[1781]: time="2025-01-30T13:24:58.725844974Z" level=error msg="Failed to destroy network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.726596 containerd[1781]: time="2025-01-30T13:24:58.726517413Z" level=error msg="encountered an error cleaning up failed sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.726787 containerd[1781]: time="2025-01-30T13:24:58.726601213Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.727144 kubelet[3455]: E0130 13:24:58.727051 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.727144 kubelet[3455]: E0130 13:24:58.727119 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:58.727144 kubelet[3455]: E0130 13:24:58.727140 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:58.727258 kubelet[3455]: E0130 13:24:58.727180 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" Jan 30 13:24:58.729758 containerd[1781]: time="2025-01-30T13:24:58.729397373Z" level=error msg="Failed to destroy network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.730503 containerd[1781]: time="2025-01-30T13:24:58.730445333Z" level=error msg="encountered an error cleaning up failed sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.730622 containerd[1781]: time="2025-01-30T13:24:58.730526693Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:3,} failed, error" error="failed to setup network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.732263 kubelet[3455]: E0130 13:24:58.732133 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.732263 kubelet[3455]: E0130 13:24:58.732205 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:58.732263 kubelet[3455]: E0130 13:24:58.732223 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:58.732409 kubelet[3455]: E0130 13:24:58.732265 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" podUID="ce7cd8b7-6707-4478-a092-50a5a608f5e0" Jan 30 13:24:58.734970 containerd[1781]: time="2025-01-30T13:24:58.734681052Z" level=error msg="Failed to destroy network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.735529 containerd[1781]: time="2025-01-30T13:24:58.735454732Z" level=error msg="encountered an error cleaning up failed sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.736055 containerd[1781]: time="2025-01-30T13:24:58.735537252Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:3,} failed, error" error="failed to setup network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.736153 kubelet[3455]: E0130 13:24:58.735858 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:58.736153 kubelet[3455]: E0130 13:24:58.735909 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:58.736153 kubelet[3455]: E0130 13:24:58.735952 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:58.736538 kubelet[3455]: E0130 13:24:58.736005 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5hs8w" podUID="0b84efc9-a798-4f8e-8d4f-6bba2eb764e3" Jan 30 13:24:59.162998 kubelet[3455]: I0130 13:24:59.162958 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e" Jan 30 13:24:59.165519 containerd[1781]: time="2025-01-30T13:24:59.165468814Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\"" Jan 30 13:24:59.165791 containerd[1781]: time="2025-01-30T13:24:59.165675014Z" level=info msg="Ensure that sandbox 00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e in task-service has been cleanup successfully" Jan 30 13:24:59.166435 containerd[1781]: time="2025-01-30T13:24:59.166249414Z" level=info msg="TearDown network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" successfully" Jan 30 13:24:59.166435 containerd[1781]: time="2025-01-30T13:24:59.166274214Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" returns successfully" Jan 30 13:24:59.172950 containerd[1781]: time="2025-01-30T13:24:59.172883693Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\"" Jan 30 13:24:59.173144 containerd[1781]: time="2025-01-30T13:24:59.173040173Z" level=info msg="TearDown network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" successfully" Jan 30 13:24:59.173144 containerd[1781]: time="2025-01-30T13:24:59.173052493Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" returns successfully" Jan 30 13:24:59.175930 containerd[1781]: time="2025-01-30T13:24:59.175890892Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" Jan 30 13:24:59.176262 containerd[1781]: time="2025-01-30T13:24:59.176204892Z" level=info msg="TearDown network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" successfully" Jan 30 13:24:59.176262 containerd[1781]: time="2025-01-30T13:24:59.176223052Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" returns successfully" Jan 30 13:24:59.178643 containerd[1781]: time="2025-01-30T13:24:59.178486012Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:24:59.180819 containerd[1781]: time="2025-01-30T13:24:59.180785691Z" level=info msg="TearDown network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" successfully" Jan 30 13:24:59.180966 containerd[1781]: time="2025-01-30T13:24:59.180939651Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" returns successfully" Jan 30 13:24:59.181217 kubelet[3455]: I0130 13:24:59.181190 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67" Jan 30 13:24:59.182875 containerd[1781]: time="2025-01-30T13:24:59.182846771Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\"" Jan 30 13:24:59.183318 containerd[1781]: time="2025-01-30T13:24:59.183270891Z" level=info msg="Ensure that sandbox 29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67 in task-service has been cleanup successfully" Jan 30 13:24:59.184141 containerd[1781]: time="2025-01-30T13:24:59.184111851Z" level=info msg="TearDown network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" successfully" Jan 30 13:24:59.184366 containerd[1781]: time="2025-01-30T13:24:59.184231811Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" returns successfully" Jan 30 13:24:59.185465 containerd[1781]: time="2025-01-30T13:24:59.185418370Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\"" Jan 30 13:24:59.186953 containerd[1781]: time="2025-01-30T13:24:59.185933890Z" level=info msg="TearDown network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" successfully" Jan 30 13:24:59.186953 containerd[1781]: time="2025-01-30T13:24:59.186816850Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" returns successfully" Jan 30 13:24:59.187989 containerd[1781]: time="2025-01-30T13:24:59.187902490Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:4,}" Jan 30 13:24:59.188960 containerd[1781]: time="2025-01-30T13:24:59.188920210Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" Jan 30 13:24:59.189035 containerd[1781]: time="2025-01-30T13:24:59.189013410Z" level=info msg="TearDown network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" successfully" Jan 30 13:24:59.189035 containerd[1781]: time="2025-01-30T13:24:59.189023890Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" returns successfully" Jan 30 13:24:59.191359 containerd[1781]: time="2025-01-30T13:24:59.191301289Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:24:59.192029 containerd[1781]: time="2025-01-30T13:24:59.191993009Z" level=info msg="TearDown network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" successfully" Jan 30 13:24:59.192194 containerd[1781]: time="2025-01-30T13:24:59.192111289Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" returns successfully" Jan 30 13:24:59.194387 containerd[1781]: time="2025-01-30T13:24:59.194060729Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:4,}" Jan 30 13:24:59.194781 kubelet[3455]: I0130 13:24:59.194757 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e" Jan 30 13:24:59.196179 containerd[1781]: time="2025-01-30T13:24:59.196111808Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\"" Jan 30 13:24:59.197615 containerd[1781]: time="2025-01-30T13:24:59.197573088Z" level=info msg="Ensure that sandbox a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e in task-service has been cleanup successfully" Jan 30 13:24:59.199007 containerd[1781]: time="2025-01-30T13:24:59.198955728Z" level=info msg="TearDown network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" successfully" Jan 30 13:24:59.199007 containerd[1781]: time="2025-01-30T13:24:59.199005528Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" returns successfully" Jan 30 13:24:59.199978 containerd[1781]: time="2025-01-30T13:24:59.199869968Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\"" Jan 30 13:24:59.200475 containerd[1781]: time="2025-01-30T13:24:59.200073568Z" level=info msg="TearDown network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" successfully" Jan 30 13:24:59.200475 containerd[1781]: time="2025-01-30T13:24:59.200092088Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" returns successfully" Jan 30 13:24:59.201405 containerd[1781]: time="2025-01-30T13:24:59.201376167Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" Jan 30 13:24:59.201607 containerd[1781]: time="2025-01-30T13:24:59.201571847Z" level=info msg="TearDown network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" successfully" Jan 30 13:24:59.201812 containerd[1781]: time="2025-01-30T13:24:59.201793847Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" returns successfully" Jan 30 13:24:59.202479 containerd[1781]: time="2025-01-30T13:24:59.202438247Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:24:59.202692 containerd[1781]: time="2025-01-30T13:24:59.202533447Z" level=info msg="TearDown network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" successfully" Jan 30 13:24:59.202692 containerd[1781]: time="2025-01-30T13:24:59.202556087Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" returns successfully" Jan 30 13:24:59.205245 kubelet[3455]: I0130 13:24:59.205212 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f" Jan 30 13:24:59.205737 containerd[1781]: time="2025-01-30T13:24:59.205563327Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:4,}" Jan 30 13:24:59.208010 containerd[1781]: time="2025-01-30T13:24:59.207785246Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\"" Jan 30 13:24:59.208477 containerd[1781]: time="2025-01-30T13:24:59.208308246Z" level=info msg="Ensure that sandbox 5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f in task-service has been cleanup successfully" Jan 30 13:24:59.208894 containerd[1781]: time="2025-01-30T13:24:59.208813166Z" level=info msg="TearDown network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" successfully" Jan 30 13:24:59.208894 containerd[1781]: time="2025-01-30T13:24:59.208835166Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" returns successfully" Jan 30 13:24:59.210517 containerd[1781]: time="2025-01-30T13:24:59.210329406Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\"" Jan 30 13:24:59.210610 containerd[1781]: time="2025-01-30T13:24:59.210542206Z" level=info msg="TearDown network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" successfully" Jan 30 13:24:59.210610 containerd[1781]: time="2025-01-30T13:24:59.210558286Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" returns successfully" Jan 30 13:24:59.211755 containerd[1781]: time="2025-01-30T13:24:59.211237446Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" Jan 30 13:24:59.211755 containerd[1781]: time="2025-01-30T13:24:59.211316446Z" level=info msg="TearDown network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" successfully" Jan 30 13:24:59.211755 containerd[1781]: time="2025-01-30T13:24:59.211325566Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" returns successfully" Jan 30 13:24:59.211974 containerd[1781]: time="2025-01-30T13:24:59.211944726Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:24:59.212133 containerd[1781]: time="2025-01-30T13:24:59.212106485Z" level=info msg="TearDown network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" successfully" Jan 30 13:24:59.212133 containerd[1781]: time="2025-01-30T13:24:59.212129765Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" returns successfully" Jan 30 13:24:59.214309 containerd[1781]: time="2025-01-30T13:24:59.214062325Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:4,}" Jan 30 13:24:59.215693 kubelet[3455]: I0130 13:24:59.215655 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2" Jan 30 13:24:59.216833 containerd[1781]: time="2025-01-30T13:24:59.216655485Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\"" Jan 30 13:24:59.217177 containerd[1781]: time="2025-01-30T13:24:59.217104405Z" level=info msg="Ensure that sandbox 3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2 in task-service has been cleanup successfully" Jan 30 13:24:59.217564 containerd[1781]: time="2025-01-30T13:24:59.217438045Z" level=info msg="TearDown network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" successfully" Jan 30 13:24:59.217564 containerd[1781]: time="2025-01-30T13:24:59.217458285Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" returns successfully" Jan 30 13:24:59.219603 containerd[1781]: time="2025-01-30T13:24:59.219342644Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\"" Jan 30 13:24:59.221813 containerd[1781]: time="2025-01-30T13:24:59.219694244Z" level=info msg="TearDown network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" successfully" Jan 30 13:24:59.221813 containerd[1781]: time="2025-01-30T13:24:59.219709724Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" returns successfully" Jan 30 13:24:59.221813 containerd[1781]: time="2025-01-30T13:24:59.220910564Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" Jan 30 13:24:59.221813 containerd[1781]: time="2025-01-30T13:24:59.221022844Z" level=info msg="TearDown network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" successfully" Jan 30 13:24:59.221813 containerd[1781]: time="2025-01-30T13:24:59.221034244Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" returns successfully" Jan 30 13:24:59.222750 containerd[1781]: time="2025-01-30T13:24:59.222471284Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:24:59.222750 containerd[1781]: time="2025-01-30T13:24:59.222551724Z" level=info msg="TearDown network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" successfully" Jan 30 13:24:59.222750 containerd[1781]: time="2025-01-30T13:24:59.222560524Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" returns successfully" Jan 30 13:24:59.224848 kubelet[3455]: I0130 13:24:59.224586 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08" Jan 30 13:24:59.225906 containerd[1781]: time="2025-01-30T13:24:59.225629363Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:4,}" Jan 30 13:24:59.227262 containerd[1781]: time="2025-01-30T13:24:59.226745203Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\"" Jan 30 13:24:59.227262 containerd[1781]: time="2025-01-30T13:24:59.226912923Z" level=info msg="Ensure that sandbox 3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08 in task-service has been cleanup successfully" Jan 30 13:24:59.228007 containerd[1781]: time="2025-01-30T13:24:59.227970803Z" level=info msg="TearDown network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" successfully" Jan 30 13:24:59.228096 containerd[1781]: time="2025-01-30T13:24:59.228014203Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" returns successfully" Jan 30 13:24:59.229218 containerd[1781]: time="2025-01-30T13:24:59.229047842Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\"" Jan 30 13:24:59.229218 containerd[1781]: time="2025-01-30T13:24:59.229146282Z" level=info msg="TearDown network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" successfully" Jan 30 13:24:59.229218 containerd[1781]: time="2025-01-30T13:24:59.229158362Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" returns successfully" Jan 30 13:24:59.229941 containerd[1781]: time="2025-01-30T13:24:59.229917602Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" Jan 30 13:24:59.230215 containerd[1781]: time="2025-01-30T13:24:59.230151082Z" level=info msg="TearDown network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" successfully" Jan 30 13:24:59.230215 containerd[1781]: time="2025-01-30T13:24:59.230167522Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" returns successfully" Jan 30 13:24:59.231350 containerd[1781]: time="2025-01-30T13:24:59.231317242Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:24:59.231473 containerd[1781]: time="2025-01-30T13:24:59.231432402Z" level=info msg="TearDown network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" successfully" Jan 30 13:24:59.231473 containerd[1781]: time="2025-01-30T13:24:59.231455762Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" returns successfully" Jan 30 13:24:59.232668 containerd[1781]: time="2025-01-30T13:24:59.232626882Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:4,}" Jan 30 13:24:59.376413 systemd[1]: run-netns-cni\x2d17aaf25e\x2d7630\x2df546\x2d06c0\x2d950a1268a889.mount: Deactivated successfully. Jan 30 13:24:59.376671 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e-shm.mount: Deactivated successfully. Jan 30 13:24:59.376905 systemd[1]: run-netns-cni\x2d18453b3b\x2d60ea\x2d82eb\x2d4129\x2d9bafd3713ed2.mount: Deactivated successfully. Jan 30 13:24:59.376963 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67-shm.mount: Deactivated successfully. Jan 30 13:24:59.409490 containerd[1781]: time="2025-01-30T13:24:59.409426130Z" level=error msg="Failed to destroy network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.414034 containerd[1781]: time="2025-01-30T13:24:59.413489129Z" level=error msg="encountered an error cleaning up failed sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.413591 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da-shm.mount: Deactivated successfully. Jan 30 13:24:59.418788 containerd[1781]: time="2025-01-30T13:24:59.418137048Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.418963 kubelet[3455]: E0130 13:24:59.418409 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.418963 kubelet[3455]: E0130 13:24:59.418462 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:59.418963 kubelet[3455]: E0130 13:24:59.418481 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:24:59.419052 kubelet[3455]: E0130 13:24:59.418533 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" podUID="ce7cd8b7-6707-4478-a092-50a5a608f5e0" Jan 30 13:24:59.503238 containerd[1781]: time="2025-01-30T13:24:59.502898553Z" level=error msg="Failed to destroy network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.505102 containerd[1781]: time="2025-01-30T13:24:59.505042792Z" level=error msg="encountered an error cleaning up failed sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.505226 containerd[1781]: time="2025-01-30T13:24:59.505127792Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.505495 kubelet[3455]: E0130 13:24:59.505335 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.505495 kubelet[3455]: E0130 13:24:59.505397 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:59.505495 kubelet[3455]: E0130 13:24:59.505417 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:24:59.505889 kubelet[3455]: E0130 13:24:59.505455 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:24:59.657826 containerd[1781]: time="2025-01-30T13:24:59.657740485Z" level=error msg="Failed to destroy network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.660119 containerd[1781]: time="2025-01-30T13:24:59.659972204Z" level=error msg="encountered an error cleaning up failed sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.660119 containerd[1781]: time="2025-01-30T13:24:59.660071124Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.660672 kubelet[3455]: E0130 13:24:59.660488 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.660672 kubelet[3455]: E0130 13:24:59.660648 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:59.660795 kubelet[3455]: E0130 13:24:59.660690 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:24:59.661015 kubelet[3455]: E0130 13:24:59.660839 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" Jan 30 13:24:59.687424 containerd[1781]: time="2025-01-30T13:24:59.687226999Z" level=error msg="Failed to destroy network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.687798 containerd[1781]: time="2025-01-30T13:24:59.687768879Z" level=error msg="encountered an error cleaning up failed sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.688017 containerd[1781]: time="2025-01-30T13:24:59.687911799Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.688253 kubelet[3455]: E0130 13:24:59.688218 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.690486 kubelet[3455]: E0130 13:24:59.688356 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:59.690486 kubelet[3455]: E0130 13:24:59.688382 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:24:59.690486 kubelet[3455]: E0130 13:24:59.688431 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5djx9" podUID="efc446ea-7180-4d70-91d5-8919459a6efa" Jan 30 13:24:59.717186 containerd[1781]: time="2025-01-30T13:24:59.717126074Z" level=error msg="Failed to destroy network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.719046 containerd[1781]: time="2025-01-30T13:24:59.718988434Z" level=error msg="encountered an error cleaning up failed sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.719181 containerd[1781]: time="2025-01-30T13:24:59.719071714Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:4,} failed, error" error="failed to setup network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.721026 kubelet[3455]: E0130 13:24:59.719296 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.721026 kubelet[3455]: E0130 13:24:59.719356 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:59.721026 kubelet[3455]: E0130 13:24:59.719377 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:24:59.721318 kubelet[3455]: E0130 13:24:59.719422 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5hs8w" podUID="0b84efc9-a798-4f8e-8d4f-6bba2eb764e3" Jan 30 13:24:59.725486 containerd[1781]: time="2025-01-30T13:24:59.725411193Z" level=error msg="Failed to destroy network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.725904 containerd[1781]: time="2025-01-30T13:24:59.725867192Z" level=error msg="encountered an error cleaning up failed sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.725965 containerd[1781]: time="2025-01-30T13:24:59.725937912Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:4,} failed, error" error="failed to setup network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.726772 kubelet[3455]: E0130 13:24:59.726169 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:24:59.726772 kubelet[3455]: E0130 13:24:59.726222 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:59.726772 kubelet[3455]: E0130 13:24:59.726247 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:24:59.726907 kubelet[3455]: E0130 13:24:59.726291 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" podUID="376f6955-c43d-49c5-a9d0-4891ee1dc8f6" Jan 30 13:25:00.232033 kubelet[3455]: I0130 13:25:00.231226 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813" Jan 30 13:25:00.233132 containerd[1781]: time="2025-01-30T13:25:00.232027741Z" level=info msg="StopPodSandbox for \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\"" Jan 30 13:25:00.233132 containerd[1781]: time="2025-01-30T13:25:00.232210981Z" level=info msg="Ensure that sandbox f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813 in task-service has been cleanup successfully" Jan 30 13:25:00.234041 containerd[1781]: time="2025-01-30T13:25:00.233927380Z" level=info msg="TearDown network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" successfully" Jan 30 13:25:00.234041 containerd[1781]: time="2025-01-30T13:25:00.233970100Z" level=info msg="StopPodSandbox for \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" returns successfully" Jan 30 13:25:00.234834 containerd[1781]: time="2025-01-30T13:25:00.234637100Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\"" Jan 30 13:25:00.234834 containerd[1781]: time="2025-01-30T13:25:00.234761820Z" level=info msg="TearDown network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" successfully" Jan 30 13:25:00.234834 containerd[1781]: time="2025-01-30T13:25:00.234772460Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" returns successfully" Jan 30 13:25:00.235738 containerd[1781]: time="2025-01-30T13:25:00.235465380Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\"" Jan 30 13:25:00.235738 containerd[1781]: time="2025-01-30T13:25:00.235557740Z" level=info msg="TearDown network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" successfully" Jan 30 13:25:00.235738 containerd[1781]: time="2025-01-30T13:25:00.235568540Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" returns successfully" Jan 30 13:25:00.236666 containerd[1781]: time="2025-01-30T13:25:00.236551500Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" Jan 30 13:25:00.236666 containerd[1781]: time="2025-01-30T13:25:00.236645180Z" level=info msg="TearDown network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" successfully" Jan 30 13:25:00.236666 containerd[1781]: time="2025-01-30T13:25:00.236656420Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" returns successfully" Jan 30 13:25:00.237334 kubelet[3455]: I0130 13:25:00.237230 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84" Jan 30 13:25:00.238488 containerd[1781]: time="2025-01-30T13:25:00.238433140Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:25:00.238566 containerd[1781]: time="2025-01-30T13:25:00.238529260Z" level=info msg="TearDown network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" successfully" Jan 30 13:25:00.238566 containerd[1781]: time="2025-01-30T13:25:00.238539740Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" returns successfully" Jan 30 13:25:00.239229 containerd[1781]: time="2025-01-30T13:25:00.238983060Z" level=info msg="StopPodSandbox for \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\"" Jan 30 13:25:00.239229 containerd[1781]: time="2025-01-30T13:25:00.239149700Z" level=info msg="Ensure that sandbox e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84 in task-service has been cleanup successfully" Jan 30 13:25:00.239497 containerd[1781]: time="2025-01-30T13:25:00.239427140Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:5,}" Jan 30 13:25:00.239826 containerd[1781]: time="2025-01-30T13:25:00.239799699Z" level=info msg="TearDown network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" successfully" Jan 30 13:25:00.239871 containerd[1781]: time="2025-01-30T13:25:00.239830459Z" level=info msg="StopPodSandbox for \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" returns successfully" Jan 30 13:25:00.240521 containerd[1781]: time="2025-01-30T13:25:00.240492539Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\"" Jan 30 13:25:00.240581 containerd[1781]: time="2025-01-30T13:25:00.240570259Z" level=info msg="TearDown network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" successfully" Jan 30 13:25:00.240606 containerd[1781]: time="2025-01-30T13:25:00.240580339Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" returns successfully" Jan 30 13:25:00.242307 containerd[1781]: time="2025-01-30T13:25:00.242271499Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\"" Jan 30 13:25:00.242394 containerd[1781]: time="2025-01-30T13:25:00.242357459Z" level=info msg="TearDown network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" successfully" Jan 30 13:25:00.242394 containerd[1781]: time="2025-01-30T13:25:00.242368499Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" returns successfully" Jan 30 13:25:00.244148 containerd[1781]: time="2025-01-30T13:25:00.243911699Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" Jan 30 13:25:00.244148 containerd[1781]: time="2025-01-30T13:25:00.244028099Z" level=info msg="TearDown network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" successfully" Jan 30 13:25:00.244148 containerd[1781]: time="2025-01-30T13:25:00.244054019Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" returns successfully" Jan 30 13:25:00.245169 containerd[1781]: time="2025-01-30T13:25:00.245108138Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:25:00.245534 containerd[1781]: time="2025-01-30T13:25:00.245220458Z" level=info msg="TearDown network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" successfully" Jan 30 13:25:00.245534 containerd[1781]: time="2025-01-30T13:25:00.245232658Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" returns successfully" Jan 30 13:25:00.247770 containerd[1781]: time="2025-01-30T13:25:00.247634898Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:5,}" Jan 30 13:25:00.249270 kubelet[3455]: I0130 13:25:00.248660 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da" Jan 30 13:25:00.249893 containerd[1781]: time="2025-01-30T13:25:00.249244178Z" level=info msg="StopPodSandbox for \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\"" Jan 30 13:25:00.249893 containerd[1781]: time="2025-01-30T13:25:00.249439458Z" level=info msg="Ensure that sandbox efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da in task-service has been cleanup successfully" Jan 30 13:25:00.253509 containerd[1781]: time="2025-01-30T13:25:00.253456257Z" level=info msg="TearDown network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" successfully" Jan 30 13:25:00.253509 containerd[1781]: time="2025-01-30T13:25:00.253506577Z" level=info msg="StopPodSandbox for \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" returns successfully" Jan 30 13:25:00.256403 containerd[1781]: time="2025-01-30T13:25:00.256351896Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\"" Jan 30 13:25:00.256532 containerd[1781]: time="2025-01-30T13:25:00.256478416Z" level=info msg="TearDown network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" successfully" Jan 30 13:25:00.256532 containerd[1781]: time="2025-01-30T13:25:00.256490416Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" returns successfully" Jan 30 13:25:00.260830 containerd[1781]: time="2025-01-30T13:25:00.260637096Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\"" Jan 30 13:25:00.260830 containerd[1781]: time="2025-01-30T13:25:00.260788776Z" level=info msg="TearDown network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" successfully" Jan 30 13:25:00.260830 containerd[1781]: time="2025-01-30T13:25:00.260801016Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" returns successfully" Jan 30 13:25:00.261591 containerd[1781]: time="2025-01-30T13:25:00.261495736Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" Jan 30 13:25:00.262473 containerd[1781]: time="2025-01-30T13:25:00.261650935Z" level=info msg="TearDown network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" successfully" Jan 30 13:25:00.262473 containerd[1781]: time="2025-01-30T13:25:00.261682015Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" returns successfully" Jan 30 13:25:00.262473 containerd[1781]: time="2025-01-30T13:25:00.262218055Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:25:00.262473 containerd[1781]: time="2025-01-30T13:25:00.262317175Z" level=info msg="TearDown network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" successfully" Jan 30 13:25:00.262473 containerd[1781]: time="2025-01-30T13:25:00.262327415Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" returns successfully" Jan 30 13:25:00.266175 containerd[1781]: time="2025-01-30T13:25:00.265538495Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:5,}" Jan 30 13:25:00.266941 kubelet[3455]: I0130 13:25:00.266921 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119" Jan 30 13:25:00.271911 containerd[1781]: time="2025-01-30T13:25:00.271810854Z" level=info msg="StopPodSandbox for \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\"" Jan 30 13:25:00.274387 containerd[1781]: time="2025-01-30T13:25:00.273537933Z" level=info msg="Ensure that sandbox 382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119 in task-service has been cleanup successfully" Jan 30 13:25:00.274387 containerd[1781]: time="2025-01-30T13:25:00.273756093Z" level=info msg="TearDown network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" successfully" Jan 30 13:25:00.274387 containerd[1781]: time="2025-01-30T13:25:00.273785133Z" level=info msg="StopPodSandbox for \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" returns successfully" Jan 30 13:25:00.275698 containerd[1781]: time="2025-01-30T13:25:00.275117053Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\"" Jan 30 13:25:00.275698 containerd[1781]: time="2025-01-30T13:25:00.275503453Z" level=info msg="TearDown network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" successfully" Jan 30 13:25:00.275698 containerd[1781]: time="2025-01-30T13:25:00.275522013Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" returns successfully" Jan 30 13:25:00.276895 containerd[1781]: time="2025-01-30T13:25:00.276529613Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\"" Jan 30 13:25:00.276895 containerd[1781]: time="2025-01-30T13:25:00.276669093Z" level=info msg="TearDown network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" successfully" Jan 30 13:25:00.276895 containerd[1781]: time="2025-01-30T13:25:00.276685853Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" returns successfully" Jan 30 13:25:00.278425 containerd[1781]: time="2025-01-30T13:25:00.278385932Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" Jan 30 13:25:00.278521 containerd[1781]: time="2025-01-30T13:25:00.278491932Z" level=info msg="TearDown network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" successfully" Jan 30 13:25:00.278521 containerd[1781]: time="2025-01-30T13:25:00.278503372Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" returns successfully" Jan 30 13:25:00.279636 kubelet[3455]: I0130 13:25:00.279614 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed" Jan 30 13:25:00.280943 containerd[1781]: time="2025-01-30T13:25:00.280617452Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:25:00.281167 containerd[1781]: time="2025-01-30T13:25:00.280945892Z" level=info msg="TearDown network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" successfully" Jan 30 13:25:00.281167 containerd[1781]: time="2025-01-30T13:25:00.280962612Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" returns successfully" Jan 30 13:25:00.282471 containerd[1781]: time="2025-01-30T13:25:00.281969692Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:5,}" Jan 30 13:25:00.283608 containerd[1781]: time="2025-01-30T13:25:00.283089932Z" level=info msg="StopPodSandbox for \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\"" Jan 30 13:25:00.284592 containerd[1781]: time="2025-01-30T13:25:00.283633012Z" level=info msg="Ensure that sandbox b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed in task-service has been cleanup successfully" Jan 30 13:25:00.284592 containerd[1781]: time="2025-01-30T13:25:00.284212891Z" level=info msg="TearDown network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" successfully" Jan 30 13:25:00.284592 containerd[1781]: time="2025-01-30T13:25:00.284349931Z" level=info msg="StopPodSandbox for \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" returns successfully" Jan 30 13:25:00.285655 containerd[1781]: time="2025-01-30T13:25:00.285619811Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\"" Jan 30 13:25:00.285760 containerd[1781]: time="2025-01-30T13:25:00.285750011Z" level=info msg="TearDown network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" successfully" Jan 30 13:25:00.285800 containerd[1781]: time="2025-01-30T13:25:00.285762931Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" returns successfully" Jan 30 13:25:00.287891 containerd[1781]: time="2025-01-30T13:25:00.287131131Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\"" Jan 30 13:25:00.287891 containerd[1781]: time="2025-01-30T13:25:00.287232931Z" level=info msg="TearDown network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" successfully" Jan 30 13:25:00.287891 containerd[1781]: time="2025-01-30T13:25:00.287243451Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" returns successfully" Jan 30 13:25:00.288481 containerd[1781]: time="2025-01-30T13:25:00.288285811Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" Jan 30 13:25:00.288481 containerd[1781]: time="2025-01-30T13:25:00.288411851Z" level=info msg="TearDown network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" successfully" Jan 30 13:25:00.288481 containerd[1781]: time="2025-01-30T13:25:00.288421331Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" returns successfully" Jan 30 13:25:00.288675 kubelet[3455]: I0130 13:25:00.288636 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636" Jan 30 13:25:00.289760 containerd[1781]: time="2025-01-30T13:25:00.289444250Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:25:00.290037 containerd[1781]: time="2025-01-30T13:25:00.289943970Z" level=info msg="TearDown network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" successfully" Jan 30 13:25:00.290037 containerd[1781]: time="2025-01-30T13:25:00.289966170Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" returns successfully" Jan 30 13:25:00.291974 containerd[1781]: time="2025-01-30T13:25:00.291474050Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:5,}" Jan 30 13:25:00.294282 containerd[1781]: time="2025-01-30T13:25:00.291563810Z" level=info msg="StopPodSandbox for \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\"" Jan 30 13:25:00.294282 containerd[1781]: time="2025-01-30T13:25:00.292987290Z" level=info msg="Ensure that sandbox 364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636 in task-service has been cleanup successfully" Jan 30 13:25:00.294645 containerd[1781]: time="2025-01-30T13:25:00.294568970Z" level=info msg="TearDown network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" successfully" Jan 30 13:25:00.295025 containerd[1781]: time="2025-01-30T13:25:00.295002569Z" level=info msg="StopPodSandbox for \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" returns successfully" Jan 30 13:25:00.296302 containerd[1781]: time="2025-01-30T13:25:00.296107089Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\"" Jan 30 13:25:00.296857 containerd[1781]: time="2025-01-30T13:25:00.296823489Z" level=info msg="TearDown network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" successfully" Jan 30 13:25:00.296857 containerd[1781]: time="2025-01-30T13:25:00.296851769Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" returns successfully" Jan 30 13:25:00.297741 containerd[1781]: time="2025-01-30T13:25:00.297633969Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\"" Jan 30 13:25:00.297969 containerd[1781]: time="2025-01-30T13:25:00.297936649Z" level=info msg="TearDown network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" successfully" Jan 30 13:25:00.297969 containerd[1781]: time="2025-01-30T13:25:00.297965569Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" returns successfully" Jan 30 13:25:00.298936 containerd[1781]: time="2025-01-30T13:25:00.298806209Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" Jan 30 13:25:00.298936 containerd[1781]: time="2025-01-30T13:25:00.298904889Z" level=info msg="TearDown network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" successfully" Jan 30 13:25:00.299132 containerd[1781]: time="2025-01-30T13:25:00.298915209Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" returns successfully" Jan 30 13:25:00.300122 containerd[1781]: time="2025-01-30T13:25:00.299889369Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:25:00.300609 containerd[1781]: time="2025-01-30T13:25:00.300580048Z" level=info msg="TearDown network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" successfully" Jan 30 13:25:00.300609 containerd[1781]: time="2025-01-30T13:25:00.300604768Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" returns successfully" Jan 30 13:25:00.302019 containerd[1781]: time="2025-01-30T13:25:00.301932088Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:5,}" Jan 30 13:25:00.371332 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed-shm.mount: Deactivated successfully. Jan 30 13:25:00.371458 systemd[1]: run-netns-cni\x2dd19dbd33\x2d2103\x2d382b\x2dbffd\x2dc5184ab53238.mount: Deactivated successfully. Jan 30 13:25:00.371509 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119-shm.mount: Deactivated successfully. Jan 30 13:25:00.371561 systemd[1]: run-netns-cni\x2d25c606e1\x2d5e07\x2d0776\x2d8624\x2d37f069fdb4fd.mount: Deactivated successfully. Jan 30 13:25:00.371613 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84-shm.mount: Deactivated successfully. Jan 30 13:25:00.371672 systemd[1]: run-netns-cni\x2d58acea69\x2d1095\x2daf18\x2dd108\x2db5a8c5017820.mount: Deactivated successfully. Jan 30 13:25:00.371752 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813-shm.mount: Deactivated successfully. Jan 30 13:25:00.371820 systemd[1]: run-netns-cni\x2d8e484a19\x2d8fb9\x2d06de\x2dd376\x2d5c9f32ea9d2f.mount: Deactivated successfully. Jan 30 13:25:00.552620 containerd[1781]: time="2025-01-30T13:25:00.552218763Z" level=error msg="Failed to destroy network for sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.557497 containerd[1781]: time="2025-01-30T13:25:00.557350202Z" level=error msg="encountered an error cleaning up failed sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.559936 containerd[1781]: time="2025-01-30T13:25:00.557442002Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.560745 kubelet[3455]: E0130 13:25:00.560164 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.560745 kubelet[3455]: E0130 13:25:00.560298 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:25:00.560745 kubelet[3455]: E0130 13:25:00.560318 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:25:00.560896 kubelet[3455]: E0130 13:25:00.560369 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:25:00.706920 containerd[1781]: time="2025-01-30T13:25:00.706863495Z" level=error msg="Failed to destroy network for sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.710902 containerd[1781]: time="2025-01-30T13:25:00.710830214Z" level=error msg="encountered an error cleaning up failed sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.711052 containerd[1781]: time="2025-01-30T13:25:00.710929414Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.711201 kubelet[3455]: E0130 13:25:00.711162 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.711272 kubelet[3455]: E0130 13:25:00.711218 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:25:00.711272 kubelet[3455]: E0130 13:25:00.711245 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:25:00.711391 kubelet[3455]: E0130 13:25:00.711285 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5djx9" podUID="efc446ea-7180-4d70-91d5-8919459a6efa" Jan 30 13:25:00.726122 containerd[1781]: time="2025-01-30T13:25:00.726067091Z" level=error msg="Failed to destroy network for sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.727632 containerd[1781]: time="2025-01-30T13:25:00.727572091Z" level=error msg="encountered an error cleaning up failed sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.727780 containerd[1781]: time="2025-01-30T13:25:00.727658931Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:5,} failed, error" error="failed to setup network for sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.728108 containerd[1781]: time="2025-01-30T13:25:00.727980851Z" level=error msg="Failed to destroy network for sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.730728 kubelet[3455]: E0130 13:25:00.730390 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.730728 kubelet[3455]: E0130 13:25:00.730552 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:25:00.730728 kubelet[3455]: E0130 13:25:00.730606 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:25:00.731190 containerd[1781]: time="2025-01-30T13:25:00.731080450Z" level=error msg="encountered an error cleaning up failed sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.731190 containerd[1781]: time="2025-01-30T13:25:00.731165290Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.731820 kubelet[3455]: E0130 13:25:00.730914 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" podUID="376f6955-c43d-49c5-a9d0-4891ee1dc8f6" Jan 30 13:25:00.733222 kubelet[3455]: E0130 13:25:00.733090 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.733222 kubelet[3455]: E0130 13:25:00.733143 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:25:00.733222 kubelet[3455]: E0130 13:25:00.733161 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:25:00.733406 kubelet[3455]: E0130 13:25:00.733200 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" Jan 30 13:25:00.740747 containerd[1781]: time="2025-01-30T13:25:00.740639809Z" level=error msg="Failed to destroy network for sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.742370 containerd[1781]: time="2025-01-30T13:25:00.742315688Z" level=error msg="encountered an error cleaning up failed sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.742507 containerd[1781]: time="2025-01-30T13:25:00.742402648Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:5,} failed, error" error="failed to setup network for sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.743212 kubelet[3455]: E0130 13:25:00.742993 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.743212 kubelet[3455]: E0130 13:25:00.743055 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:25:00.743212 kubelet[3455]: E0130 13:25:00.743078 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:25:00.743581 kubelet[3455]: E0130 13:25:00.743119 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5hs8w" podUID="0b84efc9-a798-4f8e-8d4f-6bba2eb764e3" Jan 30 13:25:00.752663 containerd[1781]: time="2025-01-30T13:25:00.752551087Z" level=error msg="Failed to destroy network for sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.753272 containerd[1781]: time="2025-01-30T13:25:00.753229566Z" level=error msg="encountered an error cleaning up failed sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.753490 containerd[1781]: time="2025-01-30T13:25:00.753466966Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:5,} failed, error" error="failed to setup network for sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.753899 kubelet[3455]: E0130 13:25:00.753850 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:00.754019 kubelet[3455]: E0130 13:25:00.753918 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:25:00.754019 kubelet[3455]: E0130 13:25:00.753945 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:25:00.754019 kubelet[3455]: E0130 13:25:00.753983 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" podUID="ce7cd8b7-6707-4478-a092-50a5a608f5e0" Jan 30 13:25:01.294445 kubelet[3455]: I0130 13:25:01.294418 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e" Jan 30 13:25:01.295876 containerd[1781]: time="2025-01-30T13:25:01.295526348Z" level=info msg="StopPodSandbox for \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\"" Jan 30 13:25:01.295876 containerd[1781]: time="2025-01-30T13:25:01.295742868Z" level=info msg="Ensure that sandbox 669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e in task-service has been cleanup successfully" Jan 30 13:25:01.296775 containerd[1781]: time="2025-01-30T13:25:01.296658628Z" level=info msg="TearDown network for sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\" successfully" Jan 30 13:25:01.296980 containerd[1781]: time="2025-01-30T13:25:01.296921588Z" level=info msg="StopPodSandbox for \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\" returns successfully" Jan 30 13:25:01.297792 containerd[1781]: time="2025-01-30T13:25:01.297614428Z" level=info msg="StopPodSandbox for \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\"" Jan 30 13:25:01.297792 containerd[1781]: time="2025-01-30T13:25:01.297695188Z" level=info msg="TearDown network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" successfully" Jan 30 13:25:01.297792 containerd[1781]: time="2025-01-30T13:25:01.297704548Z" level=info msg="StopPodSandbox for \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" returns successfully" Jan 30 13:25:01.299273 containerd[1781]: time="2025-01-30T13:25:01.299250428Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\"" Jan 30 13:25:01.299510 containerd[1781]: time="2025-01-30T13:25:01.299493268Z" level=info msg="TearDown network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" successfully" Jan 30 13:25:01.299592 containerd[1781]: time="2025-01-30T13:25:01.299577828Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" returns successfully" Jan 30 13:25:01.300598 containerd[1781]: time="2025-01-30T13:25:01.300553907Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\"" Jan 30 13:25:01.301426 containerd[1781]: time="2025-01-30T13:25:01.301404307Z" level=info msg="TearDown network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" successfully" Jan 30 13:25:01.301596 containerd[1781]: time="2025-01-30T13:25:01.301465467Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" returns successfully" Jan 30 13:25:01.302216 kubelet[3455]: I0130 13:25:01.302177 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4" Jan 30 13:25:01.303248 containerd[1781]: time="2025-01-30T13:25:01.303121387Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" Jan 30 13:25:01.303248 containerd[1781]: time="2025-01-30T13:25:01.303223547Z" level=info msg="TearDown network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" successfully" Jan 30 13:25:01.303248 containerd[1781]: time="2025-01-30T13:25:01.303234627Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" returns successfully" Jan 30 13:25:01.305676 containerd[1781]: time="2025-01-30T13:25:01.305011987Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:25:01.305676 containerd[1781]: time="2025-01-30T13:25:01.305245427Z" level=info msg="TearDown network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" successfully" Jan 30 13:25:01.305676 containerd[1781]: time="2025-01-30T13:25:01.305261907Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" returns successfully" Jan 30 13:25:01.305676 containerd[1781]: time="2025-01-30T13:25:01.305351427Z" level=info msg="StopPodSandbox for \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\"" Jan 30 13:25:01.305676 containerd[1781]: time="2025-01-30T13:25:01.305492706Z" level=info msg="Ensure that sandbox 329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4 in task-service has been cleanup successfully" Jan 30 13:25:01.306308 containerd[1781]: time="2025-01-30T13:25:01.306275666Z" level=info msg="TearDown network for sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\" successfully" Jan 30 13:25:01.306308 containerd[1781]: time="2025-01-30T13:25:01.306300066Z" level=info msg="StopPodSandbox for \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\" returns successfully" Jan 30 13:25:01.308046 containerd[1781]: time="2025-01-30T13:25:01.308009826Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:6,}" Jan 30 13:25:01.308480 containerd[1781]: time="2025-01-30T13:25:01.308450386Z" level=info msg="StopPodSandbox for \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\"" Jan 30 13:25:01.308639 containerd[1781]: time="2025-01-30T13:25:01.308623986Z" level=info msg="TearDown network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" successfully" Jan 30 13:25:01.308708 containerd[1781]: time="2025-01-30T13:25:01.308695306Z" level=info msg="StopPodSandbox for \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" returns successfully" Jan 30 13:25:01.309419 containerd[1781]: time="2025-01-30T13:25:01.309232346Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\"" Jan 30 13:25:01.309419 containerd[1781]: time="2025-01-30T13:25:01.309308066Z" level=info msg="TearDown network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" successfully" Jan 30 13:25:01.309419 containerd[1781]: time="2025-01-30T13:25:01.309318786Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" returns successfully" Jan 30 13:25:01.310132 containerd[1781]: time="2025-01-30T13:25:01.310068146Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\"" Jan 30 13:25:01.310595 containerd[1781]: time="2025-01-30T13:25:01.310428866Z" level=info msg="TearDown network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" successfully" Jan 30 13:25:01.310595 containerd[1781]: time="2025-01-30T13:25:01.310447186Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" returns successfully" Jan 30 13:25:01.312187 containerd[1781]: time="2025-01-30T13:25:01.311977745Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" Jan 30 13:25:01.312187 containerd[1781]: time="2025-01-30T13:25:01.312060705Z" level=info msg="TearDown network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" successfully" Jan 30 13:25:01.312187 containerd[1781]: time="2025-01-30T13:25:01.312072305Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" returns successfully" Jan 30 13:25:01.312453 containerd[1781]: time="2025-01-30T13:25:01.312436825Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:25:01.312604 containerd[1781]: time="2025-01-30T13:25:01.312590985Z" level=info msg="TearDown network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" successfully" Jan 30 13:25:01.312657 containerd[1781]: time="2025-01-30T13:25:01.312645705Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" returns successfully" Jan 30 13:25:01.315963 containerd[1781]: time="2025-01-30T13:25:01.315895585Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:6,}" Jan 30 13:25:01.316684 kubelet[3455]: I0130 13:25:01.316375 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b" Jan 30 13:25:01.317289 containerd[1781]: time="2025-01-30T13:25:01.317264104Z" level=info msg="StopPodSandbox for \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\"" Jan 30 13:25:01.318174 containerd[1781]: time="2025-01-30T13:25:01.318106064Z" level=info msg="Ensure that sandbox 8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b in task-service has been cleanup successfully" Jan 30 13:25:01.319093 containerd[1781]: time="2025-01-30T13:25:01.318976224Z" level=info msg="TearDown network for sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\" successfully" Jan 30 13:25:01.319093 containerd[1781]: time="2025-01-30T13:25:01.319000384Z" level=info msg="StopPodSandbox for \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\" returns successfully" Jan 30 13:25:01.320946 containerd[1781]: time="2025-01-30T13:25:01.320387744Z" level=info msg="StopPodSandbox for \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\"" Jan 30 13:25:01.320946 containerd[1781]: time="2025-01-30T13:25:01.320480304Z" level=info msg="TearDown network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" successfully" Jan 30 13:25:01.320946 containerd[1781]: time="2025-01-30T13:25:01.320490544Z" level=info msg="StopPodSandbox for \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" returns successfully" Jan 30 13:25:01.322812 containerd[1781]: time="2025-01-30T13:25:01.322671863Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\"" Jan 30 13:25:01.322812 containerd[1781]: time="2025-01-30T13:25:01.322786303Z" level=info msg="TearDown network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" successfully" Jan 30 13:25:01.323196 containerd[1781]: time="2025-01-30T13:25:01.322797863Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" returns successfully" Jan 30 13:25:01.324366 containerd[1781]: time="2025-01-30T13:25:01.324296423Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\"" Jan 30 13:25:01.324522 containerd[1781]: time="2025-01-30T13:25:01.324489063Z" level=info msg="TearDown network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" successfully" Jan 30 13:25:01.324648 containerd[1781]: time="2025-01-30T13:25:01.324504303Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" returns successfully" Jan 30 13:25:01.326099 containerd[1781]: time="2025-01-30T13:25:01.326077303Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" Jan 30 13:25:01.326512 containerd[1781]: time="2025-01-30T13:25:01.326484343Z" level=info msg="TearDown network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" successfully" Jan 30 13:25:01.326512 containerd[1781]: time="2025-01-30T13:25:01.326509623Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" returns successfully" Jan 30 13:25:01.326829 kubelet[3455]: I0130 13:25:01.326698 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad" Jan 30 13:25:01.327289 containerd[1781]: time="2025-01-30T13:25:01.327158543Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:25:01.327289 containerd[1781]: time="2025-01-30T13:25:01.327245743Z" level=info msg="TearDown network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" successfully" Jan 30 13:25:01.327289 containerd[1781]: time="2025-01-30T13:25:01.327255903Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" returns successfully" Jan 30 13:25:01.329145 containerd[1781]: time="2025-01-30T13:25:01.329014302Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:6,}" Jan 30 13:25:01.330358 containerd[1781]: time="2025-01-30T13:25:01.330049702Z" level=info msg="StopPodSandbox for \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\"" Jan 30 13:25:01.330358 containerd[1781]: time="2025-01-30T13:25:01.330210702Z" level=info msg="Ensure that sandbox 76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad in task-service has been cleanup successfully" Jan 30 13:25:01.331446 containerd[1781]: time="2025-01-30T13:25:01.330370862Z" level=info msg="TearDown network for sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\" successfully" Jan 30 13:25:01.331446 containerd[1781]: time="2025-01-30T13:25:01.330383262Z" level=info msg="StopPodSandbox for \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\" returns successfully" Jan 30 13:25:01.331705 containerd[1781]: time="2025-01-30T13:25:01.331577022Z" level=info msg="StopPodSandbox for \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\"" Jan 30 13:25:01.331705 containerd[1781]: time="2025-01-30T13:25:01.331697902Z" level=info msg="TearDown network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" successfully" Jan 30 13:25:01.331793 containerd[1781]: time="2025-01-30T13:25:01.331709702Z" level=info msg="StopPodSandbox for \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" returns successfully" Jan 30 13:25:01.332849 containerd[1781]: time="2025-01-30T13:25:01.332643862Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\"" Jan 30 13:25:01.332849 containerd[1781]: time="2025-01-30T13:25:01.332750302Z" level=info msg="TearDown network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" successfully" Jan 30 13:25:01.332849 containerd[1781]: time="2025-01-30T13:25:01.332763462Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" returns successfully" Jan 30 13:25:01.335085 containerd[1781]: time="2025-01-30T13:25:01.334923141Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\"" Jan 30 13:25:01.335412 containerd[1781]: time="2025-01-30T13:25:01.335375461Z" level=info msg="TearDown network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" successfully" Jan 30 13:25:01.335498 containerd[1781]: time="2025-01-30T13:25:01.335401581Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" returns successfully" Jan 30 13:25:01.336193 containerd[1781]: time="2025-01-30T13:25:01.336141301Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" Jan 30 13:25:01.336367 containerd[1781]: time="2025-01-30T13:25:01.336276261Z" level=info msg="TearDown network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" successfully" Jan 30 13:25:01.336367 containerd[1781]: time="2025-01-30T13:25:01.336302621Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" returns successfully" Jan 30 13:25:01.336621 kubelet[3455]: I0130 13:25:01.336592 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46" Jan 30 13:25:01.337597 containerd[1781]: time="2025-01-30T13:25:01.337184661Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:25:01.338488 containerd[1781]: time="2025-01-30T13:25:01.338013221Z" level=info msg="TearDown network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" successfully" Jan 30 13:25:01.338686 containerd[1781]: time="2025-01-30T13:25:01.338600221Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" returns successfully" Jan 30 13:25:01.339571 containerd[1781]: time="2025-01-30T13:25:01.338540541Z" level=info msg="StopPodSandbox for \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\"" Jan 30 13:25:01.339638 containerd[1781]: time="2025-01-30T13:25:01.339599380Z" level=info msg="Ensure that sandbox edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46 in task-service has been cleanup successfully" Jan 30 13:25:01.340845 containerd[1781]: time="2025-01-30T13:25:01.340753220Z" level=info msg="TearDown network for sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\" successfully" Jan 30 13:25:01.340845 containerd[1781]: time="2025-01-30T13:25:01.340782380Z" level=info msg="StopPodSandbox for \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\" returns successfully" Jan 30 13:25:01.341526 containerd[1781]: time="2025-01-30T13:25:01.341448500Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:6,}" Jan 30 13:25:01.342466 containerd[1781]: time="2025-01-30T13:25:01.341501460Z" level=info msg="StopPodSandbox for \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\"" Jan 30 13:25:01.342466 containerd[1781]: time="2025-01-30T13:25:01.341954820Z" level=info msg="TearDown network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" successfully" Jan 30 13:25:01.342466 containerd[1781]: time="2025-01-30T13:25:01.342000420Z" level=info msg="StopPodSandbox for \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" returns successfully" Jan 30 13:25:01.343142 containerd[1781]: time="2025-01-30T13:25:01.343108540Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\"" Jan 30 13:25:01.343206 containerd[1781]: time="2025-01-30T13:25:01.343196100Z" level=info msg="TearDown network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" successfully" Jan 30 13:25:01.343231 containerd[1781]: time="2025-01-30T13:25:01.343207620Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" returns successfully" Jan 30 13:25:01.344293 containerd[1781]: time="2025-01-30T13:25:01.344253459Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\"" Jan 30 13:25:01.345179 containerd[1781]: time="2025-01-30T13:25:01.344375099Z" level=info msg="TearDown network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" successfully" Jan 30 13:25:01.345179 containerd[1781]: time="2025-01-30T13:25:01.344414179Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" returns successfully" Jan 30 13:25:01.345179 containerd[1781]: time="2025-01-30T13:25:01.345096619Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" Jan 30 13:25:01.345289 containerd[1781]: time="2025-01-30T13:25:01.345193179Z" level=info msg="TearDown network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" successfully" Jan 30 13:25:01.345289 containerd[1781]: time="2025-01-30T13:25:01.345204419Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" returns successfully" Jan 30 13:25:01.345660 containerd[1781]: time="2025-01-30T13:25:01.345638179Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:25:01.345892 containerd[1781]: time="2025-01-30T13:25:01.345872459Z" level=info msg="TearDown network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" successfully" Jan 30 13:25:01.345968 containerd[1781]: time="2025-01-30T13:25:01.345954699Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" returns successfully" Jan 30 13:25:01.346273 kubelet[3455]: I0130 13:25:01.346250 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438" Jan 30 13:25:01.348359 containerd[1781]: time="2025-01-30T13:25:01.348136539Z" level=info msg="StopPodSandbox for \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\"" Jan 30 13:25:01.349561 containerd[1781]: time="2025-01-30T13:25:01.349449179Z" level=info msg="Ensure that sandbox c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438 in task-service has been cleanup successfully" Jan 30 13:25:01.351045 containerd[1781]: time="2025-01-30T13:25:01.350934578Z" level=info msg="TearDown network for sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\" successfully" Jan 30 13:25:01.351045 containerd[1781]: time="2025-01-30T13:25:01.350962338Z" level=info msg="StopPodSandbox for \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\" returns successfully" Jan 30 13:25:01.351366 containerd[1781]: time="2025-01-30T13:25:01.348140939Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:6,}" Jan 30 13:25:01.351704 containerd[1781]: time="2025-01-30T13:25:01.351652498Z" level=info msg="StopPodSandbox for \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\"" Jan 30 13:25:01.351815 containerd[1781]: time="2025-01-30T13:25:01.351791258Z" level=info msg="TearDown network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" successfully" Jan 30 13:25:01.351815 containerd[1781]: time="2025-01-30T13:25:01.351809618Z" level=info msg="StopPodSandbox for \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" returns successfully" Jan 30 13:25:01.352345 containerd[1781]: time="2025-01-30T13:25:01.352195138Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\"" Jan 30 13:25:01.352345 containerd[1781]: time="2025-01-30T13:25:01.352285138Z" level=info msg="TearDown network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" successfully" Jan 30 13:25:01.352345 containerd[1781]: time="2025-01-30T13:25:01.352295538Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" returns successfully" Jan 30 13:25:01.352784 containerd[1781]: time="2025-01-30T13:25:01.352645658Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\"" Jan 30 13:25:01.353277 containerd[1781]: time="2025-01-30T13:25:01.352786178Z" level=info msg="TearDown network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" successfully" Jan 30 13:25:01.353277 containerd[1781]: time="2025-01-30T13:25:01.352798698Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" returns successfully" Jan 30 13:25:01.353606 containerd[1781]: time="2025-01-30T13:25:01.353585218Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" Jan 30 13:25:01.353839 containerd[1781]: time="2025-01-30T13:25:01.353762218Z" level=info msg="TearDown network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" successfully" Jan 30 13:25:01.353839 containerd[1781]: time="2025-01-30T13:25:01.353779258Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" returns successfully" Jan 30 13:25:01.354176 containerd[1781]: time="2025-01-30T13:25:01.354157178Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:25:01.354426 containerd[1781]: time="2025-01-30T13:25:01.354349538Z" level=info msg="TearDown network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" successfully" Jan 30 13:25:01.354426 containerd[1781]: time="2025-01-30T13:25:01.354365858Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" returns successfully" Jan 30 13:25:01.354828 containerd[1781]: time="2025-01-30T13:25:01.354808618Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:6,}" Jan 30 13:25:01.370682 systemd[1]: run-netns-cni\x2d7afe5ac8\x2d916b\x2db49e\x2d90eb\x2d32e7d2d81361.mount: Deactivated successfully. Jan 30 13:25:01.370807 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e-shm.mount: Deactivated successfully. Jan 30 13:25:01.370890 systemd[1]: run-netns-cni\x2d04742618\x2d7fc0\x2d3fe1\x2d447a\x2d9005336e67da.mount: Deactivated successfully. Jan 30 13:25:01.370939 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b-shm.mount: Deactivated successfully. Jan 30 13:25:01.695579 containerd[1781]: time="2025-01-30T13:25:01.695420436Z" level=error msg="Failed to destroy network for sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.696848 containerd[1781]: time="2025-01-30T13:25:01.696161396Z" level=error msg="encountered an error cleaning up failed sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.696848 containerd[1781]: time="2025-01-30T13:25:01.696250236Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:6,} failed, error" error="failed to setup network for sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.703438 kubelet[3455]: E0130 13:25:01.703382 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.703576 kubelet[3455]: E0130 13:25:01.703452 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:25:01.703576 kubelet[3455]: E0130 13:25:01.703474 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" Jan 30 13:25:01.703576 kubelet[3455]: E0130 13:25:01.703514 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-kube-controllers-6dd45f4c57-58wrn_calico-system(aad42f51-7748-4be3-8b1d-97665b3346fc)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" Jan 30 13:25:01.779357 containerd[1781]: time="2025-01-30T13:25:01.779310541Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:01.786569 containerd[1781]: time="2025-01-30T13:25:01.786324179Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node:v3.29.1: active requests=0, bytes read=137671762" Jan 30 13:25:01.792464 containerd[1781]: time="2025-01-30T13:25:01.792164818Z" level=error msg="Failed to destroy network for sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.793556 containerd[1781]: time="2025-01-30T13:25:01.793393898Z" level=error msg="encountered an error cleaning up failed sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.793556 containerd[1781]: time="2025-01-30T13:25:01.793466978Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:6,} failed, error" error="failed to setup network for sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.794038 kubelet[3455]: E0130 13:25:01.793875 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.794038 kubelet[3455]: E0130 13:25:01.793935 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:25:01.796970 containerd[1781]: time="2025-01-30T13:25:01.795262698Z" level=info msg="ImageCreate event name:\"sha256:680b8c280812d12c035ca9f0deedea7c761afe0f1cc65109ea2f96bf63801758\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:01.797071 kubelet[3455]: E0130 13:25:01.795804 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" Jan 30 13:25:01.797071 kubelet[3455]: E0130 13:25:01.795920 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-jmpf2_calico-apiserver(ce7cd8b7-6707-4478-a092-50a5a608f5e0)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" podUID="ce7cd8b7-6707-4478-a092-50a5a608f5e0" Jan 30 13:25:01.806871 containerd[1781]: time="2025-01-30T13:25:01.806828976Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:01.810799 containerd[1781]: time="2025-01-30T13:25:01.810431455Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node:v3.29.1\" with image id \"sha256:680b8c280812d12c035ca9f0deedea7c761afe0f1cc65109ea2f96bf63801758\", repo tag \"ghcr.io/flatcar/calico/node:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node@sha256:99c3917516efe1f807a0cfdf2d14b628b7c5cc6bd8a9ee5a253154f31756bea1\", size \"137671624\" in 5.80104927s" Jan 30 13:25:01.811141 containerd[1781]: time="2025-01-30T13:25:01.811116495Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node:v3.29.1\" returns image reference \"sha256:680b8c280812d12c035ca9f0deedea7c761afe0f1cc65109ea2f96bf63801758\"" Jan 30 13:25:01.828098 containerd[1781]: time="2025-01-30T13:25:01.828057732Z" level=info msg="CreateContainer within sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jan 30 13:25:01.874153 containerd[1781]: time="2025-01-30T13:25:01.874105884Z" level=error msg="Failed to destroy network for sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.874908 containerd[1781]: time="2025-01-30T13:25:01.874856043Z" level=error msg="encountered an error cleaning up failed sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.875097 containerd[1781]: time="2025-01-30T13:25:01.875074363Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:6,} failed, error" error="failed to setup network for sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.875396 kubelet[3455]: E0130 13:25:01.875355 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.875497 kubelet[3455]: E0130 13:25:01.875416 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:25:01.875497 kubelet[3455]: E0130 13:25:01.875435 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-system/csi-node-driver-42f6k" Jan 30 13:25:01.875497 kubelet[3455]: E0130 13:25:01.875476 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"csi-node-driver-42f6k_calico-system(17a70524-69f8-4a61-9912-a962709a69cd)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-system/csi-node-driver-42f6k" podUID="17a70524-69f8-4a61-9912-a962709a69cd" Jan 30 13:25:01.883648 containerd[1781]: time="2025-01-30T13:25:01.883426602Z" level=info msg="CreateContainer within sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\"" Jan 30 13:25:01.885471 containerd[1781]: time="2025-01-30T13:25:01.884319762Z" level=info msg="StartContainer for \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\"" Jan 30 13:25:01.895321 containerd[1781]: time="2025-01-30T13:25:01.895192720Z" level=error msg="Failed to destroy network for sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.896703 containerd[1781]: time="2025-01-30T13:25:01.896523919Z" level=error msg="encountered an error cleaning up failed sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.896703 containerd[1781]: time="2025-01-30T13:25:01.896601839Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:6,} failed, error" error="failed to setup network for sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.897749 kubelet[3455]: E0130 13:25:01.896843 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.897749 kubelet[3455]: E0130 13:25:01.896904 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:25:01.897749 kubelet[3455]: E0130 13:25:01.896926 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5djx9" Jan 30 13:25:01.897898 kubelet[3455]: E0130 13:25:01.896962 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5djx9_kube-system(efc446ea-7180-4d70-91d5-8919459a6efa)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5djx9" podUID="efc446ea-7180-4d70-91d5-8919459a6efa" Jan 30 13:25:01.900218 containerd[1781]: time="2025-01-30T13:25:01.899914759Z" level=error msg="Failed to destroy network for sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.900313 containerd[1781]: time="2025-01-30T13:25:01.900260839Z" level=error msg="encountered an error cleaning up failed sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.900347 containerd[1781]: time="2025-01-30T13:25:01.900323439Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:6,} failed, error" error="failed to setup network for sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.900564 kubelet[3455]: E0130 13:25:01.900525 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.900644 kubelet[3455]: E0130 13:25:01.900582 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:25:01.900644 kubelet[3455]: E0130 13:25:01.900601 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="kube-system/coredns-7db6d8ff4d-5hs8w" Jan 30 13:25:01.900708 kubelet[3455]: E0130 13:25:01.900642 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"coredns-7db6d8ff4d-5hs8w_kube-system(0b84efc9-a798-4f8e-8d4f-6bba2eb764e3)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="kube-system/coredns-7db6d8ff4d-5hs8w" podUID="0b84efc9-a798-4f8e-8d4f-6bba2eb764e3" Jan 30 13:25:01.907741 containerd[1781]: time="2025-01-30T13:25:01.907665357Z" level=error msg="Failed to destroy network for sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\"" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.908400 containerd[1781]: time="2025-01-30T13:25:01.908338157Z" level=error msg="encountered an error cleaning up failed sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\", marking sandbox state as SANDBOX_UNKNOWN" error="plugin type=\"calico\" failed (delete): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.908508 containerd[1781]: time="2025-01-30T13:25:01.908428517Z" level=error msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:6,} failed, error" error="failed to setup network for sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.908836 kubelet[3455]: E0130 13:25:01.908781 3455 remote_runtime.go:193] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" Jan 30 13:25:01.908925 kubelet[3455]: E0130 13:25:01.908854 3455 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:25:01.908925 kubelet[3455]: E0130 13:25:01.908872 3455 kuberuntime_manager.go:1166] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to setup network for sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\": plugin type=\"calico\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" Jan 30 13:25:01.908980 kubelet[3455]: E0130 13:25:01.908935 3455 pod_workers.go:1298] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"calico-apiserver-c978d777-x5x6w_calico-apiserver(376f6955-c43d-49c5-a9d0-4891ee1dc8f6)\\\": rpc error: code = Unknown desc = failed to setup network for sandbox \\\"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\\\": plugin type=\\\"calico\\\" failed (add): stat /var/lib/calico/nodename: no such file or directory: check that the calico/node container is running and has mounted /var/lib/calico/\"" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" podUID="376f6955-c43d-49c5-a9d0-4891ee1dc8f6" Jan 30 13:25:01.926925 systemd[1]: Started cri-containerd-faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca.scope - libcontainer container faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca. Jan 30 13:25:01.969872 containerd[1781]: time="2025-01-30T13:25:01.969785986Z" level=info msg="StartContainer for \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\" returns successfully" Jan 30 13:25:02.247537 kernel: wireguard: WireGuard 1.0.0 loaded. See www.wireguard.com for information. Jan 30 13:25:02.247695 kernel: wireguard: Copyright (C) 2015-2019 Jason A. Donenfeld . All Rights Reserved. Jan 30 13:25:02.359966 kubelet[3455]: I0130 13:25:02.359260 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b" Jan 30 13:25:02.360998 containerd[1781]: time="2025-01-30T13:25:02.360547635Z" level=info msg="StopPodSandbox for \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\"" Jan 30 13:25:02.360998 containerd[1781]: time="2025-01-30T13:25:02.360768635Z" level=info msg="Ensure that sandbox d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b in task-service has been cleanup successfully" Jan 30 13:25:02.361863 containerd[1781]: time="2025-01-30T13:25:02.361747795Z" level=info msg="TearDown network for sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\" successfully" Jan 30 13:25:02.361863 containerd[1781]: time="2025-01-30T13:25:02.361784115Z" level=info msg="StopPodSandbox for \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\" returns successfully" Jan 30 13:25:02.363161 containerd[1781]: time="2025-01-30T13:25:02.362412355Z" level=info msg="StopPodSandbox for \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\"" Jan 30 13:25:02.363161 containerd[1781]: time="2025-01-30T13:25:02.362512515Z" level=info msg="TearDown network for sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\" successfully" Jan 30 13:25:02.363161 containerd[1781]: time="2025-01-30T13:25:02.362522395Z" level=info msg="StopPodSandbox for \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\" returns successfully" Jan 30 13:25:02.363758 containerd[1781]: time="2025-01-30T13:25:02.363696715Z" level=info msg="StopPodSandbox for \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\"" Jan 30 13:25:02.364859 containerd[1781]: time="2025-01-30T13:25:02.364798435Z" level=info msg="TearDown network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" successfully" Jan 30 13:25:02.364859 containerd[1781]: time="2025-01-30T13:25:02.364825795Z" level=info msg="StopPodSandbox for \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" returns successfully" Jan 30 13:25:02.366497 containerd[1781]: time="2025-01-30T13:25:02.366191114Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\"" Jan 30 13:25:02.366497 containerd[1781]: time="2025-01-30T13:25:02.366307114Z" level=info msg="TearDown network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" successfully" Jan 30 13:25:02.366497 containerd[1781]: time="2025-01-30T13:25:02.366317154Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" returns successfully" Jan 30 13:25:02.367851 containerd[1781]: time="2025-01-30T13:25:02.367564794Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\"" Jan 30 13:25:02.367851 containerd[1781]: time="2025-01-30T13:25:02.367658554Z" level=info msg="TearDown network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" successfully" Jan 30 13:25:02.367851 containerd[1781]: time="2025-01-30T13:25:02.367670954Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" returns successfully" Jan 30 13:25:02.372148 containerd[1781]: time="2025-01-30T13:25:02.371769753Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" Jan 30 13:25:02.372148 containerd[1781]: time="2025-01-30T13:25:02.371893633Z" level=info msg="TearDown network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" successfully" Jan 30 13:25:02.372148 containerd[1781]: time="2025-01-30T13:25:02.371903993Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" returns successfully" Jan 30 13:25:02.379045 containerd[1781]: time="2025-01-30T13:25:02.376828953Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:25:02.383807 containerd[1781]: time="2025-01-30T13:25:02.381827272Z" level=info msg="TearDown network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" successfully" Jan 30 13:25:02.383807 containerd[1781]: time="2025-01-30T13:25:02.383703071Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" returns successfully" Jan 30 13:25:02.387587 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4-shm.mount: Deactivated successfully. Jan 30 13:25:02.387695 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e-shm.mount: Deactivated successfully. Jan 30 13:25:02.388978 kubelet[3455]: I0130 13:25:02.387941 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e" Jan 30 13:25:02.388845 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4284319011.mount: Deactivated successfully. Jan 30 13:25:02.396514 kubelet[3455]: I0130 13:25:02.396440 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-f6k44" podStartSLOduration=2.648200798 podStartE2EDuration="18.380049112s" podCreationTimestamp="2025-01-30 13:24:44 +0000 UTC" firstStartedPulling="2025-01-30 13:24:46.08233098 +0000 UTC m=+23.311927967" lastFinishedPulling="2025-01-30 13:25:01.814179294 +0000 UTC m=+39.043776281" observedRunningTime="2025-01-30 13:25:02.377205872 +0000 UTC m=+39.606802859" watchObservedRunningTime="2025-01-30 13:25:02.380049112 +0000 UTC m=+39.609646099" Jan 30 13:25:02.398286 containerd[1781]: time="2025-01-30T13:25:02.398211549Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:7,}" Jan 30 13:25:02.401424 containerd[1781]: time="2025-01-30T13:25:02.401258868Z" level=info msg="StopPodSandbox for \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\"" Jan 30 13:25:02.401975 containerd[1781]: time="2025-01-30T13:25:02.401850228Z" level=info msg="Ensure that sandbox 327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e in task-service has been cleanup successfully" Jan 30 13:25:02.407248 containerd[1781]: time="2025-01-30T13:25:02.407158987Z" level=info msg="TearDown network for sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\" successfully" Jan 30 13:25:02.407248 containerd[1781]: time="2025-01-30T13:25:02.407197987Z" level=info msg="StopPodSandbox for \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\" returns successfully" Jan 30 13:25:02.408406 systemd[1]: run-netns-cni\x2d719af472\x2db70b\x2d9fdd\x2de574\x2d903d720cc652.mount: Deactivated successfully. Jan 30 13:25:02.410286 containerd[1781]: time="2025-01-30T13:25:02.410105267Z" level=info msg="StopPodSandbox for \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\"" Jan 30 13:25:02.410286 containerd[1781]: time="2025-01-30T13:25:02.410212026Z" level=info msg="TearDown network for sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\" successfully" Jan 30 13:25:02.410286 containerd[1781]: time="2025-01-30T13:25:02.410223426Z" level=info msg="StopPodSandbox for \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\" returns successfully" Jan 30 13:25:02.410641 containerd[1781]: time="2025-01-30T13:25:02.410619426Z" level=info msg="StopPodSandbox for \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\"" Jan 30 13:25:02.410890 containerd[1781]: time="2025-01-30T13:25:02.410873306Z" level=info msg="TearDown network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" successfully" Jan 30 13:25:02.410966 containerd[1781]: time="2025-01-30T13:25:02.410952906Z" level=info msg="StopPodSandbox for \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" returns successfully" Jan 30 13:25:02.413228 containerd[1781]: time="2025-01-30T13:25:02.413189226Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\"" Jan 30 13:25:02.413852 containerd[1781]: time="2025-01-30T13:25:02.413802106Z" level=info msg="TearDown network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" successfully" Jan 30 13:25:02.413852 containerd[1781]: time="2025-01-30T13:25:02.413821066Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" returns successfully" Jan 30 13:25:02.415055 containerd[1781]: time="2025-01-30T13:25:02.415035266Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\"" Jan 30 13:25:02.415268 containerd[1781]: time="2025-01-30T13:25:02.415252666Z" level=info msg="TearDown network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" successfully" Jan 30 13:25:02.415354 containerd[1781]: time="2025-01-30T13:25:02.415340506Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" returns successfully" Jan 30 13:25:02.416276 kubelet[3455]: I0130 13:25:02.416253 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4" Jan 30 13:25:02.416569 containerd[1781]: time="2025-01-30T13:25:02.416438985Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" Jan 30 13:25:02.416569 containerd[1781]: time="2025-01-30T13:25:02.416515785Z" level=info msg="TearDown network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" successfully" Jan 30 13:25:02.416569 containerd[1781]: time="2025-01-30T13:25:02.416524865Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" returns successfully" Jan 30 13:25:02.417663 containerd[1781]: time="2025-01-30T13:25:02.417494785Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:25:02.417663 containerd[1781]: time="2025-01-30T13:25:02.417600545Z" level=info msg="TearDown network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" successfully" Jan 30 13:25:02.417663 containerd[1781]: time="2025-01-30T13:25:02.417611105Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" returns successfully" Jan 30 13:25:02.419200 containerd[1781]: time="2025-01-30T13:25:02.418882025Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:7,}" Jan 30 13:25:02.419650 containerd[1781]: time="2025-01-30T13:25:02.419629345Z" level=info msg="StopPodSandbox for \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\"" Jan 30 13:25:02.420787 containerd[1781]: time="2025-01-30T13:25:02.420703705Z" level=info msg="Ensure that sandbox 73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4 in task-service has been cleanup successfully" Jan 30 13:25:02.421445 containerd[1781]: time="2025-01-30T13:25:02.421423704Z" level=info msg="TearDown network for sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\" successfully" Jan 30 13:25:02.421551 containerd[1781]: time="2025-01-30T13:25:02.421535304Z" level=info msg="StopPodSandbox for \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\" returns successfully" Jan 30 13:25:02.426466 containerd[1781]: time="2025-01-30T13:25:02.426414904Z" level=info msg="StopPodSandbox for \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\"" Jan 30 13:25:02.427373 systemd[1]: run-netns-cni\x2d76bdc604\x2dbb95\x2d7d77\x2dbf22\x2d82d5395df1fe.mount: Deactivated successfully. Jan 30 13:25:02.433592 containerd[1781]: time="2025-01-30T13:25:02.433164582Z" level=info msg="TearDown network for sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\" successfully" Jan 30 13:25:02.433592 containerd[1781]: time="2025-01-30T13:25:02.433196702Z" level=info msg="StopPodSandbox for \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\" returns successfully" Jan 30 13:25:02.434511 containerd[1781]: time="2025-01-30T13:25:02.434477142Z" level=info msg="StopPodSandbox for \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\"" Jan 30 13:25:02.434620 containerd[1781]: time="2025-01-30T13:25:02.434583382Z" level=info msg="TearDown network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" successfully" Jan 30 13:25:02.434620 containerd[1781]: time="2025-01-30T13:25:02.434596702Z" level=info msg="StopPodSandbox for \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" returns successfully" Jan 30 13:25:02.436511 containerd[1781]: time="2025-01-30T13:25:02.436483462Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\"" Jan 30 13:25:02.437022 kubelet[3455]: I0130 13:25:02.436989 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494" Jan 30 13:25:02.439187 containerd[1781]: time="2025-01-30T13:25:02.438840941Z" level=info msg="TearDown network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" successfully" Jan 30 13:25:02.439187 containerd[1781]: time="2025-01-30T13:25:02.438872461Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" returns successfully" Jan 30 13:25:02.440400 containerd[1781]: time="2025-01-30T13:25:02.440361301Z" level=info msg="StopPodSandbox for \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\"" Jan 30 13:25:02.441118 containerd[1781]: time="2025-01-30T13:25:02.441081021Z" level=info msg="Ensure that sandbox 3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494 in task-service has been cleanup successfully" Jan 30 13:25:02.442004 containerd[1781]: time="2025-01-30T13:25:02.441982861Z" level=info msg="TearDown network for sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\" successfully" Jan 30 13:25:02.442494 containerd[1781]: time="2025-01-30T13:25:02.442413781Z" level=info msg="StopPodSandbox for \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\" returns successfully" Jan 30 13:25:02.446749 systemd[1]: run-netns-cni\x2dc36b8fed\x2d3a08\x2dabf6\x2d872e\x2dede51820b1a7.mount: Deactivated successfully. Jan 30 13:25:02.449829 containerd[1781]: time="2025-01-30T13:25:02.449785499Z" level=info msg="StopPodSandbox for \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\"" Jan 30 13:25:02.450045 containerd[1781]: time="2025-01-30T13:25:02.450016899Z" level=info msg="TearDown network for sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\" successfully" Jan 30 13:25:02.450045 containerd[1781]: time="2025-01-30T13:25:02.450040419Z" level=info msg="StopPodSandbox for \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\" returns successfully" Jan 30 13:25:02.452147 containerd[1781]: time="2025-01-30T13:25:02.452119619Z" level=info msg="StopPodSandbox for \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\"" Jan 30 13:25:02.452473 containerd[1781]: time="2025-01-30T13:25:02.452411659Z" level=info msg="TearDown network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" successfully" Jan 30 13:25:02.452473 containerd[1781]: time="2025-01-30T13:25:02.452454379Z" level=info msg="StopPodSandbox for \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" returns successfully" Jan 30 13:25:02.456865 containerd[1781]: time="2025-01-30T13:25:02.456025778Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\"" Jan 30 13:25:02.457548 containerd[1781]: time="2025-01-30T13:25:02.457156458Z" level=info msg="TearDown network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" successfully" Jan 30 13:25:02.457548 containerd[1781]: time="2025-01-30T13:25:02.457180218Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" returns successfully" Jan 30 13:25:02.457920 containerd[1781]: time="2025-01-30T13:25:02.457896858Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\"" Jan 30 13:25:02.458086 containerd[1781]: time="2025-01-30T13:25:02.458070938Z" level=info msg="TearDown network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" successfully" Jan 30 13:25:02.458168 containerd[1781]: time="2025-01-30T13:25:02.458155618Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" returns successfully" Jan 30 13:25:02.459075 containerd[1781]: time="2025-01-30T13:25:02.459048178Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" Jan 30 13:25:02.459272 containerd[1781]: time="2025-01-30T13:25:02.459255818Z" level=info msg="TearDown network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" successfully" Jan 30 13:25:02.459408 containerd[1781]: time="2025-01-30T13:25:02.459338378Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" returns successfully" Jan 30 13:25:02.459621 kubelet[3455]: I0130 13:25:02.459602 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53" Jan 30 13:25:02.460090 containerd[1781]: time="2025-01-30T13:25:02.459960057Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:25:02.460090 containerd[1781]: time="2025-01-30T13:25:02.460034177Z" level=info msg="TearDown network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" successfully" Jan 30 13:25:02.460090 containerd[1781]: time="2025-01-30T13:25:02.460043657Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" returns successfully" Jan 30 13:25:02.461002 containerd[1781]: time="2025-01-30T13:25:02.460979937Z" level=info msg="StopPodSandbox for \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\"" Jan 30 13:25:02.461346 containerd[1781]: time="2025-01-30T13:25:02.461308937Z" level=info msg="Ensure that sandbox fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53 in task-service has been cleanup successfully" Jan 30 13:25:02.462610 containerd[1781]: time="2025-01-30T13:25:02.462383537Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:7,}" Jan 30 13:25:02.463393 containerd[1781]: time="2025-01-30T13:25:02.463368977Z" level=info msg="TearDown network for sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\" successfully" Jan 30 13:25:02.463573 containerd[1781]: time="2025-01-30T13:25:02.463515777Z" level=info msg="StopPodSandbox for \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\" returns successfully" Jan 30 13:25:02.464638 containerd[1781]: time="2025-01-30T13:25:02.464617097Z" level=info msg="StopPodSandbox for \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\"" Jan 30 13:25:02.464889 containerd[1781]: time="2025-01-30T13:25:02.464825177Z" level=info msg="TearDown network for sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\" successfully" Jan 30 13:25:02.464889 containerd[1781]: time="2025-01-30T13:25:02.464842057Z" level=info msg="StopPodSandbox for \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\" returns successfully" Jan 30 13:25:02.465466 containerd[1781]: time="2025-01-30T13:25:02.465444296Z" level=info msg="StopPodSandbox for \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\"" Jan 30 13:25:02.465707 containerd[1781]: time="2025-01-30T13:25:02.465636656Z" level=info msg="TearDown network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" successfully" Jan 30 13:25:02.465707 containerd[1781]: time="2025-01-30T13:25:02.465652816Z" level=info msg="StopPodSandbox for \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" returns successfully" Jan 30 13:25:02.466680 containerd[1781]: time="2025-01-30T13:25:02.466540256Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\"" Jan 30 13:25:02.466680 containerd[1781]: time="2025-01-30T13:25:02.466627096Z" level=info msg="TearDown network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" successfully" Jan 30 13:25:02.466680 containerd[1781]: time="2025-01-30T13:25:02.466636696Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" returns successfully" Jan 30 13:25:02.467258 containerd[1781]: time="2025-01-30T13:25:02.467128376Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\"" Jan 30 13:25:02.467339 containerd[1781]: time="2025-01-30T13:25:02.467239336Z" level=info msg="TearDown network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" successfully" Jan 30 13:25:02.467411 containerd[1781]: time="2025-01-30T13:25:02.467376416Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" returns successfully" Jan 30 13:25:02.467813 containerd[1781]: time="2025-01-30T13:25:02.467779376Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" Jan 30 13:25:02.468035 containerd[1781]: time="2025-01-30T13:25:02.467936416Z" level=info msg="TearDown network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" successfully" Jan 30 13:25:02.468035 containerd[1781]: time="2025-01-30T13:25:02.467963456Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" returns successfully" Jan 30 13:25:02.468563 containerd[1781]: time="2025-01-30T13:25:02.468540376Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:25:02.469210 containerd[1781]: time="2025-01-30T13:25:02.469164496Z" level=info msg="TearDown network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" successfully" Jan 30 13:25:02.469210 containerd[1781]: time="2025-01-30T13:25:02.469184016Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" returns successfully" Jan 30 13:25:02.470174 containerd[1781]: time="2025-01-30T13:25:02.470143536Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:7,}" Jan 30 13:25:02.470512 containerd[1781]: time="2025-01-30T13:25:02.470387376Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\"" Jan 30 13:25:02.470512 containerd[1781]: time="2025-01-30T13:25:02.470486056Z" level=info msg="TearDown network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" successfully" Jan 30 13:25:02.470512 containerd[1781]: time="2025-01-30T13:25:02.470497016Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" returns successfully" Jan 30 13:25:02.471548 containerd[1781]: time="2025-01-30T13:25:02.471504335Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" Jan 30 13:25:02.471950 kubelet[3455]: I0130 13:25:02.471828 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102" Jan 30 13:25:02.472084 containerd[1781]: time="2025-01-30T13:25:02.472029895Z" level=info msg="TearDown network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" successfully" Jan 30 13:25:02.473269 containerd[1781]: time="2025-01-30T13:25:02.473192015Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" returns successfully" Jan 30 13:25:02.474744 containerd[1781]: time="2025-01-30T13:25:02.474089055Z" level=info msg="StopPodSandbox for \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\"" Jan 30 13:25:02.474744 containerd[1781]: time="2025-01-30T13:25:02.474239295Z" level=info msg="Ensure that sandbox caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102 in task-service has been cleanup successfully" Jan 30 13:25:02.476810 containerd[1781]: time="2025-01-30T13:25:02.476770894Z" level=info msg="TearDown network for sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\" successfully" Jan 30 13:25:02.476810 containerd[1781]: time="2025-01-30T13:25:02.476799174Z" level=info msg="StopPodSandbox for \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\" returns successfully" Jan 30 13:25:02.476977 containerd[1781]: time="2025-01-30T13:25:02.476949574Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:25:02.477048 containerd[1781]: time="2025-01-30T13:25:02.477029494Z" level=info msg="TearDown network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" successfully" Jan 30 13:25:02.477048 containerd[1781]: time="2025-01-30T13:25:02.477042854Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" returns successfully" Jan 30 13:25:02.478677 containerd[1781]: time="2025-01-30T13:25:02.478625334Z" level=info msg="StopPodSandbox for \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\"" Jan 30 13:25:02.478849 containerd[1781]: time="2025-01-30T13:25:02.478822294Z" level=info msg="TearDown network for sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\" successfully" Jan 30 13:25:02.478849 containerd[1781]: time="2025-01-30T13:25:02.478843454Z" level=info msg="StopPodSandbox for \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\" returns successfully" Jan 30 13:25:02.479519 containerd[1781]: time="2025-01-30T13:25:02.479244774Z" level=info msg="StopPodSandbox for \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\"" Jan 30 13:25:02.479519 containerd[1781]: time="2025-01-30T13:25:02.479342414Z" level=info msg="TearDown network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" successfully" Jan 30 13:25:02.479519 containerd[1781]: time="2025-01-30T13:25:02.479353174Z" level=info msg="StopPodSandbox for \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" returns successfully" Jan 30 13:25:02.480206 containerd[1781]: time="2025-01-30T13:25:02.480089974Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\"" Jan 30 13:25:02.480348 containerd[1781]: time="2025-01-30T13:25:02.480098814Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:7,}" Jan 30 13:25:02.480652 containerd[1781]: time="2025-01-30T13:25:02.480631014Z" level=info msg="TearDown network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" successfully" Jan 30 13:25:02.480764 containerd[1781]: time="2025-01-30T13:25:02.480724494Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" returns successfully" Jan 30 13:25:02.482152 containerd[1781]: time="2025-01-30T13:25:02.481981534Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\"" Jan 30 13:25:02.482152 containerd[1781]: time="2025-01-30T13:25:02.482138853Z" level=info msg="TearDown network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" successfully" Jan 30 13:25:02.482152 containerd[1781]: time="2025-01-30T13:25:02.482152013Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" returns successfully" Jan 30 13:25:02.483019 containerd[1781]: time="2025-01-30T13:25:02.482731773Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" Jan 30 13:25:02.483019 containerd[1781]: time="2025-01-30T13:25:02.482812533Z" level=info msg="TearDown network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" successfully" Jan 30 13:25:02.483019 containerd[1781]: time="2025-01-30T13:25:02.482821293Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" returns successfully" Jan 30 13:25:02.483380 containerd[1781]: time="2025-01-30T13:25:02.483322853Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:25:02.483611 containerd[1781]: time="2025-01-30T13:25:02.483547973Z" level=info msg="TearDown network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" successfully" Jan 30 13:25:02.483995 containerd[1781]: time="2025-01-30T13:25:02.483673253Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" returns successfully" Jan 30 13:25:02.484398 containerd[1781]: time="2025-01-30T13:25:02.484262813Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:7,}" Jan 30 13:25:03.371141 systemd[1]: run-netns-cni\x2d212f6191\x2d3e08\x2d299f\x2d5ae5\x2d95aa8384e1a3.mount: Deactivated successfully. Jan 30 13:25:03.371442 systemd[1]: run-netns-cni\x2dcacc5f59\x2d124d\x2d1ef7\x2d395d\x2d311d0b86f9a8.mount: Deactivated successfully. Jan 30 13:25:03.949804 kernel: bpftool[5742]: memfd_create() called without MFD_EXEC or MFD_NOEXEC_SEAL set Jan 30 13:25:05.018015 systemd-networkd[1549]: vxlan.calico: Link UP Jan 30 13:25:05.018025 systemd-networkd[1549]: vxlan.calico: Gained carrier Jan 30 13:25:07.014122 systemd-networkd[1549]: vxlan.calico: Gained IPv6LL Jan 30 13:25:13.239127 systemd-networkd[1549]: cali835a93d42b6: Link UP Jan 30 13:25:13.239288 systemd-networkd[1549]: cali835a93d42b6: Gained carrier Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.073 [INFO][5825] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0 csi-node-driver- calico-system 17a70524-69f8-4a61-9912-a962709a69cd 600 0 2025-01-30 13:24:44 +0000 UTC map[app.kubernetes.io/name:csi-node-driver controller-revision-hash:65bf684474 k8s-app:csi-node-driver name:csi-node-driver pod-template-generation:1 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:csi-node-driver] map[] [] [] []} {k8s ci-4186.1.0-a-a7402f3063 csi-node-driver-42f6k eth0 csi-node-driver [] [] [kns.calico-system ksa.calico-system.csi-node-driver] cali835a93d42b6 [] []}} ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Namespace="calico-system" Pod="csi-node-driver-42f6k" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.162 [INFO][5825] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Namespace="calico-system" Pod="csi-node-driver-42f6k" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.191 [INFO][5836] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" HandleID="k8s-pod-network.1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Workload="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.202 [INFO][5836] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" HandleID="k8s-pod-network.1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Workload="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000317540), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4186.1.0-a-a7402f3063", "pod":"csi-node-driver-42f6k", "timestamp":"2025-01-30 13:25:13.191634261 +0000 UTC"}, Hostname:"ci-4186.1.0-a-a7402f3063", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.202 [INFO][5836] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.202 [INFO][5836] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.202 [INFO][5836] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4186.1.0-a-a7402f3063' Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.204 [INFO][5836] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.208 [INFO][5836] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.213 [INFO][5836] ipam/ipam.go 489: Trying affinity for 192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.215 [INFO][5836] ipam/ipam.go 155: Attempting to load block cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.217 [INFO][5836] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.217 [INFO][5836] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.102.64/26 handle="k8s-pod-network.1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.219 [INFO][5836] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.226 [INFO][5836] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.102.64/26 handle="k8s-pod-network.1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.232 [INFO][5836] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.102.65/26] block=192.168.102.64/26 handle="k8s-pod-network.1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.232 [INFO][5836] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.102.65/26] handle="k8s-pod-network.1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.232 [INFO][5836] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:25:17.565348 containerd[1781]: 2025-01-30 13:25:13.232 [INFO][5836] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.102.65/26] IPv6=[] ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" HandleID="k8s-pod-network.1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Workload="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" Jan 30 13:25:14.565861 systemd-networkd[1549]: cali835a93d42b6: Gained IPv6LL Jan 30 13:25:17.566175 containerd[1781]: 2025-01-30 13:25:13.235 [INFO][5825] cni-plugin/k8s.go 386: Populated endpoint ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Namespace="calico-system" Pod="csi-node-driver-42f6k" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"17a70524-69f8-4a61-9912-a962709a69cd", ResourceVersion:"600", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"65bf684474", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"", Pod:"csi-node-driver-42f6k", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.102.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali835a93d42b6", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:17.566175 containerd[1781]: 2025-01-30 13:25:13.235 [INFO][5825] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.102.65/32] ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Namespace="calico-system" Pod="csi-node-driver-42f6k" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" Jan 30 13:25:17.566175 containerd[1781]: 2025-01-30 13:25:13.235 [INFO][5825] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali835a93d42b6 ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Namespace="calico-system" Pod="csi-node-driver-42f6k" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" Jan 30 13:25:17.566175 containerd[1781]: 2025-01-30 13:25:13.239 [INFO][5825] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Namespace="calico-system" Pod="csi-node-driver-42f6k" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" Jan 30 13:25:17.566175 containerd[1781]: 2025-01-30 13:25:13.239 [INFO][5825] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Namespace="calico-system" Pod="csi-node-driver-42f6k" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0", GenerateName:"csi-node-driver-", Namespace:"calico-system", SelfLink:"", UID:"17a70524-69f8-4a61-9912-a962709a69cd", ResourceVersion:"600", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"csi-node-driver", "controller-revision-hash":"65bf684474", "k8s-app":"csi-node-driver", "name":"csi-node-driver", "pod-template-generation":"1", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"csi-node-driver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad", Pod:"csi-node-driver-42f6k", Endpoint:"eth0", ServiceAccountName:"csi-node-driver", IPNetworks:[]string{"192.168.102.65/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.csi-node-driver"}, InterfaceName:"cali835a93d42b6", MAC:"b6:d3:82:52:c8:ed", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:17.566175 containerd[1781]: 2025-01-30 13:25:13.254 [INFO][5825] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad" Namespace="calico-system" Pod="csi-node-driver-42f6k" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-csi--node--driver--42f6k-eth0" Jan 30 13:25:18.668234 containerd[1781]: time="2025-01-30T13:25:18.663614025Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:25:18.668234 containerd[1781]: time="2025-01-30T13:25:18.663691425Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:25:18.668848 containerd[1781]: time="2025-01-30T13:25:18.663707625Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:18.671799 containerd[1781]: time="2025-01-30T13:25:18.669697305Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:18.678754 systemd-networkd[1549]: cali69cbe76291a: Link UP Jan 30 13:25:18.680595 systemd-networkd[1549]: cali69cbe76291a: Gained carrier Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.569 [INFO][5868] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0 calico-kube-controllers-6dd45f4c57- calico-system aad42f51-7748-4be3-8b1d-97665b3346fc 728 0 2025-01-30 13:24:44 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:6dd45f4c57 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4186.1.0-a-a7402f3063 calico-kube-controllers-6dd45f4c57-58wrn eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali69cbe76291a [] []}} ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Namespace="calico-system" Pod="calico-kube-controllers-6dd45f4c57-58wrn" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.569 [INFO][5868] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Namespace="calico-system" Pod="calico-kube-controllers-6dd45f4c57-58wrn" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.597 [INFO][5880] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.608 [INFO][5880] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400028ced0), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4186.1.0-a-a7402f3063", "pod":"calico-kube-controllers-6dd45f4c57-58wrn", "timestamp":"2025-01-30 13:25:18.597124027 +0000 UTC"}, Hostname:"ci-4186.1.0-a-a7402f3063", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.608 [INFO][5880] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.609 [INFO][5880] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.609 [INFO][5880] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4186.1.0-a-a7402f3063' Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.610 [INFO][5880] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.614 [INFO][5880] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.618 [INFO][5880] ipam/ipam.go 489: Trying affinity for 192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.620 [INFO][5880] ipam/ipam.go 155: Attempting to load block cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.626 [INFO][5880] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.626 [INFO][5880] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.102.64/26 handle="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.629 [INFO][5880] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5 Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.638 [INFO][5880] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.102.64/26 handle="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.660 [INFO][5880] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.102.66/26] block=192.168.102.64/26 handle="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.661 [INFO][5880] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.102.66/26] handle="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.662 [INFO][5880] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:25:18.727873 containerd[1781]: 2025-01-30 13:25:18.663 [INFO][5880] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.102.66/26] IPv6=[] ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:18.728419 containerd[1781]: 2025-01-30 13:25:18.672 [INFO][5868] cni-plugin/k8s.go 386: Populated endpoint ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Namespace="calico-system" Pod="calico-kube-controllers-6dd45f4c57-58wrn" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0", GenerateName:"calico-kube-controllers-6dd45f4c57-", Namespace:"calico-system", SelfLink:"", UID:"aad42f51-7748-4be3-8b1d-97665b3346fc", ResourceVersion:"728", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6dd45f4c57", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"", Pod:"calico-kube-controllers-6dd45f4c57-58wrn", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.102.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali69cbe76291a", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:18.728419 containerd[1781]: 2025-01-30 13:25:18.673 [INFO][5868] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.102.66/32] ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Namespace="calico-system" Pod="calico-kube-controllers-6dd45f4c57-58wrn" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:18.728419 containerd[1781]: 2025-01-30 13:25:18.673 [INFO][5868] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali69cbe76291a ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Namespace="calico-system" Pod="calico-kube-controllers-6dd45f4c57-58wrn" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:18.728419 containerd[1781]: 2025-01-30 13:25:18.683 [INFO][5868] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Namespace="calico-system" Pod="calico-kube-controllers-6dd45f4c57-58wrn" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:18.728419 containerd[1781]: 2025-01-30 13:25:18.694 [INFO][5868] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Namespace="calico-system" Pod="calico-kube-controllers-6dd45f4c57-58wrn" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0", GenerateName:"calico-kube-controllers-6dd45f4c57-", Namespace:"calico-system", SelfLink:"", UID:"aad42f51-7748-4be3-8b1d-97665b3346fc", ResourceVersion:"728", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 44, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"6dd45f4c57", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5", Pod:"calico-kube-controllers-6dd45f4c57-58wrn", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.102.66/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali69cbe76291a", MAC:"12:8f:95:53:37:5e", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:18.728419 containerd[1781]: 2025-01-30 13:25:18.722 [INFO][5868] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Namespace="calico-system" Pod="calico-kube-controllers-6dd45f4c57-58wrn" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:18.743969 systemd[1]: Started cri-containerd-1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad.scope - libcontainer container 1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad. Jan 30 13:25:18.829457 containerd[1781]: time="2025-01-30T13:25:18.829371822Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:csi-node-driver-42f6k,Uid:17a70524-69f8-4a61-9912-a962709a69cd,Namespace:calico-system,Attempt:7,} returns sandbox id \"1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad\"" Jan 30 13:25:18.832978 containerd[1781]: time="2025-01-30T13:25:18.832746782Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\"" Jan 30 13:25:18.854058 containerd[1781]: time="2025-01-30T13:25:18.852350581Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:25:18.854058 containerd[1781]: time="2025-01-30T13:25:18.852419701Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:25:18.854058 containerd[1781]: time="2025-01-30T13:25:18.852435181Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:18.854058 containerd[1781]: time="2025-01-30T13:25:18.852523261Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:18.915049 systemd[1]: Started cri-containerd-39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5.scope - libcontainer container 39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5. Jan 30 13:25:19.066446 containerd[1781]: time="2025-01-30T13:25:19.066399617Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-6dd45f4c57-58wrn,Uid:aad42f51-7748-4be3-8b1d-97665b3346fc,Namespace:calico-system,Attempt:7,} returns sandbox id \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\"" Jan 30 13:25:19.080345 systemd-networkd[1549]: cali4d307d4edd3: Link UP Jan 30 13:25:19.083168 systemd-networkd[1549]: cali4d307d4edd3: Gained carrier Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:18.853 [INFO][5934] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0 calico-apiserver-c978d777- calico-apiserver 376f6955-c43d-49c5-a9d0-4891ee1dc8f6 727 0 2025-01-30 13:24:45 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:c978d777 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4186.1.0-a-a7402f3063 calico-apiserver-c978d777-x5x6w eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali4d307d4edd3 [] []}} ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-x5x6w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:18.853 [INFO][5934] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-x5x6w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:18.974 [INFO][6022] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" HandleID="k8s-pod-network.2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.005 [INFO][6022] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" HandleID="k8s-pod-network.2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400047d060), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4186.1.0-a-a7402f3063", "pod":"calico-apiserver-c978d777-x5x6w", "timestamp":"2025-01-30 13:25:18.968523179 +0000 UTC"}, Hostname:"ci-4186.1.0-a-a7402f3063", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.005 [INFO][6022] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.005 [INFO][6022] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.005 [INFO][6022] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4186.1.0-a-a7402f3063' Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.009 [INFO][6022] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.017 [INFO][6022] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.021 [INFO][6022] ipam/ipam.go 489: Trying affinity for 192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.028 [INFO][6022] ipam/ipam.go 155: Attempting to load block cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.034 [INFO][6022] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.034 [INFO][6022] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.102.64/26 handle="k8s-pod-network.2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.037 [INFO][6022] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275 Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.048 [INFO][6022] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.102.64/26 handle="k8s-pod-network.2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.062 [INFO][6022] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.102.67/26] block=192.168.102.64/26 handle="k8s-pod-network.2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.062 [INFO][6022] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.102.67/26] handle="k8s-pod-network.2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.062 [INFO][6022] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:25:19.106012 containerd[1781]: 2025-01-30 13:25:19.062 [INFO][6022] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.102.67/26] IPv6=[] ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" HandleID="k8s-pod-network.2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" Jan 30 13:25:19.106684 containerd[1781]: 2025-01-30 13:25:19.074 [INFO][5934] cni-plugin/k8s.go 386: Populated endpoint ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-x5x6w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0", GenerateName:"calico-apiserver-c978d777-", Namespace:"calico-apiserver", SelfLink:"", UID:"376f6955-c43d-49c5-a9d0-4891ee1dc8f6", ResourceVersion:"727", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"c978d777", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"", Pod:"calico-apiserver-c978d777-x5x6w", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.102.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4d307d4edd3", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:19.106684 containerd[1781]: 2025-01-30 13:25:19.074 [INFO][5934] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.102.67/32] ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-x5x6w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" Jan 30 13:25:19.106684 containerd[1781]: 2025-01-30 13:25:19.074 [INFO][5934] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4d307d4edd3 ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-x5x6w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" Jan 30 13:25:19.106684 containerd[1781]: 2025-01-30 13:25:19.083 [INFO][5934] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-x5x6w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" Jan 30 13:25:19.106684 containerd[1781]: 2025-01-30 13:25:19.083 [INFO][5934] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-x5x6w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0", GenerateName:"calico-apiserver-c978d777-", Namespace:"calico-apiserver", SelfLink:"", UID:"376f6955-c43d-49c5-a9d0-4891ee1dc8f6", ResourceVersion:"727", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"c978d777", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275", Pod:"calico-apiserver-c978d777-x5x6w", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.102.67/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali4d307d4edd3", MAC:"fe:d1:ad:79:bb:1a", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:19.106684 containerd[1781]: 2025-01-30 13:25:19.103 [INFO][5934] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-x5x6w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--x5x6w-eth0" Jan 30 13:25:19.153812 containerd[1781]: time="2025-01-30T13:25:19.153686495Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:25:19.154033 containerd[1781]: time="2025-01-30T13:25:19.153779935Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:25:19.154033 containerd[1781]: time="2025-01-30T13:25:19.153795935Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:19.154033 containerd[1781]: time="2025-01-30T13:25:19.153893975Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:19.160845 systemd-networkd[1549]: cali4e2d61fecb9: Link UP Jan 30 13:25:19.163155 systemd-networkd[1549]: cali4e2d61fecb9: Gained carrier Jan 30 13:25:19.214360 systemd[1]: Started cri-containerd-2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275.scope - libcontainer container 2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275. Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:18.848 [INFO][5909] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0 coredns-7db6d8ff4d- kube-system 0b84efc9-a798-4f8e-8d4f-6bba2eb764e3 725 0 2025-01-30 13:24:36 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4186.1.0-a-a7402f3063 coredns-7db6d8ff4d-5hs8w eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] cali4e2d61fecb9 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5hs8w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:18.849 [INFO][5909] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5hs8w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:18.981 [INFO][6021] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" HandleID="k8s-pod-network.d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Workload="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.005 [INFO][6021] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" HandleID="k8s-pod-network.d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Workload="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x40002cf630), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4186.1.0-a-a7402f3063", "pod":"coredns-7db6d8ff4d-5hs8w", "timestamp":"2025-01-30 13:25:18.981211658 +0000 UTC"}, Hostname:"ci-4186.1.0-a-a7402f3063", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.006 [INFO][6021] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.062 [INFO][6021] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.062 [INFO][6021] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4186.1.0-a-a7402f3063' Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.066 [INFO][6021] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.076 [INFO][6021] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.094 [INFO][6021] ipam/ipam.go 489: Trying affinity for 192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.102 [INFO][6021] ipam/ipam.go 155: Attempting to load block cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.110 [INFO][6021] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.110 [INFO][6021] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.102.64/26 handle="k8s-pod-network.d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.116 [INFO][6021] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7 Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.131 [INFO][6021] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.102.64/26 handle="k8s-pod-network.d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.147 [INFO][6021] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.102.68/26] block=192.168.102.64/26 handle="k8s-pod-network.d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.148 [INFO][6021] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.102.68/26] handle="k8s-pod-network.d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.148 [INFO][6021] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:25:19.215607 containerd[1781]: 2025-01-30 13:25:19.148 [INFO][6021] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.102.68/26] IPv6=[] ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" HandleID="k8s-pod-network.d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Workload="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" Jan 30 13:25:19.217446 containerd[1781]: 2025-01-30 13:25:19.153 [INFO][5909] cni-plugin/k8s.go 386: Populated endpoint ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5hs8w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"0b84efc9-a798-4f8e-8d4f-6bba2eb764e3", ResourceVersion:"725", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"", Pod:"coredns-7db6d8ff4d-5hs8w", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.102.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4e2d61fecb9", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:19.217446 containerd[1781]: 2025-01-30 13:25:19.154 [INFO][5909] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.102.68/32] ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5hs8w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" Jan 30 13:25:19.217446 containerd[1781]: 2025-01-30 13:25:19.155 [INFO][5909] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali4e2d61fecb9 ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5hs8w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" Jan 30 13:25:19.217446 containerd[1781]: 2025-01-30 13:25:19.165 [INFO][5909] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5hs8w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" Jan 30 13:25:19.217446 containerd[1781]: 2025-01-30 13:25:19.167 [INFO][5909] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5hs8w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"0b84efc9-a798-4f8e-8d4f-6bba2eb764e3", ResourceVersion:"725", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7", Pod:"coredns-7db6d8ff4d-5hs8w", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.102.68/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"cali4e2d61fecb9", MAC:"0e:6c:a2:18:cc:c4", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:19.217446 containerd[1781]: 2025-01-30 13:25:19.205 [INFO][5909] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5hs8w" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5hs8w-eth0" Jan 30 13:25:19.257600 containerd[1781]: time="2025-01-30T13:25:19.257467093Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:25:19.258053 containerd[1781]: time="2025-01-30T13:25:19.257868293Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:25:19.258155 containerd[1781]: time="2025-01-30T13:25:19.258054413Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:19.258609 containerd[1781]: time="2025-01-30T13:25:19.258544013Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:19.284733 systemd-networkd[1549]: calia9b14064ce8: Link UP Jan 30 13:25:19.286708 systemd-networkd[1549]: calia9b14064ce8: Gained carrier Jan 30 13:25:19.316527 systemd[1]: Started cri-containerd-d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7.scope - libcontainer container d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7. Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:18.797 [INFO][5895] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0 coredns-7db6d8ff4d- kube-system efc446ea-7180-4d70-91d5-8919459a6efa 722 0 2025-01-30 13:24:36 +0000 UTC map[k8s-app:kube-dns pod-template-hash:7db6d8ff4d projectcalico.org/namespace:kube-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:coredns] map[] [] [] []} {k8s ci-4186.1.0-a-a7402f3063 coredns-7db6d8ff4d-5djx9 eth0 coredns [] [] [kns.kube-system ksa.kube-system.coredns] calia9b14064ce8 [{dns UDP 53 0 } {dns-tcp TCP 53 0 } {metrics TCP 9153 0 }] []}} ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5djx9" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:18.797 [INFO][5895] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5djx9" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:18.979 [INFO][5998] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" HandleID="k8s-pod-network.393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Workload="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.010 [INFO][5998] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" HandleID="k8s-pod-network.393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Workload="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400011a230), Attrs:map[string]string{"namespace":"kube-system", "node":"ci-4186.1.0-a-a7402f3063", "pod":"coredns-7db6d8ff4d-5djx9", "timestamp":"2025-01-30 13:25:18.979188658 +0000 UTC"}, Hostname:"ci-4186.1.0-a-a7402f3063", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.010 [INFO][5998] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.149 [INFO][5998] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.149 [INFO][5998] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4186.1.0-a-a7402f3063' Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.159 [INFO][5998] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.179 [INFO][5998] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.203 [INFO][5998] ipam/ipam.go 489: Trying affinity for 192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.213 [INFO][5998] ipam/ipam.go 155: Attempting to load block cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.226 [INFO][5998] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.226 [INFO][5998] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.102.64/26 handle="k8s-pod-network.393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.231 [INFO][5998] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.241 [INFO][5998] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.102.64/26 handle="k8s-pod-network.393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.268 [INFO][5998] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.102.69/26] block=192.168.102.64/26 handle="k8s-pod-network.393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.268 [INFO][5998] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.102.69/26] handle="k8s-pod-network.393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.268 [INFO][5998] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:25:19.330025 containerd[1781]: 2025-01-30 13:25:19.269 [INFO][5998] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.102.69/26] IPv6=[] ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" HandleID="k8s-pod-network.393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Workload="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" Jan 30 13:25:19.330676 containerd[1781]: 2025-01-30 13:25:19.276 [INFO][5895] cni-plugin/k8s.go 386: Populated endpoint ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5djx9" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"efc446ea-7180-4d70-91d5-8919459a6efa", ResourceVersion:"722", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"", Pod:"coredns-7db6d8ff4d-5djx9", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.102.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia9b14064ce8", MAC:"", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:19.330676 containerd[1781]: 2025-01-30 13:25:19.277 [INFO][5895] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.102.69/32] ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5djx9" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" Jan 30 13:25:19.330676 containerd[1781]: 2025-01-30 13:25:19.278 [INFO][5895] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to calia9b14064ce8 ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5djx9" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" Jan 30 13:25:19.330676 containerd[1781]: 2025-01-30 13:25:19.287 [INFO][5895] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5djx9" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" Jan 30 13:25:19.330676 containerd[1781]: 2025-01-30 13:25:19.290 [INFO][5895] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5djx9" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0", GenerateName:"coredns-7db6d8ff4d-", Namespace:"kube-system", SelfLink:"", UID:"efc446ea-7180-4d70-91d5-8919459a6efa", ResourceVersion:"722", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 36, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"k8s-app":"kube-dns", "pod-template-hash":"7db6d8ff4d", "projectcalico.org/namespace":"kube-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"coredns"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d", Pod:"coredns-7db6d8ff4d-5djx9", Endpoint:"eth0", ServiceAccountName:"coredns", IPNetworks:[]string{"192.168.102.69/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.kube-system", "ksa.kube-system.coredns"}, InterfaceName:"calia9b14064ce8", MAC:"12:74:47:d7:ac:ba", Ports:[]v3.WorkloadEndpointPort{v3.WorkloadEndpointPort{Name:"dns", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"UDP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"dns-tcp", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x35, HostPort:0x0, HostIP:""}, v3.WorkloadEndpointPort{Name:"metrics", Protocol:numorstring.Protocol{Type:1, NumVal:0x0, StrVal:"TCP"}, Port:0x23c1, HostPort:0x0, HostIP:""}}, AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:19.330676 containerd[1781]: 2025-01-30 13:25:19.317 [INFO][5895] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d" Namespace="kube-system" Pod="coredns-7db6d8ff4d-5djx9" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-coredns--7db6d8ff4d--5djx9-eth0" Jan 30 13:25:19.407393 containerd[1781]: time="2025-01-30T13:25:19.406015339Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5hs8w,Uid:0b84efc9-a798-4f8e-8d4f-6bba2eb764e3,Namespace:kube-system,Attempt:7,} returns sandbox id \"d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7\"" Jan 30 13:25:19.421806 systemd-networkd[1549]: cali585c90e981b: Link UP Jan 30 13:25:19.427877 containerd[1781]: time="2025-01-30T13:25:19.425692460Z" level=info msg="CreateContainer within sandbox \"d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 30 13:25:19.428084 systemd-networkd[1549]: cali585c90e981b: Gained carrier Jan 30 13:25:19.438950 containerd[1781]: time="2025-01-30T13:25:19.435939581Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:25:19.438950 containerd[1781]: time="2025-01-30T13:25:19.436241421Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:25:19.438950 containerd[1781]: time="2025-01-30T13:25:19.436262461Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:19.438950 containerd[1781]: time="2025-01-30T13:25:19.436846421Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:18.878 [INFO][5926] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0 calico-apiserver-c978d777- calico-apiserver ce7cd8b7-6707-4478-a092-50a5a608f5e0 729 0 2025-01-30 13:24:45 +0000 UTC map[apiserver:true app.kubernetes.io/name:calico-apiserver k8s-app:calico-apiserver pod-template-hash:c978d777 projectcalico.org/namespace:calico-apiserver projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-apiserver] map[] [] [] []} {k8s ci-4186.1.0-a-a7402f3063 calico-apiserver-c978d777-jmpf2 eth0 calico-apiserver [] [] [kns.calico-apiserver ksa.calico-apiserver.calico-apiserver] cali585c90e981b [] []}} ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-jmpf2" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:18.878 [INFO][5926] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-jmpf2" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.012 [INFO][6033] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" HandleID="k8s-pod-network.a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.035 [INFO][6033] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" HandleID="k8s-pod-network.a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x4000317430), Attrs:map[string]string{"namespace":"calico-apiserver", "node":"ci-4186.1.0-a-a7402f3063", "pod":"calico-apiserver-c978d777-jmpf2", "timestamp":"2025-01-30 13:25:19.012780538 +0000 UTC"}, Hostname:"ci-4186.1.0-a-a7402f3063", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.035 [INFO][6033] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.268 [INFO][6033] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.269 [INFO][6033] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4186.1.0-a-a7402f3063' Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.275 [INFO][6033] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.291 [INFO][6033] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.306 [INFO][6033] ipam/ipam.go 489: Trying affinity for 192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.314 [INFO][6033] ipam/ipam.go 155: Attempting to load block cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.327 [INFO][6033] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.329 [INFO][6033] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.102.64/26 handle="k8s-pod-network.a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.336 [INFO][6033] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.348 [INFO][6033] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.102.64/26 handle="k8s-pod-network.a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.376 [INFO][6033] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.102.70/26] block=192.168.102.64/26 handle="k8s-pod-network.a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.382 [INFO][6033] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.102.70/26] handle="k8s-pod-network.a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.382 [INFO][6033] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:25:19.465591 containerd[1781]: 2025-01-30 13:25:19.383 [INFO][6033] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.102.70/26] IPv6=[] ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" HandleID="k8s-pod-network.a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" Jan 30 13:25:19.466247 containerd[1781]: 2025-01-30 13:25:19.399 [INFO][5926] cni-plugin/k8s.go 386: Populated endpoint ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-jmpf2" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0", GenerateName:"calico-apiserver-c978d777-", Namespace:"calico-apiserver", SelfLink:"", UID:"ce7cd8b7-6707-4478-a092-50a5a608f5e0", ResourceVersion:"729", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"c978d777", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"", Pod:"calico-apiserver-c978d777-jmpf2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.102.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali585c90e981b", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:19.466247 containerd[1781]: 2025-01-30 13:25:19.399 [INFO][5926] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.102.70/32] ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-jmpf2" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" Jan 30 13:25:19.466247 containerd[1781]: 2025-01-30 13:25:19.399 [INFO][5926] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali585c90e981b ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-jmpf2" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" Jan 30 13:25:19.466247 containerd[1781]: 2025-01-30 13:25:19.429 [INFO][5926] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-jmpf2" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" Jan 30 13:25:19.466247 containerd[1781]: 2025-01-30 13:25:19.434 [INFO][5926] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-jmpf2" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0", GenerateName:"calico-apiserver-c978d777-", Namespace:"calico-apiserver", SelfLink:"", UID:"ce7cd8b7-6707-4478-a092-50a5a608f5e0", ResourceVersion:"729", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 24, 45, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"apiserver":"true", "app.kubernetes.io/name":"calico-apiserver", "k8s-app":"calico-apiserver", "pod-template-hash":"c978d777", "projectcalico.org/namespace":"calico-apiserver", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-apiserver"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb", Pod:"calico-apiserver-c978d777-jmpf2", Endpoint:"eth0", ServiceAccountName:"calico-apiserver", IPNetworks:[]string{"192.168.102.70/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-apiserver", "ksa.calico-apiserver.calico-apiserver"}, InterfaceName:"cali585c90e981b", MAC:"32:d6:aa:d8:2d:97", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:19.466247 containerd[1781]: 2025-01-30 13:25:19.453 [INFO][5926] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb" Namespace="calico-apiserver" Pod="calico-apiserver-c978d777-jmpf2" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--apiserver--c978d777--jmpf2-eth0" Jan 30 13:25:19.474954 systemd[1]: Started cri-containerd-393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d.scope - libcontainer container 393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d. Jan 30 13:25:19.484242 containerd[1781]: time="2025-01-30T13:25:19.484170424Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-x5x6w,Uid:376f6955-c43d-49c5-a9d0-4891ee1dc8f6,Namespace:calico-apiserver,Attempt:7,} returns sandbox id \"2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275\"" Jan 30 13:25:19.498312 containerd[1781]: time="2025-01-30T13:25:19.498187425Z" level=info msg="CreateContainer within sandbox \"d64a17947db52bbf423542456c28ab289a65d29c7f582548098fcf232eed59b7\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"af9e67b4964cee0945952365e46812239c9ff4cdac033e022d39190f8ed7ad26\"" Jan 30 13:25:19.500246 containerd[1781]: time="2025-01-30T13:25:19.500105905Z" level=info msg="StartContainer for \"af9e67b4964cee0945952365e46812239c9ff4cdac033e022d39190f8ed7ad26\"" Jan 30 13:25:19.548886 containerd[1781]: time="2025-01-30T13:25:19.541970428Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:25:19.548886 containerd[1781]: time="2025-01-30T13:25:19.542031628Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:25:19.548886 containerd[1781]: time="2025-01-30T13:25:19.542047548Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:19.548886 containerd[1781]: time="2025-01-30T13:25:19.542427988Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:19.592951 systemd[1]: Started cri-containerd-af9e67b4964cee0945952365e46812239c9ff4cdac033e022d39190f8ed7ad26.scope - libcontainer container af9e67b4964cee0945952365e46812239c9ff4cdac033e022d39190f8ed7ad26. Jan 30 13:25:19.599511 containerd[1781]: time="2025-01-30T13:25:19.599382671Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-7db6d8ff4d-5djx9,Uid:efc446ea-7180-4d70-91d5-8919459a6efa,Namespace:kube-system,Attempt:7,} returns sandbox id \"393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d\"" Jan 30 13:25:19.605929 containerd[1781]: time="2025-01-30T13:25:19.605823392Z" level=info msg="CreateContainer within sandbox \"393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Jan 30 13:25:19.624907 systemd[1]: Started cri-containerd-a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb.scope - libcontainer container a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb. Jan 30 13:25:19.652583 containerd[1781]: time="2025-01-30T13:25:19.652422835Z" level=info msg="StartContainer for \"af9e67b4964cee0945952365e46812239c9ff4cdac033e022d39190f8ed7ad26\" returns successfully" Jan 30 13:25:19.657085 containerd[1781]: time="2025-01-30T13:25:19.657011675Z" level=info msg="CreateContainer within sandbox \"393552f1fc87be392ccf87d595cdc847d0efb1958f13afc0d7bd8b44a6e4271d\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"977eae98b760e66e2adcf17adec02c39437ec3c84aba79c765edb0b6bb680b4e\"" Jan 30 13:25:19.659518 containerd[1781]: time="2025-01-30T13:25:19.659444995Z" level=info msg="StartContainer for \"977eae98b760e66e2adcf17adec02c39437ec3c84aba79c765edb0b6bb680b4e\"" Jan 30 13:25:19.717029 systemd[1]: Started cri-containerd-977eae98b760e66e2adcf17adec02c39437ec3c84aba79c765edb0b6bb680b4e.scope - libcontainer container 977eae98b760e66e2adcf17adec02c39437ec3c84aba79c765edb0b6bb680b4e. Jan 30 13:25:19.744314 containerd[1781]: time="2025-01-30T13:25:19.744260400Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-apiserver-c978d777-jmpf2,Uid:ce7cd8b7-6707-4478-a092-50a5a608f5e0,Namespace:calico-apiserver,Attempt:7,} returns sandbox id \"a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb\"" Jan 30 13:25:19.767898 containerd[1781]: time="2025-01-30T13:25:19.767839842Z" level=info msg="StartContainer for \"977eae98b760e66e2adcf17adec02c39437ec3c84aba79c765edb0b6bb680b4e\" returns successfully" Jan 30 13:25:20.325996 systemd-networkd[1549]: cali69cbe76291a: Gained IPv6LL Jan 30 13:25:20.467898 containerd[1781]: time="2025-01-30T13:25:20.467806805Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:20.472246 containerd[1781]: time="2025-01-30T13:25:20.472018126Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/csi:v3.29.1: active requests=0, bytes read=7464730" Jan 30 13:25:20.478980 containerd[1781]: time="2025-01-30T13:25:20.478650406Z" level=info msg="ImageCreate event name:\"sha256:3c11734f3001b7070e7e2b5e64938f89891cf8c44f8997e86aa23c5d5bf70163\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:20.492582 containerd[1781]: time="2025-01-30T13:25:20.492513847Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:20.494838 containerd[1781]: time="2025-01-30T13:25:20.494730247Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/csi:v3.29.1\" with image id \"sha256:3c11734f3001b7070e7e2b5e64938f89891cf8c44f8997e86aa23c5d5bf70163\", repo tag \"ghcr.io/flatcar/calico/csi:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/csi@sha256:eaa7e01fb16b603c155a67b81f16992281db7f831684c7b2081d3434587a7ff3\", size \"8834384\" in 1.661926585s" Jan 30 13:25:20.494838 containerd[1781]: time="2025-01-30T13:25:20.494781087Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/csi:v3.29.1\" returns image reference \"sha256:3c11734f3001b7070e7e2b5e64938f89891cf8c44f8997e86aa23c5d5bf70163\"" Jan 30 13:25:20.497584 containerd[1781]: time="2025-01-30T13:25:20.497117367Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\"" Jan 30 13:25:20.500671 containerd[1781]: time="2025-01-30T13:25:20.500256407Z" level=info msg="CreateContainer within sandbox \"1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad\" for container &ContainerMetadata{Name:calico-csi,Attempt:0,}" Jan 30 13:25:20.562791 containerd[1781]: time="2025-01-30T13:25:20.562628131Z" level=info msg="CreateContainer within sandbox \"1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad\" for &ContainerMetadata{Name:calico-csi,Attempt:0,} returns container id \"3dbd85da82fa98e1844b468e4e94d0170d478534c275e91f90cb2f152c576539\"" Jan 30 13:25:20.563357 containerd[1781]: time="2025-01-30T13:25:20.563302651Z" level=info msg="StartContainer for \"3dbd85da82fa98e1844b468e4e94d0170d478534c275e91f90cb2f152c576539\"" Jan 30 13:25:20.603961 systemd[1]: Started cri-containerd-3dbd85da82fa98e1844b468e4e94d0170d478534c275e91f90cb2f152c576539.scope - libcontainer container 3dbd85da82fa98e1844b468e4e94d0170d478534c275e91f90cb2f152c576539. Jan 30 13:25:20.654948 kubelet[3455]: I0130 13:25:20.654484 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-5hs8w" podStartSLOduration=44.654459337 podStartE2EDuration="44.654459337s" podCreationTimestamp="2025-01-30 13:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:25:20.652657257 +0000 UTC m=+57.882254364" watchObservedRunningTime="2025-01-30 13:25:20.654459337 +0000 UTC m=+57.884056324" Jan 30 13:25:20.683609 containerd[1781]: time="2025-01-30T13:25:20.683503459Z" level=info msg="StartContainer for \"3dbd85da82fa98e1844b468e4e94d0170d478534c275e91f90cb2f152c576539\" returns successfully" Jan 30 13:25:20.902087 systemd-networkd[1549]: cali4e2d61fecb9: Gained IPv6LL Jan 30 13:25:20.965868 systemd-networkd[1549]: cali4d307d4edd3: Gained IPv6LL Jan 30 13:25:21.221862 systemd-networkd[1549]: calia9b14064ce8: Gained IPv6LL Jan 30 13:25:21.349949 systemd-networkd[1549]: cali585c90e981b: Gained IPv6LL Jan 30 13:25:22.887067 containerd[1781]: time="2025-01-30T13:25:22.887022397Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:25:22.887443 containerd[1781]: time="2025-01-30T13:25:22.887146797Z" level=info msg="TearDown network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" successfully" Jan 30 13:25:22.887443 containerd[1781]: time="2025-01-30T13:25:22.887157317Z" level=info msg="StopPodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" returns successfully" Jan 30 13:25:22.887804 containerd[1781]: time="2025-01-30T13:25:22.887776597Z" level=info msg="RemovePodSandbox for \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:25:22.887867 containerd[1781]: time="2025-01-30T13:25:22.887807837Z" level=info msg="Forcibly stopping sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\"" Jan 30 13:25:22.887918 containerd[1781]: time="2025-01-30T13:25:22.887887517Z" level=info msg="TearDown network for sandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" successfully" Jan 30 13:25:22.897284 containerd[1781]: time="2025-01-30T13:25:22.897228997Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:22.897434 containerd[1781]: time="2025-01-30T13:25:22.897311837Z" level=info msg="RemovePodSandbox \"618d21caa98c634d8e7962a7e6b6481a915e4fe85c2324054979485fba8e6611\" returns successfully" Jan 30 13:25:22.898206 containerd[1781]: time="2025-01-30T13:25:22.898023837Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" Jan 30 13:25:22.898206 containerd[1781]: time="2025-01-30T13:25:22.898131757Z" level=info msg="TearDown network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" successfully" Jan 30 13:25:22.898206 containerd[1781]: time="2025-01-30T13:25:22.898141637Z" level=info msg="StopPodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" returns successfully" Jan 30 13:25:22.898519 containerd[1781]: time="2025-01-30T13:25:22.898493797Z" level=info msg="RemovePodSandbox for \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" Jan 30 13:25:22.898563 containerd[1781]: time="2025-01-30T13:25:22.898523357Z" level=info msg="Forcibly stopping sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\"" Jan 30 13:25:22.898656 containerd[1781]: time="2025-01-30T13:25:22.898599077Z" level=info msg="TearDown network for sandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" successfully" Jan 30 13:25:22.906407 containerd[1781]: time="2025-01-30T13:25:22.906357238Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:22.906547 containerd[1781]: time="2025-01-30T13:25:22.906432838Z" level=info msg="RemovePodSandbox \"2a28472712fca9baa9388d8307ef7b96b458f1bce602a042bbd42d72dc7cf051\" returns successfully" Jan 30 13:25:22.907094 containerd[1781]: time="2025-01-30T13:25:22.906902758Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\"" Jan 30 13:25:22.907094 containerd[1781]: time="2025-01-30T13:25:22.906999278Z" level=info msg="TearDown network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" successfully" Jan 30 13:25:22.907094 containerd[1781]: time="2025-01-30T13:25:22.907008878Z" level=info msg="StopPodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" returns successfully" Jan 30 13:25:22.907649 containerd[1781]: time="2025-01-30T13:25:22.907473238Z" level=info msg="RemovePodSandbox for \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\"" Jan 30 13:25:22.907649 containerd[1781]: time="2025-01-30T13:25:22.907512838Z" level=info msg="Forcibly stopping sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\"" Jan 30 13:25:22.907649 containerd[1781]: time="2025-01-30T13:25:22.907597878Z" level=info msg="TearDown network for sandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" successfully" Jan 30 13:25:22.921262 containerd[1781]: time="2025-01-30T13:25:22.921219559Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:22.921560 containerd[1781]: time="2025-01-30T13:25:22.921458599Z" level=info msg="RemovePodSandbox \"3b15bdc15f7b229bb8934737fbd0d97a28e8ce1226587253220b84350b4186c9\" returns successfully" Jan 30 13:25:22.922129 containerd[1781]: time="2025-01-30T13:25:22.922098919Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\"" Jan 30 13:25:22.922231 containerd[1781]: time="2025-01-30T13:25:22.922210959Z" level=info msg="TearDown network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" successfully" Jan 30 13:25:22.922231 containerd[1781]: time="2025-01-30T13:25:22.922225879Z" level=info msg="StopPodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" returns successfully" Jan 30 13:25:22.922851 containerd[1781]: time="2025-01-30T13:25:22.922821079Z" level=info msg="RemovePodSandbox for \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\"" Jan 30 13:25:22.922920 containerd[1781]: time="2025-01-30T13:25:22.922852919Z" level=info msg="Forcibly stopping sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\"" Jan 30 13:25:22.922947 containerd[1781]: time="2025-01-30T13:25:22.922928159Z" level=info msg="TearDown network for sandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" successfully" Jan 30 13:25:22.931789 containerd[1781]: time="2025-01-30T13:25:22.931734119Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:22.931919 containerd[1781]: time="2025-01-30T13:25:22.931825959Z" level=info msg="RemovePodSandbox \"3da12ff87cde2435aea9b1a5bffa8260354c1d9e717867635371bd31d16ecf08\" returns successfully" Jan 30 13:25:22.932333 containerd[1781]: time="2025-01-30T13:25:22.932304319Z" level=info msg="StopPodSandbox for \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\"" Jan 30 13:25:22.932433 containerd[1781]: time="2025-01-30T13:25:22.932410359Z" level=info msg="TearDown network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" successfully" Jan 30 13:25:22.932433 containerd[1781]: time="2025-01-30T13:25:22.932428759Z" level=info msg="StopPodSandbox for \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" returns successfully" Jan 30 13:25:22.932837 containerd[1781]: time="2025-01-30T13:25:22.932794839Z" level=info msg="RemovePodSandbox for \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\"" Jan 30 13:25:22.932837 containerd[1781]: time="2025-01-30T13:25:22.932824439Z" level=info msg="Forcibly stopping sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\"" Jan 30 13:25:22.933030 containerd[1781]: time="2025-01-30T13:25:22.932899759Z" level=info msg="TearDown network for sandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" successfully" Jan 30 13:25:22.940401 containerd[1781]: time="2025-01-30T13:25:22.940321960Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:22.940401 containerd[1781]: time="2025-01-30T13:25:22.940407560Z" level=info msg="RemovePodSandbox \"364d3f2041d23059dc4777d765be6f49299b737eabe574c1399b0676318f4636\" returns successfully" Jan 30 13:25:22.941152 containerd[1781]: time="2025-01-30T13:25:22.940882960Z" level=info msg="StopPodSandbox for \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\"" Jan 30 13:25:22.941152 containerd[1781]: time="2025-01-30T13:25:22.940980480Z" level=info msg="TearDown network for sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\" successfully" Jan 30 13:25:22.941152 containerd[1781]: time="2025-01-30T13:25:22.940990160Z" level=info msg="StopPodSandbox for \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\" returns successfully" Jan 30 13:25:22.942275 containerd[1781]: time="2025-01-30T13:25:22.942104280Z" level=info msg="RemovePodSandbox for \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\"" Jan 30 13:25:22.942275 containerd[1781]: time="2025-01-30T13:25:22.942151320Z" level=info msg="Forcibly stopping sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\"" Jan 30 13:25:22.942275 containerd[1781]: time="2025-01-30T13:25:22.942225080Z" level=info msg="TearDown network for sandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\" successfully" Jan 30 13:25:22.952598 containerd[1781]: time="2025-01-30T13:25:22.952488041Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:22.952598 containerd[1781]: time="2025-01-30T13:25:22.952568761Z" level=info msg="RemovePodSandbox \"c4475f8995a716258ce187c69128be298c09b89498012fc07625a50f5ccd0438\" returns successfully" Jan 30 13:25:22.953045 containerd[1781]: time="2025-01-30T13:25:22.953012801Z" level=info msg="StopPodSandbox for \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\"" Jan 30 13:25:22.953137 containerd[1781]: time="2025-01-30T13:25:22.953118641Z" level=info msg="TearDown network for sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\" successfully" Jan 30 13:25:22.953165 containerd[1781]: time="2025-01-30T13:25:22.953140721Z" level=info msg="StopPodSandbox for \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\" returns successfully" Jan 30 13:25:22.953863 containerd[1781]: time="2025-01-30T13:25:22.953489561Z" level=info msg="RemovePodSandbox for \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\"" Jan 30 13:25:22.953863 containerd[1781]: time="2025-01-30T13:25:22.953518321Z" level=info msg="Forcibly stopping sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\"" Jan 30 13:25:22.953863 containerd[1781]: time="2025-01-30T13:25:22.953584361Z" level=info msg="TearDown network for sandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\" successfully" Jan 30 13:25:22.968599 containerd[1781]: time="2025-01-30T13:25:22.967953762Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:22.968599 containerd[1781]: time="2025-01-30T13:25:22.968058002Z" level=info msg="RemovePodSandbox \"caef3c0ceae7ed9b21fa3510b7201d801161ee5c3d1100f0c5f0860af9a10102\" returns successfully" Jan 30 13:25:22.969678 containerd[1781]: time="2025-01-30T13:25:22.969649162Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:25:22.969996 containerd[1781]: time="2025-01-30T13:25:22.969978282Z" level=info msg="TearDown network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" successfully" Jan 30 13:25:22.970076 containerd[1781]: time="2025-01-30T13:25:22.970054402Z" level=info msg="StopPodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" returns successfully" Jan 30 13:25:22.970558 containerd[1781]: time="2025-01-30T13:25:22.970535082Z" level=info msg="RemovePodSandbox for \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:25:22.970658 containerd[1781]: time="2025-01-30T13:25:22.970644242Z" level=info msg="Forcibly stopping sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\"" Jan 30 13:25:22.970819 containerd[1781]: time="2025-01-30T13:25:22.970799922Z" level=info msg="TearDown network for sandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" successfully" Jan 30 13:25:22.982862 containerd[1781]: time="2025-01-30T13:25:22.982822962Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:22.983144 containerd[1781]: time="2025-01-30T13:25:22.983103963Z" level=info msg="RemovePodSandbox \"8b2054dab8d29aca4d516cb76baa4d6fd5e3212a7d26c137b94f6121f021cea4\" returns successfully" Jan 30 13:25:22.983920 containerd[1781]: time="2025-01-30T13:25:22.983677883Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" Jan 30 13:25:22.983920 containerd[1781]: time="2025-01-30T13:25:22.983810483Z" level=info msg="TearDown network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" successfully" Jan 30 13:25:22.983920 containerd[1781]: time="2025-01-30T13:25:22.983822723Z" level=info msg="StopPodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" returns successfully" Jan 30 13:25:22.984586 containerd[1781]: time="2025-01-30T13:25:22.984544563Z" level=info msg="RemovePodSandbox for \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" Jan 30 13:25:22.984586 containerd[1781]: time="2025-01-30T13:25:22.984583483Z" level=info msg="Forcibly stopping sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\"" Jan 30 13:25:22.985083 containerd[1781]: time="2025-01-30T13:25:22.984991683Z" level=info msg="TearDown network for sandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" successfully" Jan 30 13:25:22.997898 containerd[1781]: time="2025-01-30T13:25:22.997835203Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:22.998015 containerd[1781]: time="2025-01-30T13:25:22.997918803Z" level=info msg="RemovePodSandbox \"77f52385d01367a3addfd0f38b9274fac7a72a9710048c906f9b22219eb4774b\" returns successfully" Jan 30 13:25:22.998662 containerd[1781]: time="2025-01-30T13:25:22.998494323Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\"" Jan 30 13:25:22.998662 containerd[1781]: time="2025-01-30T13:25:22.998603203Z" level=info msg="TearDown network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" successfully" Jan 30 13:25:22.998662 containerd[1781]: time="2025-01-30T13:25:22.998616243Z" level=info msg="StopPodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" returns successfully" Jan 30 13:25:22.999170 containerd[1781]: time="2025-01-30T13:25:22.998967884Z" level=info msg="RemovePodSandbox for \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\"" Jan 30 13:25:22.999170 containerd[1781]: time="2025-01-30T13:25:22.998997244Z" level=info msg="Forcibly stopping sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\"" Jan 30 13:25:22.999170 containerd[1781]: time="2025-01-30T13:25:22.999078524Z" level=info msg="TearDown network for sandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" successfully" Jan 30 13:25:23.007161 containerd[1781]: time="2025-01-30T13:25:23.006930364Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.007161 containerd[1781]: time="2025-01-30T13:25:23.007016484Z" level=info msg="RemovePodSandbox \"38fcf144fa9c0afbde811e06729c91359ced6569d800499a474bc610e9e52f39\" returns successfully" Jan 30 13:25:23.007691 containerd[1781]: time="2025-01-30T13:25:23.007493244Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\"" Jan 30 13:25:23.007691 containerd[1781]: time="2025-01-30T13:25:23.007604244Z" level=info msg="TearDown network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" successfully" Jan 30 13:25:23.007691 containerd[1781]: time="2025-01-30T13:25:23.007615124Z" level=info msg="StopPodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" returns successfully" Jan 30 13:25:23.008096 containerd[1781]: time="2025-01-30T13:25:23.008062284Z" level=info msg="RemovePodSandbox for \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\"" Jan 30 13:25:23.008152 containerd[1781]: time="2025-01-30T13:25:23.008115484Z" level=info msg="Forcibly stopping sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\"" Jan 30 13:25:23.008260 containerd[1781]: time="2025-01-30T13:25:23.008187004Z" level=info msg="TearDown network for sandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" successfully" Jan 30 13:25:23.019201 containerd[1781]: time="2025-01-30T13:25:23.019143565Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.019386 containerd[1781]: time="2025-01-30T13:25:23.019225525Z" level=info msg="RemovePodSandbox \"29a84bb4e60dc0c1fc2f6de676a66066c77b54cfc224ba110f321ff53eabbf67\" returns successfully" Jan 30 13:25:23.019897 containerd[1781]: time="2025-01-30T13:25:23.019771245Z" level=info msg="StopPodSandbox for \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\"" Jan 30 13:25:23.019897 containerd[1781]: time="2025-01-30T13:25:23.019872205Z" level=info msg="TearDown network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" successfully" Jan 30 13:25:23.020132 containerd[1781]: time="2025-01-30T13:25:23.019882125Z" level=info msg="StopPodSandbox for \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" returns successfully" Jan 30 13:25:23.020849 containerd[1781]: time="2025-01-30T13:25:23.020344525Z" level=info msg="RemovePodSandbox for \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\"" Jan 30 13:25:23.020849 containerd[1781]: time="2025-01-30T13:25:23.020371245Z" level=info msg="Forcibly stopping sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\"" Jan 30 13:25:23.020849 containerd[1781]: time="2025-01-30T13:25:23.020447885Z" level=info msg="TearDown network for sandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" successfully" Jan 30 13:25:23.027270 containerd[1781]: time="2025-01-30T13:25:23.027219085Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.027463 containerd[1781]: time="2025-01-30T13:25:23.027297845Z" level=info msg="RemovePodSandbox \"f25f558e8f6b95f91259f89336807750b0cbd589e025d45eadae3695b16ab813\" returns successfully" Jan 30 13:25:23.027935 containerd[1781]: time="2025-01-30T13:25:23.027753085Z" level=info msg="StopPodSandbox for \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\"" Jan 30 13:25:23.027935 containerd[1781]: time="2025-01-30T13:25:23.027849125Z" level=info msg="TearDown network for sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\" successfully" Jan 30 13:25:23.027935 containerd[1781]: time="2025-01-30T13:25:23.027859605Z" level=info msg="StopPodSandbox for \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\" returns successfully" Jan 30 13:25:23.028434 containerd[1781]: time="2025-01-30T13:25:23.028400925Z" level=info msg="RemovePodSandbox for \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\"" Jan 30 13:25:23.028697 containerd[1781]: time="2025-01-30T13:25:23.028543125Z" level=info msg="Forcibly stopping sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\"" Jan 30 13:25:23.028697 containerd[1781]: time="2025-01-30T13:25:23.028604725Z" level=info msg="TearDown network for sandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\" successfully" Jan 30 13:25:23.035582 containerd[1781]: time="2025-01-30T13:25:23.035517166Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.035787 containerd[1781]: time="2025-01-30T13:25:23.035606206Z" level=info msg="RemovePodSandbox \"8a0a8f637f55d1d8bb3130ee77dfca8b96c80faa09b7d4713a8a6c1d4c76030b\" returns successfully" Jan 30 13:25:23.036424 containerd[1781]: time="2025-01-30T13:25:23.036183766Z" level=info msg="StopPodSandbox for \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\"" Jan 30 13:25:23.036424 containerd[1781]: time="2025-01-30T13:25:23.036364606Z" level=info msg="TearDown network for sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\" successfully" Jan 30 13:25:23.036424 containerd[1781]: time="2025-01-30T13:25:23.036376606Z" level=info msg="StopPodSandbox for \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\" returns successfully" Jan 30 13:25:23.037299 containerd[1781]: time="2025-01-30T13:25:23.037218406Z" level=info msg="RemovePodSandbox for \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\"" Jan 30 13:25:23.037299 containerd[1781]: time="2025-01-30T13:25:23.037257286Z" level=info msg="Forcibly stopping sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\"" Jan 30 13:25:23.037415 containerd[1781]: time="2025-01-30T13:25:23.037338966Z" level=info msg="TearDown network for sandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\" successfully" Jan 30 13:25:23.044517 containerd[1781]: time="2025-01-30T13:25:23.044462006Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.044817 containerd[1781]: time="2025-01-30T13:25:23.044530806Z" level=info msg="RemovePodSandbox \"d36aaff25c46d5ff928f8544b78f0d2432ef36f5585a55021c77736e9dd2a10b\" returns successfully" Jan 30 13:25:23.045250 containerd[1781]: time="2025-01-30T13:25:23.045032926Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:25:23.045250 containerd[1781]: time="2025-01-30T13:25:23.045146886Z" level=info msg="TearDown network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" successfully" Jan 30 13:25:23.045250 containerd[1781]: time="2025-01-30T13:25:23.045158606Z" level=info msg="StopPodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" returns successfully" Jan 30 13:25:23.045977 containerd[1781]: time="2025-01-30T13:25:23.045777166Z" level=info msg="RemovePodSandbox for \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:25:23.045977 containerd[1781]: time="2025-01-30T13:25:23.045806566Z" level=info msg="Forcibly stopping sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\"" Jan 30 13:25:23.045977 containerd[1781]: time="2025-01-30T13:25:23.045899086Z" level=info msg="TearDown network for sandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" successfully" Jan 30 13:25:23.054932 containerd[1781]: time="2025-01-30T13:25:23.054882407Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.055323 containerd[1781]: time="2025-01-30T13:25:23.054962727Z" level=info msg="RemovePodSandbox \"7adf17e145f57eb54264a33c86322a9b54c4579a9ac505319477b4a391ba6112\" returns successfully" Jan 30 13:25:23.055366 containerd[1781]: time="2025-01-30T13:25:23.055346007Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" Jan 30 13:25:23.055532 containerd[1781]: time="2025-01-30T13:25:23.055501927Z" level=info msg="TearDown network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" successfully" Jan 30 13:25:23.055532 containerd[1781]: time="2025-01-30T13:25:23.055530127Z" level=info msg="StopPodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" returns successfully" Jan 30 13:25:23.056764 containerd[1781]: time="2025-01-30T13:25:23.055917967Z" level=info msg="RemovePodSandbox for \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" Jan 30 13:25:23.056764 containerd[1781]: time="2025-01-30T13:25:23.055945247Z" level=info msg="Forcibly stopping sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\"" Jan 30 13:25:23.056764 containerd[1781]: time="2025-01-30T13:25:23.056021487Z" level=info msg="TearDown network for sandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" successfully" Jan 30 13:25:23.062774 containerd[1781]: time="2025-01-30T13:25:23.062705527Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.062887 containerd[1781]: time="2025-01-30T13:25:23.062810527Z" level=info msg="RemovePodSandbox \"a980864318561e945fdeff425a32fd92342cd67c39017663358a1218709519e1\" returns successfully" Jan 30 13:25:23.063385 containerd[1781]: time="2025-01-30T13:25:23.063354168Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\"" Jan 30 13:25:23.063475 containerd[1781]: time="2025-01-30T13:25:23.063453768Z" level=info msg="TearDown network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" successfully" Jan 30 13:25:23.063475 containerd[1781]: time="2025-01-30T13:25:23.063469048Z" level=info msg="StopPodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" returns successfully" Jan 30 13:25:23.063883 containerd[1781]: time="2025-01-30T13:25:23.063861048Z" level=info msg="RemovePodSandbox for \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\"" Jan 30 13:25:23.064742 containerd[1781]: time="2025-01-30T13:25:23.063998128Z" level=info msg="Forcibly stopping sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\"" Jan 30 13:25:23.064742 containerd[1781]: time="2025-01-30T13:25:23.064093848Z" level=info msg="TearDown network for sandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" successfully" Jan 30 13:25:23.081389 containerd[1781]: time="2025-01-30T13:25:23.081343329Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.081652 containerd[1781]: time="2025-01-30T13:25:23.081631809Z" level=info msg="RemovePodSandbox \"5c1202e506970c973c8accfe948c2cfe6db28f2946fe2adf0d7122ad8f6cdf4e\" returns successfully" Jan 30 13:25:23.082311 containerd[1781]: time="2025-01-30T13:25:23.082254809Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\"" Jan 30 13:25:23.082456 containerd[1781]: time="2025-01-30T13:25:23.082354449Z" level=info msg="TearDown network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" successfully" Jan 30 13:25:23.082456 containerd[1781]: time="2025-01-30T13:25:23.082365129Z" level=info msg="StopPodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" returns successfully" Jan 30 13:25:23.082757 containerd[1781]: time="2025-01-30T13:25:23.082658649Z" level=info msg="RemovePodSandbox for \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\"" Jan 30 13:25:23.082757 containerd[1781]: time="2025-01-30T13:25:23.082684169Z" level=info msg="Forcibly stopping sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\"" Jan 30 13:25:23.083128 containerd[1781]: time="2025-01-30T13:25:23.083107849Z" level=info msg="TearDown network for sandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" successfully" Jan 30 13:25:23.109584 containerd[1781]: time="2025-01-30T13:25:23.109455130Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.109584 containerd[1781]: time="2025-01-30T13:25:23.109533370Z" level=info msg="RemovePodSandbox \"3f8f47000c211471ec1f9175ee7896b4ea09a1401b5783d7092388f539dcb6a2\" returns successfully" Jan 30 13:25:23.110175 containerd[1781]: time="2025-01-30T13:25:23.109978410Z" level=info msg="StopPodSandbox for \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\"" Jan 30 13:25:23.110175 containerd[1781]: time="2025-01-30T13:25:23.110087250Z" level=info msg="TearDown network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" successfully" Jan 30 13:25:23.110175 containerd[1781]: time="2025-01-30T13:25:23.110096970Z" level=info msg="StopPodSandbox for \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" returns successfully" Jan 30 13:25:23.110610 containerd[1781]: time="2025-01-30T13:25:23.110454290Z" level=info msg="RemovePodSandbox for \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\"" Jan 30 13:25:23.110610 containerd[1781]: time="2025-01-30T13:25:23.110506170Z" level=info msg="Forcibly stopping sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\"" Jan 30 13:25:23.110610 containerd[1781]: time="2025-01-30T13:25:23.110562530Z" level=info msg="TearDown network for sandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" successfully" Jan 30 13:25:23.121758 containerd[1781]: time="2025-01-30T13:25:23.121661851Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.121758 containerd[1781]: time="2025-01-30T13:25:23.121760771Z" level=info msg="RemovePodSandbox \"b7ea85d729ab74b26e5b361f02d7a85a59099d2b81c8912eaec15fe4edd4b9ed\" returns successfully" Jan 30 13:25:23.122240 containerd[1781]: time="2025-01-30T13:25:23.122142011Z" level=info msg="StopPodSandbox for \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\"" Jan 30 13:25:23.122484 containerd[1781]: time="2025-01-30T13:25:23.122269251Z" level=info msg="TearDown network for sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\" successfully" Jan 30 13:25:23.122484 containerd[1781]: time="2025-01-30T13:25:23.122281731Z" level=info msg="StopPodSandbox for \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\" returns successfully" Jan 30 13:25:23.122900 containerd[1781]: time="2025-01-30T13:25:23.122683291Z" level=info msg="RemovePodSandbox for \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\"" Jan 30 13:25:23.122900 containerd[1781]: time="2025-01-30T13:25:23.122736891Z" level=info msg="Forcibly stopping sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\"" Jan 30 13:25:23.123704 containerd[1781]: time="2025-01-30T13:25:23.122972371Z" level=info msg="TearDown network for sandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\" successfully" Jan 30 13:25:23.131614 containerd[1781]: time="2025-01-30T13:25:23.131565852Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.131730 containerd[1781]: time="2025-01-30T13:25:23.131638852Z" level=info msg="RemovePodSandbox \"edf67bf64d3bc80822d7b5fe3f22ec88f5a26bb2258595a88d895435e3860f46\" returns successfully" Jan 30 13:25:23.132343 containerd[1781]: time="2025-01-30T13:25:23.132118692Z" level=info msg="StopPodSandbox for \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\"" Jan 30 13:25:23.132343 containerd[1781]: time="2025-01-30T13:25:23.132222252Z" level=info msg="TearDown network for sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\" successfully" Jan 30 13:25:23.132343 containerd[1781]: time="2025-01-30T13:25:23.132232132Z" level=info msg="StopPodSandbox for \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\" returns successfully" Jan 30 13:25:23.133283 containerd[1781]: time="2025-01-30T13:25:23.132796212Z" level=info msg="RemovePodSandbox for \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\"" Jan 30 13:25:23.133283 containerd[1781]: time="2025-01-30T13:25:23.132825052Z" level=info msg="Forcibly stopping sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\"" Jan 30 13:25:23.133283 containerd[1781]: time="2025-01-30T13:25:23.132888812Z" level=info msg="TearDown network for sandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\" successfully" Jan 30 13:25:23.143883 containerd[1781]: time="2025-01-30T13:25:23.143741053Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.143883 containerd[1781]: time="2025-01-30T13:25:23.143825013Z" level=info msg="RemovePodSandbox \"fd418699504526bd8aa6b2262d0a989146c08fa46ae41aabfe1267ce3cc21b53\" returns successfully" Jan 30 13:25:23.145239 containerd[1781]: time="2025-01-30T13:25:23.144705853Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:25:23.145239 containerd[1781]: time="2025-01-30T13:25:23.144834733Z" level=info msg="TearDown network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" successfully" Jan 30 13:25:23.145239 containerd[1781]: time="2025-01-30T13:25:23.144845573Z" level=info msg="StopPodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" returns successfully" Jan 30 13:25:23.146203 containerd[1781]: time="2025-01-30T13:25:23.146112893Z" level=info msg="RemovePodSandbox for \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:25:23.146203 containerd[1781]: time="2025-01-30T13:25:23.146153373Z" level=info msg="Forcibly stopping sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\"" Jan 30 13:25:23.146331 containerd[1781]: time="2025-01-30T13:25:23.146232413Z" level=info msg="TearDown network for sandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" successfully" Jan 30 13:25:23.155052 containerd[1781]: time="2025-01-30T13:25:23.154978853Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.155052 containerd[1781]: time="2025-01-30T13:25:23.155053413Z" level=info msg="RemovePodSandbox \"c89ed979fd7542a9e1e5b945723492f709c3bb82cf2e617aaea12e82c665f289\" returns successfully" Jan 30 13:25:23.155827 containerd[1781]: time="2025-01-30T13:25:23.155612853Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" Jan 30 13:25:23.155827 containerd[1781]: time="2025-01-30T13:25:23.155744973Z" level=info msg="TearDown network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" successfully" Jan 30 13:25:23.155827 containerd[1781]: time="2025-01-30T13:25:23.155757053Z" level=info msg="StopPodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" returns successfully" Jan 30 13:25:23.156302 containerd[1781]: time="2025-01-30T13:25:23.156265973Z" level=info msg="RemovePodSandbox for \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" Jan 30 13:25:23.156302 containerd[1781]: time="2025-01-30T13:25:23.156299733Z" level=info msg="Forcibly stopping sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\"" Jan 30 13:25:23.156391 containerd[1781]: time="2025-01-30T13:25:23.156371853Z" level=info msg="TearDown network for sandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" successfully" Jan 30 13:25:23.164747 containerd[1781]: time="2025-01-30T13:25:23.164667934Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.164943 containerd[1781]: time="2025-01-30T13:25:23.164758094Z" level=info msg="RemovePodSandbox \"d9f0496c35f28aed5b26c6264821b85ef9bcc81a463262d5bec5fc470eebed3e\" returns successfully" Jan 30 13:25:23.165608 containerd[1781]: time="2025-01-30T13:25:23.165305814Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\"" Jan 30 13:25:23.165608 containerd[1781]: time="2025-01-30T13:25:23.165410814Z" level=info msg="TearDown network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" successfully" Jan 30 13:25:23.165608 containerd[1781]: time="2025-01-30T13:25:23.165420854Z" level=info msg="StopPodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" returns successfully" Jan 30 13:25:23.165919 containerd[1781]: time="2025-01-30T13:25:23.165846174Z" level=info msg="RemovePodSandbox for \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\"" Jan 30 13:25:23.165919 containerd[1781]: time="2025-01-30T13:25:23.165872254Z" level=info msg="Forcibly stopping sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\"" Jan 30 13:25:23.165965 containerd[1781]: time="2025-01-30T13:25:23.165947054Z" level=info msg="TearDown network for sandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" successfully" Jan 30 13:25:23.174065 containerd[1781]: time="2025-01-30T13:25:23.174003614Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.174447 containerd[1781]: time="2025-01-30T13:25:23.174079854Z" level=info msg="RemovePodSandbox \"7bbce0f670a07375721625374fa076ab7f5e8ba7d31d9eafbef8aaab956a3b50\" returns successfully" Jan 30 13:25:23.174902 containerd[1781]: time="2025-01-30T13:25:23.174687774Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\"" Jan 30 13:25:23.174902 containerd[1781]: time="2025-01-30T13:25:23.174794494Z" level=info msg="TearDown network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" successfully" Jan 30 13:25:23.174902 containerd[1781]: time="2025-01-30T13:25:23.174806854Z" level=info msg="StopPodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" returns successfully" Jan 30 13:25:23.175783 containerd[1781]: time="2025-01-30T13:25:23.175223335Z" level=info msg="RemovePodSandbox for \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\"" Jan 30 13:25:23.175783 containerd[1781]: time="2025-01-30T13:25:23.175319495Z" level=info msg="Forcibly stopping sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\"" Jan 30 13:25:23.175783 containerd[1781]: time="2025-01-30T13:25:23.175399975Z" level=info msg="TearDown network for sandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" successfully" Jan 30 13:25:23.185145 containerd[1781]: time="2025-01-30T13:25:23.184989055Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.185145 containerd[1781]: time="2025-01-30T13:25:23.185059175Z" level=info msg="RemovePodSandbox \"5ec336d95cf25c043787d71b6d774a7f1932ae0c31e6dc192e9d825be7d7ff8f\" returns successfully" Jan 30 13:25:23.185774 containerd[1781]: time="2025-01-30T13:25:23.185470815Z" level=info msg="StopPodSandbox for \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\"" Jan 30 13:25:23.185774 containerd[1781]: time="2025-01-30T13:25:23.185571855Z" level=info msg="TearDown network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" successfully" Jan 30 13:25:23.185774 containerd[1781]: time="2025-01-30T13:25:23.185580735Z" level=info msg="StopPodSandbox for \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" returns successfully" Jan 30 13:25:23.185994 containerd[1781]: time="2025-01-30T13:25:23.185972695Z" level=info msg="RemovePodSandbox for \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\"" Jan 30 13:25:23.186069 containerd[1781]: time="2025-01-30T13:25:23.186056255Z" level=info msg="Forcibly stopping sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\"" Jan 30 13:25:23.186226 containerd[1781]: time="2025-01-30T13:25:23.186194735Z" level=info msg="TearDown network for sandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" successfully" Jan 30 13:25:23.198517 containerd[1781]: time="2025-01-30T13:25:23.198473496Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.198859 containerd[1781]: time="2025-01-30T13:25:23.198709696Z" level=info msg="RemovePodSandbox \"382b13d1fdc60151569ad8ec29d006906441766e6ee681eaea101d939df63119\" returns successfully" Jan 30 13:25:23.199518 containerd[1781]: time="2025-01-30T13:25:23.199352816Z" level=info msg="StopPodSandbox for \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\"" Jan 30 13:25:23.199518 containerd[1781]: time="2025-01-30T13:25:23.199448376Z" level=info msg="TearDown network for sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\" successfully" Jan 30 13:25:23.199518 containerd[1781]: time="2025-01-30T13:25:23.199459136Z" level=info msg="StopPodSandbox for \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\" returns successfully" Jan 30 13:25:23.199804 containerd[1781]: time="2025-01-30T13:25:23.199784016Z" level=info msg="RemovePodSandbox for \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\"" Jan 30 13:25:23.199905 containerd[1781]: time="2025-01-30T13:25:23.199877856Z" level=info msg="Forcibly stopping sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\"" Jan 30 13:25:23.200159 containerd[1781]: time="2025-01-30T13:25:23.200023936Z" level=info msg="TearDown network for sandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\" successfully" Jan 30 13:25:23.209818 containerd[1781]: time="2025-01-30T13:25:23.209638057Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.209818 containerd[1781]: time="2025-01-30T13:25:23.209705817Z" level=info msg="RemovePodSandbox \"76c8f20b6927a6f5553fbc7f93874cbaf2beaf4938bae5769b600a89999b18ad\" returns successfully" Jan 30 13:25:23.210692 containerd[1781]: time="2025-01-30T13:25:23.210356057Z" level=info msg="StopPodSandbox for \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\"" Jan 30 13:25:23.210692 containerd[1781]: time="2025-01-30T13:25:23.210457177Z" level=info msg="TearDown network for sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\" successfully" Jan 30 13:25:23.210692 containerd[1781]: time="2025-01-30T13:25:23.210467377Z" level=info msg="StopPodSandbox for \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\" returns successfully" Jan 30 13:25:23.211024 containerd[1781]: time="2025-01-30T13:25:23.211002097Z" level=info msg="RemovePodSandbox for \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\"" Jan 30 13:25:23.211105 containerd[1781]: time="2025-01-30T13:25:23.211092537Z" level=info msg="Forcibly stopping sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\"" Jan 30 13:25:23.211739 containerd[1781]: time="2025-01-30T13:25:23.211206057Z" level=info msg="TearDown network for sandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\" successfully" Jan 30 13:25:23.223830 containerd[1781]: time="2025-01-30T13:25:23.223695498Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.224072 containerd[1781]: time="2025-01-30T13:25:23.223859738Z" level=info msg="RemovePodSandbox \"3b8b584ca7e2ce763200e658cc53f46876714cdf961cce79347efb13c8e4d494\" returns successfully" Jan 30 13:25:23.224828 containerd[1781]: time="2025-01-30T13:25:23.224707458Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:25:23.225494 containerd[1781]: time="2025-01-30T13:25:23.225059698Z" level=info msg="TearDown network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" successfully" Jan 30 13:25:23.225494 containerd[1781]: time="2025-01-30T13:25:23.225076058Z" level=info msg="StopPodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" returns successfully" Jan 30 13:25:23.228366 containerd[1781]: time="2025-01-30T13:25:23.227898058Z" level=info msg="RemovePodSandbox for \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:25:23.228366 containerd[1781]: time="2025-01-30T13:25:23.227948258Z" level=info msg="Forcibly stopping sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\"" Jan 30 13:25:23.228902 containerd[1781]: time="2025-01-30T13:25:23.228671218Z" level=info msg="TearDown network for sandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" successfully" Jan 30 13:25:23.332727 containerd[1781]: time="2025-01-30T13:25:23.332558064Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.332727 containerd[1781]: time="2025-01-30T13:25:23.332641024Z" level=info msg="RemovePodSandbox \"7b8e47ffb6e1b16f4f561028eeb0349ceaae83945c9be3ce2422c8d49cd3ce14\" returns successfully" Jan 30 13:25:23.333626 containerd[1781]: time="2025-01-30T13:25:23.333324384Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" Jan 30 13:25:23.333626 containerd[1781]: time="2025-01-30T13:25:23.333440424Z" level=info msg="TearDown network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" successfully" Jan 30 13:25:23.333626 containerd[1781]: time="2025-01-30T13:25:23.333450264Z" level=info msg="StopPodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" returns successfully" Jan 30 13:25:23.333902 containerd[1781]: time="2025-01-30T13:25:23.333805384Z" level=info msg="RemovePodSandbox for \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" Jan 30 13:25:23.333902 containerd[1781]: time="2025-01-30T13:25:23.333832664Z" level=info msg="Forcibly stopping sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\"" Jan 30 13:25:23.333951 containerd[1781]: time="2025-01-30T13:25:23.333907064Z" level=info msg="TearDown network for sandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" successfully" Jan 30 13:25:23.343396 containerd[1781]: time="2025-01-30T13:25:23.343338865Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.343633 containerd[1781]: time="2025-01-30T13:25:23.343420065Z" level=info msg="RemovePodSandbox \"34709336279b488949da2b889506898201c9f8b9f2e9b3535971efd9e0da0203\" returns successfully" Jan 30 13:25:23.343972 containerd[1781]: time="2025-01-30T13:25:23.343943545Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\"" Jan 30 13:25:23.344062 containerd[1781]: time="2025-01-30T13:25:23.344042905Z" level=info msg="TearDown network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" successfully" Jan 30 13:25:23.344062 containerd[1781]: time="2025-01-30T13:25:23.344058065Z" level=info msg="StopPodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" returns successfully" Jan 30 13:25:23.344492 containerd[1781]: time="2025-01-30T13:25:23.344467745Z" level=info msg="RemovePodSandbox for \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\"" Jan 30 13:25:23.344548 containerd[1781]: time="2025-01-30T13:25:23.344497025Z" level=info msg="Forcibly stopping sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\"" Jan 30 13:25:23.344604 containerd[1781]: time="2025-01-30T13:25:23.344583985Z" level=info msg="TearDown network for sandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" successfully" Jan 30 13:25:23.354363 containerd[1781]: time="2025-01-30T13:25:23.354296146Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.354363 containerd[1781]: time="2025-01-30T13:25:23.354364746Z" level=info msg="RemovePodSandbox \"a89c7688625968566538974dea8116230c5ee2bab6778d0e4931c69ca9578ecf\" returns successfully" Jan 30 13:25:23.355110 containerd[1781]: time="2025-01-30T13:25:23.355083506Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\"" Jan 30 13:25:23.355210 containerd[1781]: time="2025-01-30T13:25:23.355188626Z" level=info msg="TearDown network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" successfully" Jan 30 13:25:23.355210 containerd[1781]: time="2025-01-30T13:25:23.355204786Z" level=info msg="StopPodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" returns successfully" Jan 30 13:25:23.355664 containerd[1781]: time="2025-01-30T13:25:23.355635146Z" level=info msg="RemovePodSandbox for \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\"" Jan 30 13:25:23.355799 containerd[1781]: time="2025-01-30T13:25:23.355668426Z" level=info msg="Forcibly stopping sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\"" Jan 30 13:25:23.355833 containerd[1781]: time="2025-01-30T13:25:23.355793426Z" level=info msg="TearDown network for sandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" successfully" Jan 30 13:25:23.364828 containerd[1781]: time="2025-01-30T13:25:23.364776506Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.364979 containerd[1781]: time="2025-01-30T13:25:23.364847186Z" level=info msg="RemovePodSandbox \"a6caa6ec61be59485e56cecfaf9768f89ef6925a78ae0e8d04b2fe6d676e9f1e\" returns successfully" Jan 30 13:25:23.365531 containerd[1781]: time="2025-01-30T13:25:23.365338626Z" level=info msg="StopPodSandbox for \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\"" Jan 30 13:25:23.365531 containerd[1781]: time="2025-01-30T13:25:23.365464826Z" level=info msg="TearDown network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" successfully" Jan 30 13:25:23.365531 containerd[1781]: time="2025-01-30T13:25:23.365475426Z" level=info msg="StopPodSandbox for \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" returns successfully" Jan 30 13:25:23.365960 containerd[1781]: time="2025-01-30T13:25:23.365898706Z" level=info msg="RemovePodSandbox for \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\"" Jan 30 13:25:23.366024 containerd[1781]: time="2025-01-30T13:25:23.365963906Z" level=info msg="Forcibly stopping sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\"" Jan 30 13:25:23.366048 containerd[1781]: time="2025-01-30T13:25:23.366027746Z" level=info msg="TearDown network for sandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" successfully" Jan 30 13:25:23.374956 containerd[1781]: time="2025-01-30T13:25:23.374897507Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.375145 containerd[1781]: time="2025-01-30T13:25:23.374967027Z" level=info msg="RemovePodSandbox \"e669011e3dae67afc16ea206029a7347bd53dd0c823a64b638a98e7757a30d84\" returns successfully" Jan 30 13:25:23.375759 containerd[1781]: time="2025-01-30T13:25:23.375489947Z" level=info msg="StopPodSandbox for \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\"" Jan 30 13:25:23.375759 containerd[1781]: time="2025-01-30T13:25:23.375602387Z" level=info msg="TearDown network for sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\" successfully" Jan 30 13:25:23.375759 containerd[1781]: time="2025-01-30T13:25:23.375614307Z" level=info msg="StopPodSandbox for \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\" returns successfully" Jan 30 13:25:23.376078 containerd[1781]: time="2025-01-30T13:25:23.376050747Z" level=info msg="RemovePodSandbox for \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\"" Jan 30 13:25:23.376118 containerd[1781]: time="2025-01-30T13:25:23.376083427Z" level=info msg="Forcibly stopping sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\"" Jan 30 13:25:23.376170 containerd[1781]: time="2025-01-30T13:25:23.376150667Z" level=info msg="TearDown network for sandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\" successfully" Jan 30 13:25:23.385210 containerd[1781]: time="2025-01-30T13:25:23.385154388Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.385600 containerd[1781]: time="2025-01-30T13:25:23.385221308Z" level=info msg="RemovePodSandbox \"669dc0bba6d29130cb00a007b66465ec9ebaede0d7b59a9c6db163626e662c3e\" returns successfully" Jan 30 13:25:23.386019 containerd[1781]: time="2025-01-30T13:25:23.385836308Z" level=info msg="StopPodSandbox for \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\"" Jan 30 13:25:23.386019 containerd[1781]: time="2025-01-30T13:25:23.385940268Z" level=info msg="TearDown network for sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\" successfully" Jan 30 13:25:23.386019 containerd[1781]: time="2025-01-30T13:25:23.385949348Z" level=info msg="StopPodSandbox for \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\" returns successfully" Jan 30 13:25:23.386266 containerd[1781]: time="2025-01-30T13:25:23.386212828Z" level=info msg="RemovePodSandbox for \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\"" Jan 30 13:25:23.386266 containerd[1781]: time="2025-01-30T13:25:23.386240468Z" level=info msg="Forcibly stopping sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\"" Jan 30 13:25:23.386336 containerd[1781]: time="2025-01-30T13:25:23.386300868Z" level=info msg="TearDown network for sandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\" successfully" Jan 30 13:25:23.396088 containerd[1781]: time="2025-01-30T13:25:23.395950428Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.396088 containerd[1781]: time="2025-01-30T13:25:23.396021668Z" level=info msg="RemovePodSandbox \"327536bf350f5985f6f252f8cf1582345477b75490ce7e1f026e8fb398fdde1e\" returns successfully" Jan 30 13:25:23.396996 containerd[1781]: time="2025-01-30T13:25:23.396598868Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:25:23.396996 containerd[1781]: time="2025-01-30T13:25:23.396693868Z" level=info msg="TearDown network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" successfully" Jan 30 13:25:23.396996 containerd[1781]: time="2025-01-30T13:25:23.396703628Z" level=info msg="StopPodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" returns successfully" Jan 30 13:25:23.397565 containerd[1781]: time="2025-01-30T13:25:23.397467028Z" level=info msg="RemovePodSandbox for \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:25:23.397565 containerd[1781]: time="2025-01-30T13:25:23.397522948Z" level=info msg="Forcibly stopping sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\"" Jan 30 13:25:23.397815 containerd[1781]: time="2025-01-30T13:25:23.397710068Z" level=info msg="TearDown network for sandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" successfully" Jan 30 13:25:23.405165 containerd[1781]: time="2025-01-30T13:25:23.405105509Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.405165 containerd[1781]: time="2025-01-30T13:25:23.405171789Z" level=info msg="RemovePodSandbox \"2482422c2c23869fcb90a04bafb3deb0d2bd44433b27c367939aed904af28f2e\" returns successfully" Jan 30 13:25:23.405674 containerd[1781]: time="2025-01-30T13:25:23.405617309Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" Jan 30 13:25:23.405770 containerd[1781]: time="2025-01-30T13:25:23.405741749Z" level=info msg="TearDown network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" successfully" Jan 30 13:25:23.405770 containerd[1781]: time="2025-01-30T13:25:23.405754069Z" level=info msg="StopPodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" returns successfully" Jan 30 13:25:23.406585 containerd[1781]: time="2025-01-30T13:25:23.406116829Z" level=info msg="RemovePodSandbox for \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" Jan 30 13:25:23.406585 containerd[1781]: time="2025-01-30T13:25:23.406139589Z" level=info msg="Forcibly stopping sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\"" Jan 30 13:25:23.406585 containerd[1781]: time="2025-01-30T13:25:23.406197429Z" level=info msg="TearDown network for sandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" successfully" Jan 30 13:25:23.413198 containerd[1781]: time="2025-01-30T13:25:23.413141309Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.413367 containerd[1781]: time="2025-01-30T13:25:23.413206629Z" level=info msg="RemovePodSandbox \"6507ab4978e1c90608afc313455d21951605e7b917eab66e2b453cc6f5d90895\" returns successfully" Jan 30 13:25:23.414045 containerd[1781]: time="2025-01-30T13:25:23.413899989Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\"" Jan 30 13:25:23.414045 containerd[1781]: time="2025-01-30T13:25:23.413998229Z" level=info msg="TearDown network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" successfully" Jan 30 13:25:23.414045 containerd[1781]: time="2025-01-30T13:25:23.414007309Z" level=info msg="StopPodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" returns successfully" Jan 30 13:25:23.414443 containerd[1781]: time="2025-01-30T13:25:23.414412269Z" level=info msg="RemovePodSandbox for \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\"" Jan 30 13:25:23.414495 containerd[1781]: time="2025-01-30T13:25:23.414446989Z" level=info msg="Forcibly stopping sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\"" Jan 30 13:25:23.414554 containerd[1781]: time="2025-01-30T13:25:23.414533069Z" level=info msg="TearDown network for sandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" successfully" Jan 30 13:25:23.422936 containerd[1781]: time="2025-01-30T13:25:23.422873430Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.422936 containerd[1781]: time="2025-01-30T13:25:23.422934430Z" level=info msg="RemovePodSandbox \"d4479af4d82bc456794289c2b3d018dcb852249d6c9951a306af946440d4a956\" returns successfully" Jan 30 13:25:23.423927 containerd[1781]: time="2025-01-30T13:25:23.423701110Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\"" Jan 30 13:25:23.423927 containerd[1781]: time="2025-01-30T13:25:23.423812550Z" level=info msg="TearDown network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" successfully" Jan 30 13:25:23.423927 containerd[1781]: time="2025-01-30T13:25:23.423823430Z" level=info msg="StopPodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" returns successfully" Jan 30 13:25:23.425517 containerd[1781]: time="2025-01-30T13:25:23.424337790Z" level=info msg="RemovePodSandbox for \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\"" Jan 30 13:25:23.425517 containerd[1781]: time="2025-01-30T13:25:23.424361190Z" level=info msg="Forcibly stopping sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\"" Jan 30 13:25:23.425517 containerd[1781]: time="2025-01-30T13:25:23.424427630Z" level=info msg="TearDown network for sandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" successfully" Jan 30 13:25:23.433745 containerd[1781]: time="2025-01-30T13:25:23.433622711Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.433745 containerd[1781]: time="2025-01-30T13:25:23.433710031Z" level=info msg="RemovePodSandbox \"00f080f9072191b4833cab85c8d47802f330fb9d745d15e200beb64738b79f0e\" returns successfully" Jan 30 13:25:23.434635 containerd[1781]: time="2025-01-30T13:25:23.434577471Z" level=info msg="StopPodSandbox for \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\"" Jan 30 13:25:23.434761 containerd[1781]: time="2025-01-30T13:25:23.434701511Z" level=info msg="TearDown network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" successfully" Jan 30 13:25:23.434761 containerd[1781]: time="2025-01-30T13:25:23.434732871Z" level=info msg="StopPodSandbox for \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" returns successfully" Jan 30 13:25:23.436730 containerd[1781]: time="2025-01-30T13:25:23.436671671Z" level=info msg="RemovePodSandbox for \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\"" Jan 30 13:25:23.436817 containerd[1781]: time="2025-01-30T13:25:23.436711271Z" level=info msg="Forcibly stopping sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\"" Jan 30 13:25:23.436843 containerd[1781]: time="2025-01-30T13:25:23.436816111Z" level=info msg="TearDown network for sandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" successfully" Jan 30 13:25:23.444262 containerd[1781]: time="2025-01-30T13:25:23.444113351Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.444262 containerd[1781]: time="2025-01-30T13:25:23.444195871Z" level=info msg="RemovePodSandbox \"efdc46b11748be0a53f4ddb570f55c80a3f09dad429d390083d50ceafc1070da\" returns successfully" Jan 30 13:25:23.445630 containerd[1781]: time="2025-01-30T13:25:23.445539151Z" level=info msg="StopPodSandbox for \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\"" Jan 30 13:25:23.446050 containerd[1781]: time="2025-01-30T13:25:23.445679231Z" level=info msg="TearDown network for sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\" successfully" Jan 30 13:25:23.446050 containerd[1781]: time="2025-01-30T13:25:23.445690231Z" level=info msg="StopPodSandbox for \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\" returns successfully" Jan 30 13:25:23.446181 containerd[1781]: time="2025-01-30T13:25:23.446105671Z" level=info msg="RemovePodSandbox for \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\"" Jan 30 13:25:23.446181 containerd[1781]: time="2025-01-30T13:25:23.446125591Z" level=info msg="Forcibly stopping sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\"" Jan 30 13:25:23.446360 containerd[1781]: time="2025-01-30T13:25:23.446308711Z" level=info msg="TearDown network for sandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\" successfully" Jan 30 13:25:23.455806 containerd[1781]: time="2025-01-30T13:25:23.455757272Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.455949 containerd[1781]: time="2025-01-30T13:25:23.455884592Z" level=info msg="RemovePodSandbox \"329f12f86171e85183c941946ca84b4293198377531bdad03c80321a4f216ed4\" returns successfully" Jan 30 13:25:23.456474 containerd[1781]: time="2025-01-30T13:25:23.456436472Z" level=info msg="StopPodSandbox for \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\"" Jan 30 13:25:23.456555 containerd[1781]: time="2025-01-30T13:25:23.456541592Z" level=info msg="TearDown network for sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\" successfully" Jan 30 13:25:23.456706 containerd[1781]: time="2025-01-30T13:25:23.456551992Z" level=info msg="StopPodSandbox for \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\" returns successfully" Jan 30 13:25:23.457332 containerd[1781]: time="2025-01-30T13:25:23.457150872Z" level=info msg="RemovePodSandbox for \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\"" Jan 30 13:25:23.457332 containerd[1781]: time="2025-01-30T13:25:23.457179672Z" level=info msg="Forcibly stopping sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\"" Jan 30 13:25:23.457332 containerd[1781]: time="2025-01-30T13:25:23.457251512Z" level=info msg="TearDown network for sandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\" successfully" Jan 30 13:25:23.476901 containerd[1781]: time="2025-01-30T13:25:23.476837313Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:25:23.477090 containerd[1781]: time="2025-01-30T13:25:23.476910993Z" level=info msg="RemovePodSandbox \"73aa0b53a38a4516f695669f3d515f0faac461b95f52fc59f04acb57dfe8b8d4\" returns successfully" Jan 30 13:25:24.891843 containerd[1781]: time="2025-01-30T13:25:24.891791442Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:24.893886 containerd[1781]: time="2025-01-30T13:25:24.893801082Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/kube-controllers:v3.29.1: active requests=0, bytes read=31953828" Jan 30 13:25:24.898603 containerd[1781]: time="2025-01-30T13:25:24.898540602Z" level=info msg="ImageCreate event name:\"sha256:32c335fdb9d757e7ba6a76a9cfa8d292a5a229101ae7ea37b42f53c28adf2db1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:24.903357 containerd[1781]: time="2025-01-30T13:25:24.903297242Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:24.904471 containerd[1781]: time="2025-01-30T13:25:24.903963122Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" with image id \"sha256:32c335fdb9d757e7ba6a76a9cfa8d292a5a229101ae7ea37b42f53c28adf2db1\", repo tag \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/kube-controllers@sha256:1072d6a98167a14ca361e9ce757733f9bae36d1f1c6a9621ea10934b6b1e10d9\", size \"33323450\" in 4.406797155s" Jan 30 13:25:24.904471 containerd[1781]: time="2025-01-30T13:25:24.904001882Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/kube-controllers:v3.29.1\" returns image reference \"sha256:32c335fdb9d757e7ba6a76a9cfa8d292a5a229101ae7ea37b42f53c28adf2db1\"" Jan 30 13:25:24.905829 containerd[1781]: time="2025-01-30T13:25:24.905792363Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Jan 30 13:25:24.920563 containerd[1781]: time="2025-01-30T13:25:24.920141763Z" level=info msg="CreateContainer within sandbox \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Jan 30 13:25:24.951069 containerd[1781]: time="2025-01-30T13:25:24.951013165Z" level=info msg="CreateContainer within sandbox \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2\"" Jan 30 13:25:24.951889 containerd[1781]: time="2025-01-30T13:25:24.951849005Z" level=info msg="StartContainer for \"5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2\"" Jan 30 13:25:24.986927 systemd[1]: Started cri-containerd-5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2.scope - libcontainer container 5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2. Jan 30 13:25:25.025114 containerd[1781]: time="2025-01-30T13:25:25.025039370Z" level=info msg="StartContainer for \"5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2\" returns successfully" Jan 30 13:25:25.731250 kubelet[3455]: I0130 13:25:25.731132 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="kube-system/coredns-7db6d8ff4d-5djx9" podStartSLOduration=49.731114534 podStartE2EDuration="49.731114534s" podCreationTimestamp="2025-01-30 13:24:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:25:20.70346898 +0000 UTC m=+57.933065967" watchObservedRunningTime="2025-01-30 13:25:25.731114534 +0000 UTC m=+62.960711521" Jan 30 13:25:25.732854 kubelet[3455]: I0130 13:25:25.732691 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-6dd45f4c57-58wrn" podStartSLOduration=35.899731427 podStartE2EDuration="41.732664694s" podCreationTimestamp="2025-01-30 13:24:44 +0000 UTC" firstStartedPulling="2025-01-30 13:25:19.071970376 +0000 UTC m=+56.301567363" lastFinishedPulling="2025-01-30 13:25:24.904903643 +0000 UTC m=+62.134500630" observedRunningTime="2025-01-30 13:25:25.731059894 +0000 UTC m=+62.960656881" watchObservedRunningTime="2025-01-30 13:25:25.732664694 +0000 UTC m=+62.962261681" Jan 30 13:25:27.554841 containerd[1781]: time="2025-01-30T13:25:27.554188150Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:27.556556 containerd[1781]: time="2025-01-30T13:25:27.556382590Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=39298409" Jan 30 13:25:27.559711 containerd[1781]: time="2025-01-30T13:25:27.559638190Z" level=info msg="ImageCreate event name:\"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:27.564526 containerd[1781]: time="2025-01-30T13:25:27.564441070Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:27.565305 containerd[1781]: time="2025-01-30T13:25:27.565143070Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"40668079\" in 2.659313387s" Jan 30 13:25:27.565305 containerd[1781]: time="2025-01-30T13:25:27.565171910Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\"" Jan 30 13:25:27.566730 containerd[1781]: time="2025-01-30T13:25:27.566695630Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\"" Jan 30 13:25:27.568233 containerd[1781]: time="2025-01-30T13:25:27.568208310Z" level=info msg="CreateContainer within sandbox \"2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jan 30 13:25:27.597634 containerd[1781]: time="2025-01-30T13:25:27.597583469Z" level=info msg="CreateContainer within sandbox \"2990378b79f6ca6d7d8d320b41addebcee950d6f6951ce9d9a42e8cf315a4275\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"88df5d65fd97c4db0c32584154798fd2041840141c8751ab56609d15d5f69ec5\"" Jan 30 13:25:27.599958 containerd[1781]: time="2025-01-30T13:25:27.599846429Z" level=info msg="StartContainer for \"88df5d65fd97c4db0c32584154798fd2041840141c8751ab56609d15d5f69ec5\"" Jan 30 13:25:27.634950 systemd[1]: Started cri-containerd-88df5d65fd97c4db0c32584154798fd2041840141c8751ab56609d15d5f69ec5.scope - libcontainer container 88df5d65fd97c4db0c32584154798fd2041840141c8751ab56609d15d5f69ec5. Jan 30 13:25:27.671235 containerd[1781]: time="2025-01-30T13:25:27.671094429Z" level=info msg="StartContainer for \"88df5d65fd97c4db0c32584154798fd2041840141c8751ab56609d15d5f69ec5\" returns successfully" Jan 30 13:25:27.877823 containerd[1781]: time="2025-01-30T13:25:27.877442947Z" level=info msg="ImageUpdate event name:\"ghcr.io/flatcar/calico/apiserver:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:27.882312 containerd[1781]: time="2025-01-30T13:25:27.882250147Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/apiserver:v3.29.1: active requests=0, bytes read=77" Jan 30 13:25:27.884504 containerd[1781]: time="2025-01-30T13:25:27.884190587Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" with image id \"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\", repo tag \"ghcr.io/flatcar/calico/apiserver:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/apiserver@sha256:b8c43e264fe52e0c327b0bf3ac882a0224b33bdd7f4ff58a74242da7d9b00486\", size \"40668079\" in 317.320877ms" Jan 30 13:25:27.884504 containerd[1781]: time="2025-01-30T13:25:27.884233107Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/apiserver:v3.29.1\" returns image reference \"sha256:5451b31bd8d0784796fa1204c4ec22975a270e21feadf2c5095fe41a38524c6c\"" Jan 30 13:25:27.886165 containerd[1781]: time="2025-01-30T13:25:27.885976787Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\"" Jan 30 13:25:27.887219 containerd[1781]: time="2025-01-30T13:25:27.886951987Z" level=info msg="CreateContainer within sandbox \"a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb\" for container &ContainerMetadata{Name:calico-apiserver,Attempt:0,}" Jan 30 13:25:27.924586 containerd[1781]: time="2025-01-30T13:25:27.924534347Z" level=info msg="CreateContainer within sandbox \"a36f6e10d2ee06ab358ef4865d283a4e3cccb90a3eea85d3fc1662f50781b8fb\" for &ContainerMetadata{Name:calico-apiserver,Attempt:0,} returns container id \"487555f555cae07c5095def378fe5c0e235906a2c62ca5570232c08b2a3a2c0c\"" Jan 30 13:25:27.927776 containerd[1781]: time="2025-01-30T13:25:27.925598547Z" level=info msg="StartContainer for \"487555f555cae07c5095def378fe5c0e235906a2c62ca5570232c08b2a3a2c0c\"" Jan 30 13:25:27.957929 systemd[1]: Started cri-containerd-487555f555cae07c5095def378fe5c0e235906a2c62ca5570232c08b2a3a2c0c.scope - libcontainer container 487555f555cae07c5095def378fe5c0e235906a2c62ca5570232c08b2a3a2c0c. Jan 30 13:25:28.004040 containerd[1781]: time="2025-01-30T13:25:28.003994746Z" level=info msg="StartContainer for \"487555f555cae07c5095def378fe5c0e235906a2c62ca5570232c08b2a3a2c0c\" returns successfully" Jan 30 13:25:28.727357 kubelet[3455]: I0130 13:25:28.726081 3455 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 30 13:25:28.747868 kubelet[3455]: I0130 13:25:28.747811 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-c978d777-x5x6w" podStartSLOduration=35.670028374 podStartE2EDuration="43.74779038s" podCreationTimestamp="2025-01-30 13:24:45 +0000 UTC" firstStartedPulling="2025-01-30 13:25:19.488815704 +0000 UTC m=+56.718412691" lastFinishedPulling="2025-01-30 13:25:27.56657771 +0000 UTC m=+64.796174697" observedRunningTime="2025-01-30 13:25:27.734960868 +0000 UTC m=+64.964557855" watchObservedRunningTime="2025-01-30 13:25:28.74779038 +0000 UTC m=+65.977387367" Jan 30 13:25:28.849612 kubelet[3455]: I0130 13:25:28.848999 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-apiserver/calico-apiserver-c978d777-jmpf2" podStartSLOduration=35.712263353 podStartE2EDuration="43.848979179s" podCreationTimestamp="2025-01-30 13:24:45 +0000 UTC" firstStartedPulling="2025-01-30 13:25:19.748431481 +0000 UTC m=+56.978028468" lastFinishedPulling="2025-01-30 13:25:27.885147307 +0000 UTC m=+65.114744294" observedRunningTime="2025-01-30 13:25:28.74884166 +0000 UTC m=+65.978438647" watchObservedRunningTime="2025-01-30 13:25:28.848979179 +0000 UTC m=+66.078576166" Jan 30 13:25:30.023947 containerd[1781]: time="2025-01-30T13:25:30.023884609Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:30.027539 containerd[1781]: time="2025-01-30T13:25:30.027478689Z" level=info msg="stop pulling image ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1: active requests=0, bytes read=9883368" Jan 30 13:25:30.032757 containerd[1781]: time="2025-01-30T13:25:30.032138929Z" level=info msg="ImageCreate event name:\"sha256:3eb557f7694f230afd24a75a691bcda4c0a7bfe87a981386dcd4ecf2b0701349\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:30.038680 containerd[1781]: time="2025-01-30T13:25:30.038620289Z" level=info msg="ImageCreate event name:\"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\" labels:{key:\"io.cri-containerd.image\" value:\"managed\"}" Jan 30 13:25:30.040673 containerd[1781]: time="2025-01-30T13:25:30.040619889Z" level=info msg="Pulled image \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" with image id \"sha256:3eb557f7694f230afd24a75a691bcda4c0a7bfe87a981386dcd4ecf2b0701349\", repo tag \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\", repo digest \"ghcr.io/flatcar/calico/node-driver-registrar@sha256:a338da9488cbaa83c78457c3d7354d84149969c0480e88dd768e036632ff5b76\", size \"11252974\" in 2.154606182s" Jan 30 13:25:30.040970 containerd[1781]: time="2025-01-30T13:25:30.040901809Z" level=info msg="PullImage \"ghcr.io/flatcar/calico/node-driver-registrar:v3.29.1\" returns image reference \"sha256:3eb557f7694f230afd24a75a691bcda4c0a7bfe87a981386dcd4ecf2b0701349\"" Jan 30 13:25:30.044553 containerd[1781]: time="2025-01-30T13:25:30.044506209Z" level=info msg="CreateContainer within sandbox \"1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad\" for container &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,}" Jan 30 13:25:30.076577 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2370729190.mount: Deactivated successfully. Jan 30 13:25:30.084548 containerd[1781]: time="2025-01-30T13:25:30.084064208Z" level=info msg="CreateContainer within sandbox \"1128e5c490baba2f8d54536d30c0f21c10063816c104af54bf5589d21d41caad\" for &ContainerMetadata{Name:csi-node-driver-registrar,Attempt:0,} returns container id \"bf5424f671d5cbbe138256e550d6318802c04e9fff97838f798d871e24cc90f2\"" Jan 30 13:25:30.086476 containerd[1781]: time="2025-01-30T13:25:30.085449088Z" level=info msg="StartContainer for \"bf5424f671d5cbbe138256e550d6318802c04e9fff97838f798d871e24cc90f2\"" Jan 30 13:25:30.130953 systemd[1]: Started cri-containerd-bf5424f671d5cbbe138256e550d6318802c04e9fff97838f798d871e24cc90f2.scope - libcontainer container bf5424f671d5cbbe138256e550d6318802c04e9fff97838f798d871e24cc90f2. Jan 30 13:25:30.167871 containerd[1781]: time="2025-01-30T13:25:30.167070928Z" level=info msg="StartContainer for \"bf5424f671d5cbbe138256e550d6318802c04e9fff97838f798d871e24cc90f2\" returns successfully" Jan 30 13:25:31.002322 kubelet[3455]: I0130 13:25:31.002276 3455 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: csi.tigera.io endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock versions: 1.0.0 Jan 30 13:25:31.005575 kubelet[3455]: I0130 13:25:31.005545 3455 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: csi.tigera.io at endpoint: /var/lib/kubelet/plugins/csi.tigera.io/csi.sock Jan 30 13:25:32.385837 kubelet[3455]: I0130 13:25:32.385766 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/csi-node-driver-42f6k" podStartSLOduration=37.174773962 podStartE2EDuration="48.385026149s" podCreationTimestamp="2025-01-30 13:24:44 +0000 UTC" firstStartedPulling="2025-01-30 13:25:18.831927702 +0000 UTC m=+56.061524689" lastFinishedPulling="2025-01-30 13:25:30.042179889 +0000 UTC m=+67.271776876" observedRunningTime="2025-01-30 13:25:30.751934963 +0000 UTC m=+67.981531950" watchObservedRunningTime="2025-01-30 13:25:32.385026149 +0000 UTC m=+69.614623136" Jan 30 13:25:32.389211 containerd[1781]: time="2025-01-30T13:25:32.389159069Z" level=info msg="StopContainer for \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\" with timeout 300 (s)" Jan 30 13:25:32.391430 containerd[1781]: time="2025-01-30T13:25:32.391143829Z" level=info msg="Stop container \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\" with signal terminated" Jan 30 13:25:32.565698 containerd[1781]: time="2025-01-30T13:25:32.565635787Z" level=info msg="StopContainer for \"5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2\" with timeout 30 (s)" Jan 30 13:25:32.566136 containerd[1781]: time="2025-01-30T13:25:32.566086507Z" level=info msg="Stop container \"5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2\" with signal terminated" Jan 30 13:25:32.594055 systemd[1]: cri-containerd-5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2.scope: Deactivated successfully. Jan 30 13:25:32.630294 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2-rootfs.mount: Deactivated successfully. Jan 30 13:25:32.703666 containerd[1781]: time="2025-01-30T13:25:32.703610506Z" level=info msg="StopContainer for \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\" with timeout 5 (s)" Jan 30 13:25:32.704294 containerd[1781]: time="2025-01-30T13:25:32.704248706Z" level=info msg="Stop container \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\" with signal terminated" Jan 30 13:25:32.744495 systemd[1]: cri-containerd-faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca.scope: Deactivated successfully. Jan 30 13:25:32.745117 systemd[1]: cri-containerd-faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca.scope: Consumed 1.559s CPU time. Jan 30 13:25:32.787952 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca-rootfs.mount: Deactivated successfully. Jan 30 13:25:32.791467 containerd[1781]: time="2025-01-30T13:25:32.791396105Z" level=info msg="shim disconnected" id=faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca namespace=k8s.io Jan 30 13:25:32.791870 containerd[1781]: time="2025-01-30T13:25:32.791653465Z" level=warning msg="cleaning up after shim disconnected" id=faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca namespace=k8s.io Jan 30 13:25:32.791870 containerd[1781]: time="2025-01-30T13:25:32.791671865Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:25:32.805566 containerd[1781]: time="2025-01-30T13:25:32.805444625Z" level=warning msg="cleanup warnings time=\"2025-01-30T13:25:32Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Jan 30 13:25:33.439824 containerd[1781]: time="2025-01-30T13:25:33.439623580Z" level=info msg="shim disconnected" id=5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2 namespace=k8s.io Jan 30 13:25:33.439824 containerd[1781]: time="2025-01-30T13:25:33.439791660Z" level=warning msg="cleaning up after shim disconnected" id=5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2 namespace=k8s.io Jan 30 13:25:33.439824 containerd[1781]: time="2025-01-30T13:25:33.439802180Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:25:33.444535 containerd[1781]: time="2025-01-30T13:25:33.444368860Z" level=info msg="StopContainer for \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\" returns successfully" Jan 30 13:25:33.446665 containerd[1781]: time="2025-01-30T13:25:33.446597180Z" level=info msg="StopPodSandbox for \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\"" Jan 30 13:25:33.446665 containerd[1781]: time="2025-01-30T13:25:33.446651940Z" level=info msg="Container to stop \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:25:33.446665 containerd[1781]: time="2025-01-30T13:25:33.446666700Z" level=info msg="Container to stop \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:25:33.447085 containerd[1781]: time="2025-01-30T13:25:33.446676140Z" level=info msg="Container to stop \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:25:33.459563 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618-shm.mount: Deactivated successfully. Jan 30 13:25:33.476077 systemd[1]: cri-containerd-8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618.scope: Deactivated successfully. Jan 30 13:25:33.481649 containerd[1781]: time="2025-01-30T13:25:33.481591379Z" level=info msg="StopContainer for \"5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2\" returns successfully" Jan 30 13:25:33.483314 containerd[1781]: time="2025-01-30T13:25:33.483235619Z" level=info msg="StopPodSandbox for \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\"" Jan 30 13:25:33.483314 containerd[1781]: time="2025-01-30T13:25:33.483279259Z" level=info msg="Container to stop \"5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:25:33.488442 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5-shm.mount: Deactivated successfully. Jan 30 13:25:33.506730 systemd[1]: cri-containerd-39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5.scope: Deactivated successfully. Jan 30 13:25:33.528530 containerd[1781]: time="2025-01-30T13:25:33.527997579Z" level=info msg="shim disconnected" id=8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618 namespace=k8s.io Jan 30 13:25:33.528530 containerd[1781]: time="2025-01-30T13:25:33.528062139Z" level=warning msg="cleaning up after shim disconnected" id=8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618 namespace=k8s.io Jan 30 13:25:33.528530 containerd[1781]: time="2025-01-30T13:25:33.528071019Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:25:33.528109 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618-rootfs.mount: Deactivated successfully. Jan 30 13:25:33.552782 containerd[1781]: time="2025-01-30T13:25:33.552487859Z" level=info msg="shim disconnected" id=39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5 namespace=k8s.io Jan 30 13:25:33.553849 containerd[1781]: time="2025-01-30T13:25:33.553610419Z" level=warning msg="cleaning up after shim disconnected" id=39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5 namespace=k8s.io Jan 30 13:25:33.553849 containerd[1781]: time="2025-01-30T13:25:33.553637499Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:25:33.557275 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5-rootfs.mount: Deactivated successfully. Jan 30 13:25:33.575427 containerd[1781]: time="2025-01-30T13:25:33.575369739Z" level=info msg="TearDown network for sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" successfully" Jan 30 13:25:33.575427 containerd[1781]: time="2025-01-30T13:25:33.575417699Z" level=info msg="StopPodSandbox for \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" returns successfully" Jan 30 13:25:33.634567 kubelet[3455]: I0130 13:25:33.633528 3455 topology_manager.go:215] "Topology Admit Handler" podUID="fd79e848-a90a-4c46-9841-000810035a01" podNamespace="calico-system" podName="calico-node-6sz7h" Jan 30 13:25:33.634567 kubelet[3455]: E0130 13:25:33.633602 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="788d3ec6-3317-4475-a4e0-bb103af4d1c2" containerName="flexvol-driver" Jan 30 13:25:33.634567 kubelet[3455]: E0130 13:25:33.633612 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="788d3ec6-3317-4475-a4e0-bb103af4d1c2" containerName="install-cni" Jan 30 13:25:33.634567 kubelet[3455]: E0130 13:25:33.633619 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="788d3ec6-3317-4475-a4e0-bb103af4d1c2" containerName="calico-node" Jan 30 13:25:33.639833 kubelet[3455]: I0130 13:25:33.637085 3455 memory_manager.go:354] "RemoveStaleState removing state" podUID="788d3ec6-3317-4475-a4e0-bb103af4d1c2" containerName="calico-node" Jan 30 13:25:33.653094 systemd[1]: Created slice kubepods-besteffort-podfd79e848_a90a_4c46_9841_000810035a01.slice - libcontainer container kubepods-besteffort-podfd79e848_a90a_4c46_9841_000810035a01.slice. Jan 30 13:25:33.666390 kubelet[3455]: I0130 13:25:33.665214 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/788d3ec6-3317-4475-a4e0-bb103af4d1c2-node-certs\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666390 kubelet[3455]: I0130 13:25:33.665847 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-net-dir\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666390 kubelet[3455]: I0130 13:25:33.665871 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-lib-modules\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666390 kubelet[3455]: I0130 13:25:33.665886 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-policysync\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666390 kubelet[3455]: I0130 13:25:33.665933 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdjxc\" (UniqueName: \"kubernetes.io/projected/788d3ec6-3317-4475-a4e0-bb103af4d1c2-kube-api-access-kdjxc\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666390 kubelet[3455]: I0130 13:25:33.665956 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-bin-dir\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666670 kubelet[3455]: I0130 13:25:33.665973 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-var-lib-calico\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666670 kubelet[3455]: I0130 13:25:33.666010 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-xtables-lock\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666670 kubelet[3455]: I0130 13:25:33.666030 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-var-run-calico\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666670 kubelet[3455]: I0130 13:25:33.666044 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-flexvol-driver-host\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666670 kubelet[3455]: I0130 13:25:33.666061 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/788d3ec6-3317-4475-a4e0-bb103af4d1c2-tigera-ca-bundle\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666670 kubelet[3455]: I0130 13:25:33.666098 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-log-dir\") pod \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\" (UID: \"788d3ec6-3317-4475-a4e0-bb103af4d1c2\") " Jan 30 13:25:33.666821 kubelet[3455]: I0130 13:25:33.666165 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/fd79e848-a90a-4c46-9841-000810035a01-policysync\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.666821 kubelet[3455]: I0130 13:25:33.666189 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/fd79e848-a90a-4c46-9841-000810035a01-var-lib-calico\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.666821 kubelet[3455]: I0130 13:25:33.666207 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd79e848-a90a-4c46-9841-000810035a01-lib-modules\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.666821 kubelet[3455]: I0130 13:25:33.666224 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/fd79e848-a90a-4c46-9841-000810035a01-cni-net-dir\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.666821 kubelet[3455]: I0130 13:25:33.666262 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/fd79e848-a90a-4c46-9841-000810035a01-flexvol-driver-host\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.666932 kubelet[3455]: I0130 13:25:33.666283 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/fd79e848-a90a-4c46-9841-000810035a01-xtables-lock\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.666932 kubelet[3455]: I0130 13:25:33.666302 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/fd79e848-a90a-4c46-9841-000810035a01-cni-log-dir\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.666932 kubelet[3455]: I0130 13:25:33.666807 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/fd79e848-a90a-4c46-9841-000810035a01-tigera-ca-bundle\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.666932 kubelet[3455]: I0130 13:25:33.666844 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/fd79e848-a90a-4c46-9841-000810035a01-node-certs\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.666932 kubelet[3455]: I0130 13:25:33.666861 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/fd79e848-a90a-4c46-9841-000810035a01-cni-bin-dir\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.667037 kubelet[3455]: I0130 13:25:33.666880 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/fd79e848-a90a-4c46-9841-000810035a01-var-run-calico\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.667037 kubelet[3455]: I0130 13:25:33.666897 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsltb\" (UniqueName: \"kubernetes.io/projected/fd79e848-a90a-4c46-9841-000810035a01-kube-api-access-hsltb\") pod \"calico-node-6sz7h\" (UID: \"fd79e848-a90a-4c46-9841-000810035a01\") " pod="calico-system/calico-node-6sz7h" Jan 30 13:25:33.667037 kubelet[3455]: I0130 13:25:33.666977 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-net-dir" (OuterVolumeSpecName: "cni-net-dir") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "cni-net-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:25:33.667037 kubelet[3455]: I0130 13:25:33.667023 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:25:33.667123 kubelet[3455]: I0130 13:25:33.667043 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-policysync" (OuterVolumeSpecName: "policysync") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "policysync". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:25:33.671814 systemd-networkd[1549]: cali69cbe76291a: Link DOWN Jan 30 13:25:33.671822 systemd-networkd[1549]: cali69cbe76291a: Lost carrier Jan 30 13:25:33.678269 kubelet[3455]: I0130 13:25:33.678059 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-bin-dir" (OuterVolumeSpecName: "cni-bin-dir") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "cni-bin-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:25:33.678269 kubelet[3455]: I0130 13:25:33.678111 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-var-lib-calico" (OuterVolumeSpecName: "var-lib-calico") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "var-lib-calico". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:25:33.678269 kubelet[3455]: I0130 13:25:33.678130 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:25:33.678269 kubelet[3455]: I0130 13:25:33.678148 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-var-run-calico" (OuterVolumeSpecName: "var-run-calico") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "var-run-calico". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:25:33.678269 kubelet[3455]: I0130 13:25:33.678165 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-flexvol-driver-host" (OuterVolumeSpecName: "flexvol-driver-host") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "flexvol-driver-host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:25:33.682548 kubelet[3455]: I0130 13:25:33.682465 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-log-dir" (OuterVolumeSpecName: "cni-log-dir") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "cni-log-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Jan 30 13:25:33.689642 kubelet[3455]: I0130 13:25:33.689190 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788d3ec6-3317-4475-a4e0-bb103af4d1c2-kube-api-access-kdjxc" (OuterVolumeSpecName: "kube-api-access-kdjxc") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "kube-api-access-kdjxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 30 13:25:33.691305 systemd[1]: var-lib-kubelet-pods-788d3ec6\x2d3317\x2d4475\x2da4e0\x2dbb103af4d1c2-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dkdjxc.mount: Deactivated successfully. Jan 30 13:25:33.699119 systemd[1]: var-lib-kubelet-pods-788d3ec6\x2d3317\x2d4475\x2da4e0\x2dbb103af4d1c2-volumes-kubernetes.io\x7esecret-node\x2dcerts.mount: Deactivated successfully. Jan 30 13:25:33.708182 systemd[1]: var-lib-kubelet-pods-788d3ec6\x2d3317\x2d4475\x2da4e0\x2dbb103af4d1c2-volume\x2dsubpaths-tigera\x2dca\x2dbundle-calico\x2dnode-1.mount: Deactivated successfully. Jan 30 13:25:33.710599 kubelet[3455]: I0130 13:25:33.710455 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788d3ec6-3317-4475-a4e0-bb103af4d1c2-node-certs" (OuterVolumeSpecName: "node-certs") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "node-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 30 13:25:33.710923 kubelet[3455]: I0130 13:25:33.710889 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788d3ec6-3317-4475-a4e0-bb103af4d1c2-tigera-ca-bundle" (OuterVolumeSpecName: "tigera-ca-bundle") pod "788d3ec6-3317-4475-a4e0-bb103af4d1c2" (UID: "788d3ec6-3317-4475-a4e0-bb103af4d1c2"). InnerVolumeSpecName "tigera-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 30 13:25:33.745577 kubelet[3455]: I0130 13:25:33.745432 3455 scope.go:117] "RemoveContainer" containerID="faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca" Jan 30 13:25:33.750194 containerd[1781]: time="2025-01-30T13:25:33.749817937Z" level=info msg="RemoveContainer for \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\"" Jan 30 13:25:33.751412 kubelet[3455]: I0130 13:25:33.750986 3455 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:25:33.753266 systemd[1]: cri-containerd-ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a.scope: Deactivated successfully. Jan 30 13:25:33.758628 systemd[1]: Removed slice kubepods-besteffort-pod788d3ec6_3317_4475_a4e0_bb103af4d1c2.slice - libcontainer container kubepods-besteffort-pod788d3ec6_3317_4475_a4e0_bb103af4d1c2.slice. Jan 30 13:25:33.759099 systemd[1]: kubepods-besteffort-pod788d3ec6_3317_4475_a4e0_bb103af4d1c2.slice: Consumed 1.982s CPU time. Jan 30 13:25:33.766256 containerd[1781]: time="2025-01-30T13:25:33.765681377Z" level=info msg="RemoveContainer for \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\" returns successfully" Jan 30 13:25:33.766399 kubelet[3455]: I0130 13:25:33.766056 3455 scope.go:117] "RemoveContainer" containerID="8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5" Jan 30 13:25:33.768743 kubelet[3455]: I0130 13:25:33.768014 3455 reconciler_common.go:289] "Volume detached for volume \"policysync\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-policysync\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.768743 kubelet[3455]: I0130 13:25:33.768047 3455 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-kdjxc\" (UniqueName: \"kubernetes.io/projected/788d3ec6-3317-4475-a4e0-bb103af4d1c2-kube-api-access-kdjxc\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.768743 kubelet[3455]: I0130 13:25:33.768060 3455 reconciler_common.go:289] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-xtables-lock\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.768743 kubelet[3455]: I0130 13:25:33.768072 3455 reconciler_common.go:289] "Volume detached for volume \"cni-log-dir\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-log-dir\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.768743 kubelet[3455]: I0130 13:25:33.768081 3455 reconciler_common.go:289] "Volume detached for volume \"node-certs\" (UniqueName: \"kubernetes.io/secret/788d3ec6-3317-4475-a4e0-bb103af4d1c2-node-certs\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.768743 kubelet[3455]: I0130 13:25:33.768090 3455 reconciler_common.go:289] "Volume detached for volume \"cni-net-dir\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-net-dir\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.768743 kubelet[3455]: I0130 13:25:33.768098 3455 reconciler_common.go:289] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-lib-modules\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.768743 kubelet[3455]: I0130 13:25:33.768108 3455 reconciler_common.go:289] "Volume detached for volume \"var-lib-calico\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-var-lib-calico\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.769030 kubelet[3455]: I0130 13:25:33.768116 3455 reconciler_common.go:289] "Volume detached for volume \"cni-bin-dir\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-cni-bin-dir\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.769030 kubelet[3455]: I0130 13:25:33.768125 3455 reconciler_common.go:289] "Volume detached for volume \"var-run-calico\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-var-run-calico\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.769030 kubelet[3455]: I0130 13:25:33.768134 3455 reconciler_common.go:289] "Volume detached for volume \"flexvol-driver-host\" (UniqueName: \"kubernetes.io/host-path/788d3ec6-3317-4475-a4e0-bb103af4d1c2-flexvol-driver-host\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.769030 kubelet[3455]: I0130 13:25:33.768142 3455 reconciler_common.go:289] "Volume detached for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/788d3ec6-3317-4475-a4e0-bb103af4d1c2-tigera-ca-bundle\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:33.778352 containerd[1781]: time="2025-01-30T13:25:33.777854777Z" level=info msg="RemoveContainer for \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\"" Jan 30 13:25:33.790912 containerd[1781]: time="2025-01-30T13:25:33.790864497Z" level=info msg="RemoveContainer for \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\" returns successfully" Jan 30 13:25:33.791344 kubelet[3455]: I0130 13:25:33.791321 3455 scope.go:117] "RemoveContainer" containerID="5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015" Jan 30 13:25:33.795645 containerd[1781]: time="2025-01-30T13:25:33.795523017Z" level=info msg="RemoveContainer for \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\"" Jan 30 13:25:33.816193 containerd[1781]: time="2025-01-30T13:25:33.815626697Z" level=info msg="shim disconnected" id=ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a namespace=k8s.io Jan 30 13:25:33.816193 containerd[1781]: time="2025-01-30T13:25:33.816080497Z" level=warning msg="cleaning up after shim disconnected" id=ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a namespace=k8s.io Jan 30 13:25:33.816576 containerd[1781]: time="2025-01-30T13:25:33.816092977Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:25:33.816667 containerd[1781]: time="2025-01-30T13:25:33.815995057Z" level=info msg="RemoveContainer for \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\" returns successfully" Jan 30 13:25:33.817643 kubelet[3455]: I0130 13:25:33.817499 3455 scope.go:117] "RemoveContainer" containerID="faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca" Jan 30 13:25:33.818299 containerd[1781]: time="2025-01-30T13:25:33.818210136Z" level=error msg="ContainerStatus for \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\": not found" Jan 30 13:25:33.825408 kubelet[3455]: E0130 13:25:33.825243 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\": not found" containerID="faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca" Jan 30 13:25:33.825408 kubelet[3455]: I0130 13:25:33.825308 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca"} err="failed to get container status \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\": rpc error: code = NotFound desc = an error occurred when try to find container \"faa741244aca8b3b9eac13962db2112d598a2826409849367f62bd2d86aca0ca\": not found" Jan 30 13:25:33.825408 kubelet[3455]: I0130 13:25:33.825403 3455 scope.go:117] "RemoveContainer" containerID="8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5" Jan 30 13:25:33.828586 containerd[1781]: time="2025-01-30T13:25:33.828255696Z" level=error msg="ContainerStatus for \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\": not found" Jan 30 13:25:33.829639 kubelet[3455]: E0130 13:25:33.828887 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\": not found" containerID="8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5" Jan 30 13:25:33.829639 kubelet[3455]: I0130 13:25:33.828921 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5"} err="failed to get container status \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\": rpc error: code = NotFound desc = an error occurred when try to find container \"8f72e196f3a3e776ec7c951cd9e28aee5d510661fd031786dc5226e263b26ca5\": not found" Jan 30 13:25:33.829639 kubelet[3455]: I0130 13:25:33.828953 3455 scope.go:117] "RemoveContainer" containerID="5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015" Jan 30 13:25:33.832218 containerd[1781]: time="2025-01-30T13:25:33.831390296Z" level=error msg="ContainerStatus for \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\": not found" Jan 30 13:25:33.833570 kubelet[3455]: E0130 13:25:33.833473 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\": not found" containerID="5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015" Jan 30 13:25:33.833570 kubelet[3455]: I0130 13:25:33.833535 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015"} err="failed to get container status \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\": rpc error: code = NotFound desc = an error occurred when try to find container \"5f8cc22e780cfee045351645619cb29d8b71dbf4062b73683d2adae3af594015\": not found" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.663 [INFO][6793] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.663 [INFO][6793] cni-plugin/dataplane_linux.go 559: Deleting workload's device in netns. ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" iface="eth0" netns="/var/run/netns/cni-a0f9e971-721b-79bb-8ffa-6ea40272335b" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.665 [INFO][6793] cni-plugin/dataplane_linux.go 570: Entered netns, deleting veth. ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" iface="eth0" netns="/var/run/netns/cni-a0f9e971-721b-79bb-8ffa-6ea40272335b" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.683 [INFO][6793] cni-plugin/dataplane_linux.go 604: Deleted device in netns. ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" after=18.24816ms iface="eth0" netns="/var/run/netns/cni-a0f9e971-721b-79bb-8ffa-6ea40272335b" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.683 [INFO][6793] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.683 [INFO][6793] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.726 [INFO][6802] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.727 [INFO][6802] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.727 [INFO][6802] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.823 [INFO][6802] ipam/ipam_plugin.go 431: Released address using handleID ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.824 [INFO][6802] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.830 [INFO][6802] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:25:33.841698 containerd[1781]: 2025-01-30 13:25:33.835 [INFO][6793] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:25:33.844320 containerd[1781]: time="2025-01-30T13:25:33.843662976Z" level=warning msg="cleanup warnings time=\"2025-01-30T13:25:33Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Jan 30 13:25:33.844320 containerd[1781]: time="2025-01-30T13:25:33.844125136Z" level=info msg="TearDown network for sandbox \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\" successfully" Jan 30 13:25:33.844320 containerd[1781]: time="2025-01-30T13:25:33.844147536Z" level=info msg="StopPodSandbox for \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\" returns successfully" Jan 30 13:25:33.877880 containerd[1781]: time="2025-01-30T13:25:33.877826816Z" level=info msg="StopContainer for \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\" returns successfully" Jan 30 13:25:33.879295 containerd[1781]: time="2025-01-30T13:25:33.879004216Z" level=info msg="StopPodSandbox for \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\"" Jan 30 13:25:33.879295 containerd[1781]: time="2025-01-30T13:25:33.879054776Z" level=info msg="Container to stop \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Jan 30 13:25:33.890382 systemd[1]: cri-containerd-f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330.scope: Deactivated successfully. Jan 30 13:25:33.930669 containerd[1781]: time="2025-01-30T13:25:33.930456296Z" level=info msg="shim disconnected" id=f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330 namespace=k8s.io Jan 30 13:25:33.930669 containerd[1781]: time="2025-01-30T13:25:33.930516296Z" level=warning msg="cleaning up after shim disconnected" id=f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330 namespace=k8s.io Jan 30 13:25:33.930669 containerd[1781]: time="2025-01-30T13:25:33.930528256Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:25:33.957760 containerd[1781]: time="2025-01-30T13:25:33.957537575Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6sz7h,Uid:fd79e848-a90a-4c46-9841-000810035a01,Namespace:calico-system,Attempt:0,}" Jan 30 13:25:33.965001 containerd[1781]: time="2025-01-30T13:25:33.964949855Z" level=info msg="TearDown network for sandbox \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\" successfully" Jan 30 13:25:33.965001 containerd[1781]: time="2025-01-30T13:25:33.964990615Z" level=info msg="StopPodSandbox for \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\" returns successfully" Jan 30 13:25:33.970027 kubelet[3455]: I0130 13:25:33.969061 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aad42f51-7748-4be3-8b1d-97665b3346fc-tigera-ca-bundle\") pod \"aad42f51-7748-4be3-8b1d-97665b3346fc\" (UID: \"aad42f51-7748-4be3-8b1d-97665b3346fc\") " Jan 30 13:25:33.970027 kubelet[3455]: I0130 13:25:33.969100 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mbdm\" (UniqueName: \"kubernetes.io/projected/aad42f51-7748-4be3-8b1d-97665b3346fc-kube-api-access-2mbdm\") pod \"aad42f51-7748-4be3-8b1d-97665b3346fc\" (UID: \"aad42f51-7748-4be3-8b1d-97665b3346fc\") " Jan 30 13:25:33.980342 kubelet[3455]: I0130 13:25:33.980277 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aad42f51-7748-4be3-8b1d-97665b3346fc-tigera-ca-bundle" (OuterVolumeSpecName: "tigera-ca-bundle") pod "aad42f51-7748-4be3-8b1d-97665b3346fc" (UID: "aad42f51-7748-4be3-8b1d-97665b3346fc"). InnerVolumeSpecName "tigera-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 30 13:25:33.984937 kubelet[3455]: I0130 13:25:33.984591 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aad42f51-7748-4be3-8b1d-97665b3346fc-kube-api-access-2mbdm" (OuterVolumeSpecName: "kube-api-access-2mbdm") pod "aad42f51-7748-4be3-8b1d-97665b3346fc" (UID: "aad42f51-7748-4be3-8b1d-97665b3346fc"). InnerVolumeSpecName "kube-api-access-2mbdm". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 30 13:25:34.008272 containerd[1781]: time="2025-01-30T13:25:34.007877135Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:25:34.008272 containerd[1781]: time="2025-01-30T13:25:34.008015215Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:25:34.008802 containerd[1781]: time="2025-01-30T13:25:34.008241735Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:34.009323 containerd[1781]: time="2025-01-30T13:25:34.009254735Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:34.039000 systemd[1]: Started cri-containerd-bb2ccd51875f3ce5a8a265a2ddf901728cb1e72e1625d1f6903380ac03548620.scope - libcontainer container bb2ccd51875f3ce5a8a265a2ddf901728cb1e72e1625d1f6903380ac03548620. Jan 30 13:25:34.070242 kubelet[3455]: I0130 13:25:34.069697 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-tigera-ca-bundle\") pod \"c6dbcf90-fed1-4eb3-9b55-393ac028ea12\" (UID: \"c6dbcf90-fed1-4eb3-9b55-393ac028ea12\") " Jan 30 13:25:34.070890 kubelet[3455]: I0130 13:25:34.070387 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-typha-certs\") pod \"c6dbcf90-fed1-4eb3-9b55-393ac028ea12\" (UID: \"c6dbcf90-fed1-4eb3-9b55-393ac028ea12\") " Jan 30 13:25:34.070890 kubelet[3455]: I0130 13:25:34.070432 3455 reconciler_common.go:161] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdxh2\" (UniqueName: \"kubernetes.io/projected/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-kube-api-access-wdxh2\") pod \"c6dbcf90-fed1-4eb3-9b55-393ac028ea12\" (UID: \"c6dbcf90-fed1-4eb3-9b55-393ac028ea12\") " Jan 30 13:25:34.070890 kubelet[3455]: I0130 13:25:34.070889 3455 reconciler_common.go:289] "Volume detached for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/aad42f51-7748-4be3-8b1d-97665b3346fc-tigera-ca-bundle\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:34.071020 kubelet[3455]: I0130 13:25:34.070905 3455 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-2mbdm\" (UniqueName: \"kubernetes.io/projected/aad42f51-7748-4be3-8b1d-97665b3346fc-kube-api-access-2mbdm\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:34.080840 kubelet[3455]: I0130 13:25:34.079926 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-kube-api-access-wdxh2" (OuterVolumeSpecName: "kube-api-access-wdxh2") pod "c6dbcf90-fed1-4eb3-9b55-393ac028ea12" (UID: "c6dbcf90-fed1-4eb3-9b55-393ac028ea12"). InnerVolumeSpecName "kube-api-access-wdxh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Jan 30 13:25:34.082982 kubelet[3455]: I0130 13:25:34.082113 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-typha-certs" (OuterVolumeSpecName: "typha-certs") pod "c6dbcf90-fed1-4eb3-9b55-393ac028ea12" (UID: "c6dbcf90-fed1-4eb3-9b55-393ac028ea12"). InnerVolumeSpecName "typha-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Jan 30 13:25:34.083380 kubelet[3455]: I0130 13:25:34.083276 3455 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-tigera-ca-bundle" (OuterVolumeSpecName: "tigera-ca-bundle") pod "c6dbcf90-fed1-4eb3-9b55-393ac028ea12" (UID: "c6dbcf90-fed1-4eb3-9b55-393ac028ea12"). InnerVolumeSpecName "tigera-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Jan 30 13:25:34.087526 containerd[1781]: time="2025-01-30T13:25:34.087186934Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-node-6sz7h,Uid:fd79e848-a90a-4c46-9841-000810035a01,Namespace:calico-system,Attempt:0,} returns sandbox id \"bb2ccd51875f3ce5a8a265a2ddf901728cb1e72e1625d1f6903380ac03548620\"" Jan 30 13:25:34.094629 containerd[1781]: time="2025-01-30T13:25:34.094554934Z" level=info msg="CreateContainer within sandbox \"bb2ccd51875f3ce5a8a265a2ddf901728cb1e72e1625d1f6903380ac03548620\" for container &ContainerMetadata{Name:flexvol-driver,Attempt:0,}" Jan 30 13:25:34.126000 containerd[1781]: time="2025-01-30T13:25:34.125934094Z" level=info msg="CreateContainer within sandbox \"bb2ccd51875f3ce5a8a265a2ddf901728cb1e72e1625d1f6903380ac03548620\" for &ContainerMetadata{Name:flexvol-driver,Attempt:0,} returns container id \"83c182abbe1307ed748caa7efa304f05c4556c614be99e0c2a7b330064880bf4\"" Jan 30 13:25:34.126587 containerd[1781]: time="2025-01-30T13:25:34.126489534Z" level=info msg="StartContainer for \"83c182abbe1307ed748caa7efa304f05c4556c614be99e0c2a7b330064880bf4\"" Jan 30 13:25:34.170558 systemd[1]: Started cri-containerd-83c182abbe1307ed748caa7efa304f05c4556c614be99e0c2a7b330064880bf4.scope - libcontainer container 83c182abbe1307ed748caa7efa304f05c4556c614be99e0c2a7b330064880bf4. Jan 30 13:25:34.171499 kubelet[3455]: I0130 13:25:34.171340 3455 reconciler_common.go:289] "Volume detached for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-typha-certs\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:34.171499 kubelet[3455]: I0130 13:25:34.171370 3455 reconciler_common.go:289] "Volume detached for volume \"kube-api-access-wdxh2\" (UniqueName: \"kubernetes.io/projected/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-kube-api-access-wdxh2\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:34.171499 kubelet[3455]: I0130 13:25:34.171381 3455 reconciler_common.go:289] "Volume detached for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c6dbcf90-fed1-4eb3-9b55-393ac028ea12-tigera-ca-bundle\") on node \"ci-4186.1.0-a-a7402f3063\" DevicePath \"\"" Jan 30 13:25:34.217152 containerd[1781]: time="2025-01-30T13:25:34.217027493Z" level=info msg="StartContainer for \"83c182abbe1307ed748caa7efa304f05c4556c614be99e0c2a7b330064880bf4\" returns successfully" Jan 30 13:25:34.237250 systemd[1]: cri-containerd-83c182abbe1307ed748caa7efa304f05c4556c614be99e0c2a7b330064880bf4.scope: Deactivated successfully. Jan 30 13:25:34.279694 containerd[1781]: time="2025-01-30T13:25:34.279018133Z" level=info msg="shim disconnected" id=83c182abbe1307ed748caa7efa304f05c4556c614be99e0c2a7b330064880bf4 namespace=k8s.io Jan 30 13:25:34.280203 containerd[1781]: time="2025-01-30T13:25:34.279740093Z" level=warning msg="cleaning up after shim disconnected" id=83c182abbe1307ed748caa7efa304f05c4556c614be99e0c2a7b330064880bf4 namespace=k8s.io Jan 30 13:25:34.280203 containerd[1781]: time="2025-01-30T13:25:34.279779093Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:25:34.529300 systemd[1]: var-lib-kubelet-pods-aad42f51\x2d7748\x2d4be3\x2d8b1d\x2d97665b3346fc-volume\x2dsubpaths-tigera\x2dca\x2dbundle-calico\x2dkube\x2dcontrollers-1.mount: Deactivated successfully. Jan 30 13:25:34.529418 systemd[1]: run-netns-cni\x2da0f9e971\x2d721b\x2d79bb\x2d8ffa\x2d6ea40272335b.mount: Deactivated successfully. Jan 30 13:25:34.529472 systemd[1]: var-lib-kubelet-pods-aad42f51\x2d7748\x2d4be3\x2d8b1d\x2d97665b3346fc-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d2mbdm.mount: Deactivated successfully. Jan 30 13:25:34.529521 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a-rootfs.mount: Deactivated successfully. Jan 30 13:25:34.529568 systemd[1]: var-lib-kubelet-pods-c6dbcf90\x2dfed1\x2d4eb3\x2d9b55\x2d393ac028ea12-volume\x2dsubpaths-tigera\x2dca\x2dbundle-calico\x2dtypha-1.mount: Deactivated successfully. Jan 30 13:25:34.529620 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330-rootfs.mount: Deactivated successfully. Jan 30 13:25:34.529667 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330-shm.mount: Deactivated successfully. Jan 30 13:25:34.529737 systemd[1]: var-lib-kubelet-pods-c6dbcf90\x2dfed1\x2d4eb3\x2d9b55\x2d393ac028ea12-volumes-kubernetes.io\x7esecret-typha\x2dcerts.mount: Deactivated successfully. Jan 30 13:25:34.529788 systemd[1]: var-lib-kubelet-pods-c6dbcf90\x2dfed1\x2d4eb3\x2d9b55\x2d393ac028ea12-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dwdxh2.mount: Deactivated successfully. Jan 30 13:25:34.761053 containerd[1781]: time="2025-01-30T13:25:34.760999168Z" level=info msg="CreateContainer within sandbox \"bb2ccd51875f3ce5a8a265a2ddf901728cb1e72e1625d1f6903380ac03548620\" for container &ContainerMetadata{Name:install-cni,Attempt:0,}" Jan 30 13:25:34.771483 kubelet[3455]: I0130 13:25:34.771427 3455 scope.go:117] "RemoveContainer" containerID="ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a" Jan 30 13:25:34.777273 systemd[1]: Removed slice kubepods-besteffort-podc6dbcf90_fed1_4eb3_9b55_393ac028ea12.slice - libcontainer container kubepods-besteffort-podc6dbcf90_fed1_4eb3_9b55_393ac028ea12.slice. Jan 30 13:25:34.778374 containerd[1781]: time="2025-01-30T13:25:34.778167048Z" level=info msg="RemoveContainer for \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\"" Jan 30 13:25:34.788524 systemd[1]: Removed slice kubepods-besteffort-podaad42f51_7748_4be3_8b1d_97665b3346fc.slice - libcontainer container kubepods-besteffort-podaad42f51_7748_4be3_8b1d_97665b3346fc.slice. Jan 30 13:25:34.791850 containerd[1781]: time="2025-01-30T13:25:34.791768728Z" level=info msg="RemoveContainer for \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\" returns successfully" Jan 30 13:25:34.792140 kubelet[3455]: I0130 13:25:34.792047 3455 scope.go:117] "RemoveContainer" containerID="ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a" Jan 30 13:25:34.792566 containerd[1781]: time="2025-01-30T13:25:34.792348168Z" level=error msg="ContainerStatus for \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\": not found" Jan 30 13:25:34.792796 kubelet[3455]: E0130 13:25:34.792707 3455 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\": not found" containerID="ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a" Jan 30 13:25:34.792869 kubelet[3455]: I0130 13:25:34.792803 3455 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a"} err="failed to get container status \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\": rpc error: code = NotFound desc = an error occurred when try to find container \"ed0bd9e8358d91924902d0bb6fc0f9fe595359c74d87d7064be5430ab0e10d2a\": not found" Jan 30 13:25:34.799613 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1138370360.mount: Deactivated successfully. Jan 30 13:25:34.818435 containerd[1781]: time="2025-01-30T13:25:34.818394208Z" level=info msg="CreateContainer within sandbox \"bb2ccd51875f3ce5a8a265a2ddf901728cb1e72e1625d1f6903380ac03548620\" for &ContainerMetadata{Name:install-cni,Attempt:0,} returns container id \"67d0ca83cd0a91ce1c0cc0de5c737ab59a1cd5eab32f50546b2922f2b10cbbcd\"" Jan 30 13:25:34.819351 containerd[1781]: time="2025-01-30T13:25:34.819314488Z" level=info msg="StartContainer for \"67d0ca83cd0a91ce1c0cc0de5c737ab59a1cd5eab32f50546b2922f2b10cbbcd\"" Jan 30 13:25:34.869929 systemd[1]: Started cri-containerd-67d0ca83cd0a91ce1c0cc0de5c737ab59a1cd5eab32f50546b2922f2b10cbbcd.scope - libcontainer container 67d0ca83cd0a91ce1c0cc0de5c737ab59a1cd5eab32f50546b2922f2b10cbbcd. Jan 30 13:25:34.889061 kubelet[3455]: I0130 13:25:34.889007 3455 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788d3ec6-3317-4475-a4e0-bb103af4d1c2" path="/var/lib/kubelet/pods/788d3ec6-3317-4475-a4e0-bb103af4d1c2/volumes" Jan 30 13:25:34.889502 kubelet[3455]: I0130 13:25:34.889474 3455 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" path="/var/lib/kubelet/pods/aad42f51-7748-4be3-8b1d-97665b3346fc/volumes" Jan 30 13:25:34.891090 kubelet[3455]: I0130 13:25:34.890991 3455 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c6dbcf90-fed1-4eb3-9b55-393ac028ea12" path="/var/lib/kubelet/pods/c6dbcf90-fed1-4eb3-9b55-393ac028ea12/volumes" Jan 30 13:25:34.950413 containerd[1781]: time="2025-01-30T13:25:34.950364927Z" level=info msg="StartContainer for \"67d0ca83cd0a91ce1c0cc0de5c737ab59a1cd5eab32f50546b2922f2b10cbbcd\" returns successfully" Jan 30 13:25:35.581664 systemd[1]: cri-containerd-67d0ca83cd0a91ce1c0cc0de5c737ab59a1cd5eab32f50546b2922f2b10cbbcd.scope: Deactivated successfully. Jan 30 13:25:35.606649 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-67d0ca83cd0a91ce1c0cc0de5c737ab59a1cd5eab32f50546b2922f2b10cbbcd-rootfs.mount: Deactivated successfully. Jan 30 13:25:35.625665 containerd[1781]: time="2025-01-30T13:25:35.625599174Z" level=info msg="shim disconnected" id=67d0ca83cd0a91ce1c0cc0de5c737ab59a1cd5eab32f50546b2922f2b10cbbcd namespace=k8s.io Jan 30 13:25:35.625665 containerd[1781]: time="2025-01-30T13:25:35.625652974Z" level=warning msg="cleaning up after shim disconnected" id=67d0ca83cd0a91ce1c0cc0de5c737ab59a1cd5eab32f50546b2922f2b10cbbcd namespace=k8s.io Jan 30 13:25:35.625665 containerd[1781]: time="2025-01-30T13:25:35.625663214Z" level=info msg="cleaning up dead shim" namespace=k8s.io Jan 30 13:25:35.637629 containerd[1781]: time="2025-01-30T13:25:35.637475297Z" level=warning msg="cleanup warnings time=\"2025-01-30T13:25:35Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" namespace=k8s.io Jan 30 13:25:35.783477 containerd[1781]: time="2025-01-30T13:25:35.783355419Z" level=info msg="CreateContainer within sandbox \"bb2ccd51875f3ce5a8a265a2ddf901728cb1e72e1625d1f6903380ac03548620\" for container &ContainerMetadata{Name:calico-node,Attempt:0,}" Jan 30 13:25:35.826883 containerd[1781]: time="2025-01-30T13:25:35.826815431Z" level=info msg="CreateContainer within sandbox \"bb2ccd51875f3ce5a8a265a2ddf901728cb1e72e1625d1f6903380ac03548620\" for &ContainerMetadata{Name:calico-node,Attempt:0,} returns container id \"c3c19795377f0fa397df35adf13fe36ac8b7755bbe11eb7adfc0acdce85eb4b7\"" Jan 30 13:25:35.827819 containerd[1781]: time="2025-01-30T13:25:35.827782272Z" level=info msg="StartContainer for \"c3c19795377f0fa397df35adf13fe36ac8b7755bbe11eb7adfc0acdce85eb4b7\"" Jan 30 13:25:35.857959 systemd[1]: Started cri-containerd-c3c19795377f0fa397df35adf13fe36ac8b7755bbe11eb7adfc0acdce85eb4b7.scope - libcontainer container c3c19795377f0fa397df35adf13fe36ac8b7755bbe11eb7adfc0acdce85eb4b7. Jan 30 13:25:35.910283 containerd[1781]: time="2025-01-30T13:25:35.910215455Z" level=info msg="StartContainer for \"c3c19795377f0fa397df35adf13fe36ac8b7755bbe11eb7adfc0acdce85eb4b7\" returns successfully" Jan 30 13:25:36.028501 kubelet[3455]: I0130 13:25:36.028435 3455 topology_manager.go:215] "Topology Admit Handler" podUID="596b2831-ccf6-4e62-a56c-427ece32d593" podNamespace="calico-system" podName="calico-kube-controllers-ccb9d4c56-jglzv" Jan 30 13:25:36.028929 kubelet[3455]: E0130 13:25:36.028514 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="c6dbcf90-fed1-4eb3-9b55-393ac028ea12" containerName="calico-typha" Jan 30 13:25:36.028929 kubelet[3455]: E0130 13:25:36.028524 3455 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" containerName="calico-kube-controllers" Jan 30 13:25:36.028929 kubelet[3455]: I0130 13:25:36.028550 3455 memory_manager.go:354] "RemoveStaleState removing state" podUID="aad42f51-7748-4be3-8b1d-97665b3346fc" containerName="calico-kube-controllers" Jan 30 13:25:36.028929 kubelet[3455]: I0130 13:25:36.028558 3455 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6dbcf90-fed1-4eb3-9b55-393ac028ea12" containerName="calico-typha" Jan 30 13:25:36.041987 systemd[1]: Created slice kubepods-besteffort-pod596b2831_ccf6_4e62_a56c_427ece32d593.slice - libcontainer container kubepods-besteffort-pod596b2831_ccf6_4e62_a56c_427ece32d593.slice. Jan 30 13:25:36.088236 kubelet[3455]: I0130 13:25:36.088002 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hqkm\" (UniqueName: \"kubernetes.io/projected/596b2831-ccf6-4e62-a56c-427ece32d593-kube-api-access-5hqkm\") pod \"calico-kube-controllers-ccb9d4c56-jglzv\" (UID: \"596b2831-ccf6-4e62-a56c-427ece32d593\") " pod="calico-system/calico-kube-controllers-ccb9d4c56-jglzv" Jan 30 13:25:36.088236 kubelet[3455]: I0130 13:25:36.088049 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/596b2831-ccf6-4e62-a56c-427ece32d593-tigera-ca-bundle\") pod \"calico-kube-controllers-ccb9d4c56-jglzv\" (UID: \"596b2831-ccf6-4e62-a56c-427ece32d593\") " pod="calico-system/calico-kube-controllers-ccb9d4c56-jglzv" Jan 30 13:25:36.088991 kubelet[3455]: I0130 13:25:36.088940 3455 topology_manager.go:215] "Topology Admit Handler" podUID="e50dab71-58bb-4eca-bcc6-319aebae819a" podNamespace="calico-system" podName="calico-typha-96946fff6-qrhfp" Jan 30 13:25:36.098813 systemd[1]: Created slice kubepods-besteffort-pode50dab71_58bb_4eca_bcc6_319aebae819a.slice - libcontainer container kubepods-besteffort-pode50dab71_58bb_4eca_bcc6_319aebae819a.slice. Jan 30 13:25:36.189313 kubelet[3455]: I0130 13:25:36.189167 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"typha-certs\" (UniqueName: \"kubernetes.io/secret/e50dab71-58bb-4eca-bcc6-319aebae819a-typha-certs\") pod \"calico-typha-96946fff6-qrhfp\" (UID: \"e50dab71-58bb-4eca-bcc6-319aebae819a\") " pod="calico-system/calico-typha-96946fff6-qrhfp" Jan 30 13:25:36.189313 kubelet[3455]: I0130 13:25:36.189215 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c79cd\" (UniqueName: \"kubernetes.io/projected/e50dab71-58bb-4eca-bcc6-319aebae819a-kube-api-access-c79cd\") pod \"calico-typha-96946fff6-qrhfp\" (UID: \"e50dab71-58bb-4eca-bcc6-319aebae819a\") " pod="calico-system/calico-typha-96946fff6-qrhfp" Jan 30 13:25:36.189313 kubelet[3455]: I0130 13:25:36.189251 3455 reconciler_common.go:247] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tigera-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e50dab71-58bb-4eca-bcc6-319aebae819a-tigera-ca-bundle\") pod \"calico-typha-96946fff6-qrhfp\" (UID: \"e50dab71-58bb-4eca-bcc6-319aebae819a\") " pod="calico-system/calico-typha-96946fff6-qrhfp" Jan 30 13:25:36.346195 containerd[1781]: time="2025-01-30T13:25:36.346145941Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-ccb9d4c56-jglzv,Uid:596b2831-ccf6-4e62-a56c-427ece32d593,Namespace:calico-system,Attempt:0,}" Jan 30 13:25:36.404833 containerd[1781]: time="2025-01-30T13:25:36.404481077Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-96946fff6-qrhfp,Uid:e50dab71-58bb-4eca-bcc6-319aebae819a,Namespace:calico-system,Attempt:0,}" Jan 30 13:25:36.459771 containerd[1781]: time="2025-01-30T13:25:36.459223853Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:25:36.459771 containerd[1781]: time="2025-01-30T13:25:36.459695653Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:25:36.459771 containerd[1781]: time="2025-01-30T13:25:36.459737933Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:36.460948 containerd[1781]: time="2025-01-30T13:25:36.460132413Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:36.480952 systemd[1]: Started cri-containerd-07f437ada8808cffc15c2bc095d1509c47ce3f01af23cca4a07d245483d84951.scope - libcontainer container 07f437ada8808cffc15c2bc095d1509c47ce3f01af23cca4a07d245483d84951. Jan 30 13:25:36.531045 containerd[1781]: time="2025-01-30T13:25:36.530996314Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-typha-96946fff6-qrhfp,Uid:e50dab71-58bb-4eca-bcc6-319aebae819a,Namespace:calico-system,Attempt:0,} returns sandbox id \"07f437ada8808cffc15c2bc095d1509c47ce3f01af23cca4a07d245483d84951\"" Jan 30 13:25:36.543481 containerd[1781]: time="2025-01-30T13:25:36.543238997Z" level=info msg="CreateContainer within sandbox \"07f437ada8808cffc15c2bc095d1509c47ce3f01af23cca4a07d245483d84951\" for container &ContainerMetadata{Name:calico-typha,Attempt:0,}" Jan 30 13:25:36.576649 containerd[1781]: time="2025-01-30T13:25:36.576255527Z" level=info msg="CreateContainer within sandbox \"07f437ada8808cffc15c2bc095d1509c47ce3f01af23cca4a07d245483d84951\" for &ContainerMetadata{Name:calico-typha,Attempt:0,} returns container id \"a257e14f2ec67baa3ccb5e81c75bf86f7044d78d62b95a750347341c72a26d9e\"" Jan 30 13:25:36.578292 containerd[1781]: time="2025-01-30T13:25:36.578241247Z" level=info msg="StartContainer for \"a257e14f2ec67baa3ccb5e81c75bf86f7044d78d62b95a750347341c72a26d9e\"" Jan 30 13:25:36.600217 systemd-networkd[1549]: cali43a59e7f348: Link UP Jan 30 13:25:36.600941 systemd-networkd[1549]: cali43a59e7f348: Gained carrier Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.453 [INFO][7106] cni-plugin/plugin.go 325: Calico CNI found existing endpoint: &{{WorkloadEndpoint projectcalico.org/v3} {ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0 calico-kube-controllers-ccb9d4c56- calico-system 596b2831-ccf6-4e62-a56c-427ece32d593 1097 0 2025-01-30 13:25:34 +0000 UTC map[app.kubernetes.io/name:calico-kube-controllers k8s-app:calico-kube-controllers pod-template-hash:ccb9d4c56 projectcalico.org/namespace:calico-system projectcalico.org/orchestrator:k8s projectcalico.org/serviceaccount:calico-kube-controllers] map[] [] [] []} {k8s ci-4186.1.0-a-a7402f3063 calico-kube-controllers-ccb9d4c56-jglzv eth0 calico-kube-controllers [] [] [kns.calico-system ksa.calico-system.calico-kube-controllers] cali43a59e7f348 [] []}} ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Namespace="calico-system" Pod="calico-kube-controllers-ccb9d4c56-jglzv" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.456 [INFO][7106] cni-plugin/k8s.go 77: Extracted identifiers for CmdAddK8s ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Namespace="calico-system" Pod="calico-kube-controllers-ccb9d4c56-jglzv" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.522 [INFO][7144] ipam/ipam_plugin.go 225: Calico CNI IPAM request count IPv4=1 IPv6=0 ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" HandleID="k8s-pod-network.8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.548 [INFO][7144] ipam/ipam_plugin.go 265: Auto assigning IP ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" HandleID="k8s-pod-network.8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" assignArgs=ipam.AutoAssignArgs{Num4:1, Num6:0, HandleID:(*string)(0x400011c760), Attrs:map[string]string{"namespace":"calico-system", "node":"ci-4186.1.0-a-a7402f3063", "pod":"calico-kube-controllers-ccb9d4c56-jglzv", "timestamp":"2025-01-30 13:25:36.522354031 +0000 UTC"}, Hostname:"ci-4186.1.0-a-a7402f3063", IPv4Pools:[]net.IPNet{}, IPv6Pools:[]net.IPNet{}, MaxBlocksPerHost:0, HostReservedAttrIPv4s:(*ipam.HostReservedAttr)(nil), HostReservedAttrIPv6s:(*ipam.HostReservedAttr)(nil), IntendedUse:"Workload"} Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.548 [INFO][7144] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.548 [INFO][7144] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.548 [INFO][7144] ipam/ipam.go 107: Auto-assign 1 ipv4, 0 ipv6 addrs for host 'ci-4186.1.0-a-a7402f3063' Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.554 [INFO][7144] ipam/ipam.go 660: Looking up existing affinities for host handle="k8s-pod-network.8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.561 [INFO][7144] ipam/ipam.go 372: Looking up existing affinities for host host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.567 [INFO][7144] ipam/ipam.go 489: Trying affinity for 192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.569 [INFO][7144] ipam/ipam.go 155: Attempting to load block cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.572 [INFO][7144] ipam/ipam.go 232: Affinity is confirmed and block has been loaded cidr=192.168.102.64/26 host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.574 [INFO][7144] ipam/ipam.go 1180: Attempting to assign 1 addresses from block block=192.168.102.64/26 handle="k8s-pod-network.8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.576 [INFO][7144] ipam/ipam.go 1685: Creating new handle: k8s-pod-network.8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.583 [INFO][7144] ipam/ipam.go 1203: Writing block in order to claim IPs block=192.168.102.64/26 handle="k8s-pod-network.8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.593 [INFO][7144] ipam/ipam.go 1216: Successfully claimed IPs: [192.168.102.71/26] block=192.168.102.64/26 handle="k8s-pod-network.8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.593 [INFO][7144] ipam/ipam.go 847: Auto-assigned 1 out of 1 IPv4s: [192.168.102.71/26] handle="k8s-pod-network.8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" host="ci-4186.1.0-a-a7402f3063" Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.593 [INFO][7144] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:25:36.656324 containerd[1781]: 2025-01-30 13:25:36.593 [INFO][7144] ipam/ipam_plugin.go 283: Calico CNI IPAM assigned addresses IPv4=[192.168.102.71/26] IPv6=[] ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" HandleID="k8s-pod-network.8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" Jan 30 13:25:36.659817 containerd[1781]: 2025-01-30 13:25:36.596 [INFO][7106] cni-plugin/k8s.go 386: Populated endpoint ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Namespace="calico-system" Pod="calico-kube-controllers-ccb9d4c56-jglzv" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0", GenerateName:"calico-kube-controllers-ccb9d4c56-", Namespace:"calico-system", SelfLink:"", UID:"596b2831-ccf6-4e62-a56c-427ece32d593", ResourceVersion:"1097", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 25, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"ccb9d4c56", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"", Pod:"calico-kube-controllers-ccb9d4c56-jglzv", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.102.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali43a59e7f348", MAC:"", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:36.659817 containerd[1781]: 2025-01-30 13:25:36.596 [INFO][7106] cni-plugin/k8s.go 387: Calico CNI using IPs: [192.168.102.71/32] ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Namespace="calico-system" Pod="calico-kube-controllers-ccb9d4c56-jglzv" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" Jan 30 13:25:36.659817 containerd[1781]: 2025-01-30 13:25:36.596 [INFO][7106] cni-plugin/dataplane_linux.go 69: Setting the host side veth name to cali43a59e7f348 ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Namespace="calico-system" Pod="calico-kube-controllers-ccb9d4c56-jglzv" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" Jan 30 13:25:36.659817 containerd[1781]: 2025-01-30 13:25:36.602 [INFO][7106] cni-plugin/dataplane_linux.go 508: Disabling IPv4 forwarding ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Namespace="calico-system" Pod="calico-kube-controllers-ccb9d4c56-jglzv" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" Jan 30 13:25:36.659817 containerd[1781]: 2025-01-30 13:25:36.628 [INFO][7106] cni-plugin/k8s.go 414: Added Mac, interface name, and active container ID to endpoint ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Namespace="calico-system" Pod="calico-kube-controllers-ccb9d4c56-jglzv" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" endpoint=&v3.WorkloadEndpoint{TypeMeta:v1.TypeMeta{Kind:"WorkloadEndpoint", APIVersion:"projectcalico.org/v3"}, ObjectMeta:v1.ObjectMeta{Name:"ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0", GenerateName:"calico-kube-controllers-ccb9d4c56-", Namespace:"calico-system", SelfLink:"", UID:"596b2831-ccf6-4e62-a56c-427ece32d593", ResourceVersion:"1097", Generation:0, CreationTimestamp:time.Date(2025, time.January, 30, 13, 25, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"app.kubernetes.io/name":"calico-kube-controllers", "k8s-app":"calico-kube-controllers", "pod-template-hash":"ccb9d4c56", "projectcalico.org/namespace":"calico-system", "projectcalico.org/orchestrator":"k8s", "projectcalico.org/serviceaccount":"calico-kube-controllers"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, Spec:v3.WorkloadEndpointSpec{Orchestrator:"k8s", Workload:"", Node:"ci-4186.1.0-a-a7402f3063", ContainerID:"8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d", Pod:"calico-kube-controllers-ccb9d4c56-jglzv", Endpoint:"eth0", ServiceAccountName:"calico-kube-controllers", IPNetworks:[]string{"192.168.102.71/32"}, IPNATs:[]v3.IPNAT(nil), IPv4Gateway:"", IPv6Gateway:"", Profiles:[]string{"kns.calico-system", "ksa.calico-system.calico-kube-controllers"}, InterfaceName:"cali43a59e7f348", MAC:"d2:2b:50:26:46:cd", Ports:[]v3.WorkloadEndpointPort(nil), AllowSpoofedSourcePrefixes:[]string(nil)}} Jan 30 13:25:36.659817 containerd[1781]: 2025-01-30 13:25:36.648 [INFO][7106] cni-plugin/k8s.go 500: Wrote updated endpoint to datastore ContainerID="8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d" Namespace="calico-system" Pod="calico-kube-controllers-ccb9d4c56-jglzv" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--ccb9d4c56--jglzv-eth0" Jan 30 13:25:36.659960 systemd[1]: Started cri-containerd-a257e14f2ec67baa3ccb5e81c75bf86f7044d78d62b95a750347341c72a26d9e.scope - libcontainer container a257e14f2ec67baa3ccb5e81c75bf86f7044d78d62b95a750347341c72a26d9e. Jan 30 13:25:36.692991 containerd[1781]: time="2025-01-30T13:25:36.691978360Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Jan 30 13:25:36.692991 containerd[1781]: time="2025-01-30T13:25:36.692047320Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Jan 30 13:25:36.692991 containerd[1781]: time="2025-01-30T13:25:36.692062600Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:36.692991 containerd[1781]: time="2025-01-30T13:25:36.692155680Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.pause\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Jan 30 13:25:36.724568 systemd[1]: Started cri-containerd-8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d.scope - libcontainer container 8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d. Jan 30 13:25:36.752273 containerd[1781]: time="2025-01-30T13:25:36.751474537Z" level=info msg="StartContainer for \"a257e14f2ec67baa3ccb5e81c75bf86f7044d78d62b95a750347341c72a26d9e\" returns successfully" Jan 30 13:25:36.809369 containerd[1781]: time="2025-01-30T13:25:36.809032873Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:calico-kube-controllers-ccb9d4c56-jglzv,Uid:596b2831-ccf6-4e62-a56c-427ece32d593,Namespace:calico-system,Attempt:0,} returns sandbox id \"8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d\"" Jan 30 13:25:36.835090 containerd[1781]: time="2025-01-30T13:25:36.834946601Z" level=info msg="CreateContainer within sandbox \"8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d\" for container &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,}" Jan 30 13:25:36.858072 kubelet[3455]: I0130 13:25:36.857071 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-node-6sz7h" podStartSLOduration=3.857052127 podStartE2EDuration="3.857052127s" podCreationTimestamp="2025-01-30 13:25:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:25:36.830295679 +0000 UTC m=+74.059892746" watchObservedRunningTime="2025-01-30 13:25:36.857052127 +0000 UTC m=+74.086649114" Jan 30 13:25:36.878164 containerd[1781]: time="2025-01-30T13:25:36.878107533Z" level=info msg="CreateContainer within sandbox \"8231bc68ca9725edf5216f8a765e58e6517641f479f1f7c31ece3f438d703e8d\" for &ContainerMetadata{Name:calico-kube-controllers,Attempt:0,} returns container id \"b6116d0269663fc2493916fe1b83e49ac9b4f0c8b207b8e2af0d724ff77e9b2b\"" Jan 30 13:25:36.879159 containerd[1781]: time="2025-01-30T13:25:36.879117253Z" level=info msg="StartContainer for \"b6116d0269663fc2493916fe1b83e49ac9b4f0c8b207b8e2af0d724ff77e9b2b\"" Jan 30 13:25:36.922954 systemd[1]: Started cri-containerd-b6116d0269663fc2493916fe1b83e49ac9b4f0c8b207b8e2af0d724ff77e9b2b.scope - libcontainer container b6116d0269663fc2493916fe1b83e49ac9b4f0c8b207b8e2af0d724ff77e9b2b. Jan 30 13:25:37.039573 containerd[1781]: time="2025-01-30T13:25:37.039356019Z" level=info msg="StartContainer for \"b6116d0269663fc2493916fe1b83e49ac9b4f0c8b207b8e2af0d724ff77e9b2b\" returns successfully" Jan 30 13:25:37.826899 kubelet[3455]: I0130 13:25:37.826312 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-kube-controllers-ccb9d4c56-jglzv" podStartSLOduration=3.826289325 podStartE2EDuration="3.826289325s" podCreationTimestamp="2025-01-30 13:25:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:25:37.823416924 +0000 UTC m=+75.053013871" watchObservedRunningTime="2025-01-30 13:25:37.826289325 +0000 UTC m=+75.055886312" Jan 30 13:25:37.826899 kubelet[3455]: I0130 13:25:37.826545 3455 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="calico-system/calico-typha-96946fff6-qrhfp" podStartSLOduration=5.826535405 podStartE2EDuration="5.826535405s" podCreationTimestamp="2025-01-30 13:25:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-01-30 13:25:36.858970528 +0000 UTC m=+74.088567515" watchObservedRunningTime="2025-01-30 13:25:37.826535405 +0000 UTC m=+75.056132392" Jan 30 13:25:38.053839 systemd-networkd[1549]: cali43a59e7f348: Gained IPv6LL Jan 30 13:25:51.193646 systemd[1]: Started sshd@7-10.200.20.40:22-10.200.16.10:47508.service - OpenSSH per-connection server daemon (10.200.16.10:47508). Jan 30 13:25:51.626581 kubelet[3455]: I0130 13:25:51.626542 3455 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Jan 30 13:25:51.661489 sshd[7619]: Accepted publickey for core from 10.200.16.10 port 47508 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:25:51.664210 sshd-session[7619]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:25:51.673032 systemd-logind[1717]: New session 10 of user core. Jan 30 13:25:51.677925 systemd[1]: Started session-10.scope - Session 10 of User core. Jan 30 13:25:52.117534 sshd[7622]: Connection closed by 10.200.16.10 port 47508 Jan 30 13:25:52.118066 sshd-session[7619]: pam_unix(sshd:session): session closed for user core Jan 30 13:25:52.123022 systemd[1]: sshd@7-10.200.20.40:22-10.200.16.10:47508.service: Deactivated successfully. Jan 30 13:25:52.129559 systemd[1]: session-10.scope: Deactivated successfully. Jan 30 13:25:52.132542 systemd-logind[1717]: Session 10 logged out. Waiting for processes to exit. Jan 30 13:25:52.133867 systemd-logind[1717]: Removed session 10. Jan 30 13:25:57.206445 systemd[1]: Started sshd@8-10.200.20.40:22-10.200.16.10:35928.service - OpenSSH per-connection server daemon (10.200.16.10:35928). Jan 30 13:25:57.640364 sshd[7636]: Accepted publickey for core from 10.200.16.10 port 35928 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:25:57.641808 sshd-session[7636]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:25:57.646891 systemd-logind[1717]: New session 11 of user core. Jan 30 13:25:57.650901 systemd[1]: Started session-11.scope - Session 11 of User core. Jan 30 13:25:58.042816 sshd[7638]: Connection closed by 10.200.16.10 port 35928 Jan 30 13:25:58.043425 sshd-session[7636]: pam_unix(sshd:session): session closed for user core Jan 30 13:25:58.046758 systemd[1]: sshd@8-10.200.20.40:22-10.200.16.10:35928.service: Deactivated successfully. Jan 30 13:25:58.048526 systemd[1]: session-11.scope: Deactivated successfully. Jan 30 13:25:58.050091 systemd-logind[1717]: Session 11 logged out. Waiting for processes to exit. Jan 30 13:25:58.051765 systemd-logind[1717]: Removed session 11. Jan 30 13:26:03.126025 systemd[1]: Started sshd@9-10.200.20.40:22-10.200.16.10:35936.service - OpenSSH per-connection server daemon (10.200.16.10:35936). Jan 30 13:26:03.582680 sshd[7657]: Accepted publickey for core from 10.200.16.10 port 35936 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:03.583403 sshd-session[7657]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:03.589175 systemd-logind[1717]: New session 12 of user core. Jan 30 13:26:03.595528 systemd[1]: Started session-12.scope - Session 12 of User core. Jan 30 13:26:03.989496 sshd[7659]: Connection closed by 10.200.16.10 port 35936 Jan 30 13:26:03.988603 sshd-session[7657]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:03.993502 systemd[1]: sshd@9-10.200.20.40:22-10.200.16.10:35936.service: Deactivated successfully. Jan 30 13:26:03.996798 systemd[1]: session-12.scope: Deactivated successfully. Jan 30 13:26:03.998474 systemd-logind[1717]: Session 12 logged out. Waiting for processes to exit. Jan 30 13:26:04.000094 systemd-logind[1717]: Removed session 12. Jan 30 13:26:04.076199 systemd[1]: Started sshd@10-10.200.20.40:22-10.200.16.10:35944.service - OpenSSH per-connection server daemon (10.200.16.10:35944). Jan 30 13:26:04.509735 sshd[7693]: Accepted publickey for core from 10.200.16.10 port 35944 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:04.511330 sshd-session[7693]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:04.518354 systemd-logind[1717]: New session 13 of user core. Jan 30 13:26:04.521041 systemd[1]: Started session-13.scope - Session 13 of User core. Jan 30 13:26:04.934615 sshd[7695]: Connection closed by 10.200.16.10 port 35944 Jan 30 13:26:04.935558 sshd-session[7693]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:04.938393 systemd[1]: session-13.scope: Deactivated successfully. Jan 30 13:26:04.940272 systemd[1]: sshd@10-10.200.20.40:22-10.200.16.10:35944.service: Deactivated successfully. Jan 30 13:26:04.944640 systemd-logind[1717]: Session 13 logged out. Waiting for processes to exit. Jan 30 13:26:04.946301 systemd-logind[1717]: Removed session 13. Jan 30 13:26:05.026035 systemd[1]: Started sshd@11-10.200.20.40:22-10.200.16.10:35946.service - OpenSSH per-connection server daemon (10.200.16.10:35946). Jan 30 13:26:05.476992 sshd[7703]: Accepted publickey for core from 10.200.16.10 port 35946 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:05.478430 sshd-session[7703]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:05.483391 systemd-logind[1717]: New session 14 of user core. Jan 30 13:26:05.491900 systemd[1]: Started session-14.scope - Session 14 of User core. Jan 30 13:26:05.883095 sshd[7705]: Connection closed by 10.200.16.10 port 35946 Jan 30 13:26:05.883432 sshd-session[7703]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:05.888051 systemd[1]: sshd@11-10.200.20.40:22-10.200.16.10:35946.service: Deactivated successfully. Jan 30 13:26:05.890422 systemd[1]: session-14.scope: Deactivated successfully. Jan 30 13:26:05.891367 systemd-logind[1717]: Session 14 logged out. Waiting for processes to exit. Jan 30 13:26:05.892391 systemd-logind[1717]: Removed session 14. Jan 30 13:26:10.968040 systemd[1]: Started sshd@12-10.200.20.40:22-10.200.16.10:50000.service - OpenSSH per-connection server daemon (10.200.16.10:50000). Jan 30 13:26:11.404922 sshd[7739]: Accepted publickey for core from 10.200.16.10 port 50000 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:11.405327 sshd-session[7739]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:11.410352 systemd-logind[1717]: New session 15 of user core. Jan 30 13:26:11.415919 systemd[1]: Started session-15.scope - Session 15 of User core. Jan 30 13:26:11.823828 sshd[7741]: Connection closed by 10.200.16.10 port 50000 Jan 30 13:26:11.823738 sshd-session[7739]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:11.828912 systemd-logind[1717]: Session 15 logged out. Waiting for processes to exit. Jan 30 13:26:11.829208 systemd[1]: sshd@12-10.200.20.40:22-10.200.16.10:50000.service: Deactivated successfully. Jan 30 13:26:11.831522 systemd[1]: session-15.scope: Deactivated successfully. Jan 30 13:26:11.833604 systemd-logind[1717]: Removed session 15. Jan 30 13:26:16.906820 systemd[1]: Started sshd@13-10.200.20.40:22-10.200.16.10:54022.service - OpenSSH per-connection server daemon (10.200.16.10:54022). Jan 30 13:26:17.348850 sshd[7752]: Accepted publickey for core from 10.200.16.10 port 54022 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:17.350264 sshd-session[7752]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:17.354703 systemd-logind[1717]: New session 16 of user core. Jan 30 13:26:17.357910 systemd[1]: Started session-16.scope - Session 16 of User core. Jan 30 13:26:17.774688 sshd[7754]: Connection closed by 10.200.16.10 port 54022 Jan 30 13:26:17.775474 sshd-session[7752]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:17.783562 systemd[1]: sshd@13-10.200.20.40:22-10.200.16.10:54022.service: Deactivated successfully. Jan 30 13:26:17.785574 systemd[1]: session-16.scope: Deactivated successfully. Jan 30 13:26:17.787533 systemd-logind[1717]: Session 16 logged out. Waiting for processes to exit. Jan 30 13:26:17.790534 systemd-logind[1717]: Removed session 16. Jan 30 13:26:22.861016 systemd[1]: Started sshd@14-10.200.20.40:22-10.200.16.10:54024.service - OpenSSH per-connection server daemon (10.200.16.10:54024). Jan 30 13:26:23.294071 sshd[7781]: Accepted publickey for core from 10.200.16.10 port 54024 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:23.295423 sshd-session[7781]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:23.299543 systemd-logind[1717]: New session 17 of user core. Jan 30 13:26:23.309929 systemd[1]: Started session-17.scope - Session 17 of User core. Jan 30 13:26:23.479589 kubelet[3455]: I0130 13:26:23.479035 3455 scope.go:117] "RemoveContainer" containerID="5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2" Jan 30 13:26:23.481479 containerd[1781]: time="2025-01-30T13:26:23.481131903Z" level=info msg="RemoveContainer for \"5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2\"" Jan 30 13:26:23.491710 containerd[1781]: time="2025-01-30T13:26:23.491607863Z" level=info msg="RemoveContainer for \"5302d532f4db5eb3cc89d77b282e1477a9f1c4dbcfde2476d4196e683ecf0dd2\" returns successfully" Jan 30 13:26:23.493967 containerd[1781]: time="2025-01-30T13:26:23.493673863Z" level=info msg="StopPodSandbox for \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\"" Jan 30 13:26:23.493967 containerd[1781]: time="2025-01-30T13:26:23.493791223Z" level=info msg="TearDown network for sandbox \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\" successfully" Jan 30 13:26:23.493967 containerd[1781]: time="2025-01-30T13:26:23.493804303Z" level=info msg="StopPodSandbox for \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\" returns successfully" Jan 30 13:26:23.494575 containerd[1781]: time="2025-01-30T13:26:23.494393223Z" level=info msg="RemovePodSandbox for \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\"" Jan 30 13:26:23.494575 containerd[1781]: time="2025-01-30T13:26:23.494417623Z" level=info msg="Forcibly stopping sandbox \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\"" Jan 30 13:26:23.494575 containerd[1781]: time="2025-01-30T13:26:23.494464183Z" level=info msg="TearDown network for sandbox \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\" successfully" Jan 30 13:26:23.501923 containerd[1781]: time="2025-01-30T13:26:23.501879143Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:26:23.502123 containerd[1781]: time="2025-01-30T13:26:23.502105743Z" level=info msg="RemovePodSandbox \"f6785a18251e71cb2545dbe69ccfb8d4f9a10c28f675e014b726d36b81fda330\" returns successfully" Jan 30 13:26:23.502640 containerd[1781]: time="2025-01-30T13:26:23.502618903Z" level=info msg="StopPodSandbox for \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\"" Jan 30 13:26:23.502849 containerd[1781]: time="2025-01-30T13:26:23.502834023Z" level=info msg="TearDown network for sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" successfully" Jan 30 13:26:23.502925 containerd[1781]: time="2025-01-30T13:26:23.502912463Z" level=info msg="StopPodSandbox for \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" returns successfully" Jan 30 13:26:23.503340 containerd[1781]: time="2025-01-30T13:26:23.503312823Z" level=info msg="RemovePodSandbox for \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\"" Jan 30 13:26:23.503404 containerd[1781]: time="2025-01-30T13:26:23.503342863Z" level=info msg="Forcibly stopping sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\"" Jan 30 13:26:23.503404 containerd[1781]: time="2025-01-30T13:26:23.503400223Z" level=info msg="TearDown network for sandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" successfully" Jan 30 13:26:23.510250 containerd[1781]: time="2025-01-30T13:26:23.510192703Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:26:23.510407 containerd[1781]: time="2025-01-30T13:26:23.510273223Z" level=info msg="RemovePodSandbox \"8122d32f9d77150547ffc0bde86ac463142f9e7fe4c0413f94961b75091c5618\" returns successfully" Jan 30 13:26:23.511146 containerd[1781]: time="2025-01-30T13:26:23.510841783Z" level=info msg="StopPodSandbox for \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\"" Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.561 [WARNING][7798] cni-plugin/k8s.go 566: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.561 [INFO][7798] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.561 [INFO][7798] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" iface="eth0" netns="" Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.561 [INFO][7798] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.561 [INFO][7798] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.586 [INFO][7810] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.586 [INFO][7810] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.586 [INFO][7810] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.598 [WARNING][7810] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.598 [INFO][7810] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.599 [INFO][7810] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:26:23.602698 containerd[1781]: 2025-01-30 13:26:23.601 [INFO][7798] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:26:23.602698 containerd[1781]: time="2025-01-30T13:26:23.602567743Z" level=info msg="TearDown network for sandbox \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\" successfully" Jan 30 13:26:23.602698 containerd[1781]: time="2025-01-30T13:26:23.602593743Z" level=info msg="StopPodSandbox for \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\" returns successfully" Jan 30 13:26:23.604597 containerd[1781]: time="2025-01-30T13:26:23.604228103Z" level=info msg="RemovePodSandbox for \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\"" Jan 30 13:26:23.604597 containerd[1781]: time="2025-01-30T13:26:23.604261343Z" level=info msg="Forcibly stopping sandbox \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\"" Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.657 [WARNING][7829] cni-plugin/k8s.go 566: WorkloadEndpoint does not exist in the datastore, moving forward with the clean up ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" WorkloadEndpoint="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.657 [INFO][7829] cni-plugin/k8s.go 608: Cleaning up netns ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.657 [INFO][7829] cni-plugin/dataplane_linux.go 555: CleanUpNamespace called with no netns name, ignoring. ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" iface="eth0" netns="" Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.657 [INFO][7829] cni-plugin/k8s.go 615: Releasing IP address(es) ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.657 [INFO][7829] cni-plugin/utils.go 188: Calico CNI releasing IP address ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.686 [INFO][7835] ipam/ipam_plugin.go 412: Releasing address using handleID ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.687 [INFO][7835] ipam/ipam_plugin.go 353: About to acquire host-wide IPAM lock. Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.687 [INFO][7835] ipam/ipam_plugin.go 368: Acquired host-wide IPAM lock. Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.698 [WARNING][7835] ipam/ipam_plugin.go 429: Asked to release address but it doesn't exist. Ignoring ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.698 [INFO][7835] ipam/ipam_plugin.go 440: Releasing address using workloadID ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" HandleID="k8s-pod-network.39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Workload="ci--4186.1.0--a--a7402f3063-k8s-calico--kube--controllers--6dd45f4c57--58wrn-eth0" Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.700 [INFO][7835] ipam/ipam_plugin.go 374: Released host-wide IPAM lock. Jan 30 13:26:23.706211 containerd[1781]: 2025-01-30 13:26:23.703 [INFO][7829] cni-plugin/k8s.go 621: Teardown processing complete. ContainerID="39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5" Jan 30 13:26:23.708846 containerd[1781]: time="2025-01-30T13:26:23.706690142Z" level=info msg="TearDown network for sandbox \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\" successfully" Jan 30 13:26:23.717691 containerd[1781]: time="2025-01-30T13:26:23.717640942Z" level=warning msg="Failed to get podSandbox status for container event for sandboxID \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\": an error occurred when try to find sandbox: not found. Sending the event with nil podSandboxStatus." Jan 30 13:26:23.717956 containerd[1781]: time="2025-01-30T13:26:23.717935822Z" level=info msg="RemovePodSandbox \"39314b8b7f6643c160893d2240ad2949a0724c6035c143baa0687a87d350fee5\" returns successfully" Jan 30 13:26:23.735639 sshd[7785]: Connection closed by 10.200.16.10 port 54024 Jan 30 13:26:23.737611 sshd-session[7781]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:23.742650 systemd[1]: session-17.scope: Deactivated successfully. Jan 30 13:26:23.744725 systemd-logind[1717]: Session 17 logged out. Waiting for processes to exit. Jan 30 13:26:23.746507 systemd[1]: sshd@14-10.200.20.40:22-10.200.16.10:54024.service: Deactivated successfully. Jan 30 13:26:23.751102 systemd-logind[1717]: Removed session 17. Jan 30 13:26:28.822019 systemd[1]: Started sshd@15-10.200.20.40:22-10.200.16.10:52310.service - OpenSSH per-connection server daemon (10.200.16.10:52310). Jan 30 13:26:29.258058 sshd[7845]: Accepted publickey for core from 10.200.16.10 port 52310 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:29.258576 sshd-session[7845]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:29.262765 systemd-logind[1717]: New session 18 of user core. Jan 30 13:26:29.269952 systemd[1]: Started session-18.scope - Session 18 of User core. Jan 30 13:26:29.656212 sshd[7847]: Connection closed by 10.200.16.10 port 52310 Jan 30 13:26:29.655637 sshd-session[7845]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:29.659210 systemd[1]: sshd@15-10.200.20.40:22-10.200.16.10:52310.service: Deactivated successfully. Jan 30 13:26:29.661691 systemd[1]: session-18.scope: Deactivated successfully. Jan 30 13:26:29.663317 systemd-logind[1717]: Session 18 logged out. Waiting for processes to exit. Jan 30 13:26:29.664678 systemd-logind[1717]: Removed session 18. Jan 30 13:26:29.745059 systemd[1]: Started sshd@16-10.200.20.40:22-10.200.16.10:52320.service - OpenSSH per-connection server daemon (10.200.16.10:52320). Jan 30 13:26:30.176755 sshd[7858]: Accepted publickey for core from 10.200.16.10 port 52320 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:30.178144 sshd-session[7858]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:30.182084 systemd-logind[1717]: New session 19 of user core. Jan 30 13:26:30.186887 systemd[1]: Started session-19.scope - Session 19 of User core. Jan 30 13:26:30.678943 sshd[7860]: Connection closed by 10.200.16.10 port 52320 Jan 30 13:26:30.680169 sshd-session[7858]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:30.684187 systemd[1]: sshd@16-10.200.20.40:22-10.200.16.10:52320.service: Deactivated successfully. Jan 30 13:26:30.687631 systemd[1]: session-19.scope: Deactivated successfully. Jan 30 13:26:30.688540 systemd-logind[1717]: Session 19 logged out. Waiting for processes to exit. Jan 30 13:26:30.689624 systemd-logind[1717]: Removed session 19. Jan 30 13:26:30.762043 systemd[1]: Started sshd@17-10.200.20.40:22-10.200.16.10:52330.service - OpenSSH per-connection server daemon (10.200.16.10:52330). Jan 30 13:26:31.196100 sshd[7869]: Accepted publickey for core from 10.200.16.10 port 52330 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:31.197946 sshd-session[7869]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:31.202496 systemd-logind[1717]: New session 20 of user core. Jan 30 13:26:31.211056 systemd[1]: Started session-20.scope - Session 20 of User core. Jan 30 13:26:33.393791 sshd[7871]: Connection closed by 10.200.16.10 port 52330 Jan 30 13:26:33.394278 sshd-session[7869]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:33.398457 systemd-logind[1717]: Session 20 logged out. Waiting for processes to exit. Jan 30 13:26:33.399166 systemd[1]: sshd@17-10.200.20.40:22-10.200.16.10:52330.service: Deactivated successfully. Jan 30 13:26:33.403001 systemd[1]: session-20.scope: Deactivated successfully. Jan 30 13:26:33.404219 systemd-logind[1717]: Removed session 20. Jan 30 13:26:33.478011 systemd[1]: Started sshd@18-10.200.20.40:22-10.200.16.10:52334.service - OpenSSH per-connection server daemon (10.200.16.10:52334). Jan 30 13:26:33.912733 sshd[7887]: Accepted publickey for core from 10.200.16.10 port 52334 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:33.914252 sshd-session[7887]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:33.918868 systemd-logind[1717]: New session 21 of user core. Jan 30 13:26:33.923893 systemd[1]: Started session-21.scope - Session 21 of User core. Jan 30 13:26:34.408439 sshd[7889]: Connection closed by 10.200.16.10 port 52334 Jan 30 13:26:34.407937 sshd-session[7887]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:34.410952 systemd[1]: sshd@18-10.200.20.40:22-10.200.16.10:52334.service: Deactivated successfully. Jan 30 13:26:34.413480 systemd[1]: session-21.scope: Deactivated successfully. Jan 30 13:26:34.415026 systemd-logind[1717]: Session 21 logged out. Waiting for processes to exit. Jan 30 13:26:34.416852 systemd-logind[1717]: Removed session 21. Jan 30 13:26:34.494132 systemd[1]: Started sshd@19-10.200.20.40:22-10.200.16.10:52340.service - OpenSSH per-connection server daemon (10.200.16.10:52340). Jan 30 13:26:34.928844 sshd[7920]: Accepted publickey for core from 10.200.16.10 port 52340 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:34.930385 sshd-session[7920]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:34.934931 systemd-logind[1717]: New session 22 of user core. Jan 30 13:26:34.939913 systemd[1]: Started session-22.scope - Session 22 of User core. Jan 30 13:26:35.322798 sshd[7922]: Connection closed by 10.200.16.10 port 52340 Jan 30 13:26:35.323478 sshd-session[7920]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:35.326809 systemd[1]: sshd@19-10.200.20.40:22-10.200.16.10:52340.service: Deactivated successfully. Jan 30 13:26:35.330978 systemd[1]: session-22.scope: Deactivated successfully. Jan 30 13:26:35.332833 systemd-logind[1717]: Session 22 logged out. Waiting for processes to exit. Jan 30 13:26:35.334549 systemd-logind[1717]: Removed session 22. Jan 30 13:26:40.411843 systemd[1]: Started sshd@20-10.200.20.40:22-10.200.16.10:49896.service - OpenSSH per-connection server daemon (10.200.16.10:49896). Jan 30 13:26:40.848422 sshd[7976]: Accepted publickey for core from 10.200.16.10 port 49896 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:40.850227 sshd-session[7976]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:40.854339 systemd-logind[1717]: New session 23 of user core. Jan 30 13:26:40.866031 systemd[1]: Started session-23.scope - Session 23 of User core. Jan 30 13:26:41.249747 sshd[7978]: Connection closed by 10.200.16.10 port 49896 Jan 30 13:26:41.250447 sshd-session[7976]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:41.254173 systemd-logind[1717]: Session 23 logged out. Waiting for processes to exit. Jan 30 13:26:41.254477 systemd[1]: sshd@20-10.200.20.40:22-10.200.16.10:49896.service: Deactivated successfully. Jan 30 13:26:41.257463 systemd[1]: session-23.scope: Deactivated successfully. Jan 30 13:26:41.260291 systemd-logind[1717]: Removed session 23. Jan 30 13:26:46.339023 systemd[1]: Started sshd@21-10.200.20.40:22-10.200.16.10:47070.service - OpenSSH per-connection server daemon (10.200.16.10:47070). Jan 30 13:26:46.794779 sshd[7989]: Accepted publickey for core from 10.200.16.10 port 47070 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:46.796358 sshd-session[7989]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:46.800519 systemd-logind[1717]: New session 24 of user core. Jan 30 13:26:46.806923 systemd[1]: Started session-24.scope - Session 24 of User core. Jan 30 13:26:47.191839 sshd[7991]: Connection closed by 10.200.16.10 port 47070 Jan 30 13:26:47.192367 sshd-session[7989]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:47.196064 systemd[1]: sshd@21-10.200.20.40:22-10.200.16.10:47070.service: Deactivated successfully. Jan 30 13:26:47.199699 systemd[1]: session-24.scope: Deactivated successfully. Jan 30 13:26:47.201279 systemd-logind[1717]: Session 24 logged out. Waiting for processes to exit. Jan 30 13:26:47.202293 systemd-logind[1717]: Removed session 24. Jan 30 13:26:52.281704 systemd[1]: Started sshd@22-10.200.20.40:22-10.200.16.10:47074.service - OpenSSH per-connection server daemon (10.200.16.10:47074). Jan 30 13:26:52.740748 sshd[8001]: Accepted publickey for core from 10.200.16.10 port 47074 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:52.742121 sshd-session[8001]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:52.747234 systemd-logind[1717]: New session 25 of user core. Jan 30 13:26:52.754899 systemd[1]: Started session-25.scope - Session 25 of User core. Jan 30 13:26:53.140844 sshd[8003]: Connection closed by 10.200.16.10 port 47074 Jan 30 13:26:53.141615 sshd-session[8001]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:53.145863 systemd[1]: sshd@22-10.200.20.40:22-10.200.16.10:47074.service: Deactivated successfully. Jan 30 13:26:53.148708 systemd[1]: session-25.scope: Deactivated successfully. Jan 30 13:26:53.149608 systemd-logind[1717]: Session 25 logged out. Waiting for processes to exit. Jan 30 13:26:53.151044 systemd-logind[1717]: Removed session 25. Jan 30 13:26:58.224028 systemd[1]: Started sshd@23-10.200.20.40:22-10.200.16.10:41068.service - OpenSSH per-connection server daemon (10.200.16.10:41068). Jan 30 13:26:58.655871 sshd[8013]: Accepted publickey for core from 10.200.16.10 port 41068 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:26:58.657179 sshd-session[8013]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:26:58.661375 systemd-logind[1717]: New session 26 of user core. Jan 30 13:26:58.666895 systemd[1]: Started session-26.scope - Session 26 of User core. Jan 30 13:26:59.037035 sshd[8015]: Connection closed by 10.200.16.10 port 41068 Jan 30 13:26:59.037679 sshd-session[8013]: pam_unix(sshd:session): session closed for user core Jan 30 13:26:59.041641 systemd[1]: sshd@23-10.200.20.40:22-10.200.16.10:41068.service: Deactivated successfully. Jan 30 13:26:59.044149 systemd[1]: session-26.scope: Deactivated successfully. Jan 30 13:26:59.045505 systemd-logind[1717]: Session 26 logged out. Waiting for processes to exit. Jan 30 13:26:59.047359 systemd-logind[1717]: Removed session 26. Jan 30 13:27:04.116400 systemd[1]: Started sshd@24-10.200.20.40:22-10.200.16.10:41078.service - OpenSSH per-connection server daemon (10.200.16.10:41078). Jan 30 13:27:04.563792 sshd[8056]: Accepted publickey for core from 10.200.16.10 port 41078 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:27:04.564342 sshd-session[8056]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:27:04.569481 systemd-logind[1717]: New session 27 of user core. Jan 30 13:27:04.574018 systemd[1]: Started session-27.scope - Session 27 of User core. Jan 30 13:27:04.961287 sshd[8058]: Connection closed by 10.200.16.10 port 41078 Jan 30 13:27:04.961941 sshd-session[8056]: pam_unix(sshd:session): session closed for user core Jan 30 13:27:04.965513 systemd[1]: sshd@24-10.200.20.40:22-10.200.16.10:41078.service: Deactivated successfully. Jan 30 13:27:04.967912 systemd[1]: session-27.scope: Deactivated successfully. Jan 30 13:27:04.968761 systemd-logind[1717]: Session 27 logged out. Waiting for processes to exit. Jan 30 13:27:04.969691 systemd-logind[1717]: Removed session 27. Jan 30 13:27:10.047044 systemd[1]: Started sshd@25-10.200.20.40:22-10.200.16.10:35922.service - OpenSSH per-connection server daemon (10.200.16.10:35922). Jan 30 13:27:10.490775 sshd[8091]: Accepted publickey for core from 10.200.16.10 port 35922 ssh2: RSA SHA256:C5pjVMYzONmJhds0jUZO5MZNsVxbc+yYzbKaFYsva20 Jan 30 13:27:10.492296 sshd-session[8091]: pam_unix(sshd:session): session opened for user core(uid=500) by core(uid=0) Jan 30 13:27:10.502347 systemd-logind[1717]: New session 28 of user core. Jan 30 13:27:10.510387 systemd[1]: Started session-28.scope - Session 28 of User core. Jan 30 13:27:10.907840 sshd[8093]: Connection closed by 10.200.16.10 port 35922 Jan 30 13:27:10.908525 sshd-session[8091]: pam_unix(sshd:session): session closed for user core Jan 30 13:27:10.912585 systemd-logind[1717]: Session 28 logged out. Waiting for processes to exit. Jan 30 13:27:10.912911 systemd[1]: sshd@25-10.200.20.40:22-10.200.16.10:35922.service: Deactivated successfully. Jan 30 13:27:10.916777 systemd[1]: session-28.scope: Deactivated successfully. Jan 30 13:27:10.918988 systemd-logind[1717]: Removed session 28.