Apr 12 18:28:30.008903 kernel: Booting Linux on physical CPU 0x0000000000 [0x413fd0c1] Apr 12 18:28:30.008923 kernel: Linux version 5.15.154-flatcar (build@pony-truck.infra.kinvolk.io) (aarch64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP PREEMPT Fri Apr 12 17:21:24 -00 2024 Apr 12 18:28:30.008931 kernel: earlycon: pl11 at MMIO 0x00000000effec000 (options '') Apr 12 18:28:30.008941 kernel: printk: bootconsole [pl11] enabled Apr 12 18:28:30.008947 kernel: efi: EFI v2.70 by EDK II Apr 12 18:28:30.008952 kernel: efi: ACPI 2.0=0x3fd89018 SMBIOS=0x3fd66000 SMBIOS 3.0=0x3fd64000 MEMATTR=0x3ef3c198 RNG=0x3fd89998 MEMRESERVE=0x37b33f98 Apr 12 18:28:30.008959 kernel: random: crng init done Apr 12 18:28:30.008965 kernel: ACPI: Early table checksum verification disabled Apr 12 18:28:30.008970 kernel: ACPI: RSDP 0x000000003FD89018 000024 (v02 VRTUAL) Apr 12 18:28:30.008976 kernel: ACPI: XSDT 0x000000003FD89F18 00006C (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 12 18:28:30.008985 kernel: ACPI: FACP 0x000000003FD89C18 000114 (v06 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 12 18:28:30.008992 kernel: ACPI: DSDT 0x000000003EBD2018 01DEC0 (v02 MSFTVM DSDT01 00000001 MSFT 05000000) Apr 12 18:28:30.008997 kernel: ACPI: DBG2 0x000000003FD89B18 000072 (v00 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 12 18:28:30.009003 kernel: ACPI: GTDT 0x000000003FD89D98 000060 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 12 18:28:30.009010 kernel: ACPI: OEM0 0x000000003FD89098 000064 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 12 18:28:30.009016 kernel: ACPI: SPCR 0x000000003FD89A98 000050 (v02 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 12 18:28:30.009022 kernel: ACPI: APIC 0x000000003FD89818 0000FC (v04 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 12 18:28:30.009033 kernel: ACPI: SRAT 0x000000003FD89198 000234 (v03 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 12 18:28:30.009040 kernel: ACPI: PPTT 0x000000003FD89418 000120 (v01 VRTUAL MICROSFT 00000000 MSFT 00000000) Apr 12 18:28:30.009046 kernel: ACPI: BGRT 0x000000003FD89E98 000038 (v01 VRTUAL MICROSFT 00000001 MSFT 00000001) Apr 12 18:28:30.009052 kernel: ACPI: SPCR: console: pl011,mmio32,0xeffec000,115200 Apr 12 18:28:30.009058 kernel: NUMA: Failed to initialise from firmware Apr 12 18:28:30.009064 kernel: NUMA: Faking a node at [mem 0x0000000000000000-0x00000001bfffffff] Apr 12 18:28:30.009070 kernel: NUMA: NODE_DATA [mem 0x1bf7f1900-0x1bf7f6fff] Apr 12 18:28:30.009079 kernel: Zone ranges: Apr 12 18:28:30.009085 kernel: DMA [mem 0x0000000000000000-0x00000000ffffffff] Apr 12 18:28:30.009091 kernel: DMA32 empty Apr 12 18:28:30.009098 kernel: Normal [mem 0x0000000100000000-0x00000001bfffffff] Apr 12 18:28:30.009104 kernel: Movable zone start for each node Apr 12 18:28:30.009110 kernel: Early memory node ranges Apr 12 18:28:30.009116 kernel: node 0: [mem 0x0000000000000000-0x00000000007fffff] Apr 12 18:28:30.009125 kernel: node 0: [mem 0x0000000000824000-0x000000003ec80fff] Apr 12 18:28:30.009131 kernel: node 0: [mem 0x000000003ec81000-0x000000003eca9fff] Apr 12 18:28:30.009137 kernel: node 0: [mem 0x000000003ecaa000-0x000000003fd29fff] Apr 12 18:28:30.009143 kernel: node 0: [mem 0x000000003fd2a000-0x000000003fd7dfff] Apr 12 18:28:30.009149 kernel: node 0: [mem 0x000000003fd7e000-0x000000003fd89fff] Apr 12 18:28:30.009154 kernel: node 0: [mem 0x000000003fd8a000-0x000000003fd8dfff] Apr 12 18:28:30.009160 kernel: node 0: [mem 0x000000003fd8e000-0x000000003fffffff] Apr 12 18:28:30.009169 kernel: node 0: [mem 0x0000000100000000-0x00000001bfffffff] Apr 12 18:28:30.009176 kernel: Initmem setup node 0 [mem 0x0000000000000000-0x00000001bfffffff] Apr 12 18:28:30.009185 kernel: On node 0, zone DMA: 36 pages in unavailable ranges Apr 12 18:28:30.009191 kernel: psci: probing for conduit method from ACPI. Apr 12 18:28:30.009197 kernel: psci: PSCIv1.1 detected in firmware. Apr 12 18:28:30.009204 kernel: psci: Using standard PSCI v0.2 function IDs Apr 12 18:28:30.009211 kernel: psci: MIGRATE_INFO_TYPE not supported. Apr 12 18:28:30.009220 kernel: psci: SMC Calling Convention v1.4 Apr 12 18:28:30.009226 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x0 -> Node -1 Apr 12 18:28:30.009233 kernel: ACPI: NUMA: SRAT: PXM 0 -> MPIDR 0x1 -> Node -1 Apr 12 18:28:30.009239 kernel: percpu: Embedded 30 pages/cpu s83032 r8192 d31656 u122880 Apr 12 18:28:30.009245 kernel: pcpu-alloc: s83032 r8192 d31656 u122880 alloc=30*4096 Apr 12 18:28:30.009252 kernel: pcpu-alloc: [0] 0 [0] 1 Apr 12 18:28:30.009258 kernel: Detected PIPT I-cache on CPU0 Apr 12 18:28:30.009264 kernel: CPU features: detected: GIC system register CPU interface Apr 12 18:28:30.009273 kernel: CPU features: detected: Hardware dirty bit management Apr 12 18:28:30.009280 kernel: CPU features: detected: Spectre-BHB Apr 12 18:28:30.009286 kernel: CPU features: kernel page table isolation forced ON by KASLR Apr 12 18:28:30.009294 kernel: CPU features: detected: Kernel page table isolation (KPTI) Apr 12 18:28:30.009300 kernel: CPU features: detected: ARM erratum 1418040 Apr 12 18:28:30.009306 kernel: CPU features: detected: ARM erratum 1542419 (kernel portion) Apr 12 18:28:30.009312 kernel: Built 1 zonelists, mobility grouping on. Total pages: 1032156 Apr 12 18:28:30.009319 kernel: Policy zone: Normal Apr 12 18:28:30.009327 kernel: Kernel command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=c0b96868344262519ffdb2dae3782c942008a0fecdbc0bc85d2e170bd2e8b8a8 Apr 12 18:28:30.009336 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Apr 12 18:28:30.009343 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Apr 12 18:28:30.009349 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Apr 12 18:28:30.009355 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Apr 12 18:28:30.009363 kernel: software IO TLB: mapped [mem 0x000000003abd2000-0x000000003ebd2000] (64MB) Apr 12 18:28:30.009370 kernel: Memory: 3990260K/4194160K available (9792K kernel code, 2092K rwdata, 7568K rodata, 36352K init, 777K bss, 203900K reserved, 0K cma-reserved) Apr 12 18:28:30.009387 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=2, Nodes=1 Apr 12 18:28:30.009393 kernel: trace event string verifier disabled Apr 12 18:28:30.009399 kernel: rcu: Preemptible hierarchical RCU implementation. Apr 12 18:28:30.009409 kernel: rcu: RCU event tracing is enabled. Apr 12 18:28:30.009416 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=2. Apr 12 18:28:30.009422 kernel: Trampoline variant of Tasks RCU enabled. Apr 12 18:28:30.009429 kernel: Tracing variant of Tasks RCU enabled. Apr 12 18:28:30.009435 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Apr 12 18:28:30.009442 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=2 Apr 12 18:28:30.009452 kernel: NR_IRQS: 64, nr_irqs: 64, preallocated irqs: 0 Apr 12 18:28:30.009458 kernel: GICv3: 960 SPIs implemented Apr 12 18:28:30.009464 kernel: GICv3: 0 Extended SPIs implemented Apr 12 18:28:30.009471 kernel: GICv3: Distributor has no Range Selector support Apr 12 18:28:30.009477 kernel: Root IRQ handler: gic_handle_irq Apr 12 18:28:30.009483 kernel: GICv3: 16 PPIs implemented Apr 12 18:28:30.009490 kernel: GICv3: CPU0: found redistributor 0 region 0:0x00000000effee000 Apr 12 18:28:30.011479 kernel: ITS: No ITS available, not enabling LPIs Apr 12 18:28:30.011496 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 12 18:28:30.011504 kernel: arch_timer: cp15 timer(s) running at 25.00MHz (virt). Apr 12 18:28:30.011511 kernel: clocksource: arch_sys_counter: mask: 0xffffffffffffff max_cycles: 0x5c40939b5, max_idle_ns: 440795202646 ns Apr 12 18:28:30.011517 kernel: sched_clock: 56 bits at 25MHz, resolution 40ns, wraps every 4398046511100ns Apr 12 18:28:30.011529 kernel: Console: colour dummy device 80x25 Apr 12 18:28:30.011537 kernel: printk: console [tty1] enabled Apr 12 18:28:30.011543 kernel: ACPI: Core revision 20210730 Apr 12 18:28:30.011550 kernel: Calibrating delay loop (skipped), value calculated using timer frequency.. 50.00 BogoMIPS (lpj=25000) Apr 12 18:28:30.011557 kernel: pid_max: default: 32768 minimum: 301 Apr 12 18:28:30.011564 kernel: LSM: Security Framework initializing Apr 12 18:28:30.011570 kernel: SELinux: Initializing. Apr 12 18:28:30.011577 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 12 18:28:30.011584 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Apr 12 18:28:30.011592 kernel: Hyper-V: privilege flags low 0x2e7f, high 0x3a8030, hints 0xe, misc 0x31e1 Apr 12 18:28:30.011598 kernel: Hyper-V: Host Build 10.0.22477.1369-1-0 Apr 12 18:28:30.011605 kernel: rcu: Hierarchical SRCU implementation. Apr 12 18:28:30.011611 kernel: Remapping and enabling EFI services. Apr 12 18:28:30.011618 kernel: smp: Bringing up secondary CPUs ... Apr 12 18:28:30.011624 kernel: Detected PIPT I-cache on CPU1 Apr 12 18:28:30.011631 kernel: GICv3: CPU1: found redistributor 1 region 1:0x00000000f000e000 Apr 12 18:28:30.011638 kernel: arch_timer: Enabling local workaround for ARM erratum 1418040 Apr 12 18:28:30.011644 kernel: CPU1: Booted secondary processor 0x0000000001 [0x413fd0c1] Apr 12 18:28:30.011652 kernel: smp: Brought up 1 node, 2 CPUs Apr 12 18:28:30.011659 kernel: SMP: Total of 2 processors activated. Apr 12 18:28:30.011665 kernel: CPU features: detected: 32-bit EL0 Support Apr 12 18:28:30.011672 kernel: CPU features: detected: Instruction cache invalidation not required for I/D coherence Apr 12 18:28:30.011679 kernel: CPU features: detected: Data cache clean to the PoU not required for I/D coherence Apr 12 18:28:30.011686 kernel: CPU features: detected: CRC32 instructions Apr 12 18:28:30.011692 kernel: CPU features: detected: RCpc load-acquire (LDAPR) Apr 12 18:28:30.011699 kernel: CPU features: detected: LSE atomic instructions Apr 12 18:28:30.011706 kernel: CPU features: detected: Privileged Access Never Apr 12 18:28:30.011714 kernel: CPU: All CPU(s) started at EL1 Apr 12 18:28:30.011721 kernel: alternatives: patching kernel code Apr 12 18:28:30.011732 kernel: devtmpfs: initialized Apr 12 18:28:30.011740 kernel: KASLR enabled Apr 12 18:28:30.011747 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Apr 12 18:28:30.011754 kernel: futex hash table entries: 512 (order: 3, 32768 bytes, linear) Apr 12 18:28:30.011761 kernel: pinctrl core: initialized pinctrl subsystem Apr 12 18:28:30.011768 kernel: SMBIOS 3.1.0 present. Apr 12 18:28:30.011775 kernel: DMI: Microsoft Corporation Virtual Machine/Virtual Machine, BIOS Hyper-V UEFI Release v4.1 11/28/2023 Apr 12 18:28:30.011783 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Apr 12 18:28:30.011791 kernel: DMA: preallocated 512 KiB GFP_KERNEL pool for atomic allocations Apr 12 18:28:30.011799 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA pool for atomic allocations Apr 12 18:28:30.011806 kernel: DMA: preallocated 512 KiB GFP_KERNEL|GFP_DMA32 pool for atomic allocations Apr 12 18:28:30.011813 kernel: audit: initializing netlink subsys (disabled) Apr 12 18:28:30.011820 kernel: audit: type=2000 audit(0.091:1): state=initialized audit_enabled=0 res=1 Apr 12 18:28:30.011827 kernel: thermal_sys: Registered thermal governor 'step_wise' Apr 12 18:28:30.011834 kernel: cpuidle: using governor menu Apr 12 18:28:30.011842 kernel: hw-breakpoint: found 6 breakpoint and 4 watchpoint registers. Apr 12 18:28:30.011849 kernel: ASID allocator initialised with 32768 entries Apr 12 18:28:30.011856 kernel: ACPI: bus type PCI registered Apr 12 18:28:30.011863 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Apr 12 18:28:30.011870 kernel: Serial: AMBA PL011 UART driver Apr 12 18:28:30.011877 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Apr 12 18:28:30.011884 kernel: HugeTLB registered 32.0 MiB page size, pre-allocated 0 pages Apr 12 18:28:30.011891 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Apr 12 18:28:30.011898 kernel: HugeTLB registered 64.0 KiB page size, pre-allocated 0 pages Apr 12 18:28:30.011906 kernel: cryptd: max_cpu_qlen set to 1000 Apr 12 18:28:30.011913 kernel: alg: No test for fips(ansi_cprng) (fips_ansi_cprng) Apr 12 18:28:30.011920 kernel: ACPI: Added _OSI(Module Device) Apr 12 18:28:30.011927 kernel: ACPI: Added _OSI(Processor Device) Apr 12 18:28:30.011934 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Apr 12 18:28:30.011941 kernel: ACPI: Added _OSI(Processor Aggregator Device) Apr 12 18:28:30.011948 kernel: ACPI: Added _OSI(Linux-Dell-Video) Apr 12 18:28:30.011955 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Apr 12 18:28:30.011962 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Apr 12 18:28:30.011970 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Apr 12 18:28:30.011977 kernel: ACPI: Interpreter enabled Apr 12 18:28:30.011984 kernel: ACPI: Using GIC for interrupt routing Apr 12 18:28:30.011991 kernel: ARMH0011:00: ttyAMA0 at MMIO 0xeffec000 (irq = 12, base_baud = 0) is a SBSA Apr 12 18:28:30.011998 kernel: printk: console [ttyAMA0] enabled Apr 12 18:28:30.012005 kernel: printk: bootconsole [pl11] disabled Apr 12 18:28:30.012012 kernel: ARMH0011:01: ttyAMA1 at MMIO 0xeffeb000 (irq = 13, base_baud = 0) is a SBSA Apr 12 18:28:30.012019 kernel: iommu: Default domain type: Translated Apr 12 18:28:30.012026 kernel: iommu: DMA domain TLB invalidation policy: strict mode Apr 12 18:28:30.012034 kernel: vgaarb: loaded Apr 12 18:28:30.012041 kernel: pps_core: LinuxPPS API ver. 1 registered Apr 12 18:28:30.012048 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Apr 12 18:28:30.012055 kernel: PTP clock support registered Apr 12 18:28:30.012062 kernel: Registered efivars operations Apr 12 18:28:30.012069 kernel: No ACPI PMU IRQ for CPU0 Apr 12 18:28:30.012076 kernel: No ACPI PMU IRQ for CPU1 Apr 12 18:28:30.012083 kernel: clocksource: Switched to clocksource arch_sys_counter Apr 12 18:28:30.012090 kernel: VFS: Disk quotas dquot_6.6.0 Apr 12 18:28:30.012098 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Apr 12 18:28:30.012105 kernel: pnp: PnP ACPI init Apr 12 18:28:30.012112 kernel: pnp: PnP ACPI: found 0 devices Apr 12 18:28:30.012119 kernel: NET: Registered PF_INET protocol family Apr 12 18:28:30.012126 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Apr 12 18:28:30.012133 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Apr 12 18:28:30.012140 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Apr 12 18:28:30.012147 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Apr 12 18:28:30.012154 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) Apr 12 18:28:30.012163 kernel: TCP: Hash tables configured (established 32768 bind 32768) Apr 12 18:28:30.012170 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 12 18:28:30.012177 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Apr 12 18:28:30.012184 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Apr 12 18:28:30.012191 kernel: PCI: CLS 0 bytes, default 64 Apr 12 18:28:30.012198 kernel: hw perfevents: enabled with armv8_pmuv3_0 PMU driver, 1 counters available Apr 12 18:28:30.012205 kernel: kvm [1]: HYP mode not available Apr 12 18:28:30.012212 kernel: Initialise system trusted keyrings Apr 12 18:28:30.012218 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Apr 12 18:28:30.012227 kernel: Key type asymmetric registered Apr 12 18:28:30.012233 kernel: Asymmetric key parser 'x509' registered Apr 12 18:28:30.012240 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Apr 12 18:28:30.012247 kernel: io scheduler mq-deadline registered Apr 12 18:28:30.012254 kernel: io scheduler kyber registered Apr 12 18:28:30.012261 kernel: io scheduler bfq registered Apr 12 18:28:30.012268 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Apr 12 18:28:30.012275 kernel: thunder_xcv, ver 1.0 Apr 12 18:28:30.012282 kernel: thunder_bgx, ver 1.0 Apr 12 18:28:30.012290 kernel: nicpf, ver 1.0 Apr 12 18:28:30.012297 kernel: nicvf, ver 1.0 Apr 12 18:28:30.012442 kernel: rtc-efi rtc-efi.0: registered as rtc0 Apr 12 18:28:30.012511 kernel: rtc-efi rtc-efi.0: setting system clock to 2024-04-12T18:28:29 UTC (1712946509) Apr 12 18:28:30.012520 kernel: efifb: probing for efifb Apr 12 18:28:30.012528 kernel: efifb: framebuffer at 0x40000000, using 3072k, total 3072k Apr 12 18:28:30.012535 kernel: efifb: mode is 1024x768x32, linelength=4096, pages=1 Apr 12 18:28:30.012542 kernel: efifb: scrolling: redraw Apr 12 18:28:30.012551 kernel: efifb: Truecolor: size=8:8:8:8, shift=24:16:8:0 Apr 12 18:28:30.012559 kernel: Console: switching to colour frame buffer device 128x48 Apr 12 18:28:30.012566 kernel: fb0: EFI VGA frame buffer device Apr 12 18:28:30.012573 kernel: SMCCC: SOC_ID: ARCH_SOC_ID not implemented, skipping .... Apr 12 18:28:30.012580 kernel: hid: raw HID events driver (C) Jiri Kosina Apr 12 18:28:30.012587 kernel: NET: Registered PF_INET6 protocol family Apr 12 18:28:30.012595 kernel: Segment Routing with IPv6 Apr 12 18:28:30.012602 kernel: In-situ OAM (IOAM) with IPv6 Apr 12 18:28:30.012609 kernel: NET: Registered PF_PACKET protocol family Apr 12 18:28:30.012617 kernel: Key type dns_resolver registered Apr 12 18:28:30.012624 kernel: registered taskstats version 1 Apr 12 18:28:30.012631 kernel: Loading compiled-in X.509 certificates Apr 12 18:28:30.012638 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.154-flatcar: 8c258d82bbd8df4a9da2c0ea4108142f04be6b34' Apr 12 18:28:30.012645 kernel: Key type .fscrypt registered Apr 12 18:28:30.012652 kernel: Key type fscrypt-provisioning registered Apr 12 18:28:30.012659 kernel: ima: No TPM chip found, activating TPM-bypass! Apr 12 18:28:30.012666 kernel: ima: Allocated hash algorithm: sha1 Apr 12 18:28:30.012673 kernel: ima: No architecture policies found Apr 12 18:28:30.012681 kernel: Freeing unused kernel memory: 36352K Apr 12 18:28:30.012688 kernel: Run /init as init process Apr 12 18:28:30.012694 kernel: with arguments: Apr 12 18:28:30.012701 kernel: /init Apr 12 18:28:30.012708 kernel: with environment: Apr 12 18:28:30.012715 kernel: HOME=/ Apr 12 18:28:30.012722 kernel: TERM=linux Apr 12 18:28:30.012729 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Apr 12 18:28:30.012738 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Apr 12 18:28:30.012749 systemd[1]: Detected virtualization microsoft. Apr 12 18:28:30.012756 systemd[1]: Detected architecture arm64. Apr 12 18:28:30.012763 systemd[1]: Running in initrd. Apr 12 18:28:30.012771 systemd[1]: No hostname configured, using default hostname. Apr 12 18:28:30.012778 systemd[1]: Hostname set to . Apr 12 18:28:30.012786 systemd[1]: Initializing machine ID from random generator. Apr 12 18:28:30.012793 systemd[1]: Queued start job for default target initrd.target. Apr 12 18:28:30.012802 systemd[1]: Started systemd-ask-password-console.path. Apr 12 18:28:30.012809 systemd[1]: Reached target cryptsetup.target. Apr 12 18:28:30.012817 systemd[1]: Reached target paths.target. Apr 12 18:28:30.012824 systemd[1]: Reached target slices.target. Apr 12 18:28:30.012831 systemd[1]: Reached target swap.target. Apr 12 18:28:30.012838 systemd[1]: Reached target timers.target. Apr 12 18:28:30.012846 systemd[1]: Listening on iscsid.socket. Apr 12 18:28:30.012854 systemd[1]: Listening on iscsiuio.socket. Apr 12 18:28:30.012863 systemd[1]: Listening on systemd-journald-audit.socket. Apr 12 18:28:30.012871 systemd[1]: Listening on systemd-journald-dev-log.socket. Apr 12 18:28:30.012878 systemd[1]: Listening on systemd-journald.socket. Apr 12 18:28:30.012885 systemd[1]: Listening on systemd-networkd.socket. Apr 12 18:28:30.012893 systemd[1]: Listening on systemd-udevd-control.socket. Apr 12 18:28:30.012900 systemd[1]: Listening on systemd-udevd-kernel.socket. Apr 12 18:28:30.012907 systemd[1]: Reached target sockets.target. Apr 12 18:28:30.012915 systemd[1]: Starting kmod-static-nodes.service... Apr 12 18:28:30.012922 systemd[1]: Finished network-cleanup.service. Apr 12 18:28:30.012931 systemd[1]: Starting systemd-fsck-usr.service... Apr 12 18:28:30.012939 systemd[1]: Starting systemd-journald.service... Apr 12 18:28:30.012946 systemd[1]: Starting systemd-modules-load.service... Apr 12 18:28:30.012953 systemd[1]: Starting systemd-resolved.service... Apr 12 18:28:30.012961 systemd[1]: Starting systemd-vconsole-setup.service... Apr 12 18:28:30.012972 systemd-journald[236]: Journal started Apr 12 18:28:30.013014 systemd-journald[236]: Runtime Journal (/run/log/journal/86f61ab23176412288dc6f0396f88746) is 8.0M, max 78.6M, 70.6M free. Apr 12 18:28:29.997819 systemd-modules-load[237]: Inserted module 'overlay' Apr 12 18:28:30.044161 systemd[1]: Started systemd-journald.service. Apr 12 18:28:30.044217 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Apr 12 18:28:30.036099 systemd-resolved[238]: Positive Trust Anchors: Apr 12 18:28:30.082699 kernel: audit: type=1130 audit(1712946510.048:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.082723 kernel: Bridge firewalling registered Apr 12 18:28:30.048000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.036108 systemd-resolved[238]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 12 18:28:30.114083 kernel: SCSI subsystem initialized Apr 12 18:28:30.114114 kernel: audit: type=1130 audit(1712946510.090:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.090000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.036135 systemd-resolved[238]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Apr 12 18:28:30.177190 kernel: audit: type=1130 audit(1712946510.118:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.177220 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Apr 12 18:28:30.177231 kernel: device-mapper: uevent: version 1.0.3 Apr 12 18:28:30.118000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.038274 systemd-resolved[238]: Defaulting to hostname 'linux'. Apr 12 18:28:30.209109 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Apr 12 18:28:30.209133 kernel: audit: type=1130 audit(1712946510.181:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.181000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.067950 systemd[1]: Started systemd-resolved.service. Apr 12 18:28:30.213000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.072883 systemd-modules-load[237]: Inserted module 'br_netfilter' Apr 12 18:28:30.090616 systemd[1]: Finished kmod-static-nodes.service. Apr 12 18:28:30.118525 systemd[1]: Finished systemd-fsck-usr.service. Apr 12 18:28:30.270394 kernel: audit: type=1130 audit(1712946510.213:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.270423 kernel: audit: type=1130 audit(1712946510.242:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.242000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.203103 systemd[1]: Finished systemd-vconsole-setup.service. Apr 12 18:28:30.208459 systemd-modules-load[237]: Inserted module 'dm_multipath' Apr 12 18:28:30.233983 systemd[1]: Finished systemd-modules-load.service. Apr 12 18:28:30.331009 kernel: audit: type=1130 audit(1712946510.307:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.307000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.243062 systemd[1]: Reached target nss-lookup.target. Apr 12 18:28:30.360797 kernel: audit: type=1130 audit(1712946510.335:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.335000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.270035 systemd[1]: Starting dracut-cmdline-ask.service... Apr 12 18:28:30.274956 systemd[1]: Starting systemd-sysctl.service... Apr 12 18:28:30.365000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.279863 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Apr 12 18:28:30.396155 kernel: audit: type=1130 audit(1712946510.365:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.295657 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Apr 12 18:28:30.307996 systemd[1]: Finished systemd-sysctl.service. Apr 12 18:28:30.407563 dracut-cmdline[258]: dracut-dracut-053 Apr 12 18:28:30.335474 systemd[1]: Finished dracut-cmdline-ask.service. Apr 12 18:28:30.416581 dracut-cmdline[258]: Using kernel command line parameters: rd.driver.pre=btrfs BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=tty1 console=ttyAMA0,115200n8 earlycon=pl011,0xeffec000 flatcar.first_boot=detected acpi=force flatcar.oem.id=azure flatcar.autologin verity.usrhash=c0b96868344262519ffdb2dae3782c942008a0fecdbc0bc85d2e170bd2e8b8a8 Apr 12 18:28:30.384684 systemd[1]: Starting dracut-cmdline.service... Apr 12 18:28:30.501396 kernel: Loading iSCSI transport class v2.0-870. Apr 12 18:28:30.516388 kernel: iscsi: registered transport (tcp) Apr 12 18:28:30.537293 kernel: iscsi: registered transport (qla4xxx) Apr 12 18:28:30.537348 kernel: QLogic iSCSI HBA Driver Apr 12 18:28:30.567655 systemd[1]: Finished dracut-cmdline.service. Apr 12 18:28:30.571000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:30.572892 systemd[1]: Starting dracut-pre-udev.service... Apr 12 18:28:30.622394 kernel: raid6: neonx8 gen() 13813 MB/s Apr 12 18:28:30.642384 kernel: raid6: neonx8 xor() 10829 MB/s Apr 12 18:28:30.662383 kernel: raid6: neonx4 gen() 13558 MB/s Apr 12 18:28:30.683384 kernel: raid6: neonx4 xor() 11219 MB/s Apr 12 18:28:30.703383 kernel: raid6: neonx2 gen() 13041 MB/s Apr 12 18:28:30.723383 kernel: raid6: neonx2 xor() 10618 MB/s Apr 12 18:28:30.744383 kernel: raid6: neonx1 gen() 10561 MB/s Apr 12 18:28:30.764382 kernel: raid6: neonx1 xor() 8795 MB/s Apr 12 18:28:30.784382 kernel: raid6: int64x8 gen() 6270 MB/s Apr 12 18:28:30.805383 kernel: raid6: int64x8 xor() 3545 MB/s Apr 12 18:28:30.825382 kernel: raid6: int64x4 gen() 7230 MB/s Apr 12 18:28:30.845383 kernel: raid6: int64x4 xor() 3858 MB/s Apr 12 18:28:30.866384 kernel: raid6: int64x2 gen() 6153 MB/s Apr 12 18:28:30.886383 kernel: raid6: int64x2 xor() 3320 MB/s Apr 12 18:28:30.906382 kernel: raid6: int64x1 gen() 5043 MB/s Apr 12 18:28:30.931769 kernel: raid6: int64x1 xor() 2643 MB/s Apr 12 18:28:30.931781 kernel: raid6: using algorithm neonx8 gen() 13813 MB/s Apr 12 18:28:30.931789 kernel: raid6: .... xor() 10829 MB/s, rmw enabled Apr 12 18:28:30.935811 kernel: raid6: using neon recovery algorithm Apr 12 18:28:30.956837 kernel: xor: measuring software checksum speed Apr 12 18:28:30.956850 kernel: 8regs : 17289 MB/sec Apr 12 18:28:30.961072 kernel: 32regs : 20744 MB/sec Apr 12 18:28:30.969927 kernel: arm64_neon : 27882 MB/sec Apr 12 18:28:30.969937 kernel: xor: using function: arm64_neon (27882 MB/sec) Apr 12 18:28:31.026395 kernel: Btrfs loaded, crc32c=crc32c-generic, zoned=no, fsverity=no Apr 12 18:28:31.035618 systemd[1]: Finished dracut-pre-udev.service. Apr 12 18:28:31.039000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:31.044000 audit: BPF prog-id=7 op=LOAD Apr 12 18:28:31.044000 audit: BPF prog-id=8 op=LOAD Apr 12 18:28:31.045076 systemd[1]: Starting systemd-udevd.service... Apr 12 18:28:31.062939 systemd-udevd[435]: Using default interface naming scheme 'v252'. Apr 12 18:28:31.069138 systemd[1]: Started systemd-udevd.service. Apr 12 18:28:31.074000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:31.079284 systemd[1]: Starting dracut-pre-trigger.service... Apr 12 18:28:31.097092 dracut-pre-trigger[456]: rd.md=0: removing MD RAID activation Apr 12 18:28:31.125308 systemd[1]: Finished dracut-pre-trigger.service. Apr 12 18:28:31.129000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:31.130768 systemd[1]: Starting systemd-udev-trigger.service... Apr 12 18:28:31.164547 systemd[1]: Finished systemd-udev-trigger.service. Apr 12 18:28:31.170000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:31.219396 kernel: hv_vmbus: Vmbus version:5.3 Apr 12 18:28:31.235407 kernel: hv_vmbus: registering driver hyperv_keyboard Apr 12 18:28:31.247444 kernel: input: AT Translated Set 2 keyboard as /devices/LNXSYSTM:00/LNXSYBUS:00/ACPI0004:00/VMBUS:00/d34b2567-b9b6-42b9-8778-0a4ec0b955bf/serio0/input/input0 Apr 12 18:28:31.257703 kernel: hv_vmbus: registering driver hv_storvsc Apr 12 18:28:31.257752 kernel: hv_vmbus: registering driver hid_hyperv Apr 12 18:28:31.269671 kernel: scsi host0: storvsc_host_t Apr 12 18:28:31.269868 kernel: scsi 0:0:0:0: Direct-Access Msft Virtual Disk 1.0 PQ: 0 ANSI: 5 Apr 12 18:28:31.269900 kernel: input: Microsoft Vmbus HID-compliant Mouse as /devices/0006:045E:0621.0001/input/input1 Apr 12 18:28:31.282430 kernel: scsi host1: storvsc_host_t Apr 12 18:28:31.282612 kernel: hv_vmbus: registering driver hv_netvsc Apr 12 18:28:31.294403 kernel: hid-generic 0006:045E:0621.0001: input: VIRTUAL HID v0.01 Mouse [Microsoft Vmbus HID-compliant Mouse] on Apr 12 18:28:31.294590 kernel: scsi 0:0:0:2: CD-ROM Msft Virtual DVD-ROM 1.0 PQ: 0 ANSI: 0 Apr 12 18:28:31.323416 kernel: sr 0:0:0:2: [sr0] scsi-1 drive Apr 12 18:28:31.323628 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Apr 12 18:28:31.330985 kernel: sd 0:0:0:0: [sda] 63737856 512-byte logical blocks: (32.6 GB/30.4 GiB) Apr 12 18:28:31.331122 kernel: sd 0:0:0:0: [sda] 4096-byte physical blocks Apr 12 18:28:31.339392 kernel: sd 0:0:0:0: [sda] Write Protect is off Apr 12 18:28:31.339536 kernel: sd 0:0:0:0: [sda] Mode Sense: 0f 00 10 00 Apr 12 18:28:31.339622 kernel: sd 0:0:0:0: [sda] Write cache: disabled, read cache: enabled, supports DPO and FUA Apr 12 18:28:31.346394 kernel: sr 0:0:0:2: Attached scsi CD-ROM sr0 Apr 12 18:28:31.353400 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 12 18:28:31.358405 kernel: sd 0:0:0:0: [sda] Attached SCSI disk Apr 12 18:28:31.379411 kernel: hv_netvsc 000d3a06-e869-000d-3a06-e869000d3a06 eth0: VF slot 1 added Apr 12 18:28:31.396087 kernel: hv_vmbus: registering driver hv_pci Apr 12 18:28:31.396141 kernel: hv_pci 95228e0b-778f-40cb-954a-4952eaa91641: PCI VMBus probing: Using version 0x10004 Apr 12 18:28:31.416299 kernel: hv_pci 95228e0b-778f-40cb-954a-4952eaa91641: PCI host bridge to bus 778f:00 Apr 12 18:28:31.416492 kernel: pci_bus 778f:00: root bus resource [mem 0xfc0000000-0xfc00fffff window] Apr 12 18:28:31.416597 kernel: pci_bus 778f:00: No busn resource found for root bus, will use [bus 00-ff] Apr 12 18:28:31.429913 kernel: pci 778f:00:02.0: [15b3:1018] type 00 class 0x020000 Apr 12 18:28:31.441752 kernel: pci 778f:00:02.0: reg 0x10: [mem 0xfc0000000-0xfc00fffff 64bit pref] Apr 12 18:28:31.463580 kernel: pci 778f:00:02.0: enabling Extended Tags Apr 12 18:28:31.482391 kernel: pci 778f:00:02.0: 0.000 Gb/s available PCIe bandwidth, limited by Unknown x0 link at 778f:00:02.0 (capable of 126.016 Gb/s with 8.0 GT/s PCIe x16 link) Apr 12 18:28:31.482581 kernel: pci_bus 778f:00: busn_res: [bus 00-ff] end is updated to 00 Apr 12 18:28:31.493661 kernel: pci 778f:00:02.0: BAR 0: assigned [mem 0xfc0000000-0xfc00fffff 64bit pref] Apr 12 18:28:31.535397 kernel: mlx5_core 778f:00:02.0: firmware version: 16.30.1284 Apr 12 18:28:31.698514 kernel: mlx5_core 778f:00:02.0: MLX5E: StrdRq(1) RqSz(8) StrdSz(2048) RxCqeCmprss(0) Apr 12 18:28:31.758744 kernel: hv_netvsc 000d3a06-e869-000d-3a06-e869000d3a06 eth0: VF registering: eth1 Apr 12 18:28:31.758949 kernel: mlx5_core 778f:00:02.0 eth1: joined to eth0 Apr 12 18:28:31.774412 kernel: mlx5_core 778f:00:02.0 enP30607s1: renamed from eth1 Apr 12 18:28:31.785305 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Apr 12 18:28:31.819403 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/sda6 scanned by (udev-worker) (492) Apr 12 18:28:31.832238 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Apr 12 18:28:32.010820 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Apr 12 18:28:32.076386 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Apr 12 18:28:32.082484 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Apr 12 18:28:32.096394 systemd[1]: Starting disk-uuid.service... Apr 12 18:28:32.119402 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 12 18:28:32.139411 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 12 18:28:33.138399 kernel: sda: sda1 sda2 sda3 sda4 sda6 sda7 sda9 Apr 12 18:28:33.138750 disk-uuid[558]: The operation has completed successfully. Apr 12 18:28:33.191006 systemd[1]: disk-uuid.service: Deactivated successfully. Apr 12 18:28:33.194000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.194000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.191099 systemd[1]: Finished disk-uuid.service. Apr 12 18:28:33.208320 systemd[1]: Starting verity-setup.service... Apr 12 18:28:33.249422 kernel: device-mapper: verity: sha256 using implementation "sha256-ce" Apr 12 18:28:33.543640 systemd[1]: Found device dev-mapper-usr.device. Apr 12 18:28:33.549269 systemd[1]: Mounting sysusr-usr.mount... Apr 12 18:28:33.560804 systemd[1]: Finished verity-setup.service. Apr 12 18:28:33.564000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.616405 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Apr 12 18:28:33.616765 systemd[1]: Mounted sysusr-usr.mount. Apr 12 18:28:33.620753 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Apr 12 18:28:33.621590 systemd[1]: Starting ignition-setup.service... Apr 12 18:28:33.630345 systemd[1]: Starting parse-ip-for-networkd.service... Apr 12 18:28:33.668802 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 12 18:28:33.668862 kernel: BTRFS info (device sda6): using free space tree Apr 12 18:28:33.673244 kernel: BTRFS info (device sda6): has skinny extents Apr 12 18:28:33.735156 systemd[1]: Finished parse-ip-for-networkd.service. Apr 12 18:28:33.740000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.744000 audit: BPF prog-id=9 op=LOAD Apr 12 18:28:33.745188 systemd[1]: Starting systemd-networkd.service... Apr 12 18:28:33.760578 systemd[1]: mnt-oem.mount: Deactivated successfully. Apr 12 18:28:33.768582 systemd-networkd[826]: lo: Link UP Apr 12 18:28:33.768593 systemd-networkd[826]: lo: Gained carrier Apr 12 18:28:33.777000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.769319 systemd-networkd[826]: Enumeration completed Apr 12 18:28:33.772808 systemd[1]: Started systemd-networkd.service. Apr 12 18:28:33.773334 systemd-networkd[826]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 12 18:28:33.801000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.777926 systemd[1]: Reached target network.target. Apr 12 18:28:33.786950 systemd[1]: Starting iscsiuio.service... Apr 12 18:28:33.819587 iscsid[834]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Apr 12 18:28:33.819587 iscsid[834]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log Apr 12 18:28:33.819587 iscsid[834]: into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Apr 12 18:28:33.819587 iscsid[834]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Apr 12 18:28:33.819587 iscsid[834]: If using hardware iscsi like qla4xxx this message can be ignored. Apr 12 18:28:33.819587 iscsid[834]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Apr 12 18:28:33.819587 iscsid[834]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Apr 12 18:28:33.922062 kernel: mlx5_core 778f:00:02.0 enP30607s1: Link up Apr 12 18:28:33.822000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.869000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.921000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.797494 systemd[1]: Started iscsiuio.service. Apr 12 18:28:33.808145 systemd[1]: Starting iscsid.service... Apr 12 18:28:33.816108 systemd[1]: Started iscsid.service. Apr 12 18:28:33.824540 systemd[1]: Starting dracut-initqueue.service... Apr 12 18:28:33.951436 kernel: hv_netvsc 000d3a06-e869-000d-3a06-e869000d3a06 eth0: Data path switched to VF: enP30607s1 Apr 12 18:28:33.951598 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Apr 12 18:28:33.852520 systemd[1]: Finished dracut-initqueue.service. Apr 12 18:28:33.870076 systemd[1]: Reached target remote-fs-pre.target. Apr 12 18:28:33.879329 systemd[1]: Reached target remote-cryptsetup.target. Apr 12 18:28:33.893646 systemd[1]: Reached target remote-fs.target. Apr 12 18:28:33.905995 systemd[1]: Starting dracut-pre-mount.service... Apr 12 18:28:33.917153 systemd[1]: Finished dracut-pre-mount.service. Apr 12 18:28:33.940766 systemd-networkd[826]: enP30607s1: Link UP Apr 12 18:28:33.940844 systemd-networkd[826]: eth0: Link UP Apr 12 18:28:33.960582 systemd-networkd[826]: eth0: Gained carrier Apr 12 18:28:33.974126 systemd-networkd[826]: enP30607s1: Gained carrier Apr 12 18:28:34.020238 kernel: kauditd_printk_skb: 17 callbacks suppressed Apr 12 18:28:34.020263 kernel: audit: type=1130 audit(1712946513.994:28): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.994000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:33.988090 systemd-networkd[826]: eth0: DHCPv4 address 10.200.20.18/24, gateway 10.200.20.1 acquired from 168.63.129.16 Apr 12 18:28:33.988820 systemd[1]: Finished ignition-setup.service. Apr 12 18:28:33.999247 systemd[1]: Starting ignition-fetch-offline.service... Apr 12 18:28:35.137553 systemd-networkd[826]: eth0: Gained IPv6LL Apr 12 18:28:37.060423 ignition[849]: Ignition 2.14.0 Apr 12 18:28:37.064043 ignition[849]: Stage: fetch-offline Apr 12 18:28:37.064141 ignition[849]: reading system config file "/usr/lib/ignition/base.d/base.ign" Apr 12 18:28:37.064171 ignition[849]: parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Apr 12 18:28:37.194308 ignition[849]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 12 18:28:37.194505 ignition[849]: parsed url from cmdline: "" Apr 12 18:28:37.194510 ignition[849]: no config URL provided Apr 12 18:28:37.194515 ignition[849]: reading system config file "/usr/lib/ignition/user.ign" Apr 12 18:28:37.245128 kernel: audit: type=1130 audit(1712946517.218:29): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.218000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.210032 systemd[1]: Finished ignition-fetch-offline.service. Apr 12 18:28:37.194523 ignition[849]: no config at "/usr/lib/ignition/user.ign" Apr 12 18:28:37.220252 systemd[1]: Starting ignition-fetch.service... Apr 12 18:28:37.194528 ignition[849]: failed to fetch config: resource requires networking Apr 12 18:28:37.194651 ignition[849]: Ignition finished successfully Apr 12 18:28:37.249458 ignition[855]: Ignition 2.14.0 Apr 12 18:28:37.249466 ignition[855]: Stage: fetch Apr 12 18:28:37.249580 ignition[855]: reading system config file "/usr/lib/ignition/base.d/base.ign" Apr 12 18:28:37.249598 ignition[855]: parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Apr 12 18:28:37.260953 ignition[855]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 12 18:28:37.261096 ignition[855]: parsed url from cmdline: "" Apr 12 18:28:37.261100 ignition[855]: no config URL provided Apr 12 18:28:37.261106 ignition[855]: reading system config file "/usr/lib/ignition/user.ign" Apr 12 18:28:37.261114 ignition[855]: no config at "/usr/lib/ignition/user.ign" Apr 12 18:28:37.261145 ignition[855]: GET http://169.254.169.254/metadata/instance/compute/userData?api-version=2021-01-01&format=text: attempt #1 Apr 12 18:28:37.284900 ignition[855]: GET result: OK Apr 12 18:28:37.284988 ignition[855]: config has been read from IMDS userdata Apr 12 18:28:37.285054 ignition[855]: parsing config with SHA512: c13b1edccfbc8ae34d3d2372c3b37b5962fe0404ea3b39984b9cf6ebbf65bfab4af142660085194405818387c60c5d2854c67a6413bc3c955dc6ce8654e8719f Apr 12 18:28:37.354444 unknown[855]: fetched base config from "system" Apr 12 18:28:37.354457 unknown[855]: fetched base config from "system" Apr 12 18:28:37.364000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.355187 ignition[855]: fetch: fetch complete Apr 12 18:28:37.393498 kernel: audit: type=1130 audit(1712946517.364:30): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.354463 unknown[855]: fetched user config from "azure" Apr 12 18:28:37.355192 ignition[855]: fetch: fetch passed Apr 12 18:28:37.360117 systemd[1]: Finished ignition-fetch.service. Apr 12 18:28:37.355238 ignition[855]: Ignition finished successfully Apr 12 18:28:37.409000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.365693 systemd[1]: Starting ignition-kargs.service... Apr 12 18:28:37.393979 ignition[861]: Ignition 2.14.0 Apr 12 18:28:37.440309 kernel: audit: type=1130 audit(1712946517.409:31): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.405237 systemd[1]: Finished ignition-kargs.service. Apr 12 18:28:37.480045 kernel: audit: type=1130 audit(1712946517.451:32): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.451000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.393985 ignition[861]: Stage: kargs Apr 12 18:28:37.428681 systemd[1]: Starting ignition-disks.service... Apr 12 18:28:37.394104 ignition[861]: reading system config file "/usr/lib/ignition/base.d/base.ign" Apr 12 18:28:37.447358 systemd[1]: Finished ignition-disks.service. Apr 12 18:28:37.394127 ignition[861]: parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Apr 12 18:28:37.452307 systemd[1]: Reached target initrd-root-device.target. Apr 12 18:28:37.399331 ignition[861]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 12 18:28:37.480482 systemd[1]: Reached target local-fs-pre.target. Apr 12 18:28:37.402902 ignition[861]: kargs: kargs passed Apr 12 18:28:37.487644 systemd[1]: Reached target local-fs.target. Apr 12 18:28:37.403063 ignition[861]: Ignition finished successfully Apr 12 18:28:37.496030 systemd[1]: Reached target sysinit.target. Apr 12 18:28:37.439792 ignition[867]: Ignition 2.14.0 Apr 12 18:28:37.507032 systemd[1]: Reached target basic.target. Apr 12 18:28:37.439799 ignition[867]: Stage: disks Apr 12 18:28:37.521229 systemd[1]: Starting systemd-fsck-root.service... Apr 12 18:28:37.439946 ignition[867]: reading system config file "/usr/lib/ignition/base.d/base.ign" Apr 12 18:28:37.439982 ignition[867]: parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Apr 12 18:28:37.443258 ignition[867]: no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 12 18:28:37.446313 ignition[867]: disks: disks passed Apr 12 18:28:37.446445 ignition[867]: Ignition finished successfully Apr 12 18:28:37.619593 systemd-fsck[875]: ROOT: clean, 612/7326000 files, 481074/7359488 blocks Apr 12 18:28:37.633925 systemd[1]: Finished systemd-fsck-root.service. Apr 12 18:28:37.638000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.663441 systemd[1]: Mounting sysroot.mount... Apr 12 18:28:37.670812 kernel: audit: type=1130 audit(1712946517.638:33): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:37.685396 kernel: EXT4-fs (sda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Apr 12 18:28:37.685667 systemd[1]: Mounted sysroot.mount. Apr 12 18:28:37.689466 systemd[1]: Reached target initrd-root-fs.target. Apr 12 18:28:37.729770 systemd[1]: Mounting sysroot-usr.mount... Apr 12 18:28:37.734304 systemd[1]: Starting flatcar-metadata-hostname.service... Apr 12 18:28:37.741564 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Apr 12 18:28:37.741600 systemd[1]: Reached target ignition-diskful.target. Apr 12 18:28:37.747289 systemd[1]: Mounted sysroot-usr.mount. Apr 12 18:28:37.825426 systemd[1]: Mounting sysroot-usr-share-oem.mount... Apr 12 18:28:37.830417 systemd[1]: Starting initrd-setup-root.service... Apr 12 18:28:37.851396 kernel: BTRFS: device label OEM devid 1 transid 14 /dev/sda6 scanned by mount (886) Apr 12 18:28:37.863026 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 12 18:28:37.863062 kernel: BTRFS info (device sda6): using free space tree Apr 12 18:28:37.863234 initrd-setup-root[891]: cut: /sysroot/etc/passwd: No such file or directory Apr 12 18:28:37.874658 kernel: BTRFS info (device sda6): has skinny extents Apr 12 18:28:37.879818 systemd[1]: Mounted sysroot-usr-share-oem.mount. Apr 12 18:28:37.924157 initrd-setup-root[917]: cut: /sysroot/etc/group: No such file or directory Apr 12 18:28:37.950462 initrd-setup-root[925]: cut: /sysroot/etc/shadow: No such file or directory Apr 12 18:28:37.960002 initrd-setup-root[933]: cut: /sysroot/etc/gshadow: No such file or directory Apr 12 18:28:38.689161 systemd[1]: Finished initrd-setup-root.service. Apr 12 18:28:38.693000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:38.694590 systemd[1]: Starting ignition-mount.service... Apr 12 18:28:38.720497 kernel: audit: type=1130 audit(1712946518.693:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:38.720761 systemd[1]: Starting sysroot-boot.service... Apr 12 18:28:38.726592 systemd[1]: sysusr-usr-share-oem.mount: Deactivated successfully. Apr 12 18:28:38.726742 systemd[1]: sysroot-usr-share-oem.mount: Deactivated successfully. Apr 12 18:28:38.755781 ignition[953]: INFO : Ignition 2.14.0 Apr 12 18:28:38.755781 ignition[953]: INFO : Stage: mount Apr 12 18:28:38.764907 ignition[953]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Apr 12 18:28:38.764907 ignition[953]: DEBUG : parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Apr 12 18:28:38.764907 ignition[953]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 12 18:28:38.764907 ignition[953]: INFO : mount: mount passed Apr 12 18:28:38.764907 ignition[953]: INFO : Ignition finished successfully Apr 12 18:28:38.839986 kernel: audit: type=1130 audit(1712946518.776:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:38.840013 kernel: audit: type=1130 audit(1712946518.809:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:38.776000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:38.809000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:38.765859 systemd[1]: Finished sysroot-boot.service. Apr 12 18:28:38.776812 systemd[1]: Finished ignition-mount.service. Apr 12 18:28:39.418287 coreos-metadata[885]: Apr 12 18:28:39.418 INFO Fetching http://168.63.129.16/?comp=versions: Attempt #1 Apr 12 18:28:39.427904 coreos-metadata[885]: Apr 12 18:28:39.427 INFO Fetch successful Apr 12 18:28:39.462091 coreos-metadata[885]: Apr 12 18:28:39.462 INFO Fetching http://169.254.169.254/metadata/instance/compute/name?api-version=2017-08-01&format=text: Attempt #1 Apr 12 18:28:39.486844 coreos-metadata[885]: Apr 12 18:28:39.486 INFO Fetch successful Apr 12 18:28:39.506293 coreos-metadata[885]: Apr 12 18:28:39.506 INFO wrote hostname ci-3510.3.3-a-f11561af44 to /sysroot/etc/hostname Apr 12 18:28:39.515775 systemd[1]: Finished flatcar-metadata-hostname.service. Apr 12 18:28:39.520000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:39.548975 kernel: audit: type=1130 audit(1712946519.520:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:39.544104 systemd[1]: Starting ignition-files.service... Apr 12 18:28:39.554181 systemd[1]: Mounting sysroot-usr-share-oem.mount... Apr 12 18:28:39.579750 kernel: BTRFS: device label OEM devid 1 transid 15 /dev/sda6 scanned by mount (964) Apr 12 18:28:39.579798 kernel: BTRFS info (device sda6): using crc32c (crc32c-generic) checksum algorithm Apr 12 18:28:39.579808 kernel: BTRFS info (device sda6): using free space tree Apr 12 18:28:39.588838 kernel: BTRFS info (device sda6): has skinny extents Apr 12 18:28:39.593436 systemd[1]: Mounted sysroot-usr-share-oem.mount. Apr 12 18:28:39.610998 ignition[983]: INFO : Ignition 2.14.0 Apr 12 18:28:39.610998 ignition[983]: INFO : Stage: files Apr 12 18:28:39.621768 ignition[983]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Apr 12 18:28:39.621768 ignition[983]: DEBUG : parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Apr 12 18:28:39.621768 ignition[983]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 12 18:28:39.621768 ignition[983]: DEBUG : files: compiled without relabeling support, skipping Apr 12 18:28:39.654515 ignition[983]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Apr 12 18:28:39.654515 ignition[983]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Apr 12 18:28:39.722430 ignition[983]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Apr 12 18:28:39.731457 ignition[983]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Apr 12 18:28:39.731457 ignition[983]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Apr 12 18:28:39.731457 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 12 18:28:39.731457 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET https://get.helm.sh/helm-v3.13.2-linux-arm64.tar.gz: attempt #1 Apr 12 18:28:39.730852 unknown[983]: wrote ssh authorized keys file for user: core Apr 12 18:28:40.117542 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(3): GET result: OK Apr 12 18:28:40.337014 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/opt/helm-v3.13.2-linux-arm64.tar.gz" Apr 12 18:28:40.347676 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/opt/crictl-v1.27.0-linux-arm64.tar.gz" Apr 12 18:28:40.347676 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET https://github.com/kubernetes-sigs/cri-tools/releases/download/v1.27.0/crictl-v1.27.0-linux-arm64.tar.gz: attempt #1 Apr 12 18:28:40.653275 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(4): GET result: OK Apr 12 18:28:40.808349 ignition[983]: DEBUG : files: createFilesystemsFiles: createFiles: op(4): file matches expected sum of: db062e43351a63347871e7094115be2ae3853afcd346d47f7b51141da8c3202c2df58d2e17359322f632abcb37474fd7fdb3b7aadbc5cfd5cf6d3bad040b6251 Apr 12 18:28:40.825277 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/opt/crictl-v1.27.0-linux-arm64.tar.gz" Apr 12 18:28:40.825277 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing file "/sysroot/etc/flatcar-cgroupv1" Apr 12 18:28:40.825277 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing file "/sysroot/etc/flatcar-cgroupv1" Apr 12 18:28:40.825277 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/opt/cni-plugins-linux-arm64-v1.3.0.tgz" Apr 12 18:28:40.825277 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET https://github.com/containernetworking/plugins/releases/download/v1.3.0/cni-plugins-linux-arm64-v1.3.0.tgz: attempt #1 Apr 12 18:28:41.071236 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET result: OK Apr 12 18:28:41.315985 ignition[983]: DEBUG : files: createFilesystemsFiles: createFiles: op(6): file matches expected sum of: b2b7fb74f1b3cb8928f49e5bf9d4bc686e057e837fac3caf1b366d54757921dba80d70cc010399b274d136e8dee9a25b1ad87cdfdc4ffcf42cf88f3e8f99587a Apr 12 18:28:41.332556 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/opt/cni-plugins-linux-arm64-v1.3.0.tgz" Apr 12 18:28:41.332556 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(7): [started] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 12 18:28:41.332556 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET https://github.com/cilium/cilium-cli/releases/download/v0.12.12/cilium-linux-arm64.tar.gz: attempt #1 Apr 12 18:28:41.720451 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(7): GET result: OK Apr 12 18:28:41.774261 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(7): [finished] writing file "/sysroot/opt/bin/cilium.tar.gz" Apr 12 18:28:41.784005 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(8): [started] writing file "/sysroot/opt/bin/kubeadm" Apr 12 18:28:41.784005 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET https://dl.k8s.io/release/v1.27.2/bin/linux/arm64/kubeadm: attempt #1 Apr 12 18:28:42.126932 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(8): GET result: OK Apr 12 18:28:42.449401 ignition[983]: DEBUG : files: createFilesystemsFiles: createFiles: op(8): file matches expected sum of: 45b3100984c979ba0f1c0df8f4211474c2d75ebe916e677dff5fc8e3b3697cf7a953da94e356f39684cc860dff6878b772b7514c55651c2f866d9efeef23f970 Apr 12 18:28:42.467047 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(8): [finished] writing file "/sysroot/opt/bin/kubeadm" Apr 12 18:28:42.467047 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(9): [started] writing file "/sysroot/opt/bin/kubelet" Apr 12 18:28:42.467047 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(9): GET https://dl.k8s.io/release/v1.27.2/bin/linux/arm64/kubelet: attempt #1 Apr 12 18:28:42.502157 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(9): GET result: OK Apr 12 18:28:43.106305 ignition[983]: DEBUG : files: createFilesystemsFiles: createFiles: op(9): file matches expected sum of: 71857ff499ae135fa478e1827a0ed8865e578a8d2b1e25876e914fd0beba03733801c0654bcd4c0567bafeb16887dafb2dbbe8d1116e6ea28dcd8366c142d348 Apr 12 18:28:43.123053 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(9): [finished] writing file "/sysroot/opt/bin/kubelet" Apr 12 18:28:43.123053 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(a): [started] writing file "/sysroot/etc/docker/daemon.json" Apr 12 18:28:43.123053 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(a): [finished] writing file "/sysroot/etc/docker/daemon.json" Apr 12 18:28:43.123053 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(b): [started] writing file "/sysroot/opt/bin/kubectl" Apr 12 18:28:43.123053 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET https://dl.k8s.io/release/v1.27.2/bin/linux/arm64/kubectl: attempt #1 Apr 12 18:28:43.180219 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(b): GET result: OK Apr 12 18:28:43.531711 ignition[983]: DEBUG : files: createFilesystemsFiles: createFiles: op(b): file matches expected sum of: 14be61ec35669a27acf2df0380afb85b9b42311d50ca1165718421c5f605df1119ec9ae314696a674051712e80deeaa65e62d2d62ed4d107fe99d0aaf419dafc Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(b): [finished] writing file "/sysroot/opt/bin/kubectl" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(c): [started] writing file "/sysroot/home/core/install.sh" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(c): [finished] writing file "/sysroot/home/core/install.sh" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(d): [started] writing file "/sysroot/home/core/nginx.yaml" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(d): [finished] writing file "/sysroot/home/core/nginx.yaml" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(e): [started] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(e): [finished] writing file "/sysroot/home/core/nfs-pod.yaml" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(f): [started] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(f): [finished] writing file "/sysroot/home/core/nfs-pvc.yaml" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(10): [started] writing file "/sysroot/etc/flatcar/update.conf" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(10): [finished] writing file "/sysroot/etc/flatcar/update.conf" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(11): [started] writing file "/sysroot/etc/systemd/system/nvidia.service" Apr 12 18:28:43.549901 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(11): oem config not found in "/usr/share/oem", looking on oem partition Apr 12 18:28:43.723764 kernel: BTRFS info: devid 1 device path /dev/sda6 changed to /dev/disk/by-label/OEM scanned by ignition (986) Apr 12 18:28:43.723792 kernel: audit: type=1130 audit(1712946523.645:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.645000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(11): op(12): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem635837255" Apr 12 18:28:43.723851 ignition[983]: CRITICAL : files: createFilesystemsFiles: createFiles: op(11): op(12): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem635837255": device or resource busy Apr 12 18:28:43.723851 ignition[983]: ERROR : files: createFilesystemsFiles: createFiles: op(11): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem635837255", trying btrfs: device or resource busy Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(11): op(13): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem635837255" Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(11): op(13): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem635837255" Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(11): op(14): [started] unmounting "/mnt/oem635837255" Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(11): op(14): [finished] unmounting "/mnt/oem635837255" Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(11): [finished] writing file "/sysroot/etc/systemd/system/nvidia.service" Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(15): [started] writing file "/sysroot/etc/systemd/system/waagent.service" Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(15): oem config not found in "/usr/share/oem", looking on oem partition Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(15): op(16): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3148812595" Apr 12 18:28:43.723851 ignition[983]: CRITICAL : files: createFilesystemsFiles: createFiles: op(15): op(16): [failed] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3148812595": device or resource busy Apr 12 18:28:43.723851 ignition[983]: ERROR : files: createFilesystemsFiles: createFiles: op(15): failed to mount ext4 device "/dev/disk/by-label/OEM" at "/mnt/oem3148812595", trying btrfs: device or resource busy Apr 12 18:28:43.723851 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(15): op(17): [started] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3148812595" Apr 12 18:28:44.011049 kernel: audit: type=1130 audit(1712946523.728:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.011082 kernel: audit: type=1131 audit(1712946523.728:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.011093 kernel: audit: type=1130 audit(1712946523.776:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.011104 kernel: audit: type=1130 audit(1712946523.855:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.011121 kernel: audit: type=1131 audit(1712946523.881:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.728000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.728000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.776000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.855000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.881000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.010000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.581439 systemd[1]: mnt-oem635837255.mount: Deactivated successfully. Apr 12 18:28:44.033603 kernel: audit: type=1130 audit(1712946524.010:44): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.033628 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(15): op(17): [finished] mounting "/dev/disk/by-label/OEM" at "/mnt/oem3148812595" Apr 12 18:28:44.033628 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(15): op(18): [started] unmounting "/mnt/oem3148812595" Apr 12 18:28:44.033628 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(15): op(18): [finished] unmounting "/mnt/oem3148812595" Apr 12 18:28:44.033628 ignition[983]: INFO : files: createFilesystemsFiles: createFiles: op(15): [finished] writing file "/sysroot/etc/systemd/system/waagent.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(19): [started] processing unit "waagent.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(19): [finished] processing unit "waagent.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1a): [started] processing unit "nvidia.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1a): [finished] processing unit "nvidia.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1b): [started] processing unit "containerd.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1b): op(1c): [started] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1b): op(1c): [finished] writing systemd drop-in "10-use-cgroupfs.conf" at "/sysroot/etc/systemd/system/containerd.service.d/10-use-cgroupfs.conf" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1b): [finished] processing unit "containerd.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1d): [started] processing unit "prepare-cni-plugins.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1d): op(1e): [started] writing unit "prepare-cni-plugins.service" at "/sysroot/etc/systemd/system/prepare-cni-plugins.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1d): op(1e): [finished] writing unit "prepare-cni-plugins.service" at "/sysroot/etc/systemd/system/prepare-cni-plugins.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1d): [finished] processing unit "prepare-cni-plugins.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1f): [started] processing unit "prepare-critools.service" Apr 12 18:28:44.033628 ignition[983]: INFO : files: op(1f): op(20): [started] writing unit "prepare-critools.service" at "/sysroot/etc/systemd/system/prepare-critools.service" Apr 12 18:28:44.290621 kernel: audit: type=1130 audit(1712946524.061:45): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.290655 kernel: audit: type=1131 audit(1712946524.086:46): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.061000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.086000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.149000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.620044 systemd[1]: mnt-oem3148812595.mount: Deactivated successfully. Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(1f): op(20): [finished] writing unit "prepare-critools.service" at "/sysroot/etc/systemd/system/prepare-critools.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(1f): [finished] processing unit "prepare-critools.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(21): [started] processing unit "prepare-helm.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(21): op(22): [started] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(21): op(22): [finished] writing unit "prepare-helm.service" at "/sysroot/etc/systemd/system/prepare-helm.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(21): [finished] processing unit "prepare-helm.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(23): [started] setting preset to enabled for "waagent.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(23): [finished] setting preset to enabled for "waagent.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(24): [started] setting preset to enabled for "nvidia.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(24): [finished] setting preset to enabled for "nvidia.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(25): [started] setting preset to enabled for "prepare-cni-plugins.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(25): [finished] setting preset to enabled for "prepare-cni-plugins.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(26): [started] setting preset to enabled for "prepare-critools.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(26): [finished] setting preset to enabled for "prepare-critools.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(27): [started] setting preset to enabled for "prepare-helm.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: op(27): [finished] setting preset to enabled for "prepare-helm.service" Apr 12 18:28:44.299191 ignition[983]: INFO : files: createResultFile: createFiles: op(28): [started] writing file "/sysroot/etc/.ignition-result.json" Apr 12 18:28:44.299191 ignition[983]: INFO : files: createResultFile: createFiles: op(28): [finished] writing file "/sysroot/etc/.ignition-result.json" Apr 12 18:28:44.299191 ignition[983]: INFO : files: files passed Apr 12 18:28:44.299191 ignition[983]: INFO : Ignition finished successfully Apr 12 18:28:44.303000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.328000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.341000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.358000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.374000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=flatcar-metadata-hostname comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.413000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.424000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.436000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.447000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.459000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.471000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.483000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.495000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.545982 initrd-setup-root-after-ignition[1008]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Apr 12 18:28:43.631593 systemd[1]: Finished ignition-files.service. Apr 12 18:28:44.564628 iscsid[834]: iscsid shutting down. Apr 12 18:28:44.568000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.672073 systemd[1]: Starting initrd-setup-root-after-ignition.service... Apr 12 18:28:44.596591 kernel: kauditd_printk_skb: 14 callbacks suppressed Apr 12 18:28:44.596615 kernel: audit: type=1131 audit(1712946524.568:61): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.677455 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Apr 12 18:28:43.678385 systemd[1]: Starting ignition-quench.service... Apr 12 18:28:44.612000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.702073 systemd[1]: ignition-quench.service: Deactivated successfully. Apr 12 18:28:43.702183 systemd[1]: Finished ignition-quench.service. Apr 12 18:28:44.657595 kernel: audit: type=1131 audit(1712946524.612:62): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.657619 kernel: audit: type=1131 audit(1712946524.639:63): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.639000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.762208 systemd[1]: Finished initrd-setup-root-after-ignition.service. Apr 12 18:28:44.665538 ignition[1021]: INFO : Ignition 2.14.0 Apr 12 18:28:44.665538 ignition[1021]: INFO : Stage: umount Apr 12 18:28:44.665538 ignition[1021]: INFO : reading system config file "/usr/lib/ignition/base.d/base.ign" Apr 12 18:28:44.665538 ignition[1021]: DEBUG : parsing config with SHA512: 4824fd4a4e57848da530dc2b56e2d3e9f5f19634d1c84ef29f8fc49255520728d0377a861a375d7c8cb5301ed861ff4ede4b440b074b1d6a86e23be9cefc2f63 Apr 12 18:28:44.665538 ignition[1021]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/azure" Apr 12 18:28:44.665538 ignition[1021]: INFO : umount: umount passed Apr 12 18:28:44.665538 ignition[1021]: INFO : Ignition finished successfully Apr 12 18:28:44.762537 kernel: audit: type=1131 audit(1712946524.677:64): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.762565 kernel: audit: type=1131 audit(1712946524.736:65): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.677000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.736000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.776625 systemd[1]: Reached target ignition-complete.target. Apr 12 18:28:44.786711 kernel: audit: type=1131 audit(1712946524.766:66): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.766000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.811982 systemd[1]: Starting initrd-parse-etc.service... Apr 12 18:28:44.786000 audit: BPF prog-id=6 op=UNLOAD Apr 12 18:28:43.841138 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Apr 12 18:28:44.809294 kernel: audit: type=1334 audit(1712946524.786:67): prog-id=6 op=UNLOAD Apr 12 18:28:43.841271 systemd[1]: Finished initrd-parse-etc.service. Apr 12 18:28:44.831957 kernel: audit: type=1131 audit(1712946524.813:68): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.813000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.881335 systemd[1]: Reached target initrd-fs.target. Apr 12 18:28:43.894615 systemd[1]: Reached target initrd.target. Apr 12 18:28:44.860318 kernel: audit: type=1131 audit(1712946524.842:69): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.842000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:43.926954 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Apr 12 18:28:43.937197 systemd[1]: Starting dracut-pre-pivot.service... Apr 12 18:28:44.872000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.006741 systemd[1]: Finished dracut-pre-pivot.service. Apr 12 18:28:44.032987 systemd[1]: Starting initrd-cleanup.service... Apr 12 18:28:44.057002 systemd[1]: initrd-cleanup.service: Deactivated successfully. Apr 12 18:28:44.915264 kernel: audit: type=1131 audit(1712946524.872:70): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.910000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.057099 systemd[1]: Finished initrd-cleanup.service. Apr 12 18:28:44.933000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.086953 systemd[1]: Stopped target nss-lookup.target. Apr 12 18:28:44.939000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.101203 systemd[1]: Stopped target remote-cryptsetup.target. Apr 12 18:28:44.946000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.127833 systemd[1]: Stopped target timers.target. Apr 12 18:28:44.960000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.965000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.138337 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Apr 12 18:28:44.982000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.138440 systemd[1]: Stopped dracut-pre-pivot.service. Apr 12 18:28:44.996795 kernel: hv_netvsc 000d3a06-e869-000d-3a06-e869000d3a06 eth0: Data path switched from VF: enP30607s1 Apr 12 18:28:44.992000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.992000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.149524 systemd[1]: Stopped target initrd.target. Apr 12 18:28:44.159675 systemd[1]: Stopped target basic.target. Apr 12 18:28:44.170153 systemd[1]: Stopped target ignition-complete.target. Apr 12 18:28:44.187491 systemd[1]: Stopped target ignition-diskful.target. Apr 12 18:28:44.204628 systemd[1]: Stopped target initrd-root-device.target. Apr 12 18:28:44.215797 systemd[1]: Stopped target remote-fs.target. Apr 12 18:28:44.226863 systemd[1]: Stopped target remote-fs-pre.target. Apr 12 18:28:44.242099 systemd[1]: Stopped target sysinit.target. Apr 12 18:28:44.257454 systemd[1]: Stopped target local-fs.target. Apr 12 18:28:44.268366 systemd[1]: Stopped target local-fs-pre.target. Apr 12 18:28:44.279834 systemd[1]: Stopped target swap.target. Apr 12 18:28:44.294324 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Apr 12 18:28:44.294397 systemd[1]: Stopped dracut-pre-mount.service. Apr 12 18:28:44.303315 systemd[1]: Stopped target cryptsetup.target. Apr 12 18:28:44.318107 systemd[1]: dracut-initqueue.service: Deactivated successfully. Apr 12 18:28:44.318166 systemd[1]: Stopped dracut-initqueue.service. Apr 12 18:28:44.329169 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Apr 12 18:28:44.329210 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Apr 12 18:28:44.342102 systemd[1]: ignition-files.service: Deactivated successfully. Apr 12 18:28:44.342141 systemd[1]: Stopped ignition-files.service. Apr 12 18:28:44.358992 systemd[1]: flatcar-metadata-hostname.service: Deactivated successfully. Apr 12 18:28:44.359035 systemd[1]: Stopped flatcar-metadata-hostname.service. Apr 12 18:28:44.375500 systemd[1]: Stopping ignition-mount.service... Apr 12 18:28:44.387220 systemd[1]: Stopping iscsid.service... Apr 12 18:28:44.391324 systemd[1]: Stopping sysroot-boot.service... Apr 12 18:28:45.104000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:28:44.394953 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Apr 12 18:28:44.395055 systemd[1]: Stopped systemd-udev-trigger.service. Apr 12 18:28:44.414353 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Apr 12 18:28:44.414449 systemd[1]: Stopped dracut-pre-trigger.service. Apr 12 18:28:44.425391 systemd[1]: iscsid.service: Deactivated successfully. Apr 12 18:28:44.425505 systemd[1]: Stopped iscsid.service. Apr 12 18:28:44.436732 systemd[1]: ignition-mount.service: Deactivated successfully. Apr 12 18:28:44.436817 systemd[1]: Stopped ignition-mount.service. Apr 12 18:28:44.447747 systemd[1]: ignition-disks.service: Deactivated successfully. Apr 12 18:28:44.447807 systemd[1]: Stopped ignition-disks.service. Apr 12 18:28:44.459646 systemd[1]: ignition-kargs.service: Deactivated successfully. Apr 12 18:28:45.165000 audit: BPF prog-id=5 op=UNLOAD Apr 12 18:28:45.165000 audit: BPF prog-id=4 op=UNLOAD Apr 12 18:28:45.165000 audit: BPF prog-id=3 op=UNLOAD Apr 12 18:28:45.165000 audit: BPF prog-id=8 op=UNLOAD Apr 12 18:28:45.165000 audit: BPF prog-id=7 op=UNLOAD Apr 12 18:28:44.459700 systemd[1]: Stopped ignition-kargs.service. Apr 12 18:28:44.471808 systemd[1]: ignition-fetch.service: Deactivated successfully. Apr 12 18:28:44.471861 systemd[1]: Stopped ignition-fetch.service. Apr 12 18:28:44.483498 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Apr 12 18:28:44.483545 systemd[1]: Stopped ignition-fetch-offline.service. Apr 12 18:28:44.496345 systemd[1]: Stopped target paths.target. Apr 12 18:28:44.513688 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Apr 12 18:28:44.529921 systemd[1]: Stopped systemd-ask-password-console.path. Apr 12 18:28:44.534652 systemd[1]: Stopped target slices.target. Apr 12 18:28:44.541715 systemd[1]: Stopped target sockets.target. Apr 12 18:28:44.549496 systemd[1]: iscsid.socket: Deactivated successfully. Apr 12 18:28:44.549552 systemd[1]: Closed iscsid.socket. Apr 12 18:28:44.560962 systemd[1]: ignition-setup.service: Deactivated successfully. Apr 12 18:28:44.561009 systemd[1]: Stopped ignition-setup.service. Apr 12 18:28:44.593564 systemd[1]: Stopping iscsiuio.service... Apr 12 18:28:44.600613 systemd[1]: sysroot-boot.mount: Deactivated successfully. Apr 12 18:28:44.601125 systemd[1]: iscsiuio.service: Deactivated successfully. Apr 12 18:28:44.601239 systemd[1]: Stopped iscsiuio.service. Apr 12 18:28:44.631767 systemd[1]: sysroot-boot.service: Deactivated successfully. Apr 12 18:28:44.631860 systemd[1]: Stopped sysroot-boot.service. Apr 12 18:28:44.639920 systemd[1]: Stopped target network.target. Apr 12 18:28:44.661211 systemd[1]: iscsiuio.socket: Deactivated successfully. Apr 12 18:28:44.661260 systemd[1]: Closed iscsiuio.socket. Apr 12 18:28:45.205394 systemd-journald[236]: Received SIGTERM from PID 1 (n/a). Apr 12 18:28:44.670109 systemd[1]: initrd-setup-root.service: Deactivated successfully. Apr 12 18:28:44.670157 systemd[1]: Stopped initrd-setup-root.service. Apr 12 18:28:44.697778 systemd[1]: Stopping systemd-networkd.service... Apr 12 18:28:44.708260 systemd[1]: Stopping systemd-resolved.service... Apr 12 18:28:44.724202 systemd-networkd[826]: eth0: DHCPv6 lease lost Apr 12 18:28:45.205000 audit: BPF prog-id=9 op=UNLOAD Apr 12 18:28:44.726574 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 12 18:28:44.726679 systemd[1]: Stopped systemd-networkd.service. Apr 12 18:28:44.758388 systemd[1]: systemd-resolved.service: Deactivated successfully. Apr 12 18:28:44.758499 systemd[1]: Stopped systemd-resolved.service. Apr 12 18:28:44.767141 systemd[1]: systemd-networkd.socket: Deactivated successfully. Apr 12 18:28:44.767184 systemd[1]: Closed systemd-networkd.socket. Apr 12 18:28:44.794594 systemd[1]: Stopping network-cleanup.service... Apr 12 18:28:44.804101 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Apr 12 18:28:44.804184 systemd[1]: Stopped parse-ip-for-networkd.service. Apr 12 18:28:44.834634 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 12 18:28:44.834700 systemd[1]: Stopped systemd-sysctl.service. Apr 12 18:28:44.862046 systemd[1]: systemd-modules-load.service: Deactivated successfully. Apr 12 18:28:44.862103 systemd[1]: Stopped systemd-modules-load.service. Apr 12 18:28:44.894427 systemd[1]: Stopping systemd-udevd.service... Apr 12 18:28:44.903658 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Apr 12 18:28:44.904280 systemd[1]: systemd-udevd.service: Deactivated successfully. Apr 12 18:28:44.904426 systemd[1]: Stopped systemd-udevd.service. Apr 12 18:28:44.911517 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Apr 12 18:28:44.911570 systemd[1]: Closed systemd-udevd-control.socket. Apr 12 18:28:44.919903 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Apr 12 18:28:44.919950 systemd[1]: Closed systemd-udevd-kernel.socket. Apr 12 18:28:44.925517 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Apr 12 18:28:44.925594 systemd[1]: Stopped dracut-pre-udev.service. Apr 12 18:28:44.934338 systemd[1]: dracut-cmdline.service: Deactivated successfully. Apr 12 18:28:44.934412 systemd[1]: Stopped dracut-cmdline.service. Apr 12 18:28:44.939529 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Apr 12 18:28:44.939584 systemd[1]: Stopped dracut-cmdline-ask.service. Apr 12 18:28:44.947985 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Apr 12 18:28:44.956129 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Apr 12 18:28:44.956206 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. Apr 12 18:28:44.961334 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Apr 12 18:28:44.961419 systemd[1]: Stopped kmod-static-nodes.service. Apr 12 18:28:44.965778 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Apr 12 18:28:44.965824 systemd[1]: Stopped systemd-vconsole-setup.service. Apr 12 18:28:44.984229 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Apr 12 18:28:44.984759 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Apr 12 18:28:44.984854 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Apr 12 18:28:45.096987 systemd[1]: network-cleanup.service: Deactivated successfully. Apr 12 18:28:45.097097 systemd[1]: Stopped network-cleanup.service. Apr 12 18:28:45.104455 systemd[1]: Reached target initrd-switch-root.target. Apr 12 18:28:45.114459 systemd[1]: Starting initrd-switch-root.service... Apr 12 18:28:45.163492 systemd[1]: Switching root. Apr 12 18:28:45.206680 systemd-journald[236]: Journal stopped Apr 12 18:29:00.436237 kernel: SELinux: Class mctp_socket not defined in policy. Apr 12 18:29:00.436259 kernel: SELinux: Class anon_inode not defined in policy. Apr 12 18:29:00.436270 kernel: SELinux: the above unknown classes and permissions will be allowed Apr 12 18:29:00.436280 kernel: SELinux: policy capability network_peer_controls=1 Apr 12 18:29:00.436288 kernel: SELinux: policy capability open_perms=1 Apr 12 18:29:00.436296 kernel: SELinux: policy capability extended_socket_class=1 Apr 12 18:29:00.436304 kernel: SELinux: policy capability always_check_network=0 Apr 12 18:29:00.436313 kernel: SELinux: policy capability cgroup_seclabel=1 Apr 12 18:29:00.436321 kernel: SELinux: policy capability nnp_nosuid_transition=1 Apr 12 18:29:00.436329 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Apr 12 18:29:00.436338 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Apr 12 18:29:00.436348 systemd[1]: Successfully loaded SELinux policy in 335.863ms. Apr 12 18:29:00.436358 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 26.671ms. Apr 12 18:29:00.436368 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Apr 12 18:29:00.436394 systemd[1]: Detected virtualization microsoft. Apr 12 18:29:00.436404 systemd[1]: Detected architecture arm64. Apr 12 18:29:00.436412 systemd[1]: Detected first boot. Apr 12 18:29:00.436422 systemd[1]: Hostname set to . Apr 12 18:29:00.436431 systemd[1]: Initializing machine ID from random generator. Apr 12 18:29:00.436439 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Apr 12 18:29:00.436449 kernel: kauditd_printk_skb: 18 callbacks suppressed Apr 12 18:29:00.436459 kernel: audit: type=1400 audit(1712946531.799:89): avc: denied { associate } for pid=1072 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Apr 12 18:29:00.436471 kernel: audit: type=1300 audit(1712946531.799:89): arch=c00000b7 syscall=5 success=yes exit=0 a0=400014766c a1=40000c8af8 a2=40000cea00 a3=32 items=0 ppid=1055 pid=1072 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:29:00.436480 kernel: audit: type=1327 audit(1712946531.799:89): proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Apr 12 18:29:00.436489 kernel: audit: type=1400 audit(1712946531.814:90): avc: denied { associate } for pid=1072 comm="torcx-generator" name="bin" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Apr 12 18:29:00.436499 kernel: audit: type=1300 audit(1712946531.814:90): arch=c00000b7 syscall=34 success=yes exit=0 a0=ffffffffffffff9c a1=4000022119 a2=1ed a3=0 items=2 ppid=1055 pid=1072 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:29:00.436507 kernel: audit: type=1307 audit(1712946531.814:90): cwd="/" Apr 12 18:29:00.436518 kernel: audit: type=1302 audit(1712946531.814:90): item=0 name=(null) inode=2 dev=00:29 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:00.436527 kernel: audit: type=1302 audit(1712946531.814:90): item=1 name=(null) inode=3 dev=00:29 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:00.436537 kernel: audit: type=1327 audit(1712946531.814:90): proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Apr 12 18:29:00.436546 systemd[1]: Populated /etc with preset unit settings. Apr 12 18:29:00.436555 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Apr 12 18:29:00.436564 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Apr 12 18:29:00.436575 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 12 18:29:00.436585 systemd[1]: Queued start job for default target multi-user.target. Apr 12 18:29:00.436595 systemd[1]: Created slice system-addon\x2dconfig.slice. Apr 12 18:29:00.436604 systemd[1]: Created slice system-addon\x2drun.slice. Apr 12 18:29:00.436614 systemd[1]: Created slice system-getty.slice. Apr 12 18:29:00.436623 systemd[1]: Created slice system-modprobe.slice. Apr 12 18:29:00.436632 systemd[1]: Created slice system-serial\x2dgetty.slice. Apr 12 18:29:00.436644 systemd[1]: Created slice system-system\x2dcloudinit.slice. Apr 12 18:29:00.436655 systemd[1]: Created slice system-systemd\x2dfsck.slice. Apr 12 18:29:00.436665 systemd[1]: Created slice user.slice. Apr 12 18:29:00.436674 systemd[1]: Started systemd-ask-password-console.path. Apr 12 18:29:00.436683 systemd[1]: Started systemd-ask-password-wall.path. Apr 12 18:29:00.436692 systemd[1]: Set up automount boot.automount. Apr 12 18:29:00.436701 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Apr 12 18:29:00.436710 systemd[1]: Reached target integritysetup.target. Apr 12 18:29:00.436720 systemd[1]: Reached target remote-cryptsetup.target. Apr 12 18:29:00.436730 systemd[1]: Reached target remote-fs.target. Apr 12 18:29:00.436740 systemd[1]: Reached target slices.target. Apr 12 18:29:00.436749 systemd[1]: Reached target swap.target. Apr 12 18:29:00.436758 systemd[1]: Reached target torcx.target. Apr 12 18:29:00.436768 systemd[1]: Reached target veritysetup.target. Apr 12 18:29:00.436777 systemd[1]: Listening on systemd-coredump.socket. Apr 12 18:29:00.436787 systemd[1]: Listening on systemd-initctl.socket. Apr 12 18:29:00.436796 kernel: audit: type=1400 audit(1712946539.914:91): avc: denied { audit_read } for pid=1 comm="systemd" capability=37 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Apr 12 18:29:00.436806 systemd[1]: Listening on systemd-journald-audit.socket. Apr 12 18:29:00.436816 kernel: audit: type=1335 audit(1712946539.914:92): pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Apr 12 18:29:00.436825 systemd[1]: Listening on systemd-journald-dev-log.socket. Apr 12 18:29:00.436835 systemd[1]: Listening on systemd-journald.socket. Apr 12 18:29:00.436844 systemd[1]: Listening on systemd-networkd.socket. Apr 12 18:29:00.436854 systemd[1]: Listening on systemd-udevd-control.socket. Apr 12 18:29:00.436863 systemd[1]: Listening on systemd-udevd-kernel.socket. Apr 12 18:29:00.436874 systemd[1]: Listening on systemd-userdbd.socket. Apr 12 18:29:00.436883 systemd[1]: Mounting dev-hugepages.mount... Apr 12 18:29:00.436892 systemd[1]: Mounting dev-mqueue.mount... Apr 12 18:29:00.436902 systemd[1]: Mounting media.mount... Apr 12 18:29:00.436911 systemd[1]: Mounting sys-kernel-debug.mount... Apr 12 18:29:00.436921 systemd[1]: Mounting sys-kernel-tracing.mount... Apr 12 18:29:00.436930 systemd[1]: Mounting tmp.mount... Apr 12 18:29:00.436941 systemd[1]: Starting flatcar-tmpfiles.service... Apr 12 18:29:00.436951 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Apr 12 18:29:00.436960 systemd[1]: Starting kmod-static-nodes.service... Apr 12 18:29:00.436970 systemd[1]: Starting modprobe@configfs.service... Apr 12 18:29:00.436979 systemd[1]: Starting modprobe@dm_mod.service... Apr 12 18:29:00.436988 systemd[1]: Starting modprobe@drm.service... Apr 12 18:29:00.436998 systemd[1]: Starting modprobe@efi_pstore.service... Apr 12 18:29:00.437007 systemd[1]: Starting modprobe@fuse.service... Apr 12 18:29:00.437016 systemd[1]: Starting modprobe@loop.service... Apr 12 18:29:00.437027 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Apr 12 18:29:00.437038 systemd[1]: systemd-journald.service: unit configures an IP firewall, but the local system does not support BPF/cgroup firewalling. Apr 12 18:29:00.437047 systemd[1]: (This warning is only shown for the first unit using IP firewalling.) Apr 12 18:29:00.437057 kernel: fuse: init (API version 7.34) Apr 12 18:29:00.437067 systemd[1]: Starting systemd-journald.service... Apr 12 18:29:00.437076 kernel: loop: module loaded Apr 12 18:29:00.437085 systemd[1]: Starting systemd-modules-load.service... Apr 12 18:29:00.437094 systemd[1]: Starting systemd-network-generator.service... Apr 12 18:29:00.437105 systemd[1]: Starting systemd-remount-fs.service... Apr 12 18:29:00.437115 systemd[1]: Starting systemd-udev-trigger.service... Apr 12 18:29:00.437124 systemd[1]: Mounted dev-hugepages.mount. Apr 12 18:29:00.437134 systemd[1]: Mounted dev-mqueue.mount. Apr 12 18:29:00.437143 systemd[1]: Mounted media.mount. Apr 12 18:29:00.437152 systemd[1]: Mounted sys-kernel-debug.mount. Apr 12 18:29:00.437161 systemd[1]: Mounted sys-kernel-tracing.mount. Apr 12 18:29:00.437170 systemd[1]: Mounted tmp.mount. Apr 12 18:29:00.437180 systemd[1]: Finished flatcar-tmpfiles.service. Apr 12 18:29:00.437191 kernel: audit: type=1130 audit(1712946540.373:93): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.437200 systemd[1]: Finished kmod-static-nodes.service. Apr 12 18:29:00.437210 systemd[1]: modprobe@configfs.service: Deactivated successfully. Apr 12 18:29:00.437220 kernel: audit: type=1130 audit(1712946540.402:94): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.437229 systemd[1]: Finished modprobe@configfs.service. Apr 12 18:29:00.437243 systemd-journald[1183]: Journal started Apr 12 18:29:00.437287 systemd-journald[1183]: Runtime Journal (/run/log/journal/b47025488c7e4cea957592facdcafe72) is 8.0M, max 78.6M, 70.6M free. Apr 12 18:28:59.914000 audit[1]: EVENT_LISTENER pid=1 uid=0 auid=4294967295 tty=(none) ses=4294967295 subj=system_u:system_r:kernel_t:s0 comm="systemd" exe="/usr/lib/systemd/systemd" nl-mcgrp=1 op=connect res=1 Apr 12 18:29:00.373000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.402000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.433000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Apr 12 18:29:00.458476 kernel: audit: type=1305 audit(1712946540.433:95): op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Apr 12 18:29:00.458552 systemd[1]: Started systemd-journald.service. Apr 12 18:29:00.458570 kernel: audit: type=1300 audit(1712946540.433:95): arch=c00000b7 syscall=211 success=yes exit=60 a0=3 a1=fffffd3bced0 a2=4000 a3=1 items=0 ppid=1 pid=1183 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:29:00.433000 audit[1183]: SYSCALL arch=c00000b7 syscall=211 success=yes exit=60 a0=3 a1=fffffd3bced0 a2=4000 a3=1 items=0 ppid=1 pid=1183 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:29:00.483226 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Apr 12 18:29:00.483717 systemd[1]: Finished modprobe@dm_mod.service. Apr 12 18:29:00.433000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Apr 12 18:29:00.496183 kernel: audit: type=1327 audit(1712946540.433:95): proctitle="/usr/lib/systemd/systemd-journald" Apr 12 18:29:00.496324 systemd[1]: modprobe@drm.service: Deactivated successfully. Apr 12 18:29:00.496616 systemd[1]: Finished modprobe@drm.service. Apr 12 18:29:00.448000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.518718 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Apr 12 18:29:00.519013 systemd[1]: Finished modprobe@efi_pstore.service. Apr 12 18:29:00.519174 kernel: audit: type=1130 audit(1712946540.448:96): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.519221 kernel: audit: type=1131 audit(1712946540.448:97): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.519234 kernel: audit: type=1130 audit(1712946540.458:98): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.448000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.458000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.484000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.484000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.500000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.500000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.559000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.559000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.560484 systemd[1]: modprobe@fuse.service: Deactivated successfully. Apr 12 18:29:00.560747 systemd[1]: Finished modprobe@fuse.service. Apr 12 18:29:00.565000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.565000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.566571 systemd[1]: modprobe@loop.service: Deactivated successfully. Apr 12 18:29:00.566883 systemd[1]: Finished modprobe@loop.service. Apr 12 18:29:00.571000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.571000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.572553 systemd[1]: Finished systemd-modules-load.service. Apr 12 18:29:00.577000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.578513 systemd[1]: Finished systemd-network-generator.service. Apr 12 18:29:00.584000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.584953 systemd[1]: Finished systemd-remount-fs.service. Apr 12 18:29:00.589000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.590561 systemd[1]: Finished systemd-udev-trigger.service. Apr 12 18:29:00.595000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.596931 systemd[1]: Reached target network-pre.target. Apr 12 18:29:00.604109 systemd[1]: Mounting sys-fs-fuse-connections.mount... Apr 12 18:29:00.611239 systemd[1]: Mounting sys-kernel-config.mount... Apr 12 18:29:00.616232 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Apr 12 18:29:00.618602 systemd[1]: Starting systemd-hwdb-update.service... Apr 12 18:29:00.625154 systemd[1]: Starting systemd-journal-flush.service... Apr 12 18:29:00.630187 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Apr 12 18:29:00.632035 systemd[1]: Starting systemd-random-seed.service... Apr 12 18:29:00.636839 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Apr 12 18:29:00.638717 systemd[1]: Starting systemd-sysctl.service... Apr 12 18:29:00.644156 systemd[1]: Starting systemd-sysusers.service... Apr 12 18:29:00.649852 systemd[1]: Starting systemd-udev-settle.service... Apr 12 18:29:00.656582 systemd[1]: Mounted sys-fs-fuse-connections.mount. Apr 12 18:29:00.661593 systemd[1]: Mounted sys-kernel-config.mount. Apr 12 18:29:00.671963 udevadm[1223]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation-early.service, lvm2-activation.service not to pull it in. Apr 12 18:29:00.686441 systemd[1]: Finished systemd-random-seed.service. Apr 12 18:29:00.691000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.691729 systemd[1]: Reached target first-boot-complete.target. Apr 12 18:29:00.711980 systemd-journald[1183]: Time spent on flushing to /var/log/journal/b47025488c7e4cea957592facdcafe72 is 13.787ms for 1089 entries. Apr 12 18:29:00.711980 systemd-journald[1183]: System Journal (/var/log/journal/b47025488c7e4cea957592facdcafe72) is 8.0M, max 2.6G, 2.6G free. Apr 12 18:29:00.932800 systemd-journald[1183]: Received client request to flush runtime journal. Apr 12 18:29:00.755000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:00.751124 systemd[1]: Finished systemd-sysctl.service. Apr 12 18:29:00.933929 systemd[1]: Finished systemd-journal-flush.service. Apr 12 18:29:00.938000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:01.430558 systemd[1]: Finished systemd-sysusers.service. Apr 12 18:29:01.435000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:01.437644 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Apr 12 18:29:01.953247 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Apr 12 18:29:01.958000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:02.022456 systemd[1]: Finished systemd-hwdb-update.service. Apr 12 18:29:02.027000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:02.029267 systemd[1]: Starting systemd-udevd.service... Apr 12 18:29:02.050355 systemd-udevd[1234]: Using default interface naming scheme 'v252'. Apr 12 18:29:02.284038 systemd[1]: Started systemd-udevd.service. Apr 12 18:29:02.295000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:02.309312 systemd[1]: Starting systemd-networkd.service... Apr 12 18:29:02.334982 systemd[1]: Found device dev-ttyAMA0.device. Apr 12 18:29:02.404000 audit[1235]: AVC avc: denied { confidentiality } for pid=1235 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Apr 12 18:29:02.415408 kernel: mousedev: PS/2 mouse device common for all mice Apr 12 18:29:02.415516 kernel: hv_vmbus: registering driver hv_balloon Apr 12 18:29:02.431462 kernel: hv_balloon: Using Dynamic Memory protocol version 2.0 Apr 12 18:29:02.431622 kernel: hv_balloon: Memory hot add disabled on ARM64 Apr 12 18:29:02.433501 systemd[1]: Starting systemd-userdbd.service... Apr 12 18:29:02.404000 audit[1235]: SYSCALL arch=c00000b7 syscall=105 success=yes exit=0 a0=aaaafa2d5d80 a1=aa2c a2=ffff9f7024b0 a3=aaaafa20d010 items=12 ppid=1234 pid=1235 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:29:02.404000 audit: CWD cwd="/" Apr 12 18:29:02.404000 audit: PATH item=0 name=(null) inode=6738 dev=00:0a mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=1 name=(null) inode=10766 dev=00:0a mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=2 name=(null) inode=10766 dev=00:0a mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=3 name=(null) inode=10767 dev=00:0a mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=4 name=(null) inode=10766 dev=00:0a mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=5 name=(null) inode=10768 dev=00:0a mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=6 name=(null) inode=10766 dev=00:0a mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=7 name=(null) inode=10769 dev=00:0a mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=8 name=(null) inode=10766 dev=00:0a mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=9 name=(null) inode=10770 dev=00:0a mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=10 name=(null) inode=10766 dev=00:0a mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PATH item=11 name=(null) inode=10771 dev=00:0a mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Apr 12 18:29:02.404000 audit: PROCTITLE proctitle="(udev-worker)" Apr 12 18:29:02.481606 kernel: hv_vmbus: registering driver hyperv_fb Apr 12 18:29:02.481752 kernel: hv_utils: Registering HyperV Utility Driver Apr 12 18:29:02.481775 kernel: hv_vmbus: registering driver hv_utils Apr 12 18:29:02.498824 kernel: hv_utils: Heartbeat IC version 3.0 Apr 12 18:29:02.498933 kernel: hv_utils: Shutdown IC version 3.2 Apr 12 18:29:02.498958 kernel: hv_utils: TimeSync IC version 4.0 Apr 12 18:29:02.308264 kernel: hyperv_fb: Synthvid Version major 3, minor 5 Apr 12 18:29:02.368615 kernel: hyperv_fb: Screen resolution: 1024x768, Color depth: 32, Frame buffer size: 8388608 Apr 12 18:29:02.368649 systemd-journald[1183]: Time jumped backwards, rotating. Apr 12 18:29:02.368710 kernel: Console: switching to colour dummy device 80x25 Apr 12 18:29:02.368725 kernel: Console: switching to colour frame buffer device 128x48 Apr 12 18:29:02.352000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:02.347869 systemd[1]: Started systemd-userdbd.service. Apr 12 18:29:02.485231 kernel: BTRFS info: devid 1 device path /dev/disk/by-label/OEM changed to /dev/sda6 scanned by (udev-worker) (1244) Apr 12 18:29:02.502568 systemd[1]: dev-disk-by\x2dlabel-OEM.device was skipped because of an unmet condition check (ConditionPathExists=!/usr/.noupdate). Apr 12 18:29:02.504729 systemd[1]: Finished systemd-udev-settle.service. Apr 12 18:29:02.509000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:02.512549 systemd[1]: Starting lvm2-activation-early.service... Apr 12 18:29:02.768403 systemd-networkd[1255]: lo: Link UP Apr 12 18:29:02.768710 systemd-networkd[1255]: lo: Gained carrier Apr 12 18:29:02.769193 systemd-networkd[1255]: Enumeration completed Apr 12 18:29:02.769427 systemd[1]: Started systemd-networkd.service. Apr 12 18:29:02.773000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:02.775554 systemd[1]: Starting systemd-networkd-wait-online.service... Apr 12 18:29:02.798042 systemd-networkd[1255]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 12 18:29:02.847223 kernel: mlx5_core 778f:00:02.0 enP30607s1: Link up Apr 12 18:29:02.872229 kernel: hv_netvsc 000d3a06-e869-000d-3a06-e869000d3a06 eth0: Data path switched to VF: enP30607s1 Apr 12 18:29:02.872804 systemd-networkd[1255]: enP30607s1: Link UP Apr 12 18:29:02.872942 systemd-networkd[1255]: eth0: Link UP Apr 12 18:29:02.872950 systemd-networkd[1255]: eth0: Gained carrier Apr 12 18:29:02.877472 systemd-networkd[1255]: enP30607s1: Gained carrier Apr 12 18:29:02.888323 systemd-networkd[1255]: eth0: DHCPv4 address 10.200.20.18/24, gateway 10.200.20.1 acquired from 168.63.129.16 Apr 12 18:29:03.051913 lvm[1312]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 12 18:29:03.094301 systemd[1]: Finished lvm2-activation-early.service. Apr 12 18:29:03.098000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:03.099944 systemd[1]: Reached target cryptsetup.target. Apr 12 18:29:03.105901 systemd[1]: Starting lvm2-activation.service... Apr 12 18:29:03.110522 lvm[1316]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Apr 12 18:29:03.129257 systemd[1]: Finished lvm2-activation.service. Apr 12 18:29:03.132000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:03.134289 systemd[1]: Reached target local-fs-pre.target. Apr 12 18:29:03.139469 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Apr 12 18:29:03.139500 systemd[1]: Reached target local-fs.target. Apr 12 18:29:03.144035 systemd[1]: Reached target machines.target. Apr 12 18:29:03.149697 systemd[1]: Starting ldconfig.service... Apr 12 18:29:03.153608 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Apr 12 18:29:03.153687 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Apr 12 18:29:03.155102 systemd[1]: Starting systemd-boot-update.service... Apr 12 18:29:03.160854 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Apr 12 18:29:03.168015 systemd[1]: Starting systemd-machine-id-commit.service... Apr 12 18:29:03.172704 systemd[1]: systemd-sysext.service was skipped because no trigger condition checks were met. Apr 12 18:29:03.172778 systemd[1]: ensure-sysext.service was skipped because no trigger condition checks were met. Apr 12 18:29:03.174224 systemd[1]: Starting systemd-tmpfiles-setup.service... Apr 12 18:29:03.187695 systemd-tmpfiles[1322]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Apr 12 18:29:03.206830 systemd-tmpfiles[1322]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Apr 12 18:29:03.208645 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1319 (bootctl) Apr 12 18:29:03.210033 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Apr 12 18:29:03.233327 systemd-tmpfiles[1322]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Apr 12 18:29:03.885736 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Apr 12 18:29:03.891000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.107215 systemd-fsck[1328]: fsck.fat 4.2 (2021-01-31) Apr 12 18:29:04.107215 systemd-fsck[1328]: /dev/sda1: 236 files, 117047/258078 clusters Apr 12 18:29:04.109731 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Apr 12 18:29:04.115000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.118410 systemd[1]: Mounting boot.mount... Apr 12 18:29:04.133990 systemd[1]: Mounted boot.mount. Apr 12 18:29:04.151000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.147146 systemd[1]: Finished systemd-boot-update.service. Apr 12 18:29:04.224350 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Apr 12 18:29:04.225296 systemd[1]: Finished systemd-machine-id-commit.service. Apr 12 18:29:04.229000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.807339 systemd-networkd[1255]: eth0: Gained IPv6LL Apr 12 18:29:04.814311 systemd[1]: Finished systemd-networkd-wait-online.service. Apr 12 18:29:04.819000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-wait-online comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.824095 kernel: kauditd_printk_skb: 46 callbacks suppressed Apr 12 18:29:04.824213 kernel: audit: type=1130 audit(1712946544.819:130): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd-wait-online comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.871336 systemd[1]: Finished systemd-tmpfiles-setup.service. Apr 12 18:29:04.875000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.878228 systemd[1]: Starting audit-rules.service... Apr 12 18:29:04.896897 kernel: audit: type=1130 audit(1712946544.875:131): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.898860 systemd[1]: Starting clean-ca-certificates.service... Apr 12 18:29:04.905124 systemd[1]: Starting systemd-journal-catalog-update.service... Apr 12 18:29:04.913372 systemd[1]: Starting systemd-resolved.service... Apr 12 18:29:04.919908 systemd[1]: Starting systemd-timesyncd.service... Apr 12 18:29:04.926123 systemd[1]: Starting systemd-update-utmp.service... Apr 12 18:29:04.931105 systemd[1]: Finished clean-ca-certificates.service. Apr 12 18:29:04.934000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.936542 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Apr 12 18:29:04.956553 kernel: audit: type=1130 audit(1712946544.934:132): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.974000 audit[1347]: SYSTEM_BOOT pid=1347 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.999000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:04.995811 systemd[1]: Finished systemd-update-utmp.service. Apr 12 18:29:05.022113 kernel: audit: type=1127 audit(1712946544.974:133): pid=1347 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Apr 12 18:29:05.022255 kernel: audit: type=1130 audit(1712946544.999:134): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:05.093769 systemd[1]: Started systemd-timesyncd.service. Apr 12 18:29:05.097000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-timesyncd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:05.098895 systemd[1]: Reached target time-set.target. Apr 12 18:29:05.119420 kernel: audit: type=1130 audit(1712946545.097:135): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-timesyncd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:05.123807 systemd[1]: Finished systemd-journal-catalog-update.service. Apr 12 18:29:05.128000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:05.154223 kernel: audit: type=1130 audit(1712946545.128:136): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:05.173837 systemd-resolved[1345]: Positive Trust Anchors: Apr 12 18:29:05.174183 systemd-resolved[1345]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Apr 12 18:29:05.174295 systemd-resolved[1345]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Apr 12 18:29:05.177658 systemd-resolved[1345]: Using system hostname 'ci-3510.3.3-a-f11561af44'. Apr 12 18:29:05.179422 systemd[1]: Started systemd-resolved.service. Apr 12 18:29:05.183000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:05.184099 systemd[1]: Reached target network.target. Apr 12 18:29:05.205450 systemd[1]: Reached target network-online.target. Apr 12 18:29:05.207220 kernel: audit: type=1130 audit(1712946545.183:137): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Apr 12 18:29:05.211225 systemd[1]: Reached target nss-lookup.target. Apr 12 18:29:05.480388 systemd-timesyncd[1346]: Contacted time server 69.164.213.136:123 (0.flatcar.pool.ntp.org). Apr 12 18:29:05.480826 systemd-timesyncd[1346]: Initial clock synchronization to Fri 2024-04-12 18:29:05.476477 UTC. Apr 12 18:29:05.558000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Apr 12 18:29:05.560466 augenrules[1364]: No rules Apr 12 18:29:05.558000 audit[1364]: SYSCALL arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffc94e5c40 a2=420 a3=0 items=0 ppid=1340 pid=1364 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:29:05.573382 systemd[1]: Finished audit-rules.service. Apr 12 18:29:05.599164 kernel: audit: type=1305 audit(1712946545.558:138): auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Apr 12 18:29:05.599328 kernel: audit: type=1300 audit(1712946545.558:138): arch=c00000b7 syscall=206 success=yes exit=1056 a0=3 a1=ffffc94e5c40 a2=420 a3=0 items=0 ppid=1340 pid=1364 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Apr 12 18:29:05.558000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Apr 12 18:29:13.366008 ldconfig[1318]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Apr 12 18:29:13.376005 systemd[1]: Finished ldconfig.service. Apr 12 18:29:13.382543 systemd[1]: Starting systemd-update-done.service... Apr 12 18:29:13.407996 systemd[1]: Finished systemd-update-done.service. Apr 12 18:29:13.413690 systemd[1]: Reached target sysinit.target. Apr 12 18:29:13.418458 systemd[1]: Started motdgen.path. Apr 12 18:29:13.422656 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Apr 12 18:29:13.429464 systemd[1]: Started logrotate.timer. Apr 12 18:29:13.434977 systemd[1]: Started mdadm.timer. Apr 12 18:29:13.439113 systemd[1]: Started systemd-tmpfiles-clean.timer. Apr 12 18:29:13.444340 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Apr 12 18:29:13.444372 systemd[1]: Reached target paths.target. Apr 12 18:29:13.449089 systemd[1]: Reached target timers.target. Apr 12 18:29:13.454266 systemd[1]: Listening on dbus.socket. Apr 12 18:29:13.459944 systemd[1]: Starting docker.socket... Apr 12 18:29:13.486286 systemd[1]: Listening on sshd.socket. Apr 12 18:29:13.490674 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Apr 12 18:29:13.491146 systemd[1]: Listening on docker.socket. Apr 12 18:29:13.495857 systemd[1]: Reached target sockets.target. Apr 12 18:29:13.500718 systemd[1]: Reached target basic.target. Apr 12 18:29:13.505724 systemd[1]: System is tainted: cgroupsv1 Apr 12 18:29:13.505779 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Apr 12 18:29:13.505799 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Apr 12 18:29:13.507106 systemd[1]: Starting containerd.service... Apr 12 18:29:13.512808 systemd[1]: Starting dbus.service... Apr 12 18:29:13.518385 systemd[1]: Starting enable-oem-cloudinit.service... Apr 12 18:29:13.524479 systemd[1]: Starting extend-filesystems.service... Apr 12 18:29:13.529019 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Apr 12 18:29:13.530464 systemd[1]: Starting motdgen.service... Apr 12 18:29:13.535786 systemd[1]: Started nvidia.service. Apr 12 18:29:13.541856 systemd[1]: Starting prepare-cni-plugins.service... Apr 12 18:29:13.548337 systemd[1]: Starting prepare-critools.service... Apr 12 18:29:13.554180 systemd[1]: Starting prepare-helm.service... Apr 12 18:29:13.560957 systemd[1]: Starting ssh-key-proc-cmdline.service... Apr 12 18:29:13.569726 systemd[1]: Starting sshd-keygen.service... Apr 12 18:29:13.575864 systemd[1]: Starting systemd-logind.service... Apr 12 18:29:13.582473 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Apr 12 18:29:13.582561 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Apr 12 18:29:13.584045 systemd[1]: Starting update-engine.service... Apr 12 18:29:13.590993 systemd[1]: Starting update-ssh-keys-after-ignition.service... Apr 12 18:29:13.601294 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Apr 12 18:29:13.602597 systemd[1]: Finished ssh-key-proc-cmdline.service. Apr 12 18:29:13.616900 jq[1378]: false Apr 12 18:29:13.617232 jq[1400]: true Apr 12 18:29:13.619524 extend-filesystems[1379]: Found sda Apr 12 18:29:13.623987 extend-filesystems[1379]: Found sda1 Apr 12 18:29:13.623987 extend-filesystems[1379]: Found sda2 Apr 12 18:29:13.623987 extend-filesystems[1379]: Found sda3 Apr 12 18:29:13.623987 extend-filesystems[1379]: Found usr Apr 12 18:29:13.623987 extend-filesystems[1379]: Found sda4 Apr 12 18:29:13.623987 extend-filesystems[1379]: Found sda6 Apr 12 18:29:13.623987 extend-filesystems[1379]: Found sda7 Apr 12 18:29:13.623987 extend-filesystems[1379]: Found sda9 Apr 12 18:29:13.623987 extend-filesystems[1379]: Checking size of /dev/sda9 Apr 12 18:29:13.637600 systemd[1]: motdgen.service: Deactivated successfully. Apr 12 18:29:13.637890 systemd[1]: Finished motdgen.service. Apr 12 18:29:13.644391 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Apr 12 18:29:13.644687 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Apr 12 18:29:13.692645 jq[1417]: true Apr 12 18:29:13.693878 env[1410]: time="2024-04-12T18:29:13.693834424Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Apr 12 18:29:13.730940 env[1410]: time="2024-04-12T18:29:13.730169915Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Apr 12 18:29:13.731478 env[1410]: time="2024-04-12T18:29:13.731237235Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Apr 12 18:29:13.732724 env[1410]: time="2024-04-12T18:29:13.732679537Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.154-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Apr 12 18:29:13.732770 env[1410]: time="2024-04-12T18:29:13.732724451Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Apr 12 18:29:13.733030 env[1410]: time="2024-04-12T18:29:13.733000889Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 12 18:29:13.733030 env[1410]: time="2024-04-12T18:29:13.733024646Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Apr 12 18:29:13.733089 env[1410]: time="2024-04-12T18:29:13.733037964Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Apr 12 18:29:13.733089 env[1410]: time="2024-04-12T18:29:13.733047762Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Apr 12 18:29:13.733134 env[1410]: time="2024-04-12T18:29:13.733121511Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Apr 12 18:29:13.733422 env[1410]: time="2024-04-12T18:29:13.733397789Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Apr 12 18:29:13.733583 env[1410]: time="2024-04-12T18:29:13.733557685Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Apr 12 18:29:13.733583 env[1410]: time="2024-04-12T18:29:13.733579762Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Apr 12 18:29:13.733645 env[1410]: time="2024-04-12T18:29:13.733635074Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Apr 12 18:29:13.733685 env[1410]: time="2024-04-12T18:29:13.733647952Z" level=info msg="metadata content store policy set" policy=shared Apr 12 18:29:13.748797 env[1410]: time="2024-04-12T18:29:13.748744559Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Apr 12 18:29:13.748797 env[1410]: time="2024-04-12T18:29:13.748796192Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Apr 12 18:29:13.748966 env[1410]: time="2024-04-12T18:29:13.748810909Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Apr 12 18:29:13.748966 env[1410]: time="2024-04-12T18:29:13.748855063Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Apr 12 18:29:13.748966 env[1410]: time="2024-04-12T18:29:13.748871020Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Apr 12 18:29:13.748966 env[1410]: time="2024-04-12T18:29:13.748885018Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Apr 12 18:29:13.748966 env[1410]: time="2024-04-12T18:29:13.748897256Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Apr 12 18:29:13.749306 env[1410]: time="2024-04-12T18:29:13.749283038Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Apr 12 18:29:13.749356 env[1410]: time="2024-04-12T18:29:13.749310114Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Apr 12 18:29:13.749356 env[1410]: time="2024-04-12T18:29:13.749325232Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Apr 12 18:29:13.749356 env[1410]: time="2024-04-12T18:29:13.749338390Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Apr 12 18:29:13.749356 env[1410]: time="2024-04-12T18:29:13.749351748Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Apr 12 18:29:13.749527 env[1410]: time="2024-04-12T18:29:13.749504565Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Apr 12 18:29:13.749608 env[1410]: time="2024-04-12T18:29:13.749587233Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Apr 12 18:29:13.749920 env[1410]: time="2024-04-12T18:29:13.749898706Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Apr 12 18:29:13.749957 env[1410]: time="2024-04-12T18:29:13.749931381Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.749957 env[1410]: time="2024-04-12T18:29:13.749947498Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Apr 12 18:29:13.750018 env[1410]: time="2024-04-12T18:29:13.749998771Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750047 env[1410]: time="2024-04-12T18:29:13.750017008Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750047 env[1410]: time="2024-04-12T18:29:13.750031086Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750047 env[1410]: time="2024-04-12T18:29:13.750043444Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750108 env[1410]: time="2024-04-12T18:29:13.750062761Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750108 env[1410]: time="2024-04-12T18:29:13.750075959Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750108 env[1410]: time="2024-04-12T18:29:13.750087517Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750108 env[1410]: time="2024-04-12T18:29:13.750098756Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750184 env[1410]: time="2024-04-12T18:29:13.750113033Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Apr 12 18:29:13.750285 env[1410]: time="2024-04-12T18:29:13.750263531Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750316 env[1410]: time="2024-04-12T18:29:13.750286527Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750316 env[1410]: time="2024-04-12T18:29:13.750299285Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750316 env[1410]: time="2024-04-12T18:29:13.750312443Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Apr 12 18:29:13.750373 env[1410]: time="2024-04-12T18:29:13.750327201Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Apr 12 18:29:13.750373 env[1410]: time="2024-04-12T18:29:13.750339359Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Apr 12 18:29:13.750373 env[1410]: time="2024-04-12T18:29:13.750357077Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Apr 12 18:29:13.750435 env[1410]: time="2024-04-12T18:29:13.750391512Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Apr 12 18:29:13.750651 env[1410]: time="2024-04-12T18:29:13.750597321Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:false] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:false SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Apr 12 18:29:13.766934 env[1410]: time="2024-04-12T18:29:13.750659151Z" level=info msg="Connect containerd service" Apr 12 18:29:13.766934 env[1410]: time="2024-04-12T18:29:13.750701105Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Apr 12 18:29:13.766934 env[1410]: time="2024-04-12T18:29:13.751308494Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 12 18:29:13.766934 env[1410]: time="2024-04-12T18:29:13.753109502Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Apr 12 18:29:13.766934 env[1410]: time="2024-04-12T18:29:13.753167574Z" level=info msg=serving... address=/run/containerd/containerd.sock Apr 12 18:29:13.753368 systemd[1]: Started containerd.service. Apr 12 18:29:13.765768 systemd-logind[1396]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Apr 12 18:29:13.765972 systemd-logind[1396]: New seat seat0. Apr 12 18:29:13.772533 env[1410]: time="2024-04-12T18:29:13.772462430Z" level=info msg="Start subscribing containerd event" Apr 12 18:29:13.772533 env[1410]: time="2024-04-12T18:29:13.772531699Z" level=info msg="Start recovering state" Apr 12 18:29:13.772673 env[1410]: time="2024-04-12T18:29:13.772618566Z" level=info msg="Start event monitor" Apr 12 18:29:13.772673 env[1410]: time="2024-04-12T18:29:13.772639363Z" level=info msg="Start snapshots syncer" Apr 12 18:29:13.772673 env[1410]: time="2024-04-12T18:29:13.772651161Z" level=info msg="Start cni network conf syncer for default" Apr 12 18:29:13.772673 env[1410]: time="2024-04-12T18:29:13.772659720Z" level=info msg="Start streaming server" Apr 12 18:29:13.772825 env[1410]: time="2024-04-12T18:29:13.772796499Z" level=info msg="containerd successfully booted in 0.082099s" Apr 12 18:29:13.777487 tar[1405]: linux-arm64/helm Apr 12 18:29:13.778161 tar[1403]: ./ Apr 12 18:29:13.778161 tar[1403]: ./loopback Apr 12 18:29:13.778649 tar[1404]: crictl Apr 12 18:29:13.790725 extend-filesystems[1379]: Old size kept for /dev/sda9 Apr 12 18:29:13.798321 extend-filesystems[1379]: Found sr0 Apr 12 18:29:13.797655 systemd[1]: extend-filesystems.service: Deactivated successfully. Apr 12 18:29:13.815942 dbus-daemon[1377]: [system] SELinux support is enabled Apr 12 18:29:13.797950 systemd[1]: Finished extend-filesystems.service. Apr 12 18:29:13.827610 systemd[1]: Started dbus.service. Apr 12 18:29:13.847004 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Apr 12 18:29:13.847593 dbus-daemon[1377]: [system] Successfully activated service 'org.freedesktop.systemd1' Apr 12 18:29:13.847035 systemd[1]: Reached target system-config.target. Apr 12 18:29:13.858364 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Apr 12 18:29:13.858396 systemd[1]: Reached target user-config.target. Apr 12 18:29:13.866010 systemd[1]: Started systemd-logind.service. Apr 12 18:29:13.880735 bash[1440]: Updated "/home/core/.ssh/authorized_keys" Apr 12 18:29:13.881072 systemd[1]: Finished update-ssh-keys-after-ignition.service. Apr 12 18:29:13.921013 tar[1403]: ./bandwidth Apr 12 18:29:14.031645 tar[1403]: ./ptp Apr 12 18:29:14.077802 systemd[1]: nvidia.service: Deactivated successfully. Apr 12 18:29:14.133971 tar[1403]: ./vlan Apr 12 18:29:14.215046 tar[1403]: ./host-device Apr 12 18:29:14.281111 tar[1403]: ./tuning Apr 12 18:29:14.342227 tar[1403]: ./vrf Apr 12 18:29:14.401408 tar[1403]: ./sbr Apr 12 18:29:14.460355 tar[1403]: ./tap Apr 12 18:29:14.524351 update_engine[1397]: I0412 18:29:14.509908 1397 main.cc:92] Flatcar Update Engine starting Apr 12 18:29:14.529403 tar[1403]: ./dhcp Apr 12 18:29:14.575935 systemd[1]: Started update-engine.service. Apr 12 18:29:14.583289 update_engine[1397]: I0412 18:29:14.582983 1397 update_check_scheduler.cc:74] Next update check in 2m37s Apr 12 18:29:14.586429 systemd[1]: Started locksmithd.service. Apr 12 18:29:14.704234 tar[1405]: linux-arm64/LICENSE Apr 12 18:29:14.704234 tar[1405]: linux-arm64/README.md Apr 12 18:29:14.710699 systemd[1]: Finished prepare-helm.service. Apr 12 18:29:14.719775 tar[1403]: ./static Apr 12 18:29:14.747464 tar[1403]: ./firewall Apr 12 18:29:14.800480 tar[1403]: ./macvlan Apr 12 18:29:14.800700 systemd[1]: Finished prepare-critools.service. Apr 12 18:29:14.841645 tar[1403]: ./dummy Apr 12 18:29:14.874556 tar[1403]: ./bridge Apr 12 18:29:14.910739 tar[1403]: ./ipvlan Apr 12 18:29:14.943772 tar[1403]: ./portmap Apr 12 18:29:14.975257 tar[1403]: ./host-local Apr 12 18:29:15.082722 systemd[1]: Finished prepare-cni-plugins.service. Apr 12 18:29:16.074670 sshd_keygen[1401]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Apr 12 18:29:16.093219 systemd[1]: Finished sshd-keygen.service. Apr 12 18:29:16.099881 systemd[1]: Starting issuegen.service... Apr 12 18:29:16.105549 systemd[1]: Started waagent.service. Apr 12 18:29:16.111305 systemd[1]: issuegen.service: Deactivated successfully. Apr 12 18:29:16.111545 systemd[1]: Finished issuegen.service. Apr 12 18:29:16.118675 systemd[1]: Starting systemd-user-sessions.service... Apr 12 18:29:16.145326 systemd[1]: Finished systemd-user-sessions.service. Apr 12 18:29:16.152541 systemd[1]: Started getty@tty1.service. Apr 12 18:29:16.158403 systemd[1]: Started serial-getty@ttyAMA0.service. Apr 12 18:29:16.163342 systemd[1]: Reached target getty.target. Apr 12 18:29:16.167837 systemd[1]: Reached target multi-user.target. Apr 12 18:29:16.174080 systemd[1]: Starting systemd-update-utmp-runlevel.service... Apr 12 18:29:16.182839 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Apr 12 18:29:16.183101 systemd[1]: Finished systemd-update-utmp-runlevel.service. Apr 12 18:29:16.188782 systemd[1]: Startup finished in 19.884s (kernel) + 27.624s (userspace) = 47.508s. Apr 12 18:29:16.524267 locksmithd[1498]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Apr 12 18:29:17.016378 login[1530]: pam_lastlog(login:session): file /var/log/lastlog is locked/write Apr 12 18:29:17.036705 login[1531]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Apr 12 18:29:17.118899 systemd[1]: Created slice user-500.slice. Apr 12 18:29:17.119963 systemd[1]: Starting user-runtime-dir@500.service... Apr 12 18:29:17.123382 systemd-logind[1396]: New session 2 of user core. Apr 12 18:29:17.161064 systemd[1]: Finished user-runtime-dir@500.service. Apr 12 18:29:17.163120 systemd[1]: Starting user@500.service... Apr 12 18:29:17.214713 (systemd)[1537]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:29:17.477233 systemd[1537]: Queued start job for default target default.target. Apr 12 18:29:17.477832 systemd[1537]: Reached target paths.target. Apr 12 18:29:17.477940 systemd[1537]: Reached target sockets.target. Apr 12 18:29:17.478022 systemd[1537]: Reached target timers.target. Apr 12 18:29:17.478094 systemd[1537]: Reached target basic.target. Apr 12 18:29:17.478230 systemd[1537]: Reached target default.target. Apr 12 18:29:17.478321 systemd[1]: Started user@500.service. Apr 12 18:29:17.478419 systemd[1537]: Startup finished in 256ms. Apr 12 18:29:17.479305 systemd[1]: Started session-2.scope. Apr 12 18:29:18.017101 login[1530]: pam_unix(login:session): session opened for user core(uid=500) by LOGIN(uid=0) Apr 12 18:29:18.020832 systemd-logind[1396]: New session 1 of user core. Apr 12 18:29:18.021704 systemd[1]: Started session-1.scope. Apr 12 18:29:23.360020 waagent[1526]: 2024-04-12T18:29:23.359896Z INFO Daemon Daemon Azure Linux Agent Version:2.6.0.2 Apr 12 18:29:23.366818 waagent[1526]: 2024-04-12T18:29:23.366718Z INFO Daemon Daemon OS: flatcar 3510.3.3 Apr 12 18:29:23.371703 waagent[1526]: 2024-04-12T18:29:23.371620Z INFO Daemon Daemon Python: 3.9.16 Apr 12 18:29:23.376500 waagent[1526]: 2024-04-12T18:29:23.376409Z INFO Daemon Daemon Run daemon Apr 12 18:29:23.380887 waagent[1526]: 2024-04-12T18:29:23.380812Z INFO Daemon Daemon No RDMA handler exists for distro='Flatcar Container Linux by Kinvolk' version='3510.3.3' Apr 12 18:29:23.398419 waagent[1526]: 2024-04-12T18:29:23.398261Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 1. Apr 12 18:29:23.414431 waagent[1526]: 2024-04-12T18:29:23.414272Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Apr 12 18:29:23.424628 waagent[1526]: 2024-04-12T18:29:23.424527Z INFO Daemon Daemon cloud-init is enabled: False Apr 12 18:29:23.430549 waagent[1526]: 2024-04-12T18:29:23.430439Z INFO Daemon Daemon Using waagent for provisioning Apr 12 18:29:23.436728 waagent[1526]: 2024-04-12T18:29:23.436638Z INFO Daemon Daemon Activate resource disk Apr 12 18:29:23.441730 waagent[1526]: 2024-04-12T18:29:23.441635Z INFO Daemon Daemon Searching gen1 prefix 00000000-0001 or gen2 f8b3781a-1e82-4818-a1c3-63d806ec15bb Apr 12 18:29:23.456335 waagent[1526]: 2024-04-12T18:29:23.456237Z INFO Daemon Daemon Found device: None Apr 12 18:29:23.461133 waagent[1526]: 2024-04-12T18:29:23.461038Z ERROR Daemon Daemon Failed to mount resource disk [ResourceDiskError] unable to detect disk topology Apr 12 18:29:23.469929 waagent[1526]: 2024-04-12T18:29:23.469830Z ERROR Daemon Daemon Event: name=WALinuxAgent, op=ActivateResourceDisk, message=[ResourceDiskError] unable to detect disk topology, duration=0 Apr 12 18:29:23.482244 waagent[1526]: 2024-04-12T18:29:23.482140Z INFO Daemon Daemon Clean protocol and wireserver endpoint Apr 12 18:29:23.488447 waagent[1526]: 2024-04-12T18:29:23.488355Z INFO Daemon Daemon Running default provisioning handler Apr 12 18:29:23.501872 waagent[1526]: 2024-04-12T18:29:23.501707Z INFO Daemon Daemon Unable to get cloud-init enabled status from systemctl: Command '['systemctl', 'is-enabled', 'cloud-init-local.service']' returned non-zero exit status 1. Apr 12 18:29:23.517300 waagent[1526]: 2024-04-12T18:29:23.517120Z INFO Daemon Daemon Unable to get cloud-init enabled status from service: [Errno 2] No such file or directory: 'service' Apr 12 18:29:23.527303 waagent[1526]: 2024-04-12T18:29:23.527194Z INFO Daemon Daemon cloud-init is enabled: False Apr 12 18:29:23.532953 waagent[1526]: 2024-04-12T18:29:23.532849Z INFO Daemon Daemon Copying ovf-env.xml Apr 12 18:29:23.624269 waagent[1526]: 2024-04-12T18:29:23.624027Z INFO Daemon Daemon Successfully mounted dvd Apr 12 18:29:23.716754 systemd[1]: mnt-cdrom-secure.mount: Deactivated successfully. Apr 12 18:29:23.757008 waagent[1526]: 2024-04-12T18:29:23.756855Z INFO Daemon Daemon Detect protocol endpoint Apr 12 18:29:23.762420 waagent[1526]: 2024-04-12T18:29:23.762316Z INFO Daemon Daemon Clean protocol and wireserver endpoint Apr 12 18:29:23.768461 waagent[1526]: 2024-04-12T18:29:23.768359Z INFO Daemon Daemon WireServer endpoint is not found. Rerun dhcp handler Apr 12 18:29:23.775589 waagent[1526]: 2024-04-12T18:29:23.775491Z INFO Daemon Daemon Test for route to 168.63.129.16 Apr 12 18:29:23.781341 waagent[1526]: 2024-04-12T18:29:23.781252Z INFO Daemon Daemon Route to 168.63.129.16 exists Apr 12 18:29:23.788151 waagent[1526]: 2024-04-12T18:29:23.788056Z INFO Daemon Daemon Wire server endpoint:168.63.129.16 Apr 12 18:29:23.904285 waagent[1526]: 2024-04-12T18:29:23.904183Z INFO Daemon Daemon Fabric preferred wire protocol version:2015-04-05 Apr 12 18:29:23.911958 waagent[1526]: 2024-04-12T18:29:23.911905Z INFO Daemon Daemon Wire protocol version:2012-11-30 Apr 12 18:29:23.917567 waagent[1526]: 2024-04-12T18:29:23.917479Z INFO Daemon Daemon Server preferred version:2015-04-05 Apr 12 18:29:24.495831 waagent[1526]: 2024-04-12T18:29:24.495649Z INFO Daemon Daemon Initializing goal state during protocol detection Apr 12 18:29:24.511505 waagent[1526]: 2024-04-12T18:29:24.511400Z INFO Daemon Daemon Forcing an update of the goal state.. Apr 12 18:29:24.517787 waagent[1526]: 2024-04-12T18:29:24.517689Z INFO Daemon Daemon Fetching goal state [incarnation 1] Apr 12 18:29:24.597809 waagent[1526]: 2024-04-12T18:29:24.597642Z INFO Daemon Daemon Found private key matching thumbprint 6249EF38A94AD18564DD44227B5BF6C3A6BF8395 Apr 12 18:29:24.606472 waagent[1526]: 2024-04-12T18:29:24.606372Z INFO Daemon Daemon Certificate with thumbprint E0A086D96228ACE2A0FDE0B651338E5061F47204 has no matching private key. Apr 12 18:29:24.616120 waagent[1526]: 2024-04-12T18:29:24.616032Z INFO Daemon Daemon Fetch goal state completed Apr 12 18:29:24.663445 waagent[1526]: 2024-04-12T18:29:24.663379Z INFO Daemon Daemon Fetched new vmSettings [correlation ID: 0603d440-fc1e-4642-b715-a7d4ceb4e28b New eTag: 7212713086156601767] Apr 12 18:29:24.674390 waagent[1526]: 2024-04-12T18:29:24.674296Z INFO Daemon Daemon Status Blob type 'None' is not valid, assuming BlockBlob Apr 12 18:29:24.690028 waagent[1526]: 2024-04-12T18:29:24.689929Z INFO Daemon Daemon Starting provisioning Apr 12 18:29:24.695211 waagent[1526]: 2024-04-12T18:29:24.695105Z INFO Daemon Daemon Handle ovf-env.xml. Apr 12 18:29:24.700614 waagent[1526]: 2024-04-12T18:29:24.700531Z INFO Daemon Daemon Set hostname [ci-3510.3.3-a-f11561af44] Apr 12 18:29:24.745738 waagent[1526]: 2024-04-12T18:29:24.745589Z INFO Daemon Daemon Publish hostname [ci-3510.3.3-a-f11561af44] Apr 12 18:29:24.752914 waagent[1526]: 2024-04-12T18:29:24.752759Z INFO Daemon Daemon Examine /proc/net/route for primary interface Apr 12 18:29:24.759709 waagent[1526]: 2024-04-12T18:29:24.759620Z INFO Daemon Daemon Primary interface is [eth0] Apr 12 18:29:24.777300 systemd[1]: systemd-networkd-wait-online.service: Deactivated successfully. Apr 12 18:29:24.777532 systemd[1]: Stopped systemd-networkd-wait-online.service. Apr 12 18:29:24.777592 systemd[1]: Stopping systemd-networkd-wait-online.service... Apr 12 18:29:24.777781 systemd[1]: Stopping systemd-networkd.service... Apr 12 18:29:24.781267 systemd-networkd[1255]: eth0: DHCPv6 lease lost Apr 12 18:29:24.783297 systemd[1]: systemd-networkd.service: Deactivated successfully. Apr 12 18:29:24.783568 systemd[1]: Stopped systemd-networkd.service. Apr 12 18:29:24.785662 systemd[1]: Starting systemd-networkd.service... Apr 12 18:29:24.818991 systemd-networkd[1583]: enP30607s1: Link UP Apr 12 18:29:24.819004 systemd-networkd[1583]: enP30607s1: Gained carrier Apr 12 18:29:24.820073 systemd-networkd[1583]: eth0: Link UP Apr 12 18:29:24.820085 systemd-networkd[1583]: eth0: Gained carrier Apr 12 18:29:24.820672 systemd-networkd[1583]: lo: Link UP Apr 12 18:29:24.820683 systemd-networkd[1583]: lo: Gained carrier Apr 12 18:29:24.820930 systemd-networkd[1583]: eth0: Gained IPv6LL Apr 12 18:29:24.821901 systemd-networkd[1583]: Enumeration completed Apr 12 18:29:24.822047 systemd[1]: Started systemd-networkd.service. Apr 12 18:29:24.823827 waagent[1526]: 2024-04-12T18:29:24.823663Z INFO Daemon Daemon Create user account if not exists Apr 12 18:29:24.824059 systemd[1]: Starting systemd-networkd-wait-online.service... Apr 12 18:29:24.830777 waagent[1526]: 2024-04-12T18:29:24.830681Z INFO Daemon Daemon User core already exists, skip useradd Apr 12 18:29:24.837880 systemd-networkd[1583]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Apr 12 18:29:24.838122 waagent[1526]: 2024-04-12T18:29:24.837816Z INFO Daemon Daemon Configure sudoer Apr 12 18:29:24.843485 waagent[1526]: 2024-04-12T18:29:24.843380Z INFO Daemon Daemon Configure sshd Apr 12 18:29:24.848140 waagent[1526]: 2024-04-12T18:29:24.848031Z INFO Daemon Daemon Deploy ssh public key. Apr 12 18:29:24.860338 systemd-networkd[1583]: eth0: DHCPv4 address 10.200.20.18/24, gateway 10.200.20.1 acquired from 168.63.129.16 Apr 12 18:29:24.872433 systemd[1]: Finished systemd-networkd-wait-online.service. Apr 12 18:29:26.084662 waagent[1526]: 2024-04-12T18:29:26.084567Z INFO Daemon Daemon Provisioning complete Apr 12 18:29:26.107502 waagent[1526]: 2024-04-12T18:29:26.107426Z INFO Daemon Daemon RDMA capabilities are not enabled, skipping Apr 12 18:29:26.114483 waagent[1526]: 2024-04-12T18:29:26.114381Z INFO Daemon Daemon End of log to /dev/console. The agent will now check for updates and then will process extensions. Apr 12 18:29:26.125736 waagent[1526]: 2024-04-12T18:29:26.125632Z INFO Daemon Daemon Installed Agent WALinuxAgent-2.6.0.2 is the most current agent Apr 12 18:29:26.452802 waagent[1593]: 2024-04-12T18:29:26.452691Z INFO ExtHandler ExtHandler Agent WALinuxAgent-2.6.0.2 is running as the goal state agent Apr 12 18:29:26.453728 waagent[1593]: 2024-04-12T18:29:26.453653Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 12 18:29:26.453872 waagent[1593]: 2024-04-12T18:29:26.453823Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 12 18:29:26.467104 waagent[1593]: 2024-04-12T18:29:26.466996Z INFO ExtHandler ExtHandler Forcing an update of the goal state.. Apr 12 18:29:26.467342 waagent[1593]: 2024-04-12T18:29:26.467287Z INFO ExtHandler ExtHandler Fetching goal state [incarnation 1] Apr 12 18:29:26.542917 waagent[1593]: 2024-04-12T18:29:26.542752Z INFO ExtHandler ExtHandler Found private key matching thumbprint 6249EF38A94AD18564DD44227B5BF6C3A6BF8395 Apr 12 18:29:26.543166 waagent[1593]: 2024-04-12T18:29:26.543108Z INFO ExtHandler ExtHandler Certificate with thumbprint E0A086D96228ACE2A0FDE0B651338E5061F47204 has no matching private key. Apr 12 18:29:26.543439 waagent[1593]: 2024-04-12T18:29:26.543387Z INFO ExtHandler ExtHandler Fetch goal state completed Apr 12 18:29:26.558554 waagent[1593]: 2024-04-12T18:29:26.558493Z INFO ExtHandler ExtHandler Fetched new vmSettings [correlation ID: 509dd61a-5e33-4d9d-a169-c5ea096cf7bf New eTag: 7212713086156601767] Apr 12 18:29:26.559240 waagent[1593]: 2024-04-12T18:29:26.559157Z INFO ExtHandler ExtHandler Status Blob type 'None' is not valid, assuming BlockBlob Apr 12 18:29:26.671135 waagent[1593]: 2024-04-12T18:29:26.670971Z INFO ExtHandler ExtHandler Distro: flatcar-3510.3.3; OSUtil: CoreOSUtil; AgentService: waagent; Python: 3.9.16; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Apr 12 18:29:26.698615 waagent[1593]: 2024-04-12T18:29:26.698518Z INFO ExtHandler ExtHandler WALinuxAgent-2.6.0.2 running as process 1593 Apr 12 18:29:26.702723 waagent[1593]: 2024-04-12T18:29:26.702627Z INFO ExtHandler ExtHandler Cgroup monitoring is not supported on ['flatcar', '3510.3.3', '', 'Flatcar Container Linux by Kinvolk'] Apr 12 18:29:26.704347 waagent[1593]: 2024-04-12T18:29:26.704189Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Apr 12 18:29:26.830952 waagent[1593]: 2024-04-12T18:29:26.830879Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Apr 12 18:29:26.831466 waagent[1593]: 2024-04-12T18:29:26.831402Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Apr 12 18:29:26.839841 waagent[1593]: 2024-04-12T18:29:26.839770Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Apr 12 18:29:26.840522 waagent[1593]: 2024-04-12T18:29:26.840453Z ERROR ExtHandler ExtHandler Unable to setup the persistent firewall rules: [Errno 30] Read-only file system: '/lib/systemd/system/waagent-network-setup.service' Apr 12 18:29:26.841876 waagent[1593]: 2024-04-12T18:29:26.841798Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: configuration enabled [False], cgroups enabled [False], python supported: [True] Apr 12 18:29:26.843490 waagent[1593]: 2024-04-12T18:29:26.843400Z INFO ExtHandler ExtHandler Starting env monitor service. Apr 12 18:29:26.844226 waagent[1593]: 2024-04-12T18:29:26.844139Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 12 18:29:26.844535 waagent[1593]: 2024-04-12T18:29:26.844478Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 12 18:29:26.845300 waagent[1593]: 2024-04-12T18:29:26.845191Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Apr 12 18:29:26.845739 waagent[1593]: 2024-04-12T18:29:26.845683Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Apr 12 18:29:26.845739 waagent[1593]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Apr 12 18:29:26.845739 waagent[1593]: eth0 00000000 0114C80A 0003 0 0 1024 00000000 0 0 0 Apr 12 18:29:26.845739 waagent[1593]: eth0 0014C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Apr 12 18:29:26.845739 waagent[1593]: eth0 0114C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Apr 12 18:29:26.845739 waagent[1593]: eth0 10813FA8 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Apr 12 18:29:26.845739 waagent[1593]: eth0 FEA9FEA9 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Apr 12 18:29:26.848523 waagent[1593]: 2024-04-12T18:29:26.848328Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Apr 12 18:29:26.849649 waagent[1593]: 2024-04-12T18:29:26.849567Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 12 18:29:26.849980 waagent[1593]: 2024-04-12T18:29:26.849920Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 12 18:29:26.850462 waagent[1593]: 2024-04-12T18:29:26.850383Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Apr 12 18:29:26.851055 waagent[1593]: 2024-04-12T18:29:26.850987Z INFO EnvHandler ExtHandler Configure routes Apr 12 18:29:26.851155 waagent[1593]: 2024-04-12T18:29:26.851096Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Apr 12 18:29:26.852044 waagent[1593]: 2024-04-12T18:29:26.851923Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Apr 12 18:29:26.852133 waagent[1593]: 2024-04-12T18:29:26.852067Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Apr 12 18:29:26.852415 waagent[1593]: 2024-04-12T18:29:26.852346Z INFO EnvHandler ExtHandler Gateway:None Apr 12 18:29:26.853900 waagent[1593]: 2024-04-12T18:29:26.853825Z INFO EnvHandler ExtHandler Routes:None Apr 12 18:29:26.855604 waagent[1593]: 2024-04-12T18:29:26.855544Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Apr 12 18:29:26.863541 waagent[1593]: 2024-04-12T18:29:26.863458Z INFO ExtHandler ExtHandler Checking for agent updates (family: Prod) Apr 12 18:29:26.865155 waagent[1593]: 2024-04-12T18:29:26.865075Z WARNING ExtHandler ExtHandler Fetch failed: [HttpError] HTTPS is unavailable and required Apr 12 18:29:26.869775 waagent[1593]: 2024-04-12T18:29:26.869680Z INFO ExtHandler ExtHandler [PERIODIC] Request failed using the direct channel. Error: 'NoneType' object has no attribute 'getheaders' Apr 12 18:29:26.906157 waagent[1593]: 2024-04-12T18:29:26.906009Z ERROR EnvHandler ExtHandler Failed to get the PID of the DHCP client: invalid literal for int() with base 10: 'MainPID=1583' Apr 12 18:29:26.920461 waagent[1593]: 2024-04-12T18:29:26.920382Z INFO ExtHandler ExtHandler Default channel changed to HostGA channel. Apr 12 18:29:27.006287 waagent[1593]: 2024-04-12T18:29:27.005634Z INFO MonitorHandler ExtHandler Network interfaces: Apr 12 18:29:27.006287 waagent[1593]: Executing ['ip', '-a', '-o', 'link']: Apr 12 18:29:27.006287 waagent[1593]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Apr 12 18:29:27.006287 waagent[1593]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:06:e8:69 brd ff:ff:ff:ff:ff:ff Apr 12 18:29:27.006287 waagent[1593]: 3: enP30607s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:06:e8:69 brd ff:ff:ff:ff:ff:ff\ altname enP30607p0s2 Apr 12 18:29:27.006287 waagent[1593]: Executing ['ip', '-4', '-a', '-o', 'address']: Apr 12 18:29:27.006287 waagent[1593]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Apr 12 18:29:27.006287 waagent[1593]: 2: eth0 inet 10.200.20.18/24 metric 1024 brd 10.200.20.255 scope global eth0\ valid_lft forever preferred_lft forever Apr 12 18:29:27.006287 waagent[1593]: Executing ['ip', '-6', '-a', '-o', 'address']: Apr 12 18:29:27.006287 waagent[1593]: 1: lo inet6 ::1/128 scope host \ valid_lft forever preferred_lft forever Apr 12 18:29:27.006287 waagent[1593]: 2: eth0 inet6 fe80::20d:3aff:fe06:e869/64 scope link \ valid_lft forever preferred_lft forever Apr 12 18:29:27.028913 waagent[1593]: 2024-04-12T18:29:27.028839Z INFO ExtHandler ExtHandler Agent WALinuxAgent-2.6.0.2 discovered update WALinuxAgent-2.10.0.8 -- exiting Apr 12 18:29:27.129747 waagent[1526]: 2024-04-12T18:29:27.129625Z INFO Daemon Daemon Agent WALinuxAgent-2.6.0.2 launched with command '/usr/share/oem/python/bin/python -u /usr/share/oem/bin/waagent -run-exthandlers' is successfully running Apr 12 18:29:27.134090 waagent[1526]: 2024-04-12T18:29:27.134024Z INFO Daemon Daemon Determined Agent WALinuxAgent-2.10.0.8 to be the latest agent Apr 12 18:29:28.399108 waagent[1622]: 2024-04-12T18:29:28.398979Z INFO ExtHandler ExtHandler Azure Linux Agent (Goal State Agent version 2.10.0.8) Apr 12 18:29:28.399957 waagent[1622]: 2024-04-12T18:29:28.399874Z INFO ExtHandler ExtHandler OS: flatcar 3510.3.3 Apr 12 18:29:28.400095 waagent[1622]: 2024-04-12T18:29:28.400044Z INFO ExtHandler ExtHandler Python: 3.9.16 Apr 12 18:29:28.400267 waagent[1622]: 2024-04-12T18:29:28.400183Z INFO ExtHandler ExtHandler CPU Arch: aarch64 Apr 12 18:29:28.409404 waagent[1622]: 2024-04-12T18:29:28.409243Z INFO ExtHandler ExtHandler Distro: flatcar-3510.3.3; OSUtil: CoreOSUtil; AgentService: waagent; Python: 3.9.16; Arch: aarch64; systemd: True; LISDrivers: Absent; logrotate: logrotate 3.20.1; Apr 12 18:29:28.409895 waagent[1622]: 2024-04-12T18:29:28.409823Z INFO ExtHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 12 18:29:28.410050 waagent[1622]: 2024-04-12T18:29:28.410001Z INFO ExtHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 12 18:29:28.426043 waagent[1622]: 2024-04-12T18:29:28.425925Z INFO ExtHandler Fetched a new incarnation for the WireServer goal state [incarnation 1] Apr 12 18:29:28.436422 waagent[1622]: 2024-04-12T18:29:28.436348Z INFO ExtHandler ExtHandler HostGAPlugin version: 1.0.8.149 Apr 12 18:29:28.437660 waagent[1622]: 2024-04-12T18:29:28.437582Z INFO ExtHandler Apr 12 18:29:28.437845 waagent[1622]: 2024-04-12T18:29:28.437793Z INFO ExtHandler Fetched new vmSettings [HostGAPlugin correlation ID: 5b8effd0-7737-4d80-b1eb-ecb292bc90c4 eTag: 7212713086156601767 source: Fabric] Apr 12 18:29:28.438723 waagent[1622]: 2024-04-12T18:29:28.438654Z INFO ExtHandler The vmSettings originated via Fabric; will ignore them. Apr 12 18:29:28.440055 waagent[1622]: 2024-04-12T18:29:28.439982Z INFO ExtHandler Apr 12 18:29:28.440212 waagent[1622]: 2024-04-12T18:29:28.440151Z INFO ExtHandler Fetching full goal state from the WireServer [incarnation 1] Apr 12 18:29:28.447148 waagent[1622]: 2024-04-12T18:29:28.447087Z INFO ExtHandler ExtHandler Downloading artifacts profile blob Apr 12 18:29:28.447759 waagent[1622]: 2024-04-12T18:29:28.447703Z WARNING ExtHandler ExtHandler Fetch failed: [HttpError] HTTPS is unavailable and required Apr 12 18:29:28.471719 waagent[1622]: 2024-04-12T18:29:28.471643Z INFO ExtHandler ExtHandler Default channel changed to HostGAPlugin channel. Apr 12 18:29:28.552099 waagent[1622]: 2024-04-12T18:29:28.551932Z INFO ExtHandler Downloaded certificate {'thumbprint': 'E0A086D96228ACE2A0FDE0B651338E5061F47204', 'hasPrivateKey': False} Apr 12 18:29:28.553414 waagent[1622]: 2024-04-12T18:29:28.553332Z INFO ExtHandler Downloaded certificate {'thumbprint': '6249EF38A94AD18564DD44227B5BF6C3A6BF8395', 'hasPrivateKey': True} Apr 12 18:29:28.554706 waagent[1622]: 2024-04-12T18:29:28.554622Z INFO ExtHandler Fetch goal state completed Apr 12 18:29:28.579653 waagent[1622]: 2024-04-12T18:29:28.579505Z INFO ExtHandler ExtHandler OpenSSL version: OpenSSL 3.0.7 1 Nov 2022 (Library: OpenSSL 3.0.7 1 Nov 2022) Apr 12 18:29:28.593914 waagent[1622]: 2024-04-12T18:29:28.593776Z INFO ExtHandler ExtHandler WALinuxAgent-2.10.0.8 running as process 1622 Apr 12 18:29:28.597839 waagent[1622]: 2024-04-12T18:29:28.597743Z INFO ExtHandler ExtHandler Cgroup monitoring is not supported on ['flatcar', '3510.3.3', '', 'Flatcar Container Linux by Kinvolk'] Apr 12 18:29:28.599473 waagent[1622]: 2024-04-12T18:29:28.599389Z INFO ExtHandler ExtHandler Starting setup for Persistent firewall rules Apr 12 18:29:28.605741 waagent[1622]: 2024-04-12T18:29:28.605664Z INFO ExtHandler ExtHandler Firewalld service not running/unavailable, trying to set up waagent-network-setup.service Apr 12 18:29:28.606271 waagent[1622]: 2024-04-12T18:29:28.606166Z INFO ExtHandler ExtHandler Successfully updated the Binary file /var/lib/waagent/waagent-network-setup.py for firewall setup Apr 12 18:29:28.615424 waagent[1622]: 2024-04-12T18:29:28.615344Z INFO ExtHandler ExtHandler Service: waagent-network-setup.service not enabled. Adding it now Apr 12 18:29:28.616049 waagent[1622]: 2024-04-12T18:29:28.615976Z ERROR ExtHandler ExtHandler Unable to setup the persistent firewall rules: [Errno 30] Read-only file system: '/lib/systemd/system/waagent-network-setup.service' Apr 12 18:29:28.623821 waagent[1622]: 2024-04-12T18:29:28.623665Z INFO ExtHandler ExtHandler DROP rule is not available which implies no firewall rules are set yet. Environment thread will set it up. Apr 12 18:29:28.625049 waagent[1622]: 2024-04-12T18:29:28.624960Z INFO ExtHandler ExtHandler Checking if log collection is allowed at this time [False]. All three conditions must be met: configuration enabled [True], cgroups enabled [False], python supported: [True] Apr 12 18:29:28.627251 waagent[1622]: 2024-04-12T18:29:28.627125Z INFO ExtHandler ExtHandler Starting env monitor service. Apr 12 18:29:28.627820 waagent[1622]: 2024-04-12T18:29:28.627737Z INFO MonitorHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 12 18:29:28.628493 waagent[1622]: 2024-04-12T18:29:28.628430Z INFO MonitorHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 12 18:29:28.629320 waagent[1622]: 2024-04-12T18:29:28.629234Z INFO MonitorHandler ExtHandler Monitor.NetworkConfigurationChanges is disabled. Apr 12 18:29:28.630056 waagent[1622]: 2024-04-12T18:29:28.629960Z INFO ExtHandler ExtHandler Start SendTelemetryHandler service. Apr 12 18:29:28.630705 waagent[1622]: 2024-04-12T18:29:28.630559Z INFO SendTelemetryHandler ExtHandler Successfully started the SendTelemetryHandler thread Apr 12 18:29:28.630796 waagent[1622]: 2024-04-12T18:29:28.630717Z INFO EnvHandler ExtHandler WireServer endpoint 168.63.129.16 read from file Apr 12 18:29:28.631153 waagent[1622]: 2024-04-12T18:29:28.631077Z INFO ExtHandler ExtHandler Start Extension Telemetry service. Apr 12 18:29:28.631490 waagent[1622]: 2024-04-12T18:29:28.631404Z INFO MonitorHandler ExtHandler Routing table from /proc/net/route: Apr 12 18:29:28.631490 waagent[1622]: Iface Destination Gateway Flags RefCnt Use Metric Mask MTU Window IRTT Apr 12 18:29:28.631490 waagent[1622]: eth0 00000000 0114C80A 0003 0 0 1024 00000000 0 0 0 Apr 12 18:29:28.631490 waagent[1622]: eth0 0014C80A 00000000 0001 0 0 1024 00FFFFFF 0 0 0 Apr 12 18:29:28.631490 waagent[1622]: eth0 0114C80A 00000000 0005 0 0 1024 FFFFFFFF 0 0 0 Apr 12 18:29:28.631490 waagent[1622]: eth0 10813FA8 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Apr 12 18:29:28.631490 waagent[1622]: eth0 FEA9FEA9 0114C80A 0007 0 0 1024 FFFFFFFF 0 0 0 Apr 12 18:29:28.632025 waagent[1622]: 2024-04-12T18:29:28.631940Z INFO EnvHandler ExtHandler Wire server endpoint:168.63.129.16 Apr 12 18:29:28.632836 waagent[1622]: 2024-04-12T18:29:28.632633Z INFO TelemetryEventsCollector ExtHandler Extension Telemetry pipeline enabled: True Apr 12 18:29:28.633309 waagent[1622]: 2024-04-12T18:29:28.633191Z INFO EnvHandler ExtHandler Configure routes Apr 12 18:29:28.633729 waagent[1622]: 2024-04-12T18:29:28.633656Z INFO ExtHandler ExtHandler Goal State Period: 6 sec. This indicates how often the agent checks for new goal states and reports status. Apr 12 18:29:28.636022 waagent[1622]: 2024-04-12T18:29:28.635810Z INFO EnvHandler ExtHandler Gateway:None Apr 12 18:29:28.636834 waagent[1622]: 2024-04-12T18:29:28.636744Z INFO TelemetryEventsCollector ExtHandler Successfully started the TelemetryEventsCollector thread Apr 12 18:29:28.636976 waagent[1622]: 2024-04-12T18:29:28.636883Z INFO EnvHandler ExtHandler Routes:None Apr 12 18:29:28.658965 waagent[1622]: 2024-04-12T18:29:28.658718Z INFO ExtHandler ExtHandler Downloading agent manifest Apr 12 18:29:28.682519 waagent[1622]: 2024-04-12T18:29:28.682411Z INFO ExtHandler ExtHandler Apr 12 18:29:28.684080 waagent[1622]: 2024-04-12T18:29:28.683990Z INFO MonitorHandler ExtHandler Network interfaces: Apr 12 18:29:28.684080 waagent[1622]: Executing ['ip', '-a', '-o', 'link']: Apr 12 18:29:28.684080 waagent[1622]: 1: lo: mtu 65536 qdisc noqueue state UNKNOWN mode DEFAULT group default qlen 1000\ link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 Apr 12 18:29:28.684080 waagent[1622]: 2: eth0: mtu 1500 qdisc mq state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:06:e8:69 brd ff:ff:ff:ff:ff:ff Apr 12 18:29:28.684080 waagent[1622]: 3: enP30607s1: mtu 1500 qdisc mq master eth0 state UP mode DEFAULT group default qlen 1000\ link/ether 00:0d:3a:06:e8:69 brd ff:ff:ff:ff:ff:ff\ altname enP30607p0s2 Apr 12 18:29:28.684080 waagent[1622]: Executing ['ip', '-4', '-a', '-o', 'address']: Apr 12 18:29:28.684080 waagent[1622]: 1: lo inet 127.0.0.1/8 scope host lo\ valid_lft forever preferred_lft forever Apr 12 18:29:28.684080 waagent[1622]: 2: eth0 inet 10.200.20.18/24 metric 1024 brd 10.200.20.255 scope global eth0\ valid_lft forever preferred_lft forever Apr 12 18:29:28.684080 waagent[1622]: Executing ['ip', '-6', '-a', '-o', 'address']: Apr 12 18:29:28.684080 waagent[1622]: 1: lo inet6 ::1/128 scope host \ valid_lft forever preferred_lft forever Apr 12 18:29:28.684080 waagent[1622]: 2: eth0 inet6 fe80::20d:3aff:fe06:e869/64 scope link \ valid_lft forever preferred_lft forever Apr 12 18:29:28.689053 waagent[1622]: 2024-04-12T18:29:28.688862Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState started [incarnation_1 channel: WireServer source: Fabric activity: 4e62cc5a-317e-49c8-b27d-93720ed25d93 correlation e3438e60-28a0-49cc-ad2f-6ac45ce6ecf3 created: 2024-04-12T18:27:38.409424Z] Apr 12 18:29:28.693256 waagent[1622]: 2024-04-12T18:29:28.693137Z INFO ExtHandler ExtHandler No extension handlers found, not processing anything. Apr 12 18:29:28.696661 waagent[1622]: 2024-04-12T18:29:28.696574Z INFO ExtHandler ExtHandler ProcessExtensionsGoalState completed [incarnation_1 14 ms] Apr 12 18:29:28.719459 waagent[1622]: 2024-04-12T18:29:28.719373Z INFO ExtHandler ExtHandler Looking for existing remote access users. Apr 12 18:29:28.777767 waagent[1622]: 2024-04-12T18:29:28.777657Z INFO ExtHandler ExtHandler [HEARTBEAT] Agent WALinuxAgent-2.10.0.8 is running as the goal state agent [DEBUG HeartbeatCounter: 0;HeartbeatId: C63B54E0-618E-4CDF-9797-C52DE499255F;DroppedPackets: 0;UpdateGSErrors: 0;AutoUpdate: 1] Apr 12 18:29:28.959497 waagent[1622]: 2024-04-12T18:29:28.959284Z INFO EnvHandler ExtHandler Successfully added Azure fabric firewall rules. Current Firewall rules: Apr 12 18:29:28.959497 waagent[1622]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Apr 12 18:29:28.959497 waagent[1622]: pkts bytes target prot opt in out source destination Apr 12 18:29:28.959497 waagent[1622]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Apr 12 18:29:28.959497 waagent[1622]: pkts bytes target prot opt in out source destination Apr 12 18:29:28.959497 waagent[1622]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Apr 12 18:29:28.959497 waagent[1622]: pkts bytes target prot opt in out source destination Apr 12 18:29:28.959497 waagent[1622]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Apr 12 18:29:28.959497 waagent[1622]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Apr 12 18:29:28.959497 waagent[1622]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Apr 12 18:29:28.968912 waagent[1622]: 2024-04-12T18:29:28.968773Z INFO EnvHandler ExtHandler Current Firewall rules: Apr 12 18:29:28.968912 waagent[1622]: Chain INPUT (policy ACCEPT 0 packets, 0 bytes) Apr 12 18:29:28.968912 waagent[1622]: pkts bytes target prot opt in out source destination Apr 12 18:29:28.968912 waagent[1622]: Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) Apr 12 18:29:28.968912 waagent[1622]: pkts bytes target prot opt in out source destination Apr 12 18:29:28.968912 waagent[1622]: Chain OUTPUT (policy ACCEPT 0 packets, 0 bytes) Apr 12 18:29:28.968912 waagent[1622]: pkts bytes target prot opt in out source destination Apr 12 18:29:28.968912 waagent[1622]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 tcp dpt:53 Apr 12 18:29:28.968912 waagent[1622]: 0 0 ACCEPT tcp -- * * 0.0.0.0/0 168.63.129.16 owner UID match 0 Apr 12 18:29:28.968912 waagent[1622]: 0 0 DROP tcp -- * * 0.0.0.0/0 168.63.129.16 ctstate INVALID,NEW Apr 12 18:29:28.969928 waagent[1622]: 2024-04-12T18:29:28.969872Z INFO EnvHandler ExtHandler Set block dev timeout: sda with timeout: 300 Apr 12 18:29:50.354558 kernel: hv_balloon: Max. dynamic memory size: 4096 MB Apr 12 18:29:59.715659 update_engine[1397]: I0412 18:29:59.715286 1397 update_attempter.cc:509] Updating boot flags... Apr 12 18:30:19.772073 systemd[1]: Created slice system-sshd.slice. Apr 12 18:30:19.773412 systemd[1]: Started sshd@0-10.200.20.18:22-10.200.12.6:36706.service. Apr 12 18:30:20.481901 sshd[1740]: Accepted publickey for core from 10.200.12.6 port 36706 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:30:20.500323 sshd[1740]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:30:20.505066 systemd[1]: Started session-3.scope. Apr 12 18:30:20.506093 systemd-logind[1396]: New session 3 of user core. Apr 12 18:30:20.865272 systemd[1]: Started sshd@1-10.200.20.18:22-10.200.12.6:36710.service. Apr 12 18:30:21.274396 sshd[1745]: Accepted publickey for core from 10.200.12.6 port 36710 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:30:21.275727 sshd[1745]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:30:21.279735 systemd-logind[1396]: New session 4 of user core. Apr 12 18:30:21.280166 systemd[1]: Started session-4.scope. Apr 12 18:30:21.579413 sshd[1745]: pam_unix(sshd:session): session closed for user core Apr 12 18:30:21.582728 systemd[1]: sshd@1-10.200.20.18:22-10.200.12.6:36710.service: Deactivated successfully. Apr 12 18:30:21.583710 systemd-logind[1396]: Session 4 logged out. Waiting for processes to exit. Apr 12 18:30:21.583768 systemd[1]: session-4.scope: Deactivated successfully. Apr 12 18:30:21.584826 systemd-logind[1396]: Removed session 4. Apr 12 18:30:21.652925 systemd[1]: Started sshd@2-10.200.20.18:22-10.200.12.6:36726.service. Apr 12 18:30:22.090368 sshd[1752]: Accepted publickey for core from 10.200.12.6 port 36726 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:30:22.092873 sshd[1752]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:30:22.098279 systemd-logind[1396]: New session 5 of user core. Apr 12 18:30:22.099010 systemd[1]: Started session-5.scope. Apr 12 18:30:22.401066 sshd[1752]: pam_unix(sshd:session): session closed for user core Apr 12 18:30:22.404039 systemd[1]: sshd@2-10.200.20.18:22-10.200.12.6:36726.service: Deactivated successfully. Apr 12 18:30:22.405084 systemd[1]: session-5.scope: Deactivated successfully. Apr 12 18:30:22.405099 systemd-logind[1396]: Session 5 logged out. Waiting for processes to exit. Apr 12 18:30:22.406029 systemd-logind[1396]: Removed session 5. Apr 12 18:30:22.468373 systemd[1]: Started sshd@3-10.200.20.18:22-10.200.12.6:36734.service. Apr 12 18:30:22.878319 sshd[1759]: Accepted publickey for core from 10.200.12.6 port 36734 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:30:22.880001 sshd[1759]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:30:22.883914 systemd-logind[1396]: New session 6 of user core. Apr 12 18:30:22.884440 systemd[1]: Started session-6.scope. Apr 12 18:30:23.206339 sshd[1759]: pam_unix(sshd:session): session closed for user core Apr 12 18:30:23.210436 systemd[1]: sshd@3-10.200.20.18:22-10.200.12.6:36734.service: Deactivated successfully. Apr 12 18:30:23.211274 systemd[1]: session-6.scope: Deactivated successfully. Apr 12 18:30:23.212419 systemd-logind[1396]: Session 6 logged out. Waiting for processes to exit. Apr 12 18:30:23.213295 systemd-logind[1396]: Removed session 6. Apr 12 18:30:23.279188 systemd[1]: Started sshd@4-10.200.20.18:22-10.200.12.6:36744.service. Apr 12 18:30:23.715816 sshd[1766]: Accepted publickey for core from 10.200.12.6 port 36744 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:30:23.717188 sshd[1766]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:30:23.721309 systemd-logind[1396]: New session 7 of user core. Apr 12 18:30:23.721752 systemd[1]: Started session-7.scope. Apr 12 18:30:24.345286 sudo[1770]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Apr 12 18:30:24.345487 sudo[1770]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Apr 12 18:30:25.068098 systemd[1]: Starting docker.service... Apr 12 18:30:25.124818 env[1786]: time="2024-04-12T18:30:25.124756956Z" level=info msg="Starting up" Apr 12 18:30:25.126020 env[1786]: time="2024-04-12T18:30:25.125989522Z" level=info msg="parsed scheme: \"unix\"" module=grpc Apr 12 18:30:25.126159 env[1786]: time="2024-04-12T18:30:25.126145083Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Apr 12 18:30:25.126283 env[1786]: time="2024-04-12T18:30:25.126264843Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Apr 12 18:30:25.126352 env[1786]: time="2024-04-12T18:30:25.126337644Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Apr 12 18:30:25.128189 env[1786]: time="2024-04-12T18:30:25.128156533Z" level=info msg="parsed scheme: \"unix\"" module=grpc Apr 12 18:30:25.128189 env[1786]: time="2024-04-12T18:30:25.128180893Z" level=info msg="scheme \"unix\" not registered, fallback to default scheme" module=grpc Apr 12 18:30:25.128340 env[1786]: time="2024-04-12T18:30:25.128217333Z" level=info msg="ccResolverWrapper: sending update to cc: {[{unix:///var/run/docker/libcontainerd/docker-containerd.sock 0 }] }" module=grpc Apr 12 18:30:25.128340 env[1786]: time="2024-04-12T18:30:25.128230733Z" level=info msg="ClientConn switching balancer to \"pick_first\"" module=grpc Apr 12 18:30:25.134100 systemd[1]: var-lib-docker-check\x2doverlayfs\x2dsupport3809118725-merged.mount: Deactivated successfully. Apr 12 18:30:25.260777 env[1786]: time="2024-04-12T18:30:25.260742680Z" level=warning msg="Your kernel does not support cgroup blkio weight" Apr 12 18:30:25.261141 env[1786]: time="2024-04-12T18:30:25.261076442Z" level=warning msg="Your kernel does not support cgroup blkio weight_device" Apr 12 18:30:25.261475 env[1786]: time="2024-04-12T18:30:25.261459524Z" level=info msg="Loading containers: start." Apr 12 18:30:25.439228 kernel: Initializing XFRM netlink socket Apr 12 18:30:25.488383 env[1786]: time="2024-04-12T18:30:25.488340425Z" level=info msg="Default bridge (docker0) is assigned with an IP address 172.17.0.0/16. Daemon option --bip can be used to set a preferred IP address" Apr 12 18:30:25.632293 systemd-networkd[1583]: docker0: Link UP Apr 12 18:30:25.661755 env[1786]: time="2024-04-12T18:30:25.661713817Z" level=info msg="Loading containers: done." Apr 12 18:30:25.671668 systemd[1]: var-lib-docker-overlay2-opaque\x2dbug\x2dcheck475236013-merged.mount: Deactivated successfully. Apr 12 18:30:25.693337 env[1786]: time="2024-04-12T18:30:25.693241856Z" level=warning msg="Not using native diff for overlay2, this may cause degraded performance for building images: kernel has CONFIG_OVERLAY_FS_REDIRECT_DIR enabled" storage-driver=overlay2 Apr 12 18:30:25.693673 env[1786]: time="2024-04-12T18:30:25.693654378Z" level=info msg="Docker daemon" commit=112bdf3343 graphdriver(s)=overlay2 version=20.10.23 Apr 12 18:30:25.693901 env[1786]: time="2024-04-12T18:30:25.693884819Z" level=info msg="Daemon has completed initialization" Apr 12 18:30:25.735383 systemd[1]: Started docker.service. Apr 12 18:30:25.737116 env[1786]: time="2024-04-12T18:30:25.737050276Z" level=info msg="API listen on /run/docker.sock" Apr 12 18:30:25.754192 systemd[1]: Reloading. Apr 12 18:30:25.827433 /usr/lib/systemd/system-generators/torcx-generator[1922]: time="2024-04-12T18:30:25Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.3 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.3 /var/lib/torcx/store]" Apr 12 18:30:25.827461 /usr/lib/systemd/system-generators/torcx-generator[1922]: time="2024-04-12T18:30:25Z" level=info msg="torcx already run" Apr 12 18:30:25.878577 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Apr 12 18:30:25.878597 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Apr 12 18:30:25.895808 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 12 18:30:25.969530 systemd[1]: Started kubelet.service. Apr 12 18:30:26.043624 kubelet[1981]: E0412 18:30:26.043570 1981 run.go:74] "command failed" err="failed to load kubelet config file, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory, path: /var/lib/kubelet/config.yaml" Apr 12 18:30:26.045625 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 12 18:30:26.045799 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 12 18:30:29.799714 env[1410]: time="2024-04-12T18:30:29.799670548Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.27.12\"" Apr 12 18:30:30.802708 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount911447583.mount: Deactivated successfully. Apr 12 18:30:33.122422 env[1410]: time="2024-04-12T18:30:33.122376554Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver:v1.27.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:33.130268 env[1410]: time="2024-04-12T18:30:33.130167786Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:d4d4d261fc80c6c87ea30cb7d2b1a53b684be80fb7af5e16a2c97371e669f19f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:33.137016 env[1410]: time="2024-04-12T18:30:33.136964654Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-apiserver:v1.27.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:33.145054 env[1410]: time="2024-04-12T18:30:33.145014847Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-apiserver@sha256:cf0c29f585316888225cf254949988bdbedc7ba6238bc9a24bf6f0c508c42b6c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:33.145679 env[1410]: time="2024-04-12T18:30:33.145645449Z" level=info msg="PullImage \"registry.k8s.io/kube-apiserver:v1.27.12\" returns image reference \"sha256:d4d4d261fc80c6c87ea30cb7d2b1a53b684be80fb7af5e16a2c97371e669f19f\"" Apr 12 18:30:33.155867 env[1410]: time="2024-04-12T18:30:33.155811931Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.27.12\"" Apr 12 18:30:34.867846 env[1410]: time="2024-04-12T18:30:34.867770528Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager:v1.27.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:34.874229 env[1410]: time="2024-04-12T18:30:34.874165914Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:a4a7509f59f7f027d7c434948b3b8e5463b835d28675c76c6d1ff21d2c4e8f18,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:34.880050 env[1410]: time="2024-04-12T18:30:34.880008817Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-controller-manager:v1.27.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:34.886861 env[1410]: time="2024-04-12T18:30:34.886809124Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-controller-manager@sha256:6caa3a4278e87169371d031861e49db21742bcbd8df650d7fe519a1a7f6764af,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:34.887712 env[1410]: time="2024-04-12T18:30:34.887676888Z" level=info msg="PullImage \"registry.k8s.io/kube-controller-manager:v1.27.12\" returns image reference \"sha256:a4a7509f59f7f027d7c434948b3b8e5463b835d28675c76c6d1ff21d2c4e8f18\"" Apr 12 18:30:34.897470 env[1410]: time="2024-04-12T18:30:34.897421007Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.27.12\"" Apr 12 18:30:36.106835 env[1410]: time="2024-04-12T18:30:36.106776553Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler:v1.27.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:36.140742 env[1410]: time="2024-04-12T18:30:36.140682482Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:5de6108d9220f19bcc35bf81a2879e5ff2f6506c08af260c116b803579db675b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:36.146570 env[1410]: time="2024-04-12T18:30:36.146523104Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-scheduler:v1.27.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:36.153015 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 1. Apr 12 18:30:36.153193 systemd[1]: Stopped kubelet.service. Apr 12 18:30:36.154858 systemd[1]: Started kubelet.service. Apr 12 18:30:36.156761 env[1410]: time="2024-04-12T18:30:36.156718023Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-scheduler@sha256:b8bb7b17a4f915419575ceb885e128d0bb5ea8e67cb88dbde257988b770a4dce,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:36.157292 env[1410]: time="2024-04-12T18:30:36.157257945Z" level=info msg="PullImage \"registry.k8s.io/kube-scheduler:v1.27.12\" returns image reference \"sha256:5de6108d9220f19bcc35bf81a2879e5ff2f6506c08af260c116b803579db675b\"" Apr 12 18:30:36.170438 env[1410]: time="2024-04-12T18:30:36.170375755Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.27.12\"" Apr 12 18:30:36.211305 kubelet[2010]: E0412 18:30:36.211246 2010 run.go:74] "command failed" err="failed to load kubelet config file, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory, path: /var/lib/kubelet/config.yaml" Apr 12 18:30:36.213853 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 12 18:30:36.213995 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 12 18:30:37.404780 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3313420456.mount: Deactivated successfully. Apr 12 18:30:38.003279 env[1410]: time="2024-04-12T18:30:38.003231020Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.27.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:38.015567 env[1410]: time="2024-04-12T18:30:38.015524425Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7daec180765068529c26cc4c7c989513bebbe614cbbc58beebe1db17ae177e06,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:38.027288 env[1410]: time="2024-04-12T18:30:38.027253387Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.27.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:38.041299 env[1410]: time="2024-04-12T18:30:38.041230758Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:b0539f35b586abc54ca7660f9bb8a539d010b9e07d20e9e3d529cf0ca35d4ddf,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:38.041869 env[1410]: time="2024-04-12T18:30:38.041829560Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.27.12\" returns image reference \"sha256:7daec180765068529c26cc4c7c989513bebbe614cbbc58beebe1db17ae177e06\"" Apr 12 18:30:38.052696 env[1410]: time="2024-04-12T18:30:38.052659519Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\"" Apr 12 18:30:38.752915 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2556464974.mount: Deactivated successfully. Apr 12 18:30:38.777638 env[1410]: time="2024-04-12T18:30:38.777597260Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:38.787272 env[1410]: time="2024-04-12T18:30:38.787224614Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:38.793190 env[1410]: time="2024-04-12T18:30:38.793152116Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.9,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:38.803101 env[1410]: time="2024-04-12T18:30:38.803058832Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:7031c1b283388d2c2e09b57badb803c05ebed362dc88d84b480cc47f72a21097,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:38.803704 env[1410]: time="2024-04-12T18:30:38.803675194Z" level=info msg="PullImage \"registry.k8s.io/pause:3.9\" returns image reference \"sha256:829e9de338bd5fdd3f16f68f83a9fb288fbc8453e881e5d5cfd0f6f2ff72b43e\"" Apr 12 18:30:38.813664 env[1410]: time="2024-04-12T18:30:38.813611470Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.7-0\"" Apr 12 18:30:39.714121 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2966124143.mount: Deactivated successfully. Apr 12 18:30:42.845064 env[1410]: time="2024-04-12T18:30:42.845015334Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd:3.5.7-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:42.861659 env[1410]: time="2024-04-12T18:30:42.861617789Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:24bc64e911039ecf00e263be2161797c758b7d82403ca5516ab64047a477f737,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:42.869877 env[1410]: time="2024-04-12T18:30:42.869839296Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/etcd:3.5.7-0,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:42.875311 env[1410]: time="2024-04-12T18:30:42.875262074Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/etcd@sha256:51eae8381dcb1078289fa7b4f3df2630cdc18d09fb56f8e56b41c40e191d6c83,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:42.876105 env[1410]: time="2024-04-12T18:30:42.876071836Z" level=info msg="PullImage \"registry.k8s.io/etcd:3.5.7-0\" returns image reference \"sha256:24bc64e911039ecf00e263be2161797c758b7d82403ca5516ab64047a477f737\"" Apr 12 18:30:42.885313 env[1410]: time="2024-04-12T18:30:42.885249987Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\"" Apr 12 18:30:43.666084 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount922773224.mount: Deactivated successfully. Apr 12 18:30:44.151257 env[1410]: time="2024-04-12T18:30:44.151115685Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns:v1.10.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:44.158010 env[1410]: time="2024-04-12T18:30:44.157967706Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:97e04611ad43405a2e5863ae17c6f1bc9181bdefdaa78627c432ef754a4eb108,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:44.163874 env[1410]: time="2024-04-12T18:30:44.163834205Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/coredns/coredns:v1.10.1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:44.167628 env[1410]: time="2024-04-12T18:30:44.167585417Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/coredns/coredns@sha256:a0ead06651cf580044aeb0a0feba63591858fb2e43ade8c9dea45a6a89ae7e5e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:44.168071 env[1410]: time="2024-04-12T18:30:44.168043738Z" level=info msg="PullImage \"registry.k8s.io/coredns/coredns:v1.10.1\" returns image reference \"sha256:97e04611ad43405a2e5863ae17c6f1bc9181bdefdaa78627c432ef754a4eb108\"" Apr 12 18:30:46.403000 systemd[1]: kubelet.service: Scheduled restart job, restart counter is at 2. Apr 12 18:30:46.403182 systemd[1]: Stopped kubelet.service. Apr 12 18:30:46.404757 systemd[1]: Started kubelet.service. Apr 12 18:30:46.464134 kubelet[2095]: E0412 18:30:46.464085 2095 run.go:74] "command failed" err="failed to load kubelet config file, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory, path: /var/lib/kubelet/config.yaml" Apr 12 18:30:46.465673 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Apr 12 18:30:46.465821 systemd[1]: kubelet.service: Failed with result 'exit-code'. Apr 12 18:30:48.963469 systemd[1]: Stopped kubelet.service. Apr 12 18:30:48.980227 systemd[1]: Reloading. Apr 12 18:30:49.053056 /usr/lib/systemd/system-generators/torcx-generator[2126]: time="2024-04-12T18:30:49Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.3 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.3 /var/lib/torcx/store]" Apr 12 18:30:49.057524 /usr/lib/systemd/system-generators/torcx-generator[2126]: time="2024-04-12T18:30:49Z" level=info msg="torcx already run" Apr 12 18:30:49.129599 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Apr 12 18:30:49.129781 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Apr 12 18:30:49.147424 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 12 18:30:49.230409 systemd[1]: Started kubelet.service. Apr 12 18:30:49.290976 kubelet[2191]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 12 18:30:49.290976 kubelet[2191]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Apr 12 18:30:49.290976 kubelet[2191]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 12 18:30:49.291394 kubelet[2191]: I0412 18:30:49.291060 2191 server.go:199] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 12 18:30:50.618978 kubelet[2191]: I0412 18:30:50.618933 2191 server.go:415] "Kubelet version" kubeletVersion="v1.27.2" Apr 12 18:30:50.618978 kubelet[2191]: I0412 18:30:50.618970 2191 server.go:417] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 12 18:30:50.619391 kubelet[2191]: I0412 18:30:50.619189 2191 server.go:837] "Client rotation is on, will bootstrap in background" Apr 12 18:30:50.624695 kubelet[2191]: I0412 18:30:50.624652 2191 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 12 18:30:50.625874 kubelet[2191]: E0412 18:30:50.625851 2191 certificate_manager.go:562] kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post "https://10.200.20.18:6443/apis/certificates.k8s.io/v1/certificatesigningrequests": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:50.626862 kubelet[2191]: W0412 18:30:50.626844 2191 machine.go:65] Cannot read vendor id correctly, set empty. Apr 12 18:30:50.627527 kubelet[2191]: I0412 18:30:50.627508 2191 server.go:662] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 12 18:30:50.627991 kubelet[2191]: I0412 18:30:50.627974 2191 container_manager_linux.go:266] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 12 18:30:50.628143 kubelet[2191]: I0412 18:30:50.628131 2191 container_manager_linux.go:271] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName: SystemCgroupsName: KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:cgroupfs KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[] HardEvictionThresholds:[{Signal:nodefs.available Operator:LessThan Value:{Quantity: Percentage:0.1} GracePeriod:0s MinReclaim:} {Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity: Percentage:0.05} GracePeriod:0s MinReclaim:} {Signal:imagefs.available Operator:LessThan Value:{Quantity: Percentage:0.15} GracePeriod:0s MinReclaim:} {Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:}]} QOSReserved:map[] CPUManagerPolicy:none CPUManagerPolicyOptions:map[] TopologyManagerScope:container CPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] PodPidsLimit:-1 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms TopologyManagerPolicy:none ExperimentalTopologyManagerPolicyOptions:map[]} Apr 12 18:30:50.628315 kubelet[2191]: I0412 18:30:50.628302 2191 topology_manager.go:136] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Apr 12 18:30:50.628396 kubelet[2191]: I0412 18:30:50.628386 2191 container_manager_linux.go:302] "Creating device plugin manager" Apr 12 18:30:50.628569 kubelet[2191]: I0412 18:30:50.628556 2191 state_mem.go:36] "Initialized new in-memory state store" Apr 12 18:30:50.631712 kubelet[2191]: I0412 18:30:50.631685 2191 kubelet.go:405] "Attempting to sync node with API server" Apr 12 18:30:50.631712 kubelet[2191]: I0412 18:30:50.631714 2191 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 12 18:30:50.632120 kubelet[2191]: I0412 18:30:50.632095 2191 kubelet.go:309] "Adding apiserver pod source" Apr 12 18:30:50.632176 kubelet[2191]: I0412 18:30:50.632124 2191 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 12 18:30:50.632319 kubelet[2191]: W0412 18:30:50.632279 2191 reflector.go:533] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://10.200.20.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.3-a-f11561af44&limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:50.632414 kubelet[2191]: E0412 18:30:50.632401 2191 reflector.go:148] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.200.20.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.3-a-f11561af44&limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:50.632903 kubelet[2191]: I0412 18:30:50.632884 2191 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Apr 12 18:30:50.633295 kubelet[2191]: W0412 18:30:50.633277 2191 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Apr 12 18:30:50.633788 kubelet[2191]: I0412 18:30:50.633769 2191 server.go:1168] "Started kubelet" Apr 12 18:30:50.642361 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Apr 12 18:30:50.642491 kubelet[2191]: W0412 18:30:50.637218 2191 reflector.go:533] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://10.200.20.18:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:50.642491 kubelet[2191]: E0412 18:30:50.637262 2191 reflector.go:148] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.200.20.18:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:50.642491 kubelet[2191]: E0412 18:30:50.637323 2191 event.go:289] Unable to write event: '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"ci-3510.3.3-a-f11561af44.17c59be736383677", GenerateName:"", Namespace:"default", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Node", Namespace:"", Name:"ci-3510.3.3-a-f11561af44", UID:"ci-3510.3.3-a-f11561af44", APIVersion:"", ResourceVersion:"", FieldPath:""}, Reason:"Starting", Message:"Starting kubelet.", Source:v1.EventSource{Component:"kubelet", Host:"ci-3510.3.3-a-f11561af44"}, FirstTimestamp:time.Date(2024, time.April, 12, 18, 30, 50, 633746039, time.Local), LastTimestamp:time.Date(2024, time.April, 12, 18, 30, 50, 633746039, time.Local), Count:1, Type:"Normal", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Post "https://10.200.20.18:6443/api/v1/namespaces/default/events": dial tcp 10.200.20.18:6443: connect: connection refused'(may retry after sleeping) Apr 12 18:30:50.642641 kubelet[2191]: I0412 18:30:50.637530 2191 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10 Apr 12 18:30:50.642641 kubelet[2191]: I0412 18:30:50.637808 2191 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Apr 12 18:30:50.642641 kubelet[2191]: I0412 18:30:50.638433 2191 server.go:461] "Adding debug handlers to kubelet server" Apr 12 18:30:50.643412 kubelet[2191]: E0412 18:30:50.643384 2191 cri_stats_provider.go:455] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Apr 12 18:30:50.643412 kubelet[2191]: E0412 18:30:50.643416 2191 kubelet.go:1400] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 12 18:30:50.643984 kubelet[2191]: I0412 18:30:50.643967 2191 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 12 18:30:50.646027 kubelet[2191]: I0412 18:30:50.645763 2191 volume_manager.go:284] "Starting Kubelet Volume Manager" Apr 12 18:30:50.646027 kubelet[2191]: I0412 18:30:50.645891 2191 desired_state_of_world_populator.go:145] "Desired state populator starts to run" Apr 12 18:30:50.646331 kubelet[2191]: W0412 18:30:50.646287 2191 reflector.go:533] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://10.200.20.18:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:50.646376 kubelet[2191]: E0412 18:30:50.646340 2191 reflector.go:148] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.200.20.18:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:50.647185 kubelet[2191]: E0412 18:30:50.647150 2191 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.3-a-f11561af44?timeout=10s\": dial tcp 10.200.20.18:6443: connect: connection refused" interval="200ms" Apr 12 18:30:50.745592 kubelet[2191]: I0412 18:30:50.745567 2191 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4 Apr 12 18:30:50.746806 kubelet[2191]: I0412 18:30:50.746767 2191 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6 Apr 12 18:30:50.746974 kubelet[2191]: I0412 18:30:50.746964 2191 status_manager.go:207] "Starting to sync pod status with apiserver" Apr 12 18:30:50.747218 kubelet[2191]: I0412 18:30:50.747191 2191 kubelet.go:2257] "Starting kubelet main sync loop" Apr 12 18:30:50.748314 kubelet[2191]: E0412 18:30:50.748287 2191 kubelet.go:2281] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 12 18:30:50.748413 kubelet[2191]: W0412 18:30:50.748380 2191 reflector.go:533] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://10.200.20.18:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:50.748450 kubelet[2191]: E0412 18:30:50.748428 2191 reflector.go:148] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.200.20.18:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:50.791397 kubelet[2191]: I0412 18:30:50.791357 2191 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.791893 kubelet[2191]: E0412 18:30:50.791723 2191 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.200.20.18:6443/api/v1/nodes\": dial tcp 10.200.20.18:6443: connect: connection refused" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.792342 kubelet[2191]: I0412 18:30:50.792320 2191 cpu_manager.go:214] "Starting CPU manager" policy="none" Apr 12 18:30:50.792342 kubelet[2191]: I0412 18:30:50.792340 2191 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Apr 12 18:30:50.792451 kubelet[2191]: I0412 18:30:50.792361 2191 state_mem.go:36] "Initialized new in-memory state store" Apr 12 18:30:50.796872 kubelet[2191]: I0412 18:30:50.796843 2191 policy_none.go:49] "None policy: Start" Apr 12 18:30:50.797669 kubelet[2191]: I0412 18:30:50.797649 2191 memory_manager.go:169] "Starting memorymanager" policy="None" Apr 12 18:30:50.797750 kubelet[2191]: I0412 18:30:50.797682 2191 state_mem.go:35] "Initializing new in-memory state store" Apr 12 18:30:50.807749 kubelet[2191]: I0412 18:30:50.807712 2191 manager.go:455] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 12 18:30:50.807964 kubelet[2191]: I0412 18:30:50.807941 2191 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 12 18:30:50.809671 kubelet[2191]: E0412 18:30:50.809643 2191 eviction_manager.go:262] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"ci-3510.3.3-a-f11561af44\" not found" Apr 12 18:30:50.848381 kubelet[2191]: I0412 18:30:50.848357 2191 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:30:50.848610 kubelet[2191]: E0412 18:30:50.848414 2191 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.3-a-f11561af44?timeout=10s\": dial tcp 10.200.20.18:6443: connect: connection refused" interval="400ms" Apr 12 18:30:50.850337 kubelet[2191]: I0412 18:30:50.850311 2191 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:30:50.851813 kubelet[2191]: I0412 18:30:50.851785 2191 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:30:50.947299 kubelet[2191]: I0412 18:30:50.947249 2191 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.947299 kubelet[2191]: I0412 18:30:50.947292 2191 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/faca4c8be60ec53ec91fce85ebbca199-kubeconfig\") pod \"kube-scheduler-ci-3510.3.3-a-f11561af44\" (UID: \"faca4c8be60ec53ec91fce85ebbca199\") " pod="kube-system/kube-scheduler-ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.947461 kubelet[2191]: I0412 18:30:50.947319 2191 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/16f742faf238cd425321419501894d1a-ca-certs\") pod \"kube-apiserver-ci-3510.3.3-a-f11561af44\" (UID: \"16f742faf238cd425321419501894d1a\") " pod="kube-system/kube-apiserver-ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.947461 kubelet[2191]: I0412 18:30:50.947340 2191 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-flexvolume-dir\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.947461 kubelet[2191]: I0412 18:30:50.947360 2191 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-k8s-certs\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.947461 kubelet[2191]: I0412 18:30:50.947379 2191 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-kubeconfig\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.947461 kubelet[2191]: I0412 18:30:50.947397 2191 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/16f742faf238cd425321419501894d1a-k8s-certs\") pod \"kube-apiserver-ci-3510.3.3-a-f11561af44\" (UID: \"16f742faf238cd425321419501894d1a\") " pod="kube-system/kube-apiserver-ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.947617 kubelet[2191]: I0412 18:30:50.947415 2191 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/16f742faf238cd425321419501894d1a-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510.3.3-a-f11561af44\" (UID: \"16f742faf238cd425321419501894d1a\") " pod="kube-system/kube-apiserver-ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.947617 kubelet[2191]: I0412 18:30:50.947434 2191 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-ca-certs\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.993379 kubelet[2191]: I0412 18:30:50.993350 2191 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:50.993707 kubelet[2191]: E0412 18:30:50.993688 2191 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.200.20.18:6443/api/v1/nodes\": dial tcp 10.200.20.18:6443: connect: connection refused" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:51.155974 env[1410]: time="2024-04-12T18:30:51.155638538Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510.3.3-a-f11561af44,Uid:16f742faf238cd425321419501894d1a,Namespace:kube-system,Attempt:0,}" Apr 12 18:30:51.157952 env[1410]: time="2024-04-12T18:30:51.157914224Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510.3.3-a-f11561af44,Uid:88be763039c4881497388a451291ec85,Namespace:kube-system,Attempt:0,}" Apr 12 18:30:51.160906 env[1410]: time="2024-04-12T18:30:51.160865632Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510.3.3-a-f11561af44,Uid:faca4c8be60ec53ec91fce85ebbca199,Namespace:kube-system,Attempt:0,}" Apr 12 18:30:51.249964 kubelet[2191]: E0412 18:30:51.249847 2191 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.3-a-f11561af44?timeout=10s\": dial tcp 10.200.20.18:6443: connect: connection refused" interval="800ms" Apr 12 18:30:51.395669 kubelet[2191]: I0412 18:30:51.395641 2191 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:51.396034 kubelet[2191]: E0412 18:30:51.396011 2191 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.200.20.18:6443/api/v1/nodes\": dial tcp 10.200.20.18:6443: connect: connection refused" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:51.676285 kubelet[2191]: W0412 18:30:51.676185 2191 reflector.go:533] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Service: Get "https://10.200.20.18:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:51.676285 kubelet[2191]: E0412 18:30:51.676266 2191 reflector.go:148] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Service: failed to list *v1.Service: Get "https://10.200.20.18:6443/api/v1/services?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:51.753933 kubelet[2191]: W0412 18:30:51.753877 2191 reflector.go:533] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.Node: Get "https://10.200.20.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.3-a-f11561af44&limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:51.753933 kubelet[2191]: E0412 18:30:51.753936 2191 reflector.go:148] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.Node: failed to list *v1.Node: Get "https://10.200.20.18:6443/api/v1/nodes?fieldSelector=metadata.name%3Dci-3510.3.3-a-f11561af44&limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:51.905681 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4064473821.mount: Deactivated successfully. Apr 12 18:30:51.938279 env[1410]: time="2024-04-12T18:30:51.938147713Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:51.967306 env[1410]: time="2024-04-12T18:30:51.967261231Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:51.974646 env[1410]: time="2024-04-12T18:30:51.974607131Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:51.977279 env[1410]: time="2024-04-12T18:30:51.977244498Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:51.983953 env[1410]: time="2024-04-12T18:30:51.983911916Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:51.988152 env[1410]: time="2024-04-12T18:30:51.988099847Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:51.991449 env[1410]: time="2024-04-12T18:30:51.991411936Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:51.997287 env[1410]: time="2024-04-12T18:30:51.997236671Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:52.004661 env[1410]: time="2024-04-12T18:30:52.004614811Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:52.005613 kubelet[2191]: W0412 18:30:52.005565 2191 reflector.go:533] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.CSIDriver: Get "https://10.200.20.18:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:52.005673 kubelet[2191]: E0412 18:30:52.005625 2191 reflector.go:148] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get "https://10.200.20.18:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:52.018156 env[1410]: time="2024-04-12T18:30:52.018103046Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:7d46a07936af93fcce097459055f93ab07331509aa55f4a2a90d95a3ace1850e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:52.036429 env[1410]: time="2024-04-12T18:30:52.036379894Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:52.051170 kubelet[2191]: E0412 18:30:52.051139 2191 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.3-a-f11561af44?timeout=10s\": dial tcp 10.200.20.18:6443: connect: connection refused" interval="1.6s" Apr 12 18:30:52.062874 env[1410]: time="2024-04-12T18:30:52.062826763Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:30:52.099706 kubelet[2191]: W0412 18:30:52.099669 2191 reflector.go:533] vendor/k8s.io/client-go/informers/factory.go:150: failed to list *v1.RuntimeClass: Get "https://10.200.20.18:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:52.099706 kubelet[2191]: E0412 18:30:52.099712 2191 reflector.go:148] vendor/k8s.io/client-go/informers/factory.go:150: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get "https://10.200.20.18:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 10.200.20.18:6443: connect: connection refused Apr 12 18:30:52.124151 env[1410]: time="2024-04-12T18:30:52.123810883Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:30:52.124329 env[1410]: time="2024-04-12T18:30:52.124179164Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:30:52.124329 env[1410]: time="2024-04-12T18:30:52.124229044Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:30:52.124522 env[1410]: time="2024-04-12T18:30:52.124485645Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/0c9a55a05776474a1979534b2eda99f216919be5073234f6f6727482398c3e0c pid=2230 runtime=io.containerd.runc.v2 Apr 12 18:30:52.154348 env[1410]: time="2024-04-12T18:30:52.151981117Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:30:52.154348 env[1410]: time="2024-04-12T18:30:52.152122757Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:30:52.154348 env[1410]: time="2024-04-12T18:30:52.152151637Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:30:52.154348 env[1410]: time="2024-04-12T18:30:52.152429718Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/fe7f281ed42e6ac86c3bdb0cd9506d9282ee87dbb628df25464b3838e2dbea94 pid=2258 runtime=io.containerd.runc.v2 Apr 12 18:30:52.188583 env[1410]: time="2024-04-12T18:30:52.186829888Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:30:52.188583 env[1410]: time="2024-04-12T18:30:52.186934088Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:30:52.188583 env[1410]: time="2024-04-12T18:30:52.186959968Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:30:52.188583 env[1410]: time="2024-04-12T18:30:52.187121769Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/ed60f317bfea09a2f8e26cb6ceb846a2afaee589ed3fbf72707a1a2b67fc38cf pid=2293 runtime=io.containerd.runc.v2 Apr 12 18:30:52.199377 kubelet[2191]: I0412 18:30:52.197679 2191 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:52.199680 kubelet[2191]: E0412 18:30:52.199647 2191 kubelet_node_status.go:92] "Unable to register node with API server" err="Post \"https://10.200.20.18:6443/api/v1/nodes\": dial tcp 10.200.20.18:6443: connect: connection refused" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:52.199767 env[1410]: time="2024-04-12T18:30:52.199721042Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-apiserver-ci-3510.3.3-a-f11561af44,Uid:16f742faf238cd425321419501894d1a,Namespace:kube-system,Attempt:0,} returns sandbox id \"0c9a55a05776474a1979534b2eda99f216919be5073234f6f6727482398c3e0c\"" Apr 12 18:30:52.203684 env[1410]: time="2024-04-12T18:30:52.203631812Z" level=info msg="CreateContainer within sandbox \"0c9a55a05776474a1979534b2eda99f216919be5073234f6f6727482398c3e0c\" for container &ContainerMetadata{Name:kube-apiserver,Attempt:0,}" Apr 12 18:30:52.209014 env[1410]: time="2024-04-12T18:30:52.208929066Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-controller-manager-ci-3510.3.3-a-f11561af44,Uid:88be763039c4881497388a451291ec85,Namespace:kube-system,Attempt:0,} returns sandbox id \"fe7f281ed42e6ac86c3bdb0cd9506d9282ee87dbb628df25464b3838e2dbea94\"" Apr 12 18:30:52.212531 env[1410]: time="2024-04-12T18:30:52.212490595Z" level=info msg="CreateContainer within sandbox \"fe7f281ed42e6ac86c3bdb0cd9506d9282ee87dbb628df25464b3838e2dbea94\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:0,}" Apr 12 18:30:52.248288 env[1410]: time="2024-04-12T18:30:52.248190889Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-scheduler-ci-3510.3.3-a-f11561af44,Uid:faca4c8be60ec53ec91fce85ebbca199,Namespace:kube-system,Attempt:0,} returns sandbox id \"ed60f317bfea09a2f8e26cb6ceb846a2afaee589ed3fbf72707a1a2b67fc38cf\"" Apr 12 18:30:52.252442 env[1410]: time="2024-04-12T18:30:52.252402100Z" level=info msg="CreateContainer within sandbox \"ed60f317bfea09a2f8e26cb6ceb846a2afaee589ed3fbf72707a1a2b67fc38cf\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:0,}" Apr 12 18:30:52.273574 env[1410]: time="2024-04-12T18:30:52.273525555Z" level=info msg="CreateContainer within sandbox \"0c9a55a05776474a1979534b2eda99f216919be5073234f6f6727482398c3e0c\" for &ContainerMetadata{Name:kube-apiserver,Attempt:0,} returns container id \"5437912e5729c3d374d43019f4b34c770951422c1107efed20926d851bb62dfa\"" Apr 12 18:30:52.277992 env[1410]: time="2024-04-12T18:30:52.277948527Z" level=info msg="StartContainer for \"5437912e5729c3d374d43019f4b34c770951422c1107efed20926d851bb62dfa\"" Apr 12 18:30:52.321129 env[1410]: time="2024-04-12T18:30:52.321079240Z" level=info msg="CreateContainer within sandbox \"fe7f281ed42e6ac86c3bdb0cd9506d9282ee87dbb628df25464b3838e2dbea94\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:0,} returns container id \"2e556d2e2f2dbeeff4442b1efec977fd6c1284607bfcef3452c8100d034895b4\"" Apr 12 18:30:52.322138 env[1410]: time="2024-04-12T18:30:52.322102083Z" level=info msg="StartContainer for \"2e556d2e2f2dbeeff4442b1efec977fd6c1284607bfcef3452c8100d034895b4\"" Apr 12 18:30:52.356241 env[1410]: time="2024-04-12T18:30:52.354895008Z" level=info msg="StartContainer for \"5437912e5729c3d374d43019f4b34c770951422c1107efed20926d851bb62dfa\" returns successfully" Apr 12 18:30:52.356241 env[1410]: time="2024-04-12T18:30:52.355665010Z" level=info msg="CreateContainer within sandbox \"ed60f317bfea09a2f8e26cb6ceb846a2afaee589ed3fbf72707a1a2b67fc38cf\" for &ContainerMetadata{Name:kube-scheduler,Attempt:0,} returns container id \"08e50517fecdbc04b574da35d92f251117b2edf330df19a30e54c0ebc2967f0a\"" Apr 12 18:30:52.360221 env[1410]: time="2024-04-12T18:30:52.360145942Z" level=info msg="StartContainer for \"08e50517fecdbc04b574da35d92f251117b2edf330df19a30e54c0ebc2967f0a\"" Apr 12 18:30:52.407420 env[1410]: time="2024-04-12T18:30:52.407369226Z" level=info msg="StartContainer for \"2e556d2e2f2dbeeff4442b1efec977fd6c1284607bfcef3452c8100d034895b4\" returns successfully" Apr 12 18:30:52.557621 env[1410]: time="2024-04-12T18:30:52.557512939Z" level=info msg="StartContainer for \"08e50517fecdbc04b574da35d92f251117b2edf330df19a30e54c0ebc2967f0a\" returns successfully" Apr 12 18:30:53.801817 kubelet[2191]: I0412 18:30:53.801781 2191 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:54.757660 kubelet[2191]: E0412 18:30:54.757606 2191 nodelease.go:49] "Failed to get node when trying to set owner ref to the node lease" err="nodes \"ci-3510.3.3-a-f11561af44\" not found" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:54.787694 kubelet[2191]: I0412 18:30:54.787644 2191 kubelet_node_status.go:73] "Successfully registered node" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:54.927347 kubelet[2191]: E0412 18:30:54.927311 2191 kubelet.go:1856] "Failed creating a mirror pod for" err="pods \"kube-scheduler-ci-3510.3.3-a-f11561af44\" is forbidden: no PriorityClass with name system-node-critical was found" pod="kube-system/kube-scheduler-ci-3510.3.3-a-f11561af44" Apr 12 18:30:55.639038 kubelet[2191]: I0412 18:30:55.638974 2191 apiserver.go:52] "Watching apiserver" Apr 12 18:30:55.646442 kubelet[2191]: I0412 18:30:55.646397 2191 desired_state_of_world_populator.go:153] "Finished populating initial desired state of world" Apr 12 18:30:55.677709 kubelet[2191]: I0412 18:30:55.677663 2191 reconciler.go:41] "Reconciler: start to sync state" Apr 12 18:30:57.344014 systemd[1]: Reloading. Apr 12 18:30:57.400147 /usr/lib/systemd/system-generators/torcx-generator[2478]: time="2024-04-12T18:30:57Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.3 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.3 /var/lib/torcx/store]" Apr 12 18:30:57.400185 /usr/lib/systemd/system-generators/torcx-generator[2478]: time="2024-04-12T18:30:57Z" level=info msg="torcx already run" Apr 12 18:30:57.435958 kubelet[2191]: W0412 18:30:57.435917 2191 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 12 18:30:57.500718 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Apr 12 18:30:57.500739 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Apr 12 18:30:57.519570 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Apr 12 18:30:57.561969 kubelet[2191]: W0412 18:30:57.561939 2191 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 12 18:30:57.667467 systemd[1]: Stopping kubelet.service... Apr 12 18:30:57.680693 systemd[1]: kubelet.service: Deactivated successfully. Apr 12 18:30:57.681055 systemd[1]: Stopped kubelet.service. Apr 12 18:30:57.683169 systemd[1]: Started kubelet.service. Apr 12 18:30:57.787321 kubelet[2544]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 12 18:30:57.787321 kubelet[2544]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Apr 12 18:30:57.787321 kubelet[2544]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Apr 12 18:30:57.787706 kubelet[2544]: I0412 18:30:57.787384 2544 server.go:199] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Apr 12 18:30:57.791932 kubelet[2544]: I0412 18:30:57.791892 2544 server.go:415] "Kubelet version" kubeletVersion="v1.27.2" Apr 12 18:30:57.791932 kubelet[2544]: I0412 18:30:57.791923 2544 server.go:417] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Apr 12 18:30:57.792155 kubelet[2544]: I0412 18:30:57.792135 2544 server.go:837] "Client rotation is on, will bootstrap in background" Apr 12 18:30:57.793793 kubelet[2544]: I0412 18:30:57.793770 2544 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Apr 12 18:30:57.794762 kubelet[2544]: I0412 18:30:57.794743 2544 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Apr 12 18:30:57.797094 kubelet[2544]: W0412 18:30:57.797074 2544 machine.go:65] Cannot read vendor id correctly, set empty. Apr 12 18:30:57.797808 kubelet[2544]: I0412 18:30:57.797789 2544 server.go:662] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Apr 12 18:30:57.798326 kubelet[2544]: I0412 18:30:57.798307 2544 container_manager_linux.go:266] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Apr 12 18:30:57.798402 kubelet[2544]: I0412 18:30:57.798387 2544 container_manager_linux.go:271] "Creating Container Manager object based on Node Config" nodeConfig={RuntimeCgroupsName: SystemCgroupsName: KubeletCgroupsName: KubeletOOMScoreAdj:-999 ContainerRuntime: CgroupsPerQOS:true CgroupRoot:/ CgroupDriver:cgroupfs KubeletRootDir:/var/lib/kubelet ProtectKernelDefaults:false NodeAllocatableConfig:{KubeReservedCgroupName: SystemReservedCgroupName: ReservedSystemCPUs: EnforceNodeAllocatable:map[pods:{}] KubeReserved:map[] SystemReserved:map[] HardEvictionThresholds:[{Signal:nodefs.inodesFree Operator:LessThan Value:{Quantity: Percentage:0.05} GracePeriod:0s MinReclaim:} {Signal:imagefs.available Operator:LessThan Value:{Quantity: Percentage:0.15} GracePeriod:0s MinReclaim:} {Signal:memory.available Operator:LessThan Value:{Quantity:100Mi Percentage:0} GracePeriod:0s MinReclaim:} {Signal:nodefs.available Operator:LessThan Value:{Quantity: Percentage:0.1} GracePeriod:0s MinReclaim:}]} QOSReserved:map[] CPUManagerPolicy:none CPUManagerPolicyOptions:map[] TopologyManagerScope:container CPUManagerReconcilePeriod:10s ExperimentalMemoryManagerPolicy:None ExperimentalMemoryManagerReservedMemory:[] PodPidsLimit:-1 EnforceCPULimits:true CPUCFSQuotaPeriod:100ms TopologyManagerPolicy:none ExperimentalTopologyManagerPolicyOptions:map[]} Apr 12 18:30:57.798487 kubelet[2544]: I0412 18:30:57.798409 2544 topology_manager.go:136] "Creating topology manager with policy per scope" topologyPolicyName="none" topologyScopeName="container" Apr 12 18:30:57.798487 kubelet[2544]: I0412 18:30:57.798420 2544 container_manager_linux.go:302] "Creating device plugin manager" Apr 12 18:30:57.798487 kubelet[2544]: I0412 18:30:57.798452 2544 state_mem.go:36] "Initialized new in-memory state store" Apr 12 18:30:57.803491 kubelet[2544]: I0412 18:30:57.803467 2544 kubelet.go:405] "Attempting to sync node with API server" Apr 12 18:30:57.803663 kubelet[2544]: I0412 18:30:57.803651 2544 kubelet.go:298] "Adding static pod path" path="/etc/kubernetes/manifests" Apr 12 18:30:57.803766 kubelet[2544]: I0412 18:30:57.803755 2544 kubelet.go:309] "Adding apiserver pod source" Apr 12 18:30:57.803836 kubelet[2544]: I0412 18:30:57.803826 2544 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Apr 12 18:30:57.804858 kubelet[2544]: I0412 18:30:57.804831 2544 kuberuntime_manager.go:257] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Apr 12 18:30:57.805450 kubelet[2544]: I0412 18:30:57.805425 2544 server.go:1168] "Started kubelet" Apr 12 18:30:57.807159 kubelet[2544]: I0412 18:30:57.807111 2544 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Apr 12 18:30:57.807762 kubelet[2544]: I0412 18:30:57.807728 2544 server.go:461] "Adding debug handlers to kubelet server" Apr 12 18:30:57.808836 kubelet[2544]: I0412 18:30:57.808793 2544 ratelimit.go:65] "Setting rate limiting for podresources endpoint" qps=100 burstTokens=10 Apr 12 18:30:57.809968 kubelet[2544]: I0412 18:30:57.809922 2544 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Apr 12 18:30:57.811689 kubelet[2544]: E0412 18:30:57.811670 2544 cri_stats_provider.go:455] "Failed to get the info of the filesystem with mountpoint" err="unable to find data in memory cache" mountpoint="/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs" Apr 12 18:30:57.811819 kubelet[2544]: E0412 18:30:57.811809 2544 kubelet.go:1400] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Apr 12 18:30:57.825009 kubelet[2544]: I0412 18:30:57.824971 2544 volume_manager.go:284] "Starting Kubelet Volume Manager" Apr 12 18:30:57.828854 kubelet[2544]: I0412 18:30:57.827052 2544 desired_state_of_world_populator.go:145] "Desired state populator starts to run" Apr 12 18:30:57.829213 kubelet[2544]: I0412 18:30:57.829175 2544 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv4 Apr 12 18:30:57.831442 kubelet[2544]: I0412 18:30:57.831414 2544 kubelet_network_linux.go:63] "Initialized iptables rules." protocol=IPv6 Apr 12 18:30:57.831542 kubelet[2544]: I0412 18:30:57.831454 2544 status_manager.go:207] "Starting to sync pod status with apiserver" Apr 12 18:30:57.831542 kubelet[2544]: I0412 18:30:57.831473 2544 kubelet.go:2257] "Starting kubelet main sync loop" Apr 12 18:30:57.831608 kubelet[2544]: E0412 18:30:57.831582 2544 kubelet.go:2281] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Apr 12 18:30:57.869897 sudo[2570]: root : PWD=/home/core ; USER=root ; COMMAND=/usr/bin/tar -xf /opt/bin/cilium.tar.gz -C /opt/bin Apr 12 18:30:57.870099 sudo[2570]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=0) Apr 12 18:30:57.935218 kubelet[2544]: I0412 18:30:57.935104 2544 kubelet_node_status.go:70] "Attempting to register node" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:57.939474 kubelet[2544]: E0412 18:30:57.936680 2544 kubelet.go:2281] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Apr 12 18:30:57.950769 kubelet[2544]: I0412 18:30:57.950726 2544 kubelet_node_status.go:108] "Node was previously registered" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:57.950887 kubelet[2544]: I0412 18:30:57.950826 2544 kubelet_node_status.go:73] "Successfully registered node" node="ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.004013 kubelet[2544]: I0412 18:30:58.003990 2544 cpu_manager.go:214] "Starting CPU manager" policy="none" Apr 12 18:30:58.004171 kubelet[2544]: I0412 18:30:58.004161 2544 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Apr 12 18:30:58.004338 kubelet[2544]: I0412 18:30:58.004326 2544 state_mem.go:36] "Initialized new in-memory state store" Apr 12 18:30:58.004561 kubelet[2544]: I0412 18:30:58.004549 2544 state_mem.go:88] "Updated default CPUSet" cpuSet="" Apr 12 18:30:58.004641 kubelet[2544]: I0412 18:30:58.004632 2544 state_mem.go:96] "Updated CPUSet assignments" assignments=map[] Apr 12 18:30:58.004697 kubelet[2544]: I0412 18:30:58.004688 2544 policy_none.go:49] "None policy: Start" Apr 12 18:30:58.005582 kubelet[2544]: I0412 18:30:58.005566 2544 memory_manager.go:169] "Starting memorymanager" policy="None" Apr 12 18:30:58.005703 kubelet[2544]: I0412 18:30:58.005694 2544 state_mem.go:35] "Initializing new in-memory state store" Apr 12 18:30:58.005979 kubelet[2544]: I0412 18:30:58.005967 2544 state_mem.go:75] "Updated machine memory state" Apr 12 18:30:58.007337 kubelet[2544]: I0412 18:30:58.007319 2544 manager.go:455] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Apr 12 18:30:58.008999 kubelet[2544]: I0412 18:30:58.008980 2544 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Apr 12 18:30:58.137011 kubelet[2544]: I0412 18:30:58.136969 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:30:58.137148 kubelet[2544]: I0412 18:30:58.137072 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:30:58.137148 kubelet[2544]: I0412 18:30:58.137114 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:30:58.142377 kubelet[2544]: W0412 18:30:58.142345 2544 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 12 18:30:58.166088 kubelet[2544]: W0412 18:30:58.165616 2544 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 12 18:30:58.166294 kubelet[2544]: W0412 18:30:58.166145 2544 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 12 18:30:58.166294 kubelet[2544]: E0412 18:30:58.166219 2544 kubelet.go:1856] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" already exists" pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.166440 kubelet[2544]: E0412 18:30:58.166411 2544 kubelet.go:1856] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-3510.3.3-a-f11561af44\" already exists" pod="kube-system/kube-apiserver-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.239317 kubelet[2544]: I0412 18:30:58.239191 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-k8s-certs\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.239911 kubelet[2544]: I0412 18:30:58.239890 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-kubeconfig\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.240148 kubelet[2544]: I0412 18:30:58.240124 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-usr-share-ca-certificates\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.240289 kubelet[2544]: I0412 18:30:58.240279 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/host-path/faca4c8be60ec53ec91fce85ebbca199-kubeconfig\") pod \"kube-scheduler-ci-3510.3.3-a-f11561af44\" (UID: \"faca4c8be60ec53ec91fce85ebbca199\") " pod="kube-system/kube-scheduler-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.240438 kubelet[2544]: I0412 18:30:58.240419 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-share-ca-certificates\" (UniqueName: \"kubernetes.io/host-path/16f742faf238cd425321419501894d1a-usr-share-ca-certificates\") pod \"kube-apiserver-ci-3510.3.3-a-f11561af44\" (UID: \"16f742faf238cd425321419501894d1a\") " pod="kube-system/kube-apiserver-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.240549 kubelet[2544]: I0412 18:30:58.240534 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-ca-certs\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.240667 kubelet[2544]: I0412 18:30:58.240659 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"flexvolume-dir\" (UniqueName: \"kubernetes.io/host-path/88be763039c4881497388a451291ec85-flexvolume-dir\") pod \"kube-controller-manager-ci-3510.3.3-a-f11561af44\" (UID: \"88be763039c4881497388a451291ec85\") " pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.240780 kubelet[2544]: I0412 18:30:58.240771 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/host-path/16f742faf238cd425321419501894d1a-ca-certs\") pod \"kube-apiserver-ci-3510.3.3-a-f11561af44\" (UID: \"16f742faf238cd425321419501894d1a\") " pod="kube-system/kube-apiserver-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.240884 kubelet[2544]: I0412 18:30:58.240875 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"k8s-certs\" (UniqueName: \"kubernetes.io/host-path/16f742faf238cd425321419501894d1a-k8s-certs\") pod \"kube-apiserver-ci-3510.3.3-a-f11561af44\" (UID: \"16f742faf238cd425321419501894d1a\") " pod="kube-system/kube-apiserver-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.412876 sudo[2570]: pam_unix(sudo:session): session closed for user root Apr 12 18:30:58.806719 kubelet[2544]: I0412 18:30:58.806678 2544 apiserver.go:52] "Watching apiserver" Apr 12 18:30:58.827748 kubelet[2544]: I0412 18:30:58.827708 2544 desired_state_of_world_populator.go:153] "Finished populating initial desired state of world" Apr 12 18:30:58.842376 kubelet[2544]: I0412 18:30:58.842344 2544 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" podStartSLOduration=1.8422868270000001 podCreationTimestamp="2024-04-12 18:30:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:30:58.833410326 +0000 UTC m=+1.147119107" watchObservedRunningTime="2024-04-12 18:30:58.842286827 +0000 UTC m=+1.155995608" Apr 12 18:30:58.842520 kubelet[2544]: I0412 18:30:58.842437 2544 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-scheduler-ci-3510.3.3-a-f11561af44" podStartSLOduration=0.842422987 podCreationTimestamp="2024-04-12 18:30:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:30:58.841635625 +0000 UTC m=+1.155344366" watchObservedRunningTime="2024-04-12 18:30:58.842422987 +0000 UTC m=+1.156131768" Apr 12 18:30:58.844948 kubelet[2544]: I0412 18:30:58.844910 2544 reconciler.go:41] "Reconciler: start to sync state" Apr 12 18:30:58.934618 kubelet[2544]: W0412 18:30:58.934576 2544 warnings.go:70] metadata.name: this is used in the Pod's hostname, which can result in surprising behavior; a DNS label is recommended: [must not contain dots] Apr 12 18:30:58.934777 kubelet[2544]: E0412 18:30:58.934658 2544 kubelet.go:1856] "Failed creating a mirror pod for" err="pods \"kube-apiserver-ci-3510.3.3-a-f11561af44\" already exists" pod="kube-system/kube-apiserver-ci-3510.3.3-a-f11561af44" Apr 12 18:30:58.940236 kubelet[2544]: I0412 18:30:58.939838 2544 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-apiserver-ci-3510.3.3-a-f11561af44" podStartSLOduration=1.939802052 podCreationTimestamp="2024-04-12 18:30:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:30:58.851397688 +0000 UTC m=+1.165106469" watchObservedRunningTime="2024-04-12 18:30:58.939802052 +0000 UTC m=+1.253510833" Apr 12 18:31:00.985942 sudo[1770]: pam_unix(sudo:session): session closed for user root Apr 12 18:31:01.054418 sshd[1766]: pam_unix(sshd:session): session closed for user core Apr 12 18:31:01.057415 systemd-logind[1396]: Session 7 logged out. Waiting for processes to exit. Apr 12 18:31:01.057728 systemd[1]: sshd@4-10.200.20.18:22-10.200.12.6:36744.service: Deactivated successfully. Apr 12 18:31:01.058892 systemd[1]: session-7.scope: Deactivated successfully. Apr 12 18:31:01.059703 systemd-logind[1396]: Removed session 7. Apr 12 18:31:10.427065 kubelet[2544]: I0412 18:31:10.427040 2544 kuberuntime_manager.go:1460] "Updating runtime config through cri with podcidr" CIDR="192.168.0.0/24" Apr 12 18:31:10.427998 env[1410]: time="2024-04-12T18:31:10.427902357Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Apr 12 18:31:10.428319 kubelet[2544]: I0412 18:31:10.428114 2544 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.0.0/24" Apr 12 18:31:11.312195 kubelet[2544]: I0412 18:31:11.312139 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:31:11.317080 kubelet[2544]: I0412 18:31:11.317046 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:31:11.414591 kubelet[2544]: I0412 18:31:11.414555 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cilium-run\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.414814 kubelet[2544]: I0412 18:31:11.414803 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cni-path\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.414892 kubelet[2544]: I0412 18:31:11.414881 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-host-proc-sys-net\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.414991 kubelet[2544]: I0412 18:31:11.414978 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d7962196-9459-4162-a9ec-322a70f6491c-hubble-tls\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.415146 kubelet[2544]: I0412 18:31:11.415133 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-etc-cni-netd\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.415253 kubelet[2544]: I0412 18:31:11.415242 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-lib-modules\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.415390 kubelet[2544]: I0412 18:31:11.415376 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-xtables-lock\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.415495 kubelet[2544]: I0412 18:31:11.415481 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d7962196-9459-4162-a9ec-322a70f6491c-clustermesh-secrets\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.415598 kubelet[2544]: I0412 18:31:11.415585 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d7962196-9459-4162-a9ec-322a70f6491c-cilium-config-path\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.415690 kubelet[2544]: I0412 18:31:11.415681 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/07575488-a6d1-46ff-9339-ed562fe010b5-kube-proxy\") pod \"kube-proxy-bkxbp\" (UID: \"07575488-a6d1-46ff-9339-ed562fe010b5\") " pod="kube-system/kube-proxy-bkxbp" Apr 12 18:31:11.415779 kubelet[2544]: I0412 18:31:11.415769 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-bpf-maps\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.415855 kubelet[2544]: I0412 18:31:11.415846 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-hostproc\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.415935 kubelet[2544]: I0412 18:31:11.415926 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/07575488-a6d1-46ff-9339-ed562fe010b5-xtables-lock\") pod \"kube-proxy-bkxbp\" (UID: \"07575488-a6d1-46ff-9339-ed562fe010b5\") " pod="kube-system/kube-proxy-bkxbp" Apr 12 18:31:11.416023 kubelet[2544]: I0412 18:31:11.416013 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/07575488-a6d1-46ff-9339-ed562fe010b5-lib-modules\") pod \"kube-proxy-bkxbp\" (UID: \"07575488-a6d1-46ff-9339-ed562fe010b5\") " pod="kube-system/kube-proxy-bkxbp" Apr 12 18:31:11.416121 kubelet[2544]: I0412 18:31:11.416109 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv4pb\" (UniqueName: \"kubernetes.io/projected/d7962196-9459-4162-a9ec-322a70f6491c-kube-api-access-kv4pb\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.416254 kubelet[2544]: I0412 18:31:11.416243 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttkkh\" (UniqueName: \"kubernetes.io/projected/07575488-a6d1-46ff-9339-ed562fe010b5-kube-api-access-ttkkh\") pod \"kube-proxy-bkxbp\" (UID: \"07575488-a6d1-46ff-9339-ed562fe010b5\") " pod="kube-system/kube-proxy-bkxbp" Apr 12 18:31:11.416364 kubelet[2544]: I0412 18:31:11.416347 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cilium-cgroup\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.416468 kubelet[2544]: I0412 18:31:11.416457 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-host-proc-sys-kernel\") pod \"cilium-5mw89\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " pod="kube-system/cilium-5mw89" Apr 12 18:31:11.479566 kubelet[2544]: I0412 18:31:11.479526 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:31:11.517187 kubelet[2544]: I0412 18:31:11.517155 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdwwh\" (UniqueName: \"kubernetes.io/projected/03d8e334-f4e0-4ae8-85df-87f70aa7db1f-kube-api-access-jdwwh\") pod \"cilium-operator-574c4bb98d-g6v6h\" (UID: \"03d8e334-f4e0-4ae8-85df-87f70aa7db1f\") " pod="kube-system/cilium-operator-574c4bb98d-g6v6h" Apr 12 18:31:11.517592 kubelet[2544]: I0412 18:31:11.517575 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/03d8e334-f4e0-4ae8-85df-87f70aa7db1f-cilium-config-path\") pod \"cilium-operator-574c4bb98d-g6v6h\" (UID: \"03d8e334-f4e0-4ae8-85df-87f70aa7db1f\") " pod="kube-system/cilium-operator-574c4bb98d-g6v6h" Apr 12 18:31:11.617124 env[1410]: time="2024-04-12T18:31:11.616348853Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bkxbp,Uid:07575488-a6d1-46ff-9339-ed562fe010b5,Namespace:kube-system,Attempt:0,}" Apr 12 18:31:11.621004 env[1410]: time="2024-04-12T18:31:11.620760021Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-5mw89,Uid:d7962196-9459-4162-a9ec-322a70f6491c,Namespace:kube-system,Attempt:0,}" Apr 12 18:31:11.693980 env[1410]: time="2024-04-12T18:31:11.693869793Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:31:11.693980 env[1410]: time="2024-04-12T18:31:11.693914913Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:31:11.693980 env[1410]: time="2024-04-12T18:31:11.693925513Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:31:11.694182 env[1410]: time="2024-04-12T18:31:11.686928061Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:31:11.694182 env[1410]: time="2024-04-12T18:31:11.686968381Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:31:11.694182 env[1410]: time="2024-04-12T18:31:11.686979141Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:31:11.694182 env[1410]: time="2024-04-12T18:31:11.687126061Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/3c3f3b2c96b2607e9f1b2b33a9300ce65b2b289c66530f534daf1b0564eaa25c pid=2632 runtime=io.containerd.runc.v2 Apr 12 18:31:11.695853 env[1410]: time="2024-04-12T18:31:11.695793477Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467 pid=2649 runtime=io.containerd.runc.v2 Apr 12 18:31:11.750075 env[1410]: time="2024-04-12T18:31:11.750034135Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-5mw89,Uid:d7962196-9459-4162-a9ec-322a70f6491c,Namespace:kube-system,Attempt:0,} returns sandbox id \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\"" Apr 12 18:31:11.753875 env[1410]: time="2024-04-12T18:31:11.753487621Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Apr 12 18:31:11.757684 env[1410]: time="2024-04-12T18:31:11.757646429Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-bkxbp,Uid:07575488-a6d1-46ff-9339-ed562fe010b5,Namespace:kube-system,Attempt:0,} returns sandbox id \"3c3f3b2c96b2607e9f1b2b33a9300ce65b2b289c66530f534daf1b0564eaa25c\"" Apr 12 18:31:11.761251 env[1410]: time="2024-04-12T18:31:11.761186435Z" level=info msg="CreateContainer within sandbox \"3c3f3b2c96b2607e9f1b2b33a9300ce65b2b289c66530f534daf1b0564eaa25c\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Apr 12 18:31:11.785522 env[1410]: time="2024-04-12T18:31:11.785482999Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-574c4bb98d-g6v6h,Uid:03d8e334-f4e0-4ae8-85df-87f70aa7db1f,Namespace:kube-system,Attempt:0,}" Apr 12 18:31:11.823888 env[1410]: time="2024-04-12T18:31:11.823840149Z" level=info msg="CreateContainer within sandbox \"3c3f3b2c96b2607e9f1b2b33a9300ce65b2b289c66530f534daf1b0564eaa25c\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"9e6c184da8cc7183e23b7874487c29c5d81990b9f20be77a40f78c11f8a9500c\"" Apr 12 18:31:11.826480 env[1410]: time="2024-04-12T18:31:11.826439514Z" level=info msg="StartContainer for \"9e6c184da8cc7183e23b7874487c29c5d81990b9f20be77a40f78c11f8a9500c\"" Apr 12 18:31:11.860751 env[1410]: time="2024-04-12T18:31:11.860672776Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:31:11.860892 env[1410]: time="2024-04-12T18:31:11.860759576Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:31:11.860892 env[1410]: time="2024-04-12T18:31:11.860804496Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:31:11.861025 env[1410]: time="2024-04-12T18:31:11.860983057Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408 pid=2741 runtime=io.containerd.runc.v2 Apr 12 18:31:11.897387 env[1410]: time="2024-04-12T18:31:11.897266082Z" level=info msg="StartContainer for \"9e6c184da8cc7183e23b7874487c29c5d81990b9f20be77a40f78c11f8a9500c\" returns successfully" Apr 12 18:31:11.956021 kubelet[2544]: I0412 18:31:11.955964 2544 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-bkxbp" podStartSLOduration=0.955924069 podCreationTimestamp="2024-04-12 18:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:31:11.955260428 +0000 UTC m=+14.268969209" watchObservedRunningTime="2024-04-12 18:31:11.955924069 +0000 UTC m=+14.269632810" Apr 12 18:31:11.974896 env[1410]: time="2024-04-12T18:31:11.974848663Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-574c4bb98d-g6v6h,Uid:03d8e334-f4e0-4ae8-85df-87f70aa7db1f,Namespace:kube-system,Attempt:0,} returns sandbox id \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\"" Apr 12 18:31:16.848561 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2905746087.mount: Deactivated successfully. Apr 12 18:31:30.550030 env[1410]: time="2024-04-12T18:31:30.549942276Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:31:30.601013 env[1410]: time="2024-04-12T18:31:30.600955586Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:31:30.646457 env[1410]: time="2024-04-12T18:31:30.646413609Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:31:30.646896 env[1410]: time="2024-04-12T18:31:30.646864050Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:b69cb5ebb22d9b4f9c460a6587a0c4285d57a2bff59e4e439ad065a3f684948f\"" Apr 12 18:31:30.649091 env[1410]: time="2024-04-12T18:31:30.649034053Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Apr 12 18:31:30.651743 env[1410]: time="2024-04-12T18:31:30.651706816Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 12 18:31:30.865013 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4011282117.mount: Deactivated successfully. Apr 12 18:31:31.014316 env[1410]: time="2024-04-12T18:31:31.014274594Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73\"" Apr 12 18:31:31.016726 env[1410]: time="2024-04-12T18:31:31.016409517Z" level=info msg="StartContainer for \"dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73\"" Apr 12 18:31:31.111873 env[1410]: time="2024-04-12T18:31:31.111787767Z" level=info msg="StartContainer for \"dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73\" returns successfully" Apr 12 18:31:31.858232 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73-rootfs.mount: Deactivated successfully. Apr 12 18:31:37.779978 env[1410]: time="2024-04-12T18:31:37.779905561Z" level=info msg="shim disconnected" id=dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73 Apr 12 18:31:37.779978 env[1410]: time="2024-04-12T18:31:37.779975521Z" level=warning msg="cleaning up after shim disconnected" id=dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73 namespace=k8s.io Apr 12 18:31:37.779978 env[1410]: time="2024-04-12T18:31:37.779986161Z" level=info msg="cleaning up dead shim" Apr 12 18:31:37.788623 env[1410]: time="2024-04-12T18:31:37.788566292Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:31:37Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2958 runtime=io.containerd.runc.v2\n" Apr 12 18:31:38.004001 env[1410]: time="2024-04-12T18:31:38.003706765Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Apr 12 18:31:38.039683 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount4229221737.mount: Deactivated successfully. Apr 12 18:31:38.045768 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1926218942.mount: Deactivated successfully. Apr 12 18:31:38.072425 env[1410]: time="2024-04-12T18:31:38.072379571Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd\"" Apr 12 18:31:38.073898 env[1410]: time="2024-04-12T18:31:38.073016492Z" level=info msg="StartContainer for \"d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd\"" Apr 12 18:31:38.130248 env[1410]: time="2024-04-12T18:31:38.128555082Z" level=info msg="StartContainer for \"d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd\" returns successfully" Apr 12 18:31:38.134965 systemd[1]: systemd-sysctl.service: Deactivated successfully. Apr 12 18:31:38.135247 systemd[1]: Stopped systemd-sysctl.service. Apr 12 18:31:38.135667 systemd[1]: Stopping systemd-sysctl.service... Apr 12 18:31:38.139410 systemd[1]: Starting systemd-sysctl.service... Apr 12 18:31:38.149521 systemd[1]: Finished systemd-sysctl.service. Apr 12 18:31:38.181344 env[1410]: time="2024-04-12T18:31:38.181290108Z" level=info msg="shim disconnected" id=d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd Apr 12 18:31:38.181667 env[1410]: time="2024-04-12T18:31:38.181647909Z" level=warning msg="cleaning up after shim disconnected" id=d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd namespace=k8s.io Apr 12 18:31:38.181753 env[1410]: time="2024-04-12T18:31:38.181739789Z" level=info msg="cleaning up dead shim" Apr 12 18:31:38.188797 env[1410]: time="2024-04-12T18:31:38.188753478Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:31:38Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3024 runtime=io.containerd.runc.v2\n" Apr 12 18:31:39.006195 env[1410]: time="2024-04-12T18:31:39.005617903Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Apr 12 18:31:39.037155 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd-rootfs.mount: Deactivated successfully. Apr 12 18:31:39.057741 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1043692270.mount: Deactivated successfully. Apr 12 18:31:39.080790 env[1410]: time="2024-04-12T18:31:39.080735956Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"bbcda24e149f4c1a7654a4b90eb62279addf7693ef6b75d9cbe1a1049340e5f9\"" Apr 12 18:31:39.081696 env[1410]: time="2024-04-12T18:31:39.081665757Z" level=info msg="StartContainer for \"bbcda24e149f4c1a7654a4b90eb62279addf7693ef6b75d9cbe1a1049340e5f9\"" Apr 12 18:31:39.133213 env[1410]: time="2024-04-12T18:31:39.133162741Z" level=info msg="StartContainer for \"bbcda24e149f4c1a7654a4b90eb62279addf7693ef6b75d9cbe1a1049340e5f9\" returns successfully" Apr 12 18:31:39.174288 env[1410]: time="2024-04-12T18:31:39.174242752Z" level=info msg="shim disconnected" id=bbcda24e149f4c1a7654a4b90eb62279addf7693ef6b75d9cbe1a1049340e5f9 Apr 12 18:31:39.174522 env[1410]: time="2024-04-12T18:31:39.174504233Z" level=warning msg="cleaning up after shim disconnected" id=bbcda24e149f4c1a7654a4b90eb62279addf7693ef6b75d9cbe1a1049340e5f9 namespace=k8s.io Apr 12 18:31:39.174588 env[1410]: time="2024-04-12T18:31:39.174576033Z" level=info msg="cleaning up dead shim" Apr 12 18:31:39.182806 env[1410]: time="2024-04-12T18:31:39.182758363Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:31:39Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3083 runtime=io.containerd.runc.v2\n" Apr 12 18:31:40.012115 env[1410]: time="2024-04-12T18:31:40.012068993Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Apr 12 18:31:40.056112 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1216366413.mount: Deactivated successfully. Apr 12 18:31:40.083287 env[1410]: time="2024-04-12T18:31:40.083227721Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9\"" Apr 12 18:31:40.087599 env[1410]: time="2024-04-12T18:31:40.086613485Z" level=info msg="StartContainer for \"983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9\"" Apr 12 18:31:40.142226 env[1410]: time="2024-04-12T18:31:40.142139033Z" level=info msg="StartContainer for \"983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9\" returns successfully" Apr 12 18:31:40.187919 env[1410]: time="2024-04-12T18:31:40.187864290Z" level=info msg="shim disconnected" id=983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9 Apr 12 18:31:40.187919 env[1410]: time="2024-04-12T18:31:40.187922290Z" level=warning msg="cleaning up after shim disconnected" id=983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9 namespace=k8s.io Apr 12 18:31:40.188165 env[1410]: time="2024-04-12T18:31:40.187932290Z" level=info msg="cleaning up dead shim" Apr 12 18:31:40.195758 env[1410]: time="2024-04-12T18:31:40.195691659Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:31:40Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3140 runtime=io.containerd.runc.v2\n" Apr 12 18:31:41.019658 env[1410]: time="2024-04-12T18:31:41.019615152Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Apr 12 18:31:41.037059 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9-rootfs.mount: Deactivated successfully. Apr 12 18:31:41.053949 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3963118205.mount: Deactivated successfully. Apr 12 18:31:41.060114 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1000756110.mount: Deactivated successfully. Apr 12 18:31:41.076569 env[1410]: time="2024-04-12T18:31:41.076521742Z" level=info msg="CreateContainer within sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\"" Apr 12 18:31:41.079065 env[1410]: time="2024-04-12T18:31:41.079015825Z" level=info msg="StartContainer for \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\"" Apr 12 18:31:41.142390 env[1410]: time="2024-04-12T18:31:41.142173062Z" level=info msg="StartContainer for \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\" returns successfully" Apr 12 18:31:41.219270 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Apr 12 18:31:41.304305 kubelet[2544]: I0412 18:31:41.304108 2544 kubelet_node_status.go:493] "Fast updating node status as it just became ready" Apr 12 18:31:41.329755 kubelet[2544]: I0412 18:31:41.329708 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:31:41.337258 kubelet[2544]: I0412 18:31:41.337166 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:31:41.391788 kubelet[2544]: I0412 18:31:41.391754 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8da05403-37bd-4f64-8e88-9e6f6cd44b0c-config-volume\") pod \"coredns-5d78c9869d-jhmdp\" (UID: \"8da05403-37bd-4f64-8e88-9e6f6cd44b0c\") " pod="kube-system/coredns-5d78c9869d-jhmdp" Apr 12 18:31:41.392041 kubelet[2544]: I0412 18:31:41.392016 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxcd9\" (UniqueName: \"kubernetes.io/projected/8da05403-37bd-4f64-8e88-9e6f6cd44b0c-kube-api-access-lxcd9\") pod \"coredns-5d78c9869d-jhmdp\" (UID: \"8da05403-37bd-4f64-8e88-9e6f6cd44b0c\") " pod="kube-system/coredns-5d78c9869d-jhmdp" Apr 12 18:31:41.392150 kubelet[2544]: I0412 18:31:41.392139 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h4z8\" (UniqueName: \"kubernetes.io/projected/224ac271-914e-4d7d-a7e7-35eb26c0da5b-kube-api-access-9h4z8\") pod \"coredns-5d78c9869d-z7gcw\" (UID: \"224ac271-914e-4d7d-a7e7-35eb26c0da5b\") " pod="kube-system/coredns-5d78c9869d-z7gcw" Apr 12 18:31:41.392272 kubelet[2544]: I0412 18:31:41.392261 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/224ac271-914e-4d7d-a7e7-35eb26c0da5b-config-volume\") pod \"coredns-5d78c9869d-z7gcw\" (UID: \"224ac271-914e-4d7d-a7e7-35eb26c0da5b\") " pod="kube-system/coredns-5d78c9869d-z7gcw" Apr 12 18:31:41.634216 env[1410]: time="2024-04-12T18:31:41.634075101Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5d78c9869d-z7gcw,Uid:224ac271-914e-4d7d-a7e7-35eb26c0da5b,Namespace:kube-system,Attempt:0,}" Apr 12 18:31:41.640763 env[1410]: time="2024-04-12T18:31:41.640695229Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5d78c9869d-jhmdp,Uid:8da05403-37bd-4f64-8e88-9e6f6cd44b0c,Namespace:kube-system,Attempt:0,}" Apr 12 18:31:41.760226 kernel: WARNING: Unprivileged eBPF is enabled, data leaks possible via Spectre v2 BHB attacks! Apr 12 18:31:42.051489 kubelet[2544]: I0412 18:31:42.051453 2544 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-5mw89" podStartSLOduration=12.154984495 podCreationTimestamp="2024-04-12 18:31:11 +0000 UTC" firstStartedPulling="2024-04-12 18:31:11.751551658 +0000 UTC m=+14.065260439" lastFinishedPulling="2024-04-12 18:31:30.647958531 +0000 UTC m=+32.961667392" observedRunningTime="2024-04-12 18:31:42.051081688 +0000 UTC m=+44.364790469" watchObservedRunningTime="2024-04-12 18:31:42.051391448 +0000 UTC m=+44.365100229" Apr 12 18:31:45.147147 env[1410]: time="2024-04-12T18:31:45.147098464Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:31:45.157261 env[1410]: time="2024-04-12T18:31:45.157208796Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:31:45.162663 env[1410]: time="2024-04-12T18:31:45.162623922Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Apr 12 18:31:45.163289 env[1410]: time="2024-04-12T18:31:45.163252003Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:59357949c22410bca94f8bb5a7a7f73d575949bc16ddc4bd8c740843d4254180\"" Apr 12 18:31:45.167463 env[1410]: time="2024-04-12T18:31:45.167422288Z" level=info msg="CreateContainer within sandbox \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Apr 12 18:31:45.223469 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1109557873.mount: Deactivated successfully. Apr 12 18:31:45.243582 env[1410]: time="2024-04-12T18:31:45.243528177Z" level=info msg="CreateContainer within sandbox \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0\"" Apr 12 18:31:45.244388 env[1410]: time="2024-04-12T18:31:45.244355978Z" level=info msg="StartContainer for \"d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0\"" Apr 12 18:31:45.294192 env[1410]: time="2024-04-12T18:31:45.294132876Z" level=info msg="StartContainer for \"d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0\" returns successfully" Apr 12 18:31:45.346686 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3222385423.mount: Deactivated successfully. Apr 12 18:31:49.416864 systemd-networkd[1583]: cilium_host: Link UP Apr 12 18:31:49.416958 systemd-networkd[1583]: cilium_net: Link UP Apr 12 18:31:49.422680 systemd-networkd[1583]: cilium_net: Gained carrier Apr 12 18:31:49.428287 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Apr 12 18:31:49.428442 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Apr 12 18:31:49.429386 systemd-networkd[1583]: cilium_host: Gained carrier Apr 12 18:31:49.580116 systemd-networkd[1583]: cilium_vxlan: Link UP Apr 12 18:31:49.580122 systemd-networkd[1583]: cilium_vxlan: Gained carrier Apr 12 18:31:49.743355 systemd-networkd[1583]: cilium_host: Gained IPv6LL Apr 12 18:31:49.868247 kernel: NET: Registered PF_ALG protocol family Apr 12 18:31:49.992335 systemd-networkd[1583]: cilium_net: Gained IPv6LL Apr 12 18:31:50.620963 systemd-networkd[1583]: lxc_health: Link UP Apr 12 18:31:50.636278 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Apr 12 18:31:50.636541 systemd-networkd[1583]: lxc_health: Gained carrier Apr 12 18:31:50.748837 systemd-networkd[1583]: lxc8d055e2958ee: Link UP Apr 12 18:31:50.762303 kernel: eth0: renamed from tmpfc68d Apr 12 18:31:50.772300 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc8d055e2958ee: link becomes ready Apr 12 18:31:50.772740 systemd-networkd[1583]: lxc8d055e2958ee: Gained carrier Apr 12 18:31:51.015378 systemd-networkd[1583]: cilium_vxlan: Gained IPv6LL Apr 12 18:31:51.208706 systemd-networkd[1583]: lxcc2a2a2942b45: Link UP Apr 12 18:31:51.219244 kernel: eth0: renamed from tmp1c899 Apr 12 18:31:51.231289 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxcc2a2a2942b45: link becomes ready Apr 12 18:31:51.232875 systemd-networkd[1583]: lxcc2a2a2942b45: Gained carrier Apr 12 18:31:51.662553 kubelet[2544]: I0412 18:31:51.662503 2544 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-574c4bb98d-g6v6h" podStartSLOduration=7.474764463 podCreationTimestamp="2024-04-12 18:31:11 +0000 UTC" firstStartedPulling="2024-04-12 18:31:11.976457546 +0000 UTC m=+14.290166327" lastFinishedPulling="2024-04-12 18:31:45.164151244 +0000 UTC m=+47.477860025" observedRunningTime="2024-04-12 18:31:46.081669039 +0000 UTC m=+48.395377820" watchObservedRunningTime="2024-04-12 18:31:51.662458161 +0000 UTC m=+53.976166942" Apr 12 18:31:51.730375 update_engine[1397]: I0412 18:31:51.730329 1397 prefs.cc:52] certificate-report-to-send-update not present in /var/lib/update_engine/prefs Apr 12 18:31:51.730375 update_engine[1397]: I0412 18:31:51.730370 1397 prefs.cc:52] certificate-report-to-send-download not present in /var/lib/update_engine/prefs Apr 12 18:31:51.730743 update_engine[1397]: I0412 18:31:51.730487 1397 prefs.cc:52] aleph-version not present in /var/lib/update_engine/prefs Apr 12 18:31:51.730861 update_engine[1397]: I0412 18:31:51.730821 1397 omaha_request_params.cc:62] Current group set to lts Apr 12 18:31:51.730955 update_engine[1397]: I0412 18:31:51.730937 1397 update_attempter.cc:499] Already updated boot flags. Skipping. Apr 12 18:31:51.730955 update_engine[1397]: I0412 18:31:51.730947 1397 update_attempter.cc:643] Scheduling an action processor start. Apr 12 18:31:51.731016 update_engine[1397]: I0412 18:31:51.730962 1397 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 12 18:31:51.731016 update_engine[1397]: I0412 18:31:51.730985 1397 prefs.cc:52] previous-version not present in /var/lib/update_engine/prefs Apr 12 18:31:51.731410 locksmithd[1498]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_CHECKING_FOR_UPDATE" NewVersion=0.0.0 NewSize=0 Apr 12 18:31:51.732111 update_engine[1397]: I0412 18:31:51.732079 1397 omaha_request_action.cc:270] Posting an Omaha request to disabled Apr 12 18:31:51.732111 update_engine[1397]: I0412 18:31:51.732103 1397 omaha_request_action.cc:271] Request: Apr 12 18:31:51.732111 update_engine[1397]: Apr 12 18:31:51.732111 update_engine[1397]: Apr 12 18:31:51.732111 update_engine[1397]: Apr 12 18:31:51.732111 update_engine[1397]: Apr 12 18:31:51.732111 update_engine[1397]: Apr 12 18:31:51.732111 update_engine[1397]: Apr 12 18:31:51.732111 update_engine[1397]: Apr 12 18:31:51.732111 update_engine[1397]: Apr 12 18:31:51.732111 update_engine[1397]: I0412 18:31:51.732110 1397 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 12 18:31:51.733000 update_engine[1397]: I0412 18:31:51.732966 1397 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 12 18:31:51.733208 update_engine[1397]: I0412 18:31:51.733173 1397 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 12 18:31:51.837762 update_engine[1397]: E0412 18:31:51.837718 1397 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 12 18:31:51.837902 update_engine[1397]: I0412 18:31:51.837832 1397 libcurl_http_fetcher.cc:283] No HTTP response, retry 1 Apr 12 18:31:52.039422 systemd-networkd[1583]: lxc8d055e2958ee: Gained IPv6LL Apr 12 18:31:52.487356 systemd-networkd[1583]: lxcc2a2a2942b45: Gained IPv6LL Apr 12 18:31:52.615321 systemd-networkd[1583]: lxc_health: Gained IPv6LL Apr 12 18:31:54.909760 env[1410]: time="2024-04-12T18:31:54.907090454Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:31:54.909760 env[1410]: time="2024-04-12T18:31:54.907145654Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:31:54.909760 env[1410]: time="2024-04-12T18:31:54.907156734Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:31:54.909760 env[1410]: time="2024-04-12T18:31:54.907319694Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/fc68de01a90c9b882d144cbdc57876f7cccae0f53e515bc2ae71155e4cb5ddec pid=3729 runtime=io.containerd.runc.v2 Apr 12 18:31:54.956635 env[1410]: time="2024-04-12T18:31:54.956548588Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:31:54.956783 env[1410]: time="2024-04-12T18:31:54.956637348Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:31:54.956783 env[1410]: time="2024-04-12T18:31:54.956662068Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:31:54.957148 env[1410]: time="2024-04-12T18:31:54.957081388Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/1c89985ee6fc8ecd137c478a450643cc5096463e952665cfb27ca8a9220615d9 pid=3763 runtime=io.containerd.runc.v2 Apr 12 18:31:55.030012 env[1410]: time="2024-04-12T18:31:55.029946828Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5d78c9869d-z7gcw,Uid:224ac271-914e-4d7d-a7e7-35eb26c0da5b,Namespace:kube-system,Attempt:0,} returns sandbox id \"1c89985ee6fc8ecd137c478a450643cc5096463e952665cfb27ca8a9220615d9\"" Apr 12 18:31:55.038394 env[1410]: time="2024-04-12T18:31:55.036552515Z" level=info msg="CreateContainer within sandbox \"1c89985ee6fc8ecd137c478a450643cc5096463e952665cfb27ca8a9220615d9\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 12 18:31:55.047882 env[1410]: time="2024-04-12T18:31:55.047829527Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:coredns-5d78c9869d-jhmdp,Uid:8da05403-37bd-4f64-8e88-9e6f6cd44b0c,Namespace:kube-system,Attempt:0,} returns sandbox id \"fc68de01a90c9b882d144cbdc57876f7cccae0f53e515bc2ae71155e4cb5ddec\"" Apr 12 18:31:55.054713 env[1410]: time="2024-04-12T18:31:55.054673094Z" level=info msg="CreateContainer within sandbox \"fc68de01a90c9b882d144cbdc57876f7cccae0f53e515bc2ae71155e4cb5ddec\" for container &ContainerMetadata{Name:coredns,Attempt:0,}" Apr 12 18:31:55.149159 env[1410]: time="2024-04-12T18:31:55.149097237Z" level=info msg="CreateContainer within sandbox \"1c89985ee6fc8ecd137c478a450643cc5096463e952665cfb27ca8a9220615d9\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"996d0dcdcfca462a042a09ede20097d96f821505207708cd902520a9e392279a\"" Apr 12 18:31:55.151002 env[1410]: time="2024-04-12T18:31:55.150963919Z" level=info msg="StartContainer for \"996d0dcdcfca462a042a09ede20097d96f821505207708cd902520a9e392279a\"" Apr 12 18:31:55.177568 env[1410]: time="2024-04-12T18:31:55.177442947Z" level=info msg="CreateContainer within sandbox \"fc68de01a90c9b882d144cbdc57876f7cccae0f53e515bc2ae71155e4cb5ddec\" for &ContainerMetadata{Name:coredns,Attempt:0,} returns container id \"b6858b24224050dbd9bf7005a23dfcaf39cf069a29ac94625c16c2b2e4466574\"" Apr 12 18:31:55.180010 env[1410]: time="2024-04-12T18:31:55.179974310Z" level=info msg="StartContainer for \"b6858b24224050dbd9bf7005a23dfcaf39cf069a29ac94625c16c2b2e4466574\"" Apr 12 18:31:55.229975 env[1410]: time="2024-04-12T18:31:55.229925444Z" level=info msg="StartContainer for \"996d0dcdcfca462a042a09ede20097d96f821505207708cd902520a9e392279a\" returns successfully" Apr 12 18:31:55.266184 env[1410]: time="2024-04-12T18:31:55.266133084Z" level=info msg="StartContainer for \"b6858b24224050dbd9bf7005a23dfcaf39cf069a29ac94625c16c2b2e4466574\" returns successfully" Apr 12 18:31:55.913631 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount371855376.mount: Deactivated successfully. Apr 12 18:31:56.071316 kubelet[2544]: I0412 18:31:56.071274 2544 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5d78c9869d-z7gcw" podStartSLOduration=45.071233515 podCreationTimestamp="2024-04-12 18:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:31:56.070610035 +0000 UTC m=+58.384318816" watchObservedRunningTime="2024-04-12 18:31:56.071233515 +0000 UTC m=+58.384942256" Apr 12 18:31:56.104234 kubelet[2544]: I0412 18:31:56.104173 2544 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/coredns-5d78c9869d-jhmdp" podStartSLOduration=45.104124951 podCreationTimestamp="2024-04-12 18:31:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:31:56.10342215 +0000 UTC m=+58.417130931" watchObservedRunningTime="2024-04-12 18:31:56.104124951 +0000 UTC m=+58.417833732" Apr 12 18:32:01.687809 update_engine[1397]: I0412 18:32:01.687712 1397 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 12 18:32:01.688233 update_engine[1397]: I0412 18:32:01.687911 1397 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 12 18:32:01.688233 update_engine[1397]: I0412 18:32:01.688103 1397 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 12 18:32:01.731994 update_engine[1397]: E0412 18:32:01.731954 1397 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 12 18:32:01.732134 update_engine[1397]: I0412 18:32:01.732062 1397 libcurl_http_fetcher.cc:283] No HTTP response, retry 2 Apr 12 18:32:11.686926 update_engine[1397]: I0412 18:32:11.686878 1397 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 12 18:32:11.687359 update_engine[1397]: I0412 18:32:11.687078 1397 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 12 18:32:11.687359 update_engine[1397]: I0412 18:32:11.687302 1397 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 12 18:32:11.725005 update_engine[1397]: E0412 18:32:11.724961 1397 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 12 18:32:11.725162 update_engine[1397]: I0412 18:32:11.725081 1397 libcurl_http_fetcher.cc:283] No HTTP response, retry 3 Apr 12 18:32:21.689306 update_engine[1397]: I0412 18:32:21.689246 1397 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 12 18:32:21.689748 update_engine[1397]: I0412 18:32:21.689492 1397 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 12 18:32:21.689748 update_engine[1397]: I0412 18:32:21.689726 1397 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 12 18:32:21.699205 update_engine[1397]: E0412 18:32:21.699141 1397 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 12 18:32:21.699356 update_engine[1397]: I0412 18:32:21.699281 1397 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 12 18:32:21.699356 update_engine[1397]: I0412 18:32:21.699290 1397 omaha_request_action.cc:621] Omaha request response: Apr 12 18:32:21.699417 update_engine[1397]: E0412 18:32:21.699398 1397 omaha_request_action.cc:640] Omaha request network transfer failed. Apr 12 18:32:21.699417 update_engine[1397]: I0412 18:32:21.699416 1397 action_processor.cc:68] ActionProcessor::ActionComplete: OmahaRequestAction action failed. Aborting processing. Apr 12 18:32:21.699507 update_engine[1397]: I0412 18:32:21.699419 1397 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 12 18:32:21.699507 update_engine[1397]: I0412 18:32:21.699422 1397 update_attempter.cc:306] Processing Done. Apr 12 18:32:21.699507 update_engine[1397]: E0412 18:32:21.699434 1397 update_attempter.cc:619] Update failed. Apr 12 18:32:21.699507 update_engine[1397]: I0412 18:32:21.699437 1397 utils.cc:600] Converting error code 2000 to kActionCodeOmahaErrorInHTTPResponse Apr 12 18:32:21.699507 update_engine[1397]: I0412 18:32:21.699440 1397 payload_state.cc:97] Updating payload state for error code: 37 (kActionCodeOmahaErrorInHTTPResponse) Apr 12 18:32:21.699507 update_engine[1397]: I0412 18:32:21.699445 1397 payload_state.cc:103] Ignoring failures until we get a valid Omaha response. Apr 12 18:32:21.699687 update_engine[1397]: I0412 18:32:21.699540 1397 action_processor.cc:36] ActionProcessor::StartProcessing: OmahaRequestAction Apr 12 18:32:21.699687 update_engine[1397]: I0412 18:32:21.699559 1397 omaha_request_action.cc:270] Posting an Omaha request to disabled Apr 12 18:32:21.699687 update_engine[1397]: I0412 18:32:21.699563 1397 omaha_request_action.cc:271] Request: Apr 12 18:32:21.699687 update_engine[1397]: Apr 12 18:32:21.699687 update_engine[1397]: Apr 12 18:32:21.699687 update_engine[1397]: Apr 12 18:32:21.699687 update_engine[1397]: Apr 12 18:32:21.699687 update_engine[1397]: Apr 12 18:32:21.699687 update_engine[1397]: Apr 12 18:32:21.699687 update_engine[1397]: I0412 18:32:21.699566 1397 libcurl_http_fetcher.cc:47] Starting/Resuming transfer Apr 12 18:32:21.699956 update_engine[1397]: I0412 18:32:21.699699 1397 libcurl_http_fetcher.cc:151] Setting up curl options for HTTP Apr 12 18:32:21.699956 update_engine[1397]: I0412 18:32:21.699925 1397 libcurl_http_fetcher.cc:449] Setting up timeout source: 1 seconds. Apr 12 18:32:21.700315 locksmithd[1498]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_REPORTING_ERROR_EVENT" NewVersion=0.0.0 NewSize=0 Apr 12 18:32:22.073193 update_engine[1397]: E0412 18:32:22.073072 1397 libcurl_http_fetcher.cc:266] Unable to get http response code: Could not resolve host: disabled Apr 12 18:32:22.073193 update_engine[1397]: I0412 18:32:22.073182 1397 libcurl_http_fetcher.cc:297] Transfer resulted in an error (0), 0 bytes downloaded Apr 12 18:32:22.073193 update_engine[1397]: I0412 18:32:22.073189 1397 omaha_request_action.cc:621] Omaha request response: Apr 12 18:32:22.073395 update_engine[1397]: I0412 18:32:22.073225 1397 action_processor.cc:65] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 12 18:32:22.073395 update_engine[1397]: I0412 18:32:22.073230 1397 action_processor.cc:73] ActionProcessor::ActionComplete: finished last action of type OmahaRequestAction Apr 12 18:32:22.073395 update_engine[1397]: I0412 18:32:22.073235 1397 update_attempter.cc:306] Processing Done. Apr 12 18:32:22.073395 update_engine[1397]: I0412 18:32:22.073240 1397 update_attempter.cc:310] Error event sent. Apr 12 18:32:22.073395 update_engine[1397]: I0412 18:32:22.073251 1397 update_check_scheduler.cc:74] Next update check in 44m58s Apr 12 18:32:22.074341 locksmithd[1498]: LastCheckedTime=0 Progress=0 CurrentOperation="UPDATE_STATUS_IDLE" NewVersion=0.0.0 NewSize=0 Apr 12 18:34:03.319849 systemd[1]: Started sshd@5-10.200.20.18:22-10.200.12.6:48618.service. Apr 12 18:34:03.725992 sshd[3902]: Accepted publickey for core from 10.200.12.6 port 48618 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:03.727722 sshd[3902]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:03.734796 systemd-logind[1396]: New session 8 of user core. Apr 12 18:34:03.735536 systemd[1]: Started session-8.scope. Apr 12 18:34:04.417428 sshd[3902]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:04.419912 systemd[1]: sshd@5-10.200.20.18:22-10.200.12.6:48618.service: Deactivated successfully. Apr 12 18:34:04.421139 systemd[1]: session-8.scope: Deactivated successfully. Apr 12 18:34:04.421140 systemd-logind[1396]: Session 8 logged out. Waiting for processes to exit. Apr 12 18:34:04.422369 systemd-logind[1396]: Removed session 8. Apr 12 18:34:09.484560 systemd[1]: Started sshd@6-10.200.20.18:22-10.200.12.6:55008.service. Apr 12 18:34:09.888284 sshd[3918]: Accepted publickey for core from 10.200.12.6 port 55008 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:09.889921 sshd[3918]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:09.895091 systemd[1]: Started session-9.scope. Apr 12 18:34:09.896108 systemd-logind[1396]: New session 9 of user core. Apr 12 18:34:10.244428 sshd[3918]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:10.247484 systemd-logind[1396]: Session 9 logged out. Waiting for processes to exit. Apr 12 18:34:10.248863 systemd[1]: sshd@6-10.200.20.18:22-10.200.12.6:55008.service: Deactivated successfully. Apr 12 18:34:10.249755 systemd[1]: session-9.scope: Deactivated successfully. Apr 12 18:34:10.251182 systemd-logind[1396]: Removed session 9. Apr 12 18:34:15.316772 systemd[1]: Started sshd@7-10.200.20.18:22-10.200.12.6:56544.service. Apr 12 18:34:15.756055 sshd[3933]: Accepted publickey for core from 10.200.12.6 port 56544 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:15.757581 sshd[3933]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:15.762117 systemd-logind[1396]: New session 10 of user core. Apr 12 18:34:15.762883 systemd[1]: Started session-10.scope. Apr 12 18:34:16.132670 sshd[3933]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:16.136786 systemd[1]: sshd@7-10.200.20.18:22-10.200.12.6:56544.service: Deactivated successfully. Apr 12 18:34:16.137983 systemd[1]: session-10.scope: Deactivated successfully. Apr 12 18:34:16.138451 systemd-logind[1396]: Session 10 logged out. Waiting for processes to exit. Apr 12 18:34:16.139633 systemd-logind[1396]: Removed session 10. Apr 12 18:34:21.199848 systemd[1]: Started sshd@8-10.200.20.18:22-10.200.12.6:56556.service. Apr 12 18:34:21.604255 sshd[3946]: Accepted publickey for core from 10.200.12.6 port 56556 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:21.605990 sshd[3946]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:21.610701 systemd[1]: Started session-11.scope. Apr 12 18:34:21.611293 systemd-logind[1396]: New session 11 of user core. Apr 12 18:34:21.974414 sshd[3946]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:21.977637 systemd-logind[1396]: Session 11 logged out. Waiting for processes to exit. Apr 12 18:34:21.978161 systemd[1]: sshd@8-10.200.20.18:22-10.200.12.6:56556.service: Deactivated successfully. Apr 12 18:34:21.979052 systemd[1]: session-11.scope: Deactivated successfully. Apr 12 18:34:21.980588 systemd-logind[1396]: Removed session 11. Apr 12 18:34:22.043635 systemd[1]: Started sshd@9-10.200.20.18:22-10.200.12.6:56562.service. Apr 12 18:34:22.454973 sshd[3960]: Accepted publickey for core from 10.200.12.6 port 56562 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:22.456830 sshd[3960]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:22.461814 systemd[1]: Started session-12.scope. Apr 12 18:34:22.462482 systemd-logind[1396]: New session 12 of user core. Apr 12 18:34:23.436639 sshd[3960]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:23.439486 systemd-logind[1396]: Session 12 logged out. Waiting for processes to exit. Apr 12 18:34:23.439636 systemd[1]: sshd@9-10.200.20.18:22-10.200.12.6:56562.service: Deactivated successfully. Apr 12 18:34:23.440596 systemd[1]: session-12.scope: Deactivated successfully. Apr 12 18:34:23.441102 systemd-logind[1396]: Removed session 12. Apr 12 18:34:23.505358 systemd[1]: Started sshd@10-10.200.20.18:22-10.200.12.6:56570.service. Apr 12 18:34:23.916737 sshd[3972]: Accepted publickey for core from 10.200.12.6 port 56570 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:23.918776 sshd[3972]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:23.925886 systemd-logind[1396]: New session 13 of user core. Apr 12 18:34:23.926396 systemd[1]: Started session-13.scope. Apr 12 18:34:24.285687 sshd[3972]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:24.289250 systemd[1]: sshd@10-10.200.20.18:22-10.200.12.6:56570.service: Deactivated successfully. Apr 12 18:34:24.291111 systemd-logind[1396]: Session 13 logged out. Waiting for processes to exit. Apr 12 18:34:24.291229 systemd[1]: session-13.scope: Deactivated successfully. Apr 12 18:34:24.292741 systemd-logind[1396]: Removed session 13. Apr 12 18:34:29.349609 systemd[1]: Started sshd@11-10.200.20.18:22-10.200.12.6:50100.service. Apr 12 18:34:29.756511 sshd[3988]: Accepted publickey for core from 10.200.12.6 port 50100 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:29.758310 sshd[3988]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:29.763091 systemd[1]: Started session-14.scope. Apr 12 18:34:29.763751 systemd-logind[1396]: New session 14 of user core. Apr 12 18:34:30.109568 sshd[3988]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:30.113479 systemd[1]: sshd@11-10.200.20.18:22-10.200.12.6:50100.service: Deactivated successfully. Apr 12 18:34:30.114399 systemd[1]: session-14.scope: Deactivated successfully. Apr 12 18:34:30.114975 systemd-logind[1396]: Session 14 logged out. Waiting for processes to exit. Apr 12 18:34:30.115800 systemd-logind[1396]: Removed session 14. Apr 12 18:34:35.176448 systemd[1]: Started sshd@12-10.200.20.18:22-10.200.12.6:33800.service. Apr 12 18:34:36.052333 sshd[4000]: Accepted publickey for core from 10.200.12.6 port 33800 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:36.305090 sshd[4000]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:36.309980 systemd[1]: Started session-15.scope. Apr 12 18:34:36.310246 systemd-logind[1396]: New session 15 of user core. Apr 12 18:34:36.611500 sshd[4000]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:36.614742 systemd[1]: sshd@12-10.200.20.18:22-10.200.12.6:33800.service: Deactivated successfully. Apr 12 18:34:36.615644 systemd[1]: session-15.scope: Deactivated successfully. Apr 12 18:34:36.616118 systemd-logind[1396]: Session 15 logged out. Waiting for processes to exit. Apr 12 18:34:36.616871 systemd-logind[1396]: Removed session 15. Apr 12 18:34:36.678195 systemd[1]: Started sshd@13-10.200.20.18:22-10.200.12.6:33810.service. Apr 12 18:34:37.087377 sshd[4013]: Accepted publickey for core from 10.200.12.6 port 33810 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:37.089348 sshd[4013]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:37.094106 systemd[1]: Started session-16.scope. Apr 12 18:34:37.094557 systemd-logind[1396]: New session 16 of user core. Apr 12 18:34:37.495075 sshd[4013]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:37.498169 systemd[1]: sshd@13-10.200.20.18:22-10.200.12.6:33810.service: Deactivated successfully. Apr 12 18:34:37.499071 systemd[1]: session-16.scope: Deactivated successfully. Apr 12 18:34:37.499801 systemd-logind[1396]: Session 16 logged out. Waiting for processes to exit. Apr 12 18:34:37.500716 systemd-logind[1396]: Removed session 16. Apr 12 18:34:37.568630 systemd[1]: Started sshd@14-10.200.20.18:22-10.200.12.6:33818.service. Apr 12 18:34:37.973514 sshd[4023]: Accepted publickey for core from 10.200.12.6 port 33818 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:37.975449 sshd[4023]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:37.979702 systemd-logind[1396]: New session 17 of user core. Apr 12 18:34:37.980290 systemd[1]: Started session-17.scope. Apr 12 18:34:39.185760 sshd[4023]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:39.188641 systemd-logind[1396]: Session 17 logged out. Waiting for processes to exit. Apr 12 18:34:39.188860 systemd[1]: sshd@14-10.200.20.18:22-10.200.12.6:33818.service: Deactivated successfully. Apr 12 18:34:39.189791 systemd[1]: session-17.scope: Deactivated successfully. Apr 12 18:34:39.190616 systemd-logind[1396]: Removed session 17. Apr 12 18:34:39.250788 systemd[1]: Started sshd@15-10.200.20.18:22-10.200.12.6:33834.service. Apr 12 18:34:39.654852 sshd[4041]: Accepted publickey for core from 10.200.12.6 port 33834 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:39.656540 sshd[4041]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:39.660255 systemd-logind[1396]: New session 18 of user core. Apr 12 18:34:39.661175 systemd[1]: Started session-18.scope. Apr 12 18:34:40.264435 sshd[4041]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:40.267755 systemd-logind[1396]: Session 18 logged out. Waiting for processes to exit. Apr 12 18:34:40.269455 systemd[1]: sshd@15-10.200.20.18:22-10.200.12.6:33834.service: Deactivated successfully. Apr 12 18:34:40.270333 systemd[1]: session-18.scope: Deactivated successfully. Apr 12 18:34:40.272091 systemd-logind[1396]: Removed session 18. Apr 12 18:34:40.330625 systemd[1]: Started sshd@16-10.200.20.18:22-10.200.12.6:33838.service. Apr 12 18:34:40.737132 sshd[4052]: Accepted publickey for core from 10.200.12.6 port 33838 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:40.738809 sshd[4052]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:40.742870 systemd-logind[1396]: New session 19 of user core. Apr 12 18:34:40.743444 systemd[1]: Started session-19.scope. Apr 12 18:34:41.098073 sshd[4052]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:41.101157 systemd-logind[1396]: Session 19 logged out. Waiting for processes to exit. Apr 12 18:34:41.101589 systemd[1]: sshd@16-10.200.20.18:22-10.200.12.6:33838.service: Deactivated successfully. Apr 12 18:34:41.102525 systemd[1]: session-19.scope: Deactivated successfully. Apr 12 18:34:41.103078 systemd-logind[1396]: Removed session 19. Apr 12 18:34:46.164847 systemd[1]: Started sshd@17-10.200.20.18:22-10.200.12.6:59818.service. Apr 12 18:34:46.568088 sshd[4067]: Accepted publickey for core from 10.200.12.6 port 59818 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:46.569936 sshd[4067]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:46.574556 systemd[1]: Started session-20.scope. Apr 12 18:34:46.574952 systemd-logind[1396]: New session 20 of user core. Apr 12 18:34:46.937427 sshd[4067]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:46.940573 systemd-logind[1396]: Session 20 logged out. Waiting for processes to exit. Apr 12 18:34:46.942083 systemd[1]: sshd@17-10.200.20.18:22-10.200.12.6:59818.service: Deactivated successfully. Apr 12 18:34:46.942985 systemd[1]: session-20.scope: Deactivated successfully. Apr 12 18:34:46.944927 systemd-logind[1396]: Removed session 20. Apr 12 18:34:52.004829 systemd[1]: Started sshd@18-10.200.20.18:22-10.200.12.6:59824.service. Apr 12 18:34:52.415751 sshd[4083]: Accepted publickey for core from 10.200.12.6 port 59824 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:52.417591 sshd[4083]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:52.421573 systemd-logind[1396]: New session 21 of user core. Apr 12 18:34:52.422160 systemd[1]: Started session-21.scope. Apr 12 18:34:52.772897 sshd[4083]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:52.775879 systemd-logind[1396]: Session 21 logged out. Waiting for processes to exit. Apr 12 18:34:52.776125 systemd[1]: sshd@18-10.200.20.18:22-10.200.12.6:59824.service: Deactivated successfully. Apr 12 18:34:52.777089 systemd[1]: session-21.scope: Deactivated successfully. Apr 12 18:34:52.777685 systemd-logind[1396]: Removed session 21. Apr 12 18:34:57.839304 systemd[1]: Started sshd@19-10.200.20.18:22-10.200.12.6:41874.service. Apr 12 18:34:58.244740 sshd[4096]: Accepted publickey for core from 10.200.12.6 port 41874 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:34:58.246534 sshd[4096]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:34:58.252125 systemd[1]: Started session-22.scope. Apr 12 18:34:58.252464 systemd-logind[1396]: New session 22 of user core. Apr 12 18:34:58.609417 sshd[4096]: pam_unix(sshd:session): session closed for user core Apr 12 18:34:58.612637 systemd-logind[1396]: Session 22 logged out. Waiting for processes to exit. Apr 12 18:34:58.612963 systemd[1]: sshd@19-10.200.20.18:22-10.200.12.6:41874.service: Deactivated successfully. Apr 12 18:34:58.613824 systemd[1]: session-22.scope: Deactivated successfully. Apr 12 18:34:58.615073 systemd-logind[1396]: Removed session 22. Apr 12 18:35:03.676872 systemd[1]: Started sshd@20-10.200.20.18:22-10.200.12.6:41878.service. Apr 12 18:35:04.086105 sshd[4110]: Accepted publickey for core from 10.200.12.6 port 41878 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:35:04.087828 sshd[4110]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:35:04.091864 systemd-logind[1396]: New session 23 of user core. Apr 12 18:35:04.092460 systemd[1]: Started session-23.scope. Apr 12 18:35:04.439573 sshd[4110]: pam_unix(sshd:session): session closed for user core Apr 12 18:35:04.442737 systemd-logind[1396]: Session 23 logged out. Waiting for processes to exit. Apr 12 18:35:04.442965 systemd[1]: sshd@20-10.200.20.18:22-10.200.12.6:41878.service: Deactivated successfully. Apr 12 18:35:04.443885 systemd[1]: session-23.scope: Deactivated successfully. Apr 12 18:35:04.444399 systemd-logind[1396]: Removed session 23. Apr 12 18:35:04.510901 systemd[1]: Started sshd@21-10.200.20.18:22-10.200.12.6:41892.service. Apr 12 18:35:04.953321 sshd[4123]: Accepted publickey for core from 10.200.12.6 port 41892 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:35:04.954796 sshd[4123]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:35:04.958991 systemd-logind[1396]: New session 24 of user core. Apr 12 18:35:04.959638 systemd[1]: Started session-24.scope. Apr 12 18:35:07.188073 env[1410]: time="2024-04-12T18:35:07.188028918Z" level=info msg="StopContainer for \"d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0\" with timeout 30 (s)" Apr 12 18:35:07.189548 env[1410]: time="2024-04-12T18:35:07.189500166Z" level=info msg="Stop container \"d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0\" with signal terminated" Apr 12 18:35:07.200532 env[1410]: time="2024-04-12T18:35:07.200471026Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Apr 12 18:35:07.205885 env[1410]: time="2024-04-12T18:35:07.205852056Z" level=info msg="StopContainer for \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\" with timeout 1 (s)" Apr 12 18:35:07.206360 env[1410]: time="2024-04-12T18:35:07.206335058Z" level=info msg="Stop container \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\" with signal terminated" Apr 12 18:35:07.222130 systemd-networkd[1583]: lxc_health: Link DOWN Apr 12 18:35:07.222137 systemd-networkd[1583]: lxc_health: Lost carrier Apr 12 18:35:07.231207 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0-rootfs.mount: Deactivated successfully. Apr 12 18:35:07.262983 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073-rootfs.mount: Deactivated successfully. Apr 12 18:35:08.066666 kubelet[2544]: E0412 18:35:08.066639 2544 kubelet.go:2760] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 12 18:35:08.225120 env[1410]: time="2024-04-12T18:35:08.225060469Z" level=info msg="Kill container \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\"" Apr 12 18:35:08.382251 env[1410]: time="2024-04-12T18:35:08.381870323Z" level=info msg="shim disconnected" id=d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0 Apr 12 18:35:08.382251 env[1410]: time="2024-04-12T18:35:08.381923763Z" level=warning msg="cleaning up after shim disconnected" id=d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0 namespace=k8s.io Apr 12 18:35:08.382251 env[1410]: time="2024-04-12T18:35:08.381935123Z" level=info msg="cleaning up dead shim" Apr 12 18:35:08.382616 env[1410]: time="2024-04-12T18:35:08.381870283Z" level=info msg="shim disconnected" id=9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073 Apr 12 18:35:08.382733 env[1410]: time="2024-04-12T18:35:08.382716807Z" level=warning msg="cleaning up after shim disconnected" id=9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073 namespace=k8s.io Apr 12 18:35:08.382810 env[1410]: time="2024-04-12T18:35:08.382795968Z" level=info msg="cleaning up dead shim" Apr 12 18:35:08.390166 env[1410]: time="2024-04-12T18:35:08.390124608Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:08Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4192 runtime=io.containerd.runc.v2\n" Apr 12 18:35:08.395143 env[1410]: time="2024-04-12T18:35:08.395101595Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:08Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4194 runtime=io.containerd.runc.v2\n" Apr 12 18:35:08.399143 env[1410]: time="2024-04-12T18:35:08.399100417Z" level=info msg="StopContainer for \"d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0\" returns successfully" Apr 12 18:35:08.400069 env[1410]: time="2024-04-12T18:35:08.400009742Z" level=info msg="StopPodSandbox for \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\"" Apr 12 18:35:08.400267 env[1410]: time="2024-04-12T18:35:08.400079702Z" level=info msg="Container to stop \"d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:35:08.401897 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408-shm.mount: Deactivated successfully. Apr 12 18:35:08.405107 env[1410]: time="2024-04-12T18:35:08.405029009Z" level=info msg="StopContainer for \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\" returns successfully" Apr 12 18:35:08.405653 env[1410]: time="2024-04-12T18:35:08.405619332Z" level=info msg="StopPodSandbox for \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\"" Apr 12 18:35:08.405748 env[1410]: time="2024-04-12T18:35:08.405686693Z" level=info msg="Container to stop \"983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:35:08.405748 env[1410]: time="2024-04-12T18:35:08.405701453Z" level=info msg="Container to stop \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:35:08.405748 env[1410]: time="2024-04-12T18:35:08.405713093Z" level=info msg="Container to stop \"d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:35:08.405748 env[1410]: time="2024-04-12T18:35:08.405726813Z" level=info msg="Container to stop \"bbcda24e149f4c1a7654a4b90eb62279addf7693ef6b75d9cbe1a1049340e5f9\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:35:08.405748 env[1410]: time="2024-04-12T18:35:08.405738133Z" level=info msg="Container to stop \"dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:35:08.407470 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467-shm.mount: Deactivated successfully. Apr 12 18:35:08.441254 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467-rootfs.mount: Deactivated successfully. Apr 12 18:35:08.457811 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408-rootfs.mount: Deactivated successfully. Apr 12 18:35:08.472174 env[1410]: time="2024-04-12T18:35:08.472115495Z" level=info msg="shim disconnected" id=577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408 Apr 12 18:35:08.472918 env[1410]: time="2024-04-12T18:35:08.472886459Z" level=warning msg="cleaning up after shim disconnected" id=577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408 namespace=k8s.io Apr 12 18:35:08.473051 env[1410]: time="2024-04-12T18:35:08.473035940Z" level=info msg="cleaning up dead shim" Apr 12 18:35:08.473652 env[1410]: time="2024-04-12T18:35:08.472781698Z" level=info msg="shim disconnected" id=1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467 Apr 12 18:35:08.473731 env[1410]: time="2024-04-12T18:35:08.473655383Z" level=warning msg="cleaning up after shim disconnected" id=1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467 namespace=k8s.io Apr 12 18:35:08.473731 env[1410]: time="2024-04-12T18:35:08.473665183Z" level=info msg="cleaning up dead shim" Apr 12 18:35:08.481894 env[1410]: time="2024-04-12T18:35:08.481836667Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:08Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4257 runtime=io.containerd.runc.v2\n" Apr 12 18:35:08.482285 env[1410]: time="2024-04-12T18:35:08.482175949Z" level=info msg="TearDown network for sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" successfully" Apr 12 18:35:08.482285 env[1410]: time="2024-04-12T18:35:08.482229350Z" level=info msg="StopPodSandbox for \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" returns successfully" Apr 12 18:35:08.482583 env[1410]: time="2024-04-12T18:35:08.482552871Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:08Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4256 runtime=io.containerd.runc.v2\n" Apr 12 18:35:08.483108 env[1410]: time="2024-04-12T18:35:08.483078634Z" level=info msg="TearDown network for sandbox \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\" successfully" Apr 12 18:35:08.483237 env[1410]: time="2024-04-12T18:35:08.483217075Z" level=info msg="StopPodSandbox for \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\" returns successfully" Apr 12 18:35:08.649718 kubelet[2544]: I0412 18:35:08.648955 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cilium-cgroup\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.649718 kubelet[2544]: I0412 18:35:08.649618 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdwwh\" (UniqueName: \"kubernetes.io/projected/03d8e334-f4e0-4ae8-85df-87f70aa7db1f-kube-api-access-jdwwh\") pod \"03d8e334-f4e0-4ae8-85df-87f70aa7db1f\" (UID: \"03d8e334-f4e0-4ae8-85df-87f70aa7db1f\") " Apr 12 18:35:08.649718 kubelet[2544]: I0412 18:35:08.649678 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-host-proc-sys-net\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.649911 kubelet[2544]: I0412 18:35:08.649733 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-lib-modules\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.649911 kubelet[2544]: I0412 18:35:08.649763 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d7962196-9459-4162-a9ec-322a70f6491c-cilium-config-path\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.649911 kubelet[2544]: I0412 18:35:08.649783 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-bpf-maps\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.649911 kubelet[2544]: I0412 18:35:08.649801 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-xtables-lock\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.649911 kubelet[2544]: I0412 18:35:08.649821 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d7962196-9459-4162-a9ec-322a70f6491c-clustermesh-secrets\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.649911 kubelet[2544]: I0412 18:35:08.649841 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-host-proc-sys-kernel\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.650047 kubelet[2544]: I0412 18:35:08.649862 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kv4pb\" (UniqueName: \"kubernetes.io/projected/d7962196-9459-4162-a9ec-322a70f6491c-kube-api-access-kv4pb\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.650047 kubelet[2544]: I0412 18:35:08.649883 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d7962196-9459-4162-a9ec-322a70f6491c-hubble-tls\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.650047 kubelet[2544]: I0412 18:35:08.649901 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-etc-cni-netd\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.650047 kubelet[2544]: I0412 18:35:08.649922 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/03d8e334-f4e0-4ae8-85df-87f70aa7db1f-cilium-config-path\") pod \"03d8e334-f4e0-4ae8-85df-87f70aa7db1f\" (UID: \"03d8e334-f4e0-4ae8-85df-87f70aa7db1f\") " Apr 12 18:35:08.650047 kubelet[2544]: I0412 18:35:08.649944 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cilium-run\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.650047 kubelet[2544]: I0412 18:35:08.649961 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cni-path\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.650181 kubelet[2544]: I0412 18:35:08.649979 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-hostproc\") pod \"d7962196-9459-4162-a9ec-322a70f6491c\" (UID: \"d7962196-9459-4162-a9ec-322a70f6491c\") " Apr 12 18:35:08.650181 kubelet[2544]: I0412 18:35:08.650031 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-hostproc" (OuterVolumeSpecName: "hostproc") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.650181 kubelet[2544]: I0412 18:35:08.649039 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.650338 kubelet[2544]: I0412 18:35:08.650318 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.650440 kubelet[2544]: I0412 18:35:08.650424 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.650538 kubelet[2544]: I0412 18:35:08.650524 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.650809 kubelet[2544]: W0412 18:35:08.650773 2544 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/d7962196-9459-4162-a9ec-322a70f6491c/volumes/kubernetes.io~configmap/cilium-config-path: clearQuota called, but quotas disabled Apr 12 18:35:08.652745 kubelet[2544]: I0412 18:35:08.652717 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.652893 kubelet[2544]: I0412 18:35:08.652879 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.653353 kubelet[2544]: W0412 18:35:08.653324 2544 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/03d8e334-f4e0-4ae8-85df-87f70aa7db1f/volumes/kubernetes.io~configmap/cilium-config-path: clearQuota called, but quotas disabled Apr 12 18:35:08.654276 systemd[1]: var-lib-kubelet-pods-03d8e334\x2df4e0\x2d4ae8\x2d85df\x2d87f70aa7db1f-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2djdwwh.mount: Deactivated successfully. Apr 12 18:35:08.657161 kubelet[2544]: I0412 18:35:08.657124 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.657473 kubelet[2544]: I0412 18:35:08.657176 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cni-path" (OuterVolumeSpecName: "cni-path") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.657585 kubelet[2544]: I0412 18:35:08.657564 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/03d8e334-f4e0-4ae8-85df-87f70aa7db1f-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "03d8e334-f4e0-4ae8-85df-87f70aa7db1f" (UID: "03d8e334-f4e0-4ae8-85df-87f70aa7db1f"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Apr 12 18:35:08.657713 kubelet[2544]: I0412 18:35:08.657698 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:08.658183 kubelet[2544]: I0412 18:35:08.658157 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03d8e334-f4e0-4ae8-85df-87f70aa7db1f-kube-api-access-jdwwh" (OuterVolumeSpecName: "kube-api-access-jdwwh") pod "03d8e334-f4e0-4ae8-85df-87f70aa7db1f" (UID: "03d8e334-f4e0-4ae8-85df-87f70aa7db1f"). InnerVolumeSpecName "kube-api-access-jdwwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:35:08.658419 kubelet[2544]: I0412 18:35:08.658400 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7962196-9459-4162-a9ec-322a70f6491c-kube-api-access-kv4pb" (OuterVolumeSpecName: "kube-api-access-kv4pb") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "kube-api-access-kv4pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:35:08.658895 kubelet[2544]: I0412 18:35:08.658864 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7962196-9459-4162-a9ec-322a70f6491c-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Apr 12 18:35:08.660501 kubelet[2544]: I0412 18:35:08.660458 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7962196-9459-4162-a9ec-322a70f6491c-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Apr 12 18:35:08.661583 kubelet[2544]: I0412 18:35:08.661562 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7962196-9459-4162-a9ec-322a70f6491c-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "d7962196-9459-4162-a9ec-322a70f6491c" (UID: "d7962196-9459-4162-a9ec-322a70f6491c"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:35:08.750426 kubelet[2544]: I0412 18:35:08.750383 2544 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-host-proc-sys-net\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750426 kubelet[2544]: I0412 18:35:08.750426 2544 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-lib-modules\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750621 kubelet[2544]: I0412 18:35:08.750439 2544 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/d7962196-9459-4162-a9ec-322a70f6491c-cilium-config-path\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750621 kubelet[2544]: I0412 18:35:08.750451 2544 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-bpf-maps\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750621 kubelet[2544]: I0412 18:35:08.750463 2544 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-xtables-lock\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750621 kubelet[2544]: I0412 18:35:08.750472 2544 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/d7962196-9459-4162-a9ec-322a70f6491c-clustermesh-secrets\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750621 kubelet[2544]: I0412 18:35:08.750484 2544 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-host-proc-sys-kernel\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750621 kubelet[2544]: I0412 18:35:08.750495 2544 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/d7962196-9459-4162-a9ec-322a70f6491c-hubble-tls\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750621 kubelet[2544]: I0412 18:35:08.750505 2544 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-etc-cni-netd\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750621 kubelet[2544]: I0412 18:35:08.750515 2544 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-kv4pb\" (UniqueName: \"kubernetes.io/projected/d7962196-9459-4162-a9ec-322a70f6491c-kube-api-access-kv4pb\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750853 kubelet[2544]: I0412 18:35:08.750527 2544 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/03d8e334-f4e0-4ae8-85df-87f70aa7db1f-cilium-config-path\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750853 kubelet[2544]: I0412 18:35:08.750536 2544 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cilium-run\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750853 kubelet[2544]: I0412 18:35:08.750545 2544 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cni-path\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750853 kubelet[2544]: I0412 18:35:08.750554 2544 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-hostproc\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750853 kubelet[2544]: I0412 18:35:08.750564 2544 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/d7962196-9459-4162-a9ec-322a70f6491c-cilium-cgroup\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:08.750853 kubelet[2544]: I0412 18:35:08.750573 2544 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-jdwwh\" (UniqueName: \"kubernetes.io/projected/03d8e334-f4e0-4ae8-85df-87f70aa7db1f-kube-api-access-jdwwh\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:09.183749 sshd[4123]: pam_unix(sshd:session): session closed for user core Apr 12 18:35:09.187315 systemd[1]: sshd@21-10.200.20.18:22-10.200.12.6:41892.service: Deactivated successfully. Apr 12 18:35:09.188156 systemd[1]: session-24.scope: Deactivated successfully. Apr 12 18:35:09.189230 systemd-logind[1396]: Session 24 logged out. Waiting for processes to exit. Apr 12 18:35:09.190041 systemd-logind[1396]: Removed session 24. Apr 12 18:35:09.250627 systemd[1]: Started sshd@22-10.200.20.18:22-10.200.12.6:54166.service. Apr 12 18:35:09.401970 systemd[1]: var-lib-kubelet-pods-d7962196\x2d9459\x2d4162\x2da9ec\x2d322a70f6491c-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dkv4pb.mount: Deactivated successfully. Apr 12 18:35:09.402162 systemd[1]: var-lib-kubelet-pods-d7962196\x2d9459\x2d4162\x2da9ec\x2d322a70f6491c-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Apr 12 18:35:09.402269 systemd[1]: var-lib-kubelet-pods-d7962196\x2d9459\x2d4162\x2da9ec\x2d322a70f6491c-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Apr 12 18:35:09.438079 kubelet[2544]: I0412 18:35:09.437728 2544 scope.go:115] "RemoveContainer" containerID="9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073" Apr 12 18:35:09.444707 env[1410]: time="2024-04-12T18:35:09.444654182Z" level=info msg="RemoveContainer for \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\"" Apr 12 18:35:09.463913 env[1410]: time="2024-04-12T18:35:09.463867326Z" level=info msg="RemoveContainer for \"9bef5ebbf54b409884ce1172f66fbe4ca18dbdb4fdbb8b288e6deaf0d8471073\" returns successfully" Apr 12 18:35:09.464484 kubelet[2544]: I0412 18:35:09.464453 2544 scope.go:115] "RemoveContainer" containerID="983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9" Apr 12 18:35:09.466055 env[1410]: time="2024-04-12T18:35:09.466011697Z" level=info msg="RemoveContainer for \"983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9\"" Apr 12 18:35:09.481766 env[1410]: time="2024-04-12T18:35:09.481721423Z" level=info msg="RemoveContainer for \"983bf21253d19e7b4cb9818ee500b448f79eafa6c09c70d6cdb5f317e98c0de9\" returns successfully" Apr 12 18:35:09.482239 kubelet[2544]: I0412 18:35:09.482211 2544 scope.go:115] "RemoveContainer" containerID="bbcda24e149f4c1a7654a4b90eb62279addf7693ef6b75d9cbe1a1049340e5f9" Apr 12 18:35:09.483624 env[1410]: time="2024-04-12T18:35:09.483585233Z" level=info msg="RemoveContainer for \"bbcda24e149f4c1a7654a4b90eb62279addf7693ef6b75d9cbe1a1049340e5f9\"" Apr 12 18:35:09.499379 env[1410]: time="2024-04-12T18:35:09.499336158Z" level=info msg="RemoveContainer for \"bbcda24e149f4c1a7654a4b90eb62279addf7693ef6b75d9cbe1a1049340e5f9\" returns successfully" Apr 12 18:35:09.499603 kubelet[2544]: I0412 18:35:09.499576 2544 scope.go:115] "RemoveContainer" containerID="d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd" Apr 12 18:35:09.500973 env[1410]: time="2024-04-12T18:35:09.500935687Z" level=info msg="RemoveContainer for \"d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd\"" Apr 12 18:35:09.516037 env[1410]: time="2024-04-12T18:35:09.515988248Z" level=info msg="RemoveContainer for \"d85df0cc82d8eacc6da1366f00aeaac38c69a664cc22eb9b6cafa2eeb8656edd\" returns successfully" Apr 12 18:35:09.516412 kubelet[2544]: I0412 18:35:09.516343 2544 scope.go:115] "RemoveContainer" containerID="dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73" Apr 12 18:35:09.517769 env[1410]: time="2024-04-12T18:35:09.517732058Z" level=info msg="RemoveContainer for \"dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73\"" Apr 12 18:35:09.528860 env[1410]: time="2024-04-12T18:35:09.528815758Z" level=info msg="RemoveContainer for \"dc58f5a71e008d4cce6027810e2d5ccc86da1902037b0035647b70967ba65a73\" returns successfully" Apr 12 18:35:09.529159 kubelet[2544]: I0412 18:35:09.529131 2544 scope.go:115] "RemoveContainer" containerID="d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0" Apr 12 18:35:09.530328 env[1410]: time="2024-04-12T18:35:09.530296246Z" level=info msg="RemoveContainer for \"d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0\"" Apr 12 18:35:09.547522 env[1410]: time="2024-04-12T18:35:09.547469059Z" level=info msg="RemoveContainer for \"d1d5161d783d05b8db6cbd8a6ca27f5dd90754e16843865d4311067dc6c077e0\" returns successfully" Apr 12 18:35:09.656494 sshd[4290]: Accepted publickey for core from 10.200.12.6 port 54166 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:35:09.657983 sshd[4290]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:35:09.662104 systemd-logind[1396]: New session 25 of user core. Apr 12 18:35:09.662756 systemd[1]: Started session-25.scope. Apr 12 18:35:09.834974 kubelet[2544]: I0412 18:35:09.834390 2544 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID=03d8e334-f4e0-4ae8-85df-87f70aa7db1f path="/var/lib/kubelet/pods/03d8e334-f4e0-4ae8-85df-87f70aa7db1f/volumes" Apr 12 18:35:09.835276 kubelet[2544]: I0412 18:35:09.835254 2544 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID=d7962196-9459-4162-a9ec-322a70f6491c path="/var/lib/kubelet/pods/d7962196-9459-4162-a9ec-322a70f6491c/volumes" Apr 12 18:35:10.612393 kubelet[2544]: I0412 18:35:10.612333 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:35:10.612877 kubelet[2544]: E0412 18:35:10.612846 2544 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d7962196-9459-4162-a9ec-322a70f6491c" containerName="apply-sysctl-overwrites" Apr 12 18:35:10.612978 kubelet[2544]: E0412 18:35:10.612968 2544 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d7962196-9459-4162-a9ec-322a70f6491c" containerName="mount-bpf-fs" Apr 12 18:35:10.613052 kubelet[2544]: E0412 18:35:10.613044 2544 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="03d8e334-f4e0-4ae8-85df-87f70aa7db1f" containerName="cilium-operator" Apr 12 18:35:10.613123 kubelet[2544]: E0412 18:35:10.613103 2544 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d7962196-9459-4162-a9ec-322a70f6491c" containerName="mount-cgroup" Apr 12 18:35:10.613181 kubelet[2544]: E0412 18:35:10.613172 2544 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d7962196-9459-4162-a9ec-322a70f6491c" containerName="cilium-agent" Apr 12 18:35:10.613270 kubelet[2544]: E0412 18:35:10.613259 2544 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="d7962196-9459-4162-a9ec-322a70f6491c" containerName="clean-cilium-state" Apr 12 18:35:10.613365 kubelet[2544]: I0412 18:35:10.613355 2544 memory_manager.go:346] "RemoveStaleState removing state" podUID="d7962196-9459-4162-a9ec-322a70f6491c" containerName="cilium-agent" Apr 12 18:35:10.613434 kubelet[2544]: I0412 18:35:10.613425 2544 memory_manager.go:346] "RemoveStaleState removing state" podUID="03d8e334-f4e0-4ae8-85df-87f70aa7db1f" containerName="cilium-operator" Apr 12 18:35:10.625287 kubelet[2544]: W0412 18:35:10.625242 2544 reflector.go:533] object-"kube-system"/"cilium-config": failed to list *v1.ConfigMap: configmaps "cilium-config" is forbidden: User "system:node:ci-3510.3.3-a-f11561af44" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510.3.3-a-f11561af44' and this object Apr 12 18:35:10.625519 kubelet[2544]: E0412 18:35:10.625503 2544 reflector.go:148] object-"kube-system"/"cilium-config": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps "cilium-config" is forbidden: User "system:node:ci-3510.3.3-a-f11561af44" cannot list resource "configmaps" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510.3.3-a-f11561af44' and this object Apr 12 18:35:10.625614 kubelet[2544]: W0412 18:35:10.625263 2544 reflector.go:533] object-"kube-system"/"cilium-ipsec-keys": failed to list *v1.Secret: secrets "cilium-ipsec-keys" is forbidden: User "system:node:ci-3510.3.3-a-f11561af44" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510.3.3-a-f11561af44' and this object Apr 12 18:35:10.625702 kubelet[2544]: E0412 18:35:10.625691 2544 reflector.go:148] object-"kube-system"/"cilium-ipsec-keys": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "cilium-ipsec-keys" is forbidden: User "system:node:ci-3510.3.3-a-f11561af44" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510.3.3-a-f11561af44' and this object Apr 12 18:35:10.625796 kubelet[2544]: W0412 18:35:10.625292 2544 reflector.go:533] object-"kube-system"/"cilium-clustermesh": failed to list *v1.Secret: secrets "cilium-clustermesh" is forbidden: User "system:node:ci-3510.3.3-a-f11561af44" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510.3.3-a-f11561af44' and this object Apr 12 18:35:10.625881 kubelet[2544]: E0412 18:35:10.625871 2544 reflector.go:148] object-"kube-system"/"cilium-clustermesh": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "cilium-clustermesh" is forbidden: User "system:node:ci-3510.3.3-a-f11561af44" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510.3.3-a-f11561af44' and this object Apr 12 18:35:10.625953 kubelet[2544]: W0412 18:35:10.625338 2544 reflector.go:533] object-"kube-system"/"hubble-server-certs": failed to list *v1.Secret: secrets "hubble-server-certs" is forbidden: User "system:node:ci-3510.3.3-a-f11561af44" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510.3.3-a-f11561af44' and this object Apr 12 18:35:10.626026 kubelet[2544]: E0412 18:35:10.626017 2544 reflector.go:148] object-"kube-system"/"hubble-server-certs": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets "hubble-server-certs" is forbidden: User "system:node:ci-3510.3.3-a-f11561af44" cannot list resource "secrets" in API group "" in the namespace "kube-system": no relationship found between node 'ci-3510.3.3-a-f11561af44' and this object Apr 12 18:35:10.655873 sshd[4290]: pam_unix(sshd:session): session closed for user core Apr 12 18:35:10.658489 kubelet[2544]: I0412 18:35:10.658443 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-host-proc-sys-kernel\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.658704 kubelet[2544]: I0412 18:35:10.658692 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-xtables-lock\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.658838 kubelet[2544]: I0412 18:35:10.658827 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-hostproc\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.658943 kubelet[2544]: I0412 18:35:10.658933 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-lib-modules\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.659189 kubelet[2544]: I0412 18:35:10.659170 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-host-proc-sys-net\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.659439 kubelet[2544]: I0412 18:35:10.659421 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-run\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.659585 kubelet[2544]: I0412 18:35:10.659547 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cni-path\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.659821 kubelet[2544]: I0412 18:35:10.659802 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-etc-cni-netd\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.660020 kubelet[2544]: I0412 18:35:10.660002 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-bpf-maps\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.660799 kubelet[2544]: I0412 18:35:10.660313 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-ipsec-secrets\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.660799 kubelet[2544]: I0412 18:35:10.660725 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-clustermesh-secrets\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.660953 kubelet[2544]: I0412 18:35:10.660843 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-config-path\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.660953 kubelet[2544]: I0412 18:35:10.660933 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-cgroup\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.661643 systemd-logind[1396]: Session 25 logged out. Waiting for processes to exit. Apr 12 18:35:10.663121 systemd[1]: sshd@22-10.200.20.18:22-10.200.12.6:54166.service: Deactivated successfully. Apr 12 18:35:10.663980 systemd[1]: session-25.scope: Deactivated successfully. Apr 12 18:35:10.668305 systemd-logind[1396]: Removed session 25. Apr 12 18:35:10.722340 systemd[1]: Started sshd@23-10.200.20.18:22-10.200.12.6:54178.service. Apr 12 18:35:10.762295 kubelet[2544]: I0412 18:35:10.762249 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-77x6k\" (UniqueName: \"kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-kube-api-access-77x6k\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:10.762469 kubelet[2544]: I0412 18:35:10.762330 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-hubble-tls\") pod \"cilium-t4ht5\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " pod="kube-system/cilium-t4ht5" Apr 12 18:35:11.143681 sshd[4301]: Accepted publickey for core from 10.200.12.6 port 54178 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:35:11.144782 sshd[4301]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:35:11.149670 systemd[1]: Started session-26.scope. Apr 12 18:35:11.150155 systemd-logind[1396]: New session 26 of user core. Apr 12 18:35:11.519471 sshd[4301]: pam_unix(sshd:session): session closed for user core Apr 12 18:35:11.522390 systemd[1]: sshd@23-10.200.20.18:22-10.200.12.6:54178.service: Deactivated successfully. Apr 12 18:35:11.523973 systemd[1]: session-26.scope: Deactivated successfully. Apr 12 18:35:11.524697 systemd-logind[1396]: Session 26 logged out. Waiting for processes to exit. Apr 12 18:35:11.525563 systemd-logind[1396]: Removed session 26. Apr 12 18:35:11.590997 systemd[1]: Started sshd@24-10.200.20.18:22-10.200.12.6:54184.service. Apr 12 18:35:11.763459 kubelet[2544]: E0412 18:35:11.763406 2544 secret.go:194] Couldn't get secret kube-system/cilium-ipsec-keys: failed to sync secret cache: timed out waiting for the condition Apr 12 18:35:11.765755 kubelet[2544]: E0412 18:35:11.763524 2544 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-ipsec-secrets podName:86fca758-331c-4529-8b98-e4f87e954bd1 nodeName:}" failed. No retries permitted until 2024-04-12 18:35:12.263499172 +0000 UTC m=+254.577207953 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cilium-ipsec-secrets" (UniqueName: "kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-ipsec-secrets") pod "cilium-t4ht5" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1") : failed to sync secret cache: timed out waiting for the condition Apr 12 18:35:11.765755 kubelet[2544]: E0412 18:35:11.763828 2544 configmap.go:199] Couldn't get configMap kube-system/cilium-config: failed to sync configmap cache: timed out waiting for the condition Apr 12 18:35:11.765755 kubelet[2544]: E0412 18:35:11.763883 2544 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-config-path podName:86fca758-331c-4529-8b98-e4f87e954bd1 nodeName:}" failed. No retries permitted until 2024-04-12 18:35:12.263870014 +0000 UTC m=+254.577578795 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cilium-config-path" (UniqueName: "kubernetes.io/configmap/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-config-path") pod "cilium-t4ht5" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1") : failed to sync configmap cache: timed out waiting for the condition Apr 12 18:35:11.863794 kubelet[2544]: E0412 18:35:11.863689 2544 projected.go:267] Couldn't get secret kube-system/hubble-server-certs: failed to sync secret cache: timed out waiting for the condition Apr 12 18:35:11.863794 kubelet[2544]: E0412 18:35:11.863721 2544 projected.go:198] Error preparing data for projected volume hubble-tls for pod kube-system/cilium-t4ht5: failed to sync secret cache: timed out waiting for the condition Apr 12 18:35:11.863794 kubelet[2544]: E0412 18:35:11.863792 2544 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-hubble-tls podName:86fca758-331c-4529-8b98-e4f87e954bd1 nodeName:}" failed. No retries permitted until 2024-04-12 18:35:12.363772311 +0000 UTC m=+254.677481092 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "hubble-tls" (UniqueName: "kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-hubble-tls") pod "cilium-t4ht5" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1") : failed to sync secret cache: timed out waiting for the condition Apr 12 18:35:12.031791 sshd[4316]: Accepted publickey for core from 10.200.12.6 port 54184 ssh2: RSA SHA256:FwI9mp8Uipvmjkr+VYh+76kYXjtYhCPwjtuEb1G3LpI Apr 12 18:35:12.033681 sshd[4316]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Apr 12 18:35:12.038697 systemd-logind[1396]: New session 27 of user core. Apr 12 18:35:12.039273 systemd[1]: Started session-27.scope. Apr 12 18:35:12.417517 env[1410]: time="2024-04-12T18:35:12.417458715Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-t4ht5,Uid:86fca758-331c-4529-8b98-e4f87e954bd1,Namespace:kube-system,Attempt:0,}" Apr 12 18:35:12.456547 env[1410]: time="2024-04-12T18:35:12.456474124Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:35:12.456547 env[1410]: time="2024-04-12T18:35:12.456518924Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:35:12.456751 env[1410]: time="2024-04-12T18:35:12.456711965Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:35:12.457091 env[1410]: time="2024-04-12T18:35:12.457040087Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28 pid=4336 runtime=io.containerd.runc.v2 Apr 12 18:35:12.496430 env[1410]: time="2024-04-12T18:35:12.496374137Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-t4ht5,Uid:86fca758-331c-4529-8b98-e4f87e954bd1,Namespace:kube-system,Attempt:0,} returns sandbox id \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\"" Apr 12 18:35:12.501349 env[1410]: time="2024-04-12T18:35:12.501299324Z" level=info msg="CreateContainer within sandbox \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 12 18:35:12.541160 env[1410]: time="2024-04-12T18:35:12.541085496Z" level=info msg="CreateContainer within sandbox \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"29f51b4d5bbdb4ffbbf3ef01035f34068eabb46bea331046b6d711b988e2d597\"" Apr 12 18:35:12.543389 env[1410]: time="2024-04-12T18:35:12.542450424Z" level=info msg="StartContainer for \"29f51b4d5bbdb4ffbbf3ef01035f34068eabb46bea331046b6d711b988e2d597\"" Apr 12 18:35:12.604606 env[1410]: time="2024-04-12T18:35:12.604548396Z" level=info msg="StartContainer for \"29f51b4d5bbdb4ffbbf3ef01035f34068eabb46bea331046b6d711b988e2d597\" returns successfully" Apr 12 18:35:12.674994 env[1410]: time="2024-04-12T18:35:12.674486330Z" level=info msg="shim disconnected" id=29f51b4d5bbdb4ffbbf3ef01035f34068eabb46bea331046b6d711b988e2d597 Apr 12 18:35:12.675259 env[1410]: time="2024-04-12T18:35:12.675226174Z" level=warning msg="cleaning up after shim disconnected" id=29f51b4d5bbdb4ffbbf3ef01035f34068eabb46bea331046b6d711b988e2d597 namespace=k8s.io Apr 12 18:35:12.675259 env[1410]: time="2024-04-12T18:35:12.675253934Z" level=info msg="cleaning up dead shim" Apr 12 18:35:12.682809 env[1410]: time="2024-04-12T18:35:12.682758214Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:12Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4419 runtime=io.containerd.runc.v2\n" Apr 12 18:35:13.068266 kubelet[2544]: E0412 18:35:13.068138 2544 kubelet.go:2760] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 12 18:35:13.456866 kubelet[2544]: I0412 18:35:13.456841 2544 setters.go:548] "Node became not ready" node="ci-3510.3.3-a-f11561af44" condition={Type:Ready Status:False LastHeartbeatTime:2024-04-12 18:35:13.456770383 +0000 UTC m=+255.770479164 LastTransitionTime:2024-04-12 18:35:13.456770383 +0000 UTC m=+255.770479164 Reason:KubeletNotReady Message:container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized} Apr 12 18:35:13.462060 env[1410]: time="2024-04-12T18:35:13.462008811Z" level=info msg="StopPodSandbox for \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\"" Apr 12 18:35:13.462565 env[1410]: time="2024-04-12T18:35:13.462535654Z" level=info msg="Container to stop \"29f51b4d5bbdb4ffbbf3ef01035f34068eabb46bea331046b6d711b988e2d597\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Apr 12 18:35:13.466710 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28-shm.mount: Deactivated successfully. Apr 12 18:35:13.492809 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28-rootfs.mount: Deactivated successfully. Apr 12 18:35:13.507654 env[1410]: time="2024-04-12T18:35:13.507604014Z" level=info msg="shim disconnected" id=34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28 Apr 12 18:35:13.507920 env[1410]: time="2024-04-12T18:35:13.507901215Z" level=warning msg="cleaning up after shim disconnected" id=34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28 namespace=k8s.io Apr 12 18:35:13.508007 env[1410]: time="2024-04-12T18:35:13.507993136Z" level=info msg="cleaning up dead shim" Apr 12 18:35:13.515690 env[1410]: time="2024-04-12T18:35:13.515647336Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:13Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4451 runtime=io.containerd.runc.v2\n" Apr 12 18:35:13.516166 env[1410]: time="2024-04-12T18:35:13.516138179Z" level=info msg="TearDown network for sandbox \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\" successfully" Apr 12 18:35:13.516300 env[1410]: time="2024-04-12T18:35:13.516279540Z" level=info msg="StopPodSandbox for \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\" returns successfully" Apr 12 18:35:13.577982 kubelet[2544]: I0412 18:35:13.577950 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cni-path\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.578167 kubelet[2544]: I0412 18:35:13.578033 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cni-path" (OuterVolumeSpecName: "cni-path") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.578167 kubelet[2544]: I0412 18:35:13.578068 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-xtables-lock\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.578167 kubelet[2544]: I0412 18:35:13.578094 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-clustermesh-secrets\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.578167 kubelet[2544]: I0412 18:35:13.578127 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.580238 kubelet[2544]: I0412 18:35:13.578468 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-ipsec-secrets\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580238 kubelet[2544]: I0412 18:35:13.578508 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-hubble-tls\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580238 kubelet[2544]: I0412 18:35:13.578539 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-lib-modules\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580238 kubelet[2544]: I0412 18:35:13.578573 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-cgroup\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580238 kubelet[2544]: I0412 18:35:13.578593 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-host-proc-sys-kernel\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580238 kubelet[2544]: I0412 18:35:13.578616 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-hostproc\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580494 kubelet[2544]: I0412 18:35:13.578661 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-bpf-maps\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580494 kubelet[2544]: I0412 18:35:13.578680 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-host-proc-sys-net\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580494 kubelet[2544]: I0412 18:35:13.578703 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-config-path\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580494 kubelet[2544]: I0412 18:35:13.578722 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-run\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580494 kubelet[2544]: I0412 18:35:13.578740 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-etc-cni-netd\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580494 kubelet[2544]: I0412 18:35:13.578760 2544 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-77x6k\" (UniqueName: \"kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-kube-api-access-77x6k\") pod \"86fca758-331c-4529-8b98-e4f87e954bd1\" (UID: \"86fca758-331c-4529-8b98-e4f87e954bd1\") " Apr 12 18:35:13.580679 kubelet[2544]: I0412 18:35:13.578800 2544 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cni-path\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.580679 kubelet[2544]: I0412 18:35:13.578812 2544 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-xtables-lock\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.580679 kubelet[2544]: I0412 18:35:13.578856 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-hostproc" (OuterVolumeSpecName: "hostproc") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.580679 kubelet[2544]: I0412 18:35:13.579338 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.580679 kubelet[2544]: I0412 18:35:13.579374 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.580679 kubelet[2544]: W0412 18:35:13.579551 2544 empty_dir.go:525] Warning: Failed to clear quota on /var/lib/kubelet/pods/86fca758-331c-4529-8b98-e4f87e954bd1/volumes/kubernetes.io~configmap/cilium-config-path: clearQuota called, but quotas disabled Apr 12 18:35:13.585969 kubelet[2544]: I0412 18:35:13.583925 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Apr 12 18:35:13.585969 kubelet[2544]: I0412 18:35:13.583991 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.585969 kubelet[2544]: I0412 18:35:13.584013 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.585969 kubelet[2544]: I0412 18:35:13.584033 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.585969 kubelet[2544]: I0412 18:35:13.584266 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.584883 systemd[1]: var-lib-kubelet-pods-86fca758\x2d331c\x2d4529\x2d8b98\x2de4f87e954bd1-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Apr 12 18:35:13.586323 kubelet[2544]: I0412 18:35:13.584291 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Apr 12 18:35:13.585054 systemd[1]: var-lib-kubelet-pods-86fca758\x2d331c\x2d4529\x2d8b98\x2de4f87e954bd1-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d77x6k.mount: Deactivated successfully. Apr 12 18:35:13.587058 kubelet[2544]: I0412 18:35:13.587014 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-kube-api-access-77x6k" (OuterVolumeSpecName: "kube-api-access-77x6k") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "kube-api-access-77x6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:35:13.587644 kubelet[2544]: I0412 18:35:13.587610 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Apr 12 18:35:13.590237 systemd[1]: var-lib-kubelet-pods-86fca758\x2d331c\x2d4529\x2d8b98\x2de4f87e954bd1-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Apr 12 18:35:13.590966 kubelet[2544]: I0412 18:35:13.590365 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Apr 12 18:35:13.591269 kubelet[2544]: I0412 18:35:13.591238 2544 operation_generator.go:878] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "86fca758-331c-4529-8b98-e4f87e954bd1" (UID: "86fca758-331c-4529-8b98-e4f87e954bd1"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Apr 12 18:35:13.679543 kubelet[2544]: I0412 18:35:13.679504 2544 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-bpf-maps\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679543 kubelet[2544]: I0412 18:35:13.679542 2544 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-config-path\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679543 kubelet[2544]: I0412 18:35:13.679554 2544 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-host-proc-sys-net\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679747 kubelet[2544]: I0412 18:35:13.679565 2544 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-run\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679747 kubelet[2544]: I0412 18:35:13.679576 2544 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-etc-cni-netd\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679747 kubelet[2544]: I0412 18:35:13.679587 2544 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-77x6k\" (UniqueName: \"kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-kube-api-access-77x6k\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679747 kubelet[2544]: I0412 18:35:13.679596 2544 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-clustermesh-secrets\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679747 kubelet[2544]: I0412 18:35:13.679606 2544 reconciler_common.go:300] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-ipsec-secrets\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679747 kubelet[2544]: I0412 18:35:13.679616 2544 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/86fca758-331c-4529-8b98-e4f87e954bd1-hubble-tls\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679747 kubelet[2544]: I0412 18:35:13.679625 2544 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-cilium-cgroup\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679747 kubelet[2544]: I0412 18:35:13.679635 2544 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-host-proc-sys-kernel\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679998 kubelet[2544]: I0412 18:35:13.679647 2544 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-lib-modules\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:13.679998 kubelet[2544]: I0412 18:35:13.679656 2544 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/86fca758-331c-4529-8b98-e4f87e954bd1-hostproc\") on node \"ci-3510.3.3-a-f11561af44\" DevicePath \"\"" Apr 12 18:35:14.279751 systemd[1]: var-lib-kubelet-pods-86fca758\x2d331c\x2d4529\x2d8b98\x2de4f87e954bd1-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Apr 12 18:35:14.464601 kubelet[2544]: I0412 18:35:14.464576 2544 scope.go:115] "RemoveContainer" containerID="29f51b4d5bbdb4ffbbf3ef01035f34068eabb46bea331046b6d711b988e2d597" Apr 12 18:35:14.468265 env[1410]: time="2024-04-12T18:35:14.467898395Z" level=info msg="RemoveContainer for \"29f51b4d5bbdb4ffbbf3ef01035f34068eabb46bea331046b6d711b988e2d597\"" Apr 12 18:35:14.477990 env[1410]: time="2024-04-12T18:35:14.477690087Z" level=info msg="RemoveContainer for \"29f51b4d5bbdb4ffbbf3ef01035f34068eabb46bea331046b6d711b988e2d597\" returns successfully" Apr 12 18:35:14.500468 kubelet[2544]: I0412 18:35:14.500432 2544 topology_manager.go:212] "Topology Admit Handler" Apr 12 18:35:14.500695 kubelet[2544]: E0412 18:35:14.500681 2544 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="86fca758-331c-4529-8b98-e4f87e954bd1" containerName="mount-cgroup" Apr 12 18:35:14.500865 kubelet[2544]: I0412 18:35:14.500849 2544 memory_manager.go:346] "RemoveStaleState removing state" podUID="86fca758-331c-4529-8b98-e4f87e954bd1" containerName="mount-cgroup" Apr 12 18:35:14.586101 kubelet[2544]: I0412 18:35:14.585986 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-hostproc\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586101 kubelet[2544]: I0412 18:35:14.586047 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-lib-modules\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586101 kubelet[2544]: I0412 18:35:14.586069 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-host-proc-sys-net\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586101 kubelet[2544]: I0412 18:35:14.586095 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-cilium-run\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586338 kubelet[2544]: I0412 18:35:14.586126 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-cni-path\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586338 kubelet[2544]: I0412 18:35:14.586149 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-clustermesh-secrets\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586338 kubelet[2544]: I0412 18:35:14.586170 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-host-proc-sys-kernel\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586338 kubelet[2544]: I0412 18:35:14.586207 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-bpf-maps\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586338 kubelet[2544]: I0412 18:35:14.586230 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-hubble-tls\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586338 kubelet[2544]: I0412 18:35:14.586252 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zchpq\" (UniqueName: \"kubernetes.io/projected/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-kube-api-access-zchpq\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586473 kubelet[2544]: I0412 18:35:14.586284 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-etc-cni-netd\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586473 kubelet[2544]: I0412 18:35:14.586305 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-cilium-config-path\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586473 kubelet[2544]: I0412 18:35:14.586323 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-cilium-cgroup\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586473 kubelet[2544]: I0412 18:35:14.586348 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-xtables-lock\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.586473 kubelet[2544]: I0412 18:35:14.586369 2544 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/e949c1cc-5012-423f-9e66-7d2cc7fd0cff-cilium-ipsec-secrets\") pod \"cilium-b6hhj\" (UID: \"e949c1cc-5012-423f-9e66-7d2cc7fd0cff\") " pod="kube-system/cilium-b6hhj" Apr 12 18:35:14.805313 env[1410]: time="2024-04-12T18:35:14.804888701Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-b6hhj,Uid:e949c1cc-5012-423f-9e66-7d2cc7fd0cff,Namespace:kube-system,Attempt:0,}" Apr 12 18:35:14.846877 env[1410]: time="2024-04-12T18:35:14.846720323Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Apr 12 18:35:14.846877 env[1410]: time="2024-04-12T18:35:14.846807764Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Apr 12 18:35:14.846877 env[1410]: time="2024-04-12T18:35:14.846833204Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Apr 12 18:35:14.847256 env[1410]: time="2024-04-12T18:35:14.847167166Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e pid=4480 runtime=io.containerd.runc.v2 Apr 12 18:35:14.891477 env[1410]: time="2024-04-12T18:35:14.891430880Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-b6hhj,Uid:e949c1cc-5012-423f-9e66-7d2cc7fd0cff,Namespace:kube-system,Attempt:0,} returns sandbox id \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\"" Apr 12 18:35:14.895232 env[1410]: time="2024-04-12T18:35:14.895173140Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Apr 12 18:35:14.932779 env[1410]: time="2024-04-12T18:35:14.932719459Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"4c16b22853c40043254821f5f29f3ef2cb068e8deebbd2dcc0abf0b99ef7a264\"" Apr 12 18:35:14.933625 env[1410]: time="2024-04-12T18:35:14.933573624Z" level=info msg="StartContainer for \"4c16b22853c40043254821f5f29f3ef2cb068e8deebbd2dcc0abf0b99ef7a264\"" Apr 12 18:35:14.988092 env[1410]: time="2024-04-12T18:35:14.988046032Z" level=info msg="StartContainer for \"4c16b22853c40043254821f5f29f3ef2cb068e8deebbd2dcc0abf0b99ef7a264\" returns successfully" Apr 12 18:35:15.045389 env[1410]: time="2024-04-12T18:35:15.045335295Z" level=info msg="shim disconnected" id=4c16b22853c40043254821f5f29f3ef2cb068e8deebbd2dcc0abf0b99ef7a264 Apr 12 18:35:15.045389 env[1410]: time="2024-04-12T18:35:15.045390975Z" level=warning msg="cleaning up after shim disconnected" id=4c16b22853c40043254821f5f29f3ef2cb068e8deebbd2dcc0abf0b99ef7a264 namespace=k8s.io Apr 12 18:35:15.045620 env[1410]: time="2024-04-12T18:35:15.045403695Z" level=info msg="cleaning up dead shim" Apr 12 18:35:15.053674 env[1410]: time="2024-04-12T18:35:15.053625979Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:15Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4563 runtime=io.containerd.runc.v2\n" Apr 12 18:35:15.491096 env[1410]: time="2024-04-12T18:35:15.491051007Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Apr 12 18:35:15.532543 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount552965328.mount: Deactivated successfully. Apr 12 18:35:15.538886 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2028968442.mount: Deactivated successfully. Apr 12 18:35:15.552033 env[1410]: time="2024-04-12T18:35:15.551978089Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"8b68d9613714e5e199c844d0606ed1362fdc9376bec46facfad92db0b1d1c025\"" Apr 12 18:35:15.553304 env[1410]: time="2024-04-12T18:35:15.552535452Z" level=info msg="StartContainer for \"8b68d9613714e5e199c844d0606ed1362fdc9376bec46facfad92db0b1d1c025\"" Apr 12 18:35:15.609258 env[1410]: time="2024-04-12T18:35:15.609150950Z" level=info msg="StartContainer for \"8b68d9613714e5e199c844d0606ed1362fdc9376bec46facfad92db0b1d1c025\" returns successfully" Apr 12 18:35:15.638978 env[1410]: time="2024-04-12T18:35:15.638926027Z" level=info msg="shim disconnected" id=8b68d9613714e5e199c844d0606ed1362fdc9376bec46facfad92db0b1d1c025 Apr 12 18:35:15.638978 env[1410]: time="2024-04-12T18:35:15.638973908Z" level=warning msg="cleaning up after shim disconnected" id=8b68d9613714e5e199c844d0606ed1362fdc9376bec46facfad92db0b1d1c025 namespace=k8s.io Apr 12 18:35:15.638978 env[1410]: time="2024-04-12T18:35:15.638982948Z" level=info msg="cleaning up dead shim" Apr 12 18:35:15.646981 env[1410]: time="2024-04-12T18:35:15.646926590Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:15Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4626 runtime=io.containerd.runc.v2\ntime=\"2024-04-12T18:35:15Z\" level=warning msg=\"failed to remove runc container\" error=\"runc did not terminate successfully: exit status 255: \" runtime=io.containerd.runc.v2\n" Apr 12 18:35:15.835436 kubelet[2544]: I0412 18:35:15.835014 2544 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID=86fca758-331c-4529-8b98-e4f87e954bd1 path="/var/lib/kubelet/pods/86fca758-331c-4529-8b98-e4f87e954bd1/volumes" Apr 12 18:35:16.495434 env[1410]: time="2024-04-12T18:35:16.495386896Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Apr 12 18:35:16.538926 env[1410]: time="2024-04-12T18:35:16.538841124Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"d2e2024afaabb50b9d9ee3fdb9c649036c9a5d89c8eb000a67bfbad601123908\"" Apr 12 18:35:16.539672 env[1410]: time="2024-04-12T18:35:16.539635968Z" level=info msg="StartContainer for \"d2e2024afaabb50b9d9ee3fdb9c649036c9a5d89c8eb000a67bfbad601123908\"" Apr 12 18:35:16.600514 env[1410]: time="2024-04-12T18:35:16.600466368Z" level=info msg="StartContainer for \"d2e2024afaabb50b9d9ee3fdb9c649036c9a5d89c8eb000a67bfbad601123908\" returns successfully" Apr 12 18:35:16.641711 env[1410]: time="2024-04-12T18:35:16.641651824Z" level=info msg="shim disconnected" id=d2e2024afaabb50b9d9ee3fdb9c649036c9a5d89c8eb000a67bfbad601123908 Apr 12 18:35:16.641920 env[1410]: time="2024-04-12T18:35:16.641768185Z" level=warning msg="cleaning up after shim disconnected" id=d2e2024afaabb50b9d9ee3fdb9c649036c9a5d89c8eb000a67bfbad601123908 namespace=k8s.io Apr 12 18:35:16.641920 env[1410]: time="2024-04-12T18:35:16.641781185Z" level=info msg="cleaning up dead shim" Apr 12 18:35:16.649302 env[1410]: time="2024-04-12T18:35:16.649247384Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:16Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4686 runtime=io.containerd.runc.v2\n" Apr 12 18:35:17.279978 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-d2e2024afaabb50b9d9ee3fdb9c649036c9a5d89c8eb000a67bfbad601123908-rootfs.mount: Deactivated successfully. Apr 12 18:35:17.501508 env[1410]: time="2024-04-12T18:35:17.501466331Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Apr 12 18:35:17.534491 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3616872242.mount: Deactivated successfully. Apr 12 18:35:17.540534 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3828050755.mount: Deactivated successfully. Apr 12 18:35:17.556889 env[1410]: time="2024-04-12T18:35:17.556830581Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"9d4d0359ab40818824e5ce752bf98ea859ff0c1ff629967e4f6120c2cdebd4e8\"" Apr 12 18:35:17.559249 env[1410]: time="2024-04-12T18:35:17.557878026Z" level=info msg="StartContainer for \"9d4d0359ab40818824e5ce752bf98ea859ff0c1ff629967e4f6120c2cdebd4e8\"" Apr 12 18:35:17.607955 env[1410]: time="2024-04-12T18:35:17.607899048Z" level=info msg="StartContainer for \"9d4d0359ab40818824e5ce752bf98ea859ff0c1ff629967e4f6120c2cdebd4e8\" returns successfully" Apr 12 18:35:17.644315 env[1410]: time="2024-04-12T18:35:17.644269518Z" level=info msg="shim disconnected" id=9d4d0359ab40818824e5ce752bf98ea859ff0c1ff629967e4f6120c2cdebd4e8 Apr 12 18:35:17.644635 env[1410]: time="2024-04-12T18:35:17.644614480Z" level=warning msg="cleaning up after shim disconnected" id=9d4d0359ab40818824e5ce752bf98ea859ff0c1ff629967e4f6120c2cdebd4e8 namespace=k8s.io Apr 12 18:35:17.644714 env[1410]: time="2024-04-12T18:35:17.644701761Z" level=info msg="cleaning up dead shim" Apr 12 18:35:17.652645 env[1410]: time="2024-04-12T18:35:17.652604522Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:17Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=4744 runtime=io.containerd.runc.v2\n" Apr 12 18:35:18.069673 kubelet[2544]: E0412 18:35:18.069631 2544 kubelet.go:2760] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Apr 12 18:35:18.504977 env[1410]: time="2024-04-12T18:35:18.504932450Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Apr 12 18:35:18.582313 env[1410]: time="2024-04-12T18:35:18.582258133Z" level=info msg="CreateContainer within sandbox \"4a8fdea8c49558c6e044210237965d142d7e51699501ce37df0f6b8794c4ba9e\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"8e0253e45f85fd502a229585d8f305bd5f59845933508798e1d2d3fc42ceafa7\"" Apr 12 18:35:18.583228 env[1410]: time="2024-04-12T18:35:18.583182418Z" level=info msg="StartContainer for \"8e0253e45f85fd502a229585d8f305bd5f59845933508798e1d2d3fc42ceafa7\"" Apr 12 18:35:18.666631 env[1410]: time="2024-04-12T18:35:18.666570252Z" level=info msg="StartContainer for \"8e0253e45f85fd502a229585d8f305bd5f59845933508798e1d2d3fc42ceafa7\" returns successfully" Apr 12 18:35:19.092234 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106(gcm-aes-ce))) Apr 12 18:35:20.651736 systemd[1]: run-containerd-runc-k8s.io-8e0253e45f85fd502a229585d8f305bd5f59845933508798e1d2d3fc42ceafa7-runc.gvsoGw.mount: Deactivated successfully. Apr 12 18:35:21.698540 systemd-networkd[1583]: lxc_health: Link UP Apr 12 18:35:21.716035 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Apr 12 18:35:21.714459 systemd-networkd[1583]: lxc_health: Gained carrier Apr 12 18:35:22.855355 systemd-networkd[1583]: lxc_health: Gained IPv6LL Apr 12 18:35:22.888929 systemd[1]: run-containerd-runc-k8s.io-8e0253e45f85fd502a229585d8f305bd5f59845933508798e1d2d3fc42ceafa7-runc.A0qC1N.mount: Deactivated successfully. Apr 12 18:35:22.896259 kubelet[2544]: I0412 18:35:22.894514 2544 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-b6hhj" podStartSLOduration=8.894469633 podCreationTimestamp="2024-04-12 18:35:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-04-12 18:35:19.527657087 +0000 UTC m=+261.841365868" watchObservedRunningTime="2024-04-12 18:35:22.894469633 +0000 UTC m=+265.208178414" Apr 12 18:35:25.068349 systemd[1]: run-containerd-runc-k8s.io-8e0253e45f85fd502a229585d8f305bd5f59845933508798e1d2d3fc42ceafa7-runc.8pcgng.mount: Deactivated successfully. Apr 12 18:35:27.213612 systemd[1]: run-containerd-runc-k8s.io-8e0253e45f85fd502a229585d8f305bd5f59845933508798e1d2d3fc42ceafa7-runc.kKCpSy.mount: Deactivated successfully. Apr 12 18:35:27.338247 sshd[4316]: pam_unix(sshd:session): session closed for user core Apr 12 18:35:27.341349 systemd[1]: sshd@24-10.200.20.18:22-10.200.12.6:54184.service: Deactivated successfully. Apr 12 18:35:27.342187 systemd[1]: session-27.scope: Deactivated successfully. Apr 12 18:35:27.342597 systemd-logind[1396]: Session 27 logged out. Waiting for processes to exit. Apr 12 18:35:27.343856 systemd-logind[1396]: Removed session 27. Apr 12 18:35:42.387285 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-2e556d2e2f2dbeeff4442b1efec977fd6c1284607bfcef3452c8100d034895b4-rootfs.mount: Deactivated successfully. Apr 12 18:35:42.496043 env[1410]: time="2024-04-12T18:35:42.495977477Z" level=info msg="shim disconnected" id=2e556d2e2f2dbeeff4442b1efec977fd6c1284607bfcef3452c8100d034895b4 Apr 12 18:35:42.496043 env[1410]: time="2024-04-12T18:35:42.496035837Z" level=warning msg="cleaning up after shim disconnected" id=2e556d2e2f2dbeeff4442b1efec977fd6c1284607bfcef3452c8100d034895b4 namespace=k8s.io Apr 12 18:35:42.496043 env[1410]: time="2024-04-12T18:35:42.496045357Z" level=info msg="cleaning up dead shim" Apr 12 18:35:42.503567 env[1410]: time="2024-04-12T18:35:42.503511633Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:42Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=5414 runtime=io.containerd.runc.v2\n" Apr 12 18:35:42.561948 kubelet[2544]: I0412 18:35:42.561153 2544 scope.go:115] "RemoveContainer" containerID="2e556d2e2f2dbeeff4442b1efec977fd6c1284607bfcef3452c8100d034895b4" Apr 12 18:35:42.564621 env[1410]: time="2024-04-12T18:35:42.564581044Z" level=info msg="CreateContainer within sandbox \"fe7f281ed42e6ac86c3bdb0cd9506d9282ee87dbb628df25464b3838e2dbea94\" for container &ContainerMetadata{Name:kube-controller-manager,Attempt:1,}" Apr 12 18:35:42.595997 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1959813187.mount: Deactivated successfully. Apr 12 18:35:42.601388 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount2736988884.mount: Deactivated successfully. Apr 12 18:35:42.617466 env[1410]: time="2024-04-12T18:35:42.617390736Z" level=info msg="CreateContainer within sandbox \"fe7f281ed42e6ac86c3bdb0cd9506d9282ee87dbb628df25464b3838e2dbea94\" for &ContainerMetadata{Name:kube-controller-manager,Attempt:1,} returns container id \"aaafdbbe4d01222061c9b2be80289ccc7c69c2e769f355a0d4a298235bafd173\"" Apr 12 18:35:42.618154 env[1410]: time="2024-04-12T18:35:42.618126340Z" level=info msg="StartContainer for \"aaafdbbe4d01222061c9b2be80289ccc7c69c2e769f355a0d4a298235bafd173\"" Apr 12 18:35:42.679538 env[1410]: time="2024-04-12T18:35:42.679491032Z" level=info msg="StartContainer for \"aaafdbbe4d01222061c9b2be80289ccc7c69c2e769f355a0d4a298235bafd173\" returns successfully" Apr 12 18:35:43.702821 kubelet[2544]: E0412 18:35:43.702556 2544 controller.go:193] "Failed to update lease" err="Put \"https://10.200.20.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.3-a-f11561af44?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 12 18:35:44.170723 kubelet[2544]: E0412 18:35:44.170526 2544 controller.go:193] "Failed to update lease" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.200.20.18:54736->10.200.20.23:2379: read: connection timed out" Apr 12 18:35:48.265903 kubelet[2544]: E0412 18:35:48.265478 2544 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"kube-apiserver-ci-3510.3.3-a-f11561af44.17c59c2a1335a42e", GenerateName:"", Namespace:"kube-system", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-apiserver-ci-3510.3.3-a-f11561af44", UID:"16f742faf238cd425321419501894d1a", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-apiserver}"}, Reason:"Unhealthy", Message:"Readiness probe failed: HTTP probe failed with statuscode: 500", Source:v1.EventSource{Component:"kubelet", Host:"ci-3510.3.3-a-f11561af44"}, FirstTimestamp:time.Date(2024, time.April, 12, 18, 35, 37, 809183790, time.Local), LastTimestamp:time.Date(2024, time.April, 12, 18, 35, 37, 809183790, time.Local), Count:1, Type:"Warning", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'rpc error: code = Unavailable desc = error reading from server: read tcp 10.200.20.18:54546->10.200.20.23:2379: read: connection timed out' (will not retry!) Apr 12 18:35:49.251594 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-08e50517fecdbc04b574da35d92f251117b2edf330df19a30e54c0ebc2967f0a-rootfs.mount: Deactivated successfully. Apr 12 18:35:49.270797 env[1410]: time="2024-04-12T18:35:49.270750223Z" level=info msg="shim disconnected" id=08e50517fecdbc04b574da35d92f251117b2edf330df19a30e54c0ebc2967f0a Apr 12 18:35:49.271292 env[1410]: time="2024-04-12T18:35:49.271267625Z" level=warning msg="cleaning up after shim disconnected" id=08e50517fecdbc04b574da35d92f251117b2edf330df19a30e54c0ebc2967f0a namespace=k8s.io Apr 12 18:35:49.271360 env[1410]: time="2024-04-12T18:35:49.271347506Z" level=info msg="cleaning up dead shim" Apr 12 18:35:49.279755 env[1410]: time="2024-04-12T18:35:49.279712825Z" level=warning msg="cleanup warnings time=\"2024-04-12T18:35:49Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=5471 runtime=io.containerd.runc.v2\n" Apr 12 18:35:49.578375 kubelet[2544]: I0412 18:35:49.577742 2544 scope.go:115] "RemoveContainer" containerID="08e50517fecdbc04b574da35d92f251117b2edf330df19a30e54c0ebc2967f0a" Apr 12 18:35:49.580142 env[1410]: time="2024-04-12T18:35:49.580097548Z" level=info msg="CreateContainer within sandbox \"ed60f317bfea09a2f8e26cb6ceb846a2afaee589ed3fbf72707a1a2b67fc38cf\" for container &ContainerMetadata{Name:kube-scheduler,Attempt:1,}" Apr 12 18:35:49.608381 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount331334476.mount: Deactivated successfully. Apr 12 18:35:49.623845 env[1410]: time="2024-04-12T18:35:49.623792352Z" level=info msg="CreateContainer within sandbox \"ed60f317bfea09a2f8e26cb6ceb846a2afaee589ed3fbf72707a1a2b67fc38cf\" for &ContainerMetadata{Name:kube-scheduler,Attempt:1,} returns container id \"7b3429a83d64d57f47d133335c615a3796505f51f78c7ac3d6a96a5bbcf995ee\"" Apr 12 18:35:49.624589 env[1410]: time="2024-04-12T18:35:49.624564755Z" level=info msg="StartContainer for \"7b3429a83d64d57f47d133335c615a3796505f51f78c7ac3d6a96a5bbcf995ee\"" Apr 12 18:35:49.691329 env[1410]: time="2024-04-12T18:35:49.691278307Z" level=info msg="StartContainer for \"7b3429a83d64d57f47d133335c615a3796505f51f78c7ac3d6a96a5bbcf995ee\" returns successfully" Apr 12 18:35:53.000986 kubelet[2544]: I0412 18:35:53.000947 2544 status_manager.go:809] "Failed to get status for pod" podUID=88be763039c4881497388a451291ec85 pod="kube-system/kube-controller-manager-ci-3510.3.3-a-f11561af44" err="rpc error: code = Unavailable desc = error reading from server: read tcp 10.200.20.18:54628->10.200.20.23:2379: read: connection timed out" Apr 12 18:35:54.172759 kubelet[2544]: E0412 18:35:54.171541 2544 controller.go:193] "Failed to update lease" err="Put \"https://10.200.20.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.3-a-f11561af44?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 12 18:35:57.826262 env[1410]: time="2024-04-12T18:35:57.826216094Z" level=info msg="StopPodSandbox for \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\"" Apr 12 18:35:57.826619 env[1410]: time="2024-04-12T18:35:57.826321615Z" level=info msg="TearDown network for sandbox \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\" successfully" Apr 12 18:35:57.826619 env[1410]: time="2024-04-12T18:35:57.826367735Z" level=info msg="StopPodSandbox for \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\" returns successfully" Apr 12 18:35:57.826905 env[1410]: time="2024-04-12T18:35:57.826864337Z" level=info msg="RemovePodSandbox for \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\"" Apr 12 18:35:57.826947 env[1410]: time="2024-04-12T18:35:57.826908018Z" level=info msg="Forcibly stopping sandbox \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\"" Apr 12 18:35:57.827026 env[1410]: time="2024-04-12T18:35:57.827002658Z" level=info msg="TearDown network for sandbox \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\" successfully" Apr 12 18:35:57.849034 env[1410]: time="2024-04-12T18:35:57.848883278Z" level=info msg="RemovePodSandbox \"577962ed94a7ef65c8108688b7de63a22074d51c516cedcb0dafcadf334f8408\" returns successfully" Apr 12 18:35:57.849739 env[1410]: time="2024-04-12T18:35:57.849560721Z" level=info msg="StopPodSandbox for \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\"" Apr 12 18:35:57.849739 env[1410]: time="2024-04-12T18:35:57.849651441Z" level=info msg="TearDown network for sandbox \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\" successfully" Apr 12 18:35:57.849739 env[1410]: time="2024-04-12T18:35:57.849684282Z" level=info msg="StopPodSandbox for \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\" returns successfully" Apr 12 18:35:57.851230 env[1410]: time="2024-04-12T18:35:57.850302044Z" level=info msg="RemovePodSandbox for \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\"" Apr 12 18:35:57.851230 env[1410]: time="2024-04-12T18:35:57.850333005Z" level=info msg="Forcibly stopping sandbox \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\"" Apr 12 18:35:57.851230 env[1410]: time="2024-04-12T18:35:57.850403685Z" level=info msg="TearDown network for sandbox \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\" successfully" Apr 12 18:35:57.862119 env[1410]: time="2024-04-12T18:35:57.862068978Z" level=info msg="RemovePodSandbox \"34ad4b544037c0dbc5264b192a6b6996c674fa319ef13e2cdf50128dd4078b28\" returns successfully" Apr 12 18:35:57.862626 env[1410]: time="2024-04-12T18:35:57.862593621Z" level=info msg="StopPodSandbox for \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\"" Apr 12 18:35:57.862724 env[1410]: time="2024-04-12T18:35:57.862682661Z" level=info msg="TearDown network for sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" successfully" Apr 12 18:35:57.862761 env[1410]: time="2024-04-12T18:35:57.862720501Z" level=info msg="StopPodSandbox for \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" returns successfully" Apr 12 18:35:57.863126 env[1410]: time="2024-04-12T18:35:57.863097623Z" level=info msg="RemovePodSandbox for \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\"" Apr 12 18:35:57.863259 env[1410]: time="2024-04-12T18:35:57.863127063Z" level=info msg="Forcibly stopping sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\"" Apr 12 18:35:57.863259 env[1410]: time="2024-04-12T18:35:57.863187463Z" level=info msg="TearDown network for sandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" successfully" Apr 12 18:35:57.871775 env[1410]: time="2024-04-12T18:35:57.871706902Z" level=info msg="RemovePodSandbox \"1487f76ef89eb49adf2f4ea95f40165e9db3f950d0ee90c8f3e43940309cd467\" returns successfully" Apr 12 18:35:58.004143 kubelet[2544]: W0412 18:35:58.004108 2544 machine.go:65] Cannot read vendor id correctly, set empty. Apr 12 18:36:04.173972 kubelet[2544]: E0412 18:36:04.172589 2544 controller.go:193] "Failed to update lease" err="Put \"https://10.200.20.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.3-a-f11561af44?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 12 18:36:14.173504 kubelet[2544]: E0412 18:36:14.173470 2544 controller.go:193] "Failed to update lease" err="Put \"https://10.200.20.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.3-a-f11561af44?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Apr 12 18:36:14.173966 kubelet[2544]: I0412 18:36:14.173951 2544 controller.go:116] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Apr 12 18:36:22.268516 kubelet[2544]: E0412 18:36:22.268402 2544 event.go:280] Server rejected event '&v1.Event{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"kube-apiserver-ci-3510.3.3-a-f11561af44.17c59c2a1335a42e", GenerateName:"", Namespace:"kube-system", SelfLink:"", UID:"", ResourceVersion:"", Generation:0, CreationTimestamp:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string(nil), Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry(nil)}, InvolvedObject:v1.ObjectReference{Kind:"Pod", Namespace:"kube-system", Name:"kube-apiserver-ci-3510.3.3-a-f11561af44", UID:"16f742faf238cd425321419501894d1a", APIVersion:"v1", ResourceVersion:"", FieldPath:"spec.containers{kube-apiserver}"}, Reason:"Unhealthy", Message:"Readiness probe failed: HTTP probe failed with statuscode: 500", Source:v1.EventSource{Component:"kubelet", Host:"ci-3510.3.3-a-f11561af44"}, FirstTimestamp:time.Date(2024, time.April, 12, 18, 35, 37, 809183790, time.Local), LastTimestamp:time.Date(2024, time.April, 12, 18, 35, 41, 818338682, time.Local), Count:2, Type:"Warning", EventTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Series:(*v1.EventSeries)(nil), Action:"", Related:(*v1.ObjectReference)(nil), ReportingController:"", ReportingInstance:""}': 'Timeout: request did not complete within requested timeout - context deadline exceeded' (will not retry!) Apr 12 18:36:24.175310 kubelet[2544]: E0412 18:36:24.175265 2544 controller.go:146] "Failed to ensure lease exists, will retry" err="Get \"https://10.200.20.18:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/ci-3510.3.3-a-f11561af44?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" interval="200ms" Apr 12 18:36:25.344375 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.344701 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.367658 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.367891 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.390098 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.390391 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.412813 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.413048 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.435126 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.435425 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.449923 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.450131 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.457306 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.464540 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.472577 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.480509 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.488675 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.496401 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.503844 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.511482 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.519085 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.527138 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.535494 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.543127 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.551090 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.558796 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.566540 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.574264 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.582037 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.589681 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.597581 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.605171 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.613031 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.620575 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.628245 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.635911 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.643664 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.651583 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.659688 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.667324 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.674896 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.682334 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.689877 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.697393 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.704954 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.712429 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.720089 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.728478 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.737961 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.745601 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.753178 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.760989 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.768656 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.776233 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.783758 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.791230 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.798899 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.806339 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.813808 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.822747 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.830398 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.837973 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.845503 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.853008 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.860475 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.868048 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.875509 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.883084 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.890511 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.897942 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.905397 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.912919 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.926871 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.927707 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.935599 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.943150 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.950704 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.958175 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.965849 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.973548 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.981380 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.989195 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:25.996719 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.004293 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.011751 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.019278 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.027061 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.034599 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.042441 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.049846 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.058986 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.067247 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.074892 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.082379 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.089758 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.097262 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.104819 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.112457 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.120293 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.128091 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.135653 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.143096 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.150971 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.158295 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.166001 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.173572 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.181126 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.188701 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.196269 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.203879 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.211607 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.218842 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.226335 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.233813 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.241485 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.249812 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.257258 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.265185 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.272641 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.280424 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.288342 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.296348 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.304144 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.311988 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.319812 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.327853 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.335788 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.344509 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.352167 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.359637 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.367061 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.375130 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.382881 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.390657 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.398667 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.406635 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.414485 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.422230 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.429688 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.437239 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.444669 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.452387 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.459981 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.467747 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.475615 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.483154 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.490641 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.498133 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.505494 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.513115 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.520850 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.528357 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.535787 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.543291 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.550754 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.558333 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.565921 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.573528 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.581099 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.588607 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.596076 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.603689 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.611233 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.618852 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.626475 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.633945 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.641428 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.650363 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.656503 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.663995 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.672232 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.680156 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.687963 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.695385 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.703288 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.711058 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.719212 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.726742 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.734491 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.742431 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.751212 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.759182 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.766943 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.774395 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.782316 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.790138 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.797932 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.805366 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.813228 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.820601 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.828121 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.835494 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.843081 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.850563 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.858044 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.865547 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.873124 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.881220 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.888956 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.896762 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.904646 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.912457 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.920404 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.928103 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.935930 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.943914 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.951609 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.959554 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.967239 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.975374 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.983068 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.991140 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:26.998868 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.006499 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.014165 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.021800 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.029358 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.037439 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.045513 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.053625 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.061486 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.070739 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.079294 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.087252 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.095506 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.103136 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.111095 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.118833 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.127014 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.134822 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.142783 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.150693 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.158634 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.166529 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.174472 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.182328 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.190063 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.198122 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.206026 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.214025 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.222493 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.230253 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.238185 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.245996 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.253989 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.262055 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.270647 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.278903 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.286810 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.294681 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.302390 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.310051 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.318233 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.326236 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.333961 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.341646 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.349640 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.357468 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.365333 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.373074 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.380976 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.388734 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.396610 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.404335 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.412083 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.419713 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.427565 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.435324 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.443124 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.450867 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.458820 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.466537 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.474222 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.481815 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.489626 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.497411 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.505252 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.513164 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.520973 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.528667 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.536694 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.544685 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.552639 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.560322 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.568362 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.576498 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.585489 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.593186 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.600895 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.608549 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.616195 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.623905 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.631691 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.640099 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.648141 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.655820 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.663834 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.672173 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.680088 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.694637 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.695590 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.703295 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.710755 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.718304 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.725870 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.733386 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.740882 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.748487 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.756115 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.763597 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.770992 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.778622 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.785930 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.793698 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.801557 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.809285 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.817078 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.825016 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.836733 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.840432 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.847932 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.855595 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.863511 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.871478 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.879421 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.887823 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.896124 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.904401 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.912534 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.921394 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.929172 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.937192 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.945095 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.955011 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.962767 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.970582 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.978178 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.986134 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:27.993769 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.001322 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.008713 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.016456 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.023968 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.031787 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.039399 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.047279 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.055240 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.065025 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.073272 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.081494 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.089031 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.096620 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.104091 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.113070 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.120746 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.128655 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.136105 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.143903 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.151611 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.160550 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.166625 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.174346 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.181925 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.189383 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.196799 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.204515 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.211930 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.219889 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.227913 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.235415 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.243423 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.251168 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.258942 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.267001 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.274674 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.282379 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.289760 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.297724 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.305476 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.313259 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.321090 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.329095 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.336446 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.344312 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.351749 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.359748 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.367178 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.374546 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.382029 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.389950 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.397623 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.405193 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.413071 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.420755 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.428720 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.436571 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.444599 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.452014 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.459506 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.467130 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.475039 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.483172 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.491460 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.499267 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.507376 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.515531 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.522883 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.530807 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.538824 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.546362 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.554091 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.561537 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.569268 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.576586 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.584251 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.591880 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.599496 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.606989 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.614444 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.622622 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.633123 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.640811 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.649118 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.656848 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.673725 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.674090 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.681336 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.689066 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.696814 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.704593 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.712570 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.720583 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.728347 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#141 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001 Apr 12 18:36:28.735994 kernel: hv_storvsc f8b3781a-1e82-4818-a1c3-63d806ec15bb: tag#140 cmd 0x2a status: scsi 0x2 srb 0x4 hv 0xc0000001