Dec 13 14:19:31.091307 kernel: Linux version 5.15.173-flatcar (build@pony-truck.infra.kinvolk.io) (x86_64-cros-linux-gnu-gcc (Gentoo Hardened 11.3.1_p20221209 p3) 11.3.1 20221209, GNU ld (Gentoo 2.39 p5) 2.39.0) #1 SMP Fri Dec 13 12:55:10 -00 2024 Dec 13 14:19:31.091341 kernel: Command line: BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=8c474c3ec361ec863adbecaa85281a726e1b53f7863ecc4742be8c5f6d02a66e Dec 13 14:19:31.091352 kernel: BIOS-provided physical RAM map: Dec 13 14:19:31.091359 kernel: BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable Dec 13 14:19:31.091366 kernel: BIOS-e820: [mem 0x000000000009fc00-0x000000000009ffff] reserved Dec 13 14:19:31.091373 kernel: BIOS-e820: [mem 0x00000000000f0000-0x00000000000fffff] reserved Dec 13 14:19:31.091381 kernel: BIOS-e820: [mem 0x0000000000100000-0x000000009cfdbfff] usable Dec 13 14:19:31.091388 kernel: BIOS-e820: [mem 0x000000009cfdc000-0x000000009cffffff] reserved Dec 13 14:19:31.091401 kernel: BIOS-e820: [mem 0x00000000b0000000-0x00000000bfffffff] reserved Dec 13 14:19:31.091409 kernel: BIOS-e820: [mem 0x00000000fed1c000-0x00000000fed1ffff] reserved Dec 13 14:19:31.091416 kernel: BIOS-e820: [mem 0x00000000feffc000-0x00000000feffffff] reserved Dec 13 14:19:31.091423 kernel: BIOS-e820: [mem 0x00000000fffc0000-0x00000000ffffffff] reserved Dec 13 14:19:31.091429 kernel: BIOS-e820: [mem 0x000000fd00000000-0x000000ffffffffff] reserved Dec 13 14:19:31.091435 kernel: NX (Execute Disable) protection: active Dec 13 14:19:31.091447 kernel: SMBIOS 2.8 present. Dec 13 14:19:31.091453 kernel: DMI: QEMU Standard PC (Q35 + ICH9, 2009), BIOS 1.16.2-debian-1.16.2-1 04/01/2014 Dec 13 14:19:31.091459 kernel: Hypervisor detected: KVM Dec 13 14:19:31.091465 kernel: kvm-clock: Using msrs 4b564d01 and 4b564d00 Dec 13 14:19:31.091471 kernel: kvm-clock: cpu 0, msr 8a19a001, primary cpu clock Dec 13 14:19:31.091477 kernel: kvm-clock: using sched offset of 3225440422 cycles Dec 13 14:19:31.091484 kernel: clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns Dec 13 14:19:31.091493 kernel: tsc: Detected 2794.748 MHz processor Dec 13 14:19:31.091500 kernel: e820: update [mem 0x00000000-0x00000fff] usable ==> reserved Dec 13 14:19:31.091511 kernel: e820: remove [mem 0x000a0000-0x000fffff] usable Dec 13 14:19:31.091519 kernel: last_pfn = 0x9cfdc max_arch_pfn = 0x400000000 Dec 13 14:19:31.091526 kernel: x86/PAT: Configuration [0-7]: WB WC UC- UC WB WP UC- WT Dec 13 14:19:31.091532 kernel: Using GB pages for direct mapping Dec 13 14:19:31.091538 kernel: ACPI: Early table checksum verification disabled Dec 13 14:19:31.091544 kernel: ACPI: RSDP 0x00000000000F59D0 000014 (v00 BOCHS ) Dec 13 14:19:31.091551 kernel: ACPI: RSDT 0x000000009CFE2408 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:19:31.091557 kernel: ACPI: FACP 0x000000009CFE21E8 0000F4 (v03 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:19:31.091565 kernel: ACPI: DSDT 0x000000009CFE0040 0021A8 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:19:31.091578 kernel: ACPI: FACS 0x000000009CFE0000 000040 Dec 13 14:19:31.091587 kernel: ACPI: APIC 0x000000009CFE22DC 000090 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:19:31.091595 kernel: ACPI: HPET 0x000000009CFE236C 000038 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:19:31.091603 kernel: ACPI: MCFG 0x000000009CFE23A4 00003C (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:19:31.091610 kernel: ACPI: WAET 0x000000009CFE23E0 000028 (v01 BOCHS BXPC 00000001 BXPC 00000001) Dec 13 14:19:31.091618 kernel: ACPI: Reserving FACP table memory at [mem 0x9cfe21e8-0x9cfe22db] Dec 13 14:19:31.091627 kernel: ACPI: Reserving DSDT table memory at [mem 0x9cfe0040-0x9cfe21e7] Dec 13 14:19:31.091635 kernel: ACPI: Reserving FACS table memory at [mem 0x9cfe0000-0x9cfe003f] Dec 13 14:19:31.091651 kernel: ACPI: Reserving APIC table memory at [mem 0x9cfe22dc-0x9cfe236b] Dec 13 14:19:31.091657 kernel: ACPI: Reserving HPET table memory at [mem 0x9cfe236c-0x9cfe23a3] Dec 13 14:19:31.091664 kernel: ACPI: Reserving MCFG table memory at [mem 0x9cfe23a4-0x9cfe23df] Dec 13 14:19:31.091671 kernel: ACPI: Reserving WAET table memory at [mem 0x9cfe23e0-0x9cfe2407] Dec 13 14:19:31.091677 kernel: No NUMA configuration found Dec 13 14:19:31.091684 kernel: Faking a node at [mem 0x0000000000000000-0x000000009cfdbfff] Dec 13 14:19:31.091695 kernel: NODE_DATA(0) allocated [mem 0x9cfd6000-0x9cfdbfff] Dec 13 14:19:31.091704 kernel: Zone ranges: Dec 13 14:19:31.091713 kernel: DMA [mem 0x0000000000001000-0x0000000000ffffff] Dec 13 14:19:31.091721 kernel: DMA32 [mem 0x0000000001000000-0x000000009cfdbfff] Dec 13 14:19:31.091728 kernel: Normal empty Dec 13 14:19:31.091736 kernel: Movable zone start for each node Dec 13 14:19:31.091744 kernel: Early memory node ranges Dec 13 14:19:31.091753 kernel: node 0: [mem 0x0000000000001000-0x000000000009efff] Dec 13 14:19:31.091761 kernel: node 0: [mem 0x0000000000100000-0x000000009cfdbfff] Dec 13 14:19:31.091775 kernel: Initmem setup node 0 [mem 0x0000000000001000-0x000000009cfdbfff] Dec 13 14:19:31.091784 kernel: On node 0, zone DMA: 1 pages in unavailable ranges Dec 13 14:19:31.091790 kernel: On node 0, zone DMA: 97 pages in unavailable ranges Dec 13 14:19:31.091797 kernel: On node 0, zone DMA32: 12324 pages in unavailable ranges Dec 13 14:19:31.091804 kernel: ACPI: PM-Timer IO Port: 0x608 Dec 13 14:19:31.091851 kernel: ACPI: LAPIC_NMI (acpi_id[0xff] dfl dfl lint[0x1]) Dec 13 14:19:31.091858 kernel: IOAPIC[0]: apic_id 0, version 17, address 0xfec00000, GSI 0-23 Dec 13 14:19:31.091865 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 0 global_irq 2 dfl dfl) Dec 13 14:19:31.091872 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 5 global_irq 5 high level) Dec 13 14:19:31.091878 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 9 global_irq 9 high level) Dec 13 14:19:31.091890 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 10 global_irq 10 high level) Dec 13 14:19:31.091897 kernel: ACPI: INT_SRC_OVR (bus 0 bus_irq 11 global_irq 11 high level) Dec 13 14:19:31.091904 kernel: ACPI: Using ACPI (MADT) for SMP configuration information Dec 13 14:19:31.091913 kernel: ACPI: HPET id: 0x8086a201 base: 0xfed00000 Dec 13 14:19:31.091919 kernel: TSC deadline timer available Dec 13 14:19:31.091926 kernel: smpboot: Allowing 4 CPUs, 0 hotplug CPUs Dec 13 14:19:31.091933 kernel: kvm-guest: KVM setup pv remote TLB flush Dec 13 14:19:31.091939 kernel: kvm-guest: setup PV sched yield Dec 13 14:19:31.091946 kernel: [mem 0xc0000000-0xfed1bfff] available for PCI devices Dec 13 14:19:31.091954 kernel: Booting paravirtualized kernel on KVM Dec 13 14:19:31.091961 kernel: clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1910969940391419 ns Dec 13 14:19:31.091968 kernel: setup_percpu: NR_CPUS:512 nr_cpumask_bits:512 nr_cpu_ids:4 nr_node_ids:1 Dec 13 14:19:31.091975 kernel: percpu: Embedded 56 pages/cpu s188696 r8192 d32488 u524288 Dec 13 14:19:31.091981 kernel: pcpu-alloc: s188696 r8192 d32488 u524288 alloc=1*2097152 Dec 13 14:19:31.091988 kernel: pcpu-alloc: [0] 0 1 2 3 Dec 13 14:19:31.091994 kernel: kvm-guest: setup async PF for cpu 0 Dec 13 14:19:31.092001 kernel: kvm-guest: stealtime: cpu 0, msr 9a41c0c0 Dec 13 14:19:31.092007 kernel: kvm-guest: PV spinlocks enabled Dec 13 14:19:31.092015 kernel: PV qspinlock hash table entries: 256 (order: 0, 4096 bytes, linear) Dec 13 14:19:31.092022 kernel: Built 1 zonelists, mobility grouping on. Total pages: 632732 Dec 13 14:19:31.092030 kernel: Policy zone: DMA32 Dec 13 14:19:31.092039 kernel: Kernel command line: rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=8c474c3ec361ec863adbecaa85281a726e1b53f7863ecc4742be8c5f6d02a66e Dec 13 14:19:31.092049 kernel: Unknown kernel command line parameters "BOOT_IMAGE=/flatcar/vmlinuz-a", will be passed to user space. Dec 13 14:19:31.092058 kernel: Dentry cache hash table entries: 524288 (order: 10, 4194304 bytes, linear) Dec 13 14:19:31.092066 kernel: Inode-cache hash table entries: 262144 (order: 9, 2097152 bytes, linear) Dec 13 14:19:31.092074 kernel: mem auto-init: stack:off, heap alloc:off, heap free:off Dec 13 14:19:31.092090 kernel: Memory: 2436696K/2571752K available (12294K kernel code, 2275K rwdata, 13716K rodata, 47472K init, 4112K bss, 134796K reserved, 0K cma-reserved) Dec 13 14:19:31.092097 kernel: SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=4, Nodes=1 Dec 13 14:19:31.092104 kernel: ftrace: allocating 34549 entries in 135 pages Dec 13 14:19:31.092111 kernel: ftrace: allocated 135 pages with 4 groups Dec 13 14:19:31.092117 kernel: rcu: Hierarchical RCU implementation. Dec 13 14:19:31.092126 kernel: rcu: RCU event tracing is enabled. Dec 13 14:19:31.092145 kernel: rcu: RCU restricting CPUs from NR_CPUS=512 to nr_cpu_ids=4. Dec 13 14:19:31.092153 kernel: Rude variant of Tasks RCU enabled. Dec 13 14:19:31.092160 kernel: Tracing variant of Tasks RCU enabled. Dec 13 14:19:31.092174 kernel: rcu: RCU calculated value of scheduler-enlistment delay is 100 jiffies. Dec 13 14:19:31.092183 kernel: rcu: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=4 Dec 13 14:19:31.092191 kernel: NR_IRQS: 33024, nr_irqs: 456, preallocated irqs: 16 Dec 13 14:19:31.092200 kernel: random: crng init done Dec 13 14:19:31.092208 kernel: Console: colour VGA+ 80x25 Dec 13 14:19:31.092216 kernel: printk: console [ttyS0] enabled Dec 13 14:19:31.092225 kernel: ACPI: Core revision 20210730 Dec 13 14:19:31.092233 kernel: clocksource: hpet: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 19112604467 ns Dec 13 14:19:31.092240 kernel: APIC: Switch to symmetric I/O mode setup Dec 13 14:19:31.092252 kernel: x2apic enabled Dec 13 14:19:31.092260 kernel: Switched APIC routing to physical x2apic. Dec 13 14:19:31.092267 kernel: kvm-guest: setup PV IPIs Dec 13 14:19:31.092274 kernel: ..TIMER: vector=0x30 apic1=0 pin1=2 apic2=-1 pin2=-1 Dec 13 14:19:31.092281 kernel: tsc: Marking TSC unstable due to TSCs unsynchronized Dec 13 14:19:31.092292 kernel: Calibrating delay loop (skipped) preset value.. 5589.49 BogoMIPS (lpj=2794748) Dec 13 14:19:31.092299 kernel: x86/cpu: User Mode Instruction Prevention (UMIP) activated Dec 13 14:19:31.092305 kernel: Last level iTLB entries: 4KB 512, 2MB 255, 4MB 127 Dec 13 14:19:31.092312 kernel: Last level dTLB entries: 4KB 512, 2MB 255, 4MB 127, 1GB 0 Dec 13 14:19:31.092328 kernel: Spectre V1 : Mitigation: usercopy/swapgs barriers and __user pointer sanitization Dec 13 14:19:31.092335 kernel: Spectre V2 : Mitigation: Retpolines Dec 13 14:19:31.092342 kernel: Spectre V2 : Spectre v2 / SpectreRSB mitigation: Filling RSB on context switch Dec 13 14:19:31.092350 kernel: Spectre V2 : Spectre v2 / SpectreRSB : Filling RSB on VMEXIT Dec 13 14:19:31.092357 kernel: Spectre V2 : Enabling Speculation Barrier for firmware calls Dec 13 14:19:31.092364 kernel: RETBleed: Mitigation: untrained return thunk Dec 13 14:19:31.092371 kernel: Spectre V2 : mitigation: Enabling conditional Indirect Branch Prediction Barrier Dec 13 14:19:31.092378 kernel: Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp Dec 13 14:19:31.092386 kernel: x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' Dec 13 14:19:31.092394 kernel: x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' Dec 13 14:19:31.092401 kernel: x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' Dec 13 14:19:31.092408 kernel: x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 Dec 13 14:19:31.092415 kernel: x86/fpu: Enabled xstate features 0x7, context size is 832 bytes, using 'standard' format. Dec 13 14:19:31.092422 kernel: Freeing SMP alternatives memory: 32K Dec 13 14:19:31.092429 kernel: pid_max: default: 32768 minimum: 301 Dec 13 14:19:31.092436 kernel: LSM: Security Framework initializing Dec 13 14:19:31.092444 kernel: SELinux: Initializing. Dec 13 14:19:31.092451 kernel: Mount-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 13 14:19:31.092458 kernel: Mountpoint-cache hash table entries: 8192 (order: 4, 65536 bytes, linear) Dec 13 14:19:31.092465 kernel: smpboot: CPU0: AMD EPYC 7402P 24-Core Processor (family: 0x17, model: 0x31, stepping: 0x0) Dec 13 14:19:31.092472 kernel: Performance Events: Fam17h+ core perfctr, AMD PMU driver. Dec 13 14:19:31.092479 kernel: ... version: 0 Dec 13 14:19:31.092486 kernel: ... bit width: 48 Dec 13 14:19:31.092493 kernel: ... generic registers: 6 Dec 13 14:19:31.092502 kernel: ... value mask: 0000ffffffffffff Dec 13 14:19:31.092513 kernel: ... max period: 00007fffffffffff Dec 13 14:19:31.092522 kernel: ... fixed-purpose events: 0 Dec 13 14:19:31.092531 kernel: ... event mask: 000000000000003f Dec 13 14:19:31.092538 kernel: signal: max sigframe size: 1776 Dec 13 14:19:31.092547 kernel: rcu: Hierarchical SRCU implementation. Dec 13 14:19:31.092556 kernel: smp: Bringing up secondary CPUs ... Dec 13 14:19:31.092565 kernel: x86: Booting SMP configuration: Dec 13 14:19:31.092572 kernel: .... node #0, CPUs: #1 Dec 13 14:19:31.092579 kernel: kvm-clock: cpu 1, msr 8a19a041, secondary cpu clock Dec 13 14:19:31.092586 kernel: kvm-guest: setup async PF for cpu 1 Dec 13 14:19:31.092594 kernel: kvm-guest: stealtime: cpu 1, msr 9a49c0c0 Dec 13 14:19:31.092601 kernel: #2 Dec 13 14:19:31.092609 kernel: kvm-clock: cpu 2, msr 8a19a081, secondary cpu clock Dec 13 14:19:31.092615 kernel: kvm-guest: setup async PF for cpu 2 Dec 13 14:19:31.092622 kernel: kvm-guest: stealtime: cpu 2, msr 9a51c0c0 Dec 13 14:19:31.092630 kernel: #3 Dec 13 14:19:31.092639 kernel: kvm-clock: cpu 3, msr 8a19a0c1, secondary cpu clock Dec 13 14:19:31.092648 kernel: kvm-guest: setup async PF for cpu 3 Dec 13 14:19:31.092660 kernel: kvm-guest: stealtime: cpu 3, msr 9a59c0c0 Dec 13 14:19:31.092672 kernel: smp: Brought up 1 node, 4 CPUs Dec 13 14:19:31.092681 kernel: smpboot: Max logical packages: 1 Dec 13 14:19:31.092690 kernel: smpboot: Total of 4 processors activated (22357.98 BogoMIPS) Dec 13 14:19:31.092698 kernel: devtmpfs: initialized Dec 13 14:19:31.092705 kernel: x86/mm: Memory block size: 128MB Dec 13 14:19:31.092713 kernel: clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 1911260446275000 ns Dec 13 14:19:31.092720 kernel: futex hash table entries: 1024 (order: 4, 65536 bytes, linear) Dec 13 14:19:31.092727 kernel: pinctrl core: initialized pinctrl subsystem Dec 13 14:19:31.092734 kernel: NET: Registered PF_NETLINK/PF_ROUTE protocol family Dec 13 14:19:31.092745 kernel: audit: initializing netlink subsys (disabled) Dec 13 14:19:31.092752 kernel: audit: type=2000 audit(1734099571.311:1): state=initialized audit_enabled=0 res=1 Dec 13 14:19:31.092759 kernel: thermal_sys: Registered thermal governor 'step_wise' Dec 13 14:19:31.092766 kernel: thermal_sys: Registered thermal governor 'user_space' Dec 13 14:19:31.092773 kernel: cpuidle: using governor menu Dec 13 14:19:31.092780 kernel: ACPI: bus type PCI registered Dec 13 14:19:31.092787 kernel: acpiphp: ACPI Hot Plug PCI Controller Driver version: 0.5 Dec 13 14:19:31.092794 kernel: dca service started, version 1.12.1 Dec 13 14:19:31.092801 kernel: PCI: MMCONFIG for domain 0000 [bus 00-ff] at [mem 0xb0000000-0xbfffffff] (base 0xb0000000) Dec 13 14:19:31.092833 kernel: PCI: MMCONFIG at [mem 0xb0000000-0xbfffffff] reserved in E820 Dec 13 14:19:31.092840 kernel: PCI: Using configuration type 1 for base access Dec 13 14:19:31.092847 kernel: kprobes: kprobe jump-optimization is enabled. All kprobes are optimized if possible. Dec 13 14:19:31.092854 kernel: HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages Dec 13 14:19:31.092861 kernel: HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages Dec 13 14:19:31.092868 kernel: ACPI: Added _OSI(Module Device) Dec 13 14:19:31.092875 kernel: ACPI: Added _OSI(Processor Device) Dec 13 14:19:31.092882 kernel: ACPI: Added _OSI(3.0 _SCP Extensions) Dec 13 14:19:31.092889 kernel: ACPI: Added _OSI(Processor Aggregator Device) Dec 13 14:19:31.092897 kernel: ACPI: Added _OSI(Linux-Dell-Video) Dec 13 14:19:31.092904 kernel: ACPI: Added _OSI(Linux-Lenovo-NV-HDMI-Audio) Dec 13 14:19:31.092911 kernel: ACPI: Added _OSI(Linux-HPI-Hybrid-Graphics) Dec 13 14:19:31.092918 kernel: ACPI: 1 ACPI AML tables successfully acquired and loaded Dec 13 14:19:31.092925 kernel: ACPI: Interpreter enabled Dec 13 14:19:31.092932 kernel: ACPI: PM: (supports S0 S3 S5) Dec 13 14:19:31.092939 kernel: ACPI: Using IOAPIC for interrupt routing Dec 13 14:19:31.092946 kernel: PCI: Using host bridge windows from ACPI; if necessary, use "pci=nocrs" and report a bug Dec 13 14:19:31.092953 kernel: ACPI: Enabled 2 GPEs in block 00 to 3F Dec 13 14:19:31.092967 kernel: ACPI: PCI Root Bridge [PCI0] (domain 0000 [bus 00-ff]) Dec 13 14:19:31.093132 kernel: acpi PNP0A08:00: _OSC: OS supports [ExtendedConfig ASPM ClockPM Segments MSI HPX-Type3] Dec 13 14:19:31.093230 kernel: acpi PNP0A08:00: _OSC: platform does not support [PCIeHotplug LTR] Dec 13 14:19:31.093338 kernel: acpi PNP0A08:00: _OSC: OS now controls [PME AER PCIeCapability] Dec 13 14:19:31.093349 kernel: PCI host bridge to bus 0000:00 Dec 13 14:19:31.093444 kernel: pci_bus 0000:00: root bus resource [io 0x0000-0x0cf7 window] Dec 13 14:19:31.093526 kernel: pci_bus 0000:00: root bus resource [io 0x0d00-0xffff window] Dec 13 14:19:31.094997 kernel: pci_bus 0000:00: root bus resource [mem 0x000a0000-0x000bffff window] Dec 13 14:19:31.095121 kernel: pci_bus 0000:00: root bus resource [mem 0x9d000000-0xafffffff window] Dec 13 14:19:31.095225 kernel: pci_bus 0000:00: root bus resource [mem 0xc0000000-0xfebfffff window] Dec 13 14:19:31.095311 kernel: pci_bus 0000:00: root bus resource [mem 0x100000000-0x8ffffffff window] Dec 13 14:19:31.095398 kernel: pci_bus 0000:00: root bus resource [bus 00-ff] Dec 13 14:19:31.095531 kernel: pci 0000:00:00.0: [8086:29c0] type 00 class 0x060000 Dec 13 14:19:31.095660 kernel: pci 0000:00:01.0: [1234:1111] type 00 class 0x030000 Dec 13 14:19:31.095763 kernel: pci 0000:00:01.0: reg 0x10: [mem 0xfd000000-0xfdffffff pref] Dec 13 14:19:31.095883 kernel: pci 0000:00:01.0: reg 0x18: [mem 0xfebd0000-0xfebd0fff] Dec 13 14:19:31.095980 kernel: pci 0000:00:01.0: reg 0x30: [mem 0xfebc0000-0xfebcffff pref] Dec 13 14:19:31.096078 kernel: pci 0000:00:01.0: Video device with shadowed ROM at [mem 0x000c0000-0x000dffff] Dec 13 14:19:31.096206 kernel: pci 0000:00:02.0: [1af4:1005] type 00 class 0x00ff00 Dec 13 14:19:31.096299 kernel: pci 0000:00:02.0: reg 0x10: [io 0xc0c0-0xc0df] Dec 13 14:19:31.096401 kernel: pci 0000:00:02.0: reg 0x14: [mem 0xfebd1000-0xfebd1fff] Dec 13 14:19:31.096496 kernel: pci 0000:00:02.0: reg 0x20: [mem 0xfe000000-0xfe003fff 64bit pref] Dec 13 14:19:31.096607 kernel: pci 0000:00:03.0: [1af4:1001] type 00 class 0x010000 Dec 13 14:19:31.096701 kernel: pci 0000:00:03.0: reg 0x10: [io 0xc000-0xc07f] Dec 13 14:19:31.096795 kernel: pci 0000:00:03.0: reg 0x14: [mem 0xfebd2000-0xfebd2fff] Dec 13 14:19:31.096907 kernel: pci 0000:00:03.0: reg 0x20: [mem 0xfe004000-0xfe007fff 64bit pref] Dec 13 14:19:31.097013 kernel: pci 0000:00:04.0: [1af4:1000] type 00 class 0x020000 Dec 13 14:19:31.097122 kernel: pci 0000:00:04.0: reg 0x10: [io 0xc0e0-0xc0ff] Dec 13 14:19:31.097233 kernel: pci 0000:00:04.0: reg 0x14: [mem 0xfebd3000-0xfebd3fff] Dec 13 14:19:31.097326 kernel: pci 0000:00:04.0: reg 0x20: [mem 0xfe008000-0xfe00bfff 64bit pref] Dec 13 14:19:31.097427 kernel: pci 0000:00:04.0: reg 0x30: [mem 0xfeb80000-0xfebbffff pref] Dec 13 14:19:31.097543 kernel: pci 0000:00:1f.0: [8086:2918] type 00 class 0x060100 Dec 13 14:19:31.097643 kernel: pci 0000:00:1f.0: quirk: [io 0x0600-0x067f] claimed by ICH6 ACPI/GPIO/TCO Dec 13 14:19:31.097758 kernel: pci 0000:00:1f.2: [8086:2922] type 00 class 0x010601 Dec 13 14:19:31.097878 kernel: pci 0000:00:1f.2: reg 0x20: [io 0xc100-0xc11f] Dec 13 14:19:31.097977 kernel: pci 0000:00:1f.2: reg 0x24: [mem 0xfebd4000-0xfebd4fff] Dec 13 14:19:31.098094 kernel: pci 0000:00:1f.3: [8086:2930] type 00 class 0x0c0500 Dec 13 14:19:31.098207 kernel: pci 0000:00:1f.3: reg 0x20: [io 0x0700-0x073f] Dec 13 14:19:31.098222 kernel: ACPI: PCI: Interrupt link LNKA configured for IRQ 10 Dec 13 14:19:31.098232 kernel: ACPI: PCI: Interrupt link LNKB configured for IRQ 10 Dec 13 14:19:31.098242 kernel: ACPI: PCI: Interrupt link LNKC configured for IRQ 11 Dec 13 14:19:31.098256 kernel: ACPI: PCI: Interrupt link LNKD configured for IRQ 11 Dec 13 14:19:31.098265 kernel: ACPI: PCI: Interrupt link LNKE configured for IRQ 10 Dec 13 14:19:31.098275 kernel: ACPI: PCI: Interrupt link LNKF configured for IRQ 10 Dec 13 14:19:31.098285 kernel: ACPI: PCI: Interrupt link LNKG configured for IRQ 11 Dec 13 14:19:31.098295 kernel: ACPI: PCI: Interrupt link LNKH configured for IRQ 11 Dec 13 14:19:31.098305 kernel: ACPI: PCI: Interrupt link GSIA configured for IRQ 16 Dec 13 14:19:31.098314 kernel: ACPI: PCI: Interrupt link GSIB configured for IRQ 17 Dec 13 14:19:31.098324 kernel: ACPI: PCI: Interrupt link GSIC configured for IRQ 18 Dec 13 14:19:31.098334 kernel: ACPI: PCI: Interrupt link GSID configured for IRQ 19 Dec 13 14:19:31.098346 kernel: ACPI: PCI: Interrupt link GSIE configured for IRQ 20 Dec 13 14:19:31.098356 kernel: ACPI: PCI: Interrupt link GSIF configured for IRQ 21 Dec 13 14:19:31.098365 kernel: ACPI: PCI: Interrupt link GSIG configured for IRQ 22 Dec 13 14:19:31.098376 kernel: ACPI: PCI: Interrupt link GSIH configured for IRQ 23 Dec 13 14:19:31.098386 kernel: iommu: Default domain type: Translated Dec 13 14:19:31.098396 kernel: iommu: DMA domain TLB invalidation policy: lazy mode Dec 13 14:19:31.098504 kernel: pci 0000:00:01.0: vgaarb: setting as boot VGA device Dec 13 14:19:31.098600 kernel: pci 0000:00:01.0: vgaarb: VGA device added: decodes=io+mem,owns=io+mem,locks=none Dec 13 14:19:31.098700 kernel: pci 0000:00:01.0: vgaarb: bridge control possible Dec 13 14:19:31.098713 kernel: vgaarb: loaded Dec 13 14:19:31.098723 kernel: pps_core: LinuxPPS API ver. 1 registered Dec 13 14:19:31.098733 kernel: pps_core: Software ver. 5.3.6 - Copyright 2005-2007 Rodolfo Giometti Dec 13 14:19:31.098743 kernel: PTP clock support registered Dec 13 14:19:31.098753 kernel: PCI: Using ACPI for IRQ routing Dec 13 14:19:31.098762 kernel: PCI: pci_cache_line_size set to 64 bytes Dec 13 14:19:31.098772 kernel: e820: reserve RAM buffer [mem 0x0009fc00-0x0009ffff] Dec 13 14:19:31.098782 kernel: e820: reserve RAM buffer [mem 0x9cfdc000-0x9fffffff] Dec 13 14:19:31.098800 kernel: hpet0: at MMIO 0xfed00000, IRQs 2, 8, 0 Dec 13 14:19:31.098823 kernel: hpet0: 3 comparators, 64-bit 100.000000 MHz counter Dec 13 14:19:31.098834 kernel: clocksource: Switched to clocksource kvm-clock Dec 13 14:19:31.098844 kernel: VFS: Disk quotas dquot_6.6.0 Dec 13 14:19:31.098854 kernel: VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) Dec 13 14:19:31.098864 kernel: pnp: PnP ACPI init Dec 13 14:19:31.098990 kernel: system 00:05: [mem 0xb0000000-0xbfffffff window] has been reserved Dec 13 14:19:31.099005 kernel: pnp: PnP ACPI: found 6 devices Dec 13 14:19:31.099019 kernel: clocksource: acpi_pm: mask: 0xffffff max_cycles: 0xffffff, max_idle_ns: 2085701024 ns Dec 13 14:19:31.099029 kernel: NET: Registered PF_INET protocol family Dec 13 14:19:31.099039 kernel: IP idents hash table entries: 65536 (order: 7, 524288 bytes, linear) Dec 13 14:19:31.099049 kernel: tcp_listen_portaddr_hash hash table entries: 2048 (order: 3, 32768 bytes, linear) Dec 13 14:19:31.099059 kernel: Table-perturb hash table entries: 65536 (order: 6, 262144 bytes, linear) Dec 13 14:19:31.099069 kernel: TCP established hash table entries: 32768 (order: 6, 262144 bytes, linear) Dec 13 14:19:31.099078 kernel: TCP bind hash table entries: 32768 (order: 7, 524288 bytes, linear) Dec 13 14:19:31.099088 kernel: TCP: Hash tables configured (established 32768 bind 32768) Dec 13 14:19:31.099098 kernel: UDP hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 13 14:19:31.099110 kernel: UDP-Lite hash table entries: 2048 (order: 4, 65536 bytes, linear) Dec 13 14:19:31.099120 kernel: NET: Registered PF_UNIX/PF_LOCAL protocol family Dec 13 14:19:31.099129 kernel: NET: Registered PF_XDP protocol family Dec 13 14:19:31.099235 kernel: pci_bus 0000:00: resource 4 [io 0x0000-0x0cf7 window] Dec 13 14:19:31.099326 kernel: pci_bus 0000:00: resource 5 [io 0x0d00-0xffff window] Dec 13 14:19:31.099413 kernel: pci_bus 0000:00: resource 6 [mem 0x000a0000-0x000bffff window] Dec 13 14:19:31.099518 kernel: pci_bus 0000:00: resource 7 [mem 0x9d000000-0xafffffff window] Dec 13 14:19:31.099611 kernel: pci_bus 0000:00: resource 8 [mem 0xc0000000-0xfebfffff window] Dec 13 14:19:31.099732 kernel: pci_bus 0000:00: resource 9 [mem 0x100000000-0x8ffffffff window] Dec 13 14:19:31.099752 kernel: PCI: CLS 0 bytes, default 64 Dec 13 14:19:31.099762 kernel: Initialise system trusted keyrings Dec 13 14:19:31.099771 kernel: workingset: timestamp_bits=39 max_order=20 bucket_order=0 Dec 13 14:19:31.099790 kernel: Key type asymmetric registered Dec 13 14:19:31.099807 kernel: Asymmetric key parser 'x509' registered Dec 13 14:19:31.099830 kernel: Block layer SCSI generic (bsg) driver version 0.4 loaded (major 249) Dec 13 14:19:31.099840 kernel: io scheduler mq-deadline registered Dec 13 14:19:31.099850 kernel: io scheduler kyber registered Dec 13 14:19:31.099859 kernel: io scheduler bfq registered Dec 13 14:19:31.099878 kernel: ioatdma: Intel(R) QuickData Technology Driver 5.00 Dec 13 14:19:31.099888 kernel: ACPI: \_SB_.GSIG: Enabled at IRQ 22 Dec 13 14:19:31.099898 kernel: ACPI: \_SB_.GSIH: Enabled at IRQ 23 Dec 13 14:19:31.099908 kernel: ACPI: \_SB_.GSIE: Enabled at IRQ 20 Dec 13 14:19:31.099928 kernel: Serial: 8250/16550 driver, 4 ports, IRQ sharing enabled Dec 13 14:19:31.099944 kernel: 00:03: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a 16550A Dec 13 14:19:31.099954 kernel: i8042: PNP: PS/2 Controller [PNP0303:KBD,PNP0f13:MOU] at 0x60,0x64 irq 1,12 Dec 13 14:19:31.099964 kernel: serio: i8042 KBD port at 0x60,0x64 irq 1 Dec 13 14:19:31.099973 kernel: serio: i8042 AUX port at 0x60,0x64 irq 12 Dec 13 14:19:31.100157 kernel: rtc_cmos 00:04: RTC can wake from S4 Dec 13 14:19:31.100174 kernel: input: AT Translated Set 2 keyboard as /devices/platform/i8042/serio0/input/input0 Dec 13 14:19:31.100291 kernel: rtc_cmos 00:04: registered as rtc0 Dec 13 14:19:31.100404 kernel: rtc_cmos 00:04: setting system clock to 2024-12-13T14:19:30 UTC (1734099570) Dec 13 14:19:31.100513 kernel: rtc_cmos 00:04: alarms up to one day, y3k, 242 bytes nvram, hpet irqs Dec 13 14:19:31.100528 kernel: NET: Registered PF_INET6 protocol family Dec 13 14:19:31.100537 kernel: Segment Routing with IPv6 Dec 13 14:19:31.100547 kernel: In-situ OAM (IOAM) with IPv6 Dec 13 14:19:31.100566 kernel: NET: Registered PF_PACKET protocol family Dec 13 14:19:31.100575 kernel: Key type dns_resolver registered Dec 13 14:19:31.100585 kernel: IPI shorthand broadcast: enabled Dec 13 14:19:31.100595 kernel: sched_clock: Marking stable (499112818, 122889148)->(700937648, -78935682) Dec 13 14:19:31.100621 kernel: registered taskstats version 1 Dec 13 14:19:31.100631 kernel: Loading compiled-in X.509 certificates Dec 13 14:19:31.100641 kernel: Loaded X.509 cert 'Kinvolk GmbH: Module signing key for 5.15.173-flatcar: e1d88c9e01f5bb2adeb5b99325e46e5ca8dff115' Dec 13 14:19:31.100651 kernel: Key type .fscrypt registered Dec 13 14:19:31.100660 kernel: Key type fscrypt-provisioning registered Dec 13 14:19:31.100674 kernel: ima: No TPM chip found, activating TPM-bypass! Dec 13 14:19:31.100683 kernel: ima: Allocated hash algorithm: sha1 Dec 13 14:19:31.100707 kernel: ima: No architecture policies found Dec 13 14:19:31.100718 kernel: clk: Disabling unused clocks Dec 13 14:19:31.100727 kernel: Freeing unused kernel image (initmem) memory: 47472K Dec 13 14:19:31.100737 kernel: Write protecting the kernel read-only data: 28672k Dec 13 14:19:31.100747 kernel: Freeing unused kernel image (text/rodata gap) memory: 2040K Dec 13 14:19:31.100757 kernel: Freeing unused kernel image (rodata/data gap) memory: 620K Dec 13 14:19:31.100769 kernel: Run /init as init process Dec 13 14:19:31.100795 kernel: with arguments: Dec 13 14:19:31.100805 kernel: /init Dec 13 14:19:31.100829 kernel: with environment: Dec 13 14:19:31.100838 kernel: HOME=/ Dec 13 14:19:31.100848 kernel: TERM=linux Dec 13 14:19:31.100858 kernel: BOOT_IMAGE=/flatcar/vmlinuz-a Dec 13 14:19:31.100887 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Dec 13 14:19:31.100904 systemd[1]: Detected virtualization kvm. Dec 13 14:19:31.100928 systemd[1]: Detected architecture x86-64. Dec 13 14:19:31.100940 systemd[1]: Running in initrd. Dec 13 14:19:31.100950 systemd[1]: No hostname configured, using default hostname. Dec 13 14:19:31.100960 systemd[1]: Hostname set to . Dec 13 14:19:31.100971 systemd[1]: Initializing machine ID from VM UUID. Dec 13 14:19:31.100996 systemd[1]: Queued start job for default target initrd.target. Dec 13 14:19:31.101007 systemd[1]: Started systemd-ask-password-console.path. Dec 13 14:19:31.101017 systemd[1]: Reached target cryptsetup.target. Dec 13 14:19:31.101040 systemd[1]: Reached target paths.target. Dec 13 14:19:31.101075 systemd[1]: Reached target slices.target. Dec 13 14:19:31.101090 systemd[1]: Reached target swap.target. Dec 13 14:19:31.101110 systemd[1]: Reached target timers.target. Dec 13 14:19:31.101127 systemd[1]: Listening on iscsid.socket. Dec 13 14:19:31.101155 systemd[1]: Listening on iscsiuio.socket. Dec 13 14:19:31.101165 systemd[1]: Listening on systemd-journald-audit.socket. Dec 13 14:19:31.101190 systemd[1]: Listening on systemd-journald-dev-log.socket. Dec 13 14:19:31.101203 systemd[1]: Listening on systemd-journald.socket. Dec 13 14:19:31.101292 systemd[1]: Listening on systemd-networkd.socket. Dec 13 14:19:31.101304 systemd[1]: Listening on systemd-udevd-control.socket. Dec 13 14:19:31.101315 systemd[1]: Listening on systemd-udevd-kernel.socket. Dec 13 14:19:31.101348 systemd[1]: Reached target sockets.target. Dec 13 14:19:31.101359 systemd[1]: Starting kmod-static-nodes.service... Dec 13 14:19:31.101372 systemd[1]: Finished network-cleanup.service. Dec 13 14:19:31.101383 systemd[1]: Starting systemd-fsck-usr.service... Dec 13 14:19:31.101404 systemd[1]: Starting systemd-journald.service... Dec 13 14:19:31.101421 systemd[1]: Starting systemd-modules-load.service... Dec 13 14:19:31.101432 systemd[1]: Starting systemd-resolved.service... Dec 13 14:19:31.101442 systemd[1]: Starting systemd-vconsole-setup.service... Dec 13 14:19:31.101453 systemd[1]: Finished kmod-static-nodes.service. Dec 13 14:19:31.101478 systemd[1]: Finished systemd-fsck-usr.service. Dec 13 14:19:31.101490 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Dec 13 14:19:31.101504 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Dec 13 14:19:31.101527 systemd-journald[197]: Journal started Dec 13 14:19:31.101611 systemd-journald[197]: Runtime Journal (/run/log/journal/218af52a80cf4cffa131177f8a944895) is 6.0M, max 48.5M, 42.5M free. Dec 13 14:19:31.086445 systemd-modules-load[198]: Inserted module 'overlay' Dec 13 14:19:31.135728 kernel: bridge: filtering via arp/ip/ip6tables is no longer available by default. Update your scripts to load br_netfilter if you need this. Dec 13 14:19:31.135758 kernel: Bridge firewalling registered Dec 13 14:19:31.135768 systemd[1]: Started systemd-journald.service. Dec 13 14:19:31.135781 kernel: audit: type=1130 audit(1734099571.129:2): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.129000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.109669 systemd-resolved[199]: Positive Trust Anchors: Dec 13 14:19:31.135000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.109677 systemd-resolved[199]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 14:19:31.145287 kernel: audit: type=1130 audit(1734099571.135:3): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.145306 kernel: audit: type=1130 audit(1734099571.139:4): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.139000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.109704 systemd-resolved[199]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Dec 13 14:19:31.150079 kernel: audit: type=1130 audit(1734099571.144:5): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.144000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.112255 systemd-resolved[199]: Defaulting to hostname 'linux'. Dec 13 14:19:31.123449 systemd-modules-load[198]: Inserted module 'br_netfilter' Dec 13 14:19:31.136704 systemd[1]: Started systemd-resolved.service. Dec 13 14:19:31.140641 systemd[1]: Finished systemd-vconsole-setup.service. Dec 13 14:19:31.145862 systemd[1]: Reached target nss-lookup.target. Dec 13 14:19:31.156148 systemd[1]: Starting dracut-cmdline-ask.service... Dec 13 14:19:31.167880 kernel: SCSI subsystem initialized Dec 13 14:19:31.172102 systemd[1]: Finished dracut-cmdline-ask.service. Dec 13 14:19:31.177022 kernel: audit: type=1130 audit(1734099571.171:6): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.171000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.177026 systemd[1]: Starting dracut-cmdline.service... Dec 13 14:19:31.183724 kernel: device-mapper: core: CONFIG_IMA_DISABLE_HTABLE is disabled. Duplicate IMA measurements will not be recorded in the IMA log. Dec 13 14:19:31.183856 kernel: device-mapper: uevent: version 1.0.3 Dec 13 14:19:31.183885 kernel: device-mapper: ioctl: 4.45.0-ioctl (2021-03-22) initialised: dm-devel@redhat.com Dec 13 14:19:31.188383 systemd-modules-load[198]: Inserted module 'dm_multipath' Dec 13 14:19:31.189262 systemd[1]: Finished systemd-modules-load.service. Dec 13 14:19:31.194323 kernel: audit: type=1130 audit(1734099571.189:7): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.189000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.190945 systemd[1]: Starting systemd-sysctl.service... Dec 13 14:19:31.196108 dracut-cmdline[216]: dracut-dracut-053 Dec 13 14:19:31.199383 dracut-cmdline[216]: Using kernel command line parameters: rd.driver.pre=btrfs rootflags=rw mount.usrflags=ro BOOT_IMAGE=/flatcar/vmlinuz-a mount.usr=/dev/mapper/usr verity.usr=PARTUUID=7130c94a-213a-4e5a-8e26-6cce9662f132 rootflags=rw mount.usrflags=ro consoleblank=0 root=LABEL=ROOT console=ttyS0,115200 flatcar.first_boot=detected verity.usrhash=8c474c3ec361ec863adbecaa85281a726e1b53f7863ecc4742be8c5f6d02a66e Dec 13 14:19:31.202951 systemd[1]: Finished systemd-sysctl.service. Dec 13 14:19:31.210262 kernel: audit: type=1130 audit(1734099571.204:8): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.204000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.274875 kernel: Loading iSCSI transport class v2.0-870. Dec 13 14:19:31.292855 kernel: iscsi: registered transport (tcp) Dec 13 14:19:31.317848 kernel: iscsi: registered transport (qla4xxx) Dec 13 14:19:31.317920 kernel: QLogic iSCSI HBA Driver Dec 13 14:19:31.353779 systemd[1]: Finished dracut-cmdline.service. Dec 13 14:19:31.359403 kernel: audit: type=1130 audit(1734099571.353:9): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.353000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.359526 systemd[1]: Starting dracut-pre-udev.service... Dec 13 14:19:31.420876 kernel: raid6: avx2x4 gen() 21706 MB/s Dec 13 14:19:31.437873 kernel: raid6: avx2x4 xor() 5753 MB/s Dec 13 14:19:31.454885 kernel: raid6: avx2x2 gen() 19019 MB/s Dec 13 14:19:31.471872 kernel: raid6: avx2x2 xor() 12304 MB/s Dec 13 14:19:31.488944 kernel: raid6: avx2x1 gen() 18571 MB/s Dec 13 14:19:31.505874 kernel: raid6: avx2x1 xor() 10696 MB/s Dec 13 14:19:31.522861 kernel: raid6: sse2x4 gen() 9353 MB/s Dec 13 14:19:31.539861 kernel: raid6: sse2x4 xor() 4396 MB/s Dec 13 14:19:31.556854 kernel: raid6: sse2x2 gen() 9590 MB/s Dec 13 14:19:31.573863 kernel: raid6: sse2x2 xor() 6263 MB/s Dec 13 14:19:31.590877 kernel: raid6: sse2x1 gen() 9618 MB/s Dec 13 14:19:31.609601 kernel: raid6: sse2x1 xor() 5441 MB/s Dec 13 14:19:31.609687 kernel: raid6: using algorithm avx2x4 gen() 21706 MB/s Dec 13 14:19:31.609701 kernel: raid6: .... xor() 5753 MB/s, rmw enabled Dec 13 14:19:31.610436 kernel: raid6: using avx2x2 recovery algorithm Dec 13 14:19:31.624847 kernel: xor: automatically using best checksumming function avx Dec 13 14:19:31.722845 kernel: Btrfs loaded, crc32c=crc32c-intel, zoned=no, fsverity=no Dec 13 14:19:31.732097 systemd[1]: Finished dracut-pre-udev.service. Dec 13 14:19:31.731000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.732000 audit: BPF prog-id=7 op=LOAD Dec 13 14:19:31.735000 audit: BPF prog-id=8 op=LOAD Dec 13 14:19:31.736832 kernel: audit: type=1130 audit(1734099571.731:10): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.737343 systemd[1]: Starting systemd-udevd.service... Dec 13 14:19:31.750783 systemd-udevd[399]: Using default interface naming scheme 'v252'. Dec 13 14:19:31.755337 systemd[1]: Started systemd-udevd.service. Dec 13 14:19:31.756000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.759055 systemd[1]: Starting dracut-pre-trigger.service... Dec 13 14:19:31.772275 dracut-pre-trigger[411]: rd.md=0: removing MD RAID activation Dec 13 14:19:31.805342 systemd[1]: Finished dracut-pre-trigger.service. Dec 13 14:19:31.806000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.808841 systemd[1]: Starting systemd-udev-trigger.service... Dec 13 14:19:31.858468 systemd[1]: Finished systemd-udev-trigger.service. Dec 13 14:19:31.859000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:31.890110 kernel: virtio_blk virtio1: [vda] 19775488 512-byte logical blocks (10.1 GB/9.43 GiB) Dec 13 14:19:31.897888 kernel: GPT:Primary header thinks Alt. header is not at the end of the disk. Dec 13 14:19:31.897911 kernel: GPT:9289727 != 19775487 Dec 13 14:19:31.897923 kernel: GPT:Alternate GPT header not at the end of the disk. Dec 13 14:19:31.897934 kernel: GPT:9289727 != 19775487 Dec 13 14:19:31.897944 kernel: GPT: Use GNU Parted to correct GPT errors. Dec 13 14:19:31.897955 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 14:19:31.900836 kernel: cryptd: max_cpu_qlen set to 1000 Dec 13 14:19:31.905862 kernel: libata version 3.00 loaded. Dec 13 14:19:31.914625 kernel: AVX2 version of gcm_enc/dec engaged. Dec 13 14:19:31.914680 kernel: AES CTR mode by8 optimization enabled Dec 13 14:19:31.914694 kernel: ahci 0000:00:1f.2: version 3.0 Dec 13 14:19:31.978576 kernel: ACPI: \_SB_.GSIA: Enabled at IRQ 16 Dec 13 14:19:31.978594 kernel: ahci 0000:00:1f.2: AHCI 0001.0000 32 slots 6 ports 1.5 Gbps 0x3f impl SATA mode Dec 13 14:19:31.978702 kernel: ahci 0000:00:1f.2: flags: 64bit ncq only Dec 13 14:19:31.978804 kernel: scsi host0: ahci Dec 13 14:19:31.978953 kernel: scsi host1: ahci Dec 13 14:19:31.979075 kernel: BTRFS: device label OEM devid 1 transid 12 /dev/vda6 scanned by (udev-worker) (450) Dec 13 14:19:31.979089 kernel: scsi host2: ahci Dec 13 14:19:31.979217 kernel: scsi host3: ahci Dec 13 14:19:31.979340 kernel: scsi host4: ahci Dec 13 14:19:31.979479 kernel: scsi host5: ahci Dec 13 14:19:31.979609 kernel: ata1: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4100 irq 34 Dec 13 14:19:31.979624 kernel: ata2: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4180 irq 34 Dec 13 14:19:31.979639 kernel: ata3: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4200 irq 34 Dec 13 14:19:31.979651 kernel: ata4: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4280 irq 34 Dec 13 14:19:31.979663 kernel: ata5: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4300 irq 34 Dec 13 14:19:31.979673 kernel: ata6: SATA max UDMA/133 abar m4096@0xfebd4000 port 0xfebd4380 irq 34 Dec 13 14:19:31.966827 systemd[1]: Found device dev-disk-by\x2dpartuuid-7130c94a\x2d213a\x2d4e5a\x2d8e26\x2d6cce9662f132.device. Dec 13 14:19:32.008552 systemd[1]: Found device dev-disk-by\x2dpartlabel-USR\x2dA.device. Dec 13 14:19:32.019479 systemd[1]: Found device dev-disk-by\x2dlabel-EFI\x2dSYSTEM.device. Dec 13 14:19:32.024705 systemd[1]: Found device dev-disk-by\x2dlabel-ROOT.device. Dec 13 14:19:32.030590 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Dec 13 14:19:32.033949 systemd[1]: Starting disk-uuid.service... Dec 13 14:19:32.045072 disk-uuid[541]: Primary Header is updated. Dec 13 14:19:32.045072 disk-uuid[541]: Secondary Entries is updated. Dec 13 14:19:32.045072 disk-uuid[541]: Secondary Header is updated. Dec 13 14:19:32.048857 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 14:19:32.292127 kernel: ata3: SATA link up 1.5 Gbps (SStatus 113 SControl 300) Dec 13 14:19:32.292225 kernel: ata6: SATA link down (SStatus 0 SControl 300) Dec 13 14:19:32.292236 kernel: ata4: SATA link down (SStatus 0 SControl 300) Dec 13 14:19:32.292245 kernel: ata2: SATA link down (SStatus 0 SControl 300) Dec 13 14:19:32.293839 kernel: ata1: SATA link down (SStatus 0 SControl 300) Dec 13 14:19:32.294860 kernel: ata3.00: ATAPI: QEMU DVD-ROM, 2.5+, max UDMA/100 Dec 13 14:19:32.296175 kernel: ata3.00: applying bridge limits Dec 13 14:19:32.296839 kernel: ata5: SATA link down (SStatus 0 SControl 300) Dec 13 14:19:32.297860 kernel: ata3.00: configured for UDMA/100 Dec 13 14:19:32.298839 kernel: scsi 2:0:0:0: CD-ROM QEMU QEMU DVD-ROM 2.5+ PQ: 0 ANSI: 5 Dec 13 14:19:32.329259 kernel: sr 2:0:0:0: [sr0] scsi3-mmc drive: 4x/4x cd/rw xa/form2 tray Dec 13 14:19:32.346944 kernel: cdrom: Uniform CD-ROM driver Revision: 3.20 Dec 13 14:19:32.346970 kernel: sr 2:0:0:0: Attached scsi CD-ROM sr0 Dec 13 14:19:33.057850 kernel: vda: vda1 vda2 vda3 vda4 vda6 vda7 vda9 Dec 13 14:19:33.058508 disk-uuid[543]: The operation has completed successfully. Dec 13 14:19:33.088023 systemd[1]: disk-uuid.service: Deactivated successfully. Dec 13 14:19:33.088000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.088000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=disk-uuid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.088142 systemd[1]: Finished disk-uuid.service. Dec 13 14:19:33.093021 systemd[1]: Starting verity-setup.service... Dec 13 14:19:33.110846 kernel: device-mapper: verity: sha256 using implementation "sha256-ni" Dec 13 14:19:33.165630 systemd[1]: Found device dev-mapper-usr.device. Dec 13 14:19:33.167965 systemd[1]: Mounting sysusr-usr.mount... Dec 13 14:19:33.169917 systemd[1]: Finished verity-setup.service. Dec 13 14:19:33.169000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.258873 kernel: EXT4-fs (dm-0): mounted filesystem without journal. Opts: norecovery. Quota mode: none. Dec 13 14:19:33.259024 systemd[1]: Mounted sysusr-usr.mount. Dec 13 14:19:33.259786 systemd[1]: afterburn-network-kargs.service was skipped because no trigger condition checks were met. Dec 13 14:19:33.260706 systemd[1]: Starting ignition-setup.service... Dec 13 14:19:33.262480 systemd[1]: Starting parse-ip-for-networkd.service... Dec 13 14:19:33.271556 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 14:19:33.271613 kernel: BTRFS info (device vda6): using free space tree Dec 13 14:19:33.271626 kernel: BTRFS info (device vda6): has skinny extents Dec 13 14:19:33.281658 systemd[1]: mnt-oem.mount: Deactivated successfully. Dec 13 14:19:33.299982 systemd[1]: Finished ignition-setup.service. Dec 13 14:19:33.300000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.302122 systemd[1]: Starting ignition-fetch-offline.service... Dec 13 14:19:33.341706 systemd[1]: Finished parse-ip-for-networkd.service. Dec 13 14:19:33.340000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.343000 audit: BPF prog-id=9 op=LOAD Dec 13 14:19:33.344285 systemd[1]: Starting systemd-networkd.service... Dec 13 14:19:33.424442 systemd-networkd[722]: lo: Link UP Dec 13 14:19:33.424458 systemd-networkd[722]: lo: Gained carrier Dec 13 14:19:33.424000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.425139 systemd-networkd[722]: Enumeration completed Dec 13 14:19:33.425300 systemd[1]: Started systemd-networkd.service. Dec 13 14:19:33.425718 systemd[1]: Reached target network.target. Dec 13 14:19:33.426440 systemd-networkd[722]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 14:19:33.427800 systemd-networkd[722]: eth0: Link UP Dec 13 14:19:33.427819 systemd-networkd[722]: eth0: Gained carrier Dec 13 14:19:33.431106 systemd[1]: Starting iscsiuio.service... Dec 13 14:19:33.471294 ignition[660]: Ignition 2.14.0 Dec 13 14:19:33.471309 ignition[660]: Stage: fetch-offline Dec 13 14:19:33.471396 ignition[660]: no configs at "/usr/lib/ignition/base.d" Dec 13 14:19:33.471405 ignition[660]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:19:33.471538 ignition[660]: parsed url from cmdline: "" Dec 13 14:19:33.471542 ignition[660]: no config URL provided Dec 13 14:19:33.471547 ignition[660]: reading system config file "/usr/lib/ignition/user.ign" Dec 13 14:19:33.471553 ignition[660]: no config at "/usr/lib/ignition/user.ign" Dec 13 14:19:33.471574 ignition[660]: op(1): [started] loading QEMU firmware config module Dec 13 14:19:33.471598 ignition[660]: op(1): executing: "modprobe" "qemu_fw_cfg" Dec 13 14:19:33.477738 ignition[660]: op(1): [finished] loading QEMU firmware config module Dec 13 14:19:33.478211 ignition[660]: parsing config with SHA512: c035a16a9dac59e17df80c0373fd5119c7c9e240794ac02ce27262301537332f669688395f1aa532464ce26485b9b55226ef293bf5d7901ffbe86bfe3f243e42 Dec 13 14:19:33.535756 unknown[660]: fetched base config from "system" Dec 13 14:19:33.536427 ignition[660]: fetch-offline: fetch-offline passed Dec 13 14:19:33.535768 unknown[660]: fetched user config from "qemu" Dec 13 14:19:33.537000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.536549 ignition[660]: Ignition finished successfully Dec 13 14:19:33.537822 systemd[1]: Finished ignition-fetch-offline.service. Dec 13 14:19:33.538458 systemd[1]: ignition-fetch.service was skipped because of an unmet condition check (ConditionPathExists=!/run/ignition.json). Dec 13 14:19:33.539534 systemd[1]: Starting ignition-kargs.service... Dec 13 14:19:33.548850 systemd[1]: Started iscsiuio.service. Dec 13 14:19:33.548000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.549961 systemd-networkd[722]: eth0: DHCPv4 address 10.0.0.34/16, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 13 14:19:33.551198 systemd[1]: Starting iscsid.service... Dec 13 14:19:33.552436 ignition[729]: Ignition 2.14.0 Dec 13 14:19:33.571000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.573000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.561027 systemd[1]: Started iscsid.service. Dec 13 14:19:33.574132 iscsid[736]: iscsid: can't open InitiatorName configuration file /etc/iscsi/initiatorname.iscsi Dec 13 14:19:33.574132 iscsid[736]: iscsid: Warning: InitiatorName file /etc/iscsi/initiatorname.iscsi does not exist or does not contain a properly formatted InitiatorName. If using software iscsi (iscsi_tcp or ib_iser) or partial offload (bnx2i or cxgbi iscsi), you may not be able to log Dec 13 14:19:33.574132 iscsid[736]: into or discover targets. Please create a file /etc/iscsi/initiatorname.iscsi that contains a sting with the format: InitiatorName=iqn.yyyy-mm.[:identifier]. Dec 13 14:19:33.574132 iscsid[736]: Example: InitiatorName=iqn.2001-04.com.redhat:fc6. Dec 13 14:19:33.574132 iscsid[736]: If using hardware iscsi like qla4xxx this message can be ignored. Dec 13 14:19:33.574132 iscsid[736]: iscsid: can't open InitiatorAlias configuration file /etc/iscsi/initiatorname.iscsi Dec 13 14:19:33.574132 iscsid[736]: iscsid: can't open iscsid.safe_logout configuration file /etc/iscsi/iscsid.conf Dec 13 14:19:33.589000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.552442 ignition[729]: Stage: kargs Dec 13 14:19:33.571818 systemd[1]: Finished ignition-kargs.service. Dec 13 14:19:33.552577 ignition[729]: no configs at "/usr/lib/ignition/base.d" Dec 13 14:19:33.574067 systemd[1]: Starting dracut-initqueue.service... Dec 13 14:19:33.552589 ignition[729]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:19:33.575661 systemd[1]: Starting ignition-disks.service... Dec 13 14:19:33.553626 ignition[729]: kargs: kargs passed Dec 13 14:19:33.587426 systemd[1]: Finished dracut-initqueue.service. Dec 13 14:19:33.553676 ignition[729]: Ignition finished successfully Dec 13 14:19:33.589643 systemd[1]: Reached target remote-fs-pre.target. Dec 13 14:19:33.592102 systemd[1]: Reached target remote-cryptsetup.target. Dec 13 14:19:33.593206 systemd[1]: Reached target remote-fs.target. Dec 13 14:19:33.595442 systemd[1]: Starting dracut-pre-mount.service... Dec 13 14:19:33.627264 systemd[1]: Finished dracut-pre-mount.service. Dec 13 14:19:33.630000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.632000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.631275 systemd[1]: Finished ignition-disks.service. Dec 13 14:19:33.629057 ignition[738]: Ignition 2.14.0 Dec 13 14:19:33.632723 systemd[1]: Reached target initrd-root-device.target. Dec 13 14:19:33.629066 ignition[738]: Stage: disks Dec 13 14:19:33.633750 systemd[1]: Reached target local-fs-pre.target. Dec 13 14:19:33.629186 ignition[738]: no configs at "/usr/lib/ignition/base.d" Dec 13 14:19:33.635535 systemd[1]: Reached target local-fs.target. Dec 13 14:19:33.629195 ignition[738]: no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:19:33.636420 systemd[1]: Reached target sysinit.target. Dec 13 14:19:33.630017 ignition[738]: disks: disks passed Dec 13 14:19:33.637304 systemd[1]: Reached target basic.target. Dec 13 14:19:33.630077 ignition[738]: Ignition finished successfully Dec 13 14:19:33.639254 systemd[1]: Starting systemd-fsck-root.service... Dec 13 14:19:33.654211 systemd-fsck[759]: ROOT: clean, 621/553520 files, 56021/553472 blocks Dec 13 14:19:33.692673 systemd[1]: Finished systemd-fsck-root.service. Dec 13 14:19:33.693000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.694954 systemd[1]: Mounting sysroot.mount... Dec 13 14:19:33.709848 kernel: EXT4-fs (vda9): mounted filesystem with ordered data mode. Opts: (null). Quota mode: none. Dec 13 14:19:33.709906 systemd[1]: Mounted sysroot.mount. Dec 13 14:19:33.710840 systemd[1]: Reached target initrd-root-fs.target. Dec 13 14:19:33.713680 systemd[1]: Mounting sysroot-usr.mount... Dec 13 14:19:33.714909 systemd[1]: flatcar-metadata-hostname.service was skipped because no trigger condition checks were met. Dec 13 14:19:33.714957 systemd[1]: ignition-remount-sysroot.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/sysroot). Dec 13 14:19:33.714995 systemd[1]: Reached target ignition-diskful.target. Dec 13 14:19:33.717426 systemd[1]: Mounted sysroot-usr.mount. Dec 13 14:19:33.719629 systemd[1]: Starting initrd-setup-root.service... Dec 13 14:19:33.725067 initrd-setup-root[769]: cut: /sysroot/etc/passwd: No such file or directory Dec 13 14:19:33.730142 initrd-setup-root[777]: cut: /sysroot/etc/group: No such file or directory Dec 13 14:19:33.733575 initrd-setup-root[785]: cut: /sysroot/etc/shadow: No such file or directory Dec 13 14:19:33.737803 initrd-setup-root[793]: cut: /sysroot/etc/gshadow: No such file or directory Dec 13 14:19:33.777490 systemd[1]: Finished initrd-setup-root.service. Dec 13 14:19:33.776000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.779530 systemd[1]: Starting ignition-mount.service... Dec 13 14:19:33.780938 systemd[1]: Starting sysroot-boot.service... Dec 13 14:19:33.816001 bash[810]: umount: /sysroot/usr/share/oem: not mounted. Dec 13 14:19:33.824197 systemd[1]: Finished sysroot-boot.service. Dec 13 14:19:33.823000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:33.831404 ignition[813]: INFO : Ignition 2.14.0 Dec 13 14:19:33.831404 ignition[813]: INFO : Stage: mount Dec 13 14:19:33.833395 ignition[813]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 14:19:33.833395 ignition[813]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:19:33.836822 ignition[813]: INFO : mount: mount passed Dec 13 14:19:33.837729 ignition[813]: INFO : Ignition finished successfully Dec 13 14:19:33.839578 systemd[1]: Finished ignition-mount.service. Dec 13 14:19:33.841000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:34.181521 systemd[1]: Mounting sysroot-usr-share-oem.mount... Dec 13 14:19:34.191867 kernel: BTRFS: device label OEM devid 1 transid 13 /dev/vda6 scanned by mount (820) Dec 13 14:19:34.194709 kernel: BTRFS info (device vda6): using crc32c (crc32c-intel) checksum algorithm Dec 13 14:19:34.194739 kernel: BTRFS info (device vda6): using free space tree Dec 13 14:19:34.194752 kernel: BTRFS info (device vda6): has skinny extents Dec 13 14:19:34.199677 systemd[1]: Mounted sysroot-usr-share-oem.mount. Dec 13 14:19:34.201723 systemd[1]: Starting ignition-files.service... Dec 13 14:19:34.220628 ignition[840]: INFO : Ignition 2.14.0 Dec 13 14:19:34.220628 ignition[840]: INFO : Stage: files Dec 13 14:19:34.222579 ignition[840]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 14:19:34.222579 ignition[840]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:19:34.225626 ignition[840]: DEBUG : files: compiled without relabeling support, skipping Dec 13 14:19:34.227903 ignition[840]: INFO : files: ensureUsers: op(1): [started] creating or modifying user "core" Dec 13 14:19:34.227903 ignition[840]: DEBUG : files: ensureUsers: op(1): executing: "usermod" "--root" "/sysroot" "core" Dec 13 14:19:34.233268 ignition[840]: INFO : files: ensureUsers: op(1): [finished] creating or modifying user "core" Dec 13 14:19:34.235168 ignition[840]: INFO : files: ensureUsers: op(2): [started] adding ssh keys to user "core" Dec 13 14:19:34.236886 ignition[840]: INFO : files: ensureUsers: op(2): [finished] adding ssh keys to user "core" Dec 13 14:19:34.236189 unknown[840]: wrote ssh authorized keys file for user: core Dec 13 14:19:34.239732 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(3): [started] writing file "/sysroot/home/core/install.sh" Dec 13 14:19:34.239732 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(3): [finished] writing file "/sysroot/home/core/install.sh" Dec 13 14:19:34.239732 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(4): [started] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 14:19:34.239732 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(4): [finished] writing file "/sysroot/etc/flatcar/update.conf" Dec 13 14:19:34.239732 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(5): [started] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 14:19:34.239732 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(5): [finished] writing link "/sysroot/etc/extensions/kubernetes.raw" -> "/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 14:19:34.239732 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(6): [started] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 14:19:34.239732 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET https://github.com/flatcar/sysext-bakery/releases/download/latest/kubernetes-v1.29.2-x86-64.raw: attempt #1 Dec 13 14:19:34.619895 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(6): GET result: OK Dec 13 14:19:34.627993 systemd-networkd[722]: eth0: Gained IPv6LL Dec 13 14:19:35.315747 ignition[840]: INFO : files: createFilesystemsFiles: createFiles: op(6): [finished] writing file "/sysroot/opt/extensions/kubernetes/kubernetes-v1.29.2-x86-64.raw" Dec 13 14:19:35.315747 ignition[840]: INFO : files: op(7): [started] processing unit "coreos-metadata.service" Dec 13 14:19:35.320868 ignition[840]: INFO : files: op(7): op(8): [started] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Dec 13 14:19:35.320868 ignition[840]: INFO : files: op(7): op(8): [finished] writing unit "coreos-metadata.service" at "/sysroot/etc/systemd/system/coreos-metadata.service" Dec 13 14:19:35.320868 ignition[840]: INFO : files: op(7): [finished] processing unit "coreos-metadata.service" Dec 13 14:19:35.320868 ignition[840]: INFO : files: op(9): [started] setting preset to disabled for "coreos-metadata.service" Dec 13 14:19:35.320868 ignition[840]: INFO : files: op(9): op(a): [started] removing enablement symlink(s) for "coreos-metadata.service" Dec 13 14:19:35.398961 ignition[840]: INFO : files: op(9): op(a): [finished] removing enablement symlink(s) for "coreos-metadata.service" Dec 13 14:19:35.400874 ignition[840]: INFO : files: op(9): [finished] setting preset to disabled for "coreos-metadata.service" Dec 13 14:19:35.400874 ignition[840]: INFO : files: createResultFile: createFiles: op(b): [started] writing file "/sysroot/etc/.ignition-result.json" Dec 13 14:19:35.400874 ignition[840]: INFO : files: createResultFile: createFiles: op(b): [finished] writing file "/sysroot/etc/.ignition-result.json" Dec 13 14:19:35.400874 ignition[840]: INFO : files: files passed Dec 13 14:19:35.400874 ignition[840]: INFO : Ignition finished successfully Dec 13 14:19:35.418486 systemd[1]: Finished ignition-files.service. Dec 13 14:19:35.425960 kernel: kauditd_printk_skb: 23 callbacks suppressed Dec 13 14:19:35.425994 kernel: audit: type=1130 audit(1734099575.418:34): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.418000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.419919 systemd[1]: Starting initrd-setup-root-after-ignition.service... Dec 13 14:19:35.425921 systemd[1]: torcx-profile-populate.service was skipped because of an unmet condition check (ConditionPathExists=/sysroot/etc/torcx/next-profile). Dec 13 14:19:35.430000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.431555 initrd-setup-root-after-ignition[864]: grep: /sysroot/usr/share/oem/oem-release: No such file or directory Dec 13 14:19:35.454159 kernel: audit: type=1130 audit(1734099575.430:35): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.453000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.426625 systemd[1]: Starting ignition-quench.service... Dec 13 14:19:35.462632 kernel: audit: type=1130 audit(1734099575.453:36): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.462659 kernel: audit: type=1131 audit(1734099575.453:37): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.453000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-quench comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.462848 initrd-setup-root-after-ignition[866]: grep: /sysroot/etc/flatcar/enabled-sysext.conf: No such file or directory Dec 13 14:19:35.428097 systemd[1]: Finished initrd-setup-root-after-ignition.service. Dec 13 14:19:35.431640 systemd[1]: ignition-quench.service: Deactivated successfully. Dec 13 14:19:35.431707 systemd[1]: Finished ignition-quench.service. Dec 13 14:19:35.454310 systemd[1]: Reached target ignition-complete.target. Dec 13 14:19:35.463668 systemd[1]: Starting initrd-parse-etc.service... Dec 13 14:19:35.479069 systemd[1]: initrd-parse-etc.service: Deactivated successfully. Dec 13 14:19:35.479160 systemd[1]: Finished initrd-parse-etc.service. Dec 13 14:19:35.498418 kernel: audit: type=1130 audit(1734099575.488:38): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.498450 kernel: audit: type=1131 audit(1734099575.488:39): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.488000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.488000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-parse-etc comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.489643 systemd[1]: Reached target initrd-fs.target. Dec 13 14:19:35.498482 systemd[1]: Reached target initrd.target. Dec 13 14:19:35.499564 systemd[1]: dracut-mount.service was skipped because no trigger condition checks were met. Dec 13 14:19:35.500857 systemd[1]: Starting dracut-pre-pivot.service... Dec 13 14:19:35.512928 systemd[1]: Finished dracut-pre-pivot.service. Dec 13 14:19:35.519000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.521890 systemd[1]: Starting initrd-cleanup.service... Dec 13 14:19:35.525495 kernel: audit: type=1130 audit(1734099575.519:40): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.532534 systemd[1]: Stopped target nss-lookup.target. Dec 13 14:19:35.533797 systemd[1]: Stopped target remote-cryptsetup.target. Dec 13 14:19:35.545175 systemd[1]: Stopped target timers.target. Dec 13 14:19:35.546837 systemd[1]: dracut-pre-pivot.service: Deactivated successfully. Dec 13 14:19:35.553663 kernel: audit: type=1131 audit(1734099575.548:41): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.548000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-pivot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.546969 systemd[1]: Stopped dracut-pre-pivot.service. Dec 13 14:19:35.548513 systemd[1]: Stopped target initrd.target. Dec 13 14:19:35.553789 systemd[1]: Stopped target basic.target. Dec 13 14:19:35.555484 systemd[1]: Stopped target ignition-complete.target. Dec 13 14:19:35.557239 systemd[1]: Stopped target ignition-diskful.target. Dec 13 14:19:35.559150 systemd[1]: Stopped target initrd-root-device.target. Dec 13 14:19:35.560856 systemd[1]: Stopped target remote-fs.target. Dec 13 14:19:35.562475 systemd[1]: Stopped target remote-fs-pre.target. Dec 13 14:19:35.564152 systemd[1]: Stopped target sysinit.target. Dec 13 14:19:35.565724 systemd[1]: Stopped target local-fs.target. Dec 13 14:19:35.567262 systemd[1]: Stopped target local-fs-pre.target. Dec 13 14:19:35.582978 systemd[1]: Stopped target swap.target. Dec 13 14:19:35.590933 kernel: audit: type=1131 audit(1734099575.585:42): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.585000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.584431 systemd[1]: dracut-pre-mount.service: Deactivated successfully. Dec 13 14:19:35.584580 systemd[1]: Stopped dracut-pre-mount.service. Dec 13 14:19:35.597794 kernel: audit: type=1131 audit(1734099575.592:43): pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.592000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-initqueue comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.586211 systemd[1]: Stopped target cryptsetup.target. Dec 13 14:19:35.597000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-fetch-offline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.590997 systemd[1]: dracut-initqueue.service: Deactivated successfully. Dec 13 14:19:35.591149 systemd[1]: Stopped dracut-initqueue.service. Dec 13 14:19:35.593002 systemd[1]: ignition-fetch-offline.service: Deactivated successfully. Dec 13 14:19:35.593104 systemd[1]: Stopped ignition-fetch-offline.service. Dec 13 14:19:35.597993 systemd[1]: Stopped target paths.target. Dec 13 14:19:35.599532 systemd[1]: systemd-ask-password-console.path: Deactivated successfully. Dec 13 14:19:35.602878 systemd[1]: Stopped systemd-ask-password-console.path. Dec 13 14:19:35.616021 systemd[1]: Stopped target slices.target. Dec 13 14:19:35.621000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root-after-ignition comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.617530 systemd[1]: Stopped target sockets.target. Dec 13 14:19:35.622000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-files comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.619532 systemd[1]: initrd-setup-root-after-ignition.service: Deactivated successfully. Dec 13 14:19:35.619671 systemd[1]: Stopped initrd-setup-root-after-ignition.service. Dec 13 14:19:35.627249 iscsid[736]: iscsid shutting down. Dec 13 14:19:35.621596 systemd[1]: ignition-files.service: Deactivated successfully. Dec 13 14:19:35.621711 systemd[1]: Stopped ignition-files.service. Dec 13 14:19:35.623962 systemd[1]: Stopping ignition-mount.service... Dec 13 14:19:35.625880 systemd[1]: Stopping iscsid.service... Dec 13 14:19:35.630337 systemd[1]: Stopping sysroot-boot.service... Dec 13 14:19:35.634865 ignition[881]: INFO : Ignition 2.14.0 Dec 13 14:19:35.634865 ignition[881]: INFO : Stage: umount Dec 13 14:19:35.634865 ignition[881]: INFO : no configs at "/usr/lib/ignition/base.d" Dec 13 14:19:35.634865 ignition[881]: INFO : no config dir at "/usr/lib/ignition/base.platform.d/qemu" Dec 13 14:19:35.634865 ignition[881]: INFO : umount: umount passed Dec 13 14:19:35.634865 ignition[881]: INFO : Ignition finished successfully Dec 13 14:19:35.635000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.633742 systemd[1]: systemd-udev-trigger.service: Deactivated successfully. Dec 13 14:19:35.633955 systemd[1]: Stopped systemd-udev-trigger.service. Dec 13 14:19:35.639141 systemd[1]: dracut-pre-trigger.service: Deactivated successfully. Dec 13 14:19:35.642225 systemd[1]: Stopped dracut-pre-trigger.service. Dec 13 14:19:35.660000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.663801 systemd[1]: sysroot-boot.mount: Deactivated successfully. Dec 13 14:19:35.665349 systemd[1]: iscsid.service: Deactivated successfully. Dec 13 14:19:35.666337 systemd[1]: Stopped iscsid.service. Dec 13 14:19:35.667000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsid comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.668261 systemd[1]: ignition-mount.service: Deactivated successfully. Dec 13 14:19:35.668336 systemd[1]: Stopped ignition-mount.service. Dec 13 14:19:35.670000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-mount comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.671544 systemd[1]: iscsid.socket: Deactivated successfully. Dec 13 14:19:35.672530 systemd[1]: Closed iscsid.socket. Dec 13 14:19:35.674229 systemd[1]: ignition-disks.service: Deactivated successfully. Dec 13 14:19:35.674000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-disks comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.674274 systemd[1]: Stopped ignition-disks.service. Dec 13 14:19:35.677000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-kargs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.676187 systemd[1]: ignition-kargs.service: Deactivated successfully. Dec 13 14:19:35.676220 systemd[1]: Stopped ignition-kargs.service. Dec 13 14:19:35.679000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=ignition-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.677869 systemd[1]: ignition-setup.service: Deactivated successfully. Dec 13 14:19:35.677902 systemd[1]: Stopped ignition-setup.service. Dec 13 14:19:35.681657 systemd[1]: Stopping iscsiuio.service... Dec 13 14:19:35.684129 systemd[1]: initrd-cleanup.service: Deactivated successfully. Dec 13 14:19:35.685179 systemd[1]: Finished initrd-cleanup.service. Dec 13 14:19:35.686000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.686000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.687272 systemd[1]: iscsiuio.service: Deactivated successfully. Dec 13 14:19:35.688198 systemd[1]: Stopped iscsiuio.service. Dec 13 14:19:35.688000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=iscsiuio comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.690499 systemd[1]: Stopped target network.target. Dec 13 14:19:35.692047 systemd[1]: iscsiuio.socket: Deactivated successfully. Dec 13 14:19:35.692096 systemd[1]: Closed iscsiuio.socket. Dec 13 14:19:35.694396 systemd[1]: Stopping systemd-networkd.service... Dec 13 14:19:35.696141 systemd[1]: Stopping systemd-resolved.service... Dec 13 14:19:35.699905 systemd-networkd[722]: eth0: DHCPv6 lease lost Dec 13 14:19:35.701109 systemd[1]: systemd-networkd.service: Deactivated successfully. Dec 13 14:19:35.700000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.701212 systemd[1]: Stopped systemd-networkd.service. Dec 13 14:19:35.701752 systemd[1]: systemd-networkd.socket: Deactivated successfully. Dec 13 14:19:35.701777 systemd[1]: Closed systemd-networkd.socket. Dec 13 14:19:35.704826 systemd[1]: Stopping network-cleanup.service... Dec 13 14:19:35.706649 systemd[1]: parse-ip-for-networkd.service: Deactivated successfully. Dec 13 14:19:35.706710 systemd[1]: Stopped parse-ip-for-networkd.service. Dec 13 14:19:35.710933 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 13 14:19:35.710000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=parse-ip-for-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.710000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.710000 audit: BPF prog-id=9 op=UNLOAD Dec 13 14:19:35.710982 systemd[1]: Stopped systemd-sysctl.service. Dec 13 14:19:35.713000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.713180 systemd[1]: systemd-modules-load.service: Deactivated successfully. Dec 13 14:19:35.713231 systemd[1]: Stopped systemd-modules-load.service. Dec 13 14:19:35.713839 systemd[1]: Stopping systemd-udevd.service... Dec 13 14:19:35.721760 systemd[1]: run-credentials-systemd\x2dsysctl.service.mount: Deactivated successfully. Dec 13 14:19:35.723544 systemd[1]: systemd-resolved.service: Deactivated successfully. Dec 13 14:19:35.724537 systemd[1]: Stopped systemd-resolved.service. Dec 13 14:19:35.726000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-resolved comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.726925 systemd[1]: systemd-udevd.service: Deactivated successfully. Dec 13 14:19:35.728017 systemd[1]: Stopped systemd-udevd.service. Dec 13 14:19:35.728000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.729000 audit: BPF prog-id=6 op=UNLOAD Dec 13 14:19:35.730900 systemd[1]: systemd-udevd-control.socket: Deactivated successfully. Dec 13 14:19:35.730983 systemd[1]: Closed systemd-udevd-control.socket. Dec 13 14:19:35.734053 systemd[1]: systemd-udevd-kernel.socket: Deactivated successfully. Dec 13 14:19:35.734093 systemd[1]: Closed systemd-udevd-kernel.socket. Dec 13 14:19:35.736100 systemd[1]: dracut-pre-udev.service: Deactivated successfully. Dec 13 14:19:35.737030 systemd[1]: Stopped dracut-pre-udev.service. Dec 13 14:19:35.770000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-pre-udev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.770384 systemd[1]: dracut-cmdline.service: Deactivated successfully. Dec 13 14:19:35.770437 systemd[1]: Stopped dracut-cmdline.service. Dec 13 14:19:35.772000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.773011 systemd[1]: dracut-cmdline-ask.service: Deactivated successfully. Dec 13 14:19:35.773065 systemd[1]: Stopped dracut-cmdline-ask.service. Dec 13 14:19:35.774000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=dracut-cmdline-ask comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.776466 systemd[1]: Starting initrd-udevadm-cleanup-db.service... Dec 13 14:19:35.778379 systemd[1]: systemd-tmpfiles-setup-dev.service: Deactivated successfully. Dec 13 14:19:35.778437 systemd[1]: Stopped systemd-tmpfiles-setup-dev.service. Dec 13 14:19:35.781000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.781465 systemd[1]: kmod-static-nodes.service: Deactivated successfully. Dec 13 14:19:35.781523 systemd[1]: Stopped kmod-static-nodes.service. Dec 13 14:19:35.784000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.784334 systemd[1]: systemd-vconsole-setup.service: Deactivated successfully. Dec 13 14:19:35.784386 systemd[1]: Stopped systemd-vconsole-setup.service. Dec 13 14:19:35.786000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=systemd-vconsole-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.787978 systemd[1]: run-credentials-systemd\x2dtmpfiles\x2dsetup\x2ddev.service.mount: Deactivated successfully. Dec 13 14:19:35.789956 systemd[1]: network-cleanup.service: Deactivated successfully. Dec 13 14:19:35.791088 systemd[1]: Stopped network-cleanup.service. Dec 13 14:19:35.842000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=network-cleanup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.842635 systemd[1]: initrd-udevadm-cleanup-db.service: Deactivated successfully. Dec 13 14:19:35.842729 systemd[1]: Finished initrd-udevadm-cleanup-db.service. Dec 13 14:19:35.845000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:35.845000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-udevadm-cleanup-db comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:36.079330 systemd[1]: sysroot-boot.service: Deactivated successfully. Dec 13 14:19:36.079420 systemd[1]: Stopped sysroot-boot.service. Dec 13 14:19:36.080000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=sysroot-boot comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:36.081911 systemd[1]: Reached target initrd-switch-root.target. Dec 13 14:19:36.083650 systemd[1]: initrd-setup-root.service: Deactivated successfully. Dec 13 14:19:36.083695 systemd[1]: Stopped initrd-setup-root.service. Dec 13 14:19:36.084000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=kernel msg='unit=initrd-setup-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:36.087023 systemd[1]: Starting initrd-switch-root.service... Dec 13 14:19:36.104696 systemd[1]: Switching root. Dec 13 14:19:36.157206 systemd-journald[197]: Journal stopped Dec 13 14:19:40.809150 systemd-journald[197]: Received SIGTERM from PID 1 (systemd). Dec 13 14:19:40.809206 kernel: SELinux: Class mctp_socket not defined in policy. Dec 13 14:19:40.809222 kernel: SELinux: Class anon_inode not defined in policy. Dec 13 14:19:40.809232 kernel: SELinux: the above unknown classes and permissions will be allowed Dec 13 14:19:40.809242 kernel: SELinux: policy capability network_peer_controls=1 Dec 13 14:19:40.809251 kernel: SELinux: policy capability open_perms=1 Dec 13 14:19:40.809264 kernel: SELinux: policy capability extended_socket_class=1 Dec 13 14:19:40.809274 kernel: SELinux: policy capability always_check_network=0 Dec 13 14:19:40.809285 kernel: SELinux: policy capability cgroup_seclabel=1 Dec 13 14:19:40.809295 kernel: SELinux: policy capability nnp_nosuid_transition=1 Dec 13 14:19:40.809304 kernel: SELinux: policy capability genfs_seclabel_symlinks=0 Dec 13 14:19:40.809316 kernel: SELinux: policy capability ioctl_skip_cloexec=0 Dec 13 14:19:40.809327 systemd[1]: Successfully loaded SELinux policy in 49.165ms. Dec 13 14:19:40.809345 systemd[1]: Relabelled /dev, /dev/shm, /run, /sys/fs/cgroup in 7.216ms. Dec 13 14:19:40.809361 systemd[1]: systemd 252 running in system mode (+PAM +AUDIT +SELINUX -APPARMOR +IMA +SMACK +SECCOMP +GCRYPT -GNUTLS +OPENSSL -ACL +BLKID +CURL -ELFUTILS -FIDO2 +IDN2 -IDN +IPTC +KMOD +LIBCRYPTSETUP +LIBFDISK +PCRE2 -PWQUALITY -P11KIT -QRENCODE -TPM2 +BZIP2 +LZ4 +XZ +ZLIB +ZSTD -BPF_FRAMEWORK -XKBCOMMON +UTMP +SYSVINIT default-hierarchy=unified) Dec 13 14:19:40.809374 systemd[1]: Detected virtualization kvm. Dec 13 14:19:40.809387 systemd[1]: Detected architecture x86-64. Dec 13 14:19:40.809402 systemd[1]: Detected first boot. Dec 13 14:19:40.809417 systemd[1]: Initializing machine ID from VM UUID. Dec 13 14:19:40.809430 kernel: SELinux: Context system_u:object_r:container_file_t:s0:c1022,c1023 is not valid (left unmapped). Dec 13 14:19:40.809445 systemd[1]: Populated /etc with preset unit settings. Dec 13 14:19:40.809459 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Dec 13 14:19:40.809472 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Dec 13 14:19:40.809484 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 14:19:40.809496 kernel: kauditd_printk_skb: 48 callbacks suppressed Dec 13 14:19:40.809505 kernel: audit: type=1334 audit(1734099580.568:85): prog-id=12 op=LOAD Dec 13 14:19:40.809515 kernel: audit: type=1334 audit(1734099580.568:86): prog-id=3 op=UNLOAD Dec 13 14:19:40.809524 kernel: audit: type=1334 audit(1734099580.571:87): prog-id=13 op=LOAD Dec 13 14:19:40.809534 kernel: audit: type=1334 audit(1734099580.573:88): prog-id=14 op=LOAD Dec 13 14:19:40.809543 kernel: audit: type=1334 audit(1734099580.573:89): prog-id=4 op=UNLOAD Dec 13 14:19:40.809553 kernel: audit: type=1334 audit(1734099580.573:90): prog-id=5 op=UNLOAD Dec 13 14:19:40.809562 kernel: audit: type=1334 audit(1734099580.577:91): prog-id=15 op=LOAD Dec 13 14:19:40.809571 kernel: audit: type=1334 audit(1734099580.577:92): prog-id=12 op=UNLOAD Dec 13 14:19:40.809583 kernel: audit: type=1334 audit(1734099580.580:93): prog-id=16 op=LOAD Dec 13 14:19:40.809592 kernel: audit: type=1334 audit(1734099580.581:94): prog-id=17 op=LOAD Dec 13 14:19:40.809603 systemd[1]: initrd-switch-root.service: Deactivated successfully. Dec 13 14:19:40.809613 systemd[1]: Stopped initrd-switch-root.service. Dec 13 14:19:40.809623 systemd[1]: systemd-journald.service: Scheduled restart job, restart counter is at 1. Dec 13 14:19:40.809634 systemd[1]: Created slice system-addon\x2dconfig.slice. Dec 13 14:19:40.809644 systemd[1]: Created slice system-addon\x2drun.slice. Dec 13 14:19:40.809654 systemd[1]: Created slice system-getty.slice. Dec 13 14:19:40.809666 systemd[1]: Created slice system-modprobe.slice. Dec 13 14:19:40.809676 systemd[1]: Created slice system-serial\x2dgetty.slice. Dec 13 14:19:40.809686 systemd[1]: Created slice system-system\x2dcloudinit.slice. Dec 13 14:19:40.809697 systemd[1]: Created slice system-systemd\x2dfsck.slice. Dec 13 14:19:40.809706 systemd[1]: Created slice user.slice. Dec 13 14:19:40.809716 systemd[1]: Started systemd-ask-password-console.path. Dec 13 14:19:40.809729 systemd[1]: Started systemd-ask-password-wall.path. Dec 13 14:19:40.809739 systemd[1]: Set up automount boot.automount. Dec 13 14:19:40.809749 systemd[1]: Set up automount proc-sys-fs-binfmt_misc.automount. Dec 13 14:19:40.809759 systemd[1]: Stopped target initrd-switch-root.target. Dec 13 14:19:40.809769 systemd[1]: Stopped target initrd-fs.target. Dec 13 14:19:40.809780 systemd[1]: Stopped target initrd-root-fs.target. Dec 13 14:19:40.809790 systemd[1]: Reached target integritysetup.target. Dec 13 14:19:40.809800 systemd[1]: Reached target remote-cryptsetup.target. Dec 13 14:19:40.809824 systemd[1]: Reached target remote-fs.target. Dec 13 14:19:40.809836 systemd[1]: Reached target slices.target. Dec 13 14:19:40.809846 systemd[1]: Reached target swap.target. Dec 13 14:19:40.809856 systemd[1]: Reached target torcx.target. Dec 13 14:19:40.809866 systemd[1]: Reached target veritysetup.target. Dec 13 14:19:40.809876 systemd[1]: Listening on systemd-coredump.socket. Dec 13 14:19:40.809886 systemd[1]: Listening on systemd-initctl.socket. Dec 13 14:19:40.809896 systemd[1]: Listening on systemd-networkd.socket. Dec 13 14:19:40.809907 systemd[1]: Listening on systemd-udevd-control.socket. Dec 13 14:19:40.809918 systemd[1]: Listening on systemd-udevd-kernel.socket. Dec 13 14:19:40.809936 systemd[1]: Listening on systemd-userdbd.socket. Dec 13 14:19:40.809947 systemd[1]: Mounting dev-hugepages.mount... Dec 13 14:19:40.809957 systemd[1]: Mounting dev-mqueue.mount... Dec 13 14:19:40.809968 systemd[1]: Mounting media.mount... Dec 13 14:19:40.809978 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 14:19:40.809988 systemd[1]: Mounting sys-kernel-debug.mount... Dec 13 14:19:40.809998 systemd[1]: Mounting sys-kernel-tracing.mount... Dec 13 14:19:40.810008 systemd[1]: Mounting tmp.mount... Dec 13 14:19:40.810018 systemd[1]: Starting flatcar-tmpfiles.service... Dec 13 14:19:40.810030 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:19:40.810040 systemd[1]: Starting kmod-static-nodes.service... Dec 13 14:19:40.810051 systemd[1]: Starting modprobe@configfs.service... Dec 13 14:19:40.810062 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:19:40.810072 systemd[1]: Starting modprobe@drm.service... Dec 13 14:19:40.810082 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:19:40.810093 systemd[1]: Starting modprobe@fuse.service... Dec 13 14:19:40.810103 systemd[1]: Starting modprobe@loop.service... Dec 13 14:19:40.810113 systemd[1]: setup-nsswitch.service was skipped because of an unmet condition check (ConditionPathExists=!/etc/nsswitch.conf). Dec 13 14:19:40.810126 systemd[1]: systemd-fsck-root.service: Deactivated successfully. Dec 13 14:19:40.810136 systemd[1]: Stopped systemd-fsck-root.service. Dec 13 14:19:40.810146 systemd[1]: systemd-fsck-usr.service: Deactivated successfully. Dec 13 14:19:40.810157 kernel: loop: module loaded Dec 13 14:19:40.810166 systemd[1]: Stopped systemd-fsck-usr.service. Dec 13 14:19:40.810177 systemd[1]: Stopped systemd-journald.service. Dec 13 14:19:40.810187 kernel: fuse: init (API version 7.34) Dec 13 14:19:40.810197 systemd[1]: Starting systemd-journald.service... Dec 13 14:19:40.810207 systemd[1]: Starting systemd-modules-load.service... Dec 13 14:19:40.810219 systemd[1]: Starting systemd-network-generator.service... Dec 13 14:19:40.810229 systemd[1]: Starting systemd-remount-fs.service... Dec 13 14:19:40.810239 systemd[1]: Starting systemd-udev-trigger.service... Dec 13 14:19:40.810250 systemd[1]: verity-setup.service: Deactivated successfully. Dec 13 14:19:40.810264 systemd-journald[1003]: Journal started Dec 13 14:19:40.810304 systemd-journald[1003]: Runtime Journal (/run/log/journal/218af52a80cf4cffa131177f8a944895) is 6.0M, max 48.5M, 42.5M free. Dec 13 14:19:36.220000 audit: MAC_POLICY_LOAD auid=4294967295 ses=4294967295 lsm=selinux res=1 Dec 13 14:19:36.552000 audit[1]: AVC avc: denied { bpf } for pid=1 comm="systemd" capability=39 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Dec 13 14:19:36.552000 audit[1]: AVC avc: denied { perfmon } for pid=1 comm="systemd" capability=38 scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=capability2 permissive=1 Dec 13 14:19:36.552000 audit: BPF prog-id=10 op=LOAD Dec 13 14:19:36.552000 audit: BPF prog-id=10 op=UNLOAD Dec 13 14:19:36.552000 audit: BPF prog-id=11 op=LOAD Dec 13 14:19:36.552000 audit: BPF prog-id=11 op=UNLOAD Dec 13 14:19:36.584000 audit[914]: AVC avc: denied { associate } for pid=914 comm="torcx-generator" name="docker" dev="tmpfs" ino=2 scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 srawcon="system_u:object_r:container_file_t:s0:c1022,c1023" Dec 13 14:19:36.584000 audit[914]: SYSCALL arch=c000003e syscall=188 success=yes exit=0 a0=c0001558b2 a1=c0000d8de0 a2=c0000e10c0 a3=32 items=0 ppid=897 pid=914 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 13 14:19:36.584000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Dec 13 14:19:36.586000 audit[914]: AVC avc: denied { associate } for pid=914 comm="torcx-generator" name="usr" scontext=system_u:object_r:unlabeled_t:s0 tcontext=system_u:object_r:tmpfs_t:s0 tclass=filesystem permissive=1 Dec 13 14:19:36.586000 audit[914]: SYSCALL arch=c000003e syscall=258 success=yes exit=0 a0=ffffffffffffff9c a1=c000155989 a2=1ed a3=0 items=2 ppid=897 pid=914 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="torcx-generator" exe="/usr/lib/systemd/system-generators/torcx-generator" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 13 14:19:36.586000 audit: CWD cwd="/" Dec 13 14:19:36.586000 audit: PATH item=0 name=(null) inode=2 dev=00:1b mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:36.586000 audit: PATH item=1 name=(null) inode=3 dev=00:1b mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:unlabeled_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:36.586000 audit: PROCTITLE proctitle=2F7573722F6C69622F73797374656D642F73797374656D2D67656E657261746F72732F746F7263782D67656E657261746F72002F72756E2F73797374656D642F67656E657261746F72002F72756E2F73797374656D642F67656E657261746F722E6561726C79002F72756E2F73797374656D642F67656E657261746F722E6C61 Dec 13 14:19:40.568000 audit: BPF prog-id=12 op=LOAD Dec 13 14:19:40.568000 audit: BPF prog-id=3 op=UNLOAD Dec 13 14:19:40.571000 audit: BPF prog-id=13 op=LOAD Dec 13 14:19:40.573000 audit: BPF prog-id=14 op=LOAD Dec 13 14:19:40.573000 audit: BPF prog-id=4 op=UNLOAD Dec 13 14:19:40.573000 audit: BPF prog-id=5 op=UNLOAD Dec 13 14:19:40.577000 audit: BPF prog-id=15 op=LOAD Dec 13 14:19:40.577000 audit: BPF prog-id=12 op=UNLOAD Dec 13 14:19:40.580000 audit: BPF prog-id=16 op=LOAD Dec 13 14:19:40.581000 audit: BPF prog-id=17 op=LOAD Dec 13 14:19:40.581000 audit: BPF prog-id=13 op=UNLOAD Dec 13 14:19:40.581000 audit: BPF prog-id=14 op=UNLOAD Dec 13 14:19:40.583000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.588000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.588000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=initrd-switch-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.595000 audit: BPF prog-id=15 op=UNLOAD Dec 13 14:19:40.782000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-root comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.786000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck-usr comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.789000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.789000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.790000 audit: BPF prog-id=18 op=LOAD Dec 13 14:19:40.790000 audit: BPF prog-id=19 op=LOAD Dec 13 14:19:40.790000 audit: BPF prog-id=20 op=LOAD Dec 13 14:19:40.790000 audit: BPF prog-id=16 op=UNLOAD Dec 13 14:19:40.790000 audit: BPF prog-id=17 op=UNLOAD Dec 13 14:19:40.806000 audit: CONFIG_CHANGE op=set audit_enabled=1 old=1 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 res=1 Dec 13 14:19:40.806000 audit[1003]: SYSCALL arch=c000003e syscall=46 success=yes exit=60 a0=4 a1=7ffcaccfa7b0 a2=4000 a3=7ffcaccfa84c items=0 ppid=1 pid=1003 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="systemd-journal" exe="/usr/lib/systemd/systemd-journald" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 13 14:19:40.806000 audit: PROCTITLE proctitle="/usr/lib/systemd/systemd-journald" Dec 13 14:19:36.583986 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.6 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.6 /var/lib/torcx/store]" Dec 13 14:19:40.567162 systemd[1]: Queued start job for default target multi-user.target. Dec 13 14:19:36.584377 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Dec 13 14:19:40.567174 systemd[1]: Unnecessary job was removed for dev-vda6.device. Dec 13 14:19:40.811590 systemd[1]: Stopped verity-setup.service. Dec 13 14:19:36.584395 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Dec 13 14:19:40.583552 systemd[1]: systemd-journald.service: Deactivated successfully. Dec 13 14:19:36.584424 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=info msg="no vendor profile selected by /etc/flatcar/docker-1.12" Dec 13 14:19:36.584433 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="skipped missing lower profile" missing profile=oem Dec 13 14:19:36.584463 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=warning msg="no next profile: unable to read profile file: open /etc/torcx/next-profile: no such file or directory" Dec 13 14:19:36.584474 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="apply configuration parsed" lower profiles (vendor/oem)="[vendor]" upper profile (user)= Dec 13 14:19:36.584663 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="mounted tmpfs" target=/run/torcx/unpack Dec 13 14:19:36.584700 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="profile found" name=docker-1.12-no path=/usr/share/torcx/profiles/docker-1.12-no.json Dec 13 14:19:36.584712 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="profile found" name=vendor path=/usr/share/torcx/profiles/vendor.json Dec 13 14:19:36.585083 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:20.10.torcx.tgz" reference=20.10 Dec 13 14:19:36.585114 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=debug msg="new archive/reference added to cache" format=tgz name=docker path="/usr/share/torcx/store/docker:com.coreos.cl.torcx.tgz" reference=com.coreos.cl Dec 13 14:19:36.585130 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store/3510.3.6: no such file or directory" path=/usr/share/oem/torcx/store/3510.3.6 Dec 13 14:19:40.811000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=verity-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:36.585142 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=info msg="store skipped" err="open /usr/share/oem/torcx/store: no such file or directory" path=/usr/share/oem/torcx/store Dec 13 14:19:36.585157 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=info msg="store skipped" err="open /var/lib/torcx/store/3510.3.6: no such file or directory" path=/var/lib/torcx/store/3510.3.6 Dec 13 14:19:36.585170 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:36Z" level=info msg="store skipped" err="open /var/lib/torcx/store: no such file or directory" path=/var/lib/torcx/store Dec 13 14:19:40.234754 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:40Z" level=debug msg="image unpacked" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Dec 13 14:19:40.235153 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:40Z" level=debug msg="binaries propagated" assets="[/bin/containerd /bin/containerd-shim /bin/ctr /bin/docker /bin/docker-containerd /bin/docker-containerd-shim /bin/docker-init /bin/docker-proxy /bin/docker-runc /bin/dockerd /bin/runc /bin/tini]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Dec 13 14:19:40.235465 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:40Z" level=debug msg="networkd units propagated" assets="[/lib/systemd/network/50-docker.network /lib/systemd/network/90-docker-veth.network]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Dec 13 14:19:40.235664 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:40Z" level=debug msg="systemd units propagated" assets="[/lib/systemd/system/containerd.service /lib/systemd/system/docker.service /lib/systemd/system/docker.socket /lib/systemd/system/sockets.target.wants /lib/systemd/system/multi-user.target.wants]" image=docker path=/run/torcx/unpack/docker reference=com.coreos.cl Dec 13 14:19:40.235712 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:40Z" level=debug msg="profile applied" sealed profile=/run/torcx/profile.json upper profile= Dec 13 14:19:40.235783 /usr/lib/systemd/system-generators/torcx-generator[914]: time="2024-12-13T14:19:40Z" level=debug msg="system state sealed" content="[TORCX_LOWER_PROFILES=\"vendor\" TORCX_UPPER_PROFILE=\"\" TORCX_PROFILE_PATH=\"/run/torcx/profile.json\" TORCX_BINDIR=\"/run/torcx/bin\" TORCX_UNPACKDIR=\"/run/torcx/unpack\"]" path=/run/metadata/torcx Dec 13 14:19:40.814849 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 14:19:40.818208 systemd[1]: Started systemd-journald.service. Dec 13 14:19:40.818000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journald comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.818952 systemd[1]: Mounted dev-hugepages.mount. Dec 13 14:19:40.819959 systemd[1]: Mounted dev-mqueue.mount. Dec 13 14:19:40.821177 systemd[1]: Mounted media.mount. Dec 13 14:19:40.822111 systemd[1]: Mounted sys-kernel-debug.mount. Dec 13 14:19:40.823061 systemd[1]: Mounted sys-kernel-tracing.mount. Dec 13 14:19:40.824024 systemd[1]: Mounted tmp.mount. Dec 13 14:19:40.825120 systemd[1]: Finished flatcar-tmpfiles.service. Dec 13 14:19:40.825000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=flatcar-tmpfiles comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.826523 systemd[1]: Finished kmod-static-nodes.service. Dec 13 14:19:40.827000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=kmod-static-nodes comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.827638 systemd[1]: modprobe@configfs.service: Deactivated successfully. Dec 13 14:19:40.827830 systemd[1]: Finished modprobe@configfs.service. Dec 13 14:19:40.827000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.827000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@configfs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.828977 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:19:40.829187 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:19:40.829000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.829000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.830314 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 14:19:40.830539 systemd[1]: Finished modprobe@drm.service. Dec 13 14:19:40.831000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.831000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@drm comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.831697 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:19:40.831902 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:19:40.832000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.832000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.833162 systemd[1]: modprobe@fuse.service: Deactivated successfully. Dec 13 14:19:40.833340 systemd[1]: Finished modprobe@fuse.service. Dec 13 14:19:40.861000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.861000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@fuse comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.862252 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:19:40.862464 systemd[1]: Finished modprobe@loop.service. Dec 13 14:19:40.863000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.863000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.863759 systemd[1]: Finished systemd-modules-load.service. Dec 13 14:19:40.863000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-modules-load comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.865116 systemd[1]: Finished systemd-network-generator.service. Dec 13 14:19:40.866000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-network-generator comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.866488 systemd[1]: Finished systemd-remount-fs.service. Dec 13 14:19:40.867000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-remount-fs comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.868830 systemd[1]: Reached target network-pre.target. Dec 13 14:19:40.871078 systemd[1]: Mounting sys-fs-fuse-connections.mount... Dec 13 14:19:40.873038 systemd[1]: Mounting sys-kernel-config.mount... Dec 13 14:19:40.874129 systemd[1]: remount-root.service was skipped because of an unmet condition check (ConditionPathIsReadWrite=!/). Dec 13 14:19:40.876713 systemd[1]: Starting systemd-hwdb-update.service... Dec 13 14:19:40.878869 systemd[1]: Starting systemd-journal-flush.service... Dec 13 14:19:40.879859 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 14:19:40.880755 systemd[1]: Starting systemd-random-seed.service... Dec 13 14:19:40.881941 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Dec 13 14:19:40.882958 systemd[1]: Starting systemd-sysctl.service... Dec 13 14:19:40.886205 systemd-journald[1003]: Time spent on flushing to /var/log/journal/218af52a80cf4cffa131177f8a944895 is 18.579ms for 1095 entries. Dec 13 14:19:40.886205 systemd-journald[1003]: System Journal (/var/log/journal/218af52a80cf4cffa131177f8a944895) is 8.0M, max 195.6M, 187.6M free. Dec 13 14:19:41.290015 systemd-journald[1003]: Received client request to flush runtime journal. Dec 13 14:19:40.897000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-trigger comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:41.008000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysctl comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:41.011000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysusers comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:41.030000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup-dev comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:41.088000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-random-seed comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:40.885152 systemd[1]: Starting systemd-sysusers.service... Dec 13 14:19:40.892218 systemd[1]: Finished systemd-udev-trigger.service. Dec 13 14:19:41.291071 udevadm[1018]: systemd-udev-settle.service is deprecated. Please fix lvm2-activation.service, lvm2-activation-early.service not to pull it in. Dec 13 14:19:40.898347 systemd[1]: Mounted sys-fs-fuse-connections.mount. Dec 13 14:19:40.899505 systemd[1]: Mounted sys-kernel-config.mount. Dec 13 14:19:40.901630 systemd[1]: Starting systemd-udev-settle.service... Dec 13 14:19:41.007326 systemd[1]: Finished systemd-sysctl.service. Dec 13 14:19:41.010286 systemd[1]: Finished systemd-sysusers.service. Dec 13 14:19:41.012484 systemd[1]: Starting systemd-tmpfiles-setup-dev.service... Dec 13 14:19:41.028957 systemd[1]: Finished systemd-tmpfiles-setup-dev.service. Dec 13 14:19:41.070324 systemd[1]: Finished systemd-random-seed.service. Dec 13 14:19:41.089099 systemd[1]: Reached target first-boot-complete.target. Dec 13 14:19:41.291592 systemd[1]: Finished systemd-journal-flush.service. Dec 13 14:19:41.293000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-flush comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:41.723846 systemd[1]: Finished systemd-hwdb-update.service. Dec 13 14:19:41.724000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-hwdb-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:41.725000 audit: BPF prog-id=21 op=LOAD Dec 13 14:19:41.725000 audit: BPF prog-id=22 op=LOAD Dec 13 14:19:41.725000 audit: BPF prog-id=7 op=UNLOAD Dec 13 14:19:41.725000 audit: BPF prog-id=8 op=UNLOAD Dec 13 14:19:41.727373 systemd[1]: Starting systemd-udevd.service... Dec 13 14:19:41.752803 systemd-udevd[1023]: Using default interface naming scheme 'v252'. Dec 13 14:19:41.776059 systemd[1]: Started systemd-udevd.service. Dec 13 14:19:41.777000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udevd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:41.777000 audit: BPF prog-id=23 op=LOAD Dec 13 14:19:41.782264 systemd[1]: Starting systemd-networkd.service... Dec 13 14:19:41.792000 audit: BPF prog-id=24 op=LOAD Dec 13 14:19:41.793000 audit: BPF prog-id=25 op=LOAD Dec 13 14:19:41.793000 audit: BPF prog-id=26 op=LOAD Dec 13 14:19:41.795202 systemd[1]: Starting systemd-userdbd.service... Dec 13 14:19:41.825000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-userdbd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:41.823968 systemd[1]: Started systemd-userdbd.service. Dec 13 14:19:41.834338 systemd[1]: Condition check resulted in dev-ttyS0.device being skipped. Dec 13 14:19:41.848319 systemd[1]: Found device dev-disk-by\x2dlabel-OEM.device. Dec 13 14:19:41.872854 kernel: input: Power Button as /devices/LNXSYSTM:00/LNXPWRBN:00/input/input2 Dec 13 14:19:41.875438 systemd-networkd[1038]: lo: Link UP Dec 13 14:19:41.875449 systemd-networkd[1038]: lo: Gained carrier Dec 13 14:19:41.876291 systemd-networkd[1038]: Enumeration completed Dec 13 14:19:41.877000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-networkd comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:41.876421 systemd-networkd[1038]: eth0: Configuring with /usr/lib/systemd/network/zz-default.network. Dec 13 14:19:41.876430 systemd[1]: Started systemd-networkd.service. Dec 13 14:19:41.877732 systemd-networkd[1038]: eth0: Link UP Dec 13 14:19:41.877740 systemd-networkd[1038]: eth0: Gained carrier Dec 13 14:19:41.878869 kernel: ACPI: button: Power Button [PWRF] Dec 13 14:19:41.889014 systemd-networkd[1038]: eth0: DHCPv4 address 10.0.0.34/16, gateway 10.0.0.1 acquired from 10.0.0.1 Dec 13 14:19:41.883000 audit[1040]: AVC avc: denied { confidentiality } for pid=1040 comm="(udev-worker)" lockdown_reason="use of tracefs" scontext=system_u:system_r:kernel_t:s0 tcontext=system_u:system_r:kernel_t:s0 tclass=lockdown permissive=1 Dec 13 14:19:41.883000 audit[1040]: SYSCALL arch=c000003e syscall=175 success=yes exit=0 a0=5648d49751e0 a1=337fc a2=7f7a4f6f5bc5 a3=5 items=110 ppid=1023 pid=1040 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="(udev-worker)" exe="/usr/bin/udevadm" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 13 14:19:41.883000 audit: CWD cwd="/" Dec 13 14:19:41.883000 audit: PATH item=0 name=(null) inode=44 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=1 name=(null) inode=14079 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=2 name=(null) inode=14079 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=3 name=(null) inode=14080 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=4 name=(null) inode=14079 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=5 name=(null) inode=14081 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=6 name=(null) inode=14079 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=7 name=(null) inode=14082 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=8 name=(null) inode=14082 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=9 name=(null) inode=14083 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=10 name=(null) inode=14082 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=11 name=(null) inode=14084 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=12 name=(null) inode=14082 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=13 name=(null) inode=14085 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=14 name=(null) inode=14082 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=15 name=(null) inode=14086 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=16 name=(null) inode=14082 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=17 name=(null) inode=14087 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=18 name=(null) inode=14079 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=19 name=(null) inode=14088 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=20 name=(null) inode=14088 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=21 name=(null) inode=14089 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=22 name=(null) inode=14088 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=23 name=(null) inode=14090 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=24 name=(null) inode=14088 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=25 name=(null) inode=14091 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=26 name=(null) inode=14088 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=27 name=(null) inode=14092 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=28 name=(null) inode=14088 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=29 name=(null) inode=14093 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=30 name=(null) inode=14079 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=31 name=(null) inode=14094 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=32 name=(null) inode=14094 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=33 name=(null) inode=14095 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=34 name=(null) inode=14094 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=35 name=(null) inode=14096 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=36 name=(null) inode=14094 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=37 name=(null) inode=14097 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=38 name=(null) inode=14094 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=39 name=(null) inode=14098 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=40 name=(null) inode=14094 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=41 name=(null) inode=14099 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=42 name=(null) inode=14079 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=43 name=(null) inode=14100 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=44 name=(null) inode=14100 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=45 name=(null) inode=14101 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=46 name=(null) inode=14100 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=47 name=(null) inode=14102 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=48 name=(null) inode=14100 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=49 name=(null) inode=14103 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=50 name=(null) inode=14100 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=51 name=(null) inode=14104 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=52 name=(null) inode=14100 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=53 name=(null) inode=14105 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=54 name=(null) inode=44 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=55 name=(null) inode=14106 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=56 name=(null) inode=14106 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=57 name=(null) inode=14107 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=58 name=(null) inode=14106 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=59 name=(null) inode=14108 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=60 name=(null) inode=14106 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=61 name=(null) inode=14109 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=62 name=(null) inode=14109 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=63 name=(null) inode=14110 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=64 name=(null) inode=14109 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=65 name=(null) inode=14111 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=66 name=(null) inode=14109 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=67 name=(null) inode=14112 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=68 name=(null) inode=14109 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=69 name=(null) inode=14113 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=70 name=(null) inode=14109 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=71 name=(null) inode=14114 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=72 name=(null) inode=14106 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=73 name=(null) inode=14115 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=74 name=(null) inode=14115 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=75 name=(null) inode=14116 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=76 name=(null) inode=14115 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=77 name=(null) inode=14117 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=78 name=(null) inode=14115 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=79 name=(null) inode=14118 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=80 name=(null) inode=14115 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=81 name=(null) inode=14119 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=82 name=(null) inode=14115 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=83 name=(null) inode=14120 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=84 name=(null) inode=14106 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=85 name=(null) inode=14121 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=86 name=(null) inode=14121 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=87 name=(null) inode=14122 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=88 name=(null) inode=14121 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=89 name=(null) inode=14123 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=90 name=(null) inode=14121 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=91 name=(null) inode=14124 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=92 name=(null) inode=14121 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=93 name=(null) inode=14125 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=94 name=(null) inode=14121 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=95 name=(null) inode=14126 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=96 name=(null) inode=14106 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=97 name=(null) inode=14127 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=98 name=(null) inode=14127 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=99 name=(null) inode=14128 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=100 name=(null) inode=14127 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=101 name=(null) inode=14129 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=102 name=(null) inode=14127 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=103 name=(null) inode=14130 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=104 name=(null) inode=14127 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=105 name=(null) inode=14131 dev=00:0b mode=0100640 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=106 name=(null) inode=14127 dev=00:0b mode=040750 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=107 name=(null) inode=14132 dev=00:0b mode=0100440 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:tracefs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=108 name=(null) inode=1 dev=00:07 mode=040700 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=PARENT cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PATH item=109 name=(null) inode=14133 dev=00:07 mode=040755 ouid=0 ogid=0 rdev=00:00 obj=system_u:object_r:debugfs_t:s0 nametype=CREATE cap_fp=0 cap_fi=0 cap_fe=0 cap_fver=0 cap_frootid=0 Dec 13 14:19:41.883000 audit: PROCTITLE proctitle="(udev-worker)" Dec 13 14:19:41.918960 kernel: i801_smbus 0000:00:1f.3: SMBus using PCI interrupt Dec 13 14:19:41.933104 kernel: i2c i2c-0: 1/1 memory slots populated (from DMI) Dec 13 14:19:41.933271 kernel: i2c i2c-0: Memory type 0x07 not supported yet, not instantiating SPD Dec 13 14:19:41.933422 kernel: input: ImExPS/2 Generic Explorer Mouse as /devices/platform/i8042/serio1/input/input3 Dec 13 14:19:42.001934 kernel: mousedev: PS/2 mouse device common for all mice Dec 13 14:19:42.029370 kernel: kvm: Nested Virtualization enabled Dec 13 14:19:42.029503 kernel: SVM: kvm: Nested Paging enabled Dec 13 14:19:42.029545 kernel: SVM: Virtual VMLOAD VMSAVE supported Dec 13 14:19:42.029560 kernel: SVM: Virtual GIF supported Dec 13 14:19:42.045831 kernel: EDAC MC: Ver: 3.0.0 Dec 13 14:19:42.073289 systemd[1]: Finished systemd-udev-settle.service. Dec 13 14:19:42.073000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-udev-settle comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.075475 systemd[1]: Starting lvm2-activation-early.service... Dec 13 14:19:42.083009 lvm[1059]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 14:19:42.109284 systemd[1]: Finished lvm2-activation-early.service. Dec 13 14:19:42.110000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation-early comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.110424 systemd[1]: Reached target cryptsetup.target. Dec 13 14:19:42.112405 systemd[1]: Starting lvm2-activation.service... Dec 13 14:19:42.116607 lvm[1060]: WARNING: Failed to connect to lvmetad. Falling back to device scanning. Dec 13 14:19:42.145122 systemd[1]: Finished lvm2-activation.service. Dec 13 14:19:42.145000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=lvm2-activation comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.146720 systemd[1]: Reached target local-fs-pre.target. Dec 13 14:19:42.147624 systemd[1]: var-lib-machines.mount was skipped because of an unmet condition check (ConditionPathExists=/var/lib/machines.raw). Dec 13 14:19:42.147654 systemd[1]: Reached target local-fs.target. Dec 13 14:19:42.148498 systemd[1]: Reached target machines.target. Dec 13 14:19:42.150617 systemd[1]: Starting ldconfig.service... Dec 13 14:19:42.151755 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:19:42.151791 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:19:42.152705 systemd[1]: Starting systemd-boot-update.service... Dec 13 14:19:42.154593 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-OEM.service... Dec 13 14:19:42.156497 systemd[1]: Starting systemd-machine-id-commit.service... Dec 13 14:19:42.158908 systemd[1]: Starting systemd-sysext.service... Dec 13 14:19:42.164177 systemd[1]: boot.automount: Got automount request for /boot, triggered by 1062 (bootctl) Dec 13 14:19:42.165754 systemd[1]: Starting systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service... Dec 13 14:19:42.182679 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-OEM.service. Dec 13 14:19:42.184000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-OEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.189930 systemd[1]: Unmounting usr-share-oem.mount... Dec 13 14:19:42.195087 systemd[1]: usr-share-oem.mount: Deactivated successfully. Dec 13 14:19:42.195334 systemd[1]: Unmounted usr-share-oem.mount. Dec 13 14:19:42.225954 kernel: loop0: detected capacity change from 0 to 211296 Dec 13 14:19:42.641716 systemd-fsck[1069]: fsck.fat 4.2 (2021-01-31) Dec 13 14:19:42.641716 systemd-fsck[1069]: /dev/vda1: 789 files, 119291/258078 clusters Dec 13 14:19:42.643561 systemd[1]: Finished systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM.service. Dec 13 14:19:42.644000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-fsck@dev-disk-by\x2dlabel-EFI\x2dSYSTEM comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.647033 systemd[1]: Mounting boot.mount... Dec 13 14:19:42.648882 kernel: squashfs: version 4.0 (2009/01/31) Phillip Lougher Dec 13 14:19:42.654445 systemd[1]: Mounted boot.mount. Dec 13 14:19:42.708848 kernel: loop1: detected capacity change from 0 to 211296 Dec 13 14:19:42.718682 systemd[1]: Finished systemd-boot-update.service. Dec 13 14:19:42.719000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-boot-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.722054 (sd-sysext)[1075]: Using extensions 'kubernetes'. Dec 13 14:19:42.722449 (sd-sysext)[1075]: Merged extensions into '/usr'. Dec 13 14:19:42.753703 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 14:19:42.755422 systemd[1]: Mounting usr-share-oem.mount... Dec 13 14:19:42.756546 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:19:42.758272 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:19:42.760793 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:19:42.764303 systemd[1]: Starting modprobe@loop.service... Dec 13 14:19:42.765412 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:19:42.765646 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:19:42.765841 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 14:19:42.768654 systemd[1]: Mounted usr-share-oem.mount. Dec 13 14:19:42.770470 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:19:42.770612 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:19:42.770000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.770000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.772355 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:19:42.772539 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:19:42.773000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.773000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.774223 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:19:42.774377 systemd[1]: Finished modprobe@loop.service. Dec 13 14:19:42.776000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.776000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@loop comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.777481 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 14:19:42.777616 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Dec 13 14:19:42.778720 systemd[1]: Finished systemd-sysext.service. Dec 13 14:19:42.778000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-sysext comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:42.781172 systemd[1]: Starting ensure-sysext.service... Dec 13 14:19:42.783006 systemd[1]: Starting systemd-tmpfiles-setup.service... Dec 13 14:19:42.793525 systemd-tmpfiles[1082]: /usr/lib/tmpfiles.d/legacy.conf:13: Duplicate line for path "/run/lock", ignoring. Dec 13 14:19:42.793631 systemd[1]: Reloading. Dec 13 14:19:42.794309 systemd-tmpfiles[1082]: /usr/lib/tmpfiles.d/provision.conf:20: Duplicate line for path "/root", ignoring. Dec 13 14:19:42.796268 systemd-tmpfiles[1082]: /usr/lib/tmpfiles.d/systemd.conf:29: Duplicate line for path "/var/lib/systemd", ignoring. Dec 13 14:19:42.863119 ldconfig[1061]: /sbin/ldconfig: /lib/ld.so.conf is not an ELF file - it has the wrong magic bytes at the start. Dec 13 14:19:42.874789 /usr/lib/systemd/system-generators/torcx-generator[1102]: time="2024-12-13T14:19:42Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.6 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.6 /var/lib/torcx/store]" Dec 13 14:19:42.874858 /usr/lib/systemd/system-generators/torcx-generator[1102]: time="2024-12-13T14:19:42Z" level=info msg="torcx already run" Dec 13 14:19:42.967255 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Dec 13 14:19:42.967273 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Dec 13 14:19:42.986510 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 14:19:43.041942 systemd[1]: etc-machine\x2did.mount: Deactivated successfully. Dec 13 14:19:43.043000 audit: BPF prog-id=27 op=LOAD Dec 13 14:19:43.043000 audit: BPF prog-id=23 op=UNLOAD Dec 13 14:19:43.045000 audit: BPF prog-id=28 op=LOAD Dec 13 14:19:43.045000 audit: BPF prog-id=24 op=UNLOAD Dec 13 14:19:43.045000 audit: BPF prog-id=29 op=LOAD Dec 13 14:19:43.045000 audit: BPF prog-id=30 op=LOAD Dec 13 14:19:43.045000 audit: BPF prog-id=25 op=UNLOAD Dec 13 14:19:43.045000 audit: BPF prog-id=26 op=UNLOAD Dec 13 14:19:43.046000 audit: BPF prog-id=31 op=LOAD Dec 13 14:19:43.046000 audit: BPF prog-id=32 op=LOAD Dec 13 14:19:43.046000 audit: BPF prog-id=21 op=UNLOAD Dec 13 14:19:43.046000 audit: BPF prog-id=22 op=UNLOAD Dec 13 14:19:43.047000 audit: BPF prog-id=33 op=LOAD Dec 13 14:19:43.047000 audit: BPF prog-id=18 op=UNLOAD Dec 13 14:19:43.047000 audit: BPF prog-id=34 op=LOAD Dec 13 14:19:43.047000 audit: BPF prog-id=35 op=LOAD Dec 13 14:19:43.047000 audit: BPF prog-id=19 op=UNLOAD Dec 13 14:19:43.047000 audit: BPF prog-id=20 op=UNLOAD Dec 13 14:19:43.051172 systemd[1]: Finished ldconfig.service. Dec 13 14:19:43.051000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=ldconfig comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.052537 systemd[1]: Finished systemd-machine-id-commit.service. Dec 13 14:19:43.052000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-machine-id-commit comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.054917 systemd[1]: Finished systemd-tmpfiles-setup.service. Dec 13 14:19:43.055000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-tmpfiles-setup comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.059156 systemd[1]: Starting audit-rules.service... Dec 13 14:19:43.061014 systemd[1]: Starting clean-ca-certificates.service... Dec 13 14:19:43.063171 systemd[1]: Starting systemd-journal-catalog-update.service... Dec 13 14:19:43.064000 audit: BPF prog-id=36 op=LOAD Dec 13 14:19:43.066327 systemd[1]: Starting systemd-resolved.service... Dec 13 14:19:43.067000 audit: BPF prog-id=37 op=LOAD Dec 13 14:19:43.070123 systemd[1]: Starting systemd-timesyncd.service... Dec 13 14:19:43.073140 systemd[1]: Starting systemd-update-utmp.service... Dec 13 14:19:43.075335 systemd[1]: Finished clean-ca-certificates.service. Dec 13 14:19:43.075000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=clean-ca-certificates comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.082000 audit[1155]: SYSTEM_BOOT pid=1155 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg=' comm="systemd-update-utmp" exe="/usr/lib/systemd/systemd-update-utmp" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.087521 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:19:43.090009 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:19:43.095565 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:19:43.099612 systemd[1]: Starting modprobe@loop.service... Dec 13 14:19:43.101427 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:19:43.101726 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:19:43.101944 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 14:19:43.104293 systemd[1]: Finished systemd-journal-catalog-update.service. Dec 13 14:19:43.104000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-journal-catalog-update comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.106259 systemd[1]: Finished systemd-update-utmp.service. Dec 13 14:19:43.106000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=systemd-update-utmp comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.108000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.108000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@dm_mod comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.110000 audit[1]: SERVICE_START pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.110000 audit[1]: SERVICE_STOP pid=1 uid=0 auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 msg='unit=modprobe@efi_pstore comm="systemd" exe="/usr/lib/systemd/systemd" hostname=? addr=? terminal=? res=success' Dec 13 14:19:43.111000 audit: CONFIG_CHANGE auid=4294967295 ses=4294967295 subj=system_u:system_r:kernel_t:s0 op=add_rule key=(null) list=5 res=1 Dec 13 14:19:43.111000 audit[1167]: SYSCALL arch=c000003e syscall=44 success=yes exit=1056 a0=3 a1=7ffe12b068f0 a2=420 a3=0 items=0 ppid=1144 pid=1167 auid=4294967295 uid=0 gid=0 euid=0 suid=0 fsuid=0 egid=0 sgid=0 fsgid=0 tty=(none) ses=4294967295 comm="auditctl" exe="/usr/sbin/auditctl" subj=system_u:system_r:kernel_t:s0 key=(null) Dec 13 14:19:43.111000 audit: PROCTITLE proctitle=2F7362696E2F617564697463746C002D52002F6574632F61756469742F61756469742E72756C6573 Dec 13 14:19:43.107955 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:19:43.113248 augenrules[1167]: No rules Dec 13 14:19:43.108112 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:19:43.109756 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:19:43.109955 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:19:43.111661 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:19:43.111831 systemd[1]: Finished modprobe@loop.service. Dec 13 14:19:43.113464 systemd[1]: Finished audit-rules.service. Dec 13 14:19:43.116018 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 14:19:43.116165 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Dec 13 14:19:43.118085 systemd[1]: Starting systemd-update-done.service... Dec 13 14:19:43.121391 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:19:43.123275 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:19:43.125727 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:19:43.128438 systemd[1]: Starting modprobe@loop.service... Dec 13 14:19:43.129465 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:19:43.129600 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:19:43.129722 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 14:19:43.130871 systemd[1]: Finished systemd-update-done.service. Dec 13 14:19:43.132587 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:19:43.132748 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:19:43.134286 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:19:43.134427 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:19:43.135950 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:19:43.136098 systemd[1]: Finished modprobe@loop.service. Dec 13 14:19:43.137784 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 14:19:43.137935 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Dec 13 14:19:43.141321 systemd[1]: ignition-delete-config.service was skipped because no trigger condition checks were met. Dec 13 14:19:43.143200 systemd[1]: Starting modprobe@dm_mod.service... Dec 13 14:19:43.145997 systemd[1]: Starting modprobe@drm.service... Dec 13 14:19:43.148618 systemd[1]: Starting modprobe@efi_pstore.service... Dec 13 14:19:43.152716 systemd[1]: Starting modprobe@loop.service... Dec 13 14:19:43.154035 systemd[1]: systemd-binfmt.service was skipped because no trigger condition checks were met. Dec 13 14:19:43.154254 systemd[1]: systemd-boot-system-token.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/LoaderFeatures-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:19:43.156159 systemd[1]: Starting systemd-networkd-wait-online.service... Dec 13 14:19:43.158214 systemd[1]: update-ca-certificates.service was skipped because of an unmet condition check (ConditionPathIsSymbolicLink=!/etc/ssl/certs/ca-certificates.crt). Dec 13 14:19:43.161651 systemd[1]: modprobe@dm_mod.service: Deactivated successfully. Dec 13 14:19:43.161845 systemd[1]: Finished modprobe@dm_mod.service. Dec 13 14:19:43.163459 systemd[1]: modprobe@drm.service: Deactivated successfully. Dec 13 14:19:43.163605 systemd[1]: Finished modprobe@drm.service. Dec 13 14:19:43.165036 systemd[1]: modprobe@efi_pstore.service: Deactivated successfully. Dec 13 14:19:43.165187 systemd[1]: Finished modprobe@efi_pstore.service. Dec 13 14:19:43.166656 systemd[1]: modprobe@loop.service: Deactivated successfully. Dec 13 14:19:43.166800 systemd[1]: Finished modprobe@loop.service. Dec 13 14:19:43.168547 systemd[1]: Started systemd-timesyncd.service. Dec 13 14:19:43.170702 systemd[1]: Reached target time-set.target. Dec 13 14:19:44.559650 systemd-timesyncd[1154]: Contacted time server 10.0.0.1:123 (10.0.0.1). Dec 13 14:19:44.559729 systemd-timesyncd[1154]: Initial clock synchronization to Fri 2024-12-13 14:19:44.559476 UTC. Dec 13 14:19:44.560121 systemd[1]: systemd-pstore.service was skipped because of an unmet condition check (ConditionDirectoryNotEmpty=/sys/fs/pstore). Dec 13 14:19:44.560172 systemd[1]: systemd-repart.service was skipped because no trigger condition checks were met. Dec 13 14:19:44.560437 systemd-resolved[1151]: Positive Trust Anchors: Dec 13 14:19:44.560454 systemd-resolved[1151]: . IN DS 20326 8 2 e06d44b80b8f1d39a95c0b0d7c65d08458e880409bbc683457104237c7f8ec8d Dec 13 14:19:44.560481 systemd-resolved[1151]: Negative trust anchors: home.arpa 10.in-addr.arpa 16.172.in-addr.arpa 17.172.in-addr.arpa 18.172.in-addr.arpa 19.172.in-addr.arpa 20.172.in-addr.arpa 21.172.in-addr.arpa 22.172.in-addr.arpa 23.172.in-addr.arpa 24.172.in-addr.arpa 25.172.in-addr.arpa 26.172.in-addr.arpa 27.172.in-addr.arpa 28.172.in-addr.arpa 29.172.in-addr.arpa 30.172.in-addr.arpa 31.172.in-addr.arpa 168.192.in-addr.arpa d.f.ip6.arpa corp home internal intranet lan local private test Dec 13 14:19:44.560598 systemd[1]: Finished ensure-sysext.service. Dec 13 14:19:44.572474 systemd-resolved[1151]: Defaulting to hostname 'linux'. Dec 13 14:19:44.574228 systemd[1]: Started systemd-resolved.service. Dec 13 14:19:44.575527 systemd[1]: Reached target network.target. Dec 13 14:19:44.576576 systemd[1]: Reached target nss-lookup.target. Dec 13 14:19:44.577624 systemd[1]: Reached target sysinit.target. Dec 13 14:19:44.578721 systemd[1]: Started motdgen.path. Dec 13 14:19:44.579723 systemd[1]: Started user-cloudinit@var-lib-flatcar\x2dinstall-user_data.path. Dec 13 14:19:44.581388 systemd[1]: Started logrotate.timer. Dec 13 14:19:44.582400 systemd[1]: Started mdadm.timer. Dec 13 14:19:44.583274 systemd[1]: Started systemd-tmpfiles-clean.timer. Dec 13 14:19:44.584410 systemd[1]: update-engine-stub.timer was skipped because of an unmet condition check (ConditionPathExists=/usr/.noupdate). Dec 13 14:19:44.584450 systemd[1]: Reached target paths.target. Dec 13 14:19:44.585434 systemd[1]: Reached target timers.target. Dec 13 14:19:44.587030 systemd[1]: Listening on dbus.socket. Dec 13 14:19:44.589478 systemd[1]: Starting docker.socket... Dec 13 14:19:44.593089 systemd[1]: Listening on sshd.socket. Dec 13 14:19:44.594024 systemd[1]: systemd-pcrphase-sysinit.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:19:44.595807 systemd[1]: Listening on docker.socket. Dec 13 14:19:44.596773 systemd[1]: Reached target sockets.target. Dec 13 14:19:44.597695 systemd[1]: Reached target basic.target. Dec 13 14:19:44.598657 systemd[1]: proc-xen.mount was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 14:19:44.598694 systemd[1]: addon-config@usr-share-oem.service was skipped because no trigger condition checks were met. Dec 13 14:19:44.598712 systemd[1]: addon-run@usr-share-oem.service was skipped because no trigger condition checks were met. Dec 13 14:19:44.599760 systemd[1]: Starting containerd.service... Dec 13 14:19:44.601594 systemd[1]: Starting dbus.service... Dec 13 14:19:44.603780 systemd[1]: Starting enable-oem-cloudinit.service... Dec 13 14:19:44.605945 systemd[1]: Starting extend-filesystems.service... Dec 13 14:19:44.607223 systemd[1]: flatcar-setup-environment.service was skipped because of an unmet condition check (ConditionPathExists=/usr/share/oem/bin/flatcar-setup-environment). Dec 13 14:19:44.608493 systemd[1]: Starting motdgen.service... Dec 13 14:19:44.613763 jq[1186]: false Dec 13 14:19:44.611386 systemd[1]: Starting ssh-key-proc-cmdline.service... Dec 13 14:19:44.613655 systemd[1]: Starting sshd-keygen.service... Dec 13 14:19:44.617313 systemd[1]: Starting systemd-logind.service... Dec 13 14:19:44.623925 systemd[1]: systemd-pcrphase.service was skipped because of an unmet condition check (ConditionPathExists=/sys/firmware/efi/efivars/StubPcrKernelImage-4a67b082-0a4c-41cf-b6c7-440b29bb8c4f). Dec 13 14:19:44.624029 systemd[1]: tcsd.service was skipped because of an unmet condition check (ConditionPathExists=/dev/tpm0). Dec 13 14:19:44.626598 systemd[1]: cgroup compatibility translation between legacy and unified hierarchy settings activated. See cgroup-compat debug messages for details. Dec 13 14:19:44.627857 systemd[1]: Starting update-engine.service... Dec 13 14:19:44.629230 extend-filesystems[1187]: Found loop1 Dec 13 14:19:44.629230 extend-filesystems[1187]: Found sr0 Dec 13 14:19:44.629230 extend-filesystems[1187]: Found vda Dec 13 14:19:44.629230 extend-filesystems[1187]: Found vda1 Dec 13 14:19:44.629230 extend-filesystems[1187]: Found vda2 Dec 13 14:19:44.629230 extend-filesystems[1187]: Found vda3 Dec 13 14:19:44.629230 extend-filesystems[1187]: Found usr Dec 13 14:19:44.629230 extend-filesystems[1187]: Found vda4 Dec 13 14:19:44.629230 extend-filesystems[1187]: Found vda6 Dec 13 14:19:44.629230 extend-filesystems[1187]: Found vda7 Dec 13 14:19:44.629230 extend-filesystems[1187]: Found vda9 Dec 13 14:19:44.686299 extend-filesystems[1187]: Checking size of /dev/vda9 Dec 13 14:19:44.686299 extend-filesystems[1187]: Resized partition /dev/vda9 Dec 13 14:19:44.692691 kernel: EXT4-fs (vda9): resizing filesystem from 553472 to 1864699 blocks Dec 13 14:19:44.666891 dbus-daemon[1185]: [system] SELinux support is enabled Dec 13 14:19:44.630274 systemd[1]: Starting update-ssh-keys-after-ignition.service... Dec 13 14:19:44.694693 extend-filesystems[1214]: resize2fs 1.46.5 (30-Dec-2021) Dec 13 14:19:44.631485 systemd[1]: xenserver-pv-version.service was skipped because of an unmet condition check (ConditionVirtualization=xen). Dec 13 14:19:44.700725 jq[1205]: true Dec 13 14:19:44.633763 systemd[1]: enable-oem-cloudinit.service: Skipped due to 'exec-condition'. Dec 13 14:19:44.634062 systemd[1]: Condition check resulted in enable-oem-cloudinit.service being skipped. Dec 13 14:19:44.701370 jq[1208]: true Dec 13 14:19:44.634461 systemd[1]: ssh-key-proc-cmdline.service: Deactivated successfully. Dec 13 14:19:44.634596 systemd[1]: Finished ssh-key-proc-cmdline.service. Dec 13 14:19:44.656083 systemd-networkd[1038]: eth0: Gained IPv6LL Dec 13 14:19:44.662302 systemd[1]: Finished systemd-networkd-wait-online.service. Dec 13 14:19:44.667273 systemd[1]: Started dbus.service. Dec 13 14:19:44.671608 systemd[1]: Reached target network-online.target. Dec 13 14:19:44.676025 systemd[1]: Starting kubelet.service... Dec 13 14:19:44.684424 systemd[1]: system-cloudinit@usr-share-oem-cloud\x2dconfig.yml.service was skipped because of an unmet condition check (ConditionFileNotEmpty=/usr/share/oem/cloud-config.yml). Dec 13 14:19:44.684454 systemd[1]: Reached target system-config.target. Dec 13 14:19:44.686205 systemd[1]: user-cloudinit-proc-cmdline.service was skipped because of an unmet condition check (ConditionKernelCommandLine=cloud-config-url). Dec 13 14:19:44.686225 systemd[1]: Reached target user-config.target. Dec 13 14:19:44.687716 systemd[1]: motdgen.service: Deactivated successfully. Dec 13 14:19:44.690627 systemd[1]: Finished motdgen.service. Dec 13 14:19:44.813619 systemd-logind[1195]: Watching system buttons on /dev/input/event1 (Power Button) Dec 13 14:19:44.813655 systemd-logind[1195]: Watching system buttons on /dev/input/event0 (AT Translated Set 2 keyboard) Dec 13 14:19:44.814020 systemd-logind[1195]: New seat seat0. Dec 13 14:19:44.816495 update_engine[1203]: I1213 14:19:44.816251 1203 main.cc:92] Flatcar Update Engine starting Dec 13 14:19:44.820957 systemd[1]: Started systemd-logind.service. Dec 13 14:19:44.825987 systemd[1]: Started update-engine.service. Dec 13 14:19:44.849459 kernel: EXT4-fs (vda9): resized filesystem to 1864699 Dec 13 14:19:44.849565 update_engine[1203]: I1213 14:19:44.827807 1203 update_check_scheduler.cc:74] Next update check in 10m29s Dec 13 14:19:44.830393 systemd[1]: Started locksmithd.service. Dec 13 14:19:44.850230 env[1209]: time="2024-12-13T14:19:44.850156925Z" level=info msg="starting containerd" revision=92b3a9d6f1b3bcc6dc74875cfdea653fe39f09c2 version=1.6.16 Dec 13 14:19:44.850649 extend-filesystems[1214]: Filesystem at /dev/vda9 is mounted on /; on-line resizing required Dec 13 14:19:44.850649 extend-filesystems[1214]: old_desc_blocks = 1, new_desc_blocks = 1 Dec 13 14:19:44.850649 extend-filesystems[1214]: The filesystem on /dev/vda9 is now 1864699 (4k) blocks long. Dec 13 14:19:44.856641 extend-filesystems[1187]: Resized filesystem in /dev/vda9 Dec 13 14:19:44.858988 bash[1236]: Updated "/home/core/.ssh/authorized_keys" Dec 13 14:19:44.854585 systemd[1]: extend-filesystems.service: Deactivated successfully. Dec 13 14:19:44.854829 systemd[1]: Finished extend-filesystems.service. Dec 13 14:19:44.857283 systemd[1]: Finished update-ssh-keys-after-ignition.service. Dec 13 14:19:44.891886 env[1209]: time="2024-12-13T14:19:44.891750548Z" level=info msg="loading plugin \"io.containerd.content.v1.content\"..." type=io.containerd.content.v1 Dec 13 14:19:44.892357 env[1209]: time="2024-12-13T14:19:44.892332228Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.aufs\"..." type=io.containerd.snapshotter.v1 Dec 13 14:19:44.895249 env[1209]: time="2024-12-13T14:19:44.895173988Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.aufs\"..." error="aufs is not supported (modprobe aufs failed: exit status 1 \"modprobe: FATAL: Module aufs not found in directory /lib/modules/5.15.173-flatcar\\n\"): skip plugin" type=io.containerd.snapshotter.v1 Dec 13 14:19:44.895354 env[1209]: time="2024-12-13T14:19:44.895328758Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." type=io.containerd.snapshotter.v1 Dec 13 14:19:44.895807 env[1209]: time="2024-12-13T14:19:44.895778962Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.btrfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.btrfs (ext4) must be a btrfs filesystem to be used with the btrfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 14:19:44.896025 env[1209]: time="2024-12-13T14:19:44.895986091Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.devmapper\"..." type=io.containerd.snapshotter.v1 Dec 13 14:19:44.896134 env[1209]: time="2024-12-13T14:19:44.896108480Z" level=warning msg="failed to load plugin io.containerd.snapshotter.v1.devmapper" error="devmapper not configured" Dec 13 14:19:44.896267 env[1209]: time="2024-12-13T14:19:44.896233885Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.native\"..." type=io.containerd.snapshotter.v1 Dec 13 14:19:44.896501 env[1209]: time="2024-12-13T14:19:44.896474607Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.overlayfs\"..." type=io.containerd.snapshotter.v1 Dec 13 14:19:44.897208 env[1209]: time="2024-12-13T14:19:44.897183797Z" level=info msg="loading plugin \"io.containerd.snapshotter.v1.zfs\"..." type=io.containerd.snapshotter.v1 Dec 13 14:19:44.897529 env[1209]: time="2024-12-13T14:19:44.897469082Z" level=info msg="skip loading plugin \"io.containerd.snapshotter.v1.zfs\"..." error="path /var/lib/containerd/io.containerd.snapshotter.v1.zfs must be a zfs filesystem to be used with the zfs snapshotter: skip plugin" type=io.containerd.snapshotter.v1 Dec 13 14:19:44.898479 env[1209]: time="2024-12-13T14:19:44.897674677Z" level=info msg="loading plugin \"io.containerd.metadata.v1.bolt\"..." type=io.containerd.metadata.v1 Dec 13 14:19:44.898660 env[1209]: time="2024-12-13T14:19:44.898620521Z" level=warning msg="could not use snapshotter devmapper in metadata plugin" error="devmapper not configured" Dec 13 14:19:44.898871 env[1209]: time="2024-12-13T14:19:44.898795630Z" level=info msg="metadata content store policy set" policy=shared Dec 13 14:19:44.906470 env[1209]: time="2024-12-13T14:19:44.906425859Z" level=info msg="loading plugin \"io.containerd.differ.v1.walking\"..." type=io.containerd.differ.v1 Dec 13 14:19:44.906711 env[1209]: time="2024-12-13T14:19:44.906686147Z" level=info msg="loading plugin \"io.containerd.event.v1.exchange\"..." type=io.containerd.event.v1 Dec 13 14:19:44.906817 env[1209]: time="2024-12-13T14:19:44.906792687Z" level=info msg="loading plugin \"io.containerd.gc.v1.scheduler\"..." type=io.containerd.gc.v1 Dec 13 14:19:44.907104 env[1209]: time="2024-12-13T14:19:44.906995688Z" level=info msg="loading plugin \"io.containerd.service.v1.introspection-service\"..." type=io.containerd.service.v1 Dec 13 14:19:44.907203 env[1209]: time="2024-12-13T14:19:44.907180905Z" level=info msg="loading plugin \"io.containerd.service.v1.containers-service\"..." type=io.containerd.service.v1 Dec 13 14:19:44.907331 env[1209]: time="2024-12-13T14:19:44.907300389Z" level=info msg="loading plugin \"io.containerd.service.v1.content-service\"..." type=io.containerd.service.v1 Dec 13 14:19:44.907499 env[1209]: time="2024-12-13T14:19:44.907473844Z" level=info msg="loading plugin \"io.containerd.service.v1.diff-service\"..." type=io.containerd.service.v1 Dec 13 14:19:44.907628 env[1209]: time="2024-12-13T14:19:44.907598628Z" level=info msg="loading plugin \"io.containerd.service.v1.images-service\"..." type=io.containerd.service.v1 Dec 13 14:19:44.907762 env[1209]: time="2024-12-13T14:19:44.907739382Z" level=info msg="loading plugin \"io.containerd.service.v1.leases-service\"..." type=io.containerd.service.v1 Dec 13 14:19:44.907885 env[1209]: time="2024-12-13T14:19:44.907862984Z" level=info msg="loading plugin \"io.containerd.service.v1.namespaces-service\"..." type=io.containerd.service.v1 Dec 13 14:19:44.907983 env[1209]: time="2024-12-13T14:19:44.907963893Z" level=info msg="loading plugin \"io.containerd.service.v1.snapshots-service\"..." type=io.containerd.service.v1 Dec 13 14:19:44.908087 env[1209]: time="2024-12-13T14:19:44.908063039Z" level=info msg="loading plugin \"io.containerd.runtime.v1.linux\"..." type=io.containerd.runtime.v1 Dec 13 14:19:44.908326 env[1209]: time="2024-12-13T14:19:44.908305294Z" level=info msg="loading plugin \"io.containerd.runtime.v2.task\"..." type=io.containerd.runtime.v2 Dec 13 14:19:44.908547 env[1209]: time="2024-12-13T14:19:44.908511711Z" level=info msg="loading plugin \"io.containerd.monitor.v1.cgroups\"..." type=io.containerd.monitor.v1 Dec 13 14:19:44.909055 env[1209]: time="2024-12-13T14:19:44.909033549Z" level=info msg="loading plugin \"io.containerd.service.v1.tasks-service\"..." type=io.containerd.service.v1 Dec 13 14:19:44.909160 env[1209]: time="2024-12-13T14:19:44.909139478Z" level=info msg="loading plugin \"io.containerd.grpc.v1.introspection\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.909247 env[1209]: time="2024-12-13T14:19:44.909227122Z" level=info msg="loading plugin \"io.containerd.internal.v1.restart\"..." type=io.containerd.internal.v1 Dec 13 14:19:44.910420 env[1209]: time="2024-12-13T14:19:44.910400543Z" level=info msg="loading plugin \"io.containerd.grpc.v1.containers\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.910592 env[1209]: time="2024-12-13T14:19:44.910573608Z" level=info msg="loading plugin \"io.containerd.grpc.v1.content\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.910677 env[1209]: time="2024-12-13T14:19:44.910657745Z" level=info msg="loading plugin \"io.containerd.grpc.v1.diff\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.911306 env[1209]: time="2024-12-13T14:19:44.911278510Z" level=info msg="loading plugin \"io.containerd.grpc.v1.events\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.911417 env[1209]: time="2024-12-13T14:19:44.911393465Z" level=info msg="loading plugin \"io.containerd.grpc.v1.healthcheck\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.911536 env[1209]: time="2024-12-13T14:19:44.911512839Z" level=info msg="loading plugin \"io.containerd.grpc.v1.images\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.911653 env[1209]: time="2024-12-13T14:19:44.911629508Z" level=info msg="loading plugin \"io.containerd.grpc.v1.leases\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.911769 env[1209]: time="2024-12-13T14:19:44.911745415Z" level=info msg="loading plugin \"io.containerd.grpc.v1.namespaces\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.911981 env[1209]: time="2024-12-13T14:19:44.911959707Z" level=info msg="loading plugin \"io.containerd.internal.v1.opt\"..." type=io.containerd.internal.v1 Dec 13 14:19:44.912349 env[1209]: time="2024-12-13T14:19:44.912310825Z" level=info msg="loading plugin \"io.containerd.grpc.v1.snapshots\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.912480 env[1209]: time="2024-12-13T14:19:44.912458683Z" level=info msg="loading plugin \"io.containerd.grpc.v1.tasks\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.912590 env[1209]: time="2024-12-13T14:19:44.912546227Z" level=info msg="loading plugin \"io.containerd.grpc.v1.version\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.912725 env[1209]: time="2024-12-13T14:19:44.912697751Z" level=info msg="loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." type=io.containerd.tracing.processor.v1 Dec 13 14:19:44.912899 env[1209]: time="2024-12-13T14:19:44.912823176Z" level=info msg="skip loading plugin \"io.containerd.tracing.processor.v1.otlp\"..." error="no OpenTelemetry endpoint: skip plugin" type=io.containerd.tracing.processor.v1 Dec 13 14:19:44.913011 env[1209]: time="2024-12-13T14:19:44.912988967Z" level=info msg="loading plugin \"io.containerd.internal.v1.tracing\"..." type=io.containerd.internal.v1 Dec 13 14:19:44.913143 env[1209]: time="2024-12-13T14:19:44.913118901Z" level=error msg="failed to initialize a tracing processor \"otlp\"" error="no OpenTelemetry endpoint: skip plugin" Dec 13 14:19:44.913273 env[1209]: time="2024-12-13T14:19:44.913253363Z" level=info msg="loading plugin \"io.containerd.grpc.v1.cri\"..." type=io.containerd.grpc.v1 Dec 13 14:19:44.913632 env[1209]: time="2024-12-13T14:19:44.913570418Z" level=info msg="Start cri plugin with config {PluginConfig:{ContainerdConfig:{Snapshotter:overlayfs DefaultRuntimeName:runc DefaultRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} UntrustedWorkloadRuntime:{Type: Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0} Runtimes:map[runc:{Type:io.containerd.runc.v2 Path: Engine: PodAnnotations:[] ContainerAnnotations:[] Root: Options:map[SystemdCgroup:true] PrivilegedWithoutHostDevices:false BaseRuntimeSpec: NetworkPluginConfDir: NetworkPluginMaxConfNum:0}] NoPivot:false DisableSnapshotAnnotations:true DiscardUnpackedLayers:false IgnoreRdtNotEnabledErrors:false} CniConfig:{NetworkPluginBinDir:/opt/cni/bin NetworkPluginConfDir:/etc/cni/net.d NetworkPluginMaxConfNum:1 NetworkPluginConfTemplate: IPPreference:} Registry:{ConfigPath: Mirrors:map[] Configs:map[] Auths:map[] Headers:map[]} ImageDecryption:{KeyModel:node} DisableTCPService:true StreamServerAddress:127.0.0.1 StreamServerPort:0 StreamIdleTimeout:4h0m0s EnableSelinux:true SelinuxCategoryRange:1024 SandboxImage:registry.k8s.io/pause:3.6 StatsCollectPeriod:10 SystemdCgroup:false EnableTLSStreaming:false X509KeyPairStreaming:{TLSCertFile: TLSKeyFile:} MaxContainerLogLineSize:16384 DisableCgroup:false DisableApparmor:false RestrictOOMScoreAdj:false MaxConcurrentDownloads:3 DisableProcMount:false UnsetSeccompProfile: TolerateMissingHugetlbController:true DisableHugetlbController:true DeviceOwnershipFromSecurityContext:false IgnoreImageDefinedVolumes:false NetNSMountsUnderStateDir:false EnableUnprivilegedPorts:false EnableUnprivilegedICMP:false} ContainerdRootDir:/var/lib/containerd ContainerdEndpoint:/run/containerd/containerd.sock RootDir:/var/lib/containerd/io.containerd.grpc.v1.cri StateDir:/run/containerd/io.containerd.grpc.v1.cri}" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.913846546Z" level=info msg="Connect containerd service" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.913904003Z" level=info msg="Get image filesystem path \"/var/lib/containerd/io.containerd.snapshotter.v1.overlayfs\"" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.914722148Z" level=error msg="failed to load cni during init, please check CRI plugin status before setting up network for pods" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.914856469Z" level=info msg="Start subscribing containerd event" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.914956056Z" level=info msg="Start recovering state" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.915024104Z" level=info msg="Start event monitor" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.915042799Z" level=info msg="Start snapshots syncer" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.915054711Z" level=info msg="Start cni network conf syncer for default" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.915069499Z" level=info msg="Start streaming server" Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.916248340Z" level=info msg=serving... address=/run/containerd/containerd.sock.ttrpc Dec 13 14:19:44.916574 env[1209]: time="2024-12-13T14:19:44.916285960Z" level=info msg=serving... address=/run/containerd/containerd.sock Dec 13 14:19:44.920044 systemd[1]: Started containerd.service. Dec 13 14:19:44.922496 env[1209]: time="2024-12-13T14:19:44.922104853Z" level=info msg="containerd successfully booted in 0.090733s" Dec 13 14:19:44.974176 locksmithd[1237]: locksmithd starting currentOperation="UPDATE_STATUS_IDLE" strategy="reboot" Dec 13 14:19:45.074179 sshd_keygen[1199]: ssh-keygen: generating new host keys: RSA ECDSA ED25519 Dec 13 14:19:45.098847 systemd[1]: Finished sshd-keygen.service. Dec 13 14:19:45.103206 systemd[1]: Starting issuegen.service... Dec 13 14:19:45.107617 systemd[1]: issuegen.service: Deactivated successfully. Dec 13 14:19:45.107801 systemd[1]: Finished issuegen.service. Dec 13 14:19:45.110855 systemd[1]: Starting systemd-user-sessions.service... Dec 13 14:19:45.129047 systemd[1]: Finished systemd-user-sessions.service. Dec 13 14:19:45.132327 systemd[1]: Started getty@tty1.service. Dec 13 14:19:45.134817 systemd[1]: Started serial-getty@ttyS0.service. Dec 13 14:19:45.136196 systemd[1]: Reached target getty.target. Dec 13 14:19:46.078532 systemd[1]: Started kubelet.service. Dec 13 14:19:46.080295 systemd[1]: Reached target multi-user.target. Dec 13 14:19:46.083164 systemd[1]: Starting systemd-update-utmp-runlevel.service... Dec 13 14:19:46.093228 systemd[1]: systemd-update-utmp-runlevel.service: Deactivated successfully. Dec 13 14:19:46.093414 systemd[1]: Finished systemd-update-utmp-runlevel.service. Dec 13 14:19:46.097574 systemd[1]: Startup finished in 898ms (kernel) + 5.257s (initrd) + 8.542s (userspace) = 14.698s. Dec 13 14:19:46.902467 kubelet[1263]: E1213 14:19:46.902327 1263 run.go:74] "command failed" err="failed to load kubelet config file, path: /var/lib/kubelet/config.yaml, error: failed to load Kubelet config file /var/lib/kubelet/config.yaml, error failed to read kubelet config file \"/var/lib/kubelet/config.yaml\", error: open /var/lib/kubelet/config.yaml: no such file or directory" Dec 13 14:19:46.904697 systemd[1]: kubelet.service: Main process exited, code=exited, status=1/FAILURE Dec 13 14:19:46.904900 systemd[1]: kubelet.service: Failed with result 'exit-code'. Dec 13 14:19:46.905293 systemd[1]: kubelet.service: Consumed 1.899s CPU time. Dec 13 14:19:53.984292 systemd[1]: Created slice system-sshd.slice. Dec 13 14:19:53.985531 systemd[1]: Started sshd@0-10.0.0.34:22-10.0.0.1:34220.service. Dec 13 14:19:54.029502 sshd[1273]: Accepted publickey for core from 10.0.0.1 port 34220 ssh2: RSA SHA256:G6GGwH/f10E2j6mIu1+COWQkyppDOEetpcI3w1A8nX8 Dec 13 14:19:54.031046 sshd[1273]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:19:54.040529 systemd-logind[1195]: New session 1 of user core. Dec 13 14:19:54.041509 systemd[1]: Created slice user-500.slice. Dec 13 14:19:54.042728 systemd[1]: Starting user-runtime-dir@500.service... Dec 13 14:19:54.051691 systemd[1]: Finished user-runtime-dir@500.service. Dec 13 14:19:54.052985 systemd[1]: Starting user@500.service... Dec 13 14:19:54.055687 (systemd)[1276]: pam_unix(systemd-user:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:19:54.154605 systemd[1276]: Queued start job for default target default.target. Dec 13 14:19:54.155234 systemd[1276]: Reached target paths.target. Dec 13 14:19:54.155261 systemd[1276]: Reached target sockets.target. Dec 13 14:19:54.155277 systemd[1276]: Reached target timers.target. Dec 13 14:19:54.155293 systemd[1276]: Reached target basic.target. Dec 13 14:19:54.155340 systemd[1276]: Reached target default.target. Dec 13 14:19:54.155372 systemd[1276]: Startup finished in 94ms. Dec 13 14:19:54.155444 systemd[1]: Started user@500.service. Dec 13 14:19:54.156699 systemd[1]: Started session-1.scope. Dec 13 14:19:54.208746 systemd[1]: Started sshd@1-10.0.0.34:22-10.0.0.1:34224.service. Dec 13 14:19:54.251556 sshd[1285]: Accepted publickey for core from 10.0.0.1 port 34224 ssh2: RSA SHA256:G6GGwH/f10E2j6mIu1+COWQkyppDOEetpcI3w1A8nX8 Dec 13 14:19:54.252701 sshd[1285]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:19:54.256732 systemd-logind[1195]: New session 2 of user core. Dec 13 14:19:54.258043 systemd[1]: Started session-2.scope. Dec 13 14:19:54.313273 sshd[1285]: pam_unix(sshd:session): session closed for user core Dec 13 14:19:54.316945 systemd[1]: Started sshd@2-10.0.0.34:22-10.0.0.1:34236.service. Dec 13 14:19:54.317467 systemd[1]: sshd@1-10.0.0.34:22-10.0.0.1:34224.service: Deactivated successfully. Dec 13 14:19:54.318043 systemd[1]: session-2.scope: Deactivated successfully. Dec 13 14:19:54.318515 systemd-logind[1195]: Session 2 logged out. Waiting for processes to exit. Dec 13 14:19:54.319274 systemd-logind[1195]: Removed session 2. Dec 13 14:19:54.356190 sshd[1290]: Accepted publickey for core from 10.0.0.1 port 34236 ssh2: RSA SHA256:G6GGwH/f10E2j6mIu1+COWQkyppDOEetpcI3w1A8nX8 Dec 13 14:19:54.357311 sshd[1290]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:19:54.360664 systemd-logind[1195]: New session 3 of user core. Dec 13 14:19:54.361458 systemd[1]: Started session-3.scope. Dec 13 14:19:54.412222 sshd[1290]: pam_unix(sshd:session): session closed for user core Dec 13 14:19:54.415229 systemd[1]: sshd@2-10.0.0.34:22-10.0.0.1:34236.service: Deactivated successfully. Dec 13 14:19:54.415790 systemd[1]: session-3.scope: Deactivated successfully. Dec 13 14:19:54.416330 systemd-logind[1195]: Session 3 logged out. Waiting for processes to exit. Dec 13 14:19:54.417509 systemd[1]: Started sshd@3-10.0.0.34:22-10.0.0.1:34240.service. Dec 13 14:19:54.418303 systemd-logind[1195]: Removed session 3. Dec 13 14:19:54.459534 sshd[1297]: Accepted publickey for core from 10.0.0.1 port 34240 ssh2: RSA SHA256:G6GGwH/f10E2j6mIu1+COWQkyppDOEetpcI3w1A8nX8 Dec 13 14:19:54.461359 sshd[1297]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:19:54.465413 systemd-logind[1195]: New session 4 of user core. Dec 13 14:19:54.466285 systemd[1]: Started session-4.scope. Dec 13 14:19:54.521936 sshd[1297]: pam_unix(sshd:session): session closed for user core Dec 13 14:19:54.525303 systemd[1]: sshd@3-10.0.0.34:22-10.0.0.1:34240.service: Deactivated successfully. Dec 13 14:19:54.525961 systemd[1]: session-4.scope: Deactivated successfully. Dec 13 14:19:54.526551 systemd-logind[1195]: Session 4 logged out. Waiting for processes to exit. Dec 13 14:19:54.527725 systemd[1]: Started sshd@4-10.0.0.34:22-10.0.0.1:34244.service. Dec 13 14:19:54.528547 systemd-logind[1195]: Removed session 4. Dec 13 14:19:54.573636 sshd[1303]: Accepted publickey for core from 10.0.0.1 port 34244 ssh2: RSA SHA256:G6GGwH/f10E2j6mIu1+COWQkyppDOEetpcI3w1A8nX8 Dec 13 14:19:54.575251 sshd[1303]: pam_unix(sshd:session): session opened for user core(uid=500) by (uid=0) Dec 13 14:19:54.579263 systemd-logind[1195]: New session 5 of user core. Dec 13 14:19:54.580222 systemd[1]: Started session-5.scope. Dec 13 14:19:54.640382 sudo[1306]: core : PWD=/home/core ; USER=root ; COMMAND=/home/core/install.sh Dec 13 14:19:54.640571 sudo[1306]: pam_unix(sudo:session): session opened for user root(uid=0) by (uid=500) Dec 13 14:19:54.655550 systemd[1]: Starting coreos-metadata.service... Dec 13 14:19:54.663208 systemd[1]: coreos-metadata.service: Deactivated successfully. Dec 13 14:19:54.663439 systemd[1]: Finished coreos-metadata.service. Dec 13 14:19:55.758268 systemd[1]: Stopped kubelet.service. Dec 13 14:19:55.758483 systemd[1]: kubelet.service: Consumed 1.899s CPU time. Dec 13 14:19:55.760626 systemd[1]: Starting kubelet.service... Dec 13 14:19:55.777248 systemd[1]: Reloading. Dec 13 14:19:55.859147 /usr/lib/systemd/system-generators/torcx-generator[1373]: time="2024-12-13T14:19:55Z" level=debug msg="common configuration parsed" base_dir=/var/lib/torcx/ conf_dir=/etc/torcx/ run_dir=/run/torcx/ store_paths="[/usr/share/torcx/store /usr/share/oem/torcx/store/3510.3.6 /usr/share/oem/torcx/store /var/lib/torcx/store/3510.3.6 /var/lib/torcx/store]" Dec 13 14:19:55.859176 /usr/lib/systemd/system-generators/torcx-generator[1373]: time="2024-12-13T14:19:55Z" level=info msg="torcx already run" Dec 13 14:19:56.083805 systemd[1]: /usr/lib/systemd/system/locksmithd.service:8: Unit uses CPUShares=; please use CPUWeight= instead. Support for CPUShares= will be removed soon. Dec 13 14:19:56.083825 systemd[1]: /usr/lib/systemd/system/locksmithd.service:9: Unit uses MemoryLimit=; please use MemoryMax= instead. Support for MemoryLimit= will be removed soon. Dec 13 14:19:56.104692 systemd[1]: /run/systemd/system/docker.socket:8: ListenStream= references a path below legacy directory /var/run/, updating /var/run/docker.sock → /run/docker.sock; please update the unit file accordingly. Dec 13 14:19:56.193965 systemd[1]: Started kubelet.service. Dec 13 14:19:56.198024 systemd[1]: Stopping kubelet.service... Dec 13 14:19:56.198616 systemd[1]: kubelet.service: Deactivated successfully. Dec 13 14:19:56.198917 systemd[1]: Stopped kubelet.service. Dec 13 14:19:56.200563 systemd[1]: Starting kubelet.service... Dec 13 14:19:56.281544 systemd[1]: Started kubelet.service. Dec 13 14:19:56.360091 kubelet[1425]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 14:19:56.360091 kubelet[1425]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 13 14:19:56.360091 kubelet[1425]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 13 14:19:56.369073 kubelet[1425]: I1213 14:19:56.368995 1425 server.go:204] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 13 14:19:56.646001 kubelet[1425]: I1213 14:19:56.645886 1425 server.go:487] "Kubelet version" kubeletVersion="v1.29.2" Dec 13 14:19:56.646001 kubelet[1425]: I1213 14:19:56.645915 1425 server.go:489] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 13 14:19:56.646205 kubelet[1425]: I1213 14:19:56.646156 1425 server.go:919] "Client rotation is on, will bootstrap in background" Dec 13 14:19:56.705544 kubelet[1425]: I1213 14:19:56.705465 1425 dynamic_cafile_content.go:157] "Starting controller" name="client-ca-bundle::/etc/kubernetes/pki/ca.crt" Dec 13 14:19:56.721269 kubelet[1425]: I1213 14:19:56.721224 1425 server.go:745] "--cgroups-per-qos enabled, but --cgroup-root was not specified. defaulting to /" Dec 13 14:19:56.723822 kubelet[1425]: I1213 14:19:56.723784 1425 container_manager_linux.go:265] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 13 14:19:56.724087 kubelet[1425]: I1213 14:19:56.724048 1425 container_manager_linux.go:270] "Creating Container Manager object based on Node Config" nodeConfig={"RuntimeCgroupsName":"","SystemCgroupsName":"","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":false,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":null,"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":-1,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null} Dec 13 14:19:56.724245 kubelet[1425]: I1213 14:19:56.724092 1425 topology_manager.go:138] "Creating topology manager with none policy" Dec 13 14:19:56.724245 kubelet[1425]: I1213 14:19:56.724105 1425 container_manager_linux.go:301] "Creating device plugin manager" Dec 13 14:19:56.724245 kubelet[1425]: I1213 14:19:56.724236 1425 state_mem.go:36] "Initialized new in-memory state store" Dec 13 14:19:56.724707 kubelet[1425]: I1213 14:19:56.724397 1425 kubelet.go:396] "Attempting to sync node with API server" Dec 13 14:19:56.724707 kubelet[1425]: I1213 14:19:56.724434 1425 kubelet.go:301] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 13 14:19:56.724707 kubelet[1425]: I1213 14:19:56.724468 1425 kubelet.go:312] "Adding apiserver pod source" Dec 13 14:19:56.724707 kubelet[1425]: I1213 14:19:56.724496 1425 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 13 14:19:56.724707 kubelet[1425]: E1213 14:19:56.724681 1425 file.go:98] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:19:56.724918 kubelet[1425]: E1213 14:19:56.724782 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:19:56.726168 kubelet[1425]: I1213 14:19:56.726129 1425 kuberuntime_manager.go:258] "Container runtime initialized" containerRuntime="containerd" version="1.6.16" apiVersion="v1" Dec 13 14:19:56.729842 kubelet[1425]: W1213 14:19:56.729788 1425 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Node: nodes "10.0.0.34" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Dec 13 14:19:56.729985 kubelet[1425]: E1213 14:19:56.729863 1425 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Node: failed to list *v1.Node: nodes "10.0.0.34" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Dec 13 14:19:56.729985 kubelet[1425]: W1213 14:19:56.729937 1425 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Dec 13 14:19:56.729985 kubelet[1425]: E1213 14:19:56.729973 1425 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Dec 13 14:19:56.731122 kubelet[1425]: I1213 14:19:56.731101 1425 kubelet.go:809] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 13 14:19:56.732197 kubelet[1425]: W1213 14:19:56.732174 1425 probe.go:268] Flexvolume plugin directory at /opt/libexec/kubernetes/kubelet-plugins/volume/exec/ does not exist. Recreating. Dec 13 14:19:56.732924 kubelet[1425]: I1213 14:19:56.732904 1425 server.go:1256] "Started kubelet" Dec 13 14:19:56.753056 kubelet[1425]: I1213 14:19:56.752887 1425 server.go:162] "Starting to listen" address="0.0.0.0" port=10250 Dec 13 14:19:56.753583 kubelet[1425]: I1213 14:19:56.733295 1425 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 13 14:19:56.754170 kubelet[1425]: I1213 14:19:56.754145 1425 server.go:233] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 13 14:19:56.755725 kernel: SELinux: Context system_u:object_r:container_file_t:s0 is not valid (left unmapped). Dec 13 14:19:56.766722 kubelet[1425]: I1213 14:19:56.766693 1425 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 13 14:19:56.767551 kubelet[1425]: I1213 14:19:56.767518 1425 volume_manager.go:291] "Starting Kubelet Volume Manager" Dec 13 14:19:56.783916 kubelet[1425]: I1213 14:19:56.783877 1425 desired_state_of_world_populator.go:151] "Desired state populator starts to run" Dec 13 14:19:56.784035 kubelet[1425]: I1213 14:19:56.784011 1425 reconciler_new.go:29] "Reconciler: start to sync state" Dec 13 14:19:56.785570 kubelet[1425]: I1213 14:19:56.785549 1425 factory.go:221] Registration of the systemd container factory successfully Dec 13 14:19:56.785983 kubelet[1425]: I1213 14:19:56.785741 1425 factory.go:219] Registration of the crio container factory failed: Get "http://%2Fvar%2Frun%2Fcrio%2Fcrio.sock/info": dial unix /var/run/crio/crio.sock: connect: no such file or directory Dec 13 14:19:56.786417 kubelet[1425]: E1213 14:19:56.786397 1425 kubelet.go:1462] "Image garbage collection failed once. Stats initialization may not have completed yet" err="invalid capacity 0 on image filesystem" Dec 13 14:19:56.788190 kubelet[1425]: I1213 14:19:56.788172 1425 factory.go:221] Registration of the containerd container factory successfully Dec 13 14:19:56.788888 kubelet[1425]: W1213 14:19:56.788819 1425 reflector.go:539] vendor/k8s.io/client-go/informers/factory.go:159: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Dec 13 14:19:56.788948 kubelet[1425]: E1213 14:19:56.788899 1425 reflector.go:147] vendor/k8s.io/client-go/informers/factory.go:159: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Dec 13 14:19:56.789155 kubelet[1425]: E1213 14:19:56.789135 1425 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"10.0.0.34\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="200ms" Dec 13 14:19:56.789297 kubelet[1425]: E1213 14:19:56.789269 1425 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.34.1810c26544694639 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.34,UID:10.0.0.34,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:10.0.0.34,},FirstTimestamp:2024-12-13 14:19:56.732876345 +0000 UTC m=+0.447361593,LastTimestamp:2024-12-13 14:19:56.732876345 +0000 UTC m=+0.447361593,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.34,}" Dec 13 14:19:56.789935 kubelet[1425]: I1213 14:19:56.789909 1425 server.go:461] "Adding debug handlers to kubelet server" Dec 13 14:19:56.791226 kubelet[1425]: E1213 14:19:56.791198 1425 event.go:346] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{10.0.0.34.1810c2654799b544 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.34,UID:10.0.0.34,APIVersion:,ResourceVersion:,FieldPath:,},Reason:InvalidDiskCapacity,Message:invalid capacity 0 on image filesystem,Source:EventSource{Component:kubelet,Host:10.0.0.34,},FirstTimestamp:2024-12-13 14:19:56.786382148 +0000 UTC m=+0.500867406,LastTimestamp:2024-12-13 14:19:56.786382148 +0000 UTC m=+0.500867406,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.34,}" Dec 13 14:19:56.800431 kubelet[1425]: I1213 14:19:56.800414 1425 cpu_manager.go:214] "Starting CPU manager" policy="none" Dec 13 14:19:56.800552 kubelet[1425]: I1213 14:19:56.800514 1425 cpu_manager.go:215] "Reconciling" reconcilePeriod="10s" Dec 13 14:19:56.800552 kubelet[1425]: I1213 14:19:56.800541 1425 state_mem.go:36] "Initialized new in-memory state store" Dec 13 14:19:56.869420 kubelet[1425]: I1213 14:19:56.869353 1425 kubelet_node_status.go:73] "Attempting to register node" node="10.0.0.34" Dec 13 14:19:57.515444 kubelet[1425]: I1213 14:19:57.515374 1425 kubelet_node_status.go:76] "Successfully registered node" node="10.0.0.34" Dec 13 14:19:57.517245 kubelet[1425]: I1213 14:19:57.517206 1425 kuberuntime_manager.go:1529] "Updating runtime config through cri with podcidr" CIDR="192.168.1.0/24" Dec 13 14:19:57.517726 env[1209]: time="2024-12-13T14:19:57.517669152Z" level=info msg="No cni config template is specified, wait for other system components to drop the config." Dec 13 14:19:57.518107 kubelet[1425]: I1213 14:19:57.518076 1425 kubelet_network.go:61] "Updating Pod CIDR" originalPodCIDR="" newPodCIDR="192.168.1.0/24" Dec 13 14:19:57.544822 kubelet[1425]: I1213 14:19:57.544777 1425 policy_none.go:49] "None policy: Start" Dec 13 14:19:57.545545 kubelet[1425]: I1213 14:19:57.545527 1425 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 13 14:19:57.545606 kubelet[1425]: I1213 14:19:57.545566 1425 state_mem.go:35] "Initializing new in-memory state store" Dec 13 14:19:57.574168 kubelet[1425]: I1213 14:19:57.574134 1425 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 13 14:19:57.575165 kubelet[1425]: I1213 14:19:57.575148 1425 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 13 14:19:57.575336 kubelet[1425]: I1213 14:19:57.575307 1425 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 13 14:19:57.575395 kubelet[1425]: I1213 14:19:57.575367 1425 kubelet.go:2329] "Starting kubelet main sync loop" Dec 13 14:19:57.575821 kubelet[1425]: E1213 14:19:57.575788 1425 kubelet.go:2353] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 13 14:19:57.637049 systemd[1]: Created slice kubepods.slice. Dec 13 14:19:57.640925 systemd[1]: Created slice kubepods-besteffort.slice. Dec 13 14:19:57.648884 kubelet[1425]: I1213 14:19:57.647861 1425 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 13 14:19:57.648884 kubelet[1425]: W1213 14:19:57.648155 1425 reflector.go:462] vendor/k8s.io/client-go/informers/factory.go:159: watch of *v1.Service ended with: very short watch: vendor/k8s.io/client-go/informers/factory.go:159: Unexpected watch close - watch lasted less than a second and no items received Dec 13 14:19:57.648884 kubelet[1425]: E1213 14:19:57.648377 1425 event.go:355] "Unable to write event (may retry after sleeping)" err="Post \"https://10.0.0.27:6443/api/v1/namespaces/default/events\": read tcp 10.0.0.34:60516->10.0.0.27:6443: use of closed network connection" event="&Event{ObjectMeta:{10.0.0.34.1810c26548633a6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:10.0.0.34,UID:10.0.0.34,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node 10.0.0.34 status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:10.0.0.34,},FirstTimestamp:2024-12-13 14:19:56.799588975 +0000 UTC m=+0.514074223,LastTimestamp:2024-12-13 14:19:56.799588975 +0000 UTC m=+0.514074223,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:10.0.0.34,}" Dec 13 14:19:57.649131 kubelet[1425]: E1213 14:19:57.648483 1425 kubelet_node_status.go:544] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2024-12-13T14:19:57Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2024-12-13T14:19:57Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2024-12-13T14:19:57Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2024-12-13T14:19:57Z\\\",\\\"type\\\":\\\"Ready\\\"}]}}\" for node \"10.0.0.34\": Patch \"https://10.0.0.27:6443/api/v1/nodes/10.0.0.34/status?timeout=10s\": read tcp 10.0.0.34:60516->10.0.0.27:6443: use of closed network connection" Dec 13 14:19:57.649131 kubelet[1425]: E1213 14:19:57.648802 1425 controller.go:145] "Failed to ensure lease exists, will retry" err="Post \"https://10.0.0.27:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases?timeout=10s\": read tcp 10.0.0.34:60516->10.0.0.27:6443: use of closed network connection" interval="400ms" Dec 13 14:19:57.649131 kubelet[1425]: W1213 14:19:57.648969 1425 reflector.go:462] vendor/k8s.io/client-go/informers/factory.go:159: watch of *v1.RuntimeClass ended with: very short watch: vendor/k8s.io/client-go/informers/factory.go:159: Unexpected watch close - watch lasted less than a second and no items received Dec 13 14:19:57.649131 kubelet[1425]: W1213 14:19:57.649083 1425 reflector.go:462] vendor/k8s.io/client-go/informers/factory.go:159: watch of *v1.CSIDriver ended with: very short watch: vendor/k8s.io/client-go/informers/factory.go:159: Unexpected watch close - watch lasted less than a second and no items received Dec 13 14:19:57.657560 systemd[1]: Created slice kubepods-burstable.slice. Dec 13 14:19:57.659233 kubelet[1425]: I1213 14:19:57.659193 1425 manager.go:479] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 13 14:19:57.659569 kubelet[1425]: I1213 14:19:57.659545 1425 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 13 14:19:57.661072 kubelet[1425]: E1213 14:19:57.661038 1425 eviction_manager.go:282] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"10.0.0.34\" not found" Dec 13 14:19:57.725316 kubelet[1425]: E1213 14:19:57.725241 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:19:57.922301 sudo[1306]: pam_unix(sudo:session): session closed for user root Dec 13 14:19:57.924234 sshd[1303]: pam_unix(sshd:session): session closed for user core Dec 13 14:19:57.926878 systemd[1]: sshd@4-10.0.0.34:22-10.0.0.1:34244.service: Deactivated successfully. Dec 13 14:19:57.927543 systemd[1]: session-5.scope: Deactivated successfully. Dec 13 14:19:57.928274 systemd-logind[1195]: Session 5 logged out. Waiting for processes to exit. Dec 13 14:19:57.929225 systemd-logind[1195]: Removed session 5. Dec 13 14:19:58.725881 kubelet[1425]: E1213 14:19:58.725803 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:19:58.725881 kubelet[1425]: I1213 14:19:58.725819 1425 apiserver.go:52] "Watching apiserver" Dec 13 14:19:58.728970 kubelet[1425]: I1213 14:19:58.728945 1425 topology_manager.go:215] "Topology Admit Handler" podUID="fd8b627b-eb28-4788-976f-4daf13ac261b" podNamespace="kube-system" podName="cilium-84rkv" Dec 13 14:19:58.729124 kubelet[1425]: I1213 14:19:58.729084 1425 topology_manager.go:215] "Topology Admit Handler" podUID="a7237ffe-7da0-46a1-af75-1b2588be01dd" podNamespace="kube-system" podName="kube-proxy-kthkk" Dec 13 14:19:58.733265 systemd[1]: Created slice kubepods-besteffort-poda7237ffe_7da0_46a1_af75_1b2588be01dd.slice. Dec 13 14:19:58.740518 systemd[1]: Created slice kubepods-burstable-podfd8b627b_eb28_4788_976f_4daf13ac261b.slice. Dec 13 14:19:58.785059 kubelet[1425]: I1213 14:19:58.785020 1425 desired_state_of_world_populator.go:159] "Finished populating initial desired state of world" Dec 13 14:19:58.796200 kubelet[1425]: I1213 14:19:58.796175 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cni-path\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796324 kubelet[1425]: I1213 14:19:58.796217 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-xtables-lock\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796324 kubelet[1425]: I1213 14:19:58.796249 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-host-proc-sys-net\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796324 kubelet[1425]: I1213 14:19:58.796276 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/fd8b627b-eb28-4788-976f-4daf13ac261b-hubble-tls\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796434 kubelet[1425]: I1213 14:19:58.796373 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fht6b\" (UniqueName: \"kubernetes.io/projected/fd8b627b-eb28-4788-976f-4daf13ac261b-kube-api-access-fht6b\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796471 kubelet[1425]: I1213 14:19:58.796434 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-proxy\" (UniqueName: \"kubernetes.io/configmap/a7237ffe-7da0-46a1-af75-1b2588be01dd-kube-proxy\") pod \"kube-proxy-kthkk\" (UID: \"a7237ffe-7da0-46a1-af75-1b2588be01dd\") " pod="kube-system/kube-proxy-kthkk" Dec 13 14:19:58.796471 kubelet[1425]: I1213 14:19:58.796470 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-run\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796534 kubelet[1425]: I1213 14:19:58.796500 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-bpf-maps\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796575 kubelet[1425]: I1213 14:19:58.796562 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-etc-cni-netd\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796623 kubelet[1425]: I1213 14:19:58.796609 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-hostproc\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796685 kubelet[1425]: I1213 14:19:58.796664 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/fd8b627b-eb28-4788-976f-4daf13ac261b-clustermesh-secrets\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796725 kubelet[1425]: I1213 14:19:58.796704 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-config-path\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796758 kubelet[1425]: I1213 14:19:58.796748 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/a7237ffe-7da0-46a1-af75-1b2588be01dd-xtables-lock\") pod \"kube-proxy-kthkk\" (UID: \"a7237ffe-7da0-46a1-af75-1b2588be01dd\") " pod="kube-system/kube-proxy-kthkk" Dec 13 14:19:58.796795 kubelet[1425]: I1213 14:19:58.796790 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a7237ffe-7da0-46a1-af75-1b2588be01dd-lib-modules\") pod \"kube-proxy-kthkk\" (UID: \"a7237ffe-7da0-46a1-af75-1b2588be01dd\") " pod="kube-system/kube-proxy-kthkk" Dec 13 14:19:58.796879 kubelet[1425]: I1213 14:19:58.796861 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wh9s\" (UniqueName: \"kubernetes.io/projected/a7237ffe-7da0-46a1-af75-1b2588be01dd-kube-api-access-5wh9s\") pod \"kube-proxy-kthkk\" (UID: \"a7237ffe-7da0-46a1-af75-1b2588be01dd\") " pod="kube-system/kube-proxy-kthkk" Dec 13 14:19:58.796917 kubelet[1425]: I1213 14:19:58.796894 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-cgroup\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796917 kubelet[1425]: I1213 14:19:58.796918 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-lib-modules\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:58.796972 kubelet[1425]: I1213 14:19:58.796942 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-host-proc-sys-kernel\") pod \"cilium-84rkv\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " pod="kube-system/cilium-84rkv" Dec 13 14:19:59.039642 kubelet[1425]: E1213 14:19:59.039516 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:19:59.040511 env[1209]: time="2024-12-13T14:19:59.040402861Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-kthkk,Uid:a7237ffe-7da0-46a1-af75-1b2588be01dd,Namespace:kube-system,Attempt:0,}" Dec 13 14:19:59.051304 kubelet[1425]: E1213 14:19:59.051279 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:19:59.051820 env[1209]: time="2024-12-13T14:19:59.051772914Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-84rkv,Uid:fd8b627b-eb28-4788-976f-4daf13ac261b,Namespace:kube-system,Attempt:0,}" Dec 13 14:19:59.727031 kubelet[1425]: E1213 14:19:59.726963 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:00.109688 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1900694430.mount: Deactivated successfully. Dec 13 14:20:00.120034 env[1209]: time="2024-12-13T14:20:00.119960617Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:00.120997 env[1209]: time="2024-12-13T14:20:00.120970661Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:00.123745 env[1209]: time="2024-12-13T14:20:00.123716421Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:00.125339 env[1209]: time="2024-12-13T14:20:00.125304078Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:6270bb605e12e581514ada5fd5b3216f727db55dc87d5889c790e4c760683fee,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:00.127025 env[1209]: time="2024-12-13T14:20:00.126966376Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:00.131122 env[1209]: time="2024-12-13T14:20:00.131083797Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause:3.6,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:00.133878 env[1209]: time="2024-12-13T14:20:00.133842771Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:00.135071 env[1209]: time="2024-12-13T14:20:00.135043423Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/pause@sha256:3d380ca8864549e74af4b29c10f9cb0956236dfb01c40ca076fb6c37253234db,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:00.178556 env[1209]: time="2024-12-13T14:20:00.178431591Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:20:00.178556 env[1209]: time="2024-12-13T14:20:00.178508686Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:20:00.178824 env[1209]: time="2024-12-13T14:20:00.178520197Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:20:00.179151 env[1209]: time="2024-12-13T14:20:00.179104724Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/6906f1f1534b02fbb4d537b92719e8e4b6f7d0b1ec5fec0b4979cbc07337eb6f pid=1487 runtime=io.containerd.runc.v2 Dec 13 14:20:00.184582 env[1209]: time="2024-12-13T14:20:00.184502286Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:20:00.184582 env[1209]: time="2024-12-13T14:20:00.184540568Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:20:00.184582 env[1209]: time="2024-12-13T14:20:00.184553502Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:20:00.184747 env[1209]: time="2024-12-13T14:20:00.184696550Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569 pid=1488 runtime=io.containerd.runc.v2 Dec 13 14:20:00.208689 systemd[1]: Started cri-containerd-5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569.scope. Dec 13 14:20:00.209856 systemd[1]: Started cri-containerd-6906f1f1534b02fbb4d537b92719e8e4b6f7d0b1ec5fec0b4979cbc07337eb6f.scope. Dec 13 14:20:00.283164 env[1209]: time="2024-12-13T14:20:00.283091816Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-84rkv,Uid:fd8b627b-eb28-4788-976f-4daf13ac261b,Namespace:kube-system,Attempt:0,} returns sandbox id \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\"" Dec 13 14:20:00.284527 kubelet[1425]: E1213 14:20:00.284502 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:00.286596 env[1209]: time="2024-12-13T14:20:00.286544532Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\"" Dec 13 14:20:00.295443 env[1209]: time="2024-12-13T14:20:00.295054561Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:kube-proxy-kthkk,Uid:a7237ffe-7da0-46a1-af75-1b2588be01dd,Namespace:kube-system,Attempt:0,} returns sandbox id \"6906f1f1534b02fbb4d537b92719e8e4b6f7d0b1ec5fec0b4979cbc07337eb6f\"" Dec 13 14:20:00.295892 kubelet[1425]: E1213 14:20:00.295867 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:00.727803 kubelet[1425]: E1213 14:20:00.727750 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:01.728440 kubelet[1425]: E1213 14:20:01.728382 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:02.729210 kubelet[1425]: E1213 14:20:02.729149 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:03.729707 kubelet[1425]: E1213 14:20:03.729638 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:04.730323 kubelet[1425]: E1213 14:20:04.730255 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:05.731143 kubelet[1425]: E1213 14:20:05.731070 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:06.732204 kubelet[1425]: E1213 14:20:06.732115 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:07.732567 kubelet[1425]: E1213 14:20:07.732501 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:08.733723 kubelet[1425]: E1213 14:20:08.733663 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:09.734735 kubelet[1425]: E1213 14:20:09.734639 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:10.735860 kubelet[1425]: E1213 14:20:10.735756 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:11.736496 kubelet[1425]: E1213 14:20:11.736411 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:12.692713 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount127162278.mount: Deactivated successfully. Dec 13 14:20:12.736858 kubelet[1425]: E1213 14:20:12.736771 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:13.737190 kubelet[1425]: E1213 14:20:13.737141 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:14.738593 kubelet[1425]: E1213 14:20:14.738143 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:15.739711 kubelet[1425]: E1213 14:20:15.739653 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:16.725312 kubelet[1425]: E1213 14:20:16.725263 1425 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:16.740550 kubelet[1425]: E1213 14:20:16.740521 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:17.741630 kubelet[1425]: E1213 14:20:17.741554 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:17.992184 env[1209]: time="2024-12-13T14:20:17.992028953Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:17.997139 env[1209]: time="2024-12-13T14:20:17.997058632Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:17.999292 env[1209]: time="2024-12-13T14:20:17.999255001Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/cilium@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:17.999682 env[1209]: time="2024-12-13T14:20:17.999651951Z" level=info msg="PullImage \"quay.io/cilium/cilium:v1.12.5@sha256:06ce2b0a0a472e73334a7504ee5c5d8b2e2d7b72ef728ad94e564740dd505be5\" returns image reference \"sha256:3e35b3e9f295e7748482d40ed499b0ff7961f1f128d479d8e6682b3245bba69b\"" Dec 13 14:20:18.000877 env[1209]: time="2024-12-13T14:20:18.000821653Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.12\"" Dec 13 14:20:18.001948 env[1209]: time="2024-12-13T14:20:18.001896019Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 13 14:20:18.028044 env[1209]: time="2024-12-13T14:20:18.027972423Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\"" Dec 13 14:20:18.028873 env[1209]: time="2024-12-13T14:20:18.028815888Z" level=info msg="StartContainer for \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\"" Dec 13 14:20:18.052196 systemd[1]: Started cri-containerd-fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b.scope. Dec 13 14:20:18.079757 env[1209]: time="2024-12-13T14:20:18.079710728Z" level=info msg="StartContainer for \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\" returns successfully" Dec 13 14:20:18.092451 systemd[1]: cri-containerd-fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b.scope: Deactivated successfully. Dec 13 14:20:18.527269 env[1209]: time="2024-12-13T14:20:18.527210497Z" level=info msg="shim disconnected" id=fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b Dec 13 14:20:18.527269 env[1209]: time="2024-12-13T14:20:18.527261394Z" level=warning msg="cleaning up after shim disconnected" id=fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b namespace=k8s.io Dec 13 14:20:18.527269 env[1209]: time="2024-12-13T14:20:18.527269770Z" level=info msg="cleaning up dead shim" Dec 13 14:20:18.533210 env[1209]: time="2024-12-13T14:20:18.533154562Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:20:18Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1607 runtime=io.containerd.runc.v2\n" Dec 13 14:20:18.610757 kubelet[1425]: E1213 14:20:18.610718 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:18.612531 env[1209]: time="2024-12-13T14:20:18.612488270Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 13 14:20:18.628057 env[1209]: time="2024-12-13T14:20:18.627989686Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\"" Dec 13 14:20:18.628427 env[1209]: time="2024-12-13T14:20:18.628400703Z" level=info msg="StartContainer for \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\"" Dec 13 14:20:18.641619 systemd[1]: Started cri-containerd-46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58.scope. Dec 13 14:20:18.667755 env[1209]: time="2024-12-13T14:20:18.667662212Z" level=info msg="StartContainer for \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\" returns successfully" Dec 13 14:20:18.675554 systemd[1]: systemd-sysctl.service: Deactivated successfully. Dec 13 14:20:18.675799 systemd[1]: Stopped systemd-sysctl.service. Dec 13 14:20:18.676005 systemd[1]: Stopping systemd-sysctl.service... Dec 13 14:20:18.677824 systemd[1]: Starting systemd-sysctl.service... Dec 13 14:20:18.678126 systemd[1]: cri-containerd-46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58.scope: Deactivated successfully. Dec 13 14:20:18.686676 systemd[1]: Finished systemd-sysctl.service. Dec 13 14:20:18.708421 env[1209]: time="2024-12-13T14:20:18.708352523Z" level=info msg="shim disconnected" id=46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58 Dec 13 14:20:18.708421 env[1209]: time="2024-12-13T14:20:18.708401457Z" level=warning msg="cleaning up after shim disconnected" id=46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58 namespace=k8s.io Dec 13 14:20:18.708421 env[1209]: time="2024-12-13T14:20:18.708431334Z" level=info msg="cleaning up dead shim" Dec 13 14:20:18.714438 env[1209]: time="2024-12-13T14:20:18.714386281Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:20:18Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1671 runtime=io.containerd.runc.v2\n" Dec 13 14:20:18.742082 kubelet[1425]: E1213 14:20:18.742026 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:19.015087 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b-rootfs.mount: Deactivated successfully. Dec 13 14:20:19.603875 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3910772838.mount: Deactivated successfully. Dec 13 14:20:19.614755 kubelet[1425]: E1213 14:20:19.614714 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:19.616554 env[1209]: time="2024-12-13T14:20:19.616498709Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 13 14:20:19.640788 env[1209]: time="2024-12-13T14:20:19.640732750Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\"" Dec 13 14:20:19.641296 env[1209]: time="2024-12-13T14:20:19.641274705Z" level=info msg="StartContainer for \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\"" Dec 13 14:20:19.658664 systemd[1]: Started cri-containerd-cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48.scope. Dec 13 14:20:19.688901 env[1209]: time="2024-12-13T14:20:19.688825663Z" level=info msg="StartContainer for \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\" returns successfully" Dec 13 14:20:19.689393 systemd[1]: cri-containerd-cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48.scope: Deactivated successfully. Dec 13 14:20:19.743060 kubelet[1425]: E1213 14:20:19.742990 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:19.826225 env[1209]: time="2024-12-13T14:20:19.826153748Z" level=info msg="shim disconnected" id=cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48 Dec 13 14:20:19.826225 env[1209]: time="2024-12-13T14:20:19.826213562Z" level=warning msg="cleaning up after shim disconnected" id=cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48 namespace=k8s.io Dec 13 14:20:19.826225 env[1209]: time="2024-12-13T14:20:19.826224222Z" level=info msg="cleaning up dead shim" Dec 13 14:20:19.833158 env[1209]: time="2024-12-13T14:20:19.833088362Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:20:19Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1725 runtime=io.containerd.runc.v2\n" Dec 13 14:20:20.185802 env[1209]: time="2024-12-13T14:20:20.185739162Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy:v1.29.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:20.187967 env[1209]: time="2024-12-13T14:20:20.187898565Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:20.189425 env[1209]: time="2024-12-13T14:20:20.189393298Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/kube-proxy:v1.29.12,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:20.191055 env[1209]: time="2024-12-13T14:20:20.191020544Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/kube-proxy@sha256:bc761494b78fa152a759457f42bc9b86ee9d18f5929bb127bd5f72f8e2112c39,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:20.191507 env[1209]: time="2024-12-13T14:20:20.191476535Z" level=info msg="PullImage \"registry.k8s.io/kube-proxy:v1.29.12\" returns image reference \"sha256:d699d5830022f9e67c3271d1c2af58eaede81e3567df82728b7d2a8bf12ed153\"" Dec 13 14:20:20.193139 env[1209]: time="2024-12-13T14:20:20.193099975Z" level=info msg="CreateContainer within sandbox \"6906f1f1534b02fbb4d537b92719e8e4b6f7d0b1ec5fec0b4979cbc07337eb6f\" for container &ContainerMetadata{Name:kube-proxy,Attempt:0,}" Dec 13 14:20:20.208369 env[1209]: time="2024-12-13T14:20:20.208329289Z" level=info msg="CreateContainer within sandbox \"6906f1f1534b02fbb4d537b92719e8e4b6f7d0b1ec5fec0b4979cbc07337eb6f\" for &ContainerMetadata{Name:kube-proxy,Attempt:0,} returns container id \"8fb4fb18370672d8caf4d2aea12c0192302da62911e2293b0500012c2f9501c1\"" Dec 13 14:20:20.208787 env[1209]: time="2024-12-13T14:20:20.208740504Z" level=info msg="StartContainer for \"8fb4fb18370672d8caf4d2aea12c0192302da62911e2293b0500012c2f9501c1\"" Dec 13 14:20:20.226915 systemd[1]: Started cri-containerd-8fb4fb18370672d8caf4d2aea12c0192302da62911e2293b0500012c2f9501c1.scope. Dec 13 14:20:20.256115 env[1209]: time="2024-12-13T14:20:20.256060143Z" level=info msg="StartContainer for \"8fb4fb18370672d8caf4d2aea12c0192302da62911e2293b0500012c2f9501c1\" returns successfully" Dec 13 14:20:20.616690 kubelet[1425]: E1213 14:20:20.616666 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:20.618321 kubelet[1425]: E1213 14:20:20.618305 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:20.619861 env[1209]: time="2024-12-13T14:20:20.619809280Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 13 14:20:20.624786 kubelet[1425]: I1213 14:20:20.624761 1425 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/kube-proxy-kthkk" podStartSLOduration=3.729301232 podStartE2EDuration="23.624691702s" podCreationTimestamp="2024-12-13 14:19:57 +0000 UTC" firstStartedPulling="2024-12-13 14:20:00.29638712 +0000 UTC m=+4.010872358" lastFinishedPulling="2024-12-13 14:20:20.19177758 +0000 UTC m=+23.906262828" observedRunningTime="2024-12-13 14:20:20.624019509 +0000 UTC m=+24.338504757" watchObservedRunningTime="2024-12-13 14:20:20.624691702 +0000 UTC m=+24.339176950" Dec 13 14:20:20.635753 env[1209]: time="2024-12-13T14:20:20.635715708Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\"" Dec 13 14:20:20.636203 env[1209]: time="2024-12-13T14:20:20.636181468Z" level=info msg="StartContainer for \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\"" Dec 13 14:20:20.648942 systemd[1]: Started cri-containerd-72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3.scope. Dec 13 14:20:20.667624 systemd[1]: cri-containerd-72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3.scope: Deactivated successfully. Dec 13 14:20:20.668630 env[1209]: time="2024-12-13T14:20:20.668582526Z" level=info msg="StartContainer for \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\" returns successfully" Dec 13 14:20:20.743136 kubelet[1425]: E1213 14:20:20.743075 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:20.833475 env[1209]: time="2024-12-13T14:20:20.833427075Z" level=info msg="shim disconnected" id=72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3 Dec 13 14:20:20.833475 env[1209]: time="2024-12-13T14:20:20.833469957Z" level=warning msg="cleaning up after shim disconnected" id=72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3 namespace=k8s.io Dec 13 14:20:20.833475 env[1209]: time="2024-12-13T14:20:20.833479876Z" level=info msg="cleaning up dead shim" Dec 13 14:20:20.839146 env[1209]: time="2024-12-13T14:20:20.839094846Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:20:20Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=1945 runtime=io.containerd.runc.v2\n" Dec 13 14:20:21.621192 kubelet[1425]: E1213 14:20:21.621164 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:21.621416 kubelet[1425]: E1213 14:20:21.621164 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:21.623119 env[1209]: time="2024-12-13T14:20:21.623085394Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 13 14:20:21.743171 kubelet[1425]: E1213 14:20:21.743124 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:21.840419 env[1209]: time="2024-12-13T14:20:21.840345189Z" level=info msg="CreateContainer within sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\"" Dec 13 14:20:21.841064 env[1209]: time="2024-12-13T14:20:21.841015968Z" level=info msg="StartContainer for \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\"" Dec 13 14:20:21.862923 systemd[1]: Started cri-containerd-58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af.scope. Dec 13 14:20:21.894414 env[1209]: time="2024-12-13T14:20:21.894266508Z" level=info msg="StartContainer for \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\" returns successfully" Dec 13 14:20:22.015337 systemd[1]: run-containerd-runc-k8s.io-58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af-runc.DnKmBZ.mount: Deactivated successfully. Dec 13 14:20:22.049951 kubelet[1425]: I1213 14:20:22.049908 1425 kubelet_node_status.go:497] "Fast updating node status as it just became ready" Dec 13 14:20:22.255858 kernel: Initializing XFRM netlink socket Dec 13 14:20:22.626673 kubelet[1425]: E1213 14:20:22.626635 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:22.670130 kubelet[1425]: I1213 14:20:22.670103 1425 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-84rkv" podStartSLOduration=7.9555790779999995 podStartE2EDuration="25.6700567s" podCreationTimestamp="2024-12-13 14:19:57 +0000 UTC" firstStartedPulling="2024-12-13 14:20:00.285699807 +0000 UTC m=+4.000185055" lastFinishedPulling="2024-12-13 14:20:18.000177408 +0000 UTC m=+21.714662677" observedRunningTime="2024-12-13 14:20:22.669616131 +0000 UTC m=+26.384101379" watchObservedRunningTime="2024-12-13 14:20:22.6700567 +0000 UTC m=+26.384541948" Dec 13 14:20:22.743761 kubelet[1425]: E1213 14:20:22.743692 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:23.628095 kubelet[1425]: E1213 14:20:23.628050 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:23.744542 kubelet[1425]: E1213 14:20:23.744475 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:23.896933 systemd-networkd[1038]: cilium_host: Link UP Dec 13 14:20:23.897044 systemd-networkd[1038]: cilium_net: Link UP Dec 13 14:20:23.898855 systemd-networkd[1038]: cilium_net: Gained carrier Dec 13 14:20:23.899947 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_net: link becomes ready Dec 13 14:20:23.900012 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): cilium_host: link becomes ready Dec 13 14:20:23.900122 systemd-networkd[1038]: cilium_host: Gained carrier Dec 13 14:20:23.900269 systemd-networkd[1038]: cilium_net: Gained IPv6LL Dec 13 14:20:23.900421 systemd-networkd[1038]: cilium_host: Gained IPv6LL Dec 13 14:20:23.984674 systemd-networkd[1038]: cilium_vxlan: Link UP Dec 13 14:20:23.984686 systemd-networkd[1038]: cilium_vxlan: Gained carrier Dec 13 14:20:24.202873 kernel: NET: Registered PF_ALG protocol family Dec 13 14:20:24.585465 kubelet[1425]: I1213 14:20:24.585423 1425 topology_manager.go:215] "Topology Admit Handler" podUID="fa721a04-0255-48e5-bd49-14c7b9e37bfd" podNamespace="default" podName="nginx-deployment-6d5f899847-sb7n2" Dec 13 14:20:24.590526 systemd[1]: Created slice kubepods-besteffort-podfa721a04_0255_48e5_bd49_14c7b9e37bfd.slice. Dec 13 14:20:24.630036 kubelet[1425]: E1213 14:20:24.629997 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:24.674677 kubelet[1425]: I1213 14:20:24.674624 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6ktc2\" (UniqueName: \"kubernetes.io/projected/fa721a04-0255-48e5-bd49-14c7b9e37bfd-kube-api-access-6ktc2\") pod \"nginx-deployment-6d5f899847-sb7n2\" (UID: \"fa721a04-0255-48e5-bd49-14c7b9e37bfd\") " pod="default/nginx-deployment-6d5f899847-sb7n2" Dec 13 14:20:24.745061 kubelet[1425]: E1213 14:20:24.745009 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:24.768085 systemd-networkd[1038]: lxc_health: Link UP Dec 13 14:20:24.797603 systemd-networkd[1038]: lxc_health: Gained carrier Dec 13 14:20:24.798118 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Dec 13 14:20:24.893461 env[1209]: time="2024-12-13T14:20:24.893324672Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-sb7n2,Uid:fa721a04-0255-48e5-bd49-14c7b9e37bfd,Namespace:default,Attempt:0,}" Dec 13 14:20:24.945356 systemd-networkd[1038]: lxc127a0dfb0e2a: Link UP Dec 13 14:20:24.956863 kernel: eth0: renamed from tmp90e63 Dec 13 14:20:24.963372 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Dec 13 14:20:24.963459 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc127a0dfb0e2a: link becomes ready Dec 13 14:20:24.963537 systemd-networkd[1038]: lxc127a0dfb0e2a: Gained carrier Dec 13 14:20:25.632574 kubelet[1425]: E1213 14:20:25.632504 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:25.745940 kubelet[1425]: E1213 14:20:25.745865 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:26.107776 systemd-networkd[1038]: cilium_vxlan: Gained IPv6LL Dec 13 14:20:26.319080 systemd-networkd[1038]: lxc127a0dfb0e2a: Gained IPv6LL Dec 13 14:20:26.633771 kubelet[1425]: E1213 14:20:26.633725 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:26.746340 kubelet[1425]: E1213 14:20:26.746254 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:27.151027 systemd-networkd[1038]: lxc_health: Gained IPv6LL Dec 13 14:20:27.634938 kubelet[1425]: E1213 14:20:27.634904 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:20:27.747111 kubelet[1425]: E1213 14:20:27.746943 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:28.748089 kubelet[1425]: E1213 14:20:28.747995 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:29.748971 kubelet[1425]: E1213 14:20:29.748931 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:30.066327 env[1209]: time="2024-12-13T14:20:30.066197962Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:20:30.066327 env[1209]: time="2024-12-13T14:20:30.066246735Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:20:30.066327 env[1209]: time="2024-12-13T14:20:30.066259819Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:20:30.067059 env[1209]: time="2024-12-13T14:20:30.066459377Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/90e630eff5a234fbd33668d512caf0ef45e35f6c81589dd3b86818359d87aa7c pid=2499 runtime=io.containerd.runc.v2 Dec 13 14:20:30.084489 systemd[1]: Started cri-containerd-90e630eff5a234fbd33668d512caf0ef45e35f6c81589dd3b86818359d87aa7c.scope. Dec 13 14:20:30.096399 systemd-resolved[1151]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 14:20:30.126171 env[1209]: time="2024-12-13T14:20:30.126101923Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nginx-deployment-6d5f899847-sb7n2,Uid:fa721a04-0255-48e5-bd49-14c7b9e37bfd,Namespace:default,Attempt:0,} returns sandbox id \"90e630eff5a234fbd33668d512caf0ef45e35f6c81589dd3b86818359d87aa7c\"" Dec 13 14:20:30.127899 env[1209]: time="2024-12-13T14:20:30.127870422Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Dec 13 14:20:30.399642 update_engine[1203]: I1213 14:20:30.399439 1203 update_attempter.cc:509] Updating boot flags... Dec 13 14:20:30.749998 kubelet[1425]: E1213 14:20:30.749826 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:31.762909 kubelet[1425]: E1213 14:20:31.750642 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:32.751440 kubelet[1425]: E1213 14:20:32.751379 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:33.420014 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3593613190.mount: Deactivated successfully. Dec 13 14:20:33.751785 kubelet[1425]: E1213 14:20:33.751651 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:34.751827 kubelet[1425]: E1213 14:20:34.751769 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:35.752906 kubelet[1425]: E1213 14:20:35.752817 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:35.910007 env[1209]: time="2024-12-13T14:20:35.909945216Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:35.911927 env[1209]: time="2024-12-13T14:20:35.911897231Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:35.913819 env[1209]: time="2024-12-13T14:20:35.913780347Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:35.915381 env[1209]: time="2024-12-13T14:20:35.915341174Z" level=info msg="ImageCreate event &ImageCreate{Name:ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:35.916821 env[1209]: time="2024-12-13T14:20:35.916758410Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\"" Dec 13 14:20:35.918635 env[1209]: time="2024-12-13T14:20:35.918592844Z" level=info msg="CreateContainer within sandbox \"90e630eff5a234fbd33668d512caf0ef45e35f6c81589dd3b86818359d87aa7c\" for container &ContainerMetadata{Name:nginx,Attempt:0,}" Dec 13 14:20:35.931934 env[1209]: time="2024-12-13T14:20:35.931885092Z" level=info msg="CreateContainer within sandbox \"90e630eff5a234fbd33668d512caf0ef45e35f6c81589dd3b86818359d87aa7c\" for &ContainerMetadata{Name:nginx,Attempt:0,} returns container id \"08ee3dee256bfe1f3baa5fdeeab0f44a3e41ba115b49874d23cedef88d191cee\"" Dec 13 14:20:35.932484 env[1209]: time="2024-12-13T14:20:35.932431584Z" level=info msg="StartContainer for \"08ee3dee256bfe1f3baa5fdeeab0f44a3e41ba115b49874d23cedef88d191cee\"" Dec 13 14:20:35.953436 systemd[1]: Started cri-containerd-08ee3dee256bfe1f3baa5fdeeab0f44a3e41ba115b49874d23cedef88d191cee.scope. Dec 13 14:20:35.973214 env[1209]: time="2024-12-13T14:20:35.973170436Z" level=info msg="StartContainer for \"08ee3dee256bfe1f3baa5fdeeab0f44a3e41ba115b49874d23cedef88d191cee\" returns successfully" Dec 13 14:20:36.661939 kubelet[1425]: I1213 14:20:36.661877 1425 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nginx-deployment-6d5f899847-sb7n2" podStartSLOduration=6.8721774060000005 podStartE2EDuration="12.661791979s" podCreationTimestamp="2024-12-13 14:20:24 +0000 UTC" firstStartedPulling="2024-12-13 14:20:30.1274473 +0000 UTC m=+33.841932548" lastFinishedPulling="2024-12-13 14:20:35.917061873 +0000 UTC m=+39.631547121" observedRunningTime="2024-12-13 14:20:36.661502574 +0000 UTC m=+40.375987832" watchObservedRunningTime="2024-12-13 14:20:36.661791979 +0000 UTC m=+40.376277228" Dec 13 14:20:36.725080 kubelet[1425]: E1213 14:20:36.725006 1425 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:36.753425 kubelet[1425]: E1213 14:20:36.753401 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:37.753610 kubelet[1425]: E1213 14:20:37.753522 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:38.754216 kubelet[1425]: E1213 14:20:38.754154 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:39.755232 kubelet[1425]: E1213 14:20:39.755130 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:40.755779 kubelet[1425]: E1213 14:20:40.755696 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:41.756860 kubelet[1425]: E1213 14:20:41.756787 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:42.179014 kubelet[1425]: I1213 14:20:42.178946 1425 topology_manager.go:215] "Topology Admit Handler" podUID="3ae65bfc-103a-4904-b89a-f401680e0dd2" podNamespace="default" podName="nfs-server-provisioner-0" Dec 13 14:20:42.184800 systemd[1]: Created slice kubepods-besteffort-pod3ae65bfc_103a_4904_b89a_f401680e0dd2.slice. Dec 13 14:20:42.270995 kubelet[1425]: I1213 14:20:42.270949 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data\" (UniqueName: \"kubernetes.io/empty-dir/3ae65bfc-103a-4904-b89a-f401680e0dd2-data\") pod \"nfs-server-provisioner-0\" (UID: \"3ae65bfc-103a-4904-b89a-f401680e0dd2\") " pod="default/nfs-server-provisioner-0" Dec 13 14:20:42.271224 kubelet[1425]: I1213 14:20:42.271015 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb8zj\" (UniqueName: \"kubernetes.io/projected/3ae65bfc-103a-4904-b89a-f401680e0dd2-kube-api-access-lb8zj\") pod \"nfs-server-provisioner-0\" (UID: \"3ae65bfc-103a-4904-b89a-f401680e0dd2\") " pod="default/nfs-server-provisioner-0" Dec 13 14:20:42.487732 env[1209]: time="2024-12-13T14:20:42.487606997Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:3ae65bfc-103a-4904-b89a-f401680e0dd2,Namespace:default,Attempt:0,}" Dec 13 14:20:42.521047 systemd-networkd[1038]: lxc3f1f9efdf002: Link UP Dec 13 14:20:42.531870 kernel: eth0: renamed from tmpdc7a2 Dec 13 14:20:42.539688 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Dec 13 14:20:42.539931 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc3f1f9efdf002: link becomes ready Dec 13 14:20:42.540018 systemd-networkd[1038]: lxc3f1f9efdf002: Gained carrier Dec 13 14:20:42.700120 env[1209]: time="2024-12-13T14:20:42.700017226Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:20:42.700120 env[1209]: time="2024-12-13T14:20:42.700067350Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:20:42.700120 env[1209]: time="2024-12-13T14:20:42.700079313Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:20:42.700671 env[1209]: time="2024-12-13T14:20:42.700597650Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/dc7a265b5c6c9bd2913f5b60b283c53eb519a87dc8c1156ea3fc282f351949fa pid=2643 runtime=io.containerd.runc.v2 Dec 13 14:20:42.723389 systemd[1]: Started cri-containerd-dc7a265b5c6c9bd2913f5b60b283c53eb519a87dc8c1156ea3fc282f351949fa.scope. Dec 13 14:20:42.738899 systemd-resolved[1151]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 14:20:42.757474 kubelet[1425]: E1213 14:20:42.757416 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:42.797476 env[1209]: time="2024-12-13T14:20:42.797415227Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:nfs-server-provisioner-0,Uid:3ae65bfc-103a-4904-b89a-f401680e0dd2,Namespace:default,Attempt:0,} returns sandbox id \"dc7a265b5c6c9bd2913f5b60b283c53eb519a87dc8c1156ea3fc282f351949fa\"" Dec 13 14:20:42.798812 env[1209]: time="2024-12-13T14:20:42.798781109Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\"" Dec 13 14:20:43.758325 kubelet[1425]: E1213 14:20:43.758271 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:43.855091 systemd-networkd[1038]: lxc3f1f9efdf002: Gained IPv6LL Dec 13 14:20:44.759191 kubelet[1425]: E1213 14:20:44.759120 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:45.485305 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1724116254.mount: Deactivated successfully. Dec 13 14:20:45.760285 kubelet[1425]: E1213 14:20:45.760152 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:46.761081 kubelet[1425]: E1213 14:20:46.761022 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:47.762245 kubelet[1425]: E1213 14:20:47.762173 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:47.767528 env[1209]: time="2024-12-13T14:20:47.767485554Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:47.769492 env[1209]: time="2024-12-13T14:20:47.769464756Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:47.771466 env[1209]: time="2024-12-13T14:20:47.771442167Z" level=info msg="ImageUpdate event &ImageUpdate{Name:registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:47.773166 env[1209]: time="2024-12-13T14:20:47.773134270Z" level=info msg="ImageCreate event &ImageCreate{Name:registry.k8s.io/sig-storage/nfs-provisioner@sha256:c825f3d5e28bde099bd7a3daace28772d412c9157ad47fa752a9ad0baafc118d,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:47.773741 env[1209]: time="2024-12-13T14:20:47.773697419Z" level=info msg="PullImage \"registry.k8s.io/sig-storage/nfs-provisioner:v4.0.8\" returns image reference \"sha256:fd0b16f70b66b72bcb2f91d556fa33eba02729c44ffc5f2c16130e7f9fbed3c4\"" Dec 13 14:20:47.775690 env[1209]: time="2024-12-13T14:20:47.775654091Z" level=info msg="CreateContainer within sandbox \"dc7a265b5c6c9bd2913f5b60b283c53eb519a87dc8c1156ea3fc282f351949fa\" for container &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,}" Dec 13 14:20:47.790647 env[1209]: time="2024-12-13T14:20:47.790586903Z" level=info msg="CreateContainer within sandbox \"dc7a265b5c6c9bd2913f5b60b283c53eb519a87dc8c1156ea3fc282f351949fa\" for &ContainerMetadata{Name:nfs-server-provisioner,Attempt:0,} returns container id \"757a3c83e331ebdf33e73efa58bfa25e702f1fce6771a225b20c10b2d1956f42\"" Dec 13 14:20:47.791163 env[1209]: time="2024-12-13T14:20:47.791124485Z" level=info msg="StartContainer for \"757a3c83e331ebdf33e73efa58bfa25e702f1fce6771a225b20c10b2d1956f42\"" Dec 13 14:20:47.808351 systemd[1]: Started cri-containerd-757a3c83e331ebdf33e73efa58bfa25e702f1fce6771a225b20c10b2d1956f42.scope. Dec 13 14:20:47.830354 env[1209]: time="2024-12-13T14:20:47.830307228Z" level=info msg="StartContainer for \"757a3c83e331ebdf33e73efa58bfa25e702f1fce6771a225b20c10b2d1956f42\" returns successfully" Dec 13 14:20:48.763182 kubelet[1425]: E1213 14:20:48.763101 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:49.764349 kubelet[1425]: E1213 14:20:49.764276 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:50.765363 kubelet[1425]: E1213 14:20:50.765291 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:51.765640 kubelet[1425]: E1213 14:20:51.765574 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:52.766461 kubelet[1425]: E1213 14:20:52.766392 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:53.767511 kubelet[1425]: E1213 14:20:53.767440 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:54.768140 kubelet[1425]: E1213 14:20:54.768053 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:55.768695 kubelet[1425]: E1213 14:20:55.768622 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:56.724651 kubelet[1425]: E1213 14:20:56.724572 1425 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:56.768825 kubelet[1425]: E1213 14:20:56.768790 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:57.435058 kubelet[1425]: I1213 14:20:57.435015 1425 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/nfs-server-provisioner-0" podStartSLOduration=10.459491363 podStartE2EDuration="15.43495309s" podCreationTimestamp="2024-12-13 14:20:42 +0000 UTC" firstStartedPulling="2024-12-13 14:20:42.798521991 +0000 UTC m=+46.513007239" lastFinishedPulling="2024-12-13 14:20:47.773983718 +0000 UTC m=+51.488468966" observedRunningTime="2024-12-13 14:20:48.691546138 +0000 UTC m=+52.406031386" watchObservedRunningTime="2024-12-13 14:20:57.43495309 +0000 UTC m=+61.149438338" Dec 13 14:20:57.435444 kubelet[1425]: I1213 14:20:57.435423 1425 topology_manager.go:215] "Topology Admit Handler" podUID="272e1cd5-5939-4831-abe4-6c9cf0777863" podNamespace="default" podName="test-pod-1" Dec 13 14:20:57.441193 systemd[1]: Created slice kubepods-besteffort-pod272e1cd5_5939_4831_abe4_6c9cf0777863.slice. Dec 13 14:20:57.542434 kubelet[1425]: I1213 14:20:57.542354 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-eb1c3c74-de7d-41c3-b4fd-bf490ac292d7\" (UniqueName: \"kubernetes.io/nfs/272e1cd5-5939-4831-abe4-6c9cf0777863-pvc-eb1c3c74-de7d-41c3-b4fd-bf490ac292d7\") pod \"test-pod-1\" (UID: \"272e1cd5-5939-4831-abe4-6c9cf0777863\") " pod="default/test-pod-1" Dec 13 14:20:57.542434 kubelet[1425]: I1213 14:20:57.542440 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fc77f\" (UniqueName: \"kubernetes.io/projected/272e1cd5-5939-4831-abe4-6c9cf0777863-kube-api-access-fc77f\") pod \"test-pod-1\" (UID: \"272e1cd5-5939-4831-abe4-6c9cf0777863\") " pod="default/test-pod-1" Dec 13 14:20:57.669894 kernel: FS-Cache: Loaded Dec 13 14:20:57.726191 kernel: RPC: Registered named UNIX socket transport module. Dec 13 14:20:57.726350 kernel: RPC: Registered udp transport module. Dec 13 14:20:57.726390 kernel: RPC: Registered tcp transport module. Dec 13 14:20:57.727232 kernel: RPC: Registered tcp NFSv4.1 backchannel transport module. Dec 13 14:20:57.770057 kubelet[1425]: E1213 14:20:57.770002 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:57.812864 kernel: FS-Cache: Netfs 'nfs' registered for caching Dec 13 14:20:57.993517 kernel: NFS: Registering the id_resolver key type Dec 13 14:20:57.993699 kernel: Key type id_resolver registered Dec 13 14:20:57.993723 kernel: Key type id_legacy registered Dec 13 14:20:58.038260 nfsidmap[2766]: nss_getpwnam: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Dec 13 14:20:58.042076 nfsidmap[2769]: nss_name_to_gid: name 'root@nfs-server-provisioner.default.svc.cluster.local' does not map into domain 'localdomain' Dec 13 14:20:58.344622 env[1209]: time="2024-12-13T14:20:58.344456225Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:272e1cd5-5939-4831-abe4-6c9cf0777863,Namespace:default,Attempt:0,}" Dec 13 14:20:58.377655 systemd-networkd[1038]: lxc40b1ca45dec8: Link UP Dec 13 14:20:58.383877 kernel: eth0: renamed from tmpb4fde Dec 13 14:20:58.392703 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): eth0: link becomes ready Dec 13 14:20:58.392859 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc40b1ca45dec8: link becomes ready Dec 13 14:20:58.392283 systemd-networkd[1038]: lxc40b1ca45dec8: Gained carrier Dec 13 14:20:58.561551 env[1209]: time="2024-12-13T14:20:58.561463082Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:20:58.561551 env[1209]: time="2024-12-13T14:20:58.561510150Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:20:58.561551 env[1209]: time="2024-12-13T14:20:58.561523535Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:20:58.561936 env[1209]: time="2024-12-13T14:20:58.561662546Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/b4fded686c3dff199593e499304c3f2ba5a77fb81f51aacea741f87e3933c625 pid=2803 runtime=io.containerd.runc.v2 Dec 13 14:20:58.576823 systemd[1]: Started cri-containerd-b4fded686c3dff199593e499304c3f2ba5a77fb81f51aacea741f87e3933c625.scope. Dec 13 14:20:58.590973 systemd-resolved[1151]: Failed to determine the local hostname and LLMNR/mDNS names, ignoring: No such device or address Dec 13 14:20:58.613270 env[1209]: time="2024-12-13T14:20:58.613133923Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:test-pod-1,Uid:272e1cd5-5939-4831-abe4-6c9cf0777863,Namespace:default,Attempt:0,} returns sandbox id \"b4fded686c3dff199593e499304c3f2ba5a77fb81f51aacea741f87e3933c625\"" Dec 13 14:20:58.615259 env[1209]: time="2024-12-13T14:20:58.615225100Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\"" Dec 13 14:20:58.770746 kubelet[1425]: E1213 14:20:58.770680 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:20:58.979856 env[1209]: time="2024-12-13T14:20:58.979713566Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:58.984767 env[1209]: time="2024-12-13T14:20:58.984735699Z" level=info msg="ImageUpdate event &ImageUpdate{Name:sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:58.986709 env[1209]: time="2024-12-13T14:20:58.986675172Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx:latest,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:58.988629 env[1209]: time="2024-12-13T14:20:58.988600168Z" level=info msg="ImageUpdate event &ImageUpdate{Name:ghcr.io/flatcar/nginx@sha256:e04edf30a4ea4c5a4107110797c72d3ee8a654415f00acd4019be17218afd9a1,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:20:58.989216 env[1209]: time="2024-12-13T14:20:58.989183362Z" level=info msg="PullImage \"ghcr.io/flatcar/nginx:latest\" returns image reference \"sha256:fa0a8cea5e76ad962111c39c85bb312edaf5b89eccd8f404eeea66c9759641e3\"" Dec 13 14:20:58.990814 env[1209]: time="2024-12-13T14:20:58.990780422Z" level=info msg="CreateContainer within sandbox \"b4fded686c3dff199593e499304c3f2ba5a77fb81f51aacea741f87e3933c625\" for container &ContainerMetadata{Name:test,Attempt:0,}" Dec 13 14:20:59.006084 env[1209]: time="2024-12-13T14:20:59.006037057Z" level=info msg="CreateContainer within sandbox \"b4fded686c3dff199593e499304c3f2ba5a77fb81f51aacea741f87e3933c625\" for &ContainerMetadata{Name:test,Attempt:0,} returns container id \"aff1986b7cc7a1fceb2dfe5ae0f07ef7dab2ca796c7e11e9e6d3b88e8e46b175\"" Dec 13 14:20:59.006653 env[1209]: time="2024-12-13T14:20:59.006614853Z" level=info msg="StartContainer for \"aff1986b7cc7a1fceb2dfe5ae0f07ef7dab2ca796c7e11e9e6d3b88e8e46b175\"" Dec 13 14:20:59.023134 systemd[1]: Started cri-containerd-aff1986b7cc7a1fceb2dfe5ae0f07ef7dab2ca796c7e11e9e6d3b88e8e46b175.scope. Dec 13 14:20:59.047564 env[1209]: time="2024-12-13T14:20:59.047514836Z" level=info msg="StartContainer for \"aff1986b7cc7a1fceb2dfe5ae0f07ef7dab2ca796c7e11e9e6d3b88e8e46b175\" returns successfully" Dec 13 14:20:59.663112 systemd-networkd[1038]: lxc40b1ca45dec8: Gained IPv6LL Dec 13 14:20:59.714403 kubelet[1425]: I1213 14:20:59.714362 1425 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="default/test-pod-1" podStartSLOduration=17.33969211 podStartE2EDuration="17.71432008s" podCreationTimestamp="2024-12-13 14:20:42 +0000 UTC" firstStartedPulling="2024-12-13 14:20:58.614821683 +0000 UTC m=+62.329306931" lastFinishedPulling="2024-12-13 14:20:58.989449643 +0000 UTC m=+62.703934901" observedRunningTime="2024-12-13 14:20:59.714159288 +0000 UTC m=+63.428644546" watchObservedRunningTime="2024-12-13 14:20:59.71432008 +0000 UTC m=+63.428805328" Dec 13 14:20:59.771783 kubelet[1425]: E1213 14:20:59.771711 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:00.772423 kubelet[1425]: E1213 14:21:00.772370 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:01.773003 kubelet[1425]: E1213 14:21:01.772922 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:02.773523 kubelet[1425]: E1213 14:21:02.773460 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:03.774711 kubelet[1425]: E1213 14:21:03.774647 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:04.774933 kubelet[1425]: E1213 14:21:04.774822 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:04.813645 env[1209]: time="2024-12-13T14:21:04.813527704Z" level=error msg="failed to reload cni configuration after receiving fs change event(\"/etc/cni/net.d/05-cilium.conf\": REMOVE)" error="cni config load failed: no network config found in /etc/cni/net.d: cni plugin not initialized: failed to load cni config" Dec 13 14:21:04.820935 env[1209]: time="2024-12-13T14:21:04.820886119Z" level=info msg="StopContainer for \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\" with timeout 2 (s)" Dec 13 14:21:04.821264 env[1209]: time="2024-12-13T14:21:04.821221668Z" level=info msg="Stop container \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\" with signal terminated" Dec 13 14:21:04.827785 systemd-networkd[1038]: lxc_health: Link DOWN Dec 13 14:21:04.827808 systemd-networkd[1038]: lxc_health: Lost carrier Dec 13 14:21:04.872387 systemd[1]: cri-containerd-58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af.scope: Deactivated successfully. Dec 13 14:21:04.872780 systemd[1]: cri-containerd-58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af.scope: Consumed 9.190s CPU time. Dec 13 14:21:04.890070 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af-rootfs.mount: Deactivated successfully. Dec 13 14:21:05.001003 env[1209]: time="2024-12-13T14:21:05.000912306Z" level=info msg="shim disconnected" id=58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af Dec 13 14:21:05.001003 env[1209]: time="2024-12-13T14:21:05.000988589Z" level=warning msg="cleaning up after shim disconnected" id=58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af namespace=k8s.io Dec 13 14:21:05.001003 env[1209]: time="2024-12-13T14:21:05.000998738Z" level=info msg="cleaning up dead shim" Dec 13 14:21:05.007935 env[1209]: time="2024-12-13T14:21:05.007888062Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:21:05Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2933 runtime=io.containerd.runc.v2\n" Dec 13 14:21:05.076277 env[1209]: time="2024-12-13T14:21:05.076184689Z" level=info msg="StopContainer for \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\" returns successfully" Dec 13 14:21:05.076934 env[1209]: time="2024-12-13T14:21:05.076906704Z" level=info msg="StopPodSandbox for \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\"" Dec 13 14:21:05.077010 env[1209]: time="2024-12-13T14:21:05.076976505Z" level=info msg="Container to stop \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:21:05.077010 env[1209]: time="2024-12-13T14:21:05.076994238Z" level=info msg="Container to stop \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:21:05.077010 env[1209]: time="2024-12-13T14:21:05.077006110Z" level=info msg="Container to stop \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:21:05.077153 env[1209]: time="2024-12-13T14:21:05.077021770Z" level=info msg="Container to stop \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:21:05.077153 env[1209]: time="2024-12-13T14:21:05.077035285Z" level=info msg="Container to stop \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\" must be in running or unknown state, current state \"CONTAINER_EXITED\"" Dec 13 14:21:05.079113 systemd[1]: run-containerd-io.containerd.grpc.v1.cri-sandboxes-5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569-shm.mount: Deactivated successfully. Dec 13 14:21:05.083185 systemd[1]: cri-containerd-5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569.scope: Deactivated successfully. Dec 13 14:21:05.099308 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569-rootfs.mount: Deactivated successfully. Dec 13 14:21:05.105148 env[1209]: time="2024-12-13T14:21:05.105092515Z" level=info msg="shim disconnected" id=5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569 Dec 13 14:21:05.105148 env[1209]: time="2024-12-13T14:21:05.105144142Z" level=warning msg="cleaning up after shim disconnected" id=5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569 namespace=k8s.io Dec 13 14:21:05.105345 env[1209]: time="2024-12-13T14:21:05.105153420Z" level=info msg="cleaning up dead shim" Dec 13 14:21:05.112494 env[1209]: time="2024-12-13T14:21:05.112434918Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:21:05Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=2964 runtime=io.containerd.runc.v2\n" Dec 13 14:21:05.112803 env[1209]: time="2024-12-13T14:21:05.112764527Z" level=info msg="TearDown network for sandbox \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" successfully" Dec 13 14:21:05.112849 env[1209]: time="2024-12-13T14:21:05.112803831Z" level=info msg="StopPodSandbox for \"5f661ffc677e3ec75a286b58e7dd6a5c5ba8ec379104f04d590180a9c0973569\" returns successfully" Dec 13 14:21:05.289039 kubelet[1425]: I1213 14:21:05.288966 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-lib-modules\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289278 kubelet[1425]: I1213 14:21:05.289039 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-host-proc-sys-net\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289278 kubelet[1425]: I1213 14:21:05.289094 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.289278 kubelet[1425]: I1213 14:21:05.289145 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.289278 kubelet[1425]: I1213 14:21:05.289122 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-xtables-lock\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289278 kubelet[1425]: I1213 14:21:05.289219 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fht6b\" (UniqueName: \"kubernetes.io/projected/fd8b627b-eb28-4788-976f-4daf13ac261b-kube-api-access-fht6b\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289506 kubelet[1425]: I1213 14:21:05.289108 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.289506 kubelet[1425]: I1213 14:21:05.289244 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-cgroup\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289506 kubelet[1425]: I1213 14:21:05.289267 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/fd8b627b-eb28-4788-976f-4daf13ac261b-hubble-tls\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289506 kubelet[1425]: I1213 14:21:05.289286 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-bpf-maps\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289506 kubelet[1425]: I1213 14:21:05.289336 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-hostproc\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289506 kubelet[1425]: I1213 14:21:05.289345 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.289741 kubelet[1425]: I1213 14:21:05.289355 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/fd8b627b-eb28-4788-976f-4daf13ac261b-clustermesh-secrets\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289741 kubelet[1425]: I1213 14:21:05.289416 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-config-path\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289741 kubelet[1425]: I1213 14:21:05.289441 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-host-proc-sys-kernel\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289741 kubelet[1425]: I1213 14:21:05.289465 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cni-path\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289741 kubelet[1425]: I1213 14:21:05.289489 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-run\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.289741 kubelet[1425]: I1213 14:21:05.289514 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-etc-cni-netd\") pod \"fd8b627b-eb28-4788-976f-4daf13ac261b\" (UID: \"fd8b627b-eb28-4788-976f-4daf13ac261b\") " Dec 13 14:21:05.290032 kubelet[1425]: I1213 14:21:05.289555 1425 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-lib-modules\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.290032 kubelet[1425]: I1213 14:21:05.289574 1425 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-host-proc-sys-net\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.290032 kubelet[1425]: I1213 14:21:05.289588 1425 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-xtables-lock\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.290032 kubelet[1425]: I1213 14:21:05.289600 1425 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-cgroup\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.290032 kubelet[1425]: I1213 14:21:05.289622 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.290032 kubelet[1425]: I1213 14:21:05.289735 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cni-path" (OuterVolumeSpecName: "cni-path") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.290255 kubelet[1425]: I1213 14:21:05.289765 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.290255 kubelet[1425]: I1213 14:21:05.289856 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.290255 kubelet[1425]: I1213 14:21:05.289882 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.293209 kubelet[1425]: I1213 14:21:05.293159 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd8b627b-eb28-4788-976f-4daf13ac261b-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 14:21:05.293324 kubelet[1425]: I1213 14:21:05.293240 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-hostproc" (OuterVolumeSpecName: "hostproc") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:05.293454 kubelet[1425]: I1213 14:21:05.293411 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd8b627b-eb28-4788-976f-4daf13ac261b-kube-api-access-fht6b" (OuterVolumeSpecName: "kube-api-access-fht6b") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "kube-api-access-fht6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 14:21:05.293618 systemd[1]: var-lib-kubelet-pods-fd8b627b\x2deb28\x2d4788\x2d976f\x2d4daf13ac261b-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Dec 13 14:21:05.293977 kubelet[1425]: I1213 14:21:05.293952 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fd8b627b-eb28-4788-976f-4daf13ac261b-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 14:21:05.294251 kubelet[1425]: I1213 14:21:05.294228 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "fd8b627b-eb28-4788-976f-4daf13ac261b" (UID: "fd8b627b-eb28-4788-976f-4daf13ac261b"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 14:21:05.390550 kubelet[1425]: I1213 14:21:05.390408 1425 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/fd8b627b-eb28-4788-976f-4daf13ac261b-hubble-tls\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.390550 kubelet[1425]: I1213 14:21:05.390448 1425 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-bpf-maps\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.390550 kubelet[1425]: I1213 14:21:05.390459 1425 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-config-path\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.390550 kubelet[1425]: I1213 14:21:05.390475 1425 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-host-proc-sys-kernel\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.390550 kubelet[1425]: I1213 14:21:05.390485 1425 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cni-path\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.390550 kubelet[1425]: I1213 14:21:05.390495 1425 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-cilium-run\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.390550 kubelet[1425]: I1213 14:21:05.390503 1425 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-etc-cni-netd\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.390550 kubelet[1425]: I1213 14:21:05.390532 1425 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/fd8b627b-eb28-4788-976f-4daf13ac261b-hostproc\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.390930 kubelet[1425]: I1213 14:21:05.390543 1425 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/fd8b627b-eb28-4788-976f-4daf13ac261b-clustermesh-secrets\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.390930 kubelet[1425]: I1213 14:21:05.390558 1425 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-fht6b\" (UniqueName: \"kubernetes.io/projected/fd8b627b-eb28-4788-976f-4daf13ac261b-kube-api-access-fht6b\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:05.582118 systemd[1]: Removed slice kubepods-burstable-podfd8b627b_eb28_4788_976f_4daf13ac261b.slice. Dec 13 14:21:05.582224 systemd[1]: kubepods-burstable-podfd8b627b_eb28_4788_976f_4daf13ac261b.slice: Consumed 9.333s CPU time. Dec 13 14:21:05.721070 kubelet[1425]: I1213 14:21:05.720674 1425 scope.go:117] "RemoveContainer" containerID="58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af" Dec 13 14:21:05.722111 env[1209]: time="2024-12-13T14:21:05.722069068Z" level=info msg="RemoveContainer for \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\"" Dec 13 14:21:05.725890 env[1209]: time="2024-12-13T14:21:05.725823686Z" level=info msg="RemoveContainer for \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\" returns successfully" Dec 13 14:21:05.726121 kubelet[1425]: I1213 14:21:05.726099 1425 scope.go:117] "RemoveContainer" containerID="72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3" Dec 13 14:21:05.727172 env[1209]: time="2024-12-13T14:21:05.727132854Z" level=info msg="RemoveContainer for \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\"" Dec 13 14:21:05.730270 env[1209]: time="2024-12-13T14:21:05.730233273Z" level=info msg="RemoveContainer for \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\" returns successfully" Dec 13 14:21:05.730453 kubelet[1425]: I1213 14:21:05.730418 1425 scope.go:117] "RemoveContainer" containerID="cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48" Dec 13 14:21:05.732184 env[1209]: time="2024-12-13T14:21:05.732148559Z" level=info msg="RemoveContainer for \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\"" Dec 13 14:21:05.735670 env[1209]: time="2024-12-13T14:21:05.735635785Z" level=info msg="RemoveContainer for \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\" returns successfully" Dec 13 14:21:05.735869 kubelet[1425]: I1213 14:21:05.735853 1425 scope.go:117] "RemoveContainer" containerID="46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58" Dec 13 14:21:05.736854 env[1209]: time="2024-12-13T14:21:05.736807285Z" level=info msg="RemoveContainer for \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\"" Dec 13 14:21:05.740252 env[1209]: time="2024-12-13T14:21:05.740195946Z" level=info msg="RemoveContainer for \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\" returns successfully" Dec 13 14:21:05.740462 kubelet[1425]: I1213 14:21:05.740425 1425 scope.go:117] "RemoveContainer" containerID="fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b" Dec 13 14:21:05.741422 env[1209]: time="2024-12-13T14:21:05.741387553Z" level=info msg="RemoveContainer for \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\"" Dec 13 14:21:05.744287 env[1209]: time="2024-12-13T14:21:05.744246389Z" level=info msg="RemoveContainer for \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\" returns successfully" Dec 13 14:21:05.744495 kubelet[1425]: I1213 14:21:05.744467 1425 scope.go:117] "RemoveContainer" containerID="58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af" Dec 13 14:21:05.744795 env[1209]: time="2024-12-13T14:21:05.744689511Z" level=error msg="ContainerStatus for \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\": not found" Dec 13 14:21:05.745006 kubelet[1425]: E1213 14:21:05.744983 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\": not found" containerID="58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af" Dec 13 14:21:05.745070 kubelet[1425]: I1213 14:21:05.745055 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af"} err="failed to get container status \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\": rpc error: code = NotFound desc = an error occurred when try to find container \"58200ff73c14fe5560c1efb3073bdf8aec56f6eacef13e2d60431a44fede04af\": not found" Dec 13 14:21:05.745070 kubelet[1425]: I1213 14:21:05.745068 1425 scope.go:117] "RemoveContainer" containerID="72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3" Dec 13 14:21:05.745295 env[1209]: time="2024-12-13T14:21:05.745244102Z" level=error msg="ContainerStatus for \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\": not found" Dec 13 14:21:05.745460 kubelet[1425]: E1213 14:21:05.745393 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\": not found" containerID="72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3" Dec 13 14:21:05.745460 kubelet[1425]: I1213 14:21:05.745410 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3"} err="failed to get container status \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\": rpc error: code = NotFound desc = an error occurred when try to find container \"72fb1706063c056bad5b444101f6fa2a48ad917076c49ea801cd263a7c4047e3\": not found" Dec 13 14:21:05.745460 kubelet[1425]: I1213 14:21:05.745420 1425 scope.go:117] "RemoveContainer" containerID="cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48" Dec 13 14:21:05.745663 env[1209]: time="2024-12-13T14:21:05.745584591Z" level=error msg="ContainerStatus for \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\": not found" Dec 13 14:21:05.745898 kubelet[1425]: E1213 14:21:05.745878 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\": not found" containerID="cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48" Dec 13 14:21:05.745898 kubelet[1425]: I1213 14:21:05.745900 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48"} err="failed to get container status \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\": rpc error: code = NotFound desc = an error occurred when try to find container \"cf42048309bf70429c076b9eae78b546c6ff0d2489c82ee0a7875ac8e9d36f48\": not found" Dec 13 14:21:05.746017 kubelet[1425]: I1213 14:21:05.745908 1425 scope.go:117] "RemoveContainer" containerID="46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58" Dec 13 14:21:05.746127 env[1209]: time="2024-12-13T14:21:05.746069281Z" level=error msg="ContainerStatus for \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\": not found" Dec 13 14:21:05.746242 kubelet[1425]: E1213 14:21:05.746222 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\": not found" containerID="46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58" Dec 13 14:21:05.746242 kubelet[1425]: I1213 14:21:05.746244 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58"} err="failed to get container status \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\": rpc error: code = NotFound desc = an error occurred when try to find container \"46669e0661bd85d3f33bef5969f7109d13861f0706172ca7ae220a4ef553dd58\": not found" Dec 13 14:21:05.746351 kubelet[1425]: I1213 14:21:05.746252 1425 scope.go:117] "RemoveContainer" containerID="fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b" Dec 13 14:21:05.746465 env[1209]: time="2024-12-13T14:21:05.746404751Z" level=error msg="ContainerStatus for \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\" failed" error="rpc error: code = NotFound desc = an error occurred when try to find container \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\": not found" Dec 13 14:21:05.746607 kubelet[1425]: E1213 14:21:05.746581 1425 remote_runtime.go:432] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = an error occurred when try to find container \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\": not found" containerID="fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b" Dec 13 14:21:05.746668 kubelet[1425]: I1213 14:21:05.746634 1425 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"containerd","ID":"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b"} err="failed to get container status \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\": rpc error: code = NotFound desc = an error occurred when try to find container \"fbcc0766c8e23e766a41b14a249f3dc9af605ceec5b59a0f4568bfd1782cf66b\": not found" Dec 13 14:21:05.775948 kubelet[1425]: E1213 14:21:05.775904 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:05.797891 systemd[1]: var-lib-kubelet-pods-fd8b627b\x2deb28\x2d4788\x2d976f\x2d4daf13ac261b-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2dfht6b.mount: Deactivated successfully. Dec 13 14:21:05.798025 systemd[1]: var-lib-kubelet-pods-fd8b627b\x2deb28\x2d4788\x2d976f\x2d4daf13ac261b-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Dec 13 14:21:06.777184 kubelet[1425]: E1213 14:21:06.777074 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:07.239936 kubelet[1425]: I1213 14:21:07.239776 1425 topology_manager.go:215] "Topology Admit Handler" podUID="adb00e62-58a3-432c-bf8b-2ba1d7421208" podNamespace="kube-system" podName="cilium-operator-5cc964979-nhbfl" Dec 13 14:21:07.239936 kubelet[1425]: E1213 14:21:07.239897 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="fd8b627b-eb28-4788-976f-4daf13ac261b" containerName="clean-cilium-state" Dec 13 14:21:07.239936 kubelet[1425]: E1213 14:21:07.239914 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="fd8b627b-eb28-4788-976f-4daf13ac261b" containerName="mount-cgroup" Dec 13 14:21:07.239936 kubelet[1425]: E1213 14:21:07.239923 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="fd8b627b-eb28-4788-976f-4daf13ac261b" containerName="apply-sysctl-overwrites" Dec 13 14:21:07.239936 kubelet[1425]: E1213 14:21:07.239933 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="fd8b627b-eb28-4788-976f-4daf13ac261b" containerName="mount-bpf-fs" Dec 13 14:21:07.239936 kubelet[1425]: E1213 14:21:07.239941 1425 cpu_manager.go:395] "RemoveStaleState: removing container" podUID="fd8b627b-eb28-4788-976f-4daf13ac261b" containerName="cilium-agent" Dec 13 14:21:07.240318 kubelet[1425]: I1213 14:21:07.239972 1425 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd8b627b-eb28-4788-976f-4daf13ac261b" containerName="cilium-agent" Dec 13 14:21:07.240318 kubelet[1425]: I1213 14:21:07.240254 1425 topology_manager.go:215] "Topology Admit Handler" podUID="02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" podNamespace="kube-system" podName="cilium-gql8q" Dec 13 14:21:07.246460 systemd[1]: Created slice kubepods-besteffort-podadb00e62_58a3_432c_bf8b_2ba1d7421208.slice. Dec 13 14:21:07.250878 systemd[1]: Created slice kubepods-burstable-pod02d4381b_b6b9_4940_9cb0_c1b6c2ef3aa1.slice. Dec 13 14:21:07.404790 kubelet[1425]: I1213 14:21:07.404679 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-clustermesh-secrets\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.404790 kubelet[1425]: I1213 14:21:07.404765 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-ipsec-secrets\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.404790 kubelet[1425]: I1213 14:21:07.404805 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-hubble-tls\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405133 kubelet[1425]: I1213 14:21:07.404860 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-host-proc-sys-net\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405133 kubelet[1425]: I1213 14:21:07.404893 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-lib-modules\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405133 kubelet[1425]: I1213 14:21:07.404973 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-bpf-maps\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405133 kubelet[1425]: I1213 14:21:07.405011 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-cgroup\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405308 kubelet[1425]: I1213 14:21:07.405133 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-run\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405308 kubelet[1425]: I1213 14:21:07.405202 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-xtables-lock\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405308 kubelet[1425]: I1213 14:21:07.405232 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6h949\" (UniqueName: \"kubernetes.io/projected/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-kube-api-access-6h949\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405308 kubelet[1425]: I1213 14:21:07.405267 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/adb00e62-58a3-432c-bf8b-2ba1d7421208-cilium-config-path\") pod \"cilium-operator-5cc964979-nhbfl\" (UID: \"adb00e62-58a3-432c-bf8b-2ba1d7421208\") " pod="kube-system/cilium-operator-5cc964979-nhbfl" Dec 13 14:21:07.405475 kubelet[1425]: I1213 14:21:07.405316 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndr4q\" (UniqueName: \"kubernetes.io/projected/adb00e62-58a3-432c-bf8b-2ba1d7421208-kube-api-access-ndr4q\") pod \"cilium-operator-5cc964979-nhbfl\" (UID: \"adb00e62-58a3-432c-bf8b-2ba1d7421208\") " pod="kube-system/cilium-operator-5cc964979-nhbfl" Dec 13 14:21:07.405475 kubelet[1425]: I1213 14:21:07.405367 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-etc-cni-netd\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405475 kubelet[1425]: I1213 14:21:07.405394 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-config-path\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405475 kubelet[1425]: I1213 14:21:07.405423 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-hostproc\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405475 kubelet[1425]: I1213 14:21:07.405466 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cni-path\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.405681 kubelet[1425]: I1213 14:21:07.405501 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-host-proc-sys-kernel\") pod \"cilium-gql8q\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " pod="kube-system/cilium-gql8q" Dec 13 14:21:07.503435 kubelet[1425]: E1213 14:21:07.503289 1425 pod_workers.go:1298] "Error syncing pod, skipping" err="unmounted volumes=[bpf-maps cilium-cgroup cilium-config-path cilium-ipsec-secrets cilium-run clustermesh-secrets cni-path etc-cni-netd host-proc-sys-kernel host-proc-sys-net hostproc hubble-tls kube-api-access-6h949 lib-modules xtables-lock], unattached volumes=[], failed to process volumes=[]: context canceled" pod="kube-system/cilium-gql8q" podUID="02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" Dec 13 14:21:07.549496 kubelet[1425]: E1213 14:21:07.549450 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:07.550201 env[1209]: time="2024-12-13T14:21:07.550161256Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5cc964979-nhbfl,Uid:adb00e62-58a3-432c-bf8b-2ba1d7421208,Namespace:kube-system,Attempt:0,}" Dec 13 14:21:07.563954 env[1209]: time="2024-12-13T14:21:07.563882221Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:21:07.563954 env[1209]: time="2024-12-13T14:21:07.563925342Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:21:07.563954 env[1209]: time="2024-12-13T14:21:07.563936352Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:21:07.564189 env[1209]: time="2024-12-13T14:21:07.564116620Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/fb2c719b3caf002635795f8562e5d4b4a2d583bbc01fe6ea4ad648bc9ae5744c pid=2991 runtime=io.containerd.runc.v2 Dec 13 14:21:07.575720 systemd[1]: Started cri-containerd-fb2c719b3caf002635795f8562e5d4b4a2d583bbc01fe6ea4ad648bc9ae5744c.scope. Dec 13 14:21:07.577943 kubelet[1425]: I1213 14:21:07.577918 1425 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="fd8b627b-eb28-4788-976f-4daf13ac261b" path="/var/lib/kubelet/pods/fd8b627b-eb28-4788-976f-4daf13ac261b/volumes" Dec 13 14:21:07.616316 env[1209]: time="2024-12-13T14:21:07.615285223Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-operator-5cc964979-nhbfl,Uid:adb00e62-58a3-432c-bf8b-2ba1d7421208,Namespace:kube-system,Attempt:0,} returns sandbox id \"fb2c719b3caf002635795f8562e5d4b4a2d583bbc01fe6ea4ad648bc9ae5744c\"" Dec 13 14:21:07.616483 kubelet[1425]: E1213 14:21:07.616006 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:07.617306 env[1209]: time="2024-12-13T14:21:07.617266362Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\"" Dec 13 14:21:07.672862 kubelet[1425]: E1213 14:21:07.672801 1425 kubelet.go:2892] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Dec 13 14:21:07.778026 kubelet[1425]: E1213 14:21:07.777868 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:07.908644 kubelet[1425]: I1213 14:21:07.908571 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-xtables-lock\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.908644 kubelet[1425]: I1213 14:21:07.908637 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-host-proc-sys-net\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.908929 kubelet[1425]: I1213 14:21:07.908678 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6h949\" (UniqueName: \"kubernetes.io/projected/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-kube-api-access-6h949\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.908929 kubelet[1425]: I1213 14:21:07.908707 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-host-proc-sys-kernel\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.908929 kubelet[1425]: I1213 14:21:07.908714 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-xtables-lock" (OuterVolumeSpecName: "xtables-lock") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "xtables-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.908929 kubelet[1425]: I1213 14:21:07.908747 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-cgroup\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.908929 kubelet[1425]: I1213 14:21:07.908776 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-ipsec-secrets\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.908929 kubelet[1425]: I1213 14:21:07.908798 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-lib-modules\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.909077 kubelet[1425]: I1213 14:21:07.908823 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cni-path\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.909077 kubelet[1425]: I1213 14:21:07.908878 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-clustermesh-secrets\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.909077 kubelet[1425]: I1213 14:21:07.908911 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-hubble-tls\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.909077 kubelet[1425]: I1213 14:21:07.908934 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-bpf-maps\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.909077 kubelet[1425]: I1213 14:21:07.908958 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-hostproc\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.909077 kubelet[1425]: I1213 14:21:07.908981 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-run\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.909272 kubelet[1425]: I1213 14:21:07.909005 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-etc-cni-netd\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.909272 kubelet[1425]: I1213 14:21:07.909034 1425 reconciler_common.go:172] "operationExecutor.UnmountVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-config-path\") pod \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\" (UID: \"02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1\") " Dec 13 14:21:07.909272 kubelet[1425]: I1213 14:21:07.909080 1425 reconciler_common.go:300] "Volume detached for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-xtables-lock\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:07.909427 kubelet[1425]: I1213 14:21:07.909241 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-host-proc-sys-net" (OuterVolumeSpecName: "host-proc-sys-net") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "host-proc-sys-net". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.909427 kubelet[1425]: I1213 14:21:07.909342 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-hostproc" (OuterVolumeSpecName: "hostproc") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "hostproc". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.911576 kubelet[1425]: I1213 14:21:07.909928 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-host-proc-sys-kernel" (OuterVolumeSpecName: "host-proc-sys-kernel") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "host-proc-sys-kernel". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.911576 kubelet[1425]: I1213 14:21:07.909974 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.911576 kubelet[1425]: I1213 14:21:07.910000 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cni-path" (OuterVolumeSpecName: "cni-path") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "cni-path". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.911576 kubelet[1425]: I1213 14:21:07.910032 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-run" (OuterVolumeSpecName: "cilium-run") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "cilium-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.911576 kubelet[1425]: I1213 14:21:07.910056 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-cgroup" (OuterVolumeSpecName: "cilium-cgroup") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "cilium-cgroup". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.911759 kubelet[1425]: I1213 14:21:07.910084 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-bpf-maps" (OuterVolumeSpecName: "bpf-maps") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "bpf-maps". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.911759 kubelet[1425]: I1213 14:21:07.910114 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-etc-cni-netd" (OuterVolumeSpecName: "etc-cni-netd") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "etc-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 13 14:21:07.911759 kubelet[1425]: I1213 14:21:07.911620 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-config-path" (OuterVolumeSpecName: "cilium-config-path") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "cilium-config-path". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 13 14:21:07.912485 kubelet[1425]: I1213 14:21:07.912464 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-hubble-tls" (OuterVolumeSpecName: "hubble-tls") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "hubble-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 14:21:07.912640 kubelet[1425]: I1213 14:21:07.912610 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-clustermesh-secrets" (OuterVolumeSpecName: "clustermesh-secrets") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "clustermesh-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 14:21:07.913922 kubelet[1425]: I1213 14:21:07.913882 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-ipsec-secrets" (OuterVolumeSpecName: "cilium-ipsec-secrets") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "cilium-ipsec-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 13 14:21:07.914036 kubelet[1425]: I1213 14:21:07.914001 1425 operation_generator.go:887] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-kube-api-access-6h949" (OuterVolumeSpecName: "kube-api-access-6h949") pod "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" (UID: "02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1"). InnerVolumeSpecName "kube-api-access-6h949". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 13 14:21:08.009308 kubelet[1425]: I1213 14:21:08.009246 1425 reconciler_common.go:300] "Volume detached for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-cgroup\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009308 kubelet[1425]: I1213 14:21:08.009292 1425 reconciler_common.go:300] "Volume detached for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-ipsec-secrets\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009308 kubelet[1425]: I1213 14:21:08.009304 1425 reconciler_common.go:300] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-lib-modules\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009308 kubelet[1425]: I1213 14:21:08.009316 1425 reconciler_common.go:300] "Volume detached for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-bpf-maps\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009308 kubelet[1425]: I1213 14:21:08.009325 1425 reconciler_common.go:300] "Volume detached for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-hostproc\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009308 kubelet[1425]: I1213 14:21:08.009334 1425 reconciler_common.go:300] "Volume detached for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cni-path\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009655 kubelet[1425]: I1213 14:21:08.009346 1425 reconciler_common.go:300] "Volume detached for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-clustermesh-secrets\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009655 kubelet[1425]: I1213 14:21:08.009356 1425 reconciler_common.go:300] "Volume detached for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-hubble-tls\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009655 kubelet[1425]: I1213 14:21:08.009366 1425 reconciler_common.go:300] "Volume detached for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-etc-cni-netd\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009655 kubelet[1425]: I1213 14:21:08.009376 1425 reconciler_common.go:300] "Volume detached for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-config-path\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009655 kubelet[1425]: I1213 14:21:08.009385 1425 reconciler_common.go:300] "Volume detached for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-cilium-run\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009655 kubelet[1425]: I1213 14:21:08.009394 1425 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-host-proc-sys-kernel\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009655 kubelet[1425]: I1213 14:21:08.009403 1425 reconciler_common.go:300] "Volume detached for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-host-proc-sys-net\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.009655 kubelet[1425]: I1213 14:21:08.009413 1425 reconciler_common.go:300] "Volume detached for volume \"kube-api-access-6h949\" (UniqueName: \"kubernetes.io/projected/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1-kube-api-access-6h949\") on node \"10.0.0.34\" DevicePath \"\"" Dec 13 14:21:08.510922 systemd[1]: var-lib-kubelet-pods-02d4381b\x2db6b9\x2d4940\x2d9cb0\x2dc1b6c2ef3aa1-volumes-kubernetes.io\x7eprojected-kube\x2dapi\x2daccess\x2d6h949.mount: Deactivated successfully. Dec 13 14:21:08.511031 systemd[1]: var-lib-kubelet-pods-02d4381b\x2db6b9\x2d4940\x2d9cb0\x2dc1b6c2ef3aa1-volumes-kubernetes.io\x7esecret-cilium\x2dipsec\x2dsecrets.mount: Deactivated successfully. Dec 13 14:21:08.511108 systemd[1]: var-lib-kubelet-pods-02d4381b\x2db6b9\x2d4940\x2d9cb0\x2dc1b6c2ef3aa1-volumes-kubernetes.io\x7esecret-clustermesh\x2dsecrets.mount: Deactivated successfully. Dec 13 14:21:08.511160 systemd[1]: var-lib-kubelet-pods-02d4381b\x2db6b9\x2d4940\x2d9cb0\x2dc1b6c2ef3aa1-volumes-kubernetes.io\x7eprojected-hubble\x2dtls.mount: Deactivated successfully. Dec 13 14:21:08.730986 systemd[1]: Removed slice kubepods-burstable-pod02d4381b_b6b9_4940_9cb0_c1b6c2ef3aa1.slice. Dec 13 14:21:08.758656 kubelet[1425]: I1213 14:21:08.758597 1425 topology_manager.go:215] "Topology Admit Handler" podUID="82de25bd-c4c1-4958-bc76-5b715ecc74fe" podNamespace="kube-system" podName="cilium-n65hq" Dec 13 14:21:08.764301 systemd[1]: Created slice kubepods-burstable-pod82de25bd_c4c1_4958_bc76_5b715ecc74fe.slice. Dec 13 14:21:08.778487 kubelet[1425]: E1213 14:21:08.778430 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:08.914330 kubelet[1425]: I1213 14:21:08.914271 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-net\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-host-proc-sys-net\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914330 kubelet[1425]: I1213 14:21:08.914339 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostproc\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-hostproc\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914612 kubelet[1425]: I1213 14:21:08.914462 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-run\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-cilium-run\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914612 kubelet[1425]: I1213 14:21:08.914539 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-ipsec-secrets\" (UniqueName: \"kubernetes.io/secret/82de25bd-c4c1-4958-bc76-5b715ecc74fe-cilium-ipsec-secrets\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914612 kubelet[1425]: I1213 14:21:08.914573 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbs8p\" (UniqueName: \"kubernetes.io/projected/82de25bd-c4c1-4958-bc76-5b715ecc74fe-kube-api-access-pbs8p\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914612 kubelet[1425]: I1213 14:21:08.914598 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bpf-maps\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-bpf-maps\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914727 kubelet[1425]: I1213 14:21:08.914668 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"clustermesh-secrets\" (UniqueName: \"kubernetes.io/secret/82de25bd-c4c1-4958-bc76-5b715ecc74fe-clustermesh-secrets\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914758 kubelet[1425]: I1213 14:21:08.914723 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"xtables-lock\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-xtables-lock\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914783 kubelet[1425]: I1213 14:21:08.914769 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-lib-modules\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914827 kubelet[1425]: I1213 14:21:08.914809 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-cgroup\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-cilium-cgroup\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914905 kubelet[1425]: I1213 14:21:08.914869 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cni-netd\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-etc-cni-netd\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914905 kubelet[1425]: I1213 14:21:08.914898 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-proc-sys-kernel\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-host-proc-sys-kernel\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.914958 kubelet[1425]: I1213 14:21:08.914945 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hubble-tls\" (UniqueName: \"kubernetes.io/projected/82de25bd-c4c1-4958-bc76-5b715ecc74fe-hubble-tls\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.915000 kubelet[1425]: I1213 14:21:08.914976 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cilium-config-path\" (UniqueName: \"kubernetes.io/configmap/82de25bd-c4c1-4958-bc76-5b715ecc74fe-cilium-config-path\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:08.915027 kubelet[1425]: I1213 14:21:08.915021 1425 reconciler_common.go:258] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-path\" (UniqueName: \"kubernetes.io/host-path/82de25bd-c4c1-4958-bc76-5b715ecc74fe-cni-path\") pod \"cilium-n65hq\" (UID: \"82de25bd-c4c1-4958-bc76-5b715ecc74fe\") " pod="kube-system/cilium-n65hq" Dec 13 14:21:09.071809 kubelet[1425]: E1213 14:21:09.071758 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:09.072491 env[1209]: time="2024-12-13T14:21:09.072431897Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-n65hq,Uid:82de25bd-c4c1-4958-bc76-5b715ecc74fe,Namespace:kube-system,Attempt:0,}" Dec 13 14:21:09.086991 env[1209]: time="2024-12-13T14:21:09.086906624Z" level=info msg="loading plugin \"io.containerd.event.v1.publisher\"..." runtime=io.containerd.runc.v2 type=io.containerd.event.v1 Dec 13 14:21:09.086991 env[1209]: time="2024-12-13T14:21:09.086950316Z" level=info msg="loading plugin \"io.containerd.internal.v1.shutdown\"..." runtime=io.containerd.runc.v2 type=io.containerd.internal.v1 Dec 13 14:21:09.086991 env[1209]: time="2024-12-13T14:21:09.086961166Z" level=info msg="loading plugin \"io.containerd.ttrpc.v1.task\"..." runtime=io.containerd.runc.v2 type=io.containerd.ttrpc.v1 Dec 13 14:21:09.087218 env[1209]: time="2024-12-13T14:21:09.087107651Z" level=info msg="starting signal loop" namespace=k8s.io path=/run/containerd/io.containerd.runtime.v2.task/k8s.io/49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b pid=3040 runtime=io.containerd.runc.v2 Dec 13 14:21:09.096904 systemd[1]: Started cri-containerd-49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b.scope. Dec 13 14:21:09.119047 env[1209]: time="2024-12-13T14:21:09.118974418Z" level=info msg="RunPodSandbox for &PodSandboxMetadata{Name:cilium-n65hq,Uid:82de25bd-c4c1-4958-bc76-5b715ecc74fe,Namespace:kube-system,Attempt:0,} returns sandbox id \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\"" Dec 13 14:21:09.119885 kubelet[1425]: E1213 14:21:09.119857 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:09.122240 env[1209]: time="2024-12-13T14:21:09.122198950Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for container &ContainerMetadata{Name:mount-cgroup,Attempt:0,}" Dec 13 14:21:09.135966 env[1209]: time="2024-12-13T14:21:09.135887370Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for &ContainerMetadata{Name:mount-cgroup,Attempt:0,} returns container id \"783de563bdd6107d94c0a0bc7659f9ee3e3db77e1b07a985d9eda65ab66823a0\"" Dec 13 14:21:09.136458 env[1209]: time="2024-12-13T14:21:09.136429778Z" level=info msg="StartContainer for \"783de563bdd6107d94c0a0bc7659f9ee3e3db77e1b07a985d9eda65ab66823a0\"" Dec 13 14:21:09.155583 systemd[1]: Started cri-containerd-783de563bdd6107d94c0a0bc7659f9ee3e3db77e1b07a985d9eda65ab66823a0.scope. Dec 13 14:21:09.182422 env[1209]: time="2024-12-13T14:21:09.182370609Z" level=info msg="StartContainer for \"783de563bdd6107d94c0a0bc7659f9ee3e3db77e1b07a985d9eda65ab66823a0\" returns successfully" Dec 13 14:21:09.190749 systemd[1]: cri-containerd-783de563bdd6107d94c0a0bc7659f9ee3e3db77e1b07a985d9eda65ab66823a0.scope: Deactivated successfully. Dec 13 14:21:09.219314 env[1209]: time="2024-12-13T14:21:09.219258791Z" level=info msg="shim disconnected" id=783de563bdd6107d94c0a0bc7659f9ee3e3db77e1b07a985d9eda65ab66823a0 Dec 13 14:21:09.219314 env[1209]: time="2024-12-13T14:21:09.219302082Z" level=warning msg="cleaning up after shim disconnected" id=783de563bdd6107d94c0a0bc7659f9ee3e3db77e1b07a985d9eda65ab66823a0 namespace=k8s.io Dec 13 14:21:09.219314 env[1209]: time="2024-12-13T14:21:09.219311460Z" level=info msg="cleaning up dead shim" Dec 13 14:21:09.227548 env[1209]: time="2024-12-13T14:21:09.227501139Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:21:09Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3124 runtime=io.containerd.runc.v2\n" Dec 13 14:21:09.320567 kubelet[1425]: I1213 14:21:09.320507 1425 setters.go:568] "Node became not ready" node="10.0.0.34" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2024-12-13T14:21:09Z","lastTransitionTime":"2024-12-13T14:21:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized"} Dec 13 14:21:09.583196 kubelet[1425]: I1213 14:21:09.583140 1425 kubelet_volumes.go:161] "Cleaned up orphaned pod volumes dir" podUID="02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1" path="/var/lib/kubelet/pods/02d4381b-b6b9-4940-9cb0-c1b6c2ef3aa1/volumes" Dec 13 14:21:09.736467 kubelet[1425]: E1213 14:21:09.734520 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:09.735965 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3840019073.mount: Deactivated successfully. Dec 13 14:21:09.742123 env[1209]: time="2024-12-13T14:21:09.742040660Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for container &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,}" Dec 13 14:21:09.759132 env[1209]: time="2024-12-13T14:21:09.759070561Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for &ContainerMetadata{Name:apply-sysctl-overwrites,Attempt:0,} returns container id \"868f464a0072c59c5812872468ef786c6dbb7512faf94604e8024ebaadd68410\"" Dec 13 14:21:09.759866 env[1209]: time="2024-12-13T14:21:09.759814427Z" level=info msg="StartContainer for \"868f464a0072c59c5812872468ef786c6dbb7512faf94604e8024ebaadd68410\"" Dec 13 14:21:09.776237 systemd[1]: Started cri-containerd-868f464a0072c59c5812872468ef786c6dbb7512faf94604e8024ebaadd68410.scope. Dec 13 14:21:09.779459 kubelet[1425]: E1213 14:21:09.779426 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:09.802823 env[1209]: time="2024-12-13T14:21:09.802767301Z" level=info msg="StartContainer for \"868f464a0072c59c5812872468ef786c6dbb7512faf94604e8024ebaadd68410\" returns successfully" Dec 13 14:21:09.808589 systemd[1]: cri-containerd-868f464a0072c59c5812872468ef786c6dbb7512faf94604e8024ebaadd68410.scope: Deactivated successfully. Dec 13 14:21:09.841284 env[1209]: time="2024-12-13T14:21:09.841129499Z" level=info msg="shim disconnected" id=868f464a0072c59c5812872468ef786c6dbb7512faf94604e8024ebaadd68410 Dec 13 14:21:09.841284 env[1209]: time="2024-12-13T14:21:09.841204029Z" level=warning msg="cleaning up after shim disconnected" id=868f464a0072c59c5812872468ef786c6dbb7512faf94604e8024ebaadd68410 namespace=k8s.io Dec 13 14:21:09.841284 env[1209]: time="2024-12-13T14:21:09.841219077Z" level=info msg="cleaning up dead shim" Dec 13 14:21:09.849622 env[1209]: time="2024-12-13T14:21:09.849571863Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:21:09Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3187 runtime=io.containerd.runc.v2\n" Dec 13 14:21:10.381002 env[1209]: time="2024-12-13T14:21:10.380929274Z" level=info msg="ImageCreate event &ImageCreate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:21:10.383434 env[1209]: time="2024-12-13T14:21:10.383375654Z" level=info msg="ImageCreate event &ImageCreate{Name:sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:21:10.385231 env[1209]: time="2024-12-13T14:21:10.385174490Z" level=info msg="ImageUpdate event &ImageUpdate{Name:quay.io/cilium/operator-generic@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e,Labels:map[string]string{io.cri-containerd.image: managed,},XXX_unrecognized:[],}" Dec 13 14:21:10.385646 env[1209]: time="2024-12-13T14:21:10.385579741Z" level=info msg="PullImage \"quay.io/cilium/operator-generic:v1.12.5@sha256:b296eb7f0f7656a5cc19724f40a8a7121b7fd725278b7d61dc91fe0b7ffd7c0e\" returns image reference \"sha256:ed355de9f59fe391dbe53f3c7c7a60baab3c3a9b7549aa54d10b87fff7dacf7c\"" Dec 13 14:21:10.387487 env[1209]: time="2024-12-13T14:21:10.387455090Z" level=info msg="CreateContainer within sandbox \"fb2c719b3caf002635795f8562e5d4b4a2d583bbc01fe6ea4ad648bc9ae5744c\" for container &ContainerMetadata{Name:cilium-operator,Attempt:0,}" Dec 13 14:21:10.400287 env[1209]: time="2024-12-13T14:21:10.400217171Z" level=info msg="CreateContainer within sandbox \"fb2c719b3caf002635795f8562e5d4b4a2d583bbc01fe6ea4ad648bc9ae5744c\" for &ContainerMetadata{Name:cilium-operator,Attempt:0,} returns container id \"01dc70305c58430d47536b10e9410dc7f10a1602d2d08dc4accd9bfbb6a63d4b\"" Dec 13 14:21:10.401025 env[1209]: time="2024-12-13T14:21:10.400979422Z" level=info msg="StartContainer for \"01dc70305c58430d47536b10e9410dc7f10a1602d2d08dc4accd9bfbb6a63d4b\"" Dec 13 14:21:10.417535 systemd[1]: Started cri-containerd-01dc70305c58430d47536b10e9410dc7f10a1602d2d08dc4accd9bfbb6a63d4b.scope. Dec 13 14:21:10.446303 env[1209]: time="2024-12-13T14:21:10.446239861Z" level=info msg="StartContainer for \"01dc70305c58430d47536b10e9410dc7f10a1602d2d08dc4accd9bfbb6a63d4b\" returns successfully" Dec 13 14:21:10.735012 kubelet[1425]: E1213 14:21:10.734857 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:10.737091 kubelet[1425]: E1213 14:21:10.737058 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:10.737468 env[1209]: time="2024-12-13T14:21:10.737415981Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for container &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,}" Dec 13 14:21:10.779855 kubelet[1425]: E1213 14:21:10.779760 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:10.892555 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount1036572137.mount: Deactivated successfully. Dec 13 14:21:10.894801 env[1209]: time="2024-12-13T14:21:10.894747866Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for &ContainerMetadata{Name:mount-bpf-fs,Attempt:0,} returns container id \"05a8156def97305630f87e16788e03f1c30ceeb41a95bc51b794ad431e84ff88\"" Dec 13 14:21:10.895404 env[1209]: time="2024-12-13T14:21:10.895371475Z" level=info msg="StartContainer for \"05a8156def97305630f87e16788e03f1c30ceeb41a95bc51b794ad431e84ff88\"" Dec 13 14:21:10.914318 systemd[1]: Started cri-containerd-05a8156def97305630f87e16788e03f1c30ceeb41a95bc51b794ad431e84ff88.scope. Dec 13 14:21:10.942298 env[1209]: time="2024-12-13T14:21:10.941999743Z" level=info msg="StartContainer for \"05a8156def97305630f87e16788e03f1c30ceeb41a95bc51b794ad431e84ff88\" returns successfully" Dec 13 14:21:10.950245 systemd[1]: cri-containerd-05a8156def97305630f87e16788e03f1c30ceeb41a95bc51b794ad431e84ff88.scope: Deactivated successfully. Dec 13 14:21:10.989071 env[1209]: time="2024-12-13T14:21:10.988925218Z" level=info msg="shim disconnected" id=05a8156def97305630f87e16788e03f1c30ceeb41a95bc51b794ad431e84ff88 Dec 13 14:21:10.989071 env[1209]: time="2024-12-13T14:21:10.988980251Z" level=warning msg="cleaning up after shim disconnected" id=05a8156def97305630f87e16788e03f1c30ceeb41a95bc51b794ad431e84ff88 namespace=k8s.io Dec 13 14:21:10.989071 env[1209]: time="2024-12-13T14:21:10.988990099Z" level=info msg="cleaning up dead shim" Dec 13 14:21:10.997112 env[1209]: time="2024-12-13T14:21:10.997052580Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:21:10Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3282 runtime=io.containerd.runc.v2\n" Dec 13 14:21:11.511381 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-05a8156def97305630f87e16788e03f1c30ceeb41a95bc51b794ad431e84ff88-rootfs.mount: Deactivated successfully. Dec 13 14:21:11.741866 kubelet[1425]: E1213 14:21:11.741442 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:11.741866 kubelet[1425]: E1213 14:21:11.741695 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:11.743927 env[1209]: time="2024-12-13T14:21:11.743857013Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for container &ContainerMetadata{Name:clean-cilium-state,Attempt:0,}" Dec 13 14:21:11.757575 kubelet[1425]: I1213 14:21:11.757526 1425 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-operator-5cc964979-nhbfl" podStartSLOduration=1.988536758 podStartE2EDuration="4.757472685s" podCreationTimestamp="2024-12-13 14:21:07 +0000 UTC" firstStartedPulling="2024-12-13 14:21:07.61696717 +0000 UTC m=+71.331452428" lastFinishedPulling="2024-12-13 14:21:10.385903107 +0000 UTC m=+74.100388355" observedRunningTime="2024-12-13 14:21:10.818042383 +0000 UTC m=+74.532527621" watchObservedRunningTime="2024-12-13 14:21:11.757472685 +0000 UTC m=+75.471957933" Dec 13 14:21:11.761249 systemd[1]: var-lib-containerd-tmpmounts-containerd\x2dmount3985234417.mount: Deactivated successfully. Dec 13 14:21:11.761868 env[1209]: time="2024-12-13T14:21:11.761726557Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for &ContainerMetadata{Name:clean-cilium-state,Attempt:0,} returns container id \"8c6267b4dbf8d7d12a04dbee3a4ea6edd5564dbe63ad87c3c6a705a65a398acf\"" Dec 13 14:21:11.762683 env[1209]: time="2024-12-13T14:21:11.762624161Z" level=info msg="StartContainer for \"8c6267b4dbf8d7d12a04dbee3a4ea6edd5564dbe63ad87c3c6a705a65a398acf\"" Dec 13 14:21:11.780498 kubelet[1425]: E1213 14:21:11.780456 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:11.783990 systemd[1]: Started cri-containerd-8c6267b4dbf8d7d12a04dbee3a4ea6edd5564dbe63ad87c3c6a705a65a398acf.scope. Dec 13 14:21:11.818476 systemd[1]: cri-containerd-8c6267b4dbf8d7d12a04dbee3a4ea6edd5564dbe63ad87c3c6a705a65a398acf.scope: Deactivated successfully. Dec 13 14:21:11.818934 env[1209]: time="2024-12-13T14:21:11.818889019Z" level=info msg="StartContainer for \"8c6267b4dbf8d7d12a04dbee3a4ea6edd5564dbe63ad87c3c6a705a65a398acf\" returns successfully" Dec 13 14:21:11.848460 env[1209]: time="2024-12-13T14:21:11.848392776Z" level=info msg="shim disconnected" id=8c6267b4dbf8d7d12a04dbee3a4ea6edd5564dbe63ad87c3c6a705a65a398acf Dec 13 14:21:11.848460 env[1209]: time="2024-12-13T14:21:11.848455393Z" level=warning msg="cleaning up after shim disconnected" id=8c6267b4dbf8d7d12a04dbee3a4ea6edd5564dbe63ad87c3c6a705a65a398acf namespace=k8s.io Dec 13 14:21:11.848747 env[1209]: time="2024-12-13T14:21:11.848470772Z" level=info msg="cleaning up dead shim" Dec 13 14:21:11.856177 env[1209]: time="2024-12-13T14:21:11.856122702Z" level=warning msg="cleanup warnings time=\"2024-12-13T14:21:11Z\" level=info msg=\"starting signal loop\" namespace=k8s.io pid=3337 runtime=io.containerd.runc.v2\n" Dec 13 14:21:12.511488 systemd[1]: run-containerd-runc-k8s.io-8c6267b4dbf8d7d12a04dbee3a4ea6edd5564dbe63ad87c3c6a705a65a398acf-runc.nxNRkj.mount: Deactivated successfully. Dec 13 14:21:12.511584 systemd[1]: run-containerd-io.containerd.runtime.v2.task-k8s.io-8c6267b4dbf8d7d12a04dbee3a4ea6edd5564dbe63ad87c3c6a705a65a398acf-rootfs.mount: Deactivated successfully. Dec 13 14:21:12.674525 kubelet[1425]: E1213 14:21:12.674471 1425 kubelet.go:2892] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: cni plugin not initialized" Dec 13 14:21:12.745711 kubelet[1425]: E1213 14:21:12.745674 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:12.747481 env[1209]: time="2024-12-13T14:21:12.747433470Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for container &ContainerMetadata{Name:cilium-agent,Attempt:0,}" Dec 13 14:21:12.781475 kubelet[1425]: E1213 14:21:12.781007 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:12.841261 env[1209]: time="2024-12-13T14:21:12.841185149Z" level=info msg="CreateContainer within sandbox \"49a5df5feb26f8aba5e1ad6fc0a15677fb4eaa1a15cdb93963d1f200460adb1b\" for &ContainerMetadata{Name:cilium-agent,Attempt:0,} returns container id \"0300fa4d03d1a1b1ae4aafbb941986c69f58dde012058571d86af80884771464\"" Dec 13 14:21:12.842093 env[1209]: time="2024-12-13T14:21:12.842052056Z" level=info msg="StartContainer for \"0300fa4d03d1a1b1ae4aafbb941986c69f58dde012058571d86af80884771464\"" Dec 13 14:21:12.857947 systemd[1]: Started cri-containerd-0300fa4d03d1a1b1ae4aafbb941986c69f58dde012058571d86af80884771464.scope. Dec 13 14:21:12.900377 env[1209]: time="2024-12-13T14:21:12.900297866Z" level=info msg="StartContainer for \"0300fa4d03d1a1b1ae4aafbb941986c69f58dde012058571d86af80884771464\" returns successfully" Dec 13 14:21:13.352879 kernel: alg: No test for seqiv(rfc4106(gcm(aes))) (seqiv(rfc4106-gcm-aesni)) Dec 13 14:21:13.751770 kubelet[1425]: E1213 14:21:13.751629 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:13.766608 kubelet[1425]: I1213 14:21:13.766555 1425 pod_startup_latency_tracker.go:102] "Observed pod startup duration" pod="kube-system/cilium-n65hq" podStartSLOduration=5.766509957 podStartE2EDuration="5.766509957s" podCreationTimestamp="2024-12-13 14:21:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2024-12-13 14:21:13.766334958 +0000 UTC m=+77.480820226" watchObservedRunningTime="2024-12-13 14:21:13.766509957 +0000 UTC m=+77.480995195" Dec 13 14:21:13.781514 kubelet[1425]: E1213 14:21:13.781439 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:14.781761 kubelet[1425]: E1213 14:21:14.781697 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:15.073515 kubelet[1425]: E1213 14:21:15.073456 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:15.772776 systemd[1]: run-containerd-runc-k8s.io-0300fa4d03d1a1b1ae4aafbb941986c69f58dde012058571d86af80884771464-runc.wlteCc.mount: Deactivated successfully. Dec 13 14:21:15.782478 kubelet[1425]: E1213 14:21:15.782440 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:16.030549 systemd-networkd[1038]: lxc_health: Link UP Dec 13 14:21:16.041092 kernel: IPv6: ADDRCONF(NETDEV_CHANGE): lxc_health: link becomes ready Dec 13 14:21:16.040917 systemd-networkd[1038]: lxc_health: Gained carrier Dec 13 14:21:16.725090 kubelet[1425]: E1213 14:21:16.725029 1425 file.go:104] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:16.782670 kubelet[1425]: E1213 14:21:16.782616 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:17.073667 kubelet[1425]: E1213 14:21:17.073624 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:17.647248 systemd-networkd[1038]: lxc_health: Gained IPv6LL Dec 13 14:21:17.757416 kubelet[1425]: E1213 14:21:17.757373 1425 dns.go:153] "Nameserver limits exceeded" err="Nameserver limits were exceeded, some nameservers have been omitted, the applied nameserver line is: 1.1.1.1 1.0.0.1 8.8.8.8" Dec 13 14:21:17.783105 kubelet[1425]: E1213 14:21:17.783058 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:17.865742 systemd[1]: run-containerd-runc-k8s.io-0300fa4d03d1a1b1ae4aafbb941986c69f58dde012058571d86af80884771464-runc.GZqYoM.mount: Deactivated successfully. Dec 13 14:21:18.784285 kubelet[1425]: E1213 14:21:18.784192 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:19.786912 kubelet[1425]: E1213 14:21:19.786869 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:19.968525 systemd[1]: run-containerd-runc-k8s.io-0300fa4d03d1a1b1ae4aafbb941986c69f58dde012058571d86af80884771464-runc.QZphxI.mount: Deactivated successfully. Dec 13 14:21:20.788582 kubelet[1425]: E1213 14:21:20.788444 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:21.788895 kubelet[1425]: E1213 14:21:21.788819 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:22.060728 systemd[1]: run-containerd-runc-k8s.io-0300fa4d03d1a1b1ae4aafbb941986c69f58dde012058571d86af80884771464-runc.uznusM.mount: Deactivated successfully. Dec 13 14:21:22.789720 kubelet[1425]: E1213 14:21:22.789617 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests" Dec 13 14:21:23.790414 kubelet[1425]: E1213 14:21:23.790340 1425 file_linux.go:61] "Unable to read config path" err="path does not exist, ignoring" path="/etc/kubernetes/manifests"