Job:
#OCPBUGS-32517issue42 hours agoMissing worker nodes on metal Verified
Mon 2024-04-22 05:33:53 UTC localhost.localdomain master-bmh-update.service[12603]: Unpause all baremetal hosts
Mon 2024-04-22 05:33:53 UTC localhost.localdomain master-bmh-update.service[18264]: E0422 05:33:53.630867   18264 memcache.go:265] couldn't get current server API group list: Get "https://localhost:6443/api?timeout=32s": dial tcp [::1]:6443: connect: connection refused
Mon 2024-04-22 05:33:53 UTC localhost.localdomain master-bmh-update.service[18264]: E0422 05:33:53.631351   18264 memcache.go:265] couldn't get current server API group list: Get "https://localhost:6443/api?timeout=32s": dial tcp [::1]:6443: connect: connection refused

... 4 lines not shown

#OCPBUGS-27755issue9 days agoopenshift-kube-apiserver down and is not being restarted New
Issue 15736514: openshift-kube-apiserver down and is not being restarted
Description: Description of problem:
 {code:none}
 SNO cluster, this is the second time that the issue happens. 
 
 Error like the following are reported:
 
 ~~~
 failed to fetch token: Post "https://api-int.<cluster>:6443/api/v1/namespaces/openshift-cluster-storage-operator/serviceaccounts/cluster-storage-operator/token": dial tcp <ip>:6443: connect: connection refused
 ~~~
 
 Checking the pods logs, kube-apiserver pod is terminated and is not being restarted again:
 
 ~~~
 2024-01-13T09:41:40.931716166Z I0113 09:41:40.931584       1 main.go:213] Received signal terminated. Forwarding to sub-process "hyperkube".
 ~~~{code}
 Version-Release number of selected component (if applicable):
 {code:none}
    4.13.13 {code}
 How reproducible:
 {code:none}
     Not reproducible but has happened twice{code}
 Steps to Reproduce:
 {code:none}
     1.
     2.
     3.
     {code}
 Actual results:
 {code:none}
     API is not available and kube-apiserver is not being restarted{code}
 Expected results:
 {code:none}
     We would expect to see kube-apiserver restarts{code}
 Additional info:
 {code:none}
    {code}
Status: New
#OCPBUGS-30631issue2 weeks agoSNO (RT kernel) sosreport crash the SNO node CLOSED
Issue 15865131: SNO (RT kernel) sosreport crash the SNO node
Description: Description of problem:
 {code:none}
 sosreport collection causes SNO XR11 node crash.
 {code}
 Version-Release number of selected component (if applicable):
 {code:none}
 - RHOCP    : 4.12.30
 - kernel   : 4.18.0-372.69.1.rt7.227.el8_6.x86_64
 - platform : x86_64{code}
 How reproducible:
 {code:none}
 sh-4.4# chrt -rr 99 toolbox
 .toolboxrc file detected, overriding defaults...
 Checking if there is a newer version of ocpdalmirror.xxx.yyy:8443/rhel8/support-tools-zzz-feb available...
 Container 'toolbox-root' already exists. Trying to start...
 (To remove the container and start with a fresh toolbox, run: sudo podman rm 'toolbox-root')
 toolbox-root
 Container started successfully. To exit, type 'exit'.
 [root@node /]# which sos
 /usr/sbin/sos
 logger: socket /dev/log: No such file or directory
 [root@node /]# taskset -c 29-31,61-63 sos report --batch -n networking,kernel,processor -k crio.all=on -k crio.logs=on -k podman.all=on -kpodman.logs=on
 
 sosreport (version 4.5.6)
 
 This command will collect diagnostic and configuration information from
 this Red Hat CoreOS system.
 
 An archive containing the collected information will be generated in
 /host/var/tmp/sos.c09e4f7z and may be provided to a Red Hat support
 representative.
 
 Any information provided to Red Hat will be treated in accordance with
 the published support policies at:
 
         Distribution Website : https://www.redhat.com/
         Commercial Support   : https://access.redhat.com/
 
 The generated archive may contain data considered sensitive and its
 content should be reviewed by the originating organization before being
 passed to any third party.
 
 No changes will be made to system configuration.
 
 
  Setting up archive ...
  Setting up plugins ...
 [plugin:auditd] Could not open conf file /etc/audit/auditd.conf: [Errno 2] No such file or directory: '/etc/audit/auditd.conf'
 caught exception in plugin method "system.setup()"
 writing traceback to sos_logs/system-plugin-errors.txt
 [plugin:systemd] skipped command 'resolvectl status': required services missing: systemd-resolved.
 [plugin:systemd] skipped command 'resolvectl statistics': required services missing: systemd-resolved.
  Running plugins. Please wait ...
 
   Starting 1/91  alternatives    [Running: alternatives]
   Starting 2/91  atomichost      [Running: alternatives atomichost]
   Starting 3/91  auditd          [Running: alternatives atomichost auditd]
   Starting 4/91  block           [Running: alternatives atomichost auditd block]
   Starting 5/91  boot            [Running: alternatives auditd block boot]
   Starting 6/91  cgroups         [Running: auditd block boot cgroups]
   Starting 7/91  chrony          [Running: auditd block cgroups chrony]
   Starting 8/91  cifs            [Running: auditd block cgroups cifs]
   Starting 9/91  conntrack       [Running: auditd block cgroups conntrack]
   Starting 10/91 console         [Running: block cgroups conntrack console]
   Starting 11/91 container_log   [Running: block cgroups conntrack container_log]
   Starting 12/91 containers_common [Running: block cgroups conntrack containers_common]
   Starting 13/91 crio            [Running: block cgroups conntrack crio]
   Starting 14/91 crypto          [Running: cgroups conntrack crio crypto]
   Starting 15/91 date            [Running: cgroups conntrack crio date]
   Starting 16/91 dbus            [Running: cgroups conntrack crio dbus]
   Starting 17/91 devicemapper    [Running: cgroups conntrack crio devicemapper]
   Starting 18/91 devices         [Running: cgroups conntrack crio devices]
   Starting 19/91 dracut          [Running: cgroups conntrack crio dracut]
   Starting 20/91 ebpf            [Running: cgroups conntrack crio ebpf]
   Starting 21/91 etcd            [Running: cgroups crio ebpf etcd]
   Starting 22/91 filesys         [Running: cgroups crio ebpf filesys]
   Starting 23/91 firewall_tables [Running: cgroups crio filesys firewall_tables]
   Starting 24/91 fwupd           [Running: cgroups crio filesys fwupd]
   Starting 25/91 gluster         [Running: cgroups crio filesys gluster]
   Starting 26/91 grub2           [Running: cgroups crio filesys grub2]
   Starting 27/91 gssproxy        [Running: cgroups crio grub2 gssproxy]
   Starting 28/91 hardware        [Running: cgroups crio grub2 hardware]
   Starting 29/91 host            [Running: cgroups crio hardware host]
   Starting 30/91 hts             [Running: cgroups crio hardware hts]
   Starting 31/91 i18n            [Running: cgroups crio hardware i18n]
   Starting 32/91 iscsi           [Running: cgroups crio hardware iscsi]
   Starting 33/91 jars            [Running: cgroups crio hardware jars]
   Starting 34/91 kdump           [Running: cgroups crio hardware kdump]
   Starting 35/91 kernelrt        [Running: cgroups crio hardware kernelrt]
   Starting 36/91 keyutils        [Running: cgroups crio hardware keyutils]
   Starting 37/91 krb5            [Running: cgroups crio hardware krb5]
   Starting 38/91 kvm             [Running: cgroups crio hardware kvm]
   Starting 39/91 ldap            [Running: cgroups crio kvm ldap]
   Starting 40/91 libraries       [Running: cgroups crio kvm libraries]
   Starting 41/91 libvirt         [Running: cgroups crio kvm libvirt]
   Starting 42/91 login           [Running: cgroups crio kvm login]
   Starting 43/91 logrotate       [Running: cgroups crio kvm logrotate]
   Starting 44/91 logs            [Running: cgroups crio kvm logs]
   Starting 45/91 lvm2            [Running: cgroups crio logs lvm2]
   Starting 46/91 md              [Running: cgroups crio logs md]
   Starting 47/91 memory          [Running: cgroups crio logs memory]
   Starting 48/91 microshift_ovn  [Running: cgroups crio logs microshift_ovn]
   Starting 49/91 multipath       [Running: cgroups crio logs multipath]
   Starting 50/91 networkmanager  [Running: cgroups crio logs networkmanager]
 
 Removing debug pod ...
 error: unable to delete the debug pod "ransno1ransnomavdallabcom-debug": Delete "https://api.ransno.mavdallab.com:6443/api/v1/namespaces/openshift-debug-mt82m/pods/ransno1ransnomavdallabcom-debug": dial tcp 10.71.136.144:6443: connect: connection refused
 {code}
 Steps to Reproduce:
 {code:none}
 Launch a debug pod and the procedure above and it crash the node{code}
 Actual results:
 {code:none}
 Node crash{code}
 Expected results:
 {code:none}
 Node does not crash{code}
 Additional info:
 {code:none}
 We have two vmcore on the associated SFDC ticket.
 This system use a RT kernel.
 Using an out of tree ice driver 1.13.7 (probably from 22 dec 2023)
 
 [  103.681608] ice: module unloaded
 [  103.830535] ice: loading out-of-tree module taints kernel.
 [  103.831106] ice: module verification failed: signature and/or required key missing - tainting kernel
 [  103.841005] ice: Intel(R) Ethernet Connection E800 Series Linux Driver - version 1.13.7
 [  103.841017] ice: Copyright (C) 2018-2023 Intel Corporation
 
 
 With the following kernel command line 
 
 Command line: BOOT_IMAGE=(hd0,gpt3)/ostree/rhcos-f2c287e549b45a742b62e4f748bc2faae6ca907d24bb1e029e4985bc01649033/vmlinuz-4.18.0-372.69.1.rt7.227.el8_6.x86_64 ignition.platform.id=metal ostree=/ostree/boot.1/rhcos/f2c287e549b45a742b62e4f748bc2faae6ca907d24bb1e029e4985bc01649033/0 root=UUID=3e8bda80-5cf4-4c46-b139-4c84cb006354 rw rootflags=prjquota boot=UUID=1d0512c2-3f92-42c5-b26d-709ff9350b81 intel_iommu=on iommu=pt firmware_class.path=/var/lib/firmware skew_tick=1 nohz=on rcu_nocbs=3-31,35-63 tuned.non_isolcpus=00000007,00000007 systemd.cpu_affinity=0,1,2,32,33,34 intel_iommu=on iommu=pt isolcpus=managed_irq,3-31,35-63 nohz_full=3-31,35-63 tsc=nowatchdog nosoftlockup nmi_watchdog=0 mce=off rcutree.kthread_prio=11 default_hugepagesz=1G rcupdate.rcu_normal_after_boot=0 efi=runtime module_blacklist=irdma intel_pstate=passive intel_idle.max_cstate=0 crashkernel=256M
 
 
 
 vmcore1 show issue with the ice driver 
 
 crash vmcore tmp/vmlinux
 
 
       KERNEL: tmp/vmlinux  [TAINTED]
     DUMPFILE: vmcore  [PARTIAL DUMP]
         CPUS: 64
         DATE: Thu Mar  7 17:16:57 CET 2024
       UPTIME: 02:44:28
 LOAD AVERAGE: 24.97, 25.47, 25.46
        TASKS: 5324
     NODENAME: aaa.bbb.ccc
      RELEASE: 4.18.0-372.69.1.rt7.227.el8_6.x86_64
      VERSION: #1 SMP PREEMPT_RT Fri Aug 4 00:21:46 EDT 2023
      MACHINE: x86_64  (1500 Mhz)
       MEMORY: 127.3 GB
        PANIC: "Kernel panic - not syncing:"
          PID: 693
      COMMAND: "khungtaskd"
         TASK: ff4d1890260d4000  [THREAD_INFO: ff4d1890260d4000]
          CPU: 0
        STATE: TASK_RUNNING (PANIC)
 
 crash> ps|grep sos                                                                                                                                                                                                                                                                                                           
   449071  363440  31  ff4d189005f68000  IN   0.2  506428 314484  sos                                                                                                                                                                                                                                                         
   451043  363440  63  ff4d188943a9c000  IN   0.2  506428 314484  sos                                                                                                                                                                                                                                                         
   494099  363440  29  ff4d187f941f4000  UN   0.2  506428 314484  sos     
 
  8457.517696] ------------[ cut here ]------------
 [ 8457.517698] NETDEV WATCHDOG: ens3f1 (ice): transmit queue 35 timed out
 [ 8457.517711] WARNING: CPU: 33 PID: 349 at net/sched/sch_generic.c:472 dev_watchdog+0x270/0x300
 [ 8457.517718] Modules linked in: binfmt_misc macvlan pci_pf_stub iavf vfio_pci vfio_virqfd vfio_iommu_type1 vfio vhost_net vhost vhost_iotlb tap tun xt_addrtype nf_conntrack_netlink ip6t_REJECT nf_reject_ipv6 ipt_REJECT nf_reject_ipv4 xt_nat xt_CT tcp_diag inet_diag ip6t_MASQUERADE xt_mark ice(OE) xt_conntrack ipt_MASQUERADE nft_counter xt_comment nft_compat veth nft_chain_nat nf_tables overlay bridge 8021q garp mrp stp llc nfnetlink_cttimeout nfnetlink openvswitch nf_conncount nf_nat nf_conntrack nf_defrag_ipv6 nf_defrag_ipv4 ext4 mbcache jbd2 intel_rapl_msr iTCO_wdt iTCO_vendor_support dell_smbios wmi_bmof dell_wmi_descriptor dcdbas kvm_intel kvm irqbypass intel_rapl_common i10nm_edac nfit libnvdimm x86_pkg_temp_thermal intel_powerclamp coretemp rapl ipmi_ssif intel_cstate intel_uncore dm_thin_pool pcspkr isst_if_mbox_pci dm_persistent_data dm_bio_prison dm_bufio isst_if_mmio isst_if_common mei_me i2c_i801 joydev mei intel_pmt wmi acpi_ipmi ipmi_si acpi_power_meter sctp ip6_udp_tunnel
 [ 8457.517770]  udp_tunnel ip_tables xfs libcrc32c i40e sd_mod t10_pi sg bnxt_re ib_uverbs ib_core crct10dif_pclmul crc32_pclmul crc32c_intel ghash_clmulni_intel bnxt_en ahci libahci libata dm_multipath dm_mirror dm_region_hash dm_log dm_mod ipmi_devintf ipmi_msghandler fuse [last unloaded: ice]
 [ 8457.517784] Red Hat flags: eBPF/rawtrace
 [ 8457.517787] CPU: 33 PID: 349 Comm: ktimers/33 Kdump: loaded Tainted: G           OE    --------- -  - 4.18.0-372.69.1.rt7.227.el8_6.x86_64 #1
 [ 8457.517789] Hardware name: Dell Inc. PowerEdge XR11/0P2RNT, BIOS 1.12.1 09/13/2023
 [ 8457.517790] RIP: 0010:dev_watchdog+0x270/0x300
 [ 8457.517793] Code: 17 00 e9 f0 fe ff ff 4c 89 e7 c6 05 c6 03 34 01 01 e8 14 43 fa ff 89 d9 4c 89 e6 48 c7 c7 90 37 98 9a 48 89 c2 e8 1d be 88 ff <0f> 0b eb ad 65 8b 05 05 13 fb 65 89 c0 48 0f a3 05 1b ab 36 01 73
 [ 8457.517795] RSP: 0018:ff7aeb55c73c7d78 EFLAGS: 00010286
 [ 8457.517797] RAX: 0000000000000000 RBX: 0000000000000023 RCX: 0000000000000001
 [ 8457.517798] RDX: 0000000000000000 RSI: ffffffff9a908557 RDI: 00000000ffffffff
 [ 8457.517799] RBP: 0000000000000021 R08: ffffffff9ae6b3a0 R09: 00080000000000ff
 [ 8457.517800] R10: 000000006443a462 R11: 0000000000000036 R12: ff4d187f4d1f4000
 [ 8457.517801] R13: ff4d187f4d20df00 R14: ff4d187f4d1f44a0 R15: 0000000000000080
 [ 8457.517803] FS:  0000000000000000(0000) GS:ff4d18967a040000(0000) knlGS:0000000000000000
 [ 8457.517804] CS:  0010 DS: 0000 ES: 0000 CR0: 0000000080050033
 [ 8457.517805] CR2: 00007fc47c649974 CR3: 00000019a441a005 CR4: 0000000000771ea0
 [ 8457.517806] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
 [ 8457.517807] DR3: 0000000000000000 DR6: 00000000fffe0ff0 DR7: 0000000000000400
 [ 8457.517808] PKRU: 55555554
 [ 8457.517810] Call Trace:
 [ 8457.517813]  ? test_ti_thread_flag.constprop.50+0x10/0x10
 [ 8457.517816]  ? test_ti_thread_flag.constprop.50+0x10/0x10
 [ 8457.517818]  call_timer_fn+0x32/0x1d0
 [ 8457.517822]  ? test_ti_thread_flag.constprop.50+0x10/0x10
 [ 8457.517825]  run_timer_softirq+0x1fc/0x640
 [ 8457.517828]  ? _raw_spin_unlock_irq+0x1d/0x60
 [ 8457.517833]  ? finish_task_switch+0xea/0x320
 [ 8457.517836]  ? __switch_to+0x10c/0x4d0
 [ 8457.517840]  __do_softirq+0xa5/0x33f
 [ 8457.517844]  run_timersd+0x61/0xb0
 [ 8457.517848]  smpboot_thread_fn+0x1c1/0x2b0
 [ 8457.517851]  ? smpboot_register_percpu_thread_cpumask+0x140/0x140
 [ 8457.517853]  kthread+0x151/0x170
 [ 8457.517856]  ? set_kthread_struct+0x50/0x50
 [ 8457.517858]  ret_from_fork+0x1f/0x40
 [ 8457.517861] ---[ end trace 0000000000000002 ]---
 [ 8458.520445] ice 0000:8a:00.1 ens3f1: tx_timeout: VSI_num: 14, Q 35, NTC: 0x99, HW_HEAD: 0x14, NTU: 0x15, INT: 0x0
 [ 8458.520451] ice 0000:8a:00.1 ens3f1: tx_timeout recovery level 1, txqueue 35
 [ 8506.139246] ice 0000:8a:00.1: PTP reset successful
 [ 8506.437047] ice 0000:8a:00.1: VSI rebuilt. VSI index 0, type ICE_VSI_PF
 [ 8506.445482] ice 0000:8a:00.1: VSI rebuilt. VSI index 1, type ICE_VSI_CTRL
 [ 8540.459707] ice 0000:8a:00.1 ens3f1: tx_timeout: VSI_num: 14, Q 35, NTC: 0xe3, HW_HEAD: 0xe7, NTU: 0xe8, INT: 0x0
 [ 8540.459714] ice 0000:8a:00.1 ens3f1: tx_timeout recovery level 1, txqueue 35
 [ 8563.891356] ice 0000:8a:00.1: PTP reset successful
 ~~~
 
 Second vmcore on the same node show issue with the SSD drive
 
 $ crash vmcore-2 tmp/vmlinux
 
       KERNEL: tmp/vmlinux  [TAINTED]
     DUMPFILE: vmcore-2  [PARTIAL DUMP]
         CPUS: 64
         DATE: Thu Mar  7 14:29:31 CET 2024
       UPTIME: 1 days, 07:19:52
 LOAD AVERAGE: 25.55, 26.42, 28.30
        TASKS: 5409
     NODENAME: aaa.bbb.ccc
      RELEASE: 4.18.0-372.69.1.rt7.227.el8_6.x86_64
      VERSION: #1 SMP PREEMPT_RT Fri Aug 4 00:21:46 EDT 2023
      MACHINE: x86_64  (1500 Mhz)
       MEMORY: 127.3 GB
        PANIC: "Kernel panic - not syncing:"
          PID: 696
      COMMAND: "khungtaskd"
         TASK: ff2b35ed48d30000  [THREAD_INFO: ff2b35ed48d30000]
          CPU: 34
        STATE: TASK_RUNNING (PANIC)
 
 crash> ps |grep sos
   719784  718369  62  ff2b35ff00830000  IN   0.4 1215636 563388  sos
   721740  718369  61  ff2b3605579f8000  IN   0.4 1215636 563388  sos
   721742  718369  63  ff2b35fa5eb9c000  IN   0.4 1215636 563388  sos
   721744  718369  30  ff2b3603367fc000  IN   0.4 1215636 563388  sos
   721746  718369  29  ff2b360557944000  IN   0.4 1215636 563388  sos
   743356  718369  62  ff2b36042c8e0000  IN   0.4 1215636 563388  sos
   743818  718369  29  ff2b35f6186d0000  IN   0.4 1215636 563388  sos
   748518  718369  61  ff2b3602cfb84000  IN   0.4 1215636 563388  sos
   748884  718369  62  ff2b360713418000  UN   0.4 1215636 563388  sos
 
 crash> dmesg
 
 [111871.309883] ata3.00: exception Emask 0x0 SAct 0x3ff8 SErr 0x0 action 0x6 frozen
 [111871.309889] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309891] ata3.00: cmd 61/40:18:28:47:4b/00:00:00:00:00/40 tag 3 ncq dma 32768 out
                          res 40/00:00:00:00:00/00:00:00:00:00/00 Emask 0x4 (timeout)
 [111871.309895] ata3.00: status: { DRDY }
 [111871.309897] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309904] ata3.00: cmd 61/40:20:68:47:4b/00:00:00:00:00/40 tag 4 ncq dma 32768 out
                          res 40/00:00:00:00:00/00:00:00:00:00/00 Emask 0x4 (timeout)
 [111871.309908] ata3.00: status: { DRDY }
 [111871.309909] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309910] ata3.00: cmd 61/40:28:a8:47:4b/00:00:00:00:00/40 tag 5 ncq dma 32768 out
                          res 40/00:ff:00:00:00/00:00:00:00:00/40 Emask 0x4 (timeout)
 [111871.309913] ata3.00: status: { DRDY }
 [111871.309914] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309915] ata3.00: cmd 61/40:30:e8:47:4b/00:00:00:00:00/40 tag 6 ncq dma 32768 out
                          res 40/00:01:00:00:00/00:00:00:00:00/00 Emask 0x4 (timeout)
 [111871.309918] ata3.00: status: { DRDY }
 [111871.309919] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309919] ata3.00: cmd 61/70:38:48:37:2b/00:00:1c:00:00/40 tag 7 ncq dma 57344 out
                          res 40/00:00:00:00:00/00:00:00:00:00/00 Emask 0x4 (timeout)
 [111871.309922] ata3.00: status: { DRDY }
 [111871.309923] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309924] ata3.00: cmd 61/20:40:78:29:0c/00:00:19:00:00/40 tag 8 ncq dma 16384 out
                          res 40/00:ff:00:00:00/00:00:00:00:00/40 Emask 0x4 (timeout)
 [111871.309927] ata3.00: status: { DRDY }
 [111871.309928] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309929] ata3.00: cmd 61/08:48:08:0c:c0/00:00:1c:00:00/40 tag 9 ncq dma 4096 out
                          res 40/00:ff:00:00:00/00:00:00:00:00/40 Emask 0x4 (timeout)
 [111871.309932] ata3.00: status: { DRDY }
 [111871.309933] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309934] ata3.00: cmd 61/40:50:28:48:4b/00:00:00:00:00/40 tag 10 ncq dma 32768 out
                          res 40/00:01:00:00:00/00:00:00:00:00/40 Emask 0x4 (timeout)
 [111871.309937] ata3.00: status: { DRDY }
 [111871.309938] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309939] ata3.00: cmd 61/40:58:68:48:4b/00:00:00:00:00/40 tag 11 ncq dma 32768 out
                          res 40/00:00:00:00:00/00:00:00:00:00/00 Emask 0x4 (timeout)
 [111871.309942] ata3.00: status: { DRDY }
 [111871.309943] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309944] ata3.00: cmd 61/40:60:a8:48:4b/00:00:00:00:00/40 tag 12 ncq dma 32768 out
                          res 40/00:00:00:00:00/00:00:00:00:00/00 Emask 0x4 (timeout)
 [111871.309946] ata3.00: status: { DRDY }
 [111871.309947] ata3.00: failed command: WRITE FPDMA QUEUED
 [111871.309948] ata3.00: cmd 61/40:68:e8:48:4b/00:00:00:00:00/40 tag 13 ncq dma 32768 out
                          res 40/00:01:00:00:00/00:00:00:00:00/40 Emask 0x4 (timeout)
 [111871.309951] ata3.00: status: { DRDY }
 [111871.309953] ata3: hard resetting link
 ...
 ...
 ...
 [112789.787310] INFO: task sos:748884 blocked for more than 600 seconds.                                                                                                                                                                                                                                                     
 [112789.787314]       Tainted: G           OE    --------- -  - 4.18.0-372.69.1.rt7.227.el8_6.x86_64 #1                                                                                                                                                                                                                      
 [112789.787316] "echo 0 > /proc/sys/kernel/hung_task_timeout_secs" disables this message.                                                                                                                                                                                                                                    
 [112789.787316] task:sos             state:D stack:    0 pid:748884 ppid:718369 flags:0x00084080                                                                                                                                                                                                                             
 [112789.787320] Call Trace:                                                                                                                                                                                                                                                                                                  
 [112789.787323]  __schedule+0x37b/0x8e0                                                                                                                                                                                                                                                                                      
 [112789.787330]  schedule+0x6c/0x120                                                                                                                                                                                                                                                                                         
 [112789.787333]  schedule_timeout+0x2b7/0x410                                                                                                                                                                                                                                                                                
 [112789.787336]  ? enqueue_entity+0x130/0x790                                                                                                                                                                                                                                                                                
 [112789.787340]  wait_for_completion+0x84/0xf0                                                                                                                                                                                                                                                                               
 [112789.787343]  flush_work+0x120/0x1d0                                                                                                                                                                                                                                                                                      
 [112789.787347]  ? flush_workqueue_prep_pwqs+0x130/0x130                                                                                                                                                                                                                                                                     
 [112789.787350]  schedule_on_each_cpu+0xa7/0xe0                                                                                                                                                                                                                                                                              
 [112789.787353]  vmstat_refresh+0x22/0xa0                                                                                                                                                                                                                                                                                    
 [112789.787357]  proc_sys_call_handler+0x174/0x1d0                                                                                                                                                                                                                                                                           
 [112789.787361]  vfs_read+0x91/0x150                                                                                                                                                                                                                                                                                         
 [112789.787364]  ksys_read+0x52/0xc0                                                                                                                                                                                                                                                                                         
 [112789.787366]  do_syscall_64+0x87/0x1b0                                                                                                                                                                                                                                                                                    
 [112789.787369]  entry_SYSCALL_64_after_hwframe+0x61/0xc6                                                                                                                                                                                                                                                                    
 [112789.787372] RIP: 0033:0x7f2dca8c2ab4                                                                                                                                                                                                                                                                                     
 [112789.787378] Code: Unable to access opcode bytes at RIP 0x7f2dca8c2a8a.                                                                                                                                                                                                                                                   
 [112789.787378] RSP: 002b:00007f2dbbffc5e0 EFLAGS: 00000246 ORIG_RAX: 0000000000000000                                                                                                                                                                                                                                       
 [112789.787380] RAX: ffffffffffffffda RBX: 0000000000000008 RCX: 00007f2dca8c2ab4                                                                                                                                                                                                                                            
 [112789.787382] RDX: 0000000000004000 RSI: 00007f2db402b5a0 RDI: 0000000000000008                                                                                                                                                                                                                                            
 [112789.787383] RBP: 00007f2db402b5a0 R08: 0000000000000000 R09: 00007f2dcace27bb                                                                                                                                                                                                                                            
 [112789.787383] R10: 0000000000000000 R11: 0000000000000246 R12: 0000000000004000                                                                                                                                                                                                                                            
 [112789.787384] R13: 0000000000000008 R14: 00007f2db402b5a0 R15: 00007f2da4001a90                                                                                                                                                                                                                                            
 [112789.787418] NMI backtrace for cpu 34    {code}
Status: CLOSED
#OCPBUGS-33157issue42 hours agoIPv6 metal-ipi jobs: master-bmh-update loosing access to API Verified
Issue 15978085: IPv6 metal-ipi jobs: master-bmh-update loosing access to API
Description: The last 4 IPv6 jobs are failing on the same error
 
 https://prow.ci.openshift.org/job-history/gs/test-platform-results/logs/periodic-ci-openshift-release-master-nightly-4.16-e2e-metal-ipi-ovn-ipv6
 master-bmh-update.log looses access to the the API when trying to get/update the BMH details
 
 https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-master-nightly-4.16-e2e-metal-ipi-ovn-ipv6/1785492737169035264
 
 
 
 {noformat}
 May 01 03:32:23 localhost.localdomain master-bmh-update.sh[4663]: Waiting for 3 masters to become provisioned
 May 01 03:32:23 localhost.localdomain master-bmh-update.sh[24484]: E0501 03:32:23.531242   24484 memcache.go:265] couldn't get current server API group list: Get "https://api-int.ostest.test.metalkube.org:6443/api?timeout=32s": dial tcp [fd2e:6f44:5dd8:c956::5]:6443: connect: connection refused
 May 01 03:32:23 localhost.localdomain master-bmh-update.sh[24484]: E0501 03:32:23.531808   24484 memcache.go:265] couldn't get current server API group list: Get "https://api-int.ostest.test.metalkube.org:6443/api?timeout=32s": dial tcp [fd2e:6f44:5dd8:c956::5]:6443: connect: connection refused
 May 01 03:32:23 localhost.localdomain master-bmh-update.sh[24484]: E0501 03:32:23.533281   24484 memcache.go:265] couldn't get current server API group list: Get "https://api-int.ostest.test.metalkube.org:6443/api?timeout=32s": dial tcp [fd2e:6f44:5dd8:c956::5]:6443: connect: connection refused
 May 01 03:32:23 localhost.localdomain master-bmh-update.sh[24484]: E0501 03:32:23.533630   24484 memcache.go:265] couldn't get current server API group list: Get "https://api-int.ostest.test.metalkube.org:6443/api?timeout=32s": dial tcp [fd2e:6f44:5dd8:c956::5]:6443: connect: connection refused
 May 01 03:32:23 localhost.localdomain master-bmh-update.sh[24484]: E0501 03:32:23.535180   24484 memcache.go:265] couldn't get current server API group list: Get "https://api-int.ostest.test.metalkube.org:6443/api?timeout=32s": dial tcp [fd2e:6f44:5dd8:c956::5]:6443: connect: connection refused
 May 01 03:32:23 localhost.localdomain master-bmh-update.sh[24484]: The connection to the server api-int.ostest.test.metalkube.org:6443 was refused - did you specify the right host or port?
 {noformat}
Status: Verified
{noformat}
May 01 02:49:40 localhost.localdomain master-bmh-update.sh[12448]: E0501 02:49:40.429468   12448 memcache.go:265] couldn't get current server API group list: Get "https://api-int.ostest.test.metalkube.org:6443/api?timeout=32s": dial tcp [fd2e:6f44:5dd8:c956::5]:6443: connect: connection refused
{noformat}
#OCPBUGS-32375issue10 days agoUnsuccessful cluster installation with 4.15 nightlies on s390x using ABI CLOSED
Issue 15945005: Unsuccessful cluster installation with 4.15 nightlies on s390x using ABI
Description: When used the latest s390x release builds in 4.15 nightly stream for Agent Based Installation of SNO on IBM Z KVM, installation is failing at the end while watching cluster operators even though the DNS and HA Proxy configurations are perfect as the same setup is working with 4.15.x stable release image builds 
 
 Below is the error encountered multiple times when used "release:s390x-latest" image while booting the cluster. This image is used during the boot through OPENSHIFT_INSATLL_RELEASE_IMAGE_OVERRIDE while the binary is fetched using the latest stable builds from here : [https://mirror.openshift.com/pub/openshift-v4/s390x/clients/ocp/latest/] for which the version would be around 4.15.x 
 
 *release-image:*
 {code:java}
 registry.build01.ci.openshift.org/ci-op-cdkdqnqn/release@sha256:c6eb4affa5c44d2ad220d7064e92270a30df5f26d221e35664f4d5547a835617
 {code}
  ** 
 
 *PROW CI Build :* [https://prow.ci.openshift.org/view/gs/test-platform-results/pr-logs/pull/openshift_release/47965/rehearse-47965-periodic-ci-openshift-multiarch-master-nightly-4.15-e2e-agent-ibmz-sno/1780162365824700416] 
 
 *Error:* 
 {code:java}
 '/root/agent-sno/openshift-install wait-for install-complete --dir /root/agent-sno/ --log-level debug'
 Warning: Permanently added '128.168.142.71' (ED25519) to the list of known hosts.
 level=debug msg=OpenShift Installer 4.15.8
 level=debug msg=Built from commit f4f5d0ee0f7591fd9ddf03ac337c804608102919
 level=debug msg=Loading Install Config...
 level=debug msg=  Loading SSH Key...
 level=debug msg=  Loading Base Domain...
 level=debug msg=    Loading Platform...
 level=debug msg=  Loading Cluster Name...
 level=debug msg=    Loading Base Domain...
 level=debug msg=    Loading Platform...
 level=debug msg=  Loading Pull Secret...
 level=debug msg=  Loading Platform...
 level=debug msg=Loading Agent Config...
 level=debug msg=Using Agent Config loaded from state file
 level=warning msg=An agent configuration was detected but this command is not the agent wait-for command
 level=info msg=Waiting up to 40m0s (until 10:15AM UTC) for the cluster at https://api.agent-sno.abi-ci.com:6443 to initialize...
 W0416 09:35:51.793770    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:35:51.793827    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:35:53.127917    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:35:53.127946    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:35:54.760896    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:35:54.761058    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:36:00.790136    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:36:00.790175    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:36:08.516333    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:36:08.516445    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:36:31.442291    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:36:31.442336    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:37:03.033971    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:37:03.034049    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:37:42.025487    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:37:42.025538    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:38:32.148607    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:38:32.148677    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:39:27.680156    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:39:27.680194    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:40:23.290839    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:40:23.290988    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:41:22.298200    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:41:22.298338    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:42:01.197417    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:42:01.197465    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:42:36.739577    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:42:36.739937    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:43:07.331029    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:43:07.331154    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:44:04.008310    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:44:04.008381    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:44:40.882938    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:44:40.882973    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:45:18.975189    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:45:18.975307    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:45:49.753584    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:45:49.753614    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:46:41.148207    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:46:41.148347    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:47:12.882965    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:47:12.883075    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:47:53.636491    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:47:53.636538    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:48:31.792077    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:48:31.792165    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:49:29.117579    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:49:29.117657    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:50:02.802033    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:50:02.802167    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:50:33.826705    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:50:33.826859    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:51:16.045403    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:51:16.045447    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:51:53.795710    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:51:53.795745    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:52:52.741141    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:52:52.741289    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:53:52.621642    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:53:52.621687    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:54:35.809906    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:54:35.810054    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:55:24.249298    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:55:24.249418    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:56:12.717328    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:56:12.717372    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:56:51.172375    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:56:51.172439    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:57:42.242226    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:57:42.242292    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:58:17.663810    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:58:17.663849    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 09:59:13.319754    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 09:59:13.319889    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:00:03.188117    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:00:03.188166    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:00:54.590362    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:00:54.590494    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:01:35.673592    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:01:35.673633    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:02:11.552079    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:02:11.552133    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:02:51.110525    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:02:51.110663    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:03:31.251376    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:03:31.251494    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:04:21.566895    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:04:21.566931    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:04:52.754047    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:04:52.754221    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:05:24.673675    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:05:24.673724    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:06:17.608482    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:06:17.608598    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:06:58.215116    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:06:58.215262    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:07:46.578262    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:07:46.578392    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:08:18.239710    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:08:18.239830    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:09:06.947178    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:09:06.947239    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:10:00.261401    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:10:00.261486    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:10:59.363041    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:10:59.363113    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:11:32.205551    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:11:32.205612    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:12:24.956052    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:12:24.956147    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:12:55.353860    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:12:55.354004    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:13:39.223095    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:13:39.223170    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:14:25.018278    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:14:25.018404    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 W0416 10:15:17.227351    1589 reflector.go:535] k8s.io/client-go/tools/watch/informerwatcher.go:146: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 E0416 10:15:17.227424    1589 reflector.go:147] k8s.io/client-go/tools/watch/informerwatcher.go:146: Failed to watch *v1.ClusterVersion: failed to list *v1.ClusterVersion: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusterversions?fieldSelector=metadata.name%3Dversion&limit=500&resourceVersion=0": dial tcp 10.244.64.4:6443: connect: connection refused
 level=error msg=Attempted to gather ClusterOperator status after wait failure: listing ClusterOperator objects: Get "https://api.agent-sno.abi-ci.com:6443/apis/config.openshift.io/v1/clusteroperators": dial tcp 10.244.64.4:6443: connect: connection refused
 level=error msg=Cluster initialization failed because one or more operators are not functioning properly.
 level=error msg=The cluster should be accessible for troubleshooting as detailed in the documentation linked below,
 level=error msg=https://docs.openshift.com/container-platform/latest/support/troubleshooting/troubleshooting-installations.html
 level=error msg=The 'wait-for install-complete' subcommand can then be used to continue the installation
 level=error msg=failed to initialize the cluster: timed out waiting for the condition
 {"component":"entrypoint","error":"wrapped process failed: exit status 6","file":"k8s.io/test-infra/prow/entrypoint/run.go:84","func":"k8s.io/test-infra/prow/entrypoint.Options.internalRun","level":"error","msg":"Error executing test process","severity":"error","time":"2024-04-16T10:15:51Z"}
 error: failed to execute wrapped command: exit status 6 {code}
Status: CLOSED
#OCPBUGS-31763issue10 days agogcp install cluster creation fails after 30-40 minutes New
Issue 15921939: gcp install cluster creation fails after 30-40 minutes
Description: Component Readiness has found a potential regression in install should succeed: overall.  I see this on various different platforms, but I started digging into GCP failures.  No installer log bundle is created, which seriously hinders my ability to dig further.
 
 Bootstrap succeeds, and then 30 minutes after waiting for cluster creation, it dies.
 
 From [https://prow.ci.openshift.org/view/gs/test-platform-results/logs/periodic-ci-openshift-release-master-nightly-4.16-e2e-gcp-sdn-serial/1775871000018161664]
 
 search.ci tells me this affects nearly 10% of jobs on GCP:
 
 [https://search.dptools.openshift.org/?search=Attempted+to+gather+ClusterOperator+status+after+installation+failure%3A+listing+ClusterOperator+objects.*connection+refused&maxAge=168h&context=1&type=bug%2Bissue%2Bjunit&name=.*4.16.*gcp.*&excludeName=&maxMatches=5&maxBytes=20971520&groupBy=job]
 
  
 {code:java}
 time="2024-04-04T13:27:50Z" level=info msg="Waiting up to 40m0s (until 2:07PM UTC) for the cluster at https://api.ci-op-n3pv5pn3-4e5f3.XXXXXXXXXXXXXXXXXXXXXX:6443 to initialize..."
 time="2024-04-04T14:07:50Z" level=error msg="Attempted to gather ClusterOperator status after installation failure: listing ClusterOperator objects: Get \"https://api.ci-op-n3pv5pn3-4e5f3.XXXXXXXXXXXXXXXXXXXXXX:6443/apis/config.openshift.io/v1/clusteroperators\": dial tcp 35.238.130.20:6443: connect: connection refused"
 time="2024-04-04T14:07:50Z" level=error msg="Cluster initialization failed because one or more operators are not functioning properly.\nThe cluster should be accessible for troubleshooting as detailed in the documentation linked below,\nhttps://docs.openshift.com/container-platform/latest/support/troubleshooting/troubleshooting-installations.html\nThe 'wait-for install-complete' subcommand can then be used to continue the installation"
 time="2024-04-04T14:07:50Z" level=error msg="failed to initialize the cluster: timed out waiting for the condition" {code}
  
 
 Probability of significant regression: 99.44%
 
 Sample (being evaluated) Release: 4.16
 Start Time: 2024-03-29T00:00:00Z
 End Time: 2024-04-04T23:59:59Z
 Success Rate: 68.75%
 Successes: 11
 Failures: 5
 Flakes: 0
 
 Base (historical) Release: 4.15
 Start Time: 2024-02-01T00:00:00Z
 End Time: 2024-02-28T23:59:59Z
 Success Rate: 96.30%
 Successes: 52
 Failures: 2
 Flakes: 0
 
 View the test details report at [https://sippy.dptools.openshift.org/sippy-ng/component_readiness/test_details?arch=amd64&arch=amd64&baseEndTime=2024-02-28%2023%3A59%3A59&baseRelease=4.15&baseStartTime=2024-02-01%2000%3A00%3A00&capability=Other&component=Installer%20%2F%20openshift-installer&confidence=95&environment=sdn%20upgrade-micro%20amd64%20gcp%20standard&excludeArches=arm64%2Cheterogeneous%2Cppc64le%2Cs390x&excludeClouds=openstack%2Cibmcloud%2Clibvirt%2Covirt%2Cunknown&excludeVariants=hypershift%2Cosd%2Cmicroshift%2Ctechpreview%2Csingle-node%2Cassisted%2Ccompact&groupBy=cloud%2Carch%2Cnetwork&ignoreDisruption=true&ignoreMissing=false&minFail=3&network=sdn&network=sdn&pity=5&platform=gcp&platform=gcp&sampleEndTime=2024-04-04%2023%3A59%3A59&sampleRelease=4.16&sampleStartTime=2024-03-29%2000%3A00%3A00&testId=cluster%20install%3A0cb1bb27e418491b1ffdacab58c5c8c0&testName=install%20should%20succeed%3A%20overall&upgrade=upgrade-micro&upgrade=upgrade-micro&variant=standard&variant=standard]
Status: New
#OCPBUGS-17183issue2 days ago[BUG] Assisted installer fails to create bond with active backup for single node installation New
Issue 15401516: [BUG] Assisted installer fails to create bond with active backup for single node installation
Description: Description of problem:
 {code:none}
 The assisted installer will always fail to create bond with active backup using nmstate yaml and the errors are : 
 
 ~~~ 
 Jul 26 07:11:47 <hostname> bootkube.sh[8366]: Unable to reach API_URL's https endpoint at https://xx.xx.32.40:6443/version
 Jul 26 07:11:47 <hostname> bootkube.sh[8366]: Checking validity of <hostname> of type API_INT_URL 
 Jul 26 07:11:47 <hostname> bootkube.sh[8366]: Successfully resolved API_INT_URL <hostname> 
 Jul 26 07:11:47 <hostname> bootkube.sh[8366]: Unable to reach API_INT_URL's https endpoint at https://xx.xx.32.40:6443/versionJul 26 07:12:23 <hostname> bootkube.sh[12960]: Still waiting for the Kubernetes API: 
 Get "https://localhost:6443/readyz": dial tcp [::1]:6443: connect: connection refusedJul 26 07:15:15 <hostname> bootkube.sh[15706]: The connection to the server <hostname>:6443 was refused - did you specify the right host or port? 
 Jul 26 07:15:15 <hostname> bootkube.sh[15706]: The connection to the server <hostname>:6443 was refused - did you specify the right host or port? 
  ~~~ 
 
 Where, <hostname> is the actual hostname of the node. 
 
 Adding sosreport and nmstate yaml file here : https://drive.google.com/drive/u/0/folders/19dNzKUPIMmnUls2pT_stuJxr2Dxdi5eb{code}
 Version-Release number of selected component (if applicable):
 {code:none}
 4.12 
 Dell 16g Poweredge R660{code}
 How reproducible:
 {code:none}
 Always at customer side{code}
 Steps to Reproduce:
 {code:none}
 1. Open Assisted installer UI (console.redhat.com -> assisted installer) 
 2. Add the network configs as below for host1  
 
 -----------
 interfaces:
 - name: bond99
   type: bond
   state: up
   ipv4:
     address:
     - ip: xx.xx.32.40
       prefix-length: 24
     enabled: true
   link-aggregation:
     mode: active-backup
     options:
       miimon: '140'
     port:
     - eno12399
     - eno12409
 dns-resolver:
   config:
     search:
     - xxxx
     server:
     - xx.xx.xx.xx
 routes:
   config:
     - destination: 0.0.0.0/0
       metric: 150
       next-hop-address: xx.xx.xx.xx
       next-hop-interface: bond99
       table-id: 254    
 -----------
 
 3. Enter the mac addresses of interfaces in the fields. 
 4. Generate the iso and boot the node. The node will not be able to ping/ssh. This happen everytime and reproducible.
 5. As there was no way to check (due to ssh not working) what is happening on the node, we reset root password and can see that ip address was present on bond, still ping/ssh does not work.
 6. After multiple reboots, customer was able to ssh/ping and provided sosreport and we could see above mentioned error in the journal logs in sosreport.  
  {code}
 Actual results:
 {code:none}
 Fails to install. Seems there is some issue with networking.{code}
 Expected results:
 {code:none}
 Able to proceed with installation without above mentioned issues{code}
 Additional info:
 {code:none}
 - The installation works with round robbin bond mode in 4.12. 
 - Also, the installation works with active-backup 4.10. 
 - Active-backup bond with 4.12 is failing.{code}
Status: New
#OCPBUGS-32091issue4 weeks agoCAPI-Installer leaks processes during unsuccessful installs MODIFIED
ERROR Attempted to gather debug logs after installation failure: failed to create SSH client: ssh: handshake failed: ssh: disconnect, reason 2: Too many authentication failures
ERROR Attempted to gather ClusterOperator status after installation failure: listing ClusterOperator objects: Get "https://api.gpei-0515.qe.devcluster.openshift.com:6443/apis/config.openshift.io/v1/clusteroperators": dial tcp 3.134.9.157:6443: connect: connection refused
ERROR Bootstrap failed to complete: Get "https://api.gpei-0515.qe.devcluster.openshift.com:6443/version": dial tcp 18.222.8.23:6443: connect: connection refused

... 1 lines not shown

periodic-ci-openshift-release-master-ci-4.12-upgrade-from-stable-4.11-e2e-aws-ovn-upgrade (all) - 37 runs, 43% failed, 144% of failures match = 62% impact
#1791784578998865920junit14 hours ago
May 18 12:29:03.167 E clusterversion/version changed Failing to True: ClusterOperatorDegraded: Cluster operator etcd is degraded
May 18 12:29:05.088 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-244-241.ec2.internal node/ip-10-0-244-241.ec2.internal uid/c3150950-8d37-4652-8489-b4d716f52078 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0518 12:28:58.202692       1 cmd.go:216] Using insecure, self-signed certificates\nI0518 12:28:58.221190       1 crypto.go:601] Generating new CA for check-endpoints-signer@1716035338 cert, and key in /tmp/serving-cert-2503323580/serving-signer.crt, /tmp/serving-cert-2503323580/serving-signer.key\nI0518 12:28:59.161825       1 observer_polling.go:159] Starting file observer\nW0518 12:28:59.201493       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-244-241.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0518 12:28:59.201613       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0518 12:28:59.212946       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-2503323580/tls.crt::/tmp/serving-cert-2503323580/tls.key"\nW0518 12:29:03.617218       1 requestheader_controller.go:193] Unable to get configmap/extension-apiserver-authentication in kube-system.  Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA'\nF0518 12:29:03.617324       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: configmaps "extension-apiserver-authentication" is forbidden: User "system:serviceaccount:openshift-kube-apiserver:check-endpoints" cannot get resource "configmaps" in API group "" in the namespace "kube-system"\n
May 18 12:29:07.077 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-244-241.ec2.internal node/ip-10-0-244-241.ec2.internal uid/c3150950-8d37-4652-8489-b4d716f52078 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0518 12:28:58.202692       1 cmd.go:216] Using insecure, self-signed certificates\nI0518 12:28:58.221190       1 crypto.go:601] Generating new CA for check-endpoints-signer@1716035338 cert, and key in /tmp/serving-cert-2503323580/serving-signer.crt, /tmp/serving-cert-2503323580/serving-signer.key\nI0518 12:28:59.161825       1 observer_polling.go:159] Starting file observer\nW0518 12:28:59.201493       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-244-241.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0518 12:28:59.201613       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0518 12:28:59.212946       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-2503323580/tls.crt::/tmp/serving-cert-2503323580/tls.key"\nW0518 12:29:03.617218       1 requestheader_controller.go:193] Unable to get configmap/extension-apiserver-authentication in kube-system.  Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA'\nF0518 12:29:03.617324       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: configmaps "extension-apiserver-authentication" is forbidden: User "system:serviceaccount:openshift-kube-apiserver:check-endpoints" cannot get resource "configmaps" in API group "" in the namespace "kube-system"\n

... 1 lines not shown

#1791738720953569280junit17 hours ago
May 18 09:30:53.714 E ns/openshift-ovn-kubernetes pod/ovnkube-node-znrzg node/ip-10-0-206-29.us-west-1.compute.internal uid/5079cdec-a2d8-4cca-88fc-de7269ef8a0f container/ovn-acl-logging reason/ContainerExit code/1 cause/Error cat: /run/ovn/ovn-controller.pid: No such file or directory\n
May 18 09:30:56.727 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-206-29.us-west-1.compute.internal node/ip-10-0-206-29.us-west-1.compute.internal uid/11396771-16dd-402a-9174-5ab54523fd9a container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0518 09:30:53.596813       1 cmd.go:216] Using insecure, self-signed certificates\nI0518 09:30:53.611535       1 crypto.go:601] Generating new CA for check-endpoints-signer@1716024653 cert, and key in /tmp/serving-cert-615285488/serving-signer.crt, /tmp/serving-cert-615285488/serving-signer.key\nI0518 09:30:54.018259       1 observer_polling.go:159] Starting file observer\nW0518 09:30:54.038037       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-206-29.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0518 09:30:54.038240       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0518 09:30:54.048169       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-615285488/tls.crt::/tmp/serving-cert-615285488/tls.key"\nF0518 09:30:54.324542       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 18 09:30:57.063 E clusterversion/version changed Failing to True: ClusterOperatorDegraded: Cluster operator etcd is degraded

... 2 lines not shown

#1791530883291811840junit31 hours ago
May 17 19:43:30.228 E ns/openshift-multus pod/network-metrics-daemon-gzvkk node/ip-10-0-255-143.us-west-2.compute.internal uid/3b0c273b-6b46-474c-bd4d-38fe37c8f328 container/network-metrics-daemon reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 17 19:43:31.362 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-255-143.us-west-2.compute.internal node/ip-10-0-255-143.us-west-2.compute.internal uid/95182263-2cf7-410f-a425-421cef559a6d container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0517 19:43:24.987717       1 cmd.go:216] Using insecure, self-signed certificates\nI0517 19:43:24.998022       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715975004 cert, and key in /tmp/serving-cert-4210011396/serving-signer.crt, /tmp/serving-cert-4210011396/serving-signer.key\nI0517 19:43:25.328090       1 observer_polling.go:159] Starting file observer\nW0517 19:43:25.341142       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-255-143.us-west-2.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0517 19:43:25.341302       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0517 19:43:25.352674       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-4210011396/tls.crt::/tmp/serving-cert-4210011396/tls.key"\nW0517 19:43:29.940122       1 requestheader_controller.go:193] Unable to get configmap/extension-apiserver-authentication in kube-system.  Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA'\nF0517 19:43:29.941674       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: configmaps "extension-apiserver-authentication" is forbidden: User "system:serviceaccount:openshift-kube-apiserver:check-endpoints" cannot get resource "configmaps" in API group "" in the namespace "kube-system"\n
May 17 19:43:32.430 E ns/openshift-dns pod/dns-default-96nc9 node/ip-10-0-255-143.us-west-2.compute.internal uid/5dabea60-aaf7-421b-86ed-cb281ec056d9 container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
#1791530883291811840junit31 hours ago
May 17 19:43:32.518 E ns/openshift-multus pod/multus-additional-cni-plugins-gwnbn node/ip-10-0-255-143.us-west-2.compute.internal uid/3c5351b9-6c30-4c58-b61c-6059ccf91aa6 container/kube-multus-additional-cni-plugins reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 17 19:43:33.535 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-255-143.us-west-2.compute.internal node/ip-10-0-255-143.us-west-2.compute.internal uid/95182263-2cf7-410f-a425-421cef559a6d container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0517 19:43:24.987717       1 cmd.go:216] Using insecure, self-signed certificates\nI0517 19:43:24.998022       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715975004 cert, and key in /tmp/serving-cert-4210011396/serving-signer.crt, /tmp/serving-cert-4210011396/serving-signer.key\nI0517 19:43:25.328090       1 observer_polling.go:159] Starting file observer\nW0517 19:43:25.341142       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-255-143.us-west-2.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0517 19:43:25.341302       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0517 19:43:25.352674       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-4210011396/tls.crt::/tmp/serving-cert-4210011396/tls.key"\nW0517 19:43:29.940122       1 requestheader_controller.go:193] Unable to get configmap/extension-apiserver-authentication in kube-system.  Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA'\nF0517 19:43:29.941674       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: configmaps "extension-apiserver-authentication" is forbidden: User "system:serviceaccount:openshift-kube-apiserver:check-endpoints" cannot get resource "configmaps" in API group "" in the namespace "kube-system"\n
May 17 19:43:34.546 E ns/openshift-e2e-loki pod/loki-promtail-fc8pj node/ip-10-0-255-143.us-west-2.compute.internal uid/c39c5484-25b9-4730-acbb-0634ef82cfa8 container/prod-bearer-token reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
#1790912486363369472junit2 days ago
May 16 02:48:30.261 E ns/openshift-ovn-kubernetes pod/ovnkube-master-2bdtq node/ip-10-0-142-120.us-west-1.compute.internal uid/90522fb8-c3c6-4ecf-baab-a2dd20fc505a container/nbdb reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 16 02:48:31.259 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-142-120.us-west-1.compute.internal node/ip-10-0-142-120.us-west-1.compute.internal uid/49ddb3f8-35fc-4eef-9236-693f0b40db76 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0516 02:48:29.775486       1 cmd.go:216] Using insecure, self-signed certificates\nI0516 02:48:29.789659       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715827709 cert, and key in /tmp/serving-cert-2973121713/serving-signer.crt, /tmp/serving-cert-2973121713/serving-signer.key\nI0516 02:48:30.423912       1 observer_polling.go:159] Starting file observer\nW0516 02:48:30.432737       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-142-120.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0516 02:48:30.433018       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0516 02:48:30.441170       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-2973121713/tls.crt::/tmp/serving-cert-2973121713/tls.key"\nF0516 02:48:30.762803       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 16 02:48:35.006 E ns/e2e-k8s-sig-apps-daemonset-upgrade-1265 pod/ds1-n6hgf node/ip-10-0-142-120.us-west-1.compute.internal uid/0bc82e6b-d63b-4535-9403-861a0f06dabb container/app reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)

... 2 lines not shown

#1790547465145946112junit4 days ago
May 15 02:09:27.194 E ns/openshift-cluster-csi-drivers pod/aws-ebs-csi-driver-node-npjs6 node/ip-10-0-145-42.ec2.internal uid/9f0745fc-cebd-444c-8612-585b216cc70a container/csi-liveness-probe reason/ContainerExit code/2 cause/Error
May 15 02:12:53.647 E ns/openshift-cloud-network-config-controller pod/cloud-network-config-controller-c77644bf-l6tzj node/ip-10-0-244-40.ec2.internal uid/6b2e18dd-7938-4091-b886-a28b52448dfd container/controller reason/ContainerExit code/1 cause/Error from the node workqueue\nI0515 01:23:52.005942       1 controller.go:182] Assigning key: ip-10-0-175-73.ec2.internal to node workqueue\nI0515 01:23:52.006007       1 controller.go:160] Dropping key 'ip-10-0-175-73.ec2.internal' from the node workqueue\nI0515 01:23:52.414596       1 controller.go:182] Assigning key: ip-10-0-175-73.ec2.internal to node workqueue\nI0515 01:23:52.415177       1 controller.go:160] Dropping key 'ip-10-0-175-73.ec2.internal' from the node workqueue\nI0515 01:24:30.522926       1 controller.go:182] Assigning key: ip-10-0-175-73.ec2.internal to node workqueue\nI0515 01:24:30.522957       1 controller.go:160] Dropping key 'ip-10-0-175-73.ec2.internal' from the node workqueue\nI0515 01:24:59.598125       1 controller.go:182] Assigning key: ip-10-0-175-73.ec2.internal to node workqueue\nI0515 01:24:59.598182       1 controller.go:160] Dropping key 'ip-10-0-175-73.ec2.internal' from the node workqueue\nI0515 01:25:01.970448       1 controller.go:182] Assigning key: ip-10-0-175-73.ec2.internal to node workqueue\nI0515 01:25:01.970569       1 controller.go:160] Dropping key 'ip-10-0-175-73.ec2.internal' from the node workqueue\nE0515 01:26:10.059351       1 leaderelection.go:330] error retrieving resource lock openshift-cloud-network-config-controller/cloud-network-config-controller-lock: Get "https://api-int.ci-op-q5h2qyq0-d5c74.aws-2.ci.openshift.org:6443/api/v1/namespaces/openshift-cloud-network-config-controller/configmaps/cloud-network-config-controller-lock": dial tcp 10.0.205.40:6443: connect: connection refused\nE0515 01:26:19.165041       1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Secret: unknown (get secrets)\nI0515 02:12:51.725497       1 controller.go:104] Shutting down node workers\nI0515 02:12:51.725503       1 controller.go:104] Shutting down secret workers\nI0515 02:12:51.725731       1 controller.go:104] Shutting down cloud-private-ip-config workers\nI0515 02:12:51.732326       1 main.go:161] Stopped leading, sending SIGTERM and shutting down controller\n
May 15 02:12:58.682 E ns/openshift-multus pod/multus-additional-cni-plugins-8v926 node/ip-10-0-244-40.ec2.internal uid/3e5f030a-c032-4e48-93d2-9fce618553fe container/kube-multus-additional-cni-plugins reason/ContainerExit code/143 cause/Error
#1790547465145946112junit4 days ago
May 15 02:37:01.631 E ns/openshift-dns pod/dns-default-tvbbs node/ip-10-0-244-40.ec2.internal uid/b91a1bc9-4ad0-4b12-a347-a94102de11df container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 15 02:37:03.724 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-244-40.ec2.internal node/ip-10-0-244-40.ec2.internal uid/0c8887ec-5250-45b3-831b-c68b483fa1bb container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0515 02:37:02.020661       1 cmd.go:216] Using insecure, self-signed certificates\nI0515 02:37:02.023871       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715740622 cert, and key in /tmp/serving-cert-1562679283/serving-signer.crt, /tmp/serving-cert-1562679283/serving-signer.key\nI0515 02:37:02.388079       1 observer_polling.go:159] Starting file observer\nW0515 02:37:02.402151       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-244-40.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0515 02:37:02.402271       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0515 02:37:02.402760       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-1562679283/tls.crt::/tmp/serving-cert-1562679283/tls.key"\nF0515 02:37:02.710364       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 15 02:37:04.754 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-244-40.ec2.internal node/ip-10-0-244-40.ec2.internal uid/0c8887ec-5250-45b3-831b-c68b483fa1bb container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0515 02:37:02.020661       1 cmd.go:216] Using insecure, self-signed certificates\nI0515 02:37:02.023871       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715740622 cert, and key in /tmp/serving-cert-1562679283/serving-signer.crt, /tmp/serving-cert-1562679283/serving-signer.key\nI0515 02:37:02.388079       1 observer_polling.go:159] Starting file observer\nW0515 02:37:02.402151       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-244-40.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0515 02:37:02.402271       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0515 02:37:02.402760       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-1562679283/tls.crt::/tmp/serving-cert-1562679283/tls.key"\nF0515 02:37:02.710364       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n

... 1 lines not shown

#1790772188496793600junit3 days ago
May 15 17:11:46.184 E ns/openshift-network-operator pod/network-operator-5c88dc4978-bdnhg node/ip-10-0-128-167.ec2.internal uid/c7797f48-124e-4f69-8425-9fc9fecb854a container/network-operator reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 15 17:11:56.121 E ns/openshift-cloud-network-config-controller pod/cloud-network-config-controller-589c78bcd9-89rtp node/ip-10-0-240-6.ec2.internal uid/80ec5145-0861-4f5c-b6e9-94441198cbce container/controller reason/ContainerExit code/1 cause/Error on node: ip-10-0-137-108.ec2.internal\nI0515 16:19:38.349353       1 controller.go:160] Dropping key 'ip-10-0-137-108.ec2.internal' from the node workqueue\nI0515 16:19:41.915120       1 controller.go:182] Assigning key: ip-10-0-137-108.ec2.internal to node workqueue\nI0515 16:19:41.915162       1 controller.go:160] Dropping key 'ip-10-0-137-108.ec2.internal' from the node workqueue\nI0515 16:19:42.708741       1 controller.go:182] Assigning key: ip-10-0-137-108.ec2.internal to node workqueue\nI0515 16:19:42.708944       1 controller.go:160] Dropping key 'ip-10-0-137-108.ec2.internal' from the node workqueue\nI0515 16:19:47.747173       1 controller.go:182] Assigning key: ip-10-0-137-108.ec2.internal to node workqueue\nI0515 16:19:47.747289       1 controller.go:160] Dropping key 'ip-10-0-137-108.ec2.internal' from the node workqueue\nI0515 16:20:18.438349       1 controller.go:182] Assigning key: ip-10-0-137-108.ec2.internal to node workqueue\nI0515 16:20:18.438471       1 controller.go:160] Dropping key 'ip-10-0-137-108.ec2.internal' from the node workqueue\nI0515 16:20:21.884952       1 controller.go:182] Assigning key: ip-10-0-137-108.ec2.internal to node workqueue\nI0515 16:20:21.884984       1 controller.go:160] Dropping key 'ip-10-0-137-108.ec2.internal' from the node workqueue\nE0515 16:22:38.096146       1 leaderelection.go:330] error retrieving resource lock openshift-cloud-network-config-controller/cloud-network-config-controller-lock: Get "https://api-int.ci-op-c0jcyyn5-d5c74.aws-2.ci.openshift.org:6443/api/v1/namespaces/openshift-cloud-network-config-controller/configmaps/cloud-network-config-controller-lock": dial tcp 10.0.181.51:6443: connect: connection refused\nI0515 17:11:54.602907       1 controller.go:104] Shutting down secret workers\nI0515 17:11:54.603131       1 controller.go:104] Shutting down cloud-private-ip-config workers\nI0515 17:11:54.603184       1 controller.go:104] Shutting down node workers\nI0515 17:11:54.628275       1 main.go:161] Stopped leading, sending SIGTERM and shutting down controller\n
May 15 17:12:01.881 E ns/openshift-multus pod/multus-additional-cni-plugins-6hl4r node/ip-10-0-137-108.ec2.internal uid/a18a1f5e-a3b0-4cf0-9fbb-9760dea33266 container/kube-multus-additional-cni-plugins reason/ContainerExit code/143 cause/Error
#1790772188496793600junit3 days ago
May 15 17:42:21.595 E ns/openshift-ovn-kubernetes pod/ovnkube-master-4qf2m node/ip-10-0-128-167.ec2.internal uid/2998f594-be4e-4b2c-8bde-a1f2e0a7cd60 container/ovnkube-master reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 15 17:42:24.469 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-128-167.ec2.internal node/ip-10-0-128-167.ec2.internal uid/1ec7a433-a5d6-478d-bada-e3650c025ab6 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0515 17:42:21.383221       1 cmd.go:216] Using insecure, self-signed certificates\nI0515 17:42:21.392546       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715794941 cert, and key in /tmp/serving-cert-4028241920/serving-signer.crt, /tmp/serving-cert-4028241920/serving-signer.key\nI0515 17:42:21.796096       1 observer_polling.go:159] Starting file observer\nW0515 17:42:21.822751       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-128-167.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0515 17:42:21.823274       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0515 17:42:21.837743       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-4028241920/tls.crt::/tmp/serving-cert-4028241920/tls.key"\nF0515 17:42:22.489856       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 15 17:42:25.685 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-128-167.ec2.internal node/ip-10-0-128-167.ec2.internal uid/1ec7a433-a5d6-478d-bada-e3650c025ab6 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0515 17:42:21.383221       1 cmd.go:216] Using insecure, self-signed certificates\nI0515 17:42:21.392546       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715794941 cert, and key in /tmp/serving-cert-4028241920/serving-signer.crt, /tmp/serving-cert-4028241920/serving-signer.key\nI0515 17:42:21.796096       1 observer_polling.go:159] Starting file observer\nW0515 17:42:21.822751       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-128-167.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0515 17:42:21.823274       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0515 17:42:21.837743       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-4028241920/tls.crt::/tmp/serving-cert-4028241920/tls.key"\nF0515 17:42:22.489856       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n

... 1 lines not shown

#1790499882113437696junit4 days ago
May 14 23:43:30.292 E ns/openshift-image-registry pod/node-ca-n8tx5 node/ip-10-0-146-213.us-west-2.compute.internal uid/773f7d19-3cab-4ed1-b62c-00dea87d41a9 container/node-ca reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 14 23:43:32.291 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-146-213.us-west-2.compute.internal node/ip-10-0-146-213.us-west-2.compute.internal uid/8551393d-cf00-4213-92e5-62e70c4014ee container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0514 23:43:31.225546       1 cmd.go:216] Using insecure, self-signed certificates\nI0514 23:43:31.243823       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715730211 cert, and key in /tmp/serving-cert-890527349/serving-signer.crt, /tmp/serving-cert-890527349/serving-signer.key\nI0514 23:43:31.672211       1 observer_polling.go:159] Starting file observer\nW0514 23:43:31.686791       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-146-213.us-west-2.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0514 23:43:31.686949       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0514 23:43:31.700316       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-890527349/tls.crt::/tmp/serving-cert-890527349/tls.key"\nF0514 23:43:32.046645       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 14 23:43:35.956 E ns/e2e-k8s-sig-apps-daemonset-upgrade-2472 pod/ds1-dwnnz node/ip-10-0-146-213.us-west-2.compute.internal uid/d6cc3def-b5e1-4c4d-8757-b636597b709c container/app reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)

... 2 lines not shown

#1790454847397433344junit4 days ago
May 14 20:36:37.190 E ns/openshift-ovn-kubernetes pod/ovnkube-master-k2qzn node/ip-10-0-151-84.ec2.internal uid/0688a1e8-fb6c-4df5-8de4-05faae61608f container/ovnkube-master reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 14 20:36:38.226 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-151-84.ec2.internal node/ip-10-0-151-84.ec2.internal uid/caa7e0d6-e98e-498d-a38c-db4de2900a6a container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0514 20:36:36.400141       1 cmd.go:216] Using insecure, self-signed certificates\nI0514 20:36:36.413675       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715718996 cert, and key in /tmp/serving-cert-2703579533/serving-signer.crt, /tmp/serving-cert-2703579533/serving-signer.key\nI0514 20:36:37.072778       1 observer_polling.go:159] Starting file observer\nW0514 20:36:37.094364       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-151-84.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0514 20:36:37.094524       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0514 20:36:37.095165       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-2703579533/tls.crt::/tmp/serving-cert-2703579533/tls.key"\nF0514 20:36:37.356151       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 14 20:36:39.232 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-151-84.ec2.internal node/ip-10-0-151-84.ec2.internal uid/caa7e0d6-e98e-498d-a38c-db4de2900a6a container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0514 20:36:36.400141       1 cmd.go:216] Using insecure, self-signed certificates\nI0514 20:36:36.413675       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715718996 cert, and key in /tmp/serving-cert-2703579533/serving-signer.crt, /tmp/serving-cert-2703579533/serving-signer.key\nI0514 20:36:37.072778       1 observer_polling.go:159] Starting file observer\nW0514 20:36:37.094364       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-151-84.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0514 20:36:37.094524       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0514 20:36:37.095165       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-2703579533/tls.crt::/tmp/serving-cert-2703579533/tls.key"\nF0514 20:36:37.356151       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n

... 1 lines not shown

#1789710107358007296junit6 days ago
May 12 19:14:58.429 E ns/openshift-ovn-kubernetes pod/ovnkube-master-8td7s node/ip-10-0-153-81.us-west-1.compute.internal uid/40105d52-dfa0-44ed-acb8-6a70e393bf67 container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 12 19:14:59.451 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-153-81.us-west-1.compute.internal node/ip-10-0-153-81.us-west-1.compute.internal uid/1cefda2c-650d-4db2-a574-85f172a6bc9e container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0512 19:14:56.747873       1 cmd.go:216] Using insecure, self-signed certificates\nI0512 19:14:56.765349       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715541296 cert, and key in /tmp/serving-cert-962862230/serving-signer.crt, /tmp/serving-cert-962862230/serving-signer.key\nI0512 19:14:57.099239       1 observer_polling.go:159] Starting file observer\nW0512 19:14:57.125489       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-153-81.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0512 19:14:57.125618       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0512 19:14:57.126156       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-962862230/tls.crt::/tmp/serving-cert-962862230/tls.key"\nF0512 19:14:57.558942       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 12 19:15:01.646 E ns/openshift-network-diagnostics pod/network-check-target-pb9z7 node/ip-10-0-153-81.us-west-1.compute.internal uid/2092e520-4f45-4e9c-816a-1c49a43ba4ea container/network-check-target-container reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)

... 3 lines not shown

#1790138933279985664junit5 days ago
May 13 23:40:51.028 E ns/openshift-dns pod/dns-default-nh96w node/ip-10-0-237-107.ec2.internal uid/e071ad5d-a565-4e7f-a820-cbcd8cfae865 container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 13 23:40:51.133 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-237-107.ec2.internal node/ip-10-0-237-107.ec2.internal uid/cab6867e-78f8-44b1-abd0-340eeab8df18 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0513 23:40:49.654466       1 cmd.go:216] Using insecure, self-signed certificates\nI0513 23:40:49.663094       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715643649 cert, and key in /tmp/serving-cert-704702517/serving-signer.crt, /tmp/serving-cert-704702517/serving-signer.key\nI0513 23:40:50.036179       1 observer_polling.go:159] Starting file observer\nW0513 23:40:50.043602       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-237-107.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0513 23:40:50.043751       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0513 23:40:50.045358       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-704702517/tls.crt::/tmp/serving-cert-704702517/tls.key"\nF0513 23:40:50.675124       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 13 23:40:52.070 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-237-107.ec2.internal node/ip-10-0-237-107.ec2.internal uid/cab6867e-78f8-44b1-abd0-340eeab8df18 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0513 23:40:49.654466       1 cmd.go:216] Using insecure, self-signed certificates\nI0513 23:40:49.663094       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715643649 cert, and key in /tmp/serving-cert-704702517/serving-signer.crt, /tmp/serving-cert-704702517/serving-signer.key\nI0513 23:40:50.036179       1 observer_polling.go:159] Starting file observer\nW0513 23:40:50.043602       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-237-107.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0513 23:40:50.043751       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0513 23:40:50.045358       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-704702517/tls.crt::/tmp/serving-cert-704702517/tls.key"\nF0513 23:40:50.675124       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n

... 1 lines not shown

#1789946789093183488junit5 days ago
May 13 10:28:56.240 E ns/openshift-network-operator pod/network-operator-5c88dc4978-mvlxt node/ip-10-0-176-146.ec2.internal uid/44e0e3ea-b1c2-4029-9a41-b73b2004b77e container/network-operator reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 13 10:29:05.222 E ns/openshift-cloud-network-config-controller pod/cloud-network-config-controller-577d797d8f-bxf5n node/ip-10-0-176-146.ec2.internal uid/fc05de46-b07d-466c-a91b-789b0bebf4f6 container/controller reason/ContainerExit code/1 cause/Error 33465       1 controller.go:182] Assigning key: ip-10-0-180-123.ec2.internal to node workqueue\nI0513 09:32:09.934306       1 controller.go:160] Dropping key 'ip-10-0-180-123.ec2.internal' from the node workqueue\nI0513 09:32:11.895396       1 controller.go:182] Assigning key: ip-10-0-180-123.ec2.internal to node workqueue\nI0513 09:32:11.895486       1 controller.go:160] Dropping key 'ip-10-0-180-123.ec2.internal' from the node workqueue\nI0513 09:32:30.241134       1 controller.go:182] Assigning key: ip-10-0-173-189.ec2.internal to node workqueue\nI0513 09:32:30.241163       1 controller.go:160] Dropping key 'ip-10-0-173-189.ec2.internal' from the node workqueue\nI0513 09:32:31.913502       1 controller.go:182] Assigning key: ip-10-0-173-189.ec2.internal to node workqueue\nI0513 09:32:31.913538       1 controller.go:160] Dropping key 'ip-10-0-173-189.ec2.internal' from the node workqueue\nE0513 09:35:02.286105       1 leaderelection.go:330] error retrieving resource lock openshift-cloud-network-config-controller/cloud-network-config-controller-lock: Get "https://api-int.ci-op-55nn53cn-d5c74.aws-2.ci.openshift.org:6443/api/v1/namespaces/openshift-cloud-network-config-controller/configmaps/cloud-network-config-controller-lock": dial tcp 10.0.241.50:6443: connect: connection refused\nE0513 09:43:42.821587       1 leaderelection.go:330] error retrieving resource lock openshift-cloud-network-config-controller/cloud-network-config-controller-lock: Get "https://api-int.ci-op-55nn53cn-d5c74.aws-2.ci.openshift.org:6443/api/v1/namespaces/openshift-cloud-network-config-controller/configmaps/cloud-network-config-controller-lock": dial tcp 10.0.144.55:6443: connect: connection refused\nI0513 10:29:04.293896       1 controller.go:104] Shutting down node workers\nI0513 10:29:04.293901       1 controller.go:104] Shutting down secret workers\nI0513 10:29:04.293926       1 controller.go:104] Shutting down cloud-private-ip-config workers\nI0513 10:29:04.300402       1 main.go:161] Stopped leading, sending SIGTERM and shutting down controller\n
May 13 10:29:10.935 E ns/openshift-multus pod/multus-additional-cni-plugins-w4cgr node/ip-10-0-180-123.ec2.internal uid/57eb6acf-c06f-4754-9951-9b25e8cc1443 container/kube-multus-additional-cni-plugins reason/ContainerExit code/143 cause/Error
#1789946789093183488junit5 days ago
May 13 11:00:26.566 E ns/e2e-k8s-sig-apps-daemonset-upgrade-808 pod/ds1-n6fsj node/ip-10-0-254-17.ec2.internal uid/46bc44b6-bb79-4239-8404-8903d26402b1 container/app reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 13 11:00:27.397 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-139-2.ec2.internal node/ip-10-0-139-2.ec2.internal uid/30195683-a9a8-402a-a803-339a332e1d71 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0513 11:00:21.442542       1 cmd.go:216] Using insecure, self-signed certificates\nI0513 11:00:21.453519       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715598021 cert, and key in /tmp/serving-cert-517355983/serving-signer.crt, /tmp/serving-cert-517355983/serving-signer.key\nI0513 11:00:21.961954       1 observer_polling.go:159] Starting file observer\nW0513 11:00:21.971818       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-139-2.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0513 11:00:21.971954       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0513 11:00:21.972594       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-517355983/tls.crt::/tmp/serving-cert-517355983/tls.key"\nW0513 11:00:26.242806       1 requestheader_controller.go:193] Unable to get configmap/extension-apiserver-authentication in kube-system.  Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA'\nF0513 11:00:26.242860       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: configmaps "extension-apiserver-authentication" is forbidden: User "system:serviceaccount:openshift-kube-apiserver:check-endpoints" cannot get resource "configmaps" in API group "" in the namespace "kube-system"\n
May 13 11:00:27.493 E ns/openshift-ovn-kubernetes pod/ovnkube-master-gz4hz node/ip-10-0-139-2.ec2.internal uid/9583cc89-42bf-4e6b-b910-42faf730fa02 container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
#1789358527291068416junit7 days ago
May 11 20:03:42.750 E ns/openshift-multus pod/network-metrics-daemon-tt4tk node/ip-10-0-165-87.us-east-2.compute.internal uid/de48d8e3-a0dd-42dd-b0de-3c2e6a8137d4 container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 11 20:03:42.770 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-165-87.us-east-2.compute.internal node/ip-10-0-165-87.us-east-2.compute.internal uid/8844bc39-a077-44ab-9598-65b64bf5b40e container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0511 20:03:37.028797       1 cmd.go:216] Using insecure, self-signed certificates\nI0511 20:03:37.045320       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715457817 cert, and key in /tmp/serving-cert-1021118089/serving-signer.crt, /tmp/serving-cert-1021118089/serving-signer.key\nI0511 20:03:37.427015       1 observer_polling.go:159] Starting file observer\nW0511 20:03:37.445434       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-165-87.us-east-2.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0511 20:03:37.462333       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0511 20:03:37.468127       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-1021118089/tls.crt::/tmp/serving-cert-1021118089/tls.key"\nW0511 20:03:42.356666       1 requestheader_controller.go:193] Unable to get configmap/extension-apiserver-authentication in kube-system.  Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA'\nF0511 20:03:42.356910       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: configmaps "extension-apiserver-authentication" is forbidden: User "system:serviceaccount:openshift-kube-apiserver:check-endpoints" cannot get resource "configmaps" in API group "" in the namespace "kube-system"\n
May 11 20:03:42.796 E ns/openshift-dns pod/dns-default-q24wf node/ip-10-0-165-87.us-east-2.compute.internal uid/ff84d078-29bb-4ebf-899e-6da66c587de5 container/dns reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)

... 3 lines not shown

#1789178756108128256junit7 days ago
May 11 08:04:13.908 E ns/openshift-e2e-loki pod/loki-promtail-v25xl node/ip-10-0-147-139.us-east-2.compute.internal uid/a0373167-81a1-46f0-8a0b-9d35ffe920dc container/prod-bearer-token reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 11 08:04:14.770 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-159-224.us-east-2.compute.internal node/ip-10-0-159-224.us-east-2.compute.internal uid/f03d64ca-ed5a-4102-a207-0ef0dd253919 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0511 08:04:13.788045       1 cmd.go:216] Using insecure, self-signed certificates\nI0511 08:04:13.797198       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715414653 cert, and key in /tmp/serving-cert-984293988/serving-signer.crt, /tmp/serving-cert-984293988/serving-signer.key\nI0511 08:04:14.133674       1 observer_polling.go:159] Starting file observer\nW0511 08:04:14.144507       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-159-224.us-east-2.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0511 08:04:14.144634       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0511 08:04:14.161537       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-984293988/tls.crt::/tmp/serving-cert-984293988/tls.key"\nF0511 08:04:14.507970       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 11 08:04:18.352 E ns/openshift-network-diagnostics pod/network-check-target-btxmk node/ip-10-0-159-224.us-east-2.compute.internal uid/06573b44-c462-4ede-bd74-2d2a5556a92c container/network-check-target-container reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)

... 2 lines not shown

#1789085949343305728junit8 days ago
May 11 01:52:56.023 E ns/openshift-ovn-kubernetes pod/ovnkube-master-gbnbj node/ip-10-0-144-219.ec2.internal uid/1e532d61-a050-470e-9273-ab32784a0886 container/sbdb reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 11 01:52:58.037 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-144-219.ec2.internal node/ip-10-0-144-219.ec2.internal uid/fabe8d3f-683c-46fe-b24a-1e01601e4084 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0511 01:52:56.014155       1 cmd.go:216] Using insecure, self-signed certificates\nI0511 01:52:56.018796       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715392376 cert, and key in /tmp/serving-cert-2562785010/serving-signer.crt, /tmp/serving-cert-2562785010/serving-signer.key\nI0511 01:52:56.423236       1 observer_polling.go:159] Starting file observer\nW0511 01:52:56.444766       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-144-219.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0511 01:52:56.444984       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0511 01:52:56.455968       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-2562785010/tls.crt::/tmp/serving-cert-2562785010/tls.key"\nF0511 01:52:56.866015       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 11 01:53:02.006 E ns/e2e-k8s-sig-apps-daemonset-upgrade-4558 pod/ds1-f7477 node/ip-10-0-144-219.ec2.internal uid/b69df9dc-3301-44b6-82a0-5233369b814d container/app reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)

... 3 lines not shown

#1788924084432670720junit8 days ago
May 10 15:15:33.537 E ns/openshift-cluster-csi-drivers pod/aws-ebs-csi-driver-node-5fg86 node/ip-10-0-164-192.us-west-1.compute.internal uid/61155955-b251-4323-854a-55bf6a801f1a container/csi-node-driver-registrar reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 10 15:15:37.584 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-164-192.us-west-1.compute.internal node/ip-10-0-164-192.us-west-1.compute.internal uid/1438f7a7-aa9c-44e1-810e-d4738b9218d1 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0510 15:15:35.709955       1 cmd.go:216] Using insecure, self-signed certificates\nI0510 15:15:35.721134       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715354135 cert, and key in /tmp/serving-cert-3845407413/serving-signer.crt, /tmp/serving-cert-3845407413/serving-signer.key\nI0510 15:15:36.099349       1 observer_polling.go:159] Starting file observer\nW0510 15:15:36.114063       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-164-192.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0510 15:15:36.114169       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0510 15:15:36.120173       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-3845407413/tls.crt::/tmp/serving-cert-3845407413/tls.key"\nF0510 15:15:36.515988       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 10 15:15:39.591 E ns/openshift-ovn-kubernetes pod/ovnkube-master-sxr46 node/ip-10-0-164-192.us-west-1.compute.internal uid/c26f9845-6064-4dff-8673-77fa98fe6633 container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
#1788924084432670720junit8 days ago
May 10 15:15:39.591 E ns/openshift-ovn-kubernetes pod/ovnkube-master-sxr46 node/ip-10-0-164-192.us-west-1.compute.internal uid/c26f9845-6064-4dff-8673-77fa98fe6633 container/ovn-dbchecker reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 10 15:15:39.614 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-164-192.us-west-1.compute.internal node/ip-10-0-164-192.us-west-1.compute.internal uid/1438f7a7-aa9c-44e1-810e-d4738b9218d1 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0510 15:15:35.709955       1 cmd.go:216] Using insecure, self-signed certificates\nI0510 15:15:35.721134       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715354135 cert, and key in /tmp/serving-cert-3845407413/serving-signer.crt, /tmp/serving-cert-3845407413/serving-signer.key\nI0510 15:15:36.099349       1 observer_polling.go:159] Starting file observer\nW0510 15:15:36.114063       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-164-192.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0510 15:15:36.114169       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0510 15:15:36.120173       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-3845407413/tls.crt::/tmp/serving-cert-3845407413/tls.key"\nF0510 15:15:36.515988       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 10 15:15:42.601 E ns/openshift-multus pod/multus-additional-cni-plugins-pqvcd node/ip-10-0-164-192.us-west-1.compute.internal uid/8f77b887-2b3a-4ad2-9beb-7a14bc18a3bc container/kube-multus-additional-cni-plugins reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
#1788744934191271936junit8 days ago
May 10 03:19:43.389 E ns/openshift-ovn-kubernetes pod/ovnkube-master-7g9bs node/ip-10-0-138-42.ec2.internal uid/a66fb4cb-868b-4f92-872f-d0caf2c1cde1 container/nbdb reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 10 03:19:44.409 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-138-42.ec2.internal node/ip-10-0-138-42.ec2.internal uid/a009675e-4ce8-4bd9-a38a-37652da8315e container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0510 03:19:43.267161       1 cmd.go:216] Using insecure, self-signed certificates\nI0510 03:19:43.267968       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715311183 cert, and key in /tmp/serving-cert-218490898/serving-signer.crt, /tmp/serving-cert-218490898/serving-signer.key\nI0510 03:19:43.602846       1 observer_polling.go:159] Starting file observer\nW0510 03:19:43.610963       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-138-42.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0510 03:19:43.611132       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0510 03:19:43.618292       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-218490898/tls.crt::/tmp/serving-cert-218490898/tls.key"\nF0510 03:19:43.956990       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 10 03:19:45.390 E ns/openshift-dns pod/node-resolver-tfwbx node/ip-10-0-138-42.ec2.internal uid/aa3e1387-4e94-42d4-adc0-d20e7a5757c3 container/dns-node-resolver reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)

... 3 lines not shown

#1788836529343303680junit8 days ago
May 10 08:47:30.707 E ns/openshift-cluster-csi-drivers pod/aws-ebs-csi-driver-node-g8rcv node/ip-10-0-135-143.us-east-2.compute.internal uid/11a03507-d788-4f51-823e-6f941ff28781 container/csi-driver reason/ContainerExit code/2 cause/Error
May 10 08:50:07.031 E ns/openshift-cloud-network-config-controller pod/cloud-network-config-controller-64d5b667d9-ltcsz node/ip-10-0-205-95.us-east-2.compute.internal uid/bedb6bfb-4eba-447e-8787-2c852904c874 container/controller reason/ContainerExit code/1 cause/Error  node workqueue\nI0510 08:03:48.373207       1 controller.go:182] Assigning key: ip-10-0-135-143.us-east-2.compute.internal to node workqueue\nI0510 08:03:48.373236       1 controller.go:160] Dropping key 'ip-10-0-135-143.us-east-2.compute.internal' from the node workqueue\nI0510 08:03:52.542910       1 controller.go:182] Assigning key: ip-10-0-135-143.us-east-2.compute.internal to node workqueue\nI0510 08:03:52.542971       1 controller.go:160] Dropping key 'ip-10-0-135-143.us-east-2.compute.internal' from the node workqueue\nI0510 08:03:57.592790       1 controller.go:182] Assigning key: ip-10-0-135-143.us-east-2.compute.internal to node workqueue\nI0510 08:03:57.592820       1 controller.go:160] Dropping key 'ip-10-0-135-143.us-east-2.compute.internal' from the node workqueue\nI0510 08:04:18.771097       1 controller.go:182] Assigning key: ip-10-0-135-143.us-east-2.compute.internal to node workqueue\nI0510 08:04:18.773029       1 controller.go:160] Dropping key 'ip-10-0-135-143.us-east-2.compute.internal' from the node workqueue\nI0510 08:04:23.347679       1 controller.go:182] Assigning key: ip-10-0-135-143.us-east-2.compute.internal to node workqueue\nI0510 08:04:23.347712       1 controller.go:160] Dropping key 'ip-10-0-135-143.us-east-2.compute.internal' from the node workqueue\nE0510 08:06:07.862804       1 leaderelection.go:330] error retrieving resource lock openshift-cloud-network-config-controller/cloud-network-config-controller-lock: Get "https://api-int.ci-op-vxzqlxhk-d5c74.aws-2.ci.openshift.org:6443/api/v1/namespaces/openshift-cloud-network-config-controller/configmaps/cloud-network-config-controller-lock": dial tcp 10.0.161.7:6443: connect: connection refused\nI0510 08:50:05.963708       1 controller.go:104] Shutting down secret workers\nI0510 08:50:05.963710       1 controller.go:104] Shutting down cloud-private-ip-config workers\nI0510 08:50:05.963725       1 controller.go:104] Shutting down node workers\nI0510 08:50:06.013099       1 main.go:161] Stopped leading, sending SIGTERM and shutting down controller\n
May 10 08:50:13.538 E ns/openshift-multus pod/multus-additional-cni-plugins-66nj7 node/ip-10-0-143-114.us-east-2.compute.internal uid/6edc088b-d375-43a2-8f38-d22e99379d26 container/kube-multus-additional-cni-plugins reason/ContainerExit code/143 cause/Error
#1788588369245114368junit9 days ago
May 09 16:59:58.732 E ns/openshift-multus pod/network-metrics-daemon-fwphf node/ip-10-0-176-109.ec2.internal uid/75db5fe2-b0db-408c-a47d-4b3de9e1d3e2 container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 09 17:00:00.652 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-176-109.ec2.internal node/ip-10-0-176-109.ec2.internal uid/64a62fb2-b4ef-4ff6-b05a-1f4d5a34d022 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0509 16:59:58.496387       1 cmd.go:216] Using insecure, self-signed certificates\nI0509 16:59:58.496775       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715273998 cert, and key in /tmp/serving-cert-3937875724/serving-signer.crt, /tmp/serving-cert-3937875724/serving-signer.key\nI0509 16:59:59.148938       1 observer_polling.go:159] Starting file observer\nW0509 16:59:59.219089       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-176-109.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0509 16:59:59.219224       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0509 16:59:59.219738       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-3937875724/tls.crt::/tmp/serving-cert-3937875724/tls.key"\nF0509 16:59:59.511827       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 09 17:00:02.533 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-176-109.ec2.internal node/ip-10-0-176-109.ec2.internal uid/64a62fb2-b4ef-4ff6-b05a-1f4d5a34d022 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0509 16:59:58.496387       1 cmd.go:216] Using insecure, self-signed certificates\nI0509 16:59:58.496775       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715273998 cert, and key in /tmp/serving-cert-3937875724/serving-signer.crt, /tmp/serving-cert-3937875724/serving-signer.key\nI0509 16:59:59.148938       1 observer_polling.go:159] Starting file observer\nW0509 16:59:59.219089       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-176-109.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0509 16:59:59.219224       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0509 16:59:59.219738       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-3937875724/tls.crt::/tmp/serving-cert-3937875724/tls.key"\nF0509 16:59:59.511827       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n

... 2 lines not shown

#1788388310087897088junit9 days ago
May 09 03:43:54.926 E ns/openshift-ovn-kubernetes pod/ovnkube-master-bnxv9 node/ip-10-0-135-248.ec2.internal uid/6d727409-841d-404a-b0c6-ecc7de38bcb3 container/ovn-dbchecker reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 09 03:43:56.930 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-135-248.ec2.internal node/ip-10-0-135-248.ec2.internal uid/f93f56fd-2f00-4abb-86ce-369431870dc1 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0509 03:43:55.178190       1 cmd.go:216] Using insecure, self-signed certificates\nI0509 03:43:55.189613       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715226235 cert, and key in /tmp/serving-cert-52079475/serving-signer.crt, /tmp/serving-cert-52079475/serving-signer.key\nI0509 03:43:55.606618       1 observer_polling.go:159] Starting file observer\nW0509 03:43:55.643096       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-135-248.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0509 03:43:55.643257       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0509 03:43:55.643751       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-52079475/tls.crt::/tmp/serving-cert-52079475/tls.key"\nF0509 03:43:56.182988       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 09 03:44:00.725 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-135-248.ec2.internal node/ip-10-0-135-248.ec2.internal uid/f93f56fd-2f00-4abb-86ce-369431870dc1 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0509 03:43:55.178190       1 cmd.go:216] Using insecure, self-signed certificates\nI0509 03:43:55.189613       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715226235 cert, and key in /tmp/serving-cert-52079475/serving-signer.crt, /tmp/serving-cert-52079475/serving-signer.key\nI0509 03:43:55.606618       1 observer_polling.go:159] Starting file observer\nW0509 03:43:55.643096       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-135-248.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0509 03:43:55.643257       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0509 03:43:55.643751       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-52079475/tls.crt::/tmp/serving-cert-52079475/tls.key"\nF0509 03:43:56.182988       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n

... 1 lines not shown

#1788295440110718976junit10 days ago
May 08 21:43:53.481 E ns/e2e-k8s-sig-apps-daemonset-upgrade-6999 pod/ds1-xf8hg node/ip-10-0-175-111.us-west-1.compute.internal uid/2fef1bdd-efc1-4b7e-b067-fb910e766374 container/app reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 08 21:43:53.501 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-175-111.us-west-1.compute.internal node/ip-10-0-175-111.us-west-1.compute.internal uid/304534d9-1418-4b01-81b4-fc1cfc5aab30 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0508 21:43:51.063524       1 cmd.go:216] Using insecure, self-signed certificates\nI0508 21:43:51.071140       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715204631 cert, and key in /tmp/serving-cert-3202481152/serving-signer.crt, /tmp/serving-cert-3202481152/serving-signer.key\nI0508 21:43:51.377986       1 observer_polling.go:159] Starting file observer\nW0508 21:43:51.389135       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-175-111.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0508 21:43:51.389291       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0508 21:43:51.402076       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-3202481152/tls.crt::/tmp/serving-cert-3202481152/tls.key"\nF0508 21:43:51.946239       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 08 21:43:54.517 E ns/openshift-multus pod/network-metrics-daemon-vzh2m node/ip-10-0-175-111.us-west-1.compute.internal uid/73636155-39f7-40b1-82bc-b8e93429fde7 container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
#1788295440110718976junit10 days ago
May 08 21:43:54.585 E ns/openshift-dns pod/dns-default-4mt7s node/ip-10-0-175-111.us-west-1.compute.internal uid/f5e175ba-92f5-4f15-b17f-12e09c8fdc0c container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 08 21:43:54.605 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-175-111.us-west-1.compute.internal node/ip-10-0-175-111.us-west-1.compute.internal uid/304534d9-1418-4b01-81b4-fc1cfc5aab30 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0508 21:43:51.063524       1 cmd.go:216] Using insecure, self-signed certificates\nI0508 21:43:51.071140       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715204631 cert, and key in /tmp/serving-cert-3202481152/serving-signer.crt, /tmp/serving-cert-3202481152/serving-signer.key\nI0508 21:43:51.377986       1 observer_polling.go:159] Starting file observer\nW0508 21:43:51.389135       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-175-111.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0508 21:43:51.389291       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0508 21:43:51.402076       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-3202481152/tls.crt::/tmp/serving-cert-3202481152/tls.key"\nF0508 21:43:51.946239       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 08 21:43:56.435 E ns/openshift-multus pod/multus-additional-cni-plugins-h2vzw node/ip-10-0-175-111.us-west-1.compute.internal uid/37a0d4e6-7e2e-4ebc-b08e-70f118a0e36d container/kube-multus-additional-cni-plugins reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
#1787946818605158400junit11 days ago
May 07 22:33:21.543 E ns/openshift-multus pod/network-metrics-daemon-8f79z node/ip-10-0-162-232.us-west-1.compute.internal uid/cbcf835b-6711-4f67-b7a1-219ddf8f94b8 container/kube-rbac-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 07 22:33:23.554 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-162-232.us-west-1.compute.internal node/ip-10-0-162-232.us-west-1.compute.internal uid/f87edc7d-8839-4fb9-85b0-e8b9f0ab0d5a container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0507 22:33:17.947322       1 cmd.go:216] Using insecure, self-signed certificates\nI0507 22:33:17.953882       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715121197 cert, and key in /tmp/serving-cert-1544848494/serving-signer.crt, /tmp/serving-cert-1544848494/serving-signer.key\nI0507 22:33:18.474001       1 observer_polling.go:159] Starting file observer\nW0507 22:33:18.494887       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-162-232.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0507 22:33:18.495043       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0507 22:33:18.501964       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-1544848494/tls.crt::/tmp/serving-cert-1544848494/tls.key"\nW0507 22:33:22.654039       1 requestheader_controller.go:193] Unable to get configmap/extension-apiserver-authentication in kube-system.  Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA'\nF0507 22:33:22.654114       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: configmaps "extension-apiserver-authentication" is forbidden: User "system:serviceaccount:openshift-kube-apiserver:check-endpoints" cannot get resource "configmaps" in API group "" in the namespace "kube-system"\n
May 07 22:33:23.571 E ns/openshift-multus pod/multus-additional-cni-plugins-szxrk node/ip-10-0-162-232.us-west-1.compute.internal uid/25fd987c-479b-4237-ae5c-d46e2ab587d7 container/kube-multus-additional-cni-plugins reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
#1787946818605158400junit11 days ago
May 07 22:33:24.578 E ns/openshift-e2e-loki pod/loki-promtail-qd7fw node/ip-10-0-162-232.us-west-1.compute.internal uid/a41aea9e-ba37-4430-8f3f-91ebf3b3d100 container/oauth-proxy reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 07 22:33:24.606 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-162-232.us-west-1.compute.internal node/ip-10-0-162-232.us-west-1.compute.internal uid/f87edc7d-8839-4fb9-85b0-e8b9f0ab0d5a container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0507 22:33:17.947322       1 cmd.go:216] Using insecure, self-signed certificates\nI0507 22:33:17.953882       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715121197 cert, and key in /tmp/serving-cert-1544848494/serving-signer.crt, /tmp/serving-cert-1544848494/serving-signer.key\nI0507 22:33:18.474001       1 observer_polling.go:159] Starting file observer\nW0507 22:33:18.494887       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-162-232.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0507 22:33:18.495043       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0507 22:33:18.501964       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-1544848494/tls.crt::/tmp/serving-cert-1544848494/tls.key"\nW0507 22:33:22.654039       1 requestheader_controller.go:193] Unable to get configmap/extension-apiserver-authentication in kube-system.  Usually fixed by 'kubectl create rolebinding -n kube-system ROLEBINDING_NAME --role=extension-apiserver-authentication-reader --serviceaccount=YOUR_NS:YOUR_SA'\nF0507 22:33:22.654114       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: configmaps "extension-apiserver-authentication" is forbidden: User "system:serviceaccount:openshift-kube-apiserver:check-endpoints" cannot get resource "configmaps" in API group "" in the namespace "kube-system"\n
May 07 22:33:51.645 E ns/openshift-controller-manager-operator pod/openshift-controller-manager-operator-5697866584-2jwcc node/ip-10-0-150-123.us-west-1.compute.internal uid/587e55ec-923e-4353-a42d-a73cba309bd0 container/openshift-controller-manager-operator reason/ContainerExit code/1 cause/Error .968141       1 base_controller.go:167] Shutting down StatusSyncer_openshift-controller-manager ...\nI0507 22:33:48.968146       1 base_controller.go:145] All StatusSyncer_openshift-controller-manager post start hooks have been terminated\nI0507 22:33:48.968154       1 operator.go:147] Shutting down OpenShiftControllerManagerOperator\nI0507 22:33:48.968311       1 dynamic_serving_content.go:146] "Shutting down controller" name="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key"\nI0507 22:33:48.968444       1 base_controller.go:114] Shutting down worker of UserCAObservationController controller ...\nI0507 22:33:48.968490       1 base_controller.go:104] All UserCAObservationController workers have been terminated\nI0507 22:33:48.968535       1 base_controller.go:114] Shutting down worker of ResourceSyncController controller ...\nI0507 22:33:48.968573       1 base_controller.go:104] All ResourceSyncController workers have been terminated\nI0507 22:33:48.968599       1 base_controller.go:114] Shutting down worker of ConfigObserver controller ...\nI0507 22:33:48.968626       1 base_controller.go:104] All ConfigObserver workers have been terminated\nI0507 22:33:48.968764       1 base_controller.go:114] Shutting down worker of OpenshiftControllerManagerStaticResources controller ...\nI0507 22:33:48.968810       1 base_controller.go:104] All OpenshiftControllerManagerStaticResources workers have been terminated\nI0507 22:33:48.968460       1 base_controller.go:114] Shutting down worker of StatusSyncer_openshift-controller-manager controller ...\nI0507 22:33:48.968871       1 base_controller.go:104] All StatusSyncer_openshift-controller-manager workers have been terminated\nI0507 22:33:48.968935       1 secure_serving.go:255] Stopped listening on [::]:8443\nI0507 22:33:48.968985       1 genericapiserver.go:543] "[graceful-termination] shutdown event" name="HTTPServerStoppedListening"\nW0507 22:33:48.969067       1 builder.go:109] graceful termination failed, controllers failed with error: stopped\n
#1787554361761599488junit12 days ago
May 06 19:53:19.340 E ns/openshift-cluster-csi-drivers pod/aws-ebs-csi-driver-node-8sg9r node/ip-10-0-169-81.ec2.internal uid/2bc75617-8120-45db-8d77-51627371b0b0 container/csi-liveness-probe reason/ContainerExit code/2 cause/Error
May 06 19:55:44.054 E ns/openshift-cloud-network-config-controller pod/cloud-network-config-controller-86dbd46d4f-z7m8v node/ip-10-0-159-19.ec2.internal uid/76b754f5-0676-4d4d-844d-623646144354 container/controller reason/ContainerExit code/1 cause/Error de workqueue\nI0506 19:07:48.168478       1 controller.go:182] Assigning key: ip-10-0-192-54.ec2.internal to node workqueue\nI0506 19:07:48.168606       1 controller.go:160] Dropping key 'ip-10-0-192-54.ec2.internal' from the node workqueue\nI0506 19:07:48.225899       1 controller.go:182] Assigning key: ip-10-0-192-54.ec2.internal to node workqueue\nI0506 19:07:48.226033       1 controller.go:160] Dropping key 'ip-10-0-192-54.ec2.internal' from the node workqueue\nI0506 19:08:01.392347       1 controller.go:182] Assigning key: ip-10-0-192-54.ec2.internal to node workqueue\nI0506 19:08:01.409639       1 controller.go:160] Dropping key 'ip-10-0-192-54.ec2.internal' from the node workqueue\nI0506 19:08:03.178066       1 controller.go:182] Assigning key: ip-10-0-192-54.ec2.internal to node workqueue\nI0506 19:08:03.184064       1 controller.go:160] Dropping key 'ip-10-0-192-54.ec2.internal' from the node workqueue\nE0506 19:09:03.994066       1 reflector.go:138] k8s.io/client-go/informers/factory.go:134: Failed to watch *v1.Node: unknown (get nodes)\nE0506 19:12:44.748791       1 leaderelection.go:330] error retrieving resource lock openshift-cloud-network-config-controller/cloud-network-config-controller-lock: Get "https://api-int.ci-op-nf60p02m-d5c74.aws-2.ci.openshift.org:6443/api/v1/namespaces/openshift-cloud-network-config-controller/configmaps/cloud-network-config-controller-lock": dial tcp 10.0.213.123:6443: connect: connection refused\nE0506 19:12:56.431672       1 reflector.go:138] github.com/openshift/client-go/cloudnetwork/informers/externalversions/factory.go:101: Failed to watch *v1.CloudPrivateIPConfig: unknown (get cloudprivateipconfigs.cloud.network.openshift.io)\nI0506 19:55:43.004802       1 controller.go:104] Shutting down node workers\nI0506 19:55:43.004907       1 controller.go:104] Shutting down secret workers\nI0506 19:55:43.005541       1 controller.go:104] Shutting down cloud-private-ip-config workers\nI0506 19:55:43.022918       1 main.go:161] Stopped leading, sending SIGTERM and shutting down controller\n
May 06 19:55:50.531 E ns/openshift-multus pod/multus-additional-cni-plugins-dzf7c node/ip-10-0-140-69.ec2.internal uid/763fcecc-9be2-41c6-a313-c54ef4c8fd27 container/kube-multus-additional-cni-plugins reason/ContainerExit code/143 cause/Error
#1787554361761599488junit12 days ago
May 06 20:20:21.107 E ns/openshift-ovn-kubernetes pod/ovnkube-master-7jt8r node/ip-10-0-159-19.ec2.internal uid/a35693a4-8c54-42cb-aa6b-869e5b6ad853 container/northd reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 06 20:20:24.123 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-159-19.ec2.internal node/ip-10-0-159-19.ec2.internal uid/498c843c-9c7e-46c5-882a-6f7491393b43 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0506 20:20:22.362223       1 cmd.go:216] Using insecure, self-signed certificates\nI0506 20:20:22.369350       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715026822 cert, and key in /tmp/serving-cert-2063183158/serving-signer.crt, /tmp/serving-cert-2063183158/serving-signer.key\nI0506 20:20:22.913157       1 observer_polling.go:159] Starting file observer\nW0506 20:20:22.927732       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-159-19.ec2.internal": dial tcp [::1]:6443: connect: connection refused\nI0506 20:20:22.930215       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0506 20:20:22.936107       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-2063183158/tls.crt::/tmp/serving-cert-2063183158/tls.key"\nF0506 20:20:23.396888       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 06 20:20:25.419 E ns/e2e-k8s-sig-apps-daemonset-upgrade-8336 pod/ds1-qvp4w node/ip-10-0-159-19.ec2.internal uid/5f8aff6d-43db-4434-af38-7e32a2955765 container/app reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
#1787462282045820928junit12 days ago
May 06 13:56:16.201 E ns/openshift-cluster-csi-drivers pod/aws-ebs-csi-driver-node-9jfx2 node/ip-10-0-183-190.us-west-1.compute.internal uid/da40db02-ec02-4e96-8345-725df2fac0f6 container/csi-driver reason/ContainerExit code/2 cause/Error
May 06 13:57:37.366 E ns/openshift-cloud-network-config-controller pod/cloud-network-config-controller-7b58b67569-56pmn node/ip-10-0-164-129.us-west-1.compute.internal uid/3c1286e7-72d6-447c-bdfe-ae1c3302220b container/controller reason/ContainerExit code/1 cause/Error  the node workqueue\nI0506 13:01:44.971675       1 controller.go:182] Assigning key: ip-10-0-203-71.us-west-1.compute.internal to node workqueue\nI0506 13:01:44.971791       1 controller.go:182] Assigning key: ip-10-0-168-153.us-west-1.compute.internal to node workqueue\nI0506 13:01:44.971871       1 controller.go:160] Dropping key 'ip-10-0-168-153.us-west-1.compute.internal' from the node workqueue\nI0506 13:01:44.971848       1 controller.go:160] Dropping key 'ip-10-0-203-71.us-west-1.compute.internal' from the node workqueue\nI0506 13:01:44.984023       1 controller.go:182] Assigning key: ip-10-0-178-22.us-west-1.compute.internal to node workqueue\nI0506 13:01:44.984140       1 controller.go:160] Dropping key 'ip-10-0-178-22.us-west-1.compute.internal' from the node workqueue\nI0506 13:01:45.008650       1 controller.go:182] Assigning key: ip-10-0-178-22.us-west-1.compute.internal to node workqueue\nI0506 13:01:45.008781       1 controller.go:160] Dropping key 'ip-10-0-178-22.us-west-1.compute.internal' from the node workqueue\nI0506 13:01:45.109193       1 controller.go:182] Assigning key: ip-10-0-168-153.us-west-1.compute.internal to node workqueue\nI0506 13:01:45.110066       1 controller.go:160] Dropping key 'ip-10-0-168-153.us-west-1.compute.internal' from the node workqueue\nE0506 13:02:37.178530       1 leaderelection.go:330] error retrieving resource lock openshift-cloud-network-config-controller/cloud-network-config-controller-lock: Get "https://api-int.ci-op-rpjjis3h-d5c74.aws-2.ci.openshift.org:6443/api/v1/namespaces/openshift-cloud-network-config-controller/configmaps/cloud-network-config-controller-lock": dial tcp 10.0.238.152:6443: connect: connection refused\nI0506 13:57:36.213786       1 controller.go:104] Shutting down secret workers\nI0506 13:57:36.213914       1 controller.go:104] Shutting down cloud-private-ip-config workers\nI0506 13:57:36.214672       1 controller.go:104] Shutting down node workers\nI0506 13:57:36.253091       1 main.go:161] Stopped leading, sending SIGTERM and shutting down controller\n
May 06 13:57:43.036 E ns/openshift-multus pod/multus-additional-cni-plugins-775pj node/ip-10-0-230-131.us-west-1.compute.internal uid/d3703264-91d2-4d96-9095-370bbfe143b2 container/kube-multus-additional-cni-plugins reason/ContainerExit code/143 cause/Error
#1787462282045820928junit12 days ago
May 06 14:28:41.702 E ns/openshift-cluster-csi-drivers pod/aws-ebs-csi-driver-node-chs77 node/ip-10-0-164-129.us-west-1.compute.internal uid/8bc87ba6-45ce-4af8-b893-83a33ca99c1e container/csi-driver reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)
May 06 14:28:46.814 E ns/openshift-kube-apiserver pod/kube-apiserver-ip-10-0-164-129.us-west-1.compute.internal node/ip-10-0-164-129.us-west-1.compute.internal uid/e96cc944-7474-4346-94e7-415bdfcb00e1 container/kube-apiserver-check-endpoints reason/ContainerExit code/255 cause/Error W0506 14:28:44.664514       1 cmd.go:216] Using insecure, self-signed certificates\nI0506 14:28:44.679803       1 crypto.go:601] Generating new CA for check-endpoints-signer@1715005724 cert, and key in /tmp/serving-cert-655999110/serving-signer.crt, /tmp/serving-cert-655999110/serving-signer.key\nI0506 14:28:45.277369       1 observer_polling.go:159] Starting file observer\nW0506 14:28:45.294384       1 builder.go:239] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-ip-10-0-164-129.us-west-1.compute.internal": dial tcp [::1]:6443: connect: connection refused\nI0506 14:28:45.294541       1 builder.go:271] check-endpoints version v4.0.0-alpha.0-1741-g09d7ddb-09d7ddbab\nI0506 14:28:45.295077       1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/tmp/serving-cert-655999110/tls.crt::/tmp/serving-cert-655999110/tls.key"\nF0506 14:28:45.590002       1 cmd.go:141] error initializing delegating authentication: unable to load configmap based request-header-client-ca-file: Get "https://localhost:6443/api/v1/namespaces/kube-system/configmaps/extension-apiserver-authentication": dial tcp [::1]:6443: connect: connection refused\n
May 06 14:28:48.421 E ns/openshift-network-diagnostics pod/network-check-target-5nzsz node/ip-10-0-164-129.us-west-1.compute.internal uid/dba792e5-ffb2-474c-9763-c7522c88f82f container/network-check-target-container reason/TerminationStateCleared lastState.terminated was cleared on a pod (bug https://bugzilla.redhat.com/show_bug.cgi?id=1933760 or similar)

Found in 62.16% of runs (143.75% of failures) across 37 total runs and 1 jobs (43.24% failed) in 479ms - clear search | chart view - source code located on github