On Tue, Dec 20, 2022 at 11:41 AM Casey Bodley <cbodley@xxxxxxxxxx> wrote: > thanks Yuri, rgw approved based on today's results from > > https://pulpito.ceph.com/yuriw-2022-12-20_15:27:49-rgw-pacific_16.2.11_RC2-distro-default-smithi/ > > On Mon, Dec 19, 2022 at 12:08 PM Yuri Weinstein <yweinste@xxxxxxxxxx> > wrote: > > > If you look at the pacific 16.2.8 QE validation history ( > > https://tracker.ceph.com/issues/55356), we had pacific-x, nautilus-x, > and > > pacific-p2p all green with one exception ( > > https://tracker.ceph.com/issues/51652) > > > > Now we see so many failures in this point release with references to old > > issues. > > > > Is there anything we can fix to make them less "red"? > > > > Thx > > YuriW > > > > On Thu, Dec 15, 2022 at 2:56 PM Laura Flores <lflores@xxxxxxxxxx> wrote: > > > >> I reviewed the upgrade runs: > >> > >> > >> > https://pulpito.ceph.com/yuriw-2022-12-13_15:57:57-upgrade:nautilus-x-pacific_16.2.11_RC-distro-default-smithi/ > >> > >> > https://pulpito.ceph.com/yuriw-2022-12-13_21:47:46-upgrade:nautilus-x-pacific_16.2.11_RC-distro-default-smithi/ > >> > >> > https://pulpito.ceph.com/yuriw-2022-12-13_15:58:18-upgrade:octopus-x-pacific_16.2.11_RC-distro-default-smithi/ > >> > >> > https://pulpito.ceph.com/yuriw-2022-12-14_15:41:10-upgrade:octopus-x-pacific_16.2.11_RC-distro-default-smithi/ > >> > >> Failures: > >> 1. https://tracker.ceph.com/issues/50618 -- known bug assigned to > >> Ilya; assuming it's not a big deal since it's been around for over a > year > >> > >> Details: > >> 1. qemu_xfstests_luks1 failed on xfstest 168 - Ceph - RBD > >> > >> > >> > >> > https://pulpito.ceph.com/yuriw-2022-12-13_15:58:24-upgrade:pacific-p2p-pacific_16.2.11_RC-distro-default-smithi/ > >> > >> > https://pulpito.ceph.com/yuriw-2022-12-14_15:40:37-upgrade:pacific-p2p-pacific_16.2.11_RC-distro-default-smithi/ > >> > >> Failures, unrelated: > >> 1. https://tracker.ceph.com/issues/58223 -- new failure reported by > me > >> 7 days ago; seems infrastructure related and not regression-related > >> 2. https://tracker.ceph.com/issues/52590 -- closed by Casey; must not > >> be of importance > >> 3. https://tracker.ceph.com/issues/58289 -- new failure raised by me > >> today; seems related to other "wait_for_recovery" failures, which are > >> generally not cause for concern since they're so infrequent. > >> 4. https://tracker.ceph.com/issues/51652 -- known bug from over a > year > >> ago > >> > >> Details; > >> 1. failure on `sudo fuser -v /var/lib/dpkg/lock-frontend` - > >> Infrastructure > >> 2. "[ FAILED ] CmpOmap.cmp_vals_u64_invalid_default" in > >> upgrade:pacific-p2p-pacific - Ceph - RGW > >> 3. "AssertionError: wait_for_recovery: failed before timeout expired" > >> from down pg in pacific-p2p-pacific - Ceph - RADOS > >> 4. heartbeat timeouts on filestore OSDs while deleting objects in > >> upgrade:pacific-p2p-pacific - Ceph - RADOS > >> > >> On Thu, Dec 15, 2022 at 4:34 PM Brad Hubbard <bhubbard@xxxxxxxxxx> > wrote: > >> > >>> On Fri, Dec 16, 2022 at 3:15 AM Yuri Weinstein <yweinste@xxxxxxxxxx> > >>> wrote: > >>> > > >>> > Details of this release are summarized here: > >>> > > >>> > https://tracker.ceph.com/issues/58257#note-1 > >>> > Release Notes - TBD > >>> > > >>> > Seeking approvals for: > >>> > > >>> > rados - Neha (https://github.com/ceph/ceph/pull/49431 is still being > We are still waiting on test runs for https://github.com/ceph/ceph/pull/49431 to go through. Thanks, Neha > >>> > tested and will be merged soon) > >>> > rook - Sébastien Han > >>> > cephadm - Adam > >>> > dashboard - Ernesto > >>> > rgw - Casey (rwg will be rerun on the latest SHA1) > >>> > rbd - Ilya, Deepika > >>> > krbd - Ilya, Deepika > >>> > fs - Venky, Patrick > >>> > upgrade/nautilus-x (pacific) - Neha, Laura > >>> > upgrade/octopus-x (pacific) - Neha, Laura > >>> > upgrade/pacific-p2p - Neha - Neha, Laura > >>> > powercycle - Brad > >>> > >>> The failure here is due to fallout from the recent lab issues and was > >>> fixed in main by https://github.com/ceph/ceph/pull/49021 I'm waiting > >>> to see if there are plans to backport this to pacific and quincy since > >>> that will be needed. > >>> > >>> > ceph-volume - Guillaume, Adam K > >>> > > >>> > Thx > >>> > YuriW > >>> > > >>> > _______________________________________________ > >>> > Dev mailing list -- dev@xxxxxxx > >>> > To unsubscribe send an email to dev-leave@xxxxxxx > >>> > >>> > >>> > >>> -- > >>> Cheers, > >>> Brad > >>> > >>> _______________________________________________ > >>> ceph-users mailing list -- ceph-users@xxxxxxx > >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx > >>> > >> > >> > >> -- > >> > >> Laura Flores > >> > >> She/Her/Hers > >> > >> Software Engineer, Ceph Storage > >> > >> Red Hat Inc. <https://www.redhat.com> > >> > >> Chicago, IL > >> > >> lflores@xxxxxxxxxx > >> M: +17087388804 > >> @RedHat <https://twitter.com/redhat> Red Hat > >> <https://www.linkedin.com/company/red-hat> Red Hat > >> <https://www.facebook.com/RedHatInc> > >> <https://www.redhat.com> > >> > >> _______________________________________________ > > Dev mailing list -- dev@xxxxxxx > > To unsubscribe send an email to dev-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx