Re: Problem while upgrade 17.2.6 to 17.2.7

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

These are the last lines of /var/log/ceph/cephadm.log of the active mgr machine after an error occured. As I don't feel this will be very helpfull, would you please tell us where to look ?

Best regards,

JM Fontana

2023-11-16 14:45:08,200 7f341eae8740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:46:10,406 7fca81386740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:47:12,594 7fd48f814740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:48:14,857 7fd0b24b1740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'check-host']
2023-11-16 14:48:14,990 7fd0b24b1740 INFO podman (/usr/bin/podman) version 3.0.1 is present
2023-11-16 14:48:14,992 7fd0b24b1740 INFO systemctl is present
2023-11-16 14:48:14,993 7fd0b24b1740 INFO lvcreate is present
2023-11-16 14:48:15,041 7fd0b24b1740 INFO Unit chrony.service is enabled and running
2023-11-16 14:48:15,043 7fd0b24b1740 INFO Host looks OK
2023-11-16 14:48:15,655 7f36b81fd740 DEBUG -------------------------------------------------------------------------------- cephadm ['--image', 'quay.io/ceph/ceph@sha256:56984a149e89ce282e9400ca53371ff7df74b1c7f5e979b6ec651b751931483a', '--timeout', '895', 'ls'] 2023-11-16 14:48:17,662 7f17bfc28740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:49:20,131 7fc8a9cc1740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:50:22,284 7f1a6a7eb740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:51:24,505 7f1798dd5740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:52:26,574 7f0185a55740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:53:28,630 7f9bc3fff740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:54:30,673 7fc3752d0740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:55:32,662 7fd3865f8740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:56:34,686 7f73eedd2740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:57:36,799 7fbce19d2740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:58:38,874 7f8b5be4d740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'check-host']
2023-11-16 14:58:38,983 7f8b5be4d740 INFO podman (/usr/bin/podman) version 3.0.1 is present
2023-11-16 14:58:38,985 7f8b5be4d740 INFO systemctl is present
2023-11-16 14:58:38,987 7f8b5be4d740 INFO lvcreate is present
2023-11-16 14:58:39,050 7f8b5be4d740 INFO Unit chrony.service is enabled and running
2023-11-16 14:58:39,053 7f8b5be4d740 INFO Host looks OK
2023-11-16 14:58:39,923 7f5878f1c740 DEBUG -------------------------------------------------------------------------------- cephadm ['--image', 'quay.io/ceph/ceph@sha256:56984a149e89ce282e9400ca53371ff7df74b1c7f5e979b6ec651b751931483a', '--timeout', '895', 'ls'] 2023-11-16 14:58:41,730 7fd774f12740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 14:59:44,116 7f5822228740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 15:00:46,276 7fbc86e16740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 15:01:48,291 7fec587af740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 15:02:50,500 7f6338963740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 15:02:51,882 7fbc52e2f740 DEBUG -------------------------------------------------------------------------------- cephadm ['--image', 'quay.io/ceph/ceph@sha256:56984a149e89ce282e9400ca53371ff7df74b1c7f5e979b6ec651b751931483a', '--timeout', '895', 'list-networks'] 2023-11-16 15:03:53,692 7f652d1e6740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']
2023-11-16 15:04:56,193 7f2c66ce3740 DEBUG --------------------------------------------------------------------------------
cephadm ['--timeout', '895', 'gather-facts']

Le 16/11/2023 à 12:41, Nizamudeen A a écrit :
Hello,

can you also add the mgr logs at the time of this error?

Regards,

On Thu, Nov 16, 2023 at 4:12 PM Jean-Marc FONTANA <jean-marc.fontana@xxxxxxx> wrote:

    Hello David,

    We tried what you pointed in your message. First, it was set to

    "s3, s3website, swift, swift_auth, admin, sts, iam, subpub"

    We tried to set it to "s3, s3website, swift, swift_auth, admin, sts,
    iam, subpub, notifications"

    and then to "s3, s3website, swift, swift_auth, admin, sts, iam,
    notifications",

    with no success at each time.

    We tried then

           ceph dashboard reset-rgw-api-admin-resource

    or

           ceph dashboard set-rgw-api-admin-resource XXX

    getting a 500 internal error message in a red box on the upper corner
    with the first one

    or the 404 error message with the second one.

    Thanks for your helping,

    Cordialement,

    JM Fontana


    Le 14/11/2023 à 20:53, David C. a écrit :
    > Hi Jean Marc,
    >
    > maybe look at this parameter "rgw_enable_apis", if the values
    you have
    > correspond to the default (need rgw restart) :
    >
    >
    https://docs.ceph.com/en/quincy/radosgw/config-ref/#confval-rgw_enable_apis
    >
    > ceph config get client.rgw rgw_enable_apis
    >
    > ________________________________________________________
    >
    > Cordialement,
    >
    > *David CASIER*
    >
    > ________________________________________________________
    >
    >
    >
    > Le mar. 14 nov. 2023 à 11:45, Jean-Marc FONTANA
    > <jean-marc.fontana@xxxxxxx> a écrit :
    >
    >     Hello everyone,
    >
    >     We operate two clusters that we installed with ceph-deploy in
    >     Nautilus
    >     version on Debian 10. We use them for external S3 storage
    >     (owncloud) and
    >     rbd disk images.We had them upgraded to Octopus and Pacific
    >     versions on
    >     Debian 11 and recently converted them to cephadm and upgraded to
    >     Quincy
    >     (17.2.6).
    >
    >     As we now have the orchestrator, we tried updating to 17.2.7
    using
    >     the
    >     command# ceph orch upgrade start --image
    quay.io/ceph/ceph:v17.2.7 <http://quay.io/ceph/ceph:v17.2.7>
    >     <http://quay.io/ceph/ceph:v17.2.7>
    >
    >     Everything went well, both clusters work perfectly for our use,
    >     except
    >     that the Rados gateway configuration is no longer accessible
    from the
    >     dashboard with the following error messageError connecting
    to Object
    >     Gateway: RGW REST API failed request with status code 404.
    >
    >     We tried a few solutions found on the internet (reset rgw
    >     credentials,
    >     restart rgw adnd mgr, reenable dashboard, ...), unsuccessfully.
    >
    >     Does somebody have an idea ?
    >
    >     Best regards,
    >
    >     Jean-Marc Fontana
    >     _______________________________________________
    >     ceph-users mailing list -- ceph-users@xxxxxxx
    >     To unsubscribe send an email to ceph-users-leave@xxxxxxx
    >
    _______________________________________________
    ceph-users mailing list -- ceph-users@xxxxxxx
    To unsubscribe send an email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux