Hello Simon,The nautilus leftover idea sounds quite realistic to me. When I had Nautilus/CentOS7.7 combination running, I had a lot of internal server error 500 troubleusing the Ceph Dashboard. But only when I tried to access the RadosGateway tab in the nautilus dashboard. However, I was never able to solve this myself, until I created a fresh Octopus/CentOS 8.1 installation. Full dashboard accessible without any problem.Hope you get it up-and-running by reinstallation.Success.rgds,- gw On Wed, 2020-04-29 at 06:48 +0000, Simon Sutter wrote: > Hello Gert, > I recreated the self signed certificate. > SELinux was disabled and I temporarely disabled the firewall. > It still doesn't work and there is no entry in journalctl -f. > > Somewhere there is still something from the previous nautilus or > centos7 installation, causing this problem.I think I'll have to > reinstall the node. > > I'll update you. > Thanks and kind regards, > Simon > ________________________________Von: Gert Wieberdink < > gert.wieberdink@xxxxxxxx>Gesendet: Dienstag, 28. April 2020 > 21:16:10An: Simon Sutter; ceph-users@xxxxxxx > Betreff: Re: Re: Upgrading to Octopus > Sorry for the typo: must be journalctl -f instead of syslogctl -f.-gw > > On Tue, 2020-04-28 at 19:12 +0000, Gert Wieberdink wrote: > Hello Simon,ceph-mgr and dashboard installation should be > straightforward. > These are tough ones (internal server error 500). Did you create a > self > signed cert for dashboard?Did you check firewalld (port 8443) and/or > SELinux? Does syslogctl -f show anything? > rgds,-gw > On Tue, 2020-04-28 at 12:17 +0000, Simon Sutter wrote: > > > <!-- > p > {margin-top:0; > margin-bottom:0} > --> > > > > Hello, > > > > Yes I upgraded the system to Centos8 and now I can install the > dashboard module. > But the problem now is, I cannot log in to the dashboard. > I deleted every cached file on my end and reinstalled the mgr and > dashboard several times. > If I try to log in with a wrong password, it tells me that it's > wrong, but if i use the right password, it just gives me a "500 > Internal Server Error". > > I enabled the debug mode for the mgr: ceph config set mgr > mgr/dashboard/log_level debug > > But in the /var/log/ceph/ceph-mgr.node1.log it just tells me this > generic message (ips replaced with 0.0.0.0): > > > > > > 2020-04-28T14:11:15.191+0200 7f0baba8c700 0 [dashboard DEBUG > request] [::ffff:0.0.0.0:61383] [POST] [None] /api/auth > 2020-04-28T14:11:15.282+0200 7f0bcf164700 0 log_channel(cluster) log > [DBG] : pgmap v316: 273 pgs: 273 active+clean; 2.4 TiB data, 7.1 TiB > used, 18 TiB / 25 TiB avail > 2020-04-28T14:11:15.453+0200 7f0baba8c700 0 [dashboard DEBUG > controllers.auth] Login successful > 2020-04-28T14:11:15.453+0200 7f0baba8c700 0 [dashboard ERROR > request] [::ffff:0.0.0.0:61383] [POST] [500] [0.264s] [513.0B] > [100ecd9a-5d09-419f-8b9f-31bc3d4042b4] /api/auth > 2020-04-28T14:11:15.453+0200 7f0baba8c700 0 [dashboard ERROR > request] [b'{"status": "500 Internal Server Error", "detail": "The > server encountered an unexpected condition which prevented it from > fulfilling the request.", "request_id": "100ecd9a-5d09-419f-8b9f- > 31bc3d4042b4"} > > > > > > > > > '] > 2020-04-28T14:11:15.454+0200 7f0baba8c700 0 [dashboard INFO request] > [::ffff:0.0.0.0:61383] [POST] [500] [0.264s] [513.0B] [100ecd9a-5d09- > 419f-8b9f-31bc3d4042b4] /api/auth > > > > How can I find out, where the problem is? > > > > Thanks in advance, > Simon > > > Von: > <mailto:gert.wieberdink@xxxxxxxx> > gert.wieberdink@xxxxxxxx > > < > <mailto:gert.wieberdink@xxxxxxxx> > gert.wieberdink@xxxxxxxx > > > > Gesendet: Donnerstag, 23. April 2020 20:34:58 > > An: > <mailto:ceph-users@xxxxxxx> > ceph-users@xxxxxxx > > > > Betreff: Re: Upgrading to Octopus > > > > > > Hello Simon, > > I think that Khodayar is right. I managed to install a new Ceph > cluster on CentOS 8.1. Therefore you will need the ceph-el8.repo for > the time being. For some reason, "they" left the py3 packages you > mentioned out of EPEL (as with leveldb, but this package appeared > luckily last week in EPEL). > > Please find below the ceph-el8.repo file, which you have to create in > /etc/yum.repos.d/ > > > > [copr:copr.fedorainfracloud.org:ktdreyer:ceph-el8] > > name=Copr repo for ceph-el8 owned by ktdreyer > > baseurl= > < > https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/epel-8-$basearch/ > > > https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/epel-8-$basearch/ > > > > type=rpm-md > > skip_if_unavailable=True > > gpgcheck=1 > > gpgkey= > < > https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/pubkey.gpg > > > https://download.copr.fedorainfracloud.org/results/ktdreyer/ceph-el8/pubkey.gpg > > > > repo_gpgcheck=0 > > enabled=1 > > enabled_metadata=1 > > > > This repository - and CentOS 8.x - should have been sufficient to > bring up a fresh Ceph cluster. > > Please let me know if you still have problems in configuring your > Ceph cluster. > > rgds, > > -gw > > _______________________________________________ > > ceph-users mailing list -- > <mailto:ceph-users@xxxxxxx> > ceph-users@xxxxxxx > > > > To unsubscribe send an email to > <mailto:ceph-users-leave@xxxxxxx> > ceph-users-leave@xxxxxxx > > > > > > > > > _______________________________________________ > ceph-users mailing list -- > <mailto:ceph-users@xxxxxxx> > ceph-users@xxxxxxx > > > To unsubscribe send an email to > <mailto:ceph-users-leave@xxxxxxx> > ceph-users-leave@xxxxxxx > > _______________________________________________ceph-users mailing > list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx