Hey ceph-users,
What seems odd, is my ceph.conf never had e238f5b3-7d67-4b55-8563-52008828db51 as the fsid. I even pulled from backups, and it has always been:
One of of ceph environments changed its fsid for the cluster, and I would like advice on how to get it corrected.
We added a new OSD node in hope of retiring one of the older OSD + MON nodes.
Using ceph-deploy, we unfortunately ran "ceph-deploy mon create ..." instead of mon add
The ceph.log file reported:
[ceph_deploy.conf][DEBUG ] found configuration file at: /root/.cephdeploy.conf[ceph_deploy.cli][INFO ] Invoked (1.5.23): /usr/bin/ceph-deploy mon create ceph-osd7[ceph_deploy.mon][DEBUG ] Deploying mon, cluster ceph hosts ceph-osd7[ceph_deploy.mon][DEBUG ] detecting platform for host ceph-osd7 ...[ceph-osd7][DEBUG ] connection detected need for sudo[ceph-osd7][DEBUG ] connected to host: ceph-osd7[ceph-osd7][DEBUG ] detect platform information from remote host[ceph-osd7][DEBUG ] detect machine type[ceph_deploy.mon][INFO ] distro info: Ubuntu 14.04 trusty[ceph-osd7][DEBUG ] determining if provided host has same hostname in remote[ceph-osd7][DEBUG ] get remote short hostname[ceph-osd7][DEBUG ] deploying mon to ceph-osd7[ceph-osd7][DEBUG ] get remote short hostname[ceph-osd7][DEBUG ] remote hostname: ceph-osd7[ceph-osd7][DEBUG ] write cluster configuration to /etc/ceph/{cluster}.conf[ceph-osd7][DEBUG ] create the mon path if it does not exist[ceph-osd7][DEBUG ] checking for done path: /var/lib/ceph/mon/ceph-ceph-osd7/done[ceph-osd7][DEBUG ] done path does not exist: /var/lib/ceph/mon/ceph-ceph-osd7/done[ceph-osd7][INFO ] creating keyring file: /var/lib/ceph/tmp/ceph-ceph-osd7.mon.keyring[ceph-osd7][DEBUG ] create the monitor keyring file[ceph-osd7][INFO ] Running command: sudo ceph-mon --cluster ceph --mkfs -i ceph-osd7 --keyring /var/lib/ceph/tmp/ceph-ceph-osd7.mon.keyring[ceph-osd7][DEBUG ] ceph-mon: set fsid to 6870cff2-6cbc-4e99-8615-c159ba3a0546
So, it looks like the fsid was changed from e238f5b3-7d67-4b55-8563-52008828db51 to 6870cff2-6cbc-4e99-8615-c159ba3a0546
ceph -s shows the previous fsid:
ceph -scluster e238f5b3-7d67-4b55-8563-52008828db51health HEALTH_WARNtoo few PGs per OSD (29 < min 30)1 mons down, quorum 1,2 ceph-mon,ceph-osd3monmap e9: 3 mons at {ceph-mon=10.5.68.69:6789/0,ceph-osd3=10.5.68.92:6789/0,ceph-osd6=10.5.68.35:6789/0}election epoch 416, quorum 1,2 ceph-mon,ceph-osd3mdsmap e640: 1/1/1 up {0=ceph-mon=up:active}, 1 up:standbyosdmap e6275: 58 osds: 58 up, 58 inpgmap v18874768: 848 pgs, 16 pools, 4197 GB data, 911 kobjects8691 GB used, 22190 GB / 30881 GB avail848 active+cleanclient io 0 B/s rd, 345 kB/s wr, 58 op/s
root@ceph-mon:~/RESTORE/2016-01-02/etc/ceph# cat ceph.conf[global]fsid = 6870cff2-6cbc-4e99-8615-c159ba3a0546mon_initial_members = ceph-monmon_host = 10.5.68.69,10.5.68.65,10.5.68.92auth_cluster_required = cephxauth_service_required = cephxauth_client_required = cephxfilestore_xattr_use_omap = trueosd_pool_default_size = 2public_network = 10.5.68.0/24cluster_network = 10.7.1.0/24
The cluster seems to be "up", but I'm concerned that I only have 2 monitors, I cannot add a third since authentication to the cluster fails:
2016-01-20 16:41:09.544870 7f1f238ed8c0 0 ceph version 0.94.2 (5fb85614ca8f354284c713a2f9c610860720bbf3), process ceph-mon, pid 320102016-01-20 16:41:09.765043 7f1f238ed8c0 0 mon.ceph-osd6 does not exist in monmap, will attempt to join an existing cluster2016-01-20 16:41:09.773435 7f1f238ed8c0 0 using public_addr 10.5.68.35:0/0 -> 10.5.68.35:6789/02016-01-20 16:41:09.773517 7f1f238ed8c0 0 starting mon.ceph-osd6 rank -1 at 10.5.68.35:6789/0 mon_data /var/lib/ceph/mon/ceph-ceph-osd6 fsid 6870cff2-6cbc-4e99-8615-c159ba3a05462016-01-20 16:41:09.774549 7f1f238ed8c0 1 mon.ceph-osd6@-1(probing) e0 preinit fsid 6870cff2-6cbc-4e99-8615-c159ba3a05462016-01-20 16:41:10.746413 7f1f1ec65700 0 log_channel(audit) log [DBG] : from='admin socket' entity='admin socket' cmd='mon_status' args=[]: dispatch
And yes, we need to increase our PG count. This cluster has grown from a few 2TB drives to multiple 600GB sas drives, but I don't want to touch anything else until I can get this figured out.
This is running as our Openstack VM storage, so it is not something we can simply rebuild.
Thanks,
Mike C
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com