Re: Cluster Map Problems

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi John,

my ceph.conf is a bit further down in this email.

-martin

Am 28.03.2013 23:21, schrieb John Wilkins:
Martin,

Would you mind posting your Ceph configuration file too?  I don't see
any value set for "mon_host": ""

On Thu, Mar 28, 2013 at 1:04 PM, Martin Mailand <martin@xxxxxxxxxxxx> wrote:
Hi Greg,

the dump from mon.a is attached.

-martin

On 28.03.2013 20:55, Gregory Farnum wrote:
Hmm. The monitor code for checking this all looks good to me. Can you
go to one of your monitor nodes and dump the config?
(http://ceph.com/docs/master/rados/configuration/ceph-conf/?highlight=admin%20socket#viewing-a-configuration-at-runtime)
-Greg

On Thu, Mar 28, 2013 at 12:33 PM, Martin Mailand <martin@xxxxxxxxxxxx> wrote:
Hi,

I get the same behavior an new created cluster as well, no changes to
the cluster config at all.
I stop the osd.1, after 20 seconds it got marked down. But it never get
marked out.

ceph version 0.59 (cbae6a435c62899f857775f66659de052fb0e759)

-martin

On 28.03.2013 19:48, John Wilkins wrote:
Martin,

Greg is talking about noout. With Ceph, you can specifically preclude
OSDs from being marked out when down to prevent rebalancing--e.g.,
during upgrades, short-term maintenance, etc.

http://ceph.com/docs/master/rados/operations/troubleshooting-osd/#stopping-w-out-rebalancing

On Thu, Mar 28, 2013 at 11:12 AM, Martin Mailand <martin@xxxxxxxxxxxx> wrote:
Hi Greg,

setting the osd manually out triggered the recovery.
But now it is the question, why is the osd not marked out after 300
seconds? That's a default cluster, I use the 0.59 build from your site.
And I didn't change any value, except for the crushmap.

That's my ceph.conf.

-martin

[global]
         auth cluster requierd = none
         auth service required = none
         auth client required = none
#       log file = ""
         log_max_recent=100
         log_max_new=100

[mon]
         mon data = /data/mon.$id
[mon.a]
         host = store1
         mon addr = 192.168.195.31:6789
[mon.b]
         host = store3
         mon addr = 192.168.195.33:6789
[mon.c]
         host = store5
         mon addr = 192.168.195.35:6789
[osd]
         journal aio = true
         osd data = /data/osd.$id
         osd mount options btrfs = rw,noatime,nodiratime,autodefrag
         osd mkfs options btrfs = -n 32k -l 32k

[osd.0]
         host = store1
         osd journal = /dev/sdg1
         btrfs devs = /dev/sdc
[osd.1]
         host = store1
         osd journal = /dev/sdh1
         btrfs devs = /dev/sdd
[osd.2]
         host = store1
         osd journal = /dev/sdi1
         btrfs devs = /dev/sde
[osd.3]
         host = store1
         osd journal = /dev/sdj1
         btrfs devs = /dev/sdf
[osd.4]
         host = store2
         osd journal = /dev/sdg1
         btrfs devs = /dev/sdc
[osd.5]
         host = store2
         osd journal = /dev/sdh1
         btrfs devs = /dev/sdd
[osd.6]
         host = store2
         osd journal = /dev/sdi1
         btrfs devs = /dev/sde
[osd.7]
         host = store2
         osd journal = /dev/sdj1
         btrfs devs = /dev/sdf
[osd.8]
         host = store3
         osd journal = /dev/sdg1
         btrfs devs = /dev/sdc
[osd.9]
         host = store3
         osd journal = /dev/sdh1
         btrfs devs = /dev/sdd
[osd.10]
         host = store3
         osd journal = /dev/sdi1
         btrfs devs = /dev/sde
[osd.11]
         host = store3
         osd journal = /dev/sdj1
         btrfs devs = /dev/sdf
[osd.12]
         host = store4
         osd journal = /dev/sdg1
         btrfs devs = /dev/sdc
[osd.13]
         host = store4
         osd journal = /dev/sdh1
         btrfs devs = /dev/sdd
[osd.14]
         host = store4
         osd journal = /dev/sdi1
         btrfs devs = /dev/sde
[osd.15]
         host = store4
         osd journal = /dev/sdj1
         btrfs devs = /dev/sdf
[osd.16]
         host = store5
         osd journal = /dev/sdg1
         btrfs devs = /dev/sdc
[osd.17]
         host = store5
         osd journal = /dev/sdh1
         btrfs devs = /dev/sdd
[osd.18]
         host = store5
         osd journal = /dev/sdi1
         btrfs devs = /dev/sde
[osd.19]
         host = store5
         osd journal = /dev/sdj1
         btrfs devs = /dev/sdf
[osd.20]
         host = store6
         osd journal = /dev/sdg1
         btrfs devs = /dev/sdc
[osd.21]
         host = store6
         osd journal = /dev/sdh1
         btrfs devs = /dev/sdd
[osd.22]
         host = store6
         osd journal = /dev/sdi1
         btrfs devs = /dev/sde
[osd.23]
         host = store6
         osd journal = /dev/sdj1
         btrfs devs = /dev/sdf


On 28.03.2013 19:01, Gregory Farnum wrote:
Your crush map looks fine to me. I'm saying that your ceph -s output
showed the OSD still hadn't been marked out. No data will be migrated
until it's marked out.
After ten minutes it should have been marked out, but that's based on
a number of factors you have some control over. If you just want a
quick check of your crush map you can mark it out manually, too.
-Greg
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com






_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux