Re: Activating OSDs takes forever

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



After activating OSDs, I tried to use ceph-deploy to copy the configuration file and admin ket to my admin node and my ceph nodes so that I can use ceph CLI without having to specify the monitor address and ceph.client.admin.keyring each time I execute a command but I got the following error:

 

[ceph@ceph-admin-node-centos-6-4 mycluster]$ ceph-deploy admin ceph-admin-node-centos-6-4 ceph-node1-mon-centos-6-4 ceph-node2-osd0-centos-6-4 ceph-node3-osd1-centos-6-4

[ceph_deploy.cli][INFO  ] Invoked (1.2.7): /usr/bin/ceph-deploy admin ceph-admin-node-centos-6-4 ceph-node1-mon-centos-6-4 ceph-node2-osd0-centos-6-4 ceph-node3-osd1-centos-6-4

[ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to ceph-admin-node-centos-6-4

[ceph_deploy.sudo_pushy][DEBUG ] will use a local connection with sudo

Traceback (most recent call last):

  File "/usr/bin/ceph-deploy", line 21, in <module>

    sys.exit(main())

  File "/usr/lib/python2.6/site-packages/ceph_deploy/util/decorators.py", line 83, in newfunc

    return f(*a, **kw)

  File "/usr/lib/python2.6/site-packages/ceph_deploy/cli.py", line 150, in main

    return args.func(args)

  File "/usr/lib/python2.6/site-packages/ceph_deploy/admin.py", line 40, in admin

    overwrite=args.overwrite_conf,

  File "/usr/lib/python2.6/site-packages/pushy/protocol/proxy.py", line 255, in <lambda>

    (conn.operator(type_, self, args, kwargs))

  File "/usr/lib/python2.6/site-packages/pushy/protocol/connection.py", line 66, in operator

    return self.send_request(type_, (object, args, kwargs))

  File "/usr/lib/python2.6/site-packages/pushy/protocol/baseconnection.py", line 329, in send_request

    return self.__handle(m)

  File "/usr/lib/python2.6/site-packages/pushy/protocol/baseconnection.py", line 645, in __handle

    raise e

pushy.protocol.proxy.ExceptionProxy: [Errno 2] No such file or directory: '/etc/ceph/ceph.conf.16463.tmp'

 

Has anyone seen anything like this?

 

 

 

From: Alfredo Deza [mailto:alfredo.deza@xxxxxxxxxxx]
Sent: Thursday, October 31, 2013 8:19 AM
To: Karan Singh
Cc: alistair whittle; Trivedi, Narendra; ceph-users@xxxxxxxxxxxxxx
Subject: Re: Activating OSDs takes forever

 

 

 

On Thu, Oct 31, 2013 at 8:07 AM, Karan Singh <ksingh@xxxxxx> wrote:

You are right AW , i also faced this very same error message apparently everything just works fine till now   no issues reported yet.

 

Regards

Karan Singh

 


From: "alistair whittle" <alistair.whittle@xxxxxxxxxxxx>
To: "Narendra Trivedi" <Narendra.Trivedi@xxxxxxxxxx>, ceph-users@xxxxxxxxxxxxxx
Sent: Thursday, 31 October, 2013 12:06:52 PM
Subject: Re: Activating OSDs takes forever

 

 

I got these exact same error messages on my RHEL build.   It did however seem that my OSD’s were correctly built and brought online.   Never got to the bottom of these errors though.

 

There were others on this forum that had similar issues, and theirs too seemed to be working successfully despite the apparent problems during activation.

 

From: ceph-users-bounces@xxxxxxxxxxxxxx [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Trivedi, Narendra
Sent: Wednesday, October 30, 2013 8:58 PM
To: ceph-users@xxxxxxxxxxxxxx
Subject: Activating OSDs takes forever

 

Hi All,

 

I had a pretty good run until I issued a command to activate OSDs. Now I am back with some more problemsL. My setup is exactly like the one in the official ceph documentation:

 

http://ceph.com/docs/master/start/quick-ceph-deploy/

 

That means, I am just using node2:/tmp/osd0 and node3:/tmp/osd1 as OSDs. I am running 64-bit CentOS-6-4.  

 

I was able to prepare OSDs fine but when I am activating them it just takes forver (15 mins and counting):

 

[ceph@ceph-admin-node-centos-6-4 mycluster]$ ceph-deploy osd activate ceph-node2-osd0-centos-6-4:/tmp/osd0 ceph-node3-osd1-centos-6-4:/tmp/osd1

[ceph_deploy.cli][INFO  ] Invoked (1.2.7): /usr/bin/ceph-deploy osd activate ceph-node2-osd0-centos-6-4:/tmp/osd0 ceph-node3-osd1-centos-6-4:/tmp/osd1

[ceph_deploy.osd][DEBUG ] Activating cluster ceph disks ceph-node2-osd0-centos-6-4:/tmp/osd0: ceph-node3-osd1-centos-6-4:/tmp/osd1:

[ceph_deploy.sudo_pushy][DEBUG ] will use a remote connection with sudo

[ceph_deploy.osd][INFO  ] Distro info: CentOS 6.4 Final

[ceph_deploy.osd][DEBUG ] activating host ceph-node2-osd0-centos-6-4 disk /tmp/osd0

[ceph_deploy.osd][DEBUG ] will use init type: sysvinit

[ceph-node2-osd0-centos-6-4][INFO  ] Running command: ceph-disk-activate --mark-init sysvinit --mount /tmp/osd0

[ceph-node2-osd0-centos-6-4][INFO  ] === osd.0 ===

[ceph-node2-osd0-centos-6-4][INFO  ] Starting Ceph osd.0 on ceph-node2-osd0-centos-6-4...

[ceph-node2-osd0-centos-6-4][INFO  ] starting osd.0 at :/0 osd_data /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal

[ceph-node2-osd0-centos-6-4][ERROR ] got latest monmap

[ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.059954 7faffbd537a0 -1 journal FileJournal::_open: disabling aio for non-block journal.  Use journal_force_aio to force use of aio anyway

[ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.121705 7faffbd537a0 -1 journal FileJournal::_open: disabling aio for non-block journal.  Use journal_force_aio to force use of aio anyway

[ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.122260 7faffbd537a0 -1 filestore(/tmp/osd0) could not find 23c2fcde/osd_superblock/0//-1 in index: (2) No such file or directory

[ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.204635 7faffbd537a0 -1 created object store /tmp/osd0 journal /tmp/osd0/journal for osd.0 fsid ec8c48d5-3889-433b-bf68-558f2eb39a8c

[ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.204714 7faffbd537a0 -1 auth: error reading file: /tmp/osd0/keyring: can't open /tmp/osd0/keyring: (2) No such file or directory

[ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.204938 7faffbd537a0 -1 created new key in keyring /tmp/osd0/keyring

[ceph-node2-osd0-centos-6-4][ERROR ] added key for osd.0

[ceph-node2-osd0-centos-6-4][ERROR ] create-or-move updating item name 'osd.0' weight 0 at location {host=ceph-node2-osd0-centos-6-4,root=default} to crush map

 

Why is ceph-deploy throwing bunch of errors when all I did was to create directories under /tmp on both nodes and issue a command to prepare them without any problems. I mean I can always delete the /tmp/osd0-1 and re-issue the commands again but I am wondering if anyone else has been this before?

 

 

The logging here is misleading because what is happening is that ceph-deploy interprets "remote stderr output" as ERROR logging and "remote stdout output" as DEBUG. So if the remote host
is sending output to `stderr` it will show as ERROR in ceph-deploy's logging.

Unfortunately there are tools (and libraries!) that will default to push all logging to `stderr`. There is little ceph-deploy can do here because it can't interpret if the output is _really_ an error or is it
just logging something that is not really an error but somehow it is going to `stderr`.

I will make sure to add a not in the ceph-deploy docs about this so it is a bit more clear.

 

Thanks a lot in advance!

Narendra  

 

 


This message contains information which may be confidential and/or privileged. Unless you are the intended recipient (or authorized to receive for the intended recipient), you may not read, use, copy or disclose to anyone the message or any information contained in the message. If you have received the message in error, please advise the sender by reply e-mail and delete the message and any attachment(s) thereto without retaining any copies.

_______________________________________________

This message is for information purposes only, it is not a recommendation, advice, offer or solicitation to buy or sell a product or service nor an official confirmation of any transaction. It is directed at persons who are professionals and is not intended for retail customer use. Intended for recipient only. This message is subject to the terms at: www.barclays.com/emaildisclaimer.

For important disclosures, please see: www.barclays.com/salesandtradingdisclaimer regarding market commentary from Barclays Sales and/or Trading, who are active market participants; and in respect of Barclays Research, including disclosures relating to specific issuers, please see http://publicresearch.barclays.com.

_______________________________________________

 

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

 


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

 


This message contains information which may be confidential and/or privileged. Unless you are the intended recipient (or authorized to receive for the intended recipient), you may not read, use, copy or disclose to anyone the message or any information contained in the message. If you have received the message in error, please advise the sender by reply e-mail and delete the message and any attachment(s) thereto without retaining any copies.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux