Re: Activating OSDs takes forever

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, Oct 31, 2013 at 12:01 PM, Trivedi, Narendra
<Narendra.Trivedi@xxxxxxxxxx> wrote:
> After activating OSDs, I tried to use ceph-deploy to copy the configuration
> file and admin ket to my admin node and my ceph nodes so that I can use ceph
> CLI without having to specify the monitor address and
> ceph.client.admin.keyring each time I execute a command but I got the
> following error:
>
>
>
> [ceph@ceph-admin-node-centos-6-4 mycluster]$ ceph-deploy admin
> ceph-admin-node-centos-6-4 ceph-node1-mon-centos-6-4
> ceph-node2-osd0-centos-6-4 ceph-node3-osd1-centos-6-4
>
> [ceph_deploy.cli][INFO  ] Invoked (1.2.7): /usr/bin/ceph-deploy admin
> ceph-admin-node-centos-6-4 ceph-node1-mon-centos-6-4
> ceph-node2-osd0-centos-6-4 ceph-node3-osd1-centos-6-4
>
> [ceph_deploy.admin][DEBUG ] Pushing admin keys and conf to
> ceph-admin-node-centos-6-4
>
> [ceph_deploy.sudo_pushy][DEBUG ] will use a local connection with sudo
>
> Traceback (most recent call last):
>
>   File "/usr/bin/ceph-deploy", line 21, in <module>
>
>     sys.exit(main())
>
>   File "/usr/lib/python2.6/site-packages/ceph_deploy/util/decorators.py",
> line 83, in newfunc
>
>     return f(*a, **kw)
>
>   File "/usr/lib/python2.6/site-packages/ceph_deploy/cli.py", line 150, in
> main
>
>     return args.func(args)
>
>   File "/usr/lib/python2.6/site-packages/ceph_deploy/admin.py", line 40, in
> admin
>
>     overwrite=args.overwrite_conf,
>
>   File "/usr/lib/python2.6/site-packages/pushy/protocol/proxy.py", line 255,
> in <lambda>
>
>     (conn.operator(type_, self, args, kwargs))
>
>   File "/usr/lib/python2.6/site-packages/pushy/protocol/connection.py", line
> 66, in operator
>
>     return self.send_request(type_, (object, args, kwargs))
>
>   File "/usr/lib/python2.6/site-packages/pushy/protocol/baseconnection.py",
> line 329, in send_request
>
>     return self.__handle(m)
>
>   File "/usr/lib/python2.6/site-packages/pushy/protocol/baseconnection.py",
> line 645, in __handle
>
>     raise e
>
> pushy.protocol.proxy.ExceptionProxy: [Errno 2] No such file or directory:
> '/etc/ceph/ceph.conf.16463.tmp'
>
>
>
> Has anyone seen anything like this?

Yes, this should be fixed in the upcoming ceph-deploy version. It
happens when you attempt to run ceph-deploy on a node that
doesn't have ceph installed. This happens because the `/etc/ceph`
directory doesn't exist, which is created when ceph is installed.

If ceph was installed and this is still happening is because purge or
purgedata removed that directory. The current workaround is to `mkdir
/etc/ceph`
if ceph is in fact installed.
>
>
>
>
>
>
>
> From: Alfredo Deza [mailto:alfredo.deza@xxxxxxxxxxx]
> Sent: Thursday, October 31, 2013 8:19 AM
> To: Karan Singh
> Cc: alistair whittle; Trivedi, Narendra; ceph-users@xxxxxxxxxxxxxx
>
>
> Subject: Re:  Activating OSDs takes forever
>
>
>
>
>
>
>
> On Thu, Oct 31, 2013 at 8:07 AM, Karan Singh <ksingh@xxxxxx> wrote:
>
> You are right AW , i also faced this very same error message apparently
> everything just works fine till now   no issues reported yet.
>
>
>
> Regards
>
> Karan Singh
>
>
>
> ________________________________
>
> From: "alistair whittle" <alistair.whittle@xxxxxxxxxxxx>
> To: "Narendra Trivedi" <Narendra.Trivedi@xxxxxxxxxx>,
> ceph-users@xxxxxxxxxxxxxx
> Sent: Thursday, 31 October, 2013 12:06:52 PM
> Subject: Re:  Activating OSDs takes forever
>
>
>
>
>
> I got these exact same error messages on my RHEL build.   It did however
> seem that my OSD’s were correctly built and brought online.   Never got to
> the bottom of these errors though.
>
>
>
> There were others on this forum that had similar issues, and theirs too
> seemed to be working successfully despite the apparent problems during
> activation.
>
>
>
> From: ceph-users-bounces@xxxxxxxxxxxxxx
> [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Trivedi, Narendra
> Sent: Wednesday, October 30, 2013 8:58 PM
> To: ceph-users@xxxxxxxxxxxxxx
> Subject:  Activating OSDs takes forever
>
>
>
> Hi All,
>
>
>
> I had a pretty good run until I issued a command to activate OSDs. Now I am
> back with some more problemsL. My setup is exactly like the one in the
> official ceph documentation:
>
>
>
> http://ceph.com/docs/master/start/quick-ceph-deploy/
>
>
>
> That means, I am just using node2:/tmp/osd0 and node3:/tmp/osd1 as OSDs. I
> am running 64-bit CentOS-6-4.
>
>
>
> I was able to prepare OSDs fine but when I am activating them it just takes
> forver (15 mins and counting):
>
>
>
> [ceph@ceph-admin-node-centos-6-4 mycluster]$ ceph-deploy osd activate
> ceph-node2-osd0-centos-6-4:/tmp/osd0 ceph-node3-osd1-centos-6-4:/tmp/osd1
>
> [ceph_deploy.cli][INFO  ] Invoked (1.2.7): /usr/bin/ceph-deploy osd activate
> ceph-node2-osd0-centos-6-4:/tmp/osd0 ceph-node3-osd1-centos-6-4:/tmp/osd1
>
> [ceph_deploy.osd][DEBUG ] Activating cluster ceph disks
> ceph-node2-osd0-centos-6-4:/tmp/osd0: ceph-node3-osd1-centos-6-4:/tmp/osd1:
>
> [ceph_deploy.sudo_pushy][DEBUG ] will use a remote connection with sudo
>
> [ceph_deploy.osd][INFO  ] Distro info: CentOS 6.4 Final
>
> [ceph_deploy.osd][DEBUG ] activating host ceph-node2-osd0-centos-6-4 disk
> /tmp/osd0
>
> [ceph_deploy.osd][DEBUG ] will use init type: sysvinit
>
> [ceph-node2-osd0-centos-6-4][INFO  ] Running command: ceph-disk-activate
> --mark-init sysvinit --mount /tmp/osd0
>
> [ceph-node2-osd0-centos-6-4][INFO  ] === osd.0 ===
>
> [ceph-node2-osd0-centos-6-4][INFO  ] Starting Ceph osd.0 on
> ceph-node2-osd0-centos-6-4...
>
> [ceph-node2-osd0-centos-6-4][INFO  ] starting osd.0 at :/0 osd_data
> /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal
>
> [ceph-node2-osd0-centos-6-4][ERROR ] got latest monmap
>
> [ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.059954 7faffbd537a0
> -1 journal FileJournal::_open: disabling aio for non-block journal.  Use
> journal_force_aio to force use of aio anyway
>
> [ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.121705 7faffbd537a0
> -1 journal FileJournal::_open: disabling aio for non-block journal.  Use
> journal_force_aio to force use of aio anyway
>
> [ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.122260 7faffbd537a0
> -1 filestore(/tmp/osd0) could not find 23c2fcde/osd_superblock/0//-1 in
> index: (2) No such file or directory
>
> [ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.204635 7faffbd537a0
> -1 created object store /tmp/osd0 journal /tmp/osd0/journal for osd.0 fsid
> ec8c48d5-3889-433b-bf68-558f2eb39a8c
>
> [ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.204714 7faffbd537a0
> -1 auth: error reading file: /tmp/osd0/keyring: can't open
> /tmp/osd0/keyring: (2) No such file or directory
>
> [ceph-node2-osd0-centos-6-4][ERROR ] 2013-10-30 15:34:54.204938 7faffbd537a0
> -1 created new key in keyring /tmp/osd0/keyring
>
> [ceph-node2-osd0-centos-6-4][ERROR ] added key for osd.0
>
> [ceph-node2-osd0-centos-6-4][ERROR ] create-or-move updating item name
> 'osd.0' weight 0 at location {host=ceph-node2-osd0-centos-6-4,root=default}
> to crush map
>
>
>
> Why is ceph-deploy throwing bunch of errors when all I did was to create
> directories under /tmp on both nodes and issue a command to prepare them
> without any problems. I mean I can always delete the /tmp/osd0-1 and
> re-issue the commands again but I am wondering if anyone else has been this
> before?
>
>
>
>
>
> The logging here is misleading because what is happening is that ceph-deploy
> interprets "remote stderr output" as ERROR logging and "remote stdout
> output" as DEBUG. So if the remote host
> is sending output to `stderr` it will show as ERROR in ceph-deploy's
> logging.
>
> Unfortunately there are tools (and libraries!) that will default to push all
> logging to `stderr`. There is little ceph-deploy can do here because it
> can't interpret if the output is _really_ an error or is it
> just logging something that is not really an error but somehow it is going
> to `stderr`.
>
> I will make sure to add a not in the ceph-deploy docs about this so it is a
> bit more clear.
>
>
>
> Thanks a lot in advance!
>
> Narendra
>
>
>
>
>
>
> This message contains information which may be confidential and/or
> privileged. Unless you are the intended recipient (or authorized to receive
> for the intended recipient), you may not read, use, copy or disclose to
> anyone the message or any information contained in the message. If you have
> received the message in error, please advise the sender by reply e-mail and
> delete the message and any attachment(s) thereto without retaining any
> copies.
>
> _______________________________________________
>
> This message is for information purposes only, it is not a recommendation,
> advice, offer or solicitation to buy or sell a product or service nor an
> official confirmation of any transaction. It is directed at persons who are
> professionals and is not intended for retail customer use. Intended for
> recipient only. This message is subject to the terms at:
> www.barclays.com/emaildisclaimer.
>
> For important disclosures, please see:
> www.barclays.com/salesandtradingdisclaimer regarding market commentary from
> Barclays Sales and/or Trading, who are active market participants; and in
> respect of Barclays Research, including disclosures relating to specific
> issuers, please see http://publicresearch.barclays.com.
>
> _______________________________________________
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
>
> This message contains information which may be confidential and/or
> privileged. Unless you are the intended recipient (or authorized to receive
> for the intended recipient), you may not read, use, copy or disclose to
> anyone the message or any information contained in the message. If you have
> received the message in error, please advise the sender by reply e-mail and
> delete the message and any attachment(s) thereto without retaining any
> copies.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com





[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux