thanks, running as root does give me status, but not clean.
root@xxxxxxxxxxx:~# ceph status
cluster 9059dfad-924a-425c-a20b-17dc1d53111e
health HEALTH_WARN 91 pgs degraded; 192 pgs stuck unclean; recovery 21/42 degraded (50.000%)
monmap e1: 1 mons at {jr-ceph2=10.88.26.55:6789/0}, election epoch 2, quorum 0 jr-ceph2
osdmap e10: 2 osds: 2 up, 2 in
pgmap v2715: 192 pgs: 101 active+remapped, 91 active+degraded; 9518 bytes data, 9148 MB used, 362 GB / 391 GB avail; 21/42 degraded (50.000%)
mdsmap e4: 1/1/1 up {0=jr-ceph2.XXX=up:active}
don't see anything telling in the ceph logs; Should I wait for the new quickstart?
On Mon, Sep 16, 2013 at 2:27 PM, John Wilkins <john.wilkins@xxxxxxxxxxx> wrote:
We will have a new update to the quick start this week.
--
On Mon, Sep 16, 2013 at 12:18 PM, Alfredo Deza <alfredo.deza@xxxxxxxxxxx> wrote:
> On Mon, Sep 16, 2013 at 12:58 PM, Justin Ryan <justin.ryan@xxxxxxxxxx> wrote:
>> Hi,
>>
>> I'm brand new to Ceph, attempting to follow the Getting Started guide with 2
>> VMs. I completed the Preflight without issue. I completed Storage Cluster
>> Quick Start, but have some questions:
>>
>> The Single Node Quick Start grey box -- does 'single node' mean if you're
>> running the whole thing on a single machine, if you have only one server
>> node like the diagram at the top of the page, or if you're only running one
>> OSD process? I'm not sure if I need to make the `osd crush chooseleaf type`
>> change.
>>
>> Are the LIST, ZAP, and ADD OSDS ON STANDALONE DISKS sections an alternative
>> to the MULTIPLE OSDS ON THE OS DISK (DEMO ONLY) section? I thought I set up
>> my OSDs already on /tmp/osd{0,1}.
>>
>> Moving on to the Block Device Quick Start -- it says "To use this guide, you
>> must have executed the procedures in the Object Store Quick Start guide
>> first" -- but the link to the Object Store Quick Start actually points to
>> the Storage Cluster Quick Start -- which is it?
>>
>> Most importantly, it says "Ensure your Ceph Storage Cluster is in an active
>> + clean state before working with the Ceph Block Device" --- how can tell if
>> my cluster is active+clean?? The only ceph* command on the admin node is
>> ceph-deploy, and running `ceph` on the server node:
>>
>> ceph@jr-ceph2:~$ ceph
>> 2013-09-16 16:53:10.880267 7feb96c1b700 -1 monclient(hunting): ERROR:
>> missing keyring, cannot use cephx for authentication
>> 2013-09-16 16:53:10.880271 7feb96c1b700 0 librados: client.admin
>> initialization error (2) No such file or directory
>> Error connecting to cluster: ObjectNotFound
>
> There is a ticket open for this, but you basically need super-user
> permissions here to run (any?) ceph commands.
>>
>> Thanks in advance for any help, and apologies if I missed anything obvious.
>>
>>
>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
John Wilkins
Senior Technical Writer
Intank
john.wilkins@xxxxxxxxxxx
(415) 425-9599
http://inktank.com
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com