firefly osds stuck in state booting

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Karan ,

Thanks .. that did the trick ..
The magic word was "in"
regarding rep size . I have adjusted them
my settings are
--snip--
osd pool default size =
2

osd pool default min size =
1

osd pool default pg num =
100

osd pool default pgp num = 100
--snip--

# Also in the meantime I had chance to play with ceph-deploy script too.
# Maybe it was me or probably it is a bug . I have tried twice and
everytime I have hit this

As I said before I'm using a directory as this is a test installation .

ceph-deploy osd prepare ceph2:/ceph2:/ceph2/journald <=== Works


but

--snip--
ceph-deploy osd activate
ceph2:/ceph2:/ceph2/journald

[ceph_deploy.conf][DEBUG ] found configuration file at:
/home/ceph/.cephdeploy.conf
[ceph_deploy.cli][INFO  ] Invoked (1.5.9): /usr/bin/ceph-deploy osd
activate ceph2:/ceph2:/ceph2/journald
[ceph_deploy.osd][DEBUG ] Activating cluster ceph disks
ceph2:/ceph2:/ceph2/journald
[ceph2][DEBUG ] connected to host: ceph2
[ceph2][DEBUG ] detect platform information from remote host
[ceph2][DEBUG ] detect machine type
[ceph_deploy.osd][INFO  ] Distro info: CentOS 6.5 Final
[ceph_deploy.osd][DEBUG ] activating host ceph2 disk /ceph2
[ceph_deploy.osd][DEBUG ] will use init type: sysvinit
[ceph2][INFO  ] Running command: sudo ceph-disk-activate --mark-init
sysvinit --mount /ceph2
[ceph2][WARNIN] got monmap epoch 2
[ceph2][WARNIN] 2014-07-28 11:47:04.733204 7f08d1c667a0 -1 journal
FileJournal::_open: disabling aio for non-block journal.  Use
journal_force_aio to force
 use of aio anyway
[ceph2][WARNIN] 2014-07-28 11:47:04.733400 7f08d1c667a0 -1 journal check:
ondisk fsid 00000000-0000-0000-0000-000000000000 doesn't match expected
4795daff-
d63f-415b-9824-75f0863eb14f, invalid (someone else's?) journal
[ceph2][WARNIN] 2014-07-28 11:47:04.796835 7f08d1c667a0 -1 journal
FileJournal::_open: disabling aio for non-block journal.  Use
journal_force_aio to force
 use of aio anyway
[ceph2][WARNIN] 2014-07-28 11:47:04.798944 7f08d1c667a0 -1
filestore(/ceph2) could not find 23c2fcde/osd_superblock/0//-1 in index:
(2) No such file or dir
ectory
[ceph2][WARNIN] 2014-07-28 11:47:04.874282 7f08d1c667a0 -1 created object
store /ceph2 journal /ceph2/journal for osd.1 fsid
109507ab-adf1-4eb6-aacf-092549
4e3882
[ceph2][WARNIN] 2014-07-28 11:47:04.874474 7f08d1c667a0 -1 auth: error
reading file: /ceph2/keyring: can't open /ceph2/keyring: (2) No such file
or directo
ry
[ceph2][WARNIN] 2014-07-28 11:47:04.875209 7f08d1c667a0 -1 created new key
in keyring /ceph2/keyring
[ceph2][WARNIN] added key for osd.1
[ceph2][WARNIN] ceph-disk: Error: unable to create symlink
/var/lib/ceph/osd/ceph-1 -> /ceph2
[ceph2][ERROR ] RuntimeError: command returned non-zero exit status: 1
[ceph_deploy][ERROR ] RuntimeError: Failed to execute command:
ceph-disk-activate --mark-init sysvinit --mount /ceph2
--snip--



It turns out ceph-deploy does not create the directory /var/lib/ceph/osd
and if I create them everything works.


Cheers


On Mon, Jul 28, 2014 at 9:09 AM, Karan Singh <karan.singh at csc.fi> wrote:

> The output that you have provided says that OSDs are not IN , Try the below
>
> ceph osd in osd.0
> ceph osd in osd.1
>
> service ceph start osd.0
> service ceph start osd.1
>
> If you have 1 more host with 1 disk , add it , starting Ceph Firefly
> default rep size is 3
>
>
> - Karan -
>
> On 27 Jul 2014, at 11:17, 10 minus <t10tennn at gmail.com> wrote:
>
> Hi Sage,
>
> I have dropped all unset .. and even restarted the osd
> No dice .. OSDs are still stuck .
>
>
>
> --snip--
>  ceph daemon osd.0 status                                            ?rtt
> min/avg/max/mdev = 0.095/0.120/0.236/0.015
> ms
>
> { "cluster_fsid":
> "99babb8f-c880-4b32-a227-94aa483d4871",
> ?[root at ceph2 ~]#  ceph daemon osd.1
> status
>
>   "osd_fsid":
> "1ad28bde-c23c-44ba-a3b7-0aaaafd3372e",                                ?{
> "cluster_fsid":
> "99babb8f-c880-4b32-a227-94aa483d4871",
>
>   "whoami":
> 0,                                                                       ?
> "osd_fsid":
> "becc3252-6977-47d6-87af-7b1337e591d8",
>
>   "state":
> "booting",
> ?  "whoami":
> 1,
>
>   "oldest_map":
> 1,                                                                   ?
> "state":
> "booting",
>
>   "newest_map":
> 24,                                                                  ?
> "oldest_map":
> 1,
>
>   "num_pgs":
> 0}                                                                      ?
> "newest_map":
> 21,
>
>  --snip--
>
> --snip--
> ceph osd
> tree
>
> # id    weight  type name       up/down reweight
> -1      2       root default
> -3      1               host ceph1
> 0       1                       osd.0   down    0
> -2      1               host ceph2
> 1       1                       osd.1   down    0
>
>  --snip--
>
> --snip--
>  ceph -s
>     cluster 2929fa80-0841-4cb6-a133-90b2098fc802
>      health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean
>      monmap e2: 3 mons at {ceph0=
> 10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0},
> election epoch 50, quorum 0,1,2 ceph0,ceph1,ceph2
>      osdmap e24: 2 osds: 0 up, 0 in
>       pgmap v25: 192 pgs, 3 pools, 0 bytes data, 0 objects
>             0 kB used, 0 kB / 0 kB avail
>                  192 creating
> --snip--
>
>
>
>
> On Sat, Jul 26, 2014 at 5:57 PM, Sage Weil <sweil at redhat.com> wrote:
>
>> On Sat, 26 Jul 2014, 10 minus wrote:
>> > Hi,
>> >
>> > I just setup a test ceph installation on 3 node Centos 6.5  .
>> > two of the nodes are used for hosting osds and the third acts as mon .
>> >
>> > Please note I'm using LVM so had to set up the osd using the manual
>> install
>> > guide.
>> >
>> > --snip--
>> > ceph -s
>> >     cluster 2929fa80-0841-4cb6-a133-90b2098fc802
>> >      health HEALTH_WARN 192 pgs stuck inactive; 192 pgs stuck unclean;
>> > noup,nodown,noout flag(s) set
>> >      monmap e2: 3 mons at{ceph0=
>> 10.0.12.220:6789/0,ceph1=10.0.12.221:6789/0,ceph2=10.0.12.222:6789/0
>> > }, election epoch 46, quorum 0,1,2 ceph0,ceph1,ceph2
>> >      osdmap e21: 2 osds: 0 up, 0 in
>> >             flags noup,nodown,noout
>>                     ^^^^
>>
>> Do 'ceph osd unset noup' and they should start up.  You likely also want
>> to clear nodown and noout as well.
>>
>> sage
>>
>>
>> >       pgmap v22: 192 pgs, 3 pools, 0 bytes data, 0 objects
>> >             0 kB used, 0 kB / 0 kB avail
>> >                  192 creating
>> > --snip--
>> >
>> > osd tree
>> >
>> > --snip--
>> > ceph osd tree
>> > # id    weight  type name       up/down reweight
>> > -1      2       root default
>> > -3      1               host ceph1
>> > 0       1                       osd.0   down    0
>> > -2      1               host ceph2
>> > 1       1                       osd.1   down    0
>> > --snip--
>> >
>> > --snip--
>> >  ceph daemon osd.0 status
>> > { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
>> >   "osd_fsid": "1ad28bde-c23c-44ba-a3b7-0aaaafd3372e",
>> >   "whoami": 0,
>> >   "state": "booting",
>> >   "oldest_map": 1,
>> >   "newest_map": 21,
>> >   "num_pgs": 0}
>> >
>> > --snip--
>> >
>> > --snip--
>> >  ceph daemon osd.1 status
>> > { "cluster_fsid": "99babb8f-c880-4b32-a227-94aa483d4871",
>> >   "osd_fsid": "becc3252-6977-47d6-87af-7b1337e591d8",
>> >   "whoami": 1,
>> >   "state": "booting",
>> >   "oldest_map": 1,
>> >   "newest_map": 21,
>> >   "num_pgs": 0}
>> > --snip--
>> >
>> > # Cpus are idling
>> >
>> > # does anybody know what is wrong
>> >
>> > Thanks in advance
>> >
>> >
>>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140729/934774e3/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux