Persistent Error on osd activation

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



i have meet the same issue , when i want to use prepare .
when i use --zap-disk , it is ok.   but if use prepare to define journal
device, failed
ceph-disk-prepare --zap-disk --fs-type btrfs --cluster ceph -- /dev/sdb
/dev/sdc



2014-07-01 1:00 GMT+07:00 Iban Cabrillo <cabrillo at ifca.unican.es>:

> Hi Alfredo,
>   During this morning, I have purged all the deployment.
>   I just prepared 4 SAN Servers with 4 FC-Atacched disk (2.7 TB per disk)
> each one of them.
>
>   Tomorrow I will try to deploy anew installation leaving the VMs machines
> as mons and the OSDs with this physical servers.
>
>   The local disk Is SAS RAID1, Should I put the journal over the local
> disk (SAS RAID1), or should be a better solution used the RAID5 nlSAS disk
> FC-attached (journal + data on the same disk)?
>
>   Which would be the recommended size for the journal? 10GB per disk, for
> example?
>
> Regards, I
>
>
> 2014-06-30 18:50 GMT+02:00 Alfredo Deza <alfredo.deza at inktank.com>:
>
> On Mon, Jun 30, 2014 at 11:22 AM, Iban Cabrillo <cabrillo at ifca.unican.es>
>> wrote:
>> > Hi Alfredo and folk,
>> >   Could you have a look at this?
>> >   Someone else has any idea why i am getting this error?
>> >
>> > Thanks in advance, I
>> >
>> >
>> >
>> > 2014-06-27 16:37 GMT+02:00 Iban Cabrillo <cabrillo at ifca.unican.es>:
>> >
>> >> Hi Alfredo,
>> >>  This is the complete procedure:
>> >>
>> >>
>> >>   On OSD node:
>> >>
>> >> [ceph at ceph02 ~]$ sudo parted /dev/xvdb
>> >>
>> >> GNU Parted 2.1
>> >> Using /dev/xvdb
>> >> Welcome to GNU Parted! Type 'help' to view a list of commands.
>> >> (parted) p
>> >> Model: Xen Virtual Block Device (xvd)
>> >> Disk /dev/xvdb: 107GB
>> >> Sector size (logical/physical): 512B/512B
>> >> Partition Table: gpt
>> >>
>> >> Number  Start  End  Size  File system  Name  Flags
>> >>
>> >> [ceph at ceph02 ~]$ sudo ls -la /var/lib/ceph/tmp/
>> >> total 8
>> >> drwxr-xr-x 2 root root 4096 Jun 27 16:30 .
>> >> drwxr-xr-x 7 root root 4096 Jun 26 22:30 ..
>> >> [ceph at ceph02 ~]$ sudo ls -la /var/lib/ceph/osd/
>> >> total 8
>> >> drwxr-xr-x 2 root root 4096 Jun 27 12:14 .
>> >> drwxr-xr-x 7 root root 4096 Jun 26 22:30 ..
>> >>
>> >> On ceph admin node:
>> >>
>> >> [ceph at cephadm ~]$ sudo ceph osd tree
>> >> # id    weight    type name    up/down    reweight
>> >> -1    0.14    root default
>> >> -2    0.009995        host ceph02
>> >> 1    0.009995            osd.1    DNE
>> >> -3    0.03999        host ceph04
>> >> 4    0.03999            osd.4    up    1
>> >> -4    0.09        host ceph03
>> >> 6    0.09            osd.6    up    1
>> >>
>> >>
>> >> [ceph at cephadm ceph-cloud]$ ceph-deploy osd prepare ceph02:xvdb
>> >> [ceph_deploy.conf][DEBUG ] found configuration file at:
>> >> /home/ceph/.cephdeploy.conf
>> >> [ceph_deploy.cli][INFO  ] Invoked (1.5.5): /usr/bin/ceph-deploy osd
>> >> prepare ceph02:xvdb
>> >> [ceph_deploy.osd][DEBUG ] Preparing cluster ceph disks
>> ceph02:/dev/xvdb:
>> >> [ceph02][DEBUG ] connected to host: ceph02
>> >> [ceph02][DEBUG ] detect platform information from remote host
>> >> [ceph02][DEBUG ] detect machine type
>> >> [ceph_deploy.osd][INFO  ] Distro info: Scientific Linux 6.2 Carbon
>> >> [ceph_deploy.osd][DEBUG ] Deploying osd to ceph02
>> >> [ceph02][DEBUG ] write cluster configuration to
>> /etc/ceph/{cluster}.conf
>> >> [ceph02][INFO  ] Running command: sudo udevadm trigger
>> >> --subsystem-match=block --action=add
>> >> [ceph_deploy.osd][DEBUG ] Preparing host ceph02 disk /dev/xvdb journal
>> >> None activate False
>> >> [ceph02][INFO  ] Running command: sudo ceph-disk-prepare --fs-type xfs
>> >> --cluster ceph -- /dev/xvdb
>> >> [ceph02][DEBUG ] Setting name!
>> >> [ceph02][DEBUG ] partNum is 1
>> >> [ceph02][DEBUG ] REALLY setting name!
>> >> [ceph02][DEBUG ] The operation has completed successfully.
>> >> [ceph02][DEBUG ] Setting name!
>> >> [ceph02][DEBUG ] partNum is 0
>> >> [ceph02][DEBUG ] REALLY setting name!
>> >> [ceph02][DEBUG ] The operation has completed successfully.
>> >> [ceph02][DEBUG ] meta-data=/dev/xvdb1             isize=2048
>> agcount=4,
>> >> agsize=5897919 blks
>> >> [ceph02][DEBUG ]          =                       sectsz=512   attr=2
>> >> [ceph02][DEBUG ] data     =                       bsize=4096
>> >> blocks=23591675, imaxpct=25
>> >> [ceph02][DEBUG ]          =                       sunit=0      swidth=0
>> >> blks
>> >> [ceph02][DEBUG ] naming   =version 2              bsize=4096
>> ascii-ci=0
>> >> [ceph02][DEBUG ] log      =internal log           bsize=4096
>> >> blocks=11519, version=2
>> >> [ceph02][DEBUG ]          =                       sectsz=512   sunit=0
>> >> blks, lazy-count=1
>> >> [ceph02][DEBUG ] realtime =none                   extsz=4096
>> blocks=0,
>> >> rtextents=0
>> >> [ceph02][DEBUG ] The operation has completed successfully.
>> >> [ceph02][WARNIN] INFO:ceph-disk:Will colocate journal with data on
>> >> /dev/xvdb
>> >> [ceph02][INFO  ] checking OSD status...
>> >> [ceph02][INFO  ] Running command: sudo ceph --cluster=ceph osd stat
>> >> --format=json
>> >> [ceph_deploy.osd][DEBUG ] Host ceph02 is now ready for osd use.
>> >>
>> >>
>> >>
>> >> If i make create instead of prepare do the same (create do not make the
>> >> trick prepare+activate )
>> >>
>> >>
>> >> In the OSD:
>> >>
>> >> [ceph at ceph02 ~]$ sudo parted /dev/xvdb
>> >>
>> >> GNU Parted 2.1
>> >> Using /dev/xvdb
>> >> Welcome to GNU Parted! Type 'help' to view a list of commands.
>> >> (parted) p
>> >> Model: Xen Virtual Block Device (xvd)
>> >> Disk /dev/xvdb: 107GB
>> >> Sector size (logical/physical): 512B/512B
>> >> Partition Table: gpt
>> >>
>> >> Number  Start   End     Size    File system  Name          Flags
>> >>  2      1049kB  10.7GB  10.7GB  xfs          ceph journal
>> >>  1      10.7GB  107GB   96.6GB  xfs          ceph data
>> >>
>> >> (parted) q
>> >>
>> >> [ceph at ceph02 ~]$ sudo ls -la /var/lib/ceph/osd/
>> >> total 8
>> >> drwxr-xr-x 2 root root 4096 Jun 27 12:14 .
>> >> drwxr-xr-x 7 root root 4096 Jun 26 22:30 ..
>> >> [ceph at ceph02 ~]$ sudo ls -la /var/lib/ceph/tmp/
>> >> total 8
>> >> drwxr-xr-x 2 root root 4096 Jun 27 16:32 .
>> >> drwxr-xr-x 7 root root 4096 Jun 26 22:30 ..
>> >> -rw-r--r-- 1 root root    0 Jun 27 16:32 ceph-disk.prepare.lock
>> >>
>> >> [ceph at ceph02 ~]$ sudo ceph-disk list
>> >> /dev/xvda1 other, ext4, mounted on /
>> >> /dev/xvdb :
>> >>  /dev/xvdb1 ceph data, prepared, cluster ceph, journal /dev/xvdb2
>> >>  /dev/xvdb2 ceph journal, for /dev/xvdb1
>> >>
>> >>
>> >> In the cephadm:
>> >>
>> >> [ceph at cephadm ceph-cloud]$ ceph-deploy osd activate ceph02:xvdb1
>> >> [ceph_deploy.conf][DEBUG ] found configuration file at:
>> >> /home/ceph/.cephdeploy.conf
>> >> [ceph_deploy.cli][INFO  ] Invoked (1.5.5): /usr/bin/ceph-deploy osd
>> >> activate ceph02:xvdb1
>> >> [ceph_deploy.osd][DEBUG ] Activating cluster ceph disks
>> ceph02:/dev/xvdb1:
>> >> [ceph02][DEBUG ] connected to host: ceph02
>> >> [ceph02][DEBUG ] detect platform information from remote host
>> >> [ceph02][DEBUG ] detect machine type
>> >> [ceph_deploy.osd][INFO  ] Distro info: Scientific Linux 6.2 Carbon
>> >> [ceph_deploy.osd][DEBUG ] activating host ceph02 disk /dev/xvdb1
>> >> [ceph_deploy.osd][DEBUG ] will use init type: sysvinit
>> >> [ceph02][INFO  ] Running command: sudo ceph-disk-activate --mark-init
>> >> sysvinit --mount /dev/xvdb1
>> >> [ceph02][WARNIN] got monmap epoch 2
>> >> [ceph02][WARNIN] 2014-06-27 16:35:49.948865 7f143254e7a0 -1
>> >> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) mkjournal error creating
>> journal on
>> >> /var/lib/ceph/tmp/mnt.9VZHpR/journal: (2) No such file or directory
>> >> [ceph02][WARNIN] 2014-06-27 16:35:49.948893 7f143254e7a0 -1 OSD::mkfs:
>> >> ObjectStore::mkfs failed with error -2
>> >> [ceph02][WARNIN] 2014-06-27 16:35:49.948957 7f143254e7a0 -1  ** ERROR:
>> >> error creating empty object store in /var/lib/ceph/tmp/mnt.9VZHpR: (2)
>> No
>> >> such file or directory
>> >>
>> >> [ceph02][WARNIN] ERROR:ceph-disk:Failed to activate
>> >> [ceph02][WARNIN] Traceback (most recent call last):
>> >> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 2579, in <module>
>> >> [ceph02][WARNIN]     main()
>> >> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 2557, in main
>> >> [ceph02][WARNIN]     args.func(args)
>> >> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 1910, in
>> main_activate
>> >> [ceph02][WARNIN]     init=args.mark_init,
>> >> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 1686, in
>> >> mount_activate
>> >> [ceph02][WARNIN]     (osd_id, cluster) = activate(path,
>> >> activate_key_template, init)
>> >> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 1849, in activate
>> >> [ceph02][WARNIN]     keyring=keyring,
>> >> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 1484, in mkfs
>> >> [ceph02][WARNIN]     '--keyring', os.path.join(path, 'keyring'),
>> >> [ceph02][WARNIN]   File "/usr/sbin/ceph-disk", line 303, in
>> >> command_check_call
>> >> [ceph02][WARNIN]     return subprocess.check_call(arguments)
>> >> [ceph02][WARNIN]   File "/usr/lib64/python2.6/subprocess.py", line
>> 505, in
>> >> check_call
>> >> [ceph02][WARNIN]     raise CalledProcessError(retcode, cmd)
>> >> [ceph02][WARNIN] subprocess.CalledProcessError: Command
>> >> '['/usr/bin/ceph-osd', '--cluster', 'ceph', '--mkfs', '--mkkey', '-i',
>> '0',
>> >> '--monmap', '/var/lib/ceph/tmp/mnt.9VZHpR/activate.monmap',
>> '--osd-data',
>> >> '/var/lib/ceph/tmp/mnt.9VZHpR', '--osd-journal',
>> >> '/var/lib/ceph/tmp/mnt.9VZHpR/journal', '--osd-uuid',
>> >> '5e93fa7c-b6f7-4684-981b-bf73254bd87a', '--keyring',
>> >> '/var/lib/ceph/tmp/mnt.9VZHpR/keyring']' returned non-zero exit status
>> 1
>> >> [ceph02][ERROR ] RuntimeError: command returned non-zero exit status: 1
>> >> [ceph_deploy][ERROR ] RuntimeError: Failed to execute command:
>> >> ceph-disk-activate --mark-init sysvinit --mount /dev/xvdb1
>> >>
>>
>> Is it possible you have tried to install Ceph a few times to get into
>> this state? It is odd that /var/lib/ceph/tmp doesn't exist
>> (as presented by the logs).
>>
>> Are you able to replicate this problem from scratch? I would try to
>> redo everything from a fresh host. If that is not possible, you
>> could tell ceph-deploy to wipe everything. Just as long as you do not
>> care about the data (the following will wipe all data stored in
>> ceph) you could run this:
>>
>> ceph-deploy purge ceph02 && ceph-deploy purgedata ceph02 &&
>> ceph-deploy install ceph02
>>
>>
>> >>
>> >> At OSD log:
>> >>
>> >> [ceph at ceph02 ~]$ tail -100 /var/log/ceph/ceph-osd.0.log
>> >> 2014-06-27 16:35:49.859984 7f143254e7a0  0 ceph version 0.80.1
>> >> (a38fe1169b6d2ac98b427334c12d7cf81f809b74), process ceph-osd, pid 6590
>> >> 2014-06-27 16:35:49.861265 7f143254e7a0  1
>> >> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) mkfs in
>> /var/lib/ceph/tmp/mnt.9VZHpR
>> >> 2014-06-27 16:35:49.861319 7f143254e7a0  1
>> >> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) mkfs fsid is already set to
>> >> 5e93fa7c-b6f7-4684-981b-bf73254bd87a
>> >> 2014-06-27 16:35:49.948589 7f143254e7a0  1
>> >> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) leveldb db exists/created
>> >> 2014-06-27 16:35:49.948865 7f143254e7a0 -1
>> >> filestore(/var/lib/ceph/tmp/mnt.9VZHpR) mkjournal error creating
>> journal on
>> >> /var/lib/ceph/tmp/mnt.9VZHpR/journal: (2) No such file or directory
>> >> 2014-06-27 16:35:49.948893 7f143254e7a0 -1 OSD::mkfs: ObjectStore::mkfs
>> >> failed with error -2
>> >> 2014-06-27 16:35:49.948957 7f143254e7a0 -1  ** ERROR: error creating
>> empty
>> >> object store in /var/lib/ceph/tmp/mnt.9VZHpR: (2) No such file or
>> directory
>> >> [ceph at ceph02 ~]$
>> >>
>> >> [ceph at ceph02 ~]$ sudo ceph-disk list
>> >> /dev/xvda1 other, ext4, mounted on /
>> >> /dev/xvdb :
>> >>  /dev/xvdb1 ceph data, prepared, cluster ceph, osd.0, journal
>> /dev/xvdb2
>> >>  /dev/xvdb2 ceph journal, for /dev/xvdb1
>> >>
>> >>
>> >> Thanks in advance, I
>> >>
>> >>
>> >>
>> >> 2014-06-27 15:30 GMT+02:00 Alfredo Deza <alfredo.deza at inktank.com>:
>> >>
>> >>> Can you paste the full ceph-deploy logs? there are a few reasons why
>> >>> this might be happening.
>> >>>
>> >>>
>> >>>
>> >>> On Fri, Jun 27, 2014 at 6:42 AM, Iban Cabrillo <
>> cabrillo at ifca.unican.es>
>> >>> wrote:
>> >>> > Hi,
>> >>> >
>> >>> >  I am a little frustrated. After 6 times trying to deploy a test
>> ceph
>> >>> > always
>> >>> > get the same error in the osd activation stage.
>> >>> >  The version is firefly (for el6 repo), 3 mons, 3 osds all of then
>> Xen
>> >>> > VMs.
>> >>> >
>> >>> >  The mons wake up correctly and I do not know why two osd servers
>> too
>> >>> > after
>> >>> > a lot of errors, always the same
>> >>> >
>> >>> > #ceph-deploy --verbose osd prepare ceph02:xvdb (works fine)
>> >>> >
>> >>> > [root at ceph02 ~]# parted /dev/xvdb
>> >>> > GNU Parted 2.1
>> >>> > Using /dev/xvdb
>> >>> > Welcome to GNU Parted! Type 'help' to view a list of commands.
>> >>> > (parted) p
>> >>> > Model: Xen Virtual Block Device (xvd)
>> >>> > Disk /dev/xvdb: 107GB
>> >>> > Sector size (logical/physical): 512B/512B
>> >>> > Partition Table: gpt
>> >>> >
>> >>> > Number  Start   End     Size    File system  Name          Flags
>> >>> >  2      1049kB  10,7GB  10,7GB  xfs          ceph journal
>> >>> >  1      10,7GB  107GB   96,6GB  xfs          ceph data
>> >>> >
>> >>> >
>> >>> > But the activate give us this error:
>> >>> >
>> >>> >  ceph-deploy --verbose osd activate ceph02:xvdb1:/dev/xvdb2
>> >>> >
>> >>> >   [ceph02][WARNIN] 2014-06-27 12:27:34.750160 7f123b33d7a0 -1
>> >>> > filestore(/var/lib/ceph/tmp/mnt.HacFAP) mkjournal error creating
>> >>> > journal on
>> >>> > /var/lib/ceph/tmp/mnt.HacFAP/journal: (2) No such file or directory
>> >>> > [ceph02][WARNIN] 2014-06-27 12:27:34.750281 7f123b33d7a0 -1
>> OSD::mkfs:
>> >>> > ObjectStore::mkfs failed with error -2
>> >>> > [ceph02][WARNIN] 2014-06-27 12:27:34.750416 7f123b33d7a0 -1  **
>> ERROR:
>> >>> > error
>> >>> > creating empty object store in /var/lib/ceph/tmp/mnt.HacFAP: (2) No
>> >>> > such
>> >>> > file or directory
>> >>> > [ceph02][WARNIN] ERROR:ceph-disk:Failed to activate
>> >>> >
>> >>> > Two of then following the same procedure after the same error
>> several
>> >>> > time,
>> >>> > raise up.
>> >>> >  [ceph at ceph03 ~]$ df -h
>> >>> > Filesystem      Size  Used Avail Use% Mounted on
>> >>> > /dev/xvda1      5.0G  2.2G  2.6G  46% /
>> >>> > tmpfs           935M     0  935M   0% /dev/shm
>> >>> > /dev/xvdb1       90G   37M   90G   1% /var/lib/ceph/osd/ceph-6
>> >>> >
>> >>> >
>> >>> > Any idea please?
>> >>> >
>> >>> >
>> >>> > Bertrand Russell:
>> >>> > "El problema con el mundo es que los est?pidos est?n seguros de
>> todo y
>> >>> > los
>> >>> > inteligentes est?n llenos de dudas"
>> >>> >
>> >>> > _______________________________________________
>> >>> > ceph-users mailing list
>> >>> > ceph-users at lists.ceph.com
>> >>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >>> >
>> >>
>> >>
>> >
>>
>
>
>
> --
>
> ############################################################################
> Iban Cabrillo Bartolome
> Instituto de Fisica de Cantabria (IFCA)
> Santander, Spain
> Tel: +34942200969
> PGP PUBLIC KEY:
> http://pgp.mit.edu/pks/lookup?op=get&search=0xD9DF0B3D6C8C08AC
>
> ############################################################################
> Bertrand Russell:
> *"El problema con el mundo es que los est?pidos est?n seguros de todo y
> los inteligentes est?n llenos de dudas*"
>
> _______________________________________________
> ceph-users mailing list
> ceph-users at lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ceph.com/pipermail/ceph-users-ceph.com/attachments/20140801/8ae13dfd/attachment.htm>


[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux