Re: ceph-disk activate-block: not a block device

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Also I'm just noticing osd1 is my only OSD host that even has an enabled target for ceph-disk (ceph-disk@dev-sdb2.service).

roger@osd1:~$ systemctl list-units ceph*
  UNIT                       LOAD   ACTIVE SUB     DESCRIPTION
● ceph-disk@dev-sdb2.service loaded failed failed  Ceph disk activation: /dev/sdb2
  ceph-osd@3.service         loaded active running Ceph object storage daemon osd.3
  ceph-mds.target            loaded active active  ceph target allowing to start/stop all ceph-mds@.service instances at once
  ceph-mgr.target            loaded active active  ceph target allowing to start/stop all ceph-mgr@.service instances at once
  ceph-mon.target            loaded active active  ceph target allowing to start/stop all ceph-mon@.service instances at once
  ceph-osd.target            loaded active active  ceph target allowing to start/stop all ceph-osd@.service instances at once
  ceph-radosgw.target        loaded active active  ceph target allowing to start/stop all ceph-radosgw@.service instances at once
  ceph.target                loaded active active  ceph target allowing to start/stop all ceph*@.service instances at once

roger@osd2:~$ systemctl list-units ceph*
UNIT                LOAD   ACTIVE SUB     DESCRIPTION
ceph-osd@4.service  loaded active running Ceph object storage daemon osd.4
ceph-mds.target     loaded active active  ceph target allowing to start/stop all ceph-mds@.service instances at once
ceph-mgr.target     loaded active active  ceph target allowing to start/stop all ceph-mgr@.service instances at once
ceph-mon.target     loaded active active  ceph target allowing to start/stop all ceph-mon@.service instances at once
ceph-osd.target     loaded active active  ceph target allowing to start/stop all ceph-osd@.service instances at once
ceph-radosgw.target loaded active active  ceph target allowing to start/stop all ceph-radosgw@.service instances at once
ceph.target         loaded active active  ceph target allowing to start/stop all ceph*@.service instances at once

roger@osd3:~$ systemctl list-units ceph*
UNIT                LOAD   ACTIVE SUB     DESCRIPTION
ceph-osd@0.service  loaded active running Ceph object storage daemon osd.0
ceph-mds.target     loaded active active  ceph target allowing to start/stop all ceph-mds@.service instances at once
ceph-mgr.target     loaded active active  ceph target allowing to start/stop all ceph-mgr@.service instances at once
ceph-mon.target     loaded active active  ceph target allowing to start/stop all ceph-mon@.service instances at once
ceph-osd.target     loaded active active  ceph target allowing to start/stop all ceph-osd@.service instances at once
ceph-radosgw.target loaded active active  ceph target allowing to start/stop all ceph-radosgw@.service instances at once
ceph.target         loaded active active  ceph target allowing to start/stop all ceph*@.service instances at once


On Thu, Jul 20, 2017 at 10:23 AM Roger Brown <rogerpbrown@xxxxxxxxx> wrote:
I think I need help with some OSD trouble. OSD daemons on two hosts started flapping. At length, I rebooted host osd1 (osd.3), but the OSD daemon still fails to start. Upon closer inspection, ceph-disk@dev-sdb2.service is failing to start due to, "Error: /dev/sdb2 is not a block device"

This is the command I see it failing to run: 

roger@osd1:~$ sudo /usr/sbin/ceph-disk --verbose activate-block /dev/sdb2
Traceback (most recent call last):
  File "/usr/sbin/ceph-disk", line 9, in <module>
    load_entry_point('ceph-disk==1.0.0', 'console_scripts', 'ceph-disk')()
  File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5731, in run
    main(sys.argv[1:])
  File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5682, in main
    args.func(args)
  File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5438, in <lambda>
    func=lambda args: main_activate_space(name, args),
  File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 4160, in main_activate_space
    osd_uuid = get_space_osd_uuid(name, dev)
  File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 4115, in get_space_osd_uuid
    raise Error('%s is not a block device' % path)
ceph_disk.main.Error: Error: /dev/sdb2 is not a block device

osd1 environment:
$ ceph -v
ceph version 12.1.1 (f3e663a190bf2ed12c7e3cda288b9a159572c800) luminous (rc)
$ uname -r
4.4.0-83-generic
$ lsb_release -sc
xenial

Please advise.

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux