Re: Where rocksdb on my OSD's?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi, Igor!
Service is UP.
I did not make separate devices.
blocks.db and blocks.wal are created only if they are on separate devices?

# systemctl status ceph-osd@8.service
● ceph-osd@8.service - Ceph object storage daemon osd.8                                                                                                                                      
   Loaded: loaded (/usr/lib/systemd/system/ceph-osd@.service; enabled-runtime; vendor preset: disabled)
   Active: active (running) since Sat 2019-11-09 15:56:28 MSK; 1 day 23h ago
  Process: 2676 ExecStartPre=/usr/lib/ceph/ceph-osd-prestart.sh --cluster ${CLUSTER} --id %i (code=exited, status=0/SUCCESS)
 Main PID: 2707 (ceph-osd)
   CGroup: /system.slice/system-ceph\x2dosd.slice/ceph-osd@8.service
           └─2707 /usr/bin/ceph-osd -f --cluster ceph --id 8 --setuser ceph --setgroup ceph

Nov 09 15:56:28 test-host5 systemd[1]: Starting Ceph object storage daemon osd.8...
Nov 09 15:56:28 test-host5 systemd[1]: Started Ceph object storage daemon osd.8.


11.11.2019, 15:18, "Igor Fedotov" <ifedotov@xxxxxxx>:
> Hi Andrey,
>
> this log output rather looks like some other process is using
> /var/lib/ceph/osd/ceph-8
>
> Have you stopped OSD.8 daemon?
>
> And are you sure you deployed standalone DB/WAL devices for this OSD?
>
> Thanks,
>
> Igor
>
> On 11/11/2019 3:10 PM, Andrey Groshev wrote:
>>  Hello,
>>
>>  Some time ago I deployed a ceph cluster.
>>  It works great.
>>  Today I collect some statistics and found that the BlueFs utility is not working.
>>
>>  # ceph-bluestore-tool bluefs-bdev-sizes --path /var/lib/ceph/osd/ceph-8
>>  inferring bluefs devices from bluestore path
>>    slot 1 /var/lib/ceph/osd/ceph-8/block -> /dev/dm-5
>>  unable to open /var/lib/ceph/osd/ceph-8/block: (11) Resource temporarily unavailable
>>  2019-11-11 15:03:30.665 7f4b9a427f00 -1 bdev(0x55d5b0310a80 /var/lib/ceph/osd/ceph-8/block) _lock flock failed on /var/lib/ceph/osd/ceph-8/block
>>  2019-11-11 15:03:30.665 7f4b9a427f00 -1 bdev(0x55d5b0310a80 /var/lib/ceph/osd/ceph-8/block) open failed to lock /var/lib/ceph/osd/ceph-8/block: (11) Resource temporarily unavailable
>>
>>  As far as I understand, blocks.db and blocks.wal are missing. I don’t know how it turned out.
>>
>>  # ls -l /var/lib/ceph/osd/ceph-8
>>  total 28
>>  lrwxrwxrwx 1 ceph ceph 93 Nov 9 15:56 block -> /dev/ceph-55b8a53d-1740-402a-b6f4-09d4befdd564/osd-block-c5488db7-621a-490a-88a0-904c12e8b8ed
>>  -rw------- 1 ceph ceph 37 Nov 9 15:56 ceph_fsid
>>  -rw------- 1 ceph ceph 37 Nov 9 15:56 fsid
>>  -rw------- 1 ceph ceph 55 Nov 9 15:56 keyring
>>  -rw------- 1 ceph ceph 6 Nov 9 15:56 ready
>>  -rw-r--r-- 1 ceph ceph 3 Nov 9 15:56 require_osd_release
>>  -rw------- 1 ceph ceph 10 Nov 9 15:56 type
>>  -rw------- 1 ceph ceph 2 Nov 9 15:56 whoami
>>
>>  Did as standard:
>>  ....
>>  ceph-deploy osd create test-host1:/dev/sde
>>  ceph-deploy osd create test-host2:/dev/sde
>>  ....
>>
>>  What now to do and whether it is necessary to repair it at all?
>>  Assembled on luminous, yesterday updated to Nautilus.
>>
>>  _______________________________________________
>>  ceph-users mailing list -- ceph-users@xxxxxxx
>>  To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux