On 2018/11/6 下午4:16, Mykola Golub wrote:
On Tue, Nov 06, 2018 at 09:45:01AM +0800, Dengke Du wrote:
I reconfigure the osd service from start, the journal was:
I am not quite sure I understand what you mean here.
------------------------------------------------------------------------------------------------------------------------------------------
-- Unit ceph-osd@0.service has finished starting up.
--
-- The start-up result is RESULT.
Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 7f6a27204e80
-1 Public network was set, but cluster network was not set
Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915 7f6a27204e80
-1 Using public network also for cluster network
Nov 05 18:02:36 node1 ceph-osd[4487]: starting osd.0 at - osd_data
/var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal
Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.365 7f6a27204e80
-1 journal FileJournal::_open: disabling aio for non-block journal. Use
journal_force_aio to force use of a>
Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.414 7f6a27204e80
-1 journal do_read_entry(6930432): bad header magic
Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.729 7f6a27204e80
-1 osd.0 21 log_to_monitors {default=true}
Nov 05 18:02:47 node1 nagios[3584]: Warning: Return code of 13 for check of
host 'localhost' was out of bounds.
------------------------------------------------------------------------------------------------------------------------------------------
Could you please post the full ceph-osd log somewhere? /var/log/ceph/ceph-osd.0.log
I don't have the file /var/log/ceph/ceph-osd.o.log
root@node1:~# systemctl status ceph-osd@0
● ceph-osd@0.service - Ceph object storage daemon osd.0
Loaded: loaded (/lib/systemd/system/ceph-osd@.service; disabled;
vendor preset: enabled)
Active: active (running) since Mon 2018-11-05 18:02:36 UTC; 6h ago
Main PID: 4487 (ceph-osd)
Tasks: 64
Memory: 27.0M
CGroup: /system.slice/system-ceph\x2dosd.slice/ceph-osd@0.service
└─4487 /usr/bin/ceph-osd -f --cluster ceph --id 0
Nov 05 18:02:36 node1 systemd[1]: Starting Ceph object storage daemon
osd.0...
Nov 05 18:02:36 node1 systemd[1]: Started Ceph object storage daemon osd.0.
Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915
7f6a27204e80 -1 Public network was set, but cluster network was not set
Nov 05 18:02:36 node1 ceph-osd[4487]: 2018-11-05 18:02:36.915
7f6a27204e80 -1 Using public network also for cluster network
Nov 05 18:02:36 node1 ceph-osd[4487]: starting osd.0 at - osd_data
/var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal
Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.365
7f6a27204e80 -1 journal FileJournal::_open: disabling aio for non-block
journal. Use journal_force_aio to force use of a>
Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.414
7f6a27204e80 -1 journal do_read_entry(6930432): bad header magic
Nov 05 18:02:37 node1 ceph-osd[4487]: 2018-11-05 18:02:37.729
7f6a27204e80 -1 osd.0 21 log_to_monitors {default=true}
but hang at the command: "rbd create libvirt-pool/dimage --size 10240 "
So it hungs forever now instead of returning the error?
no returning any error, just hungs
What is `ceph -s` output?
root@node1:~# ceph -s
cluster:
id: 9c1a42e1-afc2-4170-8172-96f4ebdaac68
health: HEALTH_WARN
no active mgr
services:
mon: 1 daemons, quorum 0
mgr: no daemons active
osd: 1 osds: 0 up, 0 in
data:
pools: 0 pools, 0 pgs
objects: 0 objects, 0 B
usage: 0 B used, 0 B / 0 B avail
pgs:
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com