On Thursday, January 7, 2021, Oliver Weinmann <oliver.weinmann@xxxxxx
<mailto:oliver.weinmann@xxxxxx>> wrote:
Hi,
I have a similar if not the same issue. I run armbian buster on my
odroid hc2 which is the same as a xu4 and I get the following
error, trying to build a cluster with ceph-ansible:
We have a fix for a similar issue recently. See
https://github.com/ceph/ceph/pull/38665
<https://github.com/ceph/ceph/pull/38665>. Could you give it a shot? I
will backport it to LTS branches if it helps.
ASK [ceph-osd : use ceph-volume lvm batch to create bluestore
osds] ***************************************************
Wednesday 06 January 2021 21:46:44 +0000 (0:00:00.073)
0:02:01.697 *****
fatal: [192.168.2.123]: FAILED! => changed=true
cmd:
- ceph-volume
- --cluster
- ceph
- lvm
- batch
- --bluestore
- --yes
- /dev/sda
delta: '0:00:02.979200'
end: '2021-01-06 22:46:48.049074'
msg: non-zero return code
rc: 1
start: '2021-01-06 22:46:45.069874'
stderr: |-
--> DEPRECATION NOTICE
--> You are using the legacy automatic disk sorting behavior
--> The Pacific release will change the default to --no-auto
--> passed data devices: 1 physical, 0 LVM
--> relative data size: 1.0
Running command: /usr/bin/ceph-authtool --gen-print-key
Running command: /usr/bin/ceph --cluster ceph --name
client.bootstrap-osd --keyring
/var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
8854fc6d-d637-40a9-a1b1-b8e2eeee0afd
stderr: Traceback (most recent call last):
stderr: File "/usr/bin/ceph", line 1273, in <module>
stderr: retval = main()
stderr: File "/usr/bin/ceph", line 982, in main
stderr: conffile=conffile)
stderr: File
"/usr/lib/python3/dist-packages/ceph_argparse.py", line 1320, in
run_in_thread
stderr: raise Exception("timed out")
stderr: Exception: timed out
Traceback (most recent call last):
File "/usr/sbin/ceph-volume", line 11, in <module>
load_entry_point('ceph-volume==1.0.0', 'console_scripts',
'ceph-volume')()
File "/usr/lib/python3/dist-packages/ceph_volume/main.py",
line 39, in __init__
self.main(self.argv)
File
"/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line
59, in newfunc
return f(*a, **kw)
File "/usr/lib/python3/dist-packages/ceph_volume/main.py",
line 150, in main
terminal.dispatch(self.mapper, subcommand_args)
File
"/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line
194, in dispatch
instance.main()
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/main.py",
line 42, in main
terminal.dispatch(self.mapper, self.argv)
File
"/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line
194, in dispatch
instance.main()
File
"/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line
16, in is_root
return func(*a, **kw)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py",
line 415, in main
self._execute(plan)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py",
line 434, in _execute
c.create(argparse.Namespace(**args))
File
"/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line
16, in is_root
return func(*a, **kw)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/create.py",
line 26, in create
prepare_step.safe_prepare(args)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py",
line 252, in safe_prepare
self.prepare()
File
"/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line
16, in is_root
return func(*a, **kw)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py",
line 292, in prepare
self.osd_id = prepare_utils.create_id(osd_fsid,
json.dumps(secrets), osd_id=self.args.osd_id)
File
"/usr/lib/python3/dist-packages/ceph_volume/util/prepare.py", line
173, in create_id
raise RuntimeError('Unable to create a new OSD id')
RuntimeError: Unable to create a new OSD id
stderr_lines: <omitted>
stdout: ''
stdout_lines: <omitted>
fatal: [odroidxu4]: FAILED! => changed=true
cmd:
- ceph-volume
- --cluster
- ceph
- lvm
- batch
- --bluestore
- --yes
- /dev/sda
delta: '0:00:03.510973'
end: '2021-01-06 22:46:48.514102'
msg: non-zero return code
rc: 1
start: '2021-01-06 22:46:45.003129'
stderr: |-
--> DEPRECATION NOTICE
--> You are using the legacy automatic disk sorting behavior
--> The Pacific release will change the default to --no-auto
--> passed data devices: 1 physical, 0 LVM
--> relative data size: 1.0
Running command: /usr/bin/ceph-authtool --gen-print-key
Running command: /usr/bin/ceph --cluster ceph --name
client.bootstrap-osd --keyring
/var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
4e292c82-bb4d-4581-aead-46ff635fda69
stderr: Traceback (most recent call last):
stderr: File "/usr/bin/ceph", line 1273, in <module>
stderr: retval = main()
stderr: File "/usr/bin/ceph", line 982, in main
stderr: conffile=conffile)
stderr: File
"/usr/lib/python3/dist-packages/ceph_argparse.py", line 1320, in
run_in_thread
stderr: raise Exception("timed out")
stderr: Exception: timed out
stderr: /build/ceph-Ti7FjJ/ceph-14.2.15/src/common/config.cc:
In function 'void md_config_t::set_val_default(ConfigValues&,
const ConfigTracker&, const string&, const string&)' thread
b0e3a460 time 2021-01-06 22:46:48.357354
stderr: /build/ceph-Ti7FjJ/ceph-14.2.15/src/common/config.cc:
259: FAILED ceph_assert(o)
stderr: ceph version 14.2.15
(afdd217ae5fb1ed3f60e16bd62357ca58cc650e5) nautilus (stable)
stderr: 1: (ceph::__ceph_assert_fail(char const*, char
const*, int, char const*)+0xeb) [0xb18b26a4]
stderr: 2: (ceph::__ceph_assert_fail(ceph::assert_data
const&)+0xd) [0xb18b2802]
stderr: 3: (md_config_t::set_val_default(ConfigValues&,
ConfigTracker const&, std::__cxx11::basic_string<char,
std::char_traits<char>, std::allocator<char> > const&,
std::__cxx11::basic_string<char, std::char_traits<char>,
std::allocator<char> > const&)+0x69) [0xb195be0a]
stderr: 4: (md_config_t::md_config_t(ConfigValues&,
ConfigTracker const&, bool)+0x15d31) [0xb1972ac6]
stderr: 5: (CephContext::CephContext(unsigned int,
code_environment_t, int)+0x10ef) [0xb193d090]
stderr: 6: (common_preinit(CephInitParameters const&,
code_environment_t, int)+0x7d) [0xb1956af6]
stderr: 7: (()+0x2046a) [0xb639546a]
stderr: 8: (rados_create2()+0x55) [0xb639589e]
Traceback (most recent call last):
File "/usr/sbin/ceph-volume", line 11, in <module>
load_entry_point('ceph-volume==1.0.0', 'console_scripts',
'ceph-volume')()
File "/usr/lib/python3/dist-packages/ceph_volume/main.py",
line 39, in __init__
self.main(self.argv)
File
"/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line
59, in newfunc
return f(*a, **kw)
File "/usr/lib/python3/dist-packages/ceph_volume/main.py",
line 150, in main
terminal.dispatch(self.mapper, subcommand_args)
File
"/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line
194, in dispatch
instance.main()
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/main.py",
line 42, in main
terminal.dispatch(self.mapper, self.argv)
File
"/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line
194, in dispatch
instance.main()
File
"/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line
16, in is_root
return func(*a, **kw)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py",
line 415, in main
self._execute(plan)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py",
line 434, in _execute
c.create(argparse.Namespace(**args))
File
"/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line
16, in is_root
return func(*a, **kw)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/create.py",
line 26, in create
prepare_step.safe_prepare(args)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py",
line 252, in safe_prepare
self.prepare()
File
"/usr/lib/python3/dist-packages/ceph_volume/decorators.py", line
16, in is_root
return func(*a, **kw)
File
"/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py",
line 292, in prepare
self.osd_id = prepare_utils.create_id(osd_fsid,
json.dumps(secrets), osd_id=self.args.osd_id)
File
"/usr/lib/python3/dist-packages/ceph_volume/util/prepare.py", line
173, in create_id
raise RuntimeError('Unable to create a new OSD id')
RuntimeError: Unable to create a new OSD id
stderr_lines: <omitted>
stdout: ''
stdout_lines: <omitted>
Trying to run the failed command on one of the odroid nodes:
root@odroidxu4:~# /usr/bin/ceph --cluster ceph --name
client.bootstrap-osd --keyring
/var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new
8854fc6d-d637-40a9-a1b1-b8e2eeee0afd
Traceback (most recent call last):
File "/usr/bin/ceph", line 1273, in <module>
retval = main()
File "/usr/bin/ceph", line 982, in main
conffile=conffile)
File "/usr/lib/python3/dist-packages/ceph_argparse.py", line
1320, in run_in_thread
raise Exception("timed out")
Exception: timed out
Any clues?
Best Regards,
Oliver
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
<mailto:ceph-users@xxxxxxx>
To unsubscribe send an email to ceph-users-leave@xxxxxxx
<mailto:ceph-users-leave@xxxxxxx>
--
Regards
Kefu Chai