Oliver Weinmann <oliver.weinmann@xxxxxx>于2021年1月8日 周五04:30写道: > Ok, I replaced the whole file ceph_argparse.py with the patched one from > github. Instead of an throwing an error it now seems to be stuck forever. > Or am I to impatient? I'm running > I don’t think so. In a healthy cluster, the command should complete in no more than 1 second. I just checked the revision history of ceph_argparse.py, there are a bunch of changes since the release of nautilus. My guess is that the version in master might include some bits not compatible with nautilus? So, I’d suggest only cherry-pick the change in that PR, and try again. > debian buster so this is not the latest ceph release octopus, but nautilus: > > root@odroidxu4:~# dpkg -l ceph > Desired=Unknown/Install/Remove/Purge/Hold > | > Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend > |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad) > ||/ Name Version Architecture Description > > +++-==============-=================-============-=================================== > ii ceph 14.2.15-3~bpo10+1 armhf distributed storage and > file system > Am 07.01.2021 um 13:07 schrieb kefu chai: > > > > Oliver Weinmann <oliver.weinmann@xxxxxx>于2021年1月7日 周四16:32写道: > >> Hi, >> >> thanks for the quick reply. I will test it. Do I have to recompile ceph >> in order to test it? >> > > No, you just need to apply the change of ceph_argparse.py. > > >> Am 07.01.2021 um 02:13 schrieb kefu chai: >> >> >> >> On Thursday, January 7, 2021, Oliver Weinmann <oliver.weinmann@xxxxxx> >> wrote: >> >>> Hi, >>> >>> I have a similar if not the same issue. I run armbian buster on my >>> odroid hc2 which is the same as a xu4 and I get the following error, trying >>> to build a cluster with ceph-ansible: >> >> >> We have a fix for a similar issue recently. See >> https://github.com/ceph/ceph/pull/38665. Could you give it a shot? I >> will backport it to LTS branches if it helps. >> >> >> >>> >>> ASK [ceph-osd : use ceph-volume lvm batch to create bluestore osds] >>> *************************************************** >>> Wednesday 06 January 2021 21:46:44 +0000 (0:00:00.073) 0:02:01.697 ***** >>> fatal: [192.168.2.123]: FAILED! => changed=true >>> cmd: >>> - ceph-volume >>> - --cluster >>> - ceph >>> - lvm >>> - batch >>> - --bluestore >>> - --yes >>> - /dev/sda >>> delta: '0:00:02.979200' >>> end: '2021-01-06 22:46:48.049074' >>> msg: non-zero return code >>> rc: 1 >>> start: '2021-01-06 22:46:45.069874' >>> stderr: |- >>> --> DEPRECATION NOTICE >>> --> You are using the legacy automatic disk sorting behavior >>> --> The Pacific release will change the default to --no-auto >>> --> passed data devices: 1 physical, 0 LVM >>> --> relative data size: 1.0 >>> Running command: /usr/bin/ceph-authtool --gen-print-key >>> Running command: /usr/bin/ceph --cluster ceph --name >>> client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i >>> - osd new 8854fc6d-d637-40a9-a1b1-b8e2eeee0afd >>> stderr: Traceback (most recent call last): >>> stderr: File "/usr/bin/ceph", line 1273, in <module> >>> stderr: retval = main() >>> stderr: File "/usr/bin/ceph", line 982, in main >>> stderr: conffile=conffile) >>> stderr: File "/usr/lib/python3/dist-packages/ceph_argparse.py", >>> line 1320, in run_in_thread >>> stderr: raise Exception("timed out") >>> stderr: Exception: timed out >>> Traceback (most recent call last): >>> File "/usr/sbin/ceph-volume", line 11, in <module> >>> load_entry_point('ceph-volume==1.0.0', 'console_scripts', >>> 'ceph-volume')() >>> File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line >>> 39, in __init__ >>> self.main(self.argv) >>> File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", >>> line 59, in newfunc >>> return f(*a, **kw) >>> File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line >>> 150, in main >>> terminal.dispatch(self.mapper, subcommand_args) >>> File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", >>> line 194, in dispatch >>> instance.main() >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/main.py", line 42, >>> in main >>> terminal.dispatch(self.mapper, self.argv) >>> File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", >>> line 194, in dispatch >>> instance.main() >>> File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", >>> line 16, in is_root >>> return func(*a, **kw) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py", line >>> 415, in main >>> self._execute(plan) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py", line >>> 434, in _execute >>> c.create(argparse.Namespace(**args)) >>> File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", >>> line 16, in is_root >>> return func(*a, **kw) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/create.py", line >>> 26, in create >>> prepare_step.safe_prepare(args) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py", line >>> 252, in safe_prepare >>> self.prepare() >>> File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", >>> line 16, in is_root >>> return func(*a, **kw) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py", line >>> 292, in prepare >>> self.osd_id = prepare_utils.create_id(osd_fsid, >>> json.dumps(secrets), osd_id=self.args.osd_id) >>> File "/usr/lib/python3/dist-packages/ceph_volume/util/prepare.py", >>> line 173, in create_id >>> raise RuntimeError('Unable to create a new OSD id') >>> RuntimeError: Unable to create a new OSD id >>> stderr_lines: <omitted> >>> stdout: '' >>> stdout_lines: <omitted> >>> fatal: [odroidxu4]: FAILED! => changed=true >>> cmd: >>> - ceph-volume >>> - --cluster >>> - ceph >>> - lvm >>> - batch >>> - --bluestore >>> - --yes >>> - /dev/sda >>> delta: '0:00:03.510973' >>> end: '2021-01-06 22:46:48.514102' >>> msg: non-zero return code >>> rc: 1 >>> start: '2021-01-06 22:46:45.003129' >>> stderr: |- >>> --> DEPRECATION NOTICE >>> --> You are using the legacy automatic disk sorting behavior >>> --> The Pacific release will change the default to --no-auto >>> --> passed data devices: 1 physical, 0 LVM >>> --> relative data size: 1.0 >>> Running command: /usr/bin/ceph-authtool --gen-print-key >>> Running command: /usr/bin/ceph --cluster ceph --name >>> client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i >>> - osd new 4e292c82-bb4d-4581-aead-46ff635fda69 >>> stderr: Traceback (most recent call last): >>> stderr: File "/usr/bin/ceph", line 1273, in <module> >>> stderr: retval = main() >>> stderr: File "/usr/bin/ceph", line 982, in main >>> stderr: conffile=conffile) >>> stderr: File "/usr/lib/python3/dist-packages/ceph_argparse.py", >>> line 1320, in run_in_thread >>> stderr: raise Exception("timed out") >>> stderr: Exception: timed out >>> stderr: /build/ceph-Ti7FjJ/ceph-14.2.15/src/common/config.cc: In >>> function 'void md_config_t::set_val_default(ConfigValues&, const >>> ConfigTracker&, const string&, const string&)' thread b0e3a460 time >>> 2021-01-06 22:46:48.357354 >>> stderr: /build/ceph-Ti7FjJ/ceph-14.2.15/src/common/config.cc: 259: >>> FAILED ceph_assert(o) >>> stderr: ceph version 14.2.15 >>> (afdd217ae5fb1ed3f60e16bd62357ca58cc650e5) nautilus (stable) >>> stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int, >>> char const*)+0xeb) [0xb18b26a4] >>> stderr: 2: (ceph::__ceph_assert_fail(ceph::assert_data const&)+0xd) >>> [0xb18b2802] >>> stderr: 3: (md_config_t::set_val_default(ConfigValues&, >>> ConfigTracker const&, std::__cxx11::basic_string<char, >>> std::char_traits<char>, std::allocator<char> > const&, >>> std::__cxx11::basic_string<char, std::char_traits<char>, >>> std::allocator<char> > const&)+0x69) [0xb195be0a] >>> stderr: 4: (md_config_t::md_config_t(ConfigValues&, ConfigTracker >>> const&, bool)+0x15d31) [0xb1972ac6] >>> stderr: 5: (CephContext::CephContext(unsigned int, >>> code_environment_t, int)+0x10ef) [0xb193d090] >>> stderr: 6: (common_preinit(CephInitParameters const&, >>> code_environment_t, int)+0x7d) [0xb1956af6] >>> stderr: 7: (()+0x2046a) [0xb639546a] >>> stderr: 8: (rados_create2()+0x55) [0xb639589e] >>> Traceback (most recent call last): >>> File "/usr/sbin/ceph-volume", line 11, in <module> >>> load_entry_point('ceph-volume==1.0.0', 'console_scripts', >>> 'ceph-volume')() >>> File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line >>> 39, in __init__ >>> self.main(self.argv) >>> File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", >>> line 59, in newfunc >>> return f(*a, **kw) >>> File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line >>> 150, in main >>> terminal.dispatch(self.mapper, subcommand_args) >>> File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", >>> line 194, in dispatch >>> instance.main() >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/main.py", line 42, >>> in main >>> terminal.dispatch(self.mapper, self.argv) >>> File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", >>> line 194, in dispatch >>> instance.main() >>> File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", >>> line 16, in is_root >>> return func(*a, **kw) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py", line >>> 415, in main >>> self._execute(plan) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py", line >>> 434, in _execute >>> c.create(argparse.Namespace(**args)) >>> File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", >>> line 16, in is_root >>> return func(*a, **kw) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/create.py", line >>> 26, in create >>> prepare_step.safe_prepare(args) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py", line >>> 252, in safe_prepare >>> self.prepare() >>> File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py", >>> line 16, in is_root >>> return func(*a, **kw) >>> File >>> "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py", line >>> 292, in prepare >>> self.osd_id = prepare_utils.create_id(osd_fsid, >>> json.dumps(secrets), osd_id=self.args.osd_id) >>> File "/usr/lib/python3/dist-packages/ceph_volume/util/prepare.py", >>> line 173, in create_id >>> raise RuntimeError('Unable to create a new OSD id') >>> RuntimeError: Unable to create a new OSD id >>> stderr_lines: <omitted> >>> stdout: '' >>> stdout_lines: <omitted> >>> >>> Trying to run the failed command on one of the odroid nodes: >>> >>> root@odroidxu4:~# /usr/bin/ceph --cluster ceph --name >>> client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i >>> - osd new 8854fc6d-d637-40a9-a1b1-b8e2eeee0afd >>> Traceback (most recent call last): >>> File "/usr/bin/ceph", line 1273, in <module> >>> retval = main() >>> File "/usr/bin/ceph", line 982, in main >>> conffile=conffile) >>> File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1320, in >>> run_in_thread >>> raise Exception("timed out") >>> Exception: timed out >>> >>> Any clues? >>> >>> Best Regards, >>> >>> Oliver >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@xxxxxxx >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx >>> >> >> >> -- >> Regards >> Kefu Chai >> >> -- > Regards > Kefu Chai > > -- Regards Kefu Chai _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx