Re: Ceph 15.2.3 on Ubuntu 20.04 with odroid xu4 / python thread Problem

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thursday, January 7, 2021, Oliver Weinmann <oliver.weinmann@xxxxxx>
wrote:

> Hi,
>
> I have a similar if not the same issue. I run armbian buster on my odroid
> hc2 which is the same as a xu4 and I get the following error, trying to
> build a cluster with ceph-ansible:


We have a fix for a similar issue recently. See
https://github.com/ceph/ceph/pull/38665. Could you give it a shot? I will
backport it to LTS branches if it helps.



>
> ASK [ceph-osd : use ceph-volume lvm batch to create bluestore osds]
> ***************************************************
> Wednesday 06 January 2021  21:46:44 +0000 (0:00:00.073) 0:02:01.697 *****
> fatal: [192.168.2.123]: FAILED! => changed=true
>   cmd:
>   - ceph-volume
>   - --cluster
>   - ceph
>   - lvm
>   - batch
>   - --bluestore
>   - --yes
>   - /dev/sda
>   delta: '0:00:02.979200'
>   end: '2021-01-06 22:46:48.049074'
>   msg: non-zero return code
>   rc: 1
>   start: '2021-01-06 22:46:45.069874'
>   stderr: |-
>     --> DEPRECATION NOTICE
>     --> You are using the legacy automatic disk sorting behavior
>     --> The Pacific release will change the default to --no-auto
>     --> passed data devices: 1 physical, 0 LVM
>     --> relative data size: 1.0
>     Running command: /usr/bin/ceph-authtool --gen-print-key
>     Running command: /usr/bin/ceph --cluster ceph --name
> client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring
> -i - osd new 8854fc6d-d637-40a9-a1b1-b8e2eeee0afd
>      stderr: Traceback (most recent call last):
>      stderr: File "/usr/bin/ceph", line 1273, in <module>
>      stderr: retval = main()
>      stderr: File "/usr/bin/ceph", line 982, in main
>      stderr: conffile=conffile)
>      stderr: File "/usr/lib/python3/dist-packages/ceph_argparse.py", line
> 1320, in run_in_thread
>      stderr: raise Exception("timed out")
>      stderr: Exception: timed out
>     Traceback (most recent call last):
>       File "/usr/sbin/ceph-volume", line 11, in <module>
>         load_entry_point('ceph-volume==1.0.0', 'console_scripts',
> 'ceph-volume')()
>       File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line 39,
> in __init__
>         self.main(self.argv)
>       File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py",
> line 59, in newfunc
>         return f(*a, **kw)
>       File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line
> 150, in main
>         terminal.dispatch(self.mapper, subcommand_args)
>       File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line
> 194, in dispatch
>         instance.main()
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/main.py",
> line 42, in main
>         terminal.dispatch(self.mapper, self.argv)
>       File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line
> 194, in dispatch
>         instance.main()
>       File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py",
> line 16, in is_root
>         return func(*a, **kw)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py",
> line 415, in main
>         self._execute(plan)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py",
> line 434, in _execute
>         c.create(argparse.Namespace(**args))
>       File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py",
> line 16, in is_root
>         return func(*a, **kw)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/create.py",
> line 26, in create
>         prepare_step.safe_prepare(args)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py",
> line 252, in safe_prepare
>         self.prepare()
>       File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py",
> line 16, in is_root
>         return func(*a, **kw)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py",
> line 292, in prepare
>         self.osd_id = prepare_utils.create_id(osd_fsid,
> json.dumps(secrets), osd_id=self.args.osd_id)
>       File "/usr/lib/python3/dist-packages/ceph_volume/util/prepare.py",
> line 173, in create_id
>         raise RuntimeError('Unable to create a new OSD id')
>     RuntimeError: Unable to create a new OSD id
>   stderr_lines: <omitted>
>   stdout: ''
>   stdout_lines: <omitted>
> fatal: [odroidxu4]: FAILED! => changed=true
>   cmd:
>   - ceph-volume
>   - --cluster
>   - ceph
>   - lvm
>   - batch
>   - --bluestore
>   - --yes
>   - /dev/sda
>   delta: '0:00:03.510973'
>   end: '2021-01-06 22:46:48.514102'
>   msg: non-zero return code
>   rc: 1
>   start: '2021-01-06 22:46:45.003129'
>   stderr: |-
>     --> DEPRECATION NOTICE
>     --> You are using the legacy automatic disk sorting behavior
>     --> The Pacific release will change the default to --no-auto
>     --> passed data devices: 1 physical, 0 LVM
>     --> relative data size: 1.0
>     Running command: /usr/bin/ceph-authtool --gen-print-key
>     Running command: /usr/bin/ceph --cluster ceph --name
> client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring
> -i - osd new 4e292c82-bb4d-4581-aead-46ff635fda69
>      stderr: Traceback (most recent call last):
>      stderr: File "/usr/bin/ceph", line 1273, in <module>
>      stderr: retval = main()
>      stderr: File "/usr/bin/ceph", line 982, in main
>      stderr: conffile=conffile)
>      stderr: File "/usr/lib/python3/dist-packages/ceph_argparse.py", line
> 1320, in run_in_thread
>      stderr: raise Exception("timed out")
>      stderr: Exception: timed out
>      stderr: /build/ceph-Ti7FjJ/ceph-14.2.15/src/common/config.cc: In
> function 'void md_config_t::set_val_default(ConfigValues&, const
> ConfigTracker&, const string&, const string&)' thread b0e3a460 time
> 2021-01-06 22:46:48.357354
>      stderr: /build/ceph-Ti7FjJ/ceph-14.2.15/src/common/config.cc: 259:
> FAILED ceph_assert(o)
>      stderr: ceph version 14.2.15 (afdd217ae5fb1ed3f60e16bd62357ca58cc650e5)
> nautilus (stable)
>      stderr: 1: (ceph::__ceph_assert_fail(char const*, char const*, int,
> char const*)+0xeb) [0xb18b26a4]
>      stderr: 2: (ceph::__ceph_assert_fail(ceph::assert_data const&)+0xd)
> [0xb18b2802]
>      stderr: 3: (md_config_t::set_val_default(ConfigValues&,
> ConfigTracker const&, std::__cxx11::basic_string<char,
> std::char_traits<char>, std::allocator<char> > const&,
> std::__cxx11::basic_string<char, std::char_traits<char>,
> std::allocator<char> > const&)+0x69) [0xb195be0a]
>      stderr: 4: (md_config_t::md_config_t(ConfigValues&, ConfigTracker
> const&, bool)+0x15d31) [0xb1972ac6]
>      stderr: 5: (CephContext::CephContext(unsigned int,
> code_environment_t, int)+0x10ef) [0xb193d090]
>      stderr: 6: (common_preinit(CephInitParameters const&,
> code_environment_t, int)+0x7d) [0xb1956af6]
>      stderr: 7: (()+0x2046a) [0xb639546a]
>      stderr: 8: (rados_create2()+0x55) [0xb639589e]
>     Traceback (most recent call last):
>       File "/usr/sbin/ceph-volume", line 11, in <module>
>         load_entry_point('ceph-volume==1.0.0', 'console_scripts',
> 'ceph-volume')()
>       File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line 39,
> in __init__
>         self.main(self.argv)
>       File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py",
> line 59, in newfunc
>         return f(*a, **kw)
>       File "/usr/lib/python3/dist-packages/ceph_volume/main.py", line
> 150, in main
>         terminal.dispatch(self.mapper, subcommand_args)
>       File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line
> 194, in dispatch
>         instance.main()
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/main.py",
> line 42, in main
>         terminal.dispatch(self.mapper, self.argv)
>       File "/usr/lib/python3/dist-packages/ceph_volume/terminal.py", line
> 194, in dispatch
>         instance.main()
>       File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py",
> line 16, in is_root
>         return func(*a, **kw)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py",
> line 415, in main
>         self._execute(plan)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/batch.py",
> line 434, in _execute
>         c.create(argparse.Namespace(**args))
>       File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py",
> line 16, in is_root
>         return func(*a, **kw)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/create.py",
> line 26, in create
>         prepare_step.safe_prepare(args)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py",
> line 252, in safe_prepare
>         self.prepare()
>       File "/usr/lib/python3/dist-packages/ceph_volume/decorators.py",
> line 16, in is_root
>         return func(*a, **kw)
>       File "/usr/lib/python3/dist-packages/ceph_volume/devices/lvm/prepare.py",
> line 292, in prepare
>         self.osd_id = prepare_utils.create_id(osd_fsid,
> json.dumps(secrets), osd_id=self.args.osd_id)
>       File "/usr/lib/python3/dist-packages/ceph_volume/util/prepare.py",
> line 173, in create_id
>         raise RuntimeError('Unable to create a new OSD id')
>     RuntimeError: Unable to create a new OSD id
>   stderr_lines: <omitted>
>   stdout: ''
>   stdout_lines: <omitted>
>
> Trying to run the failed command on one of the odroid nodes:
>
> root@odroidxu4:~# /usr/bin/ceph --cluster ceph --name
> client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring
> -i - osd new 8854fc6d-d637-40a9-a1b1-b8e2eeee0afd
> Traceback (most recent call last):
>   File "/usr/bin/ceph", line 1273, in <module>
>     retval = main()
>   File "/usr/bin/ceph", line 982, in main
>     conffile=conffile)
>   File "/usr/lib/python3/dist-packages/ceph_argparse.py", line 1320, in
> run_in_thread
>     raise Exception("timed out")
> Exception: timed out
>
> Any clues?
>
> Best Regards,
>
> Oliver
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>


-- 
Regards
Kefu Chai
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux