On Fri, Dec 8, 2017 at 9:49 PM, nokia ceph <nokiacephusers@xxxxxxxxx> wrote:
Hello Team,We aware that ceph-disk which is deprecated in 12.2.2 . As part of my testing, I can still using this ceph-disk utility for creating OSD's in 12.2.2Here I'm getting activation error on the second hit onwards.First occurance OSD's creating without any issue.======================================= ## ceph-disk prepare --bluestore --cluster ceph --cluster-uuid b2f1b9b9-eecc-4c17-8b92-cfa60b31c121 /dev/sde; ceph-disk activate /dev/sde1 /usr/lib/python2.7/site-packages/ceph_disk/main.py: 5653: UserWarning: ************************************************************ ******************* This tool is now deprecated in favor of ceph-volume.It is recommended to use ceph-volume for OSD deployments. For details see:************************************************************ ******************* warnings.warn(DEPRECATION_WARNING) Creating new GPT entries.The operation has completed successfully.The operation has completed successfully.The operation has completed successfully.meta-data="" isize=2048 agcount=4, agsize=6336 blks= sectsz=512 attr=2, projid32bit=1= crc=1 finobt=0, sparse=0data = bsize=4096 blocks=25344, imaxpct=25= sunit=64 swidth=64 blksnaming =version 2 bsize=4096 ascii-ci=0 ftype=1log =internal log bsize=4096 blocks=1728, version=2= sectsz=512 sunit=64 blks, lazy-count=1realtime =none extsz=4096 blocks=0, rtextents=0Warning: The kernel is still using the old partition table.The new table will be used at the next reboot.The operation has completed successfully./usr/lib/python2.7/site-packages/ceph_disk/main.py: 5685: UserWarning: ************************************************************ ******************* This tool is now deprecated in favor of ceph-volume.It is recommended to use ceph-volume for OSD deployments. For details see:************************************************************ ******************* warnings.warn(DEPRECATION_WARNING) /usr/lib/python2.7/site-packages/ceph_disk/main.py: 5653: UserWarning: ************************************************************ ******************* This tool is now deprecated in favor of ceph-volume.It is recommended to use ceph-volume for OSD deployments. For details see:************************************************************ ******************* warnings.warn(DEPRECATION_WARNING) got monmap epoch 32017-12-08 16:07:57.262854 7fe58f6e6d00 -1 key2017-12-08 16:07:57.769048 7fe58f6e6d00 -1 created object store /var/lib/ceph/tmp/mnt.dTiXMX for osd.16 fsid b2f1b9b9-eecc-4c17-8b92-cfa60b31c121 Removed symlink /run/systemd/system/ceph-osd.target.wants/ceph-osd@16. service. Created symlink from /run/systemd/system/ceph-osd.target.wants/ceph-osd@16. service to /usr/lib/systemd/system/ceph- osd@.service. ---On the second occurance. , I'm getting below issue...======================================== # ceph-disk prepare --bluestore --cluster ceph --cluster-uuid b2f1b9b9-eecc-4c17-8b92-cfa60b31c121 /dev/sde; ceph-disk activate /dev/sde1 /usr/lib/python2.7/site-packages/ceph_disk/main.py: 5653: UserWarning: ************************************************************ ******************* This tool is now deprecated in favor of ceph-volume.It is recommended to use ceph-volume for OSD deployments. For details see:************************************************************ ******************* warnings.warn(DEPRECATION_WARNING) Creating new GPT entries.The operation has completed successfully.The operation has completed successfully.The operation has completed successfully.meta-data="" isize=2048 agcount=4, agsize=6336 blks= sectsz=512 attr=2, projid32bit=1= crc=1 finobt=0, sparse=0data = bsize=4096 blocks=25344, imaxpct=25= sunit=64 swidth=64 blksnaming =version 2 bsize=4096 ascii-ci=0 ftype=1log =internal log bsize=4096 blocks=1728, version=2= sectsz=512 sunit=64 blks, lazy-count=1realtime =none extsz=4096 blocks=0, rtextents=0Warning: The kernel is still using the old partition table.The new table will be used at the next reboot.The operation has completed successfully./usr/lib/python2.7/site-packages/ceph_disk/main.py: 5685: UserWarning: ************************************************************ ******************* This tool is now deprecated in favor of ceph-volume.It is recommended to use ceph-volume for OSD deployments. For details see:************************************************************ ******************* warnings.warn(DEPRECATION_WARNING) /usr/lib/python2.7/site-packages/ceph_disk/main.py: 5653: UserWarning: ************************************************************ ******************* This tool is now deprecated in favor of ceph-volume.It is recommended to use ceph-volume for OSD deployments. For details see:************************************************************ ******************* warnings.warn(DEPRECATION_WARNING) got monmap epoch 32017-12-08 16:09:07.518454 7fa64e12fd00 -1 bluestore(/var/lib/ceph/tmp/mnt.7x0kCL/block) _check_or_set_bdev_label bdev /var/lib/ceph/tmp/mnt.7x0kCL/ block fsid 54954cfd-b7f3-4f74-9b2e- 2ef57c5143cc does not match our fsid 29262e99-12ff-4c45-9113- 8f69830a1a5e 2017-12-08 16:09:07.772688 7fa64e12fd00 -1 bluestore(/var/lib/ceph/tmp/mnt.7x0kCL) mkfs fsck found fatal error: (5) Input/output error 2017-12-08 16:09:07.772723 7fa64e12fd00 -1 OSD::mkfs: ObjectStore::mkfs failed with error (5) Input/output error2017-12-08 16:09:07.772823 7fa64e12fd00 -1 ** ERROR: error creating empty object store in /var/lib/ceph/tmp/mnt.7x0kCL: (5) Input/output errormount_activate: Failed to activateTraceback (most recent call last):File "/usr/sbin/ceph-disk", line 9, in <module>load_entry_point('ceph-disk==1.0.0', 'console_scripts', 'ceph-disk')() File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 5736, in run main(sys.argv[1:])File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 5682, in main main_catch(args.func, args)File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 5710, in main_catch func(args)File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 3761, in main_activate reactivate=args.reactivate,File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 3524, in mount_activate (osd_id, cluster) = activate(path, activate_key_template, init)File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 3701, in activate keyring=keyring,File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 3153, in mkfs '--setgroup', get_ceph_group(),File "/usr/lib/python2.7/site-packages/ceph_disk/main.py", line 570, in command_check_call return subprocess.check_call(arguments) File "/usr/lib64/python2.7/subprocess.py", line 542, in check_call raise CalledProcessError(retcode, cmd)subprocess.CalledProcessError: Command '['/usr/bin/ceph-osd', '--cluster', 'ceph', '--mkfs', '-i', u'16', '--monmap', '/var/lib/ceph/tmp/mnt.7x0kCL/activate.monmap', '--osd-data', '/var/lib/ceph/tmp/mnt.7x0kCL' , '--osd-uuid', u'29262e99-12ff-4c45-9113- 8f69830a1a5e', '--setuser', 'ceph', '--setgroup', 'ceph']' returned non-zero exit status 1 Do you think is this a bug ? Even I can reproducable it everytime on any osd's . On the first time I can recreate OSD's without any problem. second time onwards the issue happening.Need your views on this ... ?ThanksJayaram
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com