Re: osd activation under 9.1.0

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, 16 Oct 2015, Deneau, Tom wrote:
> Using 9.1.0 I am getting the error shown below at ceph-deploy osd activate time.
> 
> + ceph-deploy --overwrite-conf osd activate Intel-2P-Sandy-Bridge-04:/var/local//dev/sdf2:/dev/sdf1
> ...
> [][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd --cluster ceph --mkfs --mkkey -i 4 --monmap /var/local/\
> /dev/sdf2/activate.monmap --osd-data /var/local//dev/sdf2 --osd-journal /var/local//dev/sdf2/journal --osd-uuid 204865df-8dbf-4f26-91f2-5dfa7c3a49f8 --keyring /var/local//dev/sdf2/keyring --setuser ceph --setgroup ceph
> [][WARNIN] 2015-10-16 13:13:41.464615 7f3f40642940 -1 filestore(/var/local//dev/sdf2) mkjournal error creating journ\
> al on /var/local//dev/sdf2/journal: (13) Permission denied
> [][WARNIN] 2015-10-16 13:13:41.464635 7f3f40642940 -1 OSD::mkfs: ObjectStore::mkfs failed with error -13
> [][WARNIN] 2015-10-16 13:13:41.464669 7f3f40642940 -1  ** ERROR: error creating empty object store in /var/local//de\
> v/sdf2: (13) Permission denied
> [][WARNIN] Traceback (most recent call last):
> [][WARNIN]   File "/usr/sbin/ceph-disk", line 3576, in <module>
> [][WARNIN]     main(sys.argv[1:])
> [][WARNIN]   File "/usr/sbin/ceph-disk", line 3530, in main
> [][WARNIN]     args.func(args)
> [][WARNIN]   File "/usr/sbin/ceph-disk", line 2432, in main_activate
> [][WARNIN]     init=args.mark_init,
> [][WARNIN]   File "/usr/sbin/ceph-disk", line 2258, in activate_dir
> [][WARNIN]     (osd_id, cluster) = activate(path, activate_key_template, init)
> [][WARNIN]   File "/usr/sbin/ceph-disk", line 2360, in activate
> [][WARNIN]     keyring=keyring,
> [][WARNIN]   File "/usr/sbin/ceph-disk", line 1950, in mkfs
> [][WARNIN]     '--setgroup', get_ceph_user(),
> [][WARNIN]   File "/usr/sbin/ceph-disk", line 349, in command_check_call
> [][WARNIN]     return subprocess.check_call(arguments)
> [][WARNIN]   File "/usr/lib/python2.7/subprocess.py", line 540, in check_call
> [][WARNIN]     raise CalledProcessError(retcode, cmd)
> [][WARNIN] subprocess.CalledProcessError: Command '['/usr/bin/ceph-osd', '--cluster', 'ceph', '--mkfs', '--mkkey', '
> -i', '4', '--monmap', '/var/local//dev/sdf2/activate.monmap', '--osd-data', '/var/local//dev/sdf2', '--osd-journal', '/var/local//dev/sdf2/journal', '--osd-uuid', '204865df-8dbf-4f26-91f2-5dfa7c3a49f8', '--keyring', '/var/local//dev/sdf2/keyring', '--setuser', 'ceph', '--setgroup\
> ', 'ceph']' returned non-zero exit status 1
> [][ERROR ] RuntimeError: command returned non-zero exit status: 1
> [ceph_deploy][ERROR ] RuntimeError: Failed to execute command: ceph-disk -v activate --mark-init upstart --mount /var/local//dev/sdf2
> 
> When I look at the data disk, I see the following.  
> 
>   -rw-r--r-- 1 root ceph   210 Oct 16 13:13 activate.monmap
>   -rw-r--r-- 1 ceph ceph    37 Oct 16 13:13 ceph_fsid
>   drwxr-sr-x 3 ceph ceph  4096 Oct 16 13:13 current
>   -rw-r--r-- 1 ceph ceph    37 Oct 16 13:13 fsid
>   lrwxrwxrwx 1 root ceph     9 Oct 16 13:13 journal -> /dev/sdf1

My guess is that /dev/sdf1 needs to be chowned to user ceph.

sage


>   -rw-r--r-- 1 ceph ceph    21 Oct 16 13:13 magic
>   -rw-r--r-- 1 ceph ceph     4 Oct 16 13:13 store_version
>   -rw-r--r-- 1 ceph ceph    53 Oct 16 13:13 superblock
>   -rw-r--r-- 1 ceph ceph     2 Oct 16 13:13 whoami
> 
> (The parent directory has
>   drwxr-sr-x 3 ceph ceph  4096 Oct 16 13:13 sdf2)
> 
> I had been creating the partitions myself and then passing them to ceph-deploy osd prepare and osd activate.
> Which worked fine before 9.1.0.
> Is there some extra permissions setup I need to do for 9.1.0?
> 
> Alternatively, is there a single-node setup script for 9.1.0 that I can look at?
> 
> -- Tom Deneau
> 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux