osd activation under 9.1.0

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Using 9.1.0 I am getting the error shown below at ceph-deploy osd activate time.

+ ceph-deploy --overwrite-conf osd activate Intel-2P-Sandy-Bridge-04:/var/local//dev/sdf2:/dev/sdf1
...
[][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd --cluster ceph --mkfs --mkkey -i 4 --monmap /var/local/\
/dev/sdf2/activate.monmap --osd-data /var/local//dev/sdf2 --osd-journal /var/local//dev/sdf2/journal --osd-uuid 204865df-8dbf-4f26-91f2-5dfa7c3a49f8 --keyring /var/local//dev/sdf2/keyring --setuser ceph --setgroup ceph
[][WARNIN] 2015-10-16 13:13:41.464615 7f3f40642940 -1 filestore(/var/local//dev/sdf2) mkjournal error creating journ\
al on /var/local//dev/sdf2/journal: (13) Permission denied
[][WARNIN] 2015-10-16 13:13:41.464635 7f3f40642940 -1 OSD::mkfs: ObjectStore::mkfs failed with error -13
[][WARNIN] 2015-10-16 13:13:41.464669 7f3f40642940 -1  ** ERROR: error creating empty object store in /var/local//de\
v/sdf2: (13) Permission denied
[][WARNIN] Traceback (most recent call last):
[][WARNIN]   File "/usr/sbin/ceph-disk", line 3576, in <module>
[][WARNIN]     main(sys.argv[1:])
[][WARNIN]   File "/usr/sbin/ceph-disk", line 3530, in main
[][WARNIN]     args.func(args)
[][WARNIN]   File "/usr/sbin/ceph-disk", line 2432, in main_activate
[][WARNIN]     init=args.mark_init,
[][WARNIN]   File "/usr/sbin/ceph-disk", line 2258, in activate_dir
[][WARNIN]     (osd_id, cluster) = activate(path, activate_key_template, init)
[][WARNIN]   File "/usr/sbin/ceph-disk", line 2360, in activate
[][WARNIN]     keyring=keyring,
[][WARNIN]   File "/usr/sbin/ceph-disk", line 1950, in mkfs
[][WARNIN]     '--setgroup', get_ceph_user(),
[][WARNIN]   File "/usr/sbin/ceph-disk", line 349, in command_check_call
[][WARNIN]     return subprocess.check_call(arguments)
[][WARNIN]   File "/usr/lib/python2.7/subprocess.py", line 540, in check_call
[][WARNIN]     raise CalledProcessError(retcode, cmd)
[][WARNIN] subprocess.CalledProcessError: Command '['/usr/bin/ceph-osd', '--cluster', 'ceph', '--mkfs', '--mkkey', '
-i', '4', '--monmap', '/var/local//dev/sdf2/activate.monmap', '--osd-data', '/var/local//dev/sdf2', '--osd-journal', '/var/local//dev/sdf2/journal', '--osd-uuid', '204865df-8dbf-4f26-91f2-5dfa7c3a49f8', '--keyring', '/var/local//dev/sdf2/keyring', '--setuser', 'ceph', '--setgroup\
', 'ceph']' returned non-zero exit status 1
[][ERROR ] RuntimeError: command returned non-zero exit status: 1
[ceph_deploy][ERROR ] RuntimeError: Failed to execute command: ceph-disk -v activate --mark-init upstart --mount /var/local//dev/sdf2

When I look at the data disk, I see the following.  

  -rw-r--r-- 1 root ceph   210 Oct 16 13:13 activate.monmap
  -rw-r--r-- 1 ceph ceph    37 Oct 16 13:13 ceph_fsid
  drwxr-sr-x 3 ceph ceph  4096 Oct 16 13:13 current
  -rw-r--r-- 1 ceph ceph    37 Oct 16 13:13 fsid
  lrwxrwxrwx 1 root ceph     9 Oct 16 13:13 journal -> /dev/sdf1
  -rw-r--r-- 1 ceph ceph    21 Oct 16 13:13 magic
  -rw-r--r-- 1 ceph ceph     4 Oct 16 13:13 store_version
  -rw-r--r-- 1 ceph ceph    53 Oct 16 13:13 superblock
  -rw-r--r-- 1 ceph ceph     2 Oct 16 13:13 whoami

(The parent directory has
  drwxr-sr-x 3 ceph ceph  4096 Oct 16 13:13 sdf2)

I had been creating the partitions myself and then passing them to ceph-deploy osd prepare and osd activate.
Which worked fine before 9.1.0.
Is there some extra permissions setup I need to do for 9.1.0?

Alternatively, is there a single-node setup script for 9.1.0 that I can look at?

-- Tom Deneau

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux