Using 9.1.0 I am getting the error shown below at ceph-deploy osd activate time. + ceph-deploy --overwrite-conf osd activate Intel-2P-Sandy-Bridge-04:/var/local//dev/sdf2:/dev/sdf1 ... [][WARNIN] INFO:ceph-disk:Running command: /usr/bin/ceph-osd --cluster ceph --mkfs --mkkey -i 4 --monmap /var/local/\ /dev/sdf2/activate.monmap --osd-data /var/local//dev/sdf2 --osd-journal /var/local//dev/sdf2/journal --osd-uuid 204865df-8dbf-4f26-91f2-5dfa7c3a49f8 --keyring /var/local//dev/sdf2/keyring --setuser ceph --setgroup ceph [][WARNIN] 2015-10-16 13:13:41.464615 7f3f40642940 -1 filestore(/var/local//dev/sdf2) mkjournal error creating journ\ al on /var/local//dev/sdf2/journal: (13) Permission denied [][WARNIN] 2015-10-16 13:13:41.464635 7f3f40642940 -1 OSD::mkfs: ObjectStore::mkfs failed with error -13 [][WARNIN] 2015-10-16 13:13:41.464669 7f3f40642940 -1 ** ERROR: error creating empty object store in /var/local//de\ v/sdf2: (13) Permission denied [][WARNIN] Traceback (most recent call last): [][WARNIN] File "/usr/sbin/ceph-disk", line 3576, in <module> [][WARNIN] main(sys.argv[1:]) [][WARNIN] File "/usr/sbin/ceph-disk", line 3530, in main [][WARNIN] args.func(args) [][WARNIN] File "/usr/sbin/ceph-disk", line 2432, in main_activate [][WARNIN] init=args.mark_init, [][WARNIN] File "/usr/sbin/ceph-disk", line 2258, in activate_dir [][WARNIN] (osd_id, cluster) = activate(path, activate_key_template, init) [][WARNIN] File "/usr/sbin/ceph-disk", line 2360, in activate [][WARNIN] keyring=keyring, [][WARNIN] File "/usr/sbin/ceph-disk", line 1950, in mkfs [][WARNIN] '--setgroup', get_ceph_user(), [][WARNIN] File "/usr/sbin/ceph-disk", line 349, in command_check_call [][WARNIN] return subprocess.check_call(arguments) [][WARNIN] File "/usr/lib/python2.7/subprocess.py", line 540, in check_call [][WARNIN] raise CalledProcessError(retcode, cmd) [][WARNIN] subprocess.CalledProcessError: Command '['/usr/bin/ceph-osd', '--cluster', 'ceph', '--mkfs', '--mkkey', ' -i', '4', '--monmap', '/var/local//dev/sdf2/activate.monmap', '--osd-data', '/var/local//dev/sdf2', '--osd-journal', '/var/local//dev/sdf2/journal', '--osd-uuid', '204865df-8dbf-4f26-91f2-5dfa7c3a49f8', '--keyring', '/var/local//dev/sdf2/keyring', '--setuser', 'ceph', '--setgroup\ ', 'ceph']' returned non-zero exit status 1 [][ERROR ] RuntimeError: command returned non-zero exit status: 1 [ceph_deploy][ERROR ] RuntimeError: Failed to execute command: ceph-disk -v activate --mark-init upstart --mount /var/local//dev/sdf2 When I look at the data disk, I see the following. -rw-r--r-- 1 root ceph 210 Oct 16 13:13 activate.monmap -rw-r--r-- 1 ceph ceph 37 Oct 16 13:13 ceph_fsid drwxr-sr-x 3 ceph ceph 4096 Oct 16 13:13 current -rw-r--r-- 1 ceph ceph 37 Oct 16 13:13 fsid lrwxrwxrwx 1 root ceph 9 Oct 16 13:13 journal -> /dev/sdf1 -rw-r--r-- 1 ceph ceph 21 Oct 16 13:13 magic -rw-r--r-- 1 ceph ceph 4 Oct 16 13:13 store_version -rw-r--r-- 1 ceph ceph 53 Oct 16 13:13 superblock -rw-r--r-- 1 ceph ceph 2 Oct 16 13:13 whoami (The parent directory has drwxr-sr-x 3 ceph ceph 4096 Oct 16 13:13 sdf2) I had been creating the partitions myself and then passing them to ceph-deploy osd prepare and osd activate. Which worked fine before 9.1.0. Is there some extra permissions setup I need to do for 9.1.0? Alternatively, is there a single-node setup script for 9.1.0 that I can look at? -- Tom Deneau -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html