On Fri, Sep 27, 2013 at 4:59 PM, Piers Dawson-Damer <piers@xxxxx> wrote: > Hi, > > I'm trying to setup my first cluster, (have never manually bootstrapped a > cluster) > > Is ceph-deploy odd activate/prepare supposed to write to the master > ceph.conf file, specific entries for each OSD along the lines of > http://ceph.com/docs/master/rados/configuration/osd-config-ref/ ? No, ceph-deploy will not write specifc sections for OSDs > > I appear to have the OSDs prepared without error, but then.. no OSD entries > in master cepf.conf nor node /etc/cepf.conf > > Am I missing something? > > Thanks in advance, > > Piers Dawson-Damer > Tasmania > > > 2013-09-28 06:47:00,471 [ceph_deploy.sudo_pushy][DEBUG ] will use a remote > connection with sudo > 2013-09-28 06:47:01,205 [ceph_deploy.osd][INFO ] Distro info: Ubuntu 12.04 > precise > 2013-09-28 06:47:01,205 [ceph_deploy.osd][DEBUG ] Preparing host > storage03-vs-e2 disk /dev/sdm journal /dev/mapper/ceph_journal-osd_12 > activate True > 2013-09-28 06:47:01,206 [storage03-vs-e2][INFO ] Running command: > ceph-disk-prepare --cluster ceph -- /dev/sdm /dev/mapper/ceph_journal-osd_12 > 2013-09-28 06:47:20,247 [storage03-vs-e2][INFO ] Information: Moved > requested sector from 4194338 to 4196352 in > 2013-09-28 06:47:20,248 [storage03-vs-e2][INFO ] order to align on > 2048-sector boundaries. > 2013-09-28 06:47:20,248 [storage03-vs-e2][INFO ] Warning: The kernel is > still using the old partition table. > 2013-09-28 06:47:20,248 [storage03-vs-e2][INFO ] The new table will be used > at the next reboot. > 2013-09-28 06:47:20,248 [storage03-vs-e2][INFO ] The operation has > completed successfully. > 2013-09-28 06:47:20,248 [storage03-vs-e2][INFO ] Information: Moved > requested sector from 34 to 2048 in > 2013-09-28 06:47:20,249 [storage03-vs-e2][INFO ] order to align on > 2048-sector boundaries. > 2013-09-28 06:47:20,249 [storage03-vs-e2][INFO ] The operation has > completed successfully. > 2013-09-28 06:47:20,249 [storage03-vs-e2][INFO ] meta-data=/dev/sdm1 > isize=2048 agcount=4, agsize=183105343 blks > 2013-09-28 06:47:20,250 [storage03-vs-e2][INFO ] = > sectsz=512 attr=2, projid32bit=0 > 2013-09-28 06:47:20,250 [storage03-vs-e2][INFO ] data = > bsize=4096 blocks=732421371, imaxpct=5 > 2013-09-28 06:47:20,250 [storage03-vs-e2][INFO ] = > sunit=0 swidth=0 blks > 2013-09-28 06:47:20,250 [storage03-vs-e2][INFO ] naming =version 2 > bsize=4096 ascii-ci=0 > 2013-09-28 06:47:20,251 [storage03-vs-e2][INFO ] log =internal log > bsize=4096 blocks=357627, version=2 > 2013-09-28 06:47:20,251 [storage03-vs-e2][INFO ] = > sectsz=512 sunit=0 blks, lazy-count=1 > 2013-09-28 06:47:20,251 [storage03-vs-e2][INFO ] realtime =none > extsz=4096 blocks=0, rtextents=0 > 2013-09-28 06:47:20,251 [storage03-vs-e2][INFO ] The operation has > completed successfully. > 2013-09-28 06:47:20,252 [storage03-vs-e2][ERROR ] WARNING:ceph-disk:OSD will > not be hot-swappable if journal is not the same device as the osd data > 2013-09-28 06:47:20,266 [storage03-vs-e2][INFO ] Running command: udevadm > trigger --subsystem-match=block --action=add > 2013-09-28 06:47:20,413 [ceph_deploy.osd][DEBUG ] Host storage03-vs-e2 is > now ready for osd use. > > > > > > 2013-09-27 10:13:25,349 [storage03-vs-e2][DEBUG ] status for monitor: > mon.storage03-vs-e2 > 2013-09-27 10:13:25,349 [storage03-vs-e2][DEBUG ] { "name": > "storage03-vs-e2", > 2013-09-27 10:13:25,350 [storage03-vs-e2][DEBUG ] "rank": 2, > 2013-09-27 10:13:25,350 [storage03-vs-e2][DEBUG ] "state": "electing", > 2013-09-27 10:13:25,350 [storage03-vs-e2][DEBUG ] "election_epoch": 1, > 2013-09-27 10:13:25,351 [storage03-vs-e2][DEBUG ] "quorum": [], > 2013-09-27 10:13:25,351 [storage03-vs-e2][DEBUG ] "outside_quorum": [], > 2013-09-27 10:13:25,351 [storage03-vs-e2][DEBUG ] "extra_probe_peers": [ > 2013-09-27 10:13:25,351 [storage03-vs-e2][DEBUG ] > "172.17.181.47:6789\/0", > 2013-09-27 10:13:25,352 [storage03-vs-e2][DEBUG ] > "172.17.181.48:6789\/0"], > 2013-09-27 10:13:25,352 [storage03-vs-e2][DEBUG ] "sync_provider": [], > 2013-09-27 10:13:25,352 [storage03-vs-e2][DEBUG ] "monmap": { "epoch": 0, > 2013-09-27 10:13:25,352 [storage03-vs-e2][DEBUG ] "fsid": > "28626c0a-0266-4b80-8c06-0562bf48b793", > 2013-09-27 10:13:25,353 [storage03-vs-e2][DEBUG ] "modified": > "0.000000", > 2013-09-27 10:13:25,353 [storage03-vs-e2][DEBUG ] "created": > "0.000000", > 2013-09-27 10:13:25,353 [storage03-vs-e2][DEBUG ] "mons": [ > 2013-09-27 10:13:25,353 [storage03-vs-e2][DEBUG ] { "rank": 0, > 2013-09-27 10:13:25,354 [storage03-vs-e2][DEBUG ] "name": > "storage01-vs-e2", > 2013-09-27 10:13:25,354 [storage03-vs-e2][DEBUG ] "addr": > "172.17.181.47:6789\/0"}, > 2013-09-27 10:13:25,354 [storage03-vs-e2][DEBUG ] { "rank": 1, > 2013-09-27 10:13:25,354 [storage03-vs-e2][DEBUG ] "name": > "storage02-vs-e2", > 2013-09-27 10:13:25,355 [storage03-vs-e2][DEBUG ] "addr": > "172.17.181.48:6789\/0"}, > 2013-09-27 10:13:25,355 [storage03-vs-e2][DEBUG ] { "rank": 2, > 2013-09-27 10:13:25,355 [storage03-vs-e2][DEBUG ] "name": > "storage03-vs-e2", > 2013-09-27 10:13:25,355 [storage03-vs-e2][DEBUG ] "addr": > "172.17.181.49:6789\/0"}]}} > 2013-09-27 10:13:25,356 [storage03-vs-e2][DEBUG ] > 2013-09-27 10:13:25,356 [storage03-vs-e2][DEBUG ] > ******************************************************************************** > 2013-09-27 10:13:25,356 [storage03-vs-e2][INFO ] monitor: > mon.storage03-vs-e2 is running > > > > > > 2013-09-27 10:12:17,384 [storage03-vs-e2.groupthink.cc][INFO ] ceph-all > start/running > 2013-09-27 10:12:17,384 [storage03-vs-e2.groupthink.cc][INFO ] Setting up > ceph-fs-common (0.67.3-1precise) ... > 2013-09-27 10:12:17,384 [storage03-vs-e2.groupthink.cc][INFO ] Setting up > ceph-mds (0.67.3-1precise) ... > 2013-09-27 10:12:17,384 [storage03-vs-e2.groupthink.cc][INFO ] ceph-mds-all > start/running > 2013-09-27 10:12:17,395 [storage03-vs-e2.groupthink.cc][INFO ] Running > command: ceph --version > 2013-09-27 10:12:17,547 [storage03-vs-e2.groupthink.cc][INFO ] ceph version > 0.67.3 (408cd61584c72c0d97b774b3d8f95c6b1b06341a) > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com