Re: Problem with Ceph installation inside Amazon EC2

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

What does "ceph -w" show?

Do you get some "xauth" messages in your mon0 logfiles?

-- 
Met vriendelijke groet,

Wido den Hollander
Hoofd Systeembeheer / CSO
Telefoon Support Nederland: 0900 9633 (45 cpm)
Telefoon Support België: 0900 70312 (45 cpm)
Telefoon Direct: (+31) (0)20 50 60 104
Fax: +31 (0)20 50 60 111
E-mail: support@xxxxxxxxxxxx
Website: http://www.pcextreme.nl
Kennisbank: http://support.pcextreme.nl/
Netwerkstatus: http://nmc.pcextreme.nl


On Tue, 2010-04-20 at 13:38 +0000, Christian Baun wrote:
> Hi,
> 
> I want to test Ceph inside Amazon EC2. Therefore I compiled and installed ceph-0.19.1.tar.gz and ceph-kclient-0.19.1.tar.gz
> 
> Two Ubuntu instances are running inside EC2. 
> One server with cmon, cmds, and cosd and one client.
> 
> # cat /etc/ceph/ceph.conf
> [global]
>        pid file = /var/run/ceph/$name.pid
>        debug ms = 1
> 
> [mon]
>        mon data = /srv/ceph/mon
> 
> [mon0]
>        host = ceph01
>        mon addr = 10.194.145.203:6789
> 
> [mds]
> 
> [mds0]
>        host = ceph01
> 
> [osd]
>        sudo = true
>        osd data = /srv/ceph/osd
> 
> [osd1]
>        host = ceph01
>        btrfs devs = /dev/sdc
> 
> 
> # cat /etc/hosts
> 127.0.0.1 localhost
> 10.194.145.203   ceph01
> 
> 
> # mkcephfs -c /etc/ceph/ceph.conf --allhosts --mkbtrfs -k ~/admin.keyring
> .conf --allhosts --mkbtrfs -k ~/admin.keyring
> /usr/local/bin/monmaptool --create --clobber --add 10.194.145.203:6789 --print /tmp/monmap.9681
> /usr/local/bin/monmaptool: monmap file /tmp/monmap.9681
> /usr/local/bin/monmaptool: generated fsid 7a008165-58df-733f-deca-90faecaef60a
> epoch 1
> fsid 7a008165-58df-733f-deca-90faecaef60a
> last_changed 10.04.20 13:34:52.518583
> created 10.04.20 13:34:52.518583
> 	mon0 10.194.145.203:6789/0
> /usr/local/bin/monmaptool: writing epoch 1 to /tmp/monmap.9681 (1 monitors)
> max osd in /etc/ceph/ceph.conf is 1, num osd is 2
> /usr/local/bin/osdmaptool: osdmap file '/tmp/osdmap.9681'
> /usr/local/bin/osdmaptool: writing epoch 1 to /tmp/osdmap.9681
> Building admin keyring at /root/admin.keyring
> creating /root/admin.keyring
> Building monitor keyring with all service keys
> creating /tmp/monkeyring.9681
> importing contents of /root/admin.keyring into /tmp/monkeyring.9681
> creating /tmp/keyring.mds.0
> importing contents of /tmp/keyring.mds.0 into /tmp/monkeyring.9681
> creating /tmp/keyring.osd.1
> importing contents of /tmp/keyring.osd.1 into /tmp/monkeyring.9681
> === mon0 === 
> 10.04.20 13:34:54.805739 store(/srv/ceph/mon) mkfs
> 10.04.20 13:34:54.805853 store(/srv/ceph/mon) test -d /srv/ceph/mon && /bin/rm -rf /srv/ceph/mon ; mkdir -p /srv/ceph/mon
> 10.04.20 13:34:54.831017 mon0(starting).class v0 create_initial -- creating initial map
> 10.04.20 13:34:54.832277 mon0(starting).auth v0 create_initial -- creating initial map
> 10.04.20 13:34:54.832308 mon0(starting).auth v0 reading initial keyring 
> /usr/local/bin/mkmonfs: created monfs at /srv/ceph/mon for mon0
> admin.keyring                                    100%  119     0.1KB/s   00:00    
> === mds0 === 
> WARNING: no keyring specified for mds0
> === osd1 === 
> umount: /srv/ceph/osd: not mounted
> umount: /dev/sdc: not mounted
> 
> WARNING! - Btrfs Btrfs v0.19 IS EXPERIMENTAL
> WARNING! - see http://btrfs.wiki.kernel.org before using
> 
> fs created label (null) on /dev/sdc
> 	nodesize 4096 leafsize 4096 sectorsize 4096 size 419.97GB
> Btrfs Btrfs v0.19
> Scanning for Btrfs filesystems
> monmap.9681                                      100%  187     0.2KB/s   00:00    
>  ** WARNING: Ceph is still under heavy development, and is only suitable for **
>  **          testing and review.  Do not trust it with important data.       **
> created object store for osd1 fsid 7a008165-58df-733f-deca-90faecaef60a on /srv/ceph/osd
> WARNING: no keyring specified for osd1
> 
> 
> 
> # df
> Filesystem           1K-blocks      Used Available Use% Mounted on
> /dev/sda1             10403128   4038372   5945328  41% /
> none                   3804316       112   3804204   1% /dev
> none                   3941904         0   3941904   0% /dev/shm
> none                   3941904        60   3941844   1% /var/run
> none                   3941904         0   3941904   0% /var/lock
> none                   3941904         0   3941904   0% /lib/init/rw
> /dev/sdb             433455904    203016 411234584   1% /mnt
> /dev/sdc             440366080        32 440366048   1% /srv/ceph/osd
> 
> 
> # mount | grep ceph
> /dev/sdc on /srv/ceph/osd type btrfs (rw,flushoncommit)
> 
> 
> At the server:
> # ./ceph start
> 10.04.20 13:35:39.757563 7f6485c32720 -- :/10726 messenger.start
> 10.04.20 13:35:39.757808 7f6485c32720 -- :/10726 --> mon0 10.194.145.203:6789/0 -- auth(proto 0 30 bytes) -- ?+0 0x1e46910
> 10.04.20 13:35:39.757990 7f6482eb3710 -- :/10726 >> 10.194.145.203:6789/0 pipe(0x1e426a0 sd=-1 pgs=0 cs=0 l=0).fault first fault
> 
> 
> At the client:
> # mount -t ceph 10.194.145.203:/ /mnt/ceph/
> mount: 10.194.145.203:/: can't read superblock
> 
> 
> What can I do at this point?
> 
> Thanks in advance
>     Christian 
> 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux