hi there are two [global] sections in your ceph.conf i do not know why you didn't have admin_keyring.bin in you /data/mon0, maybe you should restart the whole cluster after you assured configure file is right 在 2011年5月23日 下午3:48,biyan chen <riby.chen@xxxxxxxxx> 写道: > ; > ; Sample ceph ceph.conf file. > ; > ; This file defines cluster membership, the various locations > ; that Ceph stores data, and any other runtime options. > > ; If a 'host' is defined for a daemon, the start/stop script will > ; verify that it matches the hostname (or else ignore it). If it is > ; not defined, it is assumed that the daemon is intended to start on > ; the current host (e.g., in a setup with a startup.conf on each > ; node). > > ; global > [global] > ; enable secure authentication > auth supported = cephx > keyring = /use/local/etc/ceph/keyring.bin > > > ; allow ourselves to open a lot of files > max open files = 131072 > > ; set up logging > "/usr/local/etc/ceph/ceph.conf" 98L, 2601C > ; > ; Sample ceph ceph.conf file. > ; > ; This file defines cluster membership, the various locations > ; that Ceph stores data, and any other runtime options. > > ; If a 'host' is defined for a daemon, the start/stop script will > ; verify that it matches the hostname (or else ignore it). If it is > ; not defined, it is assumed that the daemon is intended to start on > ; the current host (e.g., in a setup with a startup.conf on each > ; node). > > ; global > [global] > ; enable secure authentication > auth supported = cephx > keyring = /use/local/etc/ceph/keyring.bin > > > ; allow ourselves to open a lot of files > max open files = 131072 > > ; set up logging > log file = /var/log/ceph/$name.log > > ; set up pid files > pid file = /var/run/ceph/$name.pid > > ; monitors > ; You need at least one. You need at least three if you want to > ; tolerate any node failures. Always create an odd number. > [mon] > mon data = /data/mon$id > > ; logging, for debugging monitor crashes, in order of > ; their likelihood of being helpful :) > ;debug ms = 1 > ;debug mon = 20 > ;debug paxos = 20 > ;debug auth = 20 > > [mon.0] > host = ceph_mon0 > mon addr = 192.168.0.211:6789 > > ; mds > ; You need at least one. Define two to get a standby. > [mds] > ; where the mds keeps it's secret encryption keys > keyring = /usr/local/etc/ceph/keyring.$name > > ; mds logging to debug issues. > ;debug ms = 1 > ;debug mds = 20 > > [mds.alpha] > host = ceph_mds0 > > ; osd > ; You need at least one. Two if you want data to be replicated. > ; Define as many as you like. > [osd] > ; This is where the btrfs volume will be mounted. > osd data = /data/osd$id > keyring = /usr/local/etc/ceph/keyring.$name > > > ; Ideally, make this a separate disk or partition. A few > ; hundred MB should be enough; more if you have fast or many > ; disks. You can use a file under the osd data dir if need be > ; (e.g. /data/osd$id/journal), but it will be slower than a > ; separate disk or partition. > > ; This is an example of a file-based journal. > osd journal = /data/osd$id/journal > osd journal size = 1000 ; journal size, in megabytes > > ; osd logging to debug osd issues, in order of likelihood of being > ; helpful > ;debug ms = 1 > ;debug osd = 20 > ;debug filestore = 20 > ;debug journal = 20 > > [osd.0] > host = ceph_osd0 > > ; if 'btrfs devs' is not specified, you're responsible for > ; setting up the 'osd data' dir. if it is not btrfs, things > ; will behave up until you try to recover from a crash (which > ; usually fine for basic testing). > btrfs devs = /dev/sda7 > > > [osd.1] > host = ceph_osd1 > btrfs devs = /dev/sda7 > > > > log: > [root@ceph_mon0 ~]# ceph mon stat -c /usr/local/etc/ceph/ceph.conf > 2011-05-23 11:44:47.078662 7fb266bf8720 unable to authenticate as client.admin > 2011-05-23 11:44:47.079037 7fb266bf8720 ceph_tool_common_init failed. > > [root@ceph_mds0 ~]# tail /var/log/ceph/mds.alpha.log -f > 2011-05-23 23:44:53.464262 7fb0ce956710 -- 192.168.0.207:6800/29302 >> > 192.168.0.211:6789/0 pipe(0x26c1a00 sd=5 pgs=0 cs=0 l=0).fault first > fault > 2011-05-23 23:44:56.464402 7fb0ce855710 -- 192.168.0.207:6800/29302 >> > 192.168.0.211:6789/0 pipe(0x26d3280 sd=5 pgs=0 cs=0 l=0).fault first > fault > 2011-05-23 23:44:59.464492 7fb0ce956710 -- 192.168.0.207:6800/29302 >> > 192.168.0.211:6789/0 pipe(0x26c3c80 sd=5 pgs=0 cs=0 l=0).fault first > fault > 2011-05-23 23:45:02.464787 7fb0ce855710 -- 192.168.0.207:6800/29302 >> > 192.168.0.211:6789/0 pipe(0x26cec80 sd=6 pgs=0 cs=0 l=0).fault first > fault > 2011-05-23 23:45:05.464941 7fb0cfd58710 mds-1.0 ms_handle_connect on > 192.168.0.211:6789/0 > 2011-05-23 23:45:05.465405 7fb0cfd58710 cannot convert AES key for NSS: -8023 > 2011-05-23 23:45:05.465946 7fb0cfd58710 cannot convert AES key for NSS: -8023 > 2011-05-23 23:45:05.465969 7fb0cfd58710 error from decrypt -22 > 2011-05-23 23:45:05.465998 7fb0cfd58710 cephx: > verify_service_ticket_reply failed decode_decrypt with secret > AQB4R9pNEJ/lFRAAdsRa+EFMA00acC28x9Fj7A== > 2011-05-23 23:45:05.466010 7fb0cfd58710 cephx client: could not verify > service_ticket reply > > [root@ceph_osd0 ~]# tail /var/log/ceph/osd.0.log -f > 2011-05-23 23:43:45.583183 7f5b2f383710 -- 192.168.0.208:6800/30745 >> > 192.168.0.211:6789/0 pipe(0xf79280 sd=12 pgs=0 cs=0 l=0).fault first > fault > 2011-05-23 23:43:48.583357 7f5b32689710 -- 192.168.0.208:6800/30745 >> > 192.168.0.211:6789/0 pipe(0xf79000 sd=12 pgs=0 cs=0 l=0).fault first > fault > 2011-05-23 23:43:51.583412 7f5b2f383710 -- 192.168.0.208:6800/30745 >> > 192.168.0.211:6789/0 pipe(0xf79280 sd=13 pgs=0 cs=0 l=0).fault first > fault > 2011-05-23 23:43:54.583548 7f5b32689710 -- 192.168.0.208:6800/30745 >> > 192.168.0.211:6789/0 pipe(0xf79000 sd=12 pgs=0 cs=0 l=0).fault first > fault > 2011-05-23 23:43:57.583842 7f5b2f383710 -- 192.168.0.208:6800/30745 >> > 192.168.0.211:6789/0 pipe(0xf79280 sd=12 pgs=0 cs=0 l=0).fault first > fault > 2011-05-23 23:44:00.584629 7f5b34f8e710 cannot convert AES key for NSS: -8023 > 2011-05-23 23:44:00.585262 7f5b34f8e710 cannot convert AES key for NSS: -8023 > 2011-05-23 23:44:00.585290 7f5b34f8e710 error from decrypt -22 > 2011-05-23 23:44:00.585312 7f5b34f8e710 cephx: > verify_service_ticket_reply failed decode_decrypt with secret > AQAqR9pNUNJ2IBAAp16KszdsZHCwEf5IOcoSdw== > 2011-05-23 23:44:00.585322 7f5b34f8e710 cephx client: could not verify > service_ticket reply > > Do we have such problems? If you give me some help, > > thank you! > -- > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html