Re: "clocks not synchronized" warning message

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I'm experiencing the same. The default "mon lease wiggle room" is set
t .001, which gives the messages in my setup.

Right now i've set the room to 0.5, which gets rid of the messages, but
obviously this is imho not the way to go.

My nodes are running openntpd and a cron which runs ntpdate every hour.

-- 
Met vriendelijke groet,

Wido den Hollander
Hoofd Systeembeheer / CSO
Telefoon Support Nederland: 0900 9633 (45 cpm)
Telefoon Support België: 0900 70312 (45 cpm)
Telefoon Direct: (+31) (0)20 50 60 104
Fax: +31 (0)20 50 60 111
E-mail: support@xxxxxxxxxxxx
Website: http://www.pcextreme.nl
Kennisbank: http://support.pcextreme.nl/
Netwerkstatus: http://nmc.pcextreme.nl


On Thu, 2010-07-08 at 12:54 +0000, Thomas Mueller wrote:
> hi
> 
> running ceph -w I see many of those warnings flying by:
> 
> 10.07.08_14:36:49.897108   log 10.07.08_14:36:48.530057 mon1 192.168.1.52:6789/0 340 : [WRN] lease_expire from mon0 was sent from future time 10.07.08_14:36:48.886876, clocks not synchronized
> 10.07.08_14:36:49.897108   log 10.07.08_14:36:48.530076 mon1 192.168.1.52:6789/0 341 : [WRN] 
> 
> i've stopped ceph, ran ntpdate on every node, started ceph 
> - the message is still there. 
> 
> date from the nodes:
> # for i in node001 node002 node003 node004; do ssh root@$i date; done
> Thu Jul  8 14:52:20 CEST 2010
> Thu Jul  8 14:52:20 CEST 2010
> Thu Jul  8 14:52:20 CEST 2010
> Thu Jul  8 14:52:21 CEST 2010
> 
> 
> this is on ceph.git/unstable from today.
> 
> - Thomas
> 
> 
> 
> ceph.conf:
> 
> 
> [global]
> 	; enable secure authentication
> 	; auth supported = cephx
> 
> [mon]
> 	mon data = /ceph/mon$id
> 
> 	; some minimal logging (just message traffic) to aid debugging
> 	;debug ms = 1
> 
> [mon0]
> 	host = node001
> 	mon addr = 192.168.1.51:6789
> 
> [mon1]
>         host = node002
>         mon addr = 192.168.1.52:6789
> 
> [mon2]
>         host = node003
>         mon addr = 192.168.1.53:6789
> 
> [mds]
> 	; where the mds keeps it's secret encryption keys
> 	keyring = /ceph/keyring.$name
> 
> [mds.node001]
> 	host = node001
> 
> [mds.node002]
> 	host = node002
> 
> [mds.node003]
>         host = node003
> 
> ; osd
> ;  You need at least one.  Two if you want data to be replicated.
> ;  Define as many as you like.
> [osd]
> 	; This is where the btrfs volume will be mounted.
> 	osd data = /ceph/osd$id
> 
> 	; Ideally, make this a separate disk or partition.  A few GB
>  	; is usually enough; more if you have fast disks.  You can use
>  	; a file under the osd data dir if need be
>  	; (e.g. /ceph/osd$id/journal), but it will be slower than a
>  	; separate disk or partition.
> 	osd journal = /ceph/osd$id/journal
> 	osd journal size = 100
> 
> [osd0]
> 	host = node001
> 
> 	; if 'btrfs devs' is not specified, you're responsible for
> 	; setting up the 'osd data' dir.  if it is not btrfs, things
> 	; will behave up until you try to recover from a crash (which
> 	; usually fine for basic testing).
> 	;btrfs devs = /dev/sdx
> 
> [osd1]
> 	host = node002
> 	;btrfs devs = /dev/sdy
> 
> [osd2]
> 	host = node003
> 	;btrfs devs = /dev/sdx
> 
> [osd3]
> 	host = node004
> 	;btrfs devs = /dev/sdy
> 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux