Re: [Warning: Forged Email] Ceph 10.2.11 - Status not working

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



mmm wonder why the list is saying my email is forged, wonder what I have
wrong.

My email is sent via an outbound spam filter, but I was sure I had the
SPF set correctly.

Mike

On 18/12/18 10:53 am, Mike O'Connor wrote:
> Hi All
>
> I have a ceph cluster which has been working with out issues for about 2
> years now, it was upgrade about 6 month ago to 10.2.11
>
> root@blade3:/var/lib/ceph/mon# ceph status
> 2018-12-18 10:42:39.242217 7ff770471700  0 -- 10.1.5.203:0/1608630285 >>
> 10.1.5.207:6789/0 pipe(0x7ff768000c80 sd=4 :0 s=1 pgs=0 cs=0 l=1
> c=0x7ff768001f90).fault
> 2018-12-18 10:42:45.242745 7ff770471700  0 -- 10.1.5.203:0/1608630285 >>
> 10.1.5.207:6789/0 pipe(0x7ff7680051e0 sd=3 :0 s=1 pgs=0 cs=0 l=1
> c=0x7ff768002410).fault
> 2018-12-18 10:42:51.243230 7ff770471700  0 -- 10.1.5.203:0/1608630285 >>
> 10.1.5.207:6789/0 pipe(0x7ff7680051e0 sd=3 :0 s=1 pgs=0 cs=0 l=1
> c=0x7ff768002f40).fault
> 2018-12-18 10:42:54.243452 7ff770572700  0 -- 10.1.5.203:0/1608630285 >>
> 10.1.5.205:6789/0 pipe(0x7ff768000c80 sd=4 :0 s=1 pgs=0 cs=0 l=1
> c=0x7ff768008060).fault
> 2018-12-18 10:42:57.243715 7ff770471700  0 -- 10.1.5.203:0/1608630285 >>
> 10.1.5.207:6789/0 pipe(0x7ff7680051e0 sd=3 :0 s=1 pgs=0 cs=0 l=1
> c=0x7ff768003580).fault
> 2018-12-18 10:43:03.244280 7ff7781b9700  0 -- 10.1.5.203:0/1608630285 >>
> 10.1.5.205:6789/0 pipe(0x7ff7680051e0 sd=3 :0 s=1 pgs=0 cs=0 l=1
> c=0x7ff768003670).fault
>
> All system can ping each other. I simple can not see why its failing.
>
>
> ceph.conf
>
> [global]
>      auth client required = cephx
>      auth cluster required = cephx
>      auth service required = cephx
>      cluster network = 10.1.5.0/24
>      filestore xattr use omap = true
>      fsid = 42a0f015-76da-4f47-b506-da5cdacd030f
>      keyring = /etc/pve/priv/$cluster.$name.keyring
>      osd journal size = 5120
>      osd pool default min size = 1
>      public network = 10.1.5.0/24
>      mon_pg_warn_max_per_osd = 0
>
> [client]
>      rbd cache = true
> [osd]
>      keyring = /var/lib/ceph/osd/ceph-$id/keyring
>      osd max backfills = 1
>      osd recovery max active = 1
>      osd_disk_threads = 1
>      osd_disk_thread_ioprio_class = idle
>      osd_disk_thread_ioprio_priority = 7
> [mon.2]
>      host = blade5
>      mon addr = 10.1.5.205:6789
> [mon.1]
>      host = blade3
>      mon addr = 10.1.5.203:6789
> [mon.3]
>      host = blade7
>      mon addr = 10.1.5.207:6789
> [mon.0]
>      host = blade1
>      mon addr = 10.1.5.201:6789
> [mds]
>          mds data = /var/lib/ceph/mds/mds.$id
>          keyring = /var/lib/ceph/mds/mds.$id/mds.$id.keyring
> [mds.0]
>          host = blade1
> [mds.1]
>          host = blade3
> [mds.2]
>          host = blade5
> [mds.3]
>          host = blade7
>
>
> Any ideas ? more information ?
>
>
> Mike
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux