Re: ceph new osd addition and client disconnected

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Taylor,

   Details are below.

ceph -s
    cluster 944fa0af-b7be-45a9-93ff-b9907cfaee3f
     health HEALTH_OK
     monmap e2: 3 mons at {integ-hm5=192.168.112.192:6789/0,integ-hm6=192.168.112.193:6789/0,integ-hm7=192.168.112.194:6789/0}
            election epoch 526, quorum 0,1,2 integ-hm5,integ-hm6,integ-hm7
     osdmap e50127: 3 osds: 3 up, 3 in
      pgmap v2923439: 190 pgs, 2 pools, 3401 GB data, 920 kobjects
            6711 GB used, 31424 GB / 40160 GB avail
                 190 active+clean
 client io 35153 kB/s rd, 1912 kB/s wr, 672 op/s

Client is automatically unmounted in our cause.

Is it possible to change the PG_num in the production setup.

Journal stored on SATA 7.2k RPM  6GPS and 1gb network interface.

We are not configured Public and cluster as a separate network and it will be transferable via same LAN. Do we need to do this setup for better performance.

Also what is beter i/o operation setting for the crush map.


Still we are getting errors in ceph osd logs ,what need to done for this error.

2015-11-03 13:04:18.809488 7f387019c700  0 bad crc in data 3742210963 != exp 924878202
2015-11-03 13:04:18.812911 7f387019c700  0 -- 192.168.112.231:6800/49908 >> 192.168.112.192:0/1457324982 pipe(0x170d2000 sd=44 :6800 s=0 pgs=0 cs=0 l=0 c=0x1b18bf40).accept peer addr is really 192.168.112.192:0/1457324982 (socket is 192.168.112.192:47128/0)


Regards
Prabu




---- On Tue, 03 Nov 2015 12:50:40 +0530 Chris Taylor <ctaylor@xxxxxxxxxx> wrote ----

On 2015-11-02 10:19 pm, gjprabu wrote:

> Hi Taylor,
>
> I have checked DNS name and all host resolve to the correct IP. MTU
> size is 1500 in switch level configuration done. There is no firewall/
> selinux is running currently.
>
> Also we would like to know below query's which already in the thread.
>
> Regards
> Prabu
>
> ---- On Tue, 03 Nov 2015 11:20:07 +0530 CHRIS TAYLOR
> <CTAYLOR@xxxxxxxxxx> wrote ----
>
> I would double check the network configuration on the new node.
> Including hosts files and DNS names. Do all the host names resolve to
> the correct IP addresses from all hosts?
>
> "... 192.168.112.231:6800/49908 >> 192.168.113.42:0/599324131 ..."
>
> Looks like the communication between subnets is a problem. Is
> xxx.xxx.113.xxx a typo? If that's correct, check MTU sizes. Are they
> configured correctly on the switch and all NICs?
>
> Is there any iptables/firewall rules that could be blocking traffic
> between hosts?
>
> Hope that helps,
>
> Chris
>
> On 2015-11-02 9:18 pm, gjprabu wrote:
>
> Hi,
>
> Anybody please help me on this issue.
>
> Regards
> Prabu
>
> ---- On Mon, 02 Nov 2015 17:54:27 +0530 GJPRABU <GJPRABU@xxxxxxxxxxxx>
> wrote ----
>
> Hi Team,
>
> We have ceph setup with 2 OSD and replica 2 and it is mounted with
> ocfs2 clients and its working. When we added new osd all the clients
> rbd mapped device disconnected and got hanged by running rbd ls or rbd
> map command. We waited for long hours to scale the new osd size but
> peering not completed event data sync finished, but client side issue
> was persist and thought to try old osd service stop/start, after some
> time rbd mapped automatically using existing map script.
>
> After service stop/start in old osd again 3rd OSD rebuild and back
> filling started and after some time clients rbd mapped device
> disconnected and got hanged by running rbd ls or rbd map command. We
> thought to wait till to finished data sync in 3'rd OSD and its
> completed, even though client side rbd not mapped. After we restarted
> all mon and osd service and client side issue got fixed and mounted
> rbd. We suspected some issue in our setup. also attached logs for your
> reference.
>

What does 'ceph -s' look like? is the cluster HEALTH_OK?

>
> Something we are missing in our setup i don't know, highly appreciated
> if anybody help us to solve this issue.
>
> Before new osd.2 addition :
>
> osd.0 - size : 13T and used 2.7 T
> osd.1 - size : 13T and used 2.7 T
>
> After new osd addition :
> osd.0 size : 13T and used 1.8T
> osd.1 size : 13T and used 2.1T
> osd.2 size : 15T and used 2.5T
>
> rbd ls
> repo / integrepository (pg_num: 126)
> rbd / integdownloads (pg_num: 64)
>
> Also we would like to know few clarifications .
>
> If any new osd will be added whether all client will be unmounted
> automatically .
>

Clients do not need to unmount images when OSDs are added.

> While add new osd can we access ( read / write ) from client machines ?
>

Clients still have read/write access to RBD images in the cluster while
adding OSDs and during recovery.

> How much data will be added in new osd - without change any repilca /
> pg_num ?
>

The data will re-balance between OSDs automatically. I found having more
PGs help distribute the load more evenly.

> How long to take finish this process ?

Depends greatly on the hardware and configuration. Whether Journals on
SSD or spinning disks, network connectivity, max_backfills, etc.

>
> If we missed any common configuration - please share the same .

I don't see any configuration for public and cluster networks. If you
are sharing the same network for clients and object replication/recovery
the cluster re-balancing data between OSDs could cause problems with the
client traffic.

Take a look at:

>
> ceph.conf
> [global]
> fsid = 944fa0af-b7be-45a9-93ff-b9907cfaee3f
> mon_initial_members = integ-hm5, integ-hm6, integ-hm7
> mon_host = 192.168.112.192,192.168.112.193,192.168.112.194
> auth_cluster_required = cephx
> auth_service_required = cephx
> auth_client_required = cephx
> filestore_xattr_use_omap = true
> osd_pool_default_size = 2
>
> [mon]
> mon_clock_drift_allowed = .500
>
> [client]
> rbd_cache = false
>
> Current Logs from new osd also attached old logs.
>
> 2015-11-02 12:47:48.481641 7f386f691700 0 bad crc in data 3889133030 !=
> exp 2857248268
> 2015-11-02 12:47:48.482230 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc510580).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42530/0)
> 2015-11-02 12:47:48.483951 7f386f691700 0 bad crc in data 3192803598 !=
> exp 1083014631
> 2015-11-02 12:47:48.484512 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x170ea000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc516f60).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42531/0)
> 2015-11-02 12:47:48.486284 7f386f691700 0 bad crc in data 133120597 !=
> exp 393328400
> 2015-11-02 12:47:48.486777 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x16a18000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc514620).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42532/0)
> 2015-11-02 12:47:48.488624 7f386f691700 0 bad crc in data 3299720069 !=
> exp 211350069
> 2015-11-02 12:47:48.489100 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc513860).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42533/0)
> 2015-11-02 12:47:48.490911 7f386f691700 0 bad crc in data 2381447347 !=
> exp 1177846878
> 2015-11-02 12:47:48.491390 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x170ea000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc513700).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42534/0)
> 2015-11-02 12:47:48.493167 7f386f691700 0 bad crc in data 2093712440 !=
> exp 2175112954
> 2015-11-02 12:47:48.493682 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x16a18000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc514200).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42535/0)
> 2015-11-02 12:47:48.495150 7f386f691700 0 bad crc in data 3047197039 !=
> exp 38098198
> 2015-11-02 12:47:48.495679 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc510b00).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42536/0)
> 2015-11-02 12:47:48.497259 7f386f691700 0 bad crc in data 1400444622 !=
> exp 2648291990
> 2015-11-02 12:47:48.497756 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x170ea000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0x17f7b700).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42537/0)
> 2015-11-02 13:02:00.439025 7f386f691700 0 bad crc in data 4159064831 !=
> exp 903679865
> 2015-11-02 13:02:00.441337 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x16a18000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0x17f7e5c0).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:43128/0)
> 2015-11-02 13:02:00.442756 7f386f691700 0 bad crc in data 1134831440 !=
> exp 892008036
> 2015-11-02 13:02:00.443369 7f386f691700 0 -- 192.168.112.231:6800/49908
> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0x17f7ee00).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:43129/0)
> 2015-11-02 13:08:43.272527 7f387049f700 0 -- 192.168.112.231:6800/49908
> >> 192.168.112.115:0/4256128918 pipe(0x170ea000 sd=33 :6800 s=0 pgs=0 cs=0 l=0 c=0x17f7e1a0).accept peer addr is really 192.168.112.115:0/4256128918 (socket is 192.168.112.115:51660/0)
>
> Regards
> Prabu
>
> Regards
> G.J
>
> _______________________________________________
> ceph-users mailing list
>
> _______________________________________________
> ceph-users mailing list



Links:
------

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux