Hi Taylor,
Details are below.
ceph -s
cluster 944fa0af-b7be-45a9-93ff-b9907cfaee3f
health HEALTH_OK
monmap e2: 3 mons at {integ-hm5=192.168.112.192:6789/0,integ-hm6=192.168.112.193:6789/0,integ-hm7=192.168.112.194:6789/0}
election epoch 526, quorum 0,1,2 integ-hm5,integ-hm6,integ-hm7
osdmap e50127: 3 osds: 3 up, 3 in
pgmap v2923439: 190 pgs, 2 pools, 3401 GB data, 920 kobjects
6711 GB used, 31424 GB / 40160 GB avail
190 active+clean
client io 35153 kB/s rd, 1912 kB/s wr, 672 op/s
Client is automatically unmounted in our cause.
Is it possible to change the PG_num in the production setup.
Journal stored on SATA 7.2k RPM 6GPS and 1gb network interface.
We are not configured Public and cluster as a separate network and it will be transferable via same LAN. Do we need to do this setup for better performance.
Also what is beter i/o operation setting for the crush map.
Still we are getting errors in ceph osd logs ,what need to done for this error.
2015-11-03 13:04:18.809488 7f387019c700 0 bad crc in data 3742210963 != exp 924878202
2015-11-03 13:04:18.812911 7f387019c700 0 -- 192.168.112.231:6800/49908 >> 192.168.112.192:0/1457324982 pipe(0x170d2000 sd=44 :6800 s=0 pgs=0 cs=0 l=0 c=0x1b18bf40).accept peer addr is really 192.168.112.192:0/1457324982 (socket is 192.168.112.192:47128/0)
Regards
Prabu
---- On Tue, 03 Nov 2015 12:50:40 +0530 Chris Taylor <ctaylor@xxxxxxxxxx> wrote ----
On 2015-11-02 10:19 pm, gjprabu wrote:> Hi Taylor,>> I have checked DNS name and all host resolve to the correct IP. MTU> size is 1500 in switch level configuration done. There is no firewall/> selinux is running currently.>> Also we would like to know below query's which already in the thread.>> Regards> Prabu>> ---- On Tue, 03 Nov 2015 11:20:07 +0530 CHRIS TAYLOR> <CTAYLOR@xxxxxxxxxx> wrote ---->> I would double check the network configuration on the new node.> Including hosts files and DNS names. Do all the host names resolve to> the correct IP addresses from all hosts?>> "... 192.168.112.231:6800/49908 >> 192.168.113.42:0/599324131 ...">> Looks like the communication between subnets is a problem. Is> xxx.xxx.113.xxx a typo? If that's correct, check MTU sizes. Are they> configured correctly on the switch and all NICs?>> Is there any iptables/firewall rules that could be blocking traffic> between hosts?>> Hope that helps,>> Chris>> On 2015-11-02 9:18 pm, gjprabu wrote:>> Hi,>> Anybody please help me on this issue.>> Regards> Prabu>> ---- On Mon, 02 Nov 2015 17:54:27 +0530 GJPRABU <GJPRABU@xxxxxxxxxxxx>> wrote ---->> Hi Team,>> We have ceph setup with 2 OSD and replica 2 and it is mounted with> ocfs2 clients and its working. When we added new osd all the clients> rbd mapped device disconnected and got hanged by running rbd ls or rbd> map command. We waited for long hours to scale the new osd size but> peering not completed event data sync finished, but client side issue> was persist and thought to try old osd service stop/start, after some> time rbd mapped automatically using existing map script.>> After service stop/start in old osd again 3rd OSD rebuild and back> filling started and after some time clients rbd mapped device> disconnected and got hanged by running rbd ls or rbd map command. We> thought to wait till to finished data sync in 3'rd OSD and its> completed, even though client side rbd not mapped. After we restarted> all mon and osd service and client side issue got fixed and mounted> rbd. We suspected some issue in our setup. also attached logs for your> reference.>What does 'ceph -s' look like? is the cluster HEALTH_OK?>> Something we are missing in our setup i don't know, highly appreciated> if anybody help us to solve this issue.>> Before new osd.2 addition :>> osd.0 - size : 13T and used 2.7 T> osd.1 - size : 13T and used 2.7 T>> After new osd addition :> osd.0 size : 13T and used 1.8T> osd.1 size : 13T and used 2.1T> osd.2 size : 15T and used 2.5T>> rbd ls> repo / integrepository (pg_num: 126)> rbd / integdownloads (pg_num: 64)>> Also we would like to know few clarifications .>> If any new osd will be added whether all client will be unmounted> automatically .>Clients do not need to unmount images when OSDs are added.> While add new osd can we access ( read / write ) from client machines ?>Clients still have read/write access to RBD images in the cluster whileadding OSDs and during recovery.> How much data will be added in new osd - without change any repilca /> pg_num ?>The data will re-balance between OSDs automatically. I found having morePGs help distribute the load more evenly.> How long to take finish this process ?Depends greatly on the hardware and configuration. Whether Journals onSSD or spinning disks, network connectivity, max_backfills, etc.>> If we missed any common configuration - please share the same .I don't see any configuration for public and cluster networks. If youare sharing the same network for clients and object replication/recoverythe cluster re-balancing data between OSDs could cause problems with theclient traffic.Take a look at:>> ceph.conf> [global]> fsid = 944fa0af-b7be-45a9-93ff-b9907cfaee3f> mon_initial_members = integ-hm5, integ-hm6, integ-hm7> mon_host = 192.168.112.192,192.168.112.193,192.168.112.194> auth_cluster_required = cephx> auth_service_required = cephx> auth_client_required = cephx> filestore_xattr_use_omap = true> osd_pool_default_size = 2>> [mon]> mon_clock_drift_allowed = .500>> [client]> rbd_cache = false>> Current Logs from new osd also attached old logs.>> 2015-11-02 12:47:48.481641 7f386f691700 0 bad crc in data 3889133030 !=> exp 2857248268> 2015-11-02 12:47:48.482230 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc510580).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42530/0)> 2015-11-02 12:47:48.483951 7f386f691700 0 bad crc in data 3192803598 !=> exp 1083014631> 2015-11-02 12:47:48.484512 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x170ea000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc516f60).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42531/0)> 2015-11-02 12:47:48.486284 7f386f691700 0 bad crc in data 133120597 !=> exp 393328400> 2015-11-02 12:47:48.486777 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x16a18000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc514620).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42532/0)> 2015-11-02 12:47:48.488624 7f386f691700 0 bad crc in data 3299720069 !=> exp 211350069> 2015-11-02 12:47:48.489100 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc513860).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42533/0)> 2015-11-02 12:47:48.490911 7f386f691700 0 bad crc in data 2381447347 !=> exp 1177846878> 2015-11-02 12:47:48.491390 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x170ea000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc513700).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42534/0)> 2015-11-02 12:47:48.493167 7f386f691700 0 bad crc in data 2093712440 !=> exp 2175112954> 2015-11-02 12:47:48.493682 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x16a18000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc514200).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42535/0)> 2015-11-02 12:47:48.495150 7f386f691700 0 bad crc in data 3047197039 !=> exp 38098198> 2015-11-02 12:47:48.495679 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0xc510b00).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42536/0)> 2015-11-02 12:47:48.497259 7f386f691700 0 bad crc in data 1400444622 !=> exp 2648291990> 2015-11-02 12:47:48.497756 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x170ea000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0x17f7b700).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:42537/0)> 2015-11-02 13:02:00.439025 7f386f691700 0 bad crc in data 4159064831 !=> exp 903679865> 2015-11-02 13:02:00.441337 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x16a18000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0x17f7e5c0).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:43128/0)> 2015-11-02 13:02:00.442756 7f386f691700 0 bad crc in data 1134831440 !=> exp 892008036> 2015-11-02 13:02:00.443369 7f386f691700 0 -- 192.168.112.231:6800/49908> >> 192.168.113.42:0/599324131 pipe(0x170d2000 sd=28 :6800 s=0 pgs=0 cs=0 l=0 c=0x17f7ee00).accept peer addr is really 192.168.113.42:0/599324131 (socket is 192.168.113.42:43129/0)> 2015-11-02 13:08:43.272527 7f387049f700 0 -- 192.168.112.231:6800/49908> >> 192.168.112.115:0/4256128918 pipe(0x170ea000 sd=33 :6800 s=0 pgs=0 cs=0 l=0 c=0x17f7e1a0).accept peer addr is really 192.168.112.115:0/4256128918 (socket is 192.168.112.115:51660/0)>> Regards> Prabu>> Regards> G.J>> _______________________________________________> ceph-users mailing list>> _______________________________________________> ceph-users mailing listLinks:------
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com