Top posting because there are multiple questions....
1. Atin, it is expected to fail if you don't have RDMA device or if it is not configured.
2. Rafi and Dennis,
I was not able to determine from logs if it really is a RDMA bug. The brick logs suggest that brick started and even accepted clients. We should look at the brick log more deeply to see if there is a config issue somewhere.
As to why RDMA listener did not initiate on fs2 and fs3, we need to get brick logs for them too.
Thanks,
Raghavendra Talur
On Fri, Sep 30, 2016 at 3:43 PM, Mohammed Rafi K C <rkavunga@xxxxxxxxxx> wrote:
It seems like an actual bug, if you can file a bug in bugzilla, that would be great.
At least I don't see workaround for this issue, may be till the next update is available with fix, you can use either rdma alone or tcp alone volume.
Let me know whether this is acceptable, if so I can give you the steps to change the transport of an existing volume.
Regards
Rafi KC
On 09/30/2016 10:35 AM, Mohammed Rafi K C wrote:
On 09/30/2016 02:35 AM, Dennis Michael wrote:
Are there any workarounds to this? RDMA is configured on my servers.
By this, I assume your rdma setup/configuration over IPoIB is working fine.
Can you tell us what machine you are using and whether SELinux is configured on the machine or not.
Also I couldn't see any logs attached here.
Rafi KC
Dennis
On Thu, Sep 29, 2016 at 7:19 AM, Atin Mukherjee <amukherj@xxxxxxxxxx> wrote:
Is this an intended behaviour if rdma device is not configured? Please chime in with your thoughtsRaghavendra Talur/Rafi,It seems like a volume configured created with tcp,rdma transport fails to start (atleast in my local set up). The issue here is although the brick process comes up, but glusterd receives a non zero ret code from the runner interface which spawns the brick process(es).Dennis,Thanks for sharing the logs.
--
On Wed, Sep 28, 2016 at 10:22 AM, Atin Mukherjee <amukherj@xxxxxxxxxx> wrote:
Dennis,It seems like that add-brick has definitely failed and the entry is not committed into glusterd store. volume status and volume info commands are referring the in-memory data for fs4 (which exist) but post a restart they are no longer available. Could you run glusterd with debug log enabled (systemctl stop glusterd; glusterd -LDEBUG) and provide us cmd_history.log, glusterd log along with fs4 brick log files to further analyze the issue? Regarding the missing RDMA ports for fs2, fs3 brick can you cross check if glusterfs-rdma package is installed on both the nodes?
--
On Wed, Sep 28, 2016 at 7:14 AM, Ravishankar N <ravishankar@xxxxxxxxxx> wrote:
I'm not sure what's going on here. Restarting glusterd seems to change the output of gluster volume info? I also see you are using RDMA. Not sure why the RDMA ports for fs2 and fs3 are not shown in the volume status output. CC'ing some glusterd/rdma devs for pointers.On 09/27/2016 10:29 PM, Dennis Michael wrote:
[root@fs4 bricks]# gluster volume infoVolume Name: cees-dataType: DistributeVolume ID: 27d2a59c-bdac-4f66-bcd8-e6124e53a4a2 Status: StartedNumber of Bricks: 4Transport-type: tcp,rdmaBricks:Brick1: fs1:/data/brickBrick2: fs2:/data/brickBrick3: fs3:/data/brickBrick4: fs4:/data/brickOptions Reconfigured:features.quota-deem-statfs: onfeatures.inode-quota: onfeatures.quota: onperformance.readdir-ahead: on[root@fs4 bricks]# gluster volume statusStatus of volume: cees-dataGluster process TCP Port RDMA Port Online Pid------------------------------------------------------------ ------------------ Brick fs1:/data/brick 49152 49153 Y 1878Brick fs2:/data/brick 49152 0 Y 1707Brick fs3:/data/brick 49152 0 Y 4696Brick fs4:/data/brick N/A N/A N N/ANFS Server on localhost 2049 0 Y 13808Quota Daemon on localhost N/A N/A Y 13813NFS Server on fs1 2049 0 Y 6722Quota Daemon on fs1 N/A N/A Y 6730NFS Server on fs3 2049 0 Y 12553Quota Daemon on fs3 N/A N/A Y 12561NFS Server on fs2 2049 0 Y 11702Quota Daemon on fs2 N/A N/A Y 11710Task Status of Volume cees-data------------------------------------------------------------ ------------------ There are no active volume tasks[root@fs4 bricks]# ps auxww | grep glusterroot 13791 0.0 0.0 701472 19768 ? Ssl 09:06 0:00 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFOroot 13808 0.0 0.0 560236 41420 ? Ssl 09:07 0:00 /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log -S /var/run/gluster/01c6152337436 9658a62b75c582b5ac2.socket root 13813 0.0 0.0 443164 17908 ? Ssl 09:07 0:00 /usr/sbin/glusterfs -s localhost --volfile-id gluster/quotad -p /var/lib/glusterd/quotad/run/quotad.pid -l /var/log/glusterfs/quotad.log -S /var/run/gluster/3753def90f5c3 4f656513dba6a544f7d.socket --xlator-option *replicate*.data-self-heal=off --xlator-option *replicate*.metadata-self-heal =off --xlator-option *replicate*.entry-self-heal=of f root 13874 0.0 0.0 1200472 31700 ? Ssl 09:16 0:00 /usr/sbin/glusterfsd -s fs4 --volfile-id cees-data.fs4.data-brick -p /var/lib/glusterd/vols/cees-data/run/fs4-data-brick.pid -S /var/run/gluster/5203ab38be21e 1d37c04f6bdfee77d4a.socket --brick-name /data/brick -l /var/log/glusterfs/bricks/data -brick.log --xlator-option *-posix.glusterd-uuid=f04b231e -63f8-4374-91ae-17c0c623f165 --brick-port 49152 49153 --xlator-option cees-data-server.transport.rdm a.listen-port=49153 --xlator-option cees-data-server.listen-port=4 9152 --volfile-server-transport=soc ket,rdma root 13941 0.0 0.0 112648 976 pts/0 S+ 09:50 0:00 grep --color=auto gluster
[root@fs4 bricks]# systemctl restart glusterfsd glusterd
[root@fs4 bricks]# ps auxww | grep glusterroot 13808 0.0 0.0 560236 41420 ? Ssl 09:07 0:00 /usr/sbin/glusterfs -s localhost --volfile-id gluster/nfs -p /var/lib/glusterd/nfs/run/nfs.pid -l /var/log/glusterfs/nfs.log -S /var/run/gluster/01c6152337436 9658a62b75c582b5ac2.socket root 13813 0.0 0.0 443164 17908 ? Ssl 09:07 0:00 /usr/sbin/glusterfs -s localhost --volfile-id gluster/quotad -p /var/lib/glusterd/quotad/run/quotad.pid -l /var/log/glusterfs/quotad.log -S /var/run/gluster/3753def90f5c3 4f656513dba6a544f7d.socket --xlator-option *replicate*.data-self-heal=off --xlator-option *replicate*.metadata-self-heal =off --xlator-option *replicate*.entry-self-heal=of f root 13953 0.1 0.0 570740 14988 ? Ssl 09:51 0:00 /usr/sbin/glusterd -p /var/run/glusterd.pid --log-level INFOroot 13965 0.0 0.0 112648 976 pts/0 S+ 09:51 0:00 grep --color=auto gluster
[root@fs4 bricks]# gluster volume infoVolume Name: cees-dataType: DistributeVolume ID: 27d2a59c-bdac-4f66-bcd8-e6124e53a4a2 Status: StartedNumber of Bricks: 3Transport-type: tcp,rdmaBricks:Brick1: fs1:/data/brickBrick2: fs2:/data/brickBrick3: fs3:/data/brickOptions Reconfigured:performance.readdir-ahead: onfeatures.quota: onfeatures.inode-quota: onfeatures.quota-deem-statfs: on
-Ravi
[root@fs4 bricks]# gluster volume statusStatus of volume: cees-dataGluster process TCP Port RDMA Port Online Pid------------------------------------------------------------ ------------------ Brick fs1:/data/brick 49152 49153 Y 1878Brick fs2:/data/brick 49152 0 Y 1707Brick fs3:/data/brick 49152 0 Y 4696NFS Server on localhost 2049 0 Y 13968Quota Daemon on localhost N/A N/A Y 13976NFS Server on fs2 2049 0 Y 11702Quota Daemon on fs2 N/A N/A Y 11710NFS Server on fs3 2049 0 Y 12553Quota Daemon on fs3 N/A N/A Y 12561NFS Server on fs1 2049 0 Y 6722Task Status of Volume cees-data------------------------------------------------------------ ------------------ There are no active volume tasks
[root@fs4 bricks]# gluster peer statusNumber of Peers: 3
Hostname: fs1Uuid: ddc0a23e-05e5-48f7-993e-a37e43b21605 State: Peer in Cluster (Connected)
Hostname: fs2Uuid: e37108f8-d2f1-4f28-adc8-0b3d3401df29 State: Peer in Cluster (Connected)
Hostname: fs3Uuid: 19a42201-c932-44db-b1a7-8b5b1af32a36 State: Peer in Cluster (Connected)
Dennis
On Tue, Sep 27, 2016 at 9:40 AM, Ravishankar N <ravishankar@xxxxxxxxxx> wrote:
Hmm, this shows the brick has started.On 09/27/2016 09:53 PM, Dennis Michael wrote:
Yes, you are right. I mixed up the logs. I just ran the add-brick command again after cleaning up fs4 and re-installing gluster. This is the complete fs4 data-brick.log.
[root@fs1 ~]# gluster volume add-brick cees-data fs4:/data/brickvolume add-brick: failed: Commit failed on fs4. Please check log file for details.
[root@fs4 bricks]# pwd/var/log/glusterfs/bricks[root@fs4 bricks]# cat data-brick.log[2016-09-27 16:16:28.095661] I [MSGID: 100030] [glusterfsd.c:2338:main] 0-/usr/sbin/glusterfsd: Started running /usr/sbin/glusterfsd version 3.7.14 (args: /usr/sbin/glusterfsd -s fs4 --volfile-id cees-data.fs4.data-brick -p /var/lib/glusterd/vols/cees-data/run/fs4-data-brick.pid -S /var/run/gluster/5203ab38be21e 1d37c04f6bdfee77d4a.socket --brick-name /data/brick -l /var/log/glusterfs/bricks/data -brick.log --xlator-option *-posix.glusterd-uuid=f04b231e -63f8-4374-91ae-17c0c623f165 --brick-port 49152 --xlator-option cees-data-server.transport.rdm a.listen-port=49153 --xlator-option cees-data-server.listen-port=4 9152 --volfile-server-transport=soc ket,rdma) [2016-09-27 16:16:28.101547] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 1 [2016-09-27 16:16:28.104637] I [graph.c:269:gf_add_cmdline_options] 0-cees-data-server: adding option 'listen-port' for volume 'cees-data-server' with value '49152' [2016-09-27 16:16:28.104646] I [graph.c:269:gf_add_cmdline_options] 0-cees-data-server: adding option 'transport.rdma.listen-port' for volume 'cees-data-server' with value '49153' [2016-09-27 16:16:28.104662] I [graph.c:269:gf_add_cmdline_options] 0-cees-data-posix: adding option 'glusterd-uuid' for volume 'cees-data-posix' with value 'f04b231e-63f8-4374-91ae-17c0c 623f165' [2016-09-27 16:16:28.104808] I [MSGID: 115034] [server.c:403:_check_for_auth_option] 0-/data/brick: skip format check for non-addr auth option auth.login./data/brick.allow [2016-09-27 16:16:28.104814] I [MSGID: 115034] [server.c:403:_check_for_auth_option] 0-/data/brick: skip format check for non-addr auth option auth.login.18ddaf4c-ad98-4155- 9372-717eae718b4c.password [2016-09-27 16:16:28.104883] I [MSGID: 101190] [event-epoll.c:632:event_dispatch_epoll_worker] 0-epoll: Started thread with index 2 [2016-09-27 16:16:28.105479] I [rpcsvc.c:2196:rpcsvc_set_outstanding_rpc_limit] 0-rpc-service: Configured rpc.outstanding-rpc-limit with value 64 [2016-09-27 16:16:28.105532] W [MSGID: 101002] [options.c:957:xl_opt_validate] 0-cees-data-server: option 'listen-port' is deprecated, preferred is 'transport.socket.listen-port' , continuing with correction [2016-09-27 16:16:28.109456] W [socket.c:3665:reconfigure] 0-cees-data-quota: NBIO on -1 failed (Bad file descriptor)[2016-09-27 16:16:28.489255] I [MSGID: 121050] [ctr-helper.c:259:extract_ctr_options] 0-gfdbdatastore: CTR Xlator is disabled. [2016-09-27 16:16:28.489272] W [MSGID: 101105] [gfdb_sqlite3.h:239:gfdb_set_sql_params] 0-cees-data-changetimerecorder : Failed to retrieve sql-db-pagesize from params.Assigning default value: 4096 [2016-09-27 16:16:28.489278] W [MSGID: 101105] [gfdb_sqlite3.h:239:gfdb_set_sql_params] 0-cees-data-changetimerecorder : Failed to retrieve sql-db-journalmode from params.Assigning default value: wal [2016-09-27 16:16:28.489284] W [MSGID: 101105] [gfdb_sqlite3.h:239:gfdb_set_sql_params] 0-cees-data-changetimerecorder : Failed to retrieve sql-db-sync from params.Assigning default value: off [2016-09-27 16:16:28.489288] W [MSGID: 101105] [gfdb_sqlite3.h:239:gfdb_set_sql_params] 0-cees-data-changetimerecorder : Failed to retrieve sql-db-autovacuum from params.Assigning default value: none [2016-09-27 16:16:28.490431] I [trash.c:2412:init] 0-cees-data-trash: no option specified for 'eliminate', using NULL[2016-09-27 16:16:28.672814] W [graph.c:357:_log_if_unknown_option] 0-cees-data-server: option 'rpc-auth.auth-glusterfs' is not recognized [2016-09-27 16:16:28.672854] W [graph.c:357:_log_if_unknown_option] 0-cees-data-server: option 'rpc-auth.auth-unix' is not recognized [2016-09-27 16:16:28.672872] W [graph.c:357:_log_if_unknown_option] 0-cees-data-server: option 'rpc-auth.auth-null' is not recognized [2016-09-27 16:16:28.672924] W [graph.c:357:_log_if_unknown_option] 0-cees-data-quota: option 'timeout' is not recognized [2016-09-27 16:16:28.672955] W [graph.c:357:_log_if_unknown_option] 0-cees-data-trash: option 'brick-path' is not recognized Final graph:+----------------------------------------------------------- -------------------+ 1: volume cees-data-posix2: type storage/posix3: option glusterd-uuid f04b231e-63f8-4374-91ae-17c0c623f165 4: option directory /data/brick5: option volume-id 27d2a59c-bdac-4f66-bcd8-e6124e53a4a2 6: option update-link-count-parent on7: end-volume8:9: volume cees-data-trash10: type features/trash11: option trash-dir .trashcan12: option brick-path /data/brick13: option trash-internal-op off14: subvolumes cees-data-posix15: end-volume16:17: volume cees-data-changetimerecorder18: type features/changetimerecorder19: option db-type sqlite320: option hot-brick off21: option db-name brick.db22: option db-path /data/brick/.glusterfs/23: option record-exit off24: option ctr_link_consistency off25: option ctr_lookupheal_link_timeout 30026: option ctr_lookupheal_inode_timeout 30027: option record-entry on28: option ctr-enabled off29: option record-counters off30: option ctr-record-metadata-heat off31: option sql-db-cachesize 100032: option sql-db-wal-autocheckpoint 100033: subvolumes cees-data-trash34: end-volume35:36: volume cees-data-changelog37: type features/changelog38: option changelog-brick /data/brick39: option changelog-dir /data/brick/.glusterfs/changelogs 40: option changelog-barrier-timeout 12041: subvolumes cees-data-changetimerecorder42: end-volume43:44: volume cees-data-bitrot-stub45: type features/bitrot-stub46: option export /data/brick47: subvolumes cees-data-changelog48: end-volume49:50: volume cees-data-access-control51: type features/access-control52: subvolumes cees-data-bitrot-stub53: end-volume54:55: volume cees-data-locks56: type features/locks57: subvolumes cees-data-access-control58: end-volume59:60: volume cees-data-upcall61: type features/upcall62: option cache-invalidation off63: subvolumes cees-data-locks64: end-volume65:66: volume cees-data-io-threads67: type performance/io-threads68: subvolumes cees-data-upcall69: end-volume70:71: volume cees-data-marker72: type features/marker73: option volume-uuid 27d2a59c-bdac-4f66-bcd8-e6124e53a4a2 74: option timestamp-file /var/lib/glusterd/vols/cees-data/marker.tstamp 75: option quota-version 176: option xtime off77: option gsync-force-xtime off78: option quota on79: option inode-quota on80: subvolumes cees-data-io-threads81: end-volume82:83: volume cees-data-barrier84: type features/barrier85: option barrier disable86: option barrier-timeout 12087: subvolumes cees-data-marker88: end-volume89:90: volume cees-data-index91: type features/index92: option index-base /data/brick/.glusterfs/indices93: subvolumes cees-data-barrier94: end-volume95:96: volume cees-data-quota97: type features/quota98: option transport.socket.connect-path /var/run/gluster/quotad.socket99: option transport-type socket100: option transport.address-family unix101: option volume-uuid cees-data102: option server-quota on103: option timeout 0104: option deem-statfs on105: subvolumes cees-data-index106: end-volume107:108: volume cees-data-worm109: type features/worm110: option worm off111: subvolumes cees-data-quota112: end-volume113:114: volume cees-data-read-only115: type features/read-only116: option read-only off117: subvolumes cees-data-worm118: end-volume119:120: volume /data/brick121: type debug/io-stats122: option log-level INFO123: option latency-measurement off124: option count-fop-hits off125: subvolumes cees-data-read-only126: end-volume127:128: volume cees-data-server129: type protocol/server130: option transport.socket.listen-port 49152131: option rpc-auth.auth-glusterfs on132: option rpc-auth.auth-unix on133: option rpc-auth.auth-null on134: option rpc-auth-allow-insecure on135: option transport.rdma.listen-port 49153136: option transport-type tcp,rdma137: option auth.login./data/brick.allow 18ddaf4c-ad98-4155-9372-717eae718b4c 138: option auth.login.18ddaf4c-ad98-4155-9372-717eae718b4c.password 9e913e92-7de0-47f9-94ed-d08cbb 130d23 139: option auth.addr./data/brick.allow *140: subvolumes /data/brick141: end-volume142:+----------------------------------------------------------- -------------------+ [2016-09-27 16:16:30.079541] I [login.c:81:gf_auth] 0-auth/login: allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c [2016-09-27 16:16:30.079567] I [MSGID: 115029] [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted client from fs3-12560-2016/09/27-16:16:30: 47674-cees-data-client-3-0-0 (version: 3.7.14) [2016-09-27 16:16:30.081487] I [login.c:81:gf_auth] 0-auth/login: allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c [2016-09-27 16:16:30.081505] I [MSGID: 115029] [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted client from fs2-11709-2016/09/27-16:16:30: 50047-cees-data-client-3-0-0 (version: 3.7.14) [2016-09-27 16:16:30.111091] I [login.c:81:gf_auth] 0-auth/login: allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c [2016-09-27 16:16:30.111113] I [MSGID: 115029] [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted client from fs2-11701-2016/09/27-16:16:29: 24060-cees-data-client-3-0-0 (version: 3.7.14) [2016-09-27 16:16:30.112822] I [login.c:81:gf_auth] 0-auth/login: allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c [2016-09-27 16:16:30.112836] I [MSGID: 115029] [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted client from fs3-12552-2016/09/27-16:16:29: 23041-cees-data-client-3-0-0 (version: 3.7.14) [2016-09-27 16:16:31.950978] I [login.c:81:gf_auth] 0-auth/login: allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c [2016-09-27 16:16:31.950998] I [MSGID: 115029] [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted client from fs1-6721-2016/09/27-16:16:26:9 39991-cees-data-client-3-0-0 (version: 3.7.14) [2016-09-27 16:16:31.981977] I [login.c:81:gf_auth] 0-auth/login: allowed user names: 18ddaf4c-ad98-4155-9372-717eae718b4c [2016-09-27 16:16:31.981994] I [MSGID: 115029] [server-handshake.c:690:server_setvolume] 0-cees-data-server: accepted client from fs1-6729-2016/09/27-16:16:27:9 71228-cees-data-client-3-0-0 (version: 3.7.14)
Does gluster volume info on fs4 shows all 4 bricks? (I guess it does based on your first email).
Does gluster volume status on fs4 (or ps aux|grep glusterfsd) show the brick as running?
Does gluster peer status on all nodes list the other 3 nodes as connected?
If yes, you could try `service glusterd restart` on fs4 and see if if brings up the brick? I'm just shooting in the dark here for possible clues.
-Ravi
On Tue, Sep 27, 2016 at 8:46 AM, Ravishankar N <ravishankar@xxxxxxxxxx> wrote:
On 09/27/2016 09:06 PM, Dennis Michael wrote:
Unless I'm missing something, the snippet below is from glusterd's log and not the brick's as is evident from the function names.Yes, the brick log /var/log/glusterfs/bricks/data-brick.log is created on fs4, and the snippets showing the errors were from that log.
-Ravi
Dennis
On Mon, Sep 26, 2016 at 5:58 PM, Ravishankar N <ravishankar@xxxxxxxxxx> wrote:
On 09/27/2016 05:25 AM, Dennis Michael wrote:
[2016-09-26 22:44:39.254921] E [MSGID: 106005] [glusterd-utils.c:4771:glusterd_brick_start] 0-management: Unable to start brick fs4:/data/brick
[2016-09-26 22:44:39.254949] E [MSGID: 106074] [glusterd-brick-ops.c:2372:glusterd_op_add_brick] 0-glusterd: Unable to add bricks
Is the brick log created on fs4? Does it contain warnings/errors?
-Ravi
--Atin
--Atin
_______________________________________________ Gluster-users mailing list Gluster-users@xxxxxxxxxxx http://www.gluster.org/mailman/listinfo/gluster-users