Hi,
it occurs maybe the previous email was too many words and not enough
data. so will try to display the issue differently.
gluster created (single brick volume following advice from
https://lists.gluster.org/pipermail/gluster-users/2016-October/028821.html):
root@sand1lian:~# gluster volume create moogle-gluster
sand1lian.computerisms.ca:/var/GlusterBrick/moogle-gluster
Gluster was started from cli with --debug, console reports the following
with creation of the volume:
[2018-10-31 17:00:51.555918] D [MSGID: 0]
[glusterd-volume-ops.c:328:__glusterd_handle_create_volume]
0-management: Received create volume req
[2018-10-31 17:00:51.555963] D [MSGID: 0]
[glusterd-utils.c:1774:glusterd_volinfo_find] 0-management: Returning -1
[2018-10-31 17:00:51.556072] D [MSGID: 0]
[glusterd-op-sm.c:209:glusterd_generate_txn_id] 0-management:
Transaction_id = 3f5d14c9-ee08-493c-afac-d04d53c12aad
[2018-10-31 17:00:51.556090] D [MSGID: 0]
[glusterd-op-sm.c:302:glusterd_set_txn_opinfo] 0-management:
Successfully set opinfo for transaction ID :
3f5d14c9-ee08-493c-afac-d04d53c12aad
[2018-10-31 17:00:51.556099] D [MSGID: 0]
[glusterd-op-sm.c:309:glusterd_set_txn_opinfo] 0-management: Returning 0
[2018-10-31 17:00:51.556108] D [MSGID: 0]
[glusterd-syncop.c:1809:gd_sync_task_begin] 0-management: Transaction ID
: 3f5d14c9-ee08-493c-afac-d04d53c12aad
[2018-10-31 17:00:51.556127] D [MSGID: 0]
[glusterd-locks.c:573:glusterd_mgmt_v3_lock] 0-management: Trying to
acquire lock of vol moogle-gluster for
bb8c61eb-f321-4485-8a8d-ddc369ac2203 as moogle-gluster_vol
[2018-10-31 17:00:51.556293] D [MSGID: 0]
[glusterd-locks.c:657:glusterd_mgmt_v3_lock] 0-management: Lock for vol
moogle-gluster successfully held by bb8c61eb-f321-4485-8a8d-ddc369ac2203
[2018-10-31 17:00:51.556333] D [MSGID: 0]
[glusterd-utils.c:1774:glusterd_volinfo_find] 0-management: Returning -1
[2018-10-31 17:00:51.556368] D [logging.c:1998:_gf_msg_internal]
0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
About to flush least recently used log message to disk
[2018-10-31 17:00:51.556345] D [MSGID: 0]
[glusterd-utils.c:1774:glusterd_volinfo_find] 0-management: Returning -1
[2018-10-31 17:00:51.556368] D [MSGID: 0]
[glusterd-utils.c:1094:glusterd_brickinfo_new] 0-management: Returning 0
[2018-10-31 17:00:51.556608] D [MSGID: 0]
[glusterd-utils.c:1308:glusterd_brickinfo_new_from_brick] 0-management:
Returning 0
[2018-10-31 17:00:51.556656] D [MSGID: 0]
[glusterd-utils.c:678:glusterd_volinfo_new] 0-management: Returning 0
[2018-10-31 17:00:51.556669] D [MSGID: 0]
[store.c:473:gf_store_handle_destroy] 0-: Returning 0
[2018-10-31 17:00:51.556681] D [MSGID: 0]
[glusterd-utils.c:990:glusterd_volume_brickinfos_delete] 0-management:
Returning 0
[2018-10-31 17:00:51.556690] D [MSGID: 0]
[store.c:473:gf_store_handle_destroy] 0-: Returning 0
[2018-10-31 17:00:51.556699] D [logging.c:1998:_gf_msg_internal]
0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
About to flush least recently used log message to disk
The message "D [MSGID: 0] [store.c:473:gf_store_handle_destroy] 0-:
Returning 0" repeated 3 times between [2018-10-31 17:00:51.556690] and
[2018-10-31 17:00:51.556698]
[2018-10-31 17:00:51.556699] D [MSGID: 0]
[glusterd-utils.c:1042:glusterd_volinfo_delete] 0-management: Returning 0
[2018-10-31 17:00:51.556728] D [MSGID: 0]
[glusterd-utils.c:1094:glusterd_brickinfo_new] 0-management: Returning 0
[2018-10-31 17:00:51.556738] D [MSGID: 0]
[glusterd-utils.c:1308:glusterd_brickinfo_new_from_brick] 0-management:
Returning 0
[2018-10-31 17:00:51.556752] D [MSGID: 0]
[glusterd-utils.c:678:glusterd_volinfo_new] 0-management: Returning 0
[2018-10-31 17:00:51.556764] D [MSGID: 0]
[store.c:473:gf_store_handle_destroy] 0-: Returning 0
[2018-10-31 17:00:51.556772] D [MSGID: 0]
[glusterd-utils.c:990:glusterd_volume_brickinfos_delete] 0-management:
Returning 0
[2018-10-31 17:00:51.556781] D [MSGID: 0]
[store.c:473:gf_store_handle_destroy] 0-: Returning 0
[2018-10-31 17:00:51.556791] D [logging.c:1998:_gf_msg_internal]
0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
About to flush least recently used log message to disk
The message "D [MSGID: 0] [store.c:473:gf_store_handle_destroy] 0-:
Returning 0" repeated 3 times between [2018-10-31 17:00:51.556781] and
[2018-10-31 17:00:51.556790]
[2018-10-31 17:00:51.556791] D [MSGID: 0]
[glusterd-utils.c:1042:glusterd_volinfo_delete] 0-management: Returning 0
[2018-10-31 17:00:51.556818] D [MSGID: 0]
[glusterd-utils.c:1094:glusterd_brickinfo_new] 0-management: Returning 0
[2018-10-31 17:00:51.556955] D [MSGID: 0]
[glusterd-peer-utils.c:130:glusterd_peerinfo_find_by_hostname]
0-management: Unable to find friend: sand1lian.computerisms.ca
[2018-10-31 17:00:51.557033] D [MSGID: 0]
[common-utils.c:3590:gf_is_local_addr] 0-management: 192.168.25.52
[2018-10-31 17:00:51.557140] D [MSGID: 0]
[common-utils.c:3478:gf_interface_search] 0-management: 192.168.25.52 is
local address at interface eno1
[2018-10-31 17:00:51.557154] D [MSGID: 0]
[glusterd-peer-utils.c:165:glusterd_hostname_to_uuid] 0-management:
returning 0
[2018-10-31 17:00:51.557172] D [MSGID: 0]
[glusterd-utils.c:1308:glusterd_brickinfo_new_from_brick] 0-management:
Returning 0
[2018-10-31 17:00:51.557183] D [MSGID: 0]
[glusterd-utils.c:1142:glusterd_resolve_brick] 0-management: Returning 0
[2018-10-31 17:00:51.557198] D [MSGID: 0]
[glusterd-utils.c:7558:glusterd_new_brick_validate] 0-management:
returning 0
[2018-10-31 17:00:51.557207] D [MSGID: 0]
[glusterd-utils.c:1142:glusterd_resolve_brick] 0-management: Returning 0
[2018-10-31 17:00:51.557392] D [MSGID: 0]
[glusterd-peer-utils.c:130:glusterd_peerinfo_find_by_hostname]
0-management: Unable to find friend: sand1lian.computerisms.ca
[2018-10-31 17:00:51.557468] D [MSGID: 0]
[common-utils.c:3590:gf_is_local_addr] 0-management: 192.168.25.52
[2018-10-31 17:00:51.557542] D [MSGID: 0]
[common-utils.c:3478:gf_interface_search] 0-management: 192.168.25.52 is
local address at interface eno1
[2018-10-31 17:00:51.557554] D [MSGID: 0]
[glusterd-peer-utils.c:165:glusterd_hostname_to_uuid] 0-management:
returning 0
[2018-10-31 17:00:51.557573] D [MSGID: 0]
[store.c:473:gf_store_handle_destroy] 0-: Returning 0
[2018-10-31 17:00:51.557586] D [MSGID: 0]
[glusterd-volume-ops.c:1467:glusterd_op_stage_create_volume]
0-management: Returning 0
[2018-10-31 17:00:51.557595] D [MSGID: 0]
[glusterd-op-sm.c:6014:glusterd_op_stage_validate] 0-management: OP = 1.
Returning 0
[2018-10-31 17:00:51.557610] D [MSGID: 0]
[glusterd-op-sm.c:7659:glusterd_op_bricks_select] 0-management: Returning 0
[2018-10-31 17:00:51.557620] D [MSGID: 0]
[glusterd-syncop.c:1751:gd_brick_op_phase] 0-management: Sent op req to
0 bricks
[2018-10-31 17:00:51.557663] D [MSGID: 0]
[glusterd-utils.c:678:glusterd_volinfo_new] 0-management: Returning 0
[2018-10-31 17:00:51.557693] D [MSGID: 0]
[glusterd-utils.c:1094:glusterd_brickinfo_new] 0-management: Returning 0
[2018-10-31 17:00:51.557771] D [MSGID: 0]
[glusterd-peer-utils.c:130:glusterd_peerinfo_find_by_hostname]
0-management: Unable to find friend: sand1lian.computerisms.ca
[2018-10-31 17:00:51.557844] D [MSGID: 0]
[common-utils.c:3590:gf_is_local_addr] 0-management: 192.168.25.52
[2018-10-31 17:00:51.557917] D [MSGID: 0]
[common-utils.c:3478:gf_interface_search] 0-management: 192.168.25.52 is
local address at interface eno1
[2018-10-31 17:00:51.557931] D [MSGID: 0]
[glusterd-peer-utils.c:165:glusterd_hostname_to_uuid] 0-management:
returning 0
[2018-10-31 17:00:51.557947] D [MSGID: 0]
[glusterd-utils.c:1308:glusterd_brickinfo_new_from_brick] 0-management:
Returning 0
[2018-10-31 17:00:51.557957] D [MSGID: 0]
[glusterd-utils.c:1142:glusterd_resolve_brick] 0-management: Returning 0
[2018-10-31 17:00:51.558393] D [MSGID: 0]
[xlator.c:218:xlator_volopt_dynload] 0-xlator: Returning 0
[2018-10-31 17:00:51.558409] D [MSGID: 0]
[glusterd-volgen.c:3140:_get_xlator_opt_key_from_vme] 0-glusterd:
Returning 0
[2018-10-31 17:00:51.558495] W [MSGID: 101095]
[xlator.c:180:xlator_volopt_dynload] 0-xlator:
/usr/lib/x86_64-linux-gnu/glusterfs/5.0/xlator/nfs/server.so: cannot
open shared object file: No such file or directory
[2018-10-31 17:00:51.558509] D [MSGID: 0]
[xlator.c:218:xlator_volopt_dynload] 0-xlator: Returning -1
[2018-10-31 17:00:51.558566] D [MSGID: 0]
[glusterd-store.c:1107:glusterd_store_create_volume_dir] 0-management:
Returning with 0
[2018-10-31 17:00:51.558593] D [MSGID: 0]
[glusterd-store.c:1125:glusterd_store_create_volume_run_dir]
0-management: Returning with 0
[2018-10-31 17:00:51.899586] D [MSGID: 0]
[store.c:432:gf_store_handle_new] 0-: Returning 0
[2018-10-31 17:00:51.930562] D [logging.c:1998:_gf_msg_internal]
0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
About to flush least recently used log message to disk
[2018-10-31 17:00:51.930485] D [MSGID: 0]
[store.c:432:gf_store_handle_new] 0-: Returning 0
[2018-10-31 17:00:51.930561] D [MSGID: 0]
[store.c:386:gf_store_save_value] 0-management: returning: 0
[2018-10-31 17:00:51.932563] D [logging.c:1998:_gf_msg_internal]
0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
About to flush least recently used log message to disk
The message "D [MSGID: 0] [store.c:386:gf_store_save_value]
0-management: returning: 0" repeated 19 times between [2018-10-31
17:00:51.930561] and [2018-10-31 17:00:51.930794]
[2018-10-31 17:00:51.932562] D [MSGID: 0]
[store.c:432:gf_store_handle_new] 0-: Returning 0
[2018-10-31 17:00:51.932688] D [MSGID: 0]
[store.c:386:gf_store_save_value] 0-management: returning: 0
[2018-10-31 17:00:51.932709] D [MSGID: 0]
[glusterd-store.c:457:glusterd_store_snapd_write] 0-management: Returning 0
[2018-10-31 17:00:51.935196] D [MSGID: 0]
[glusterd-store.c:521:glusterd_store_perform_snapd_store] 0-management:
Returning 0
[2018-10-31 17:00:51.935226] D [MSGID: 0]
[glusterd-store.c:585:glusterd_store_snapd_info] 0-management: Returning
with 0
[2018-10-31 17:00:51.935251] D [MSGID: 0]
[glusterd-store.c:788:_storeopts] 0-management: Storing in volinfo:key=
transport.address-family, val=inet
[2018-10-31 17:00:51.935290] D [MSGID: 0]
[store.c:386:gf_store_save_value] 0-management: returning: 0
[2018-10-31 17:00:51.935314] D [MSGID: 0]
[glusterd-store.c:788:_storeopts] 0-management: Storing in volinfo:key=
nfs.disable, val=on
[2018-10-31 17:00:51.935344] D [MSGID: 0]
[store.c:386:gf_store_save_value] 0-management: returning: 0
[2018-10-31 17:00:51.935360] D [MSGID: 0]
[glusterd-store.c:1174:glusterd_store_volinfo_write] 0-management:
Returning 0
[2018-10-31 17:00:51.935382] D [MSGID: 0]
[store.c:386:gf_store_save_value] 0-management: returning: 0
[2018-10-31 17:00:51.936584] D [MSGID: 0]
[store.c:432:gf_store_handle_new] 0-: Returning 0
[2018-10-31 17:00:51.936685] D [MSGID: 0]
[store.c:386:gf_store_save_value] 0-management: returning: 0
[2018-10-31 17:00:51.936807] D [logging.c:1998:_gf_msg_internal]
0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
About to flush least recently used log message to disk
The message "D [MSGID: 0] [store.c:386:gf_store_save_value]
0-management: returning: 0" repeated 10 times between [2018-10-31
17:00:51.936685] and [2018-10-31 17:00:51.936806]
[2018-10-31 17:00:51.936807] D [MSGID: 0]
[glusterd-store.c:430:glusterd_store_brickinfo_write] 0-management:
Returning 0
[2018-10-31 17:00:51.936833] D [MSGID: 0]
[glusterd-store.c:481:glusterd_store_perform_brick_store] 0-management:
Returning 0
[2018-10-31 17:00:51.936841] D [MSGID: 0]
[glusterd-store.c:550:glusterd_store_brickinfo] 0-management: Returning
with 0
[2018-10-31 17:00:51.936848] D [MSGID: 0]
[glusterd-store.c:1394:glusterd_store_brickinfos] 0-management: Returning 0
[2018-10-31 17:00:51.936856] D [MSGID: 0]
[glusterd-store.c:1620:glusterd_store_perform_volume_store]
0-management: Returning 0
[2018-10-31 17:00:51.958353] D [MSGID: 0]
[store.c:386:gf_store_save_value] 0-management: returning: 0
[2018-10-31 17:00:51.958494] D [logging.c:1998:_gf_msg_internal]
0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
About to flush least recently used log message to disk
The message "D [MSGID: 0] [store.c:386:gf_store_save_value]
0-management: returning: 0" repeated 9 times between [2018-10-31
17:00:51.958353] and [2018-10-31 17:00:51.958493]
[2018-10-31 17:00:51.958493] D [MSGID: 0]
[glusterd-store.c:1558:glusterd_store_node_state_write] 0-management:
Returning 0
[2018-10-31 17:00:51.960449] D [MSGID: 0]
[glusterd-store.c:1592:glusterd_store_perform_node_state_store]
0-management: Returning 0
[2018-10-31 17:00:51.960683] D [MSGID: 0]
[glusterd-utils.c:2840:glusterd_volume_compute_cksum] 0-management:
Returning with 0
[2018-10-31 17:00:51.960699] D [MSGID: 0]
[glusterd-store.c:1832:glusterd_store_volinfo] 0-management: Returning 0
[2018-10-31 17:00:51.960797] D [MSGID: 0]
[glusterd-utils.c:181:_brick_for_each] 0-management: Found a brick -
sand1lian.computerisms.ca:/var/GlusterBrick/moogle-gluster
[2018-10-31 17:00:51.961200] D [MSGID: 0]
[glusterd-volgen.c:1309:server_check_marker_off] 0-glusterd: Returning 0
[2018-10-31 17:00:51.961529] D [MSGID: 0]
[glusterd-volgen.c:5816:generate_brick_volfiles] 0-management: Returning 0
[2018-10-31 17:00:51.961681] D [MSGID: 0]
[glusterd-peer-utils.c:130:glusterd_peerinfo_find_by_hostname]
0-management: Unable to find friend: sand1lian.computerisms.ca
[2018-10-31 17:00:51.961756] D [MSGID: 0]
[common-utils.c:3590:gf_is_local_addr] 0-management: 192.168.25.52
[2018-10-31 17:00:51.961832] D [MSGID: 0]
[common-utils.c:3478:gf_interface_search] 0-management: 192.168.25.52 is
local address at interface eno1
[2018-10-31 17:00:51.961846] D [MSGID: 0]
[glusterd-peer-utils.c:165:glusterd_hostname_to_uuid] 0-management:
returning 0
[2018-10-31 17:00:51.961855] D [MSGID: 0]
[glusterd-utils.c:1668:glusterd_volume_brickinfo_get] 0-management:
Found brick sand1lian.computerisms.ca:/var/GlusterBrick/moogle-gluster
in volume moogle-gluster
[2018-10-31 17:00:51.961864] D [MSGID: 0]
[glusterd-utils.c:1677:glusterd_volume_brickinfo_get] 0-management:
Returning 0
[2018-10-31 17:00:51.963126] D [MSGID: 0]
[glusterd-peer-utils.c:130:glusterd_peerinfo_find_by_hostname]
0-management: Unable to find friend: sand1lian.computerisms.ca
[2018-10-31 17:00:51.963203] D [MSGID: 0]
[common-utils.c:3590:gf_is_local_addr] 0-management: 192.168.25.52
[2018-10-31 17:00:51.963280] D [MSGID: 0]
[common-utils.c:3478:gf_interface_search] 0-management: 192.168.25.52 is
local address at interface eno1
[2018-10-31 17:00:51.963298] D [MSGID: 0]
[glusterd-peer-utils.c:165:glusterd_hostname_to_uuid] 0-management:
returning 0
[2018-10-31 17:00:51.963308] D [MSGID: 0]
[glusterd-utils.c:1668:glusterd_volume_brickinfo_get] 0-management:
Found brick sand1lian.computerisms.ca:/var/GlusterBrick/moogle-gluster
in volume moogle-gluster
[2018-10-31 17:00:51.963316] D [MSGID: 0]
[glusterd-utils.c:1677:glusterd_volume_brickinfo_get] 0-management:
Returning 0
[2018-10-31 17:00:51.964038] D [MSGID: 0]
[glusterd-peer-utils.c:130:glusterd_peerinfo_find_by_hostname]
0-management: Unable to find friend: sand1lian.computerisms.ca
[2018-10-31 17:00:51.964112] D [MSGID: 0]
[common-utils.c:3590:gf_is_local_addr] 0-management: 192.168.25.52
[2018-10-31 17:00:51.964186] D [MSGID: 0]
[common-utils.c:3478:gf_interface_search] 0-management: 192.168.25.52 is
local address at interface eno1
[2018-10-31 17:00:51.964200] D [MSGID: 0]
[glusterd-peer-utils.c:165:glusterd_hostname_to_uuid] 0-management:
returning 0
[2018-10-31 17:00:51.964211] D [MSGID: 0]
[glusterd-utils.c:1668:glusterd_volume_brickinfo_get] 0-management:
Found brick sand1lian.computerisms.ca:/var/GlusterBrick/moogle-gluster
in volume moogle-gluster
[2018-10-31 17:00:51.964226] D [MSGID: 0]
[glusterd-utils.c:1677:glusterd_volume_brickinfo_get] 0-management:
Returning 0
[2018-10-31 17:00:51.965159] D [MSGID: 0]
[glusterd-op-sm.c:6150:glusterd_op_commit_perform] 0-management: Returning 0
[2018-10-31 17:00:51.965177] D [MSGID: 0]
[glusterd-utils.c:9664:glusterd_aggr_brick_mount_dirs] 0-management: No
brick_count present
[2018-10-31 17:00:51.965193] D [MSGID: 0]
[glusterd-op-sm.c:5109:glusterd_op_modify_op_ctx] 0-management: op_ctx
modification not required
[2018-10-31 17:00:51.965219] D [MSGID: 0]
[glusterd-locks.c:785:glusterd_mgmt_v3_unlock] 0-management: Trying to
release lock of vol moogle-gluster for
bb8c61eb-f321-4485-8a8d-ddc369ac2203 as moogle-gluster_vol
[2018-10-31 17:00:51.966350] D [MSGID: 0]
[glusterd-locks.c:834:glusterd_mgmt_v3_unlock] 0-management: Lock for
vol moogle-gluster successfully released
[2018-10-31 17:00:51.966462] D [MSGID: 0]
[glusterd-utils.c:1767:glusterd_volinfo_find] 0-management: Volume
moogle-gluster found
[2018-10-31 17:00:51.966479] D [MSGID: 0]
[glusterd-utils.c:1774:glusterd_volinfo_find] 0-management: Returning 0
[2018-10-31 17:00:51.966509] D [MSGID: 0]
[glusterd-op-sm.c:248:glusterd_get_txn_opinfo] 0-management:
Successfully got opinfo for transaction ID :
3f5d14c9-ee08-493c-afac-d04d53c12aad
[2018-10-31 17:00:51.966532] D [MSGID: 0]
[glusterd-op-sm.c:252:glusterd_get_txn_opinfo] 0-management: Returning 0
[2018-10-31 17:00:51.966551] D [MSGID: 0]
[glusterd-op-sm.c:352:glusterd_clear_txn_opinfo] 0-management:
Successfully cleared opinfo for transaction ID :
3f5d14c9-ee08-493c-afac-d04d53c12aad
[2018-10-31 17:00:51.966668] D [logging.c:1998:_gf_msg_internal]
0-logging-infra: Buffer overflow of a buffer whose size limit is 5.
About to flush least recently used log message to disk
[2018-10-31 17:00:51.966561] D [MSGID: 0]
[glusterd-op-sm.c:356:glusterd_clear_txn_opinfo] 0-management: Returning 0
[2018-10-31 17:00:51.966667] D [MSGID: 0]
[glusterd-rpc-ops.c:199:glusterd_op_send_cli_response] 0-management:
Returning 0
[2018-10-31 17:00:51.968134] D [socket.c:2927:socket_event_handler]
0-transport: EPOLLERR - disconnecting (sock:7) (non-SSL)
[2018-10-31 17:00:51.968183] E [MSGID: 101191]
[event-epoll.c:671:event_dispatch_epoll_worker] 0-epoll: Failed to
dispatch handler
grep: /var/lib/glusterd/vols/moogle-gluster/bricks/*: No such file or
directory
[2018-10-31 17:00:51.975661] I [run.c:242:runner_log]
(-->/usr/lib/x86_64-linux-gnu/glusterfs/5.0/xlator/mgmt/glusterd.so(+0xe0dbe)
[0x7f3f248dbdbe]
-->/usr/lib/x86_64-linux-gnu/glusterfs/5.0/xlator/mgmt/glusterd.so(+0xe07fe)
[0x7f3f248db7fe]
-->/usr/lib/x86_64-linux-gnu/libglusterfs.so.0(runner_log+0x105)
[0x7f3f28ac35a5] ) 0-management: Ran script:
/var/lib/glusterd/hooks/1/create/post/S10selinux-label-brick.sh
--volname=moogle-gluster
[2018-10-31 17:01:12.466614] D [logging.c:1871:gf_log_flush_timeout_cbk]
0-logging-infra: Log timer timed out. About to flush outstanding
messages if present
[2018-10-31 17:01:12.466667] D
[logging.c:1833:__gf_log_inject_timer_event] 0-logging-infra: Starting
timer now. Timeout = 120, current buf size = 5
[2018-10-31 17:03:12.492414] D [logging.c:1871:gf_log_flush_timeout_cbk]
0-logging-infra: Log timer timed out. About to flush outstanding
messages if present
[2018-10-31 17:03:12.492447] D
[logging.c:1833:__gf_log_inject_timer_event] 0-logging-infra: Starting
timer now. Timeout = 120, current buf size = 5
Not sure about the unable to find friend message:
root@sand1lian:~# dig +short sand1lian.computerisms.ca
192.168.25.52
start the volume:
root@sand1lian:~# gluster v start moogle-gluster
volume start: moogle-gluster: failed: Commit failed on localhost. Please
check log file for details.
output of cli.log while issuing start command:
[2018-10-31 17:08:49.019079] I [cli.c:764:main] 0-cli: Started running
gluster with version 5.0
[2018-10-31 17:08:49.021694] W [socket.c:3365:socket_connect]
0-glusterfs: Error disabling sockopt IPV6_V6ONLY: "Operation not supported"
[2018-10-31 17:08:49.021924] W [socket.c:3365:socket_connect]
0-glusterfs: Error disabling sockopt IPV6_V6ONLY: "Operation not supported"
[2018-10-31 17:08:49.101120] I [MSGID: 101190]
[event-epoll.c:622:event_dispatch_epoll_worker] 0-epoll: Started thread
with index 1
[2018-10-31 17:08:49.101231] E [MSGID: 101191]
[event-epoll.c:671:event_dispatch_epoll_worker] 0-epoll: Failed to
dispatch handler
[2018-10-31 17:08:49.113485] I
[cli-rpc-ops.c:1419:gf_cli_start_volume_cbk] 0-cli: Received resp to
start volume
[2018-10-31 17:08:49.113626] I [input.c:31:cli_batch] 0-: Exiting with: -1
and output of brick log while starting volume:
[2018-10-31 17:08:49.107966] I [MSGID: 100030] [glusterfsd.c:2691:main]
0-/usr/sbin/glusterfsd: Started running /usr/sbin/glusterfsd version 5.0
(args: /usr/sbin/glusterfsd -s sand1lian.computerisms.ca --volfile-id
moogle-gluster.sand1lian.computerisms.ca.var-GlusterBrick-moogle-gluster
-p
/var/run/gluster/vols/moogle-gluster/sand1lian.computerisms.ca-var-GlusterBrick-moogle-gluster.pid
-S /var/run/gluster/f41bfcfaf40deb7d.socket --brick-name
/var/GlusterBrick/moogle-gluster -l
/var/log/glusterfs/bricks/var-GlusterBrick-moogle-gluster.log
--xlator-option
*-posix.glusterd-uuid=bb8c61eb-f321-4485-8a8d-ddc369ac2203
--process-name brick --brick-port 49157 --xlator-option
moogle-gluster-server.listen-port=49157)
[2018-10-31 17:08:49.112123] E [socket.c:3466:socket_connect]
0-glusterfs: connection attempt on failed, (Invalid argument)
[2018-10-31 17:08:49.112293] I [MSGID: 101190]
[event-epoll.c:622:event_dispatch_epoll_worker] 0-epoll: Started thread
with index 1
[2018-10-31 17:08:49.112374] I [glusterfsd-mgmt.c:2424:mgmt_rpc_notify]
0-glusterfsd-mgmt: disconnected from remote-host: sand1lian.computerisms.ca
[2018-10-31 17:08:49.112399] I [glusterfsd-mgmt.c:2444:mgmt_rpc_notify]
0-glusterfsd-mgmt: Exhausted all volfile servers
[2018-10-31 17:08:49.112656] W [glusterfsd.c:1481:cleanup_and_exit]
(-->/usr/lib/x86_64-linux-gnu/libgfrpc.so.0(+0xf023) [0x7f3466c12023]
-->/usr/sbin/glusterfsd(+0x1273e) [0x557f4ea6373e]
-->/usr/sbin/glusterfsd(cleanup_and_exit+0x54) [0x557f4ea5be94] ) 0-:
received signum (1), shutting down
[2018-10-31 17:08:49.112973] E [socket.c:3466:socket_connect]
0-glusterfs: connection attempt on failed, (Invalid argument)
[2018-10-31 17:08:49.112996] W [rpc-clnt.c:1683:rpc_clnt_submit]
0-glusterfs: error returned while attempting to connect to host:(null),
port:0
[2018-10-31 17:08:49.113007] I
[socket.c:3710:socket_submit_outgoing_msg] 0-glusterfs: not connected
(priv->connected = 0)
[2018-10-31 17:08:49.113016] W [rpc-clnt.c:1695:rpc_clnt_submit]
0-glusterfs: failed to submit rpc-request (unique: 0, XID: 0x2 Program:
Gluster Portmap, ProgVers: 1, Proc: 5) to rpc-transport (glusterfs)
still seeing the empty pid file and the connection attempt on failed,
(Invalid argument) as the mostly likely culprits, but have read
everything of relevance I have found on google and not discovered a
solution yet...
On 2018-10-30 9:15 p.m., Computerisms Corporation wrote:
Hi,
Fortunately I am playing in a sandbox right now, but I am good and stuck
and hoping someone can point me in the right direction.
I have been playing for about 3 months with a gluster that currently has
one brick. The idea is that I have a server with data, I need to
migrate that server onto the new gluster-capable server, then I can use
the original server to make a 2nd brick, then I will be able to make
some room on a 3rd server for an arbiter brick. So I am building and
testing to be sure it all works before I try it in production.
Yesterday morning I was plugging away at figuring out how to make stuff
work on the new gluster server when I ran into an issue trying to rm -rf
a directory and it telling me it wasn't empty when ls -al showed that it
was. This has happened to me before, and what I did to fix it before
was unmount the Glusterfs, go into the brick, delete the files, and
remount the Glusterfs. I did that and it appeared to mount fine, but
when I tried to access the gluster mount, it gave me an error that there
were too many levels of symlinks.
I spent my day yesterday trying pretty much everything I could find on
google and a few things I couldn't. In the past when stuff has gone
funny with gluster on this box, I have always shut everything down and
checked if there was a new version of gluster, and indeed there was
version 5.0 available. So I did the upgrade quite early in the day.
Sadly it didn't fix my problem, but it did give me an error that led me
to modifying my hosts file to be ipv6 resolvable. Also after that, the
only time the gluster would mount was at reboot, but always with the
symlinks error, and it wasn't really mounted as reported by mount, but
the directory could be unmounted.
Having struck out completely yesterday, today I decided to start with a
new machine. I kept a history of the commands I had used to build the
gluster a few months back and pasted them all in. Found that the 5.0
package does not enable systemd, found that I needed the ipv6 entries in
the hosts file again, and also found the same problem in that the
glusterfs would not mount, the symlinks error at reboot, and the same
log entries.
I am still pretty new with gluster, so my best may not be that good, but
as best as I can tell the issue is that the brick will not start, even
with the force option. I think the problem boils down to one or both of
two lines in the logs. In the glusterd.log I have a line:
0-: Unable to read pidfile:
/var/run/gluster/vols/moogle-gluster/sand1lian.computerisms.ca-var-GlusterBrick-moogle-gluster.pid
The file exists, and I can't see anything wrong with permissions on the
file or the file tree leading to it, but it is a zero-bit file, so I am
thinking the problem is not the file, but that it can't read the
contents of the file because there aren't any.
The other log entry is in the brick log:
0-glusterfs: connection attempt on failed, (Invalid argument)
When I looked this up, it seems in my case there should be an attempt to
connect on 127.0.0.1, but given the double space I am thinking the host
argument is null, hence the invalid argument. It occurs that maybe I
still need some other entry in my hosts file to satisfy this, but I
can't think what it would be. I have created DNS entries; dig works,
and both hostname and FQDN resolve.
I have tried to change a lot of things today, so probably things are
buggered up beyond hope right now so even if I do find the solution
maybe it won't work. will wipe the new machine and start over again
tomorrow.
I realize the post is kinda long, sorry for that, but I want to make
sure I get every thing important. In fairness, though, I could easily
double the length of this post with possibly relevant things (if you are
interested). If you are still reading, thank you so much, I would
appreciate anything, even a wild guess, as to how to move forward on this?
_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users