Re: CentOS Storage SIG Repo for 3.8 is behind 3 versions (3.8.5). Is it still being used and can I help?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Feb 15, 2017 at 02:53:26PM +0100, Pavel Szalbot wrote:
> Hi, tested it with 3.8.8 on client (CentOS) and server (Ubuntu) and
> everything is OK now.

Awesome, many thanks for testing and reporting back the results.

Niels


> 
> -ps
> 
> On Wed, Feb 15, 2017 at 11:49 AM, Pavel Szalbot <pavel.szalbot@xxxxxxxxx>
> wrote:
> 
> > Hi Daryl,
> >
> > I must have missed your reply and found out about it when reading about
> > 3.8.9 and searching in gluster-users history.
> >
> > I will test the same setup with gluster 3.8.8 i.e. libvirt
> > 2.0.0-10.el7_3.4, glusterfs 3.8.8-1.el7 and gluster 3.8.8 on servers
> > (Ubuntu) and let you know.
> >
> > This is libvirt log for instance that used gluster storage backend
> > (libvirt 2.0.0, gluster client 3.8.5 and later 3.8.7, probably 3.8.5 on
> > servers, not sure):
> >
> > [2017-01-03 17:10:58.155566] I [MSGID: 104045] [glfs-master.c:91:notify]
> > 0-gfapi: New graph 6e6f6465-342d-6d69-6372-6f312e707267 (0) coming up
> > [2017-01-03 17:10:58.155615] I [MSGID: 114020] [client.c:2356:notify]
> > 0-gv_openstack_0-client-6: parent translators are ready, attempting connect
> > on transport
> > [2017-01-03 17:10:58.186043] I [MSGID: 114020] [client.c:2356:notify]
> > 0-gv_openstack_0-client-7: parent translators are ready, attempting connect
> > on transport
> > [2017-01-03 17:10:58.186518] I [rpc-clnt.c:1947:rpc_clnt_reconfig]
> > 0-gv_openstack_0-client-6: changing port to 49156 (from 0)
> > [2017-01-03 17:10:58.215411] I [rpc-clnt.c:1947:rpc_clnt_reconfig]
> > 0-gv_openstack_0-client-7: changing port to 49153 (from 0)
> > [2017-01-03 17:10:58.243706] I [MSGID: 114057] [client-handshake.c:1446:
> > select_server_supported_programs] 0-gv_openstack_0-client-6: Using
> > Program GlusterFS 3.3, Num (1298437), Version (330)
> > [2017-01-03 17:10:58.244215] I [MSGID: 114046] [client-handshake.c:1222:client_setvolume_cbk]
> > 0-gv_openstack_0-client-6: Connected to gv_openstack_0-client-6, attached
> > to remote volume '/export/gfs_0/gv_openstack_0_brick'.
> > [2017-01-03 17:10:58.244235] I [MSGID: 114047] [client-handshake.c:1233:client_setvolume_cbk]
> > 0-gv_openstack_0-client-6: Server and Client lk-version numbers are not
> > same, reopening the fds
> > [2017-01-03 17:10:58.244318] I [MSGID: 108005]
> > [afr-common.c:4301:afr_notify] 0-gv_openstack_0-replicate-0: Subvolume
> > 'gv_openstack_0-client-6' came back up; going online.
> > [2017-01-03 17:10:58.244437] I [MSGID: 114035] [client-handshake.c:201:client_set_lk_version_cbk]
> > 0-gv_openstack_0-client-6: Server lk version = 1
> > [2017-01-03 17:10:58.246940] I [MSGID: 114057] [client-handshake.c:1446:
> > select_server_supported_programs] 0-gv_openstack_0-client-7: Using
> > Program GlusterFS 3.3, Num (1298437), Version (330)
> > [2017-01-03 17:10:58.247252] I [MSGID: 114046] [client-handshake.c:1222:client_setvolume_cbk]
> > 0-gv_openstack_0-client-7: Connected to gv_openstack_0-client-7, attached
> > to remote volume '/export/gfs_0/gv_openstack_0_brick'.
> > [2017-01-03 17:10:58.247273] I [MSGID: 114047] [client-handshake.c:1233:client_setvolume_cbk]
> > 0-gv_openstack_0-client-7: Server and Client lk-version numbers are not
> > same, reopening the fds
> > [2017-01-03 17:10:58.257855] I [MSGID: 114035] [client-handshake.c:201:client_set_lk_version_cbk]
> > 0-gv_openstack_0-client-7: Server lk version = 1
> > [2017-01-03 17:10:58.259641] I [MSGID: 104041] [glfs-resolve.c:885:__glfs_active_subvol]
> > 0-gv_openstack_0: switched to graph 6e6f6465-342d-6d69-6372-6f312e707267
> > (0)
> > [2017-01-03 17:10:58.439897] I [MSGID: 104045] [glfs-master.c:91:notify]
> > 0-gfapi: New graph 6e6f6465-342d-6d69-6372-6f312e707267 (0) coming up
> > [2017-01-03 17:10:58.439929] I [MSGID: 114020] [client.c:2356:notify]
> > 0-gv_openstack_0-client-6: parent translators are ready, attempting connect
> > on transport
> > [2017-01-03 17:10:58.519082] I [MSGID: 114020] [client.c:2356:notify]
> > 0-gv_openstack_0-client-7: parent translators are ready, attempting connect
> > on transport
> > [2017-01-03 17:10:58.519527] I [rpc-clnt.c:1947:rpc_clnt_reconfig]
> > 0-gv_openstack_0-client-6: changing port to 49156 (from 0)
> > [2017-01-03 17:10:58.550482] I [MSGID: 114057] [client-handshake.c:1446:
> > select_server_supported_programs] 0-gv_openstack_0-client-6: Using
> > Program GlusterFS 3.3, Num (1298437), Version (330)
> > [2017-01-03 17:10:58.550997] I [MSGID: 114046] [client-handshake.c:1222:client_setvolume_cbk]
> > 0-gv_openstack_0-client-6: Connected to gv_openstack_0-client-6, attached
> > to remote volume '/export/gfs_0/gv_openstack_0_brick'.
> > [2017-01-03 17:10:58.551021] I [MSGID: 114047] [client-handshake.c:1233:client_setvolume_cbk]
> > 0-gv_openstack_0-client-6: Server and Client lk-version numbers are not
> > same, reopening the fds
> > [2017-01-03 17:10:58.551089] I [MSGID: 108005]
> > [afr-common.c:4301:afr_notify] 0-gv_openstack_0-replicate-0: Subvolume
> > 'gv_openstack_0-client-6' came back up; going online.
> > [2017-01-03 17:10:58.551199] I [MSGID: 114035] [client-handshake.c:201:client_set_lk_version_cbk]
> > 0-gv_openstack_0-client-6: Server lk version = 1
> > [2017-01-03 17:10:58.554413] I [rpc-clnt.c:1947:rpc_clnt_reconfig]
> > 0-gv_openstack_0-client-7: changing port to 49153 (from 0)
> > [2017-01-03 17:10:58.600956] I [MSGID: 114057] [client-handshake.c:1446:
> > select_server_supported_programs] 0-gv_openstack_0-client-7: Using
> > Program GlusterFS 3.3, Num (1298437), Version (330)
> > [2017-01-03 17:10:58.601276] I [MSGID: 114046] [client-handshake.c:1222:client_setvolume_cbk]
> > 0-gv_openstack_0-client-7: Connected to gv_openstack_0-client-7, attached
> > to remote volume '/export/gfs_0/gv_openstack_0_brick'.
> > [2017-01-03 17:10:58.601293] I [MSGID: 114047] [client-handshake.c:1233:client_setvolume_cbk]
> > 0-gv_openstack_0-client-7: Server and Client lk-version numbers are not
> > same, reopening the fds
> > [2017-01-03 17:10:58.616249] I [MSGID: 114035] [client-handshake.c:201:client_set_lk_version_cbk]
> > 0-gv_openstack_0-client-7: Server lk version = 1
> > [2017-01-03 17:10:58.617781] I [MSGID: 104041] [glfs-resolve.c:885:__glfs_active_subvol]
> > 0-gv_openstack_0: switched to graph 6e6f6465-342d-6d69-6372-6f312e707267
> > (0)
> > warning: host doesn't support requested feature: CPUID.01H:EDX.ds [bit 21]
> > warning: host doesn't support requested feature: CPUID.01H:EDX.acpi [bit
> > 22]
> > warning: host doesn't support requested feature: CPUID.01H:EDX.ht [bit 28]
> > warning: host doesn't support requested feature: CPUID.01H:EDX.tm [bit 29]
> > warning: host doesn't support requested feature: CPUID.01H:EDX.pbe [bit 31]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.dtes64 [bit
> > 2]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.monitor
> > [bit 3]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.ds_cpl [bit
> > 4]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.vmx [bit 5]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.smx [bit 6]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.est [bit 7]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.tm2 [bit 8]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.xtpr [bit
> > 14]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.pdcm [bit
> > 15]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.dca [bit 18]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.osxsave
> > [bit 27]
> > warning: host doesn't support requested feature: CPUID.01H:EDX.ds [bit 21]
> > warning: host doesn't support requested feature: CPUID.01H:EDX.acpi [bit
> > 22]
> > warning: host doesn't support requested feature: CPUID.01H:EDX.ht [bit 28]
> > warning: host doesn't support requested feature: CPUID.01H:EDX.tm [bit 29]
> > warning: host doesn't support requested feature: CPUID.01H:EDX.pbe [bit 31]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.dtes64 [bit
> > 2]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.monitor
> > [bit 3]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.ds_cpl [bit
> > 4]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.vmx [bit 5]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.smx [bit 6]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.est [bit 7]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.tm2 [bit 8]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.xtpr [bit
> > 14]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.pdcm [bit
> > 15]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.dca [bit 18]
> > warning: host doesn't support requested feature: CPUID.01H:ECX.osxsave
> > [bit 27]
> > 2017-01-04T12:21:14.293630Z qemu-kvm: terminating on signal 15 from pid 1
> > [2017-01-04 12:21:14.396691] I [MSGID: 114021] [client.c:2365:notify]
> > 0-gv_openstack_0-client-6: current graph is no longer active, destroying
> > rpc_client
> > [2017-01-04 12:21:14.396895] I [MSGID: 114021] [client.c:2365:notify]
> > 0-gv_openstack_0-client-7: current graph is no longer active, destroying
> > rpc_client
> > [2017-01-04 12:21:14.396910] I [MSGID: 114018] [client.c:2280:client_rpc_notify]
> > 0-gv_openstack_0-client-6: disconnected from gv_openstack_0-client-6.
> > Client process will keep trying to connect to glusterd until brick's port
> > is available
> > [2017-01-04 12:21:14.396927] I [MSGID: 114018] [client.c:2280:client_rpc_notify]
> > 0-gv_openstack_0-client-7: disconnected from gv_openstack_0-client-7.
> > Client process will keep trying to connect to glusterd until brick's port
> > is available
> > [2017-01-04 12:21:14.396942] E [MSGID: 108006]
> > [afr-common.c:4323:afr_notify] 0-gv_openstack_0-replicate-0: All subvolumes
> > are down. Going offline until atleast one of them comes back up.
> > [2017-01-04 12:21:14.397274] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gfapi: size=84 max=1 total=1
> > [2017-01-04 12:21:14.397565] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gfapi: size=188 max=2 total=2
> > [2017-01-04 12:21:14.397816] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gfapi: size=140 max=2 total=79
> > [2017-01-04 12:21:14.397993] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-client-6: size=1324 max=64 total=70716
> > [2017-01-04 12:21:14.398002] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-client-7: size=1324 max=64 total=49991
> > [2017-01-04 12:21:14.398010] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-replicate-0: size=10580 max=464 total=48110
> > [2017-01-04 12:21:14.398277] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-dht: size=1148 max=0 total=0
> > [2017-01-04 12:21:14.398376] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-dht: size=3380 max=233 total=37020
> > [2017-01-04 12:21:14.398583] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-readdir-ahead: size=60 max=0 total=0
> > [2017-01-04 12:21:14.398591] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-io-cache: size=68 max=0 total=0
> > [2017-01-04 12:21:14.398636] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-io-cache: size=252 max=64 total=11093
> > [2017-01-04 12:21:14.398664] I [io-stats.c:3747:fini] 0-gv_openstack_0:
> > io-stats translator unloaded
> > [2017-01-04 12:21:14.398854] I [MSGID: 101191] [event-epoll.c:659:event_dispatch_epoll_worker]
> > 0-epoll: Exited thread with index 2
> > [2017-01-04 12:21:14.398861] I [MSGID: 101191] [event-epoll.c:659:event_dispatch_epoll_worker]
> > 0-epoll: Exited thread with index 1
> > [2017-01-04 12:21:15.240813] I [MSGID: 114021] [client.c:2365:notify]
> > 0-gv_openstack_0-client-6: current graph is no longer active, destroying
> > rpc_client
> > [2017-01-04 12:21:15.241016] I [MSGID: 114021] [client.c:2365:notify]
> > 0-gv_openstack_0-client-7: current graph is no longer active, destroying
> > rpc_client
> > [2017-01-04 12:21:15.241061] I [MSGID: 114018] [client.c:2280:client_rpc_notify]
> > 0-gv_openstack_0-client-6: disconnected from gv_openstack_0-client-6.
> > Client process will keep trying to connect to glusterd until brick's port
> > is available
> > [2017-01-04 12:21:15.241089] I [MSGID: 114018] [client.c:2280:client_rpc_notify]
> > 0-gv_openstack_0-client-7: disconnected from gv_openstack_0-client-7.
> > Client process will keep trying to connect to glusterd until brick's port
> > is available
> > [2017-01-04 12:21:15.241108] E [MSGID: 108006]
> > [afr-common.c:4323:afr_notify] 0-gv_openstack_0-replicate-0: All subvolumes
> > are down. Going offline until atleast one of them comes back up.
> > [2017-01-04 12:21:15.241511] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gfapi: size=84 max=1 total=1
> > [2017-01-04 12:21:15.241906] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gfapi: size=188 max=2 total=2
> > [2017-01-04 12:21:15.242243] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gfapi: size=140 max=2 total=155
> > [2017-01-04 12:21:15.242264] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-client-6: size=1324 max=21 total=610
> > [2017-01-04 12:21:15.242282] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-client-7: size=1324 max=21 total=1646
> > [2017-01-04 12:21:15.242303] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-replicate-0: size=10580 max=40 total=1619
> > [2017-01-04 12:21:15.242838] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-dht: size=1148 max=0 total=0
> > [2017-01-04 12:21:15.243016] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-dht: size=3380 max=20 total=1482
> > [2017-01-04 12:21:15.243288] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-readdir-ahead: size=60 max=0 total=0
> > [2017-01-04 12:21:15.243303] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-io-cache: size=68 max=0 total=0
> > [2017-01-04 12:21:15.243461] I [MSGID: 101053] [mem-pool.c:641:mem_pool_destroy]
> > 0-gv_openstack_0-io-cache: size=252 max=3 total=140
> > [2017-01-04 12:21:15.243483] I [io-stats.c:3747:fini] 0-gv_openstack_0:
> > io-stats translator unloaded
> > [2017-01-04 12:21:15.243603] I [MSGID: 101191] [event-epoll.c:659:event_dispatch_epoll_worker]
> > 0-epoll: Exited thread with index 1
> > [2017-01-04 12:21:15.243631] I [MSGID: 101191] [event-epoll.c:659:event_dispatch_epoll_worker]
> > 0-epoll: Exited thread with index 2
> > 2017-01-04 12:21:16.363+0000: shutting down
> >
> >
> >
> >
> > -ps
> >
> > On Sun, Jan 15, 2017 at 8:16 PM, Niels de Vos <ndevos@xxxxxxxxxx> wrote:
> >
> >> On Fri, Jan 13, 2017 at 11:01:38AM +0100, Pavel Szalbot wrote:
> >> > Hi, you can install 3.8.7 from centos-gluster38-test using:
> >> >
> >> > yum --enablerepo=centos-gluster38-test install glusterfs
> >> >
> >> > I am not sure how QA works for CentOS Storage SIG, but 3.8.7 works same
> >> as
> >> > 3.8.5 for me - libvirt gfapi is unfortunately broken, no other problems
> >> > detected.
> >>
> >> Could you explain a little more of how this is broken? You would
> >> probably do good to report a bug as well:
> >>
> >>   https://bugzilla.redhat.com/enter_bug.cgi?product=GlusterFS&;
> >> version=3.8&component=gfapi
> >>
> >> If the bug contain steps that we can follow to reproduce the problem you
> >> are facing, it will be easier to investigate the cause and fix it.
> >>
> >> Thanks,
> >> Niels
> >>
> >>
> >> >
> >> > Btw 3.9 is short term maintenance release (
> >> > https://lists.centos.org/pipermail/centos-devel/2016-Septemb
> >> er/015197.html).
> >> >
> >> >
> >> > -ps
> >> >
> >> > On Fri, Jan 13, 2017 at 1:18 AM, Daryl lee <daryllee@xxxxxxxxxxxx>
> >> wrote:
> >> >
> >> > > Hey Gluster Community,
> >> > >
> >> > > According to the community packages list I get the impression that 3.8
> >> > > would be released to the CentOS Storage SIG Repo, but this seems to
> >> have
> >> > > stopped with 3.8.5 and 3.9 is still missing all together.   However,
> >> 3.7 is
> >> > > still being updated and is at 3.7.8 so I am confused why the other two
> >> > > versions have stopped.
> >> > >
> >> > >
> >> > >
> >> > > I did some looking on the past posts to this list and found a
> >> conversation
> >> > > about 3.9 on the CentOS repo last year but it looks like it's still
> >> not up
> >> > > yet; possibly due to a lack of community involvement in the testing
> >> and
> >> > > reporting back to whoever the maintainer is (which we don’t know
> >> yet)?   I
> >> > > might be in a position to help since I have a test environment that
> >> mirrors
> >> > > my production environment setup that I would use for testing the patch
> >> > > anyways, I might as well provide some good to the community..   At
> >> this
> >> > > point I know to do " yum install --enablerepo=centos-gluster38-test
> >> > > glusterfs-server" but I'm not sure who to tell if it works or not,
> >> and what
> >> > > kind of info they are looking for.    If someone wanted to give me a
> >> little
> >> > > guidance that would be awesome, especially if it will save me from
> >> having
> >> > > to switch to manually downloading packages.
> >> > >
> >> > >
> >> > >
> >> > > I guess the basic question is do we expect releases to resume for 3.8
> >> on
> >> > > the CentOS Storage SIG repo or should I be looking to move to manual
> >> > > patching for 3.8.  Additionally, if the person who does the releases
> >> to the
> >> > > CentOS Storage SIG is waiting for someone to tell them it looks
> >> fine,  who
> >> > > should I contact to do so?
> >> > >
> >> > >
> >> > >
> >> > >
> >> > >
> >> > >
> >> > >
> >> > > Thanks!
> >> > >
> >> > >
> >> > >
> >> > > Daryl
> >> > >
> >> > > _______________________________________________
> >> > > Gluster-users mailing list
> >> > > Gluster-users@xxxxxxxxxxx
> >> > > http://www.gluster.org/mailman/listinfo/gluster-users
> >> > >
> >>
> >> > _______________________________________________
> >> > Gluster-users mailing list
> >> > Gluster-users@xxxxxxxxxxx
> >> > http://www.gluster.org/mailman/listinfo/gluster-users
> >>
> >>
> >

Attachment: signature.asc
Description: PGP signature

_______________________________________________
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
http://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux