Re: How to configure?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Having less disks in a raid will help you recover your raid faster on disk failure.
If you use RAID0, then when you loose a disk, gluster will have to sync 36TB of data (with a lot of small files that will take some time).
Using a 6-disk RAID5 could be a little bit more convenient but you loose space.

About the arbiter topic - if you can have a low latency and redundant connection - go for full arbiter.

You can always set a different arbiter for each subvolume like this:
Server1:/brick1
Server2:/brick1
Server3:/arbiterbrick1
Server2:/brick2
Server3:/brick2
Server1:/arbiterbrick2
….

Best Regards,
Strahil Nikolov 



On Sunday, April 23, 2023, 10:15 AM, Diego Zuccato <diego.zuccato@xxxxxxxx> wrote:

After a lot of tests and unsuccessful searching, I decided to start from
scratch: I'm going to ditch the old volume and create a new one.

I have 3 servers with 30 12TB disks each. Since I'm going to start a new
volume, could it be better to group disks in 10 3-disk (or 6 5-disk)
RAID-0 volumes to reduce the number of bricks? Redundancy would be given
by replica 2 (still undecided about arbiter vs thin-arbiter...).

Current configuration is:
root@str957-clustor00:~# gluster v info cluster_data

Volume Name: cluster_data
Type: Distributed-Replicate
Volume ID: a8caaa90-d161-45bb-a68c-278263a8531a
Status: Started
Snapshot Count: 0
Number of Bricks: 45 x (2 + 1) = 135
Transport-type: tcp
Bricks:
Brick1: clustor00:/srv/bricks/00/d
Brick2: clustor01:/srv/bricks/00/d
Brick3: clustor02:/srv/bricks/00/q (arbiter)
...
Brick133: clustor01:/srv/bricks/29/d
Brick134: clustor02:/srv/bricks/29/d
Brick135: clustor00:/srv/bricks/14/q (arbiter)
Options Reconfigured:
cluster.background-self-heal-count: 256
cluster.heal-wait-queue-length: 10000
performance.quick-read: off
cluster.entry-self-heal: on
cluster.data-self-heal-algorithm: full
cluster.metadata-self-heal: on
cluster.shd-max-threads: 2
network.inode-lru-limit: 500000
performance.md-cache-timeout: 600
performance.cache-invalidation: on
features.cache-invalidation-timeout: 600
features.cache-invalidation: on
features.quota-deem-statfs: on
performance.readdir-ahead: on
cluster.granular-entry-heal: enable
features.scrub: Active
features.bitrot: on
cluster.lookup-optimize: on
performance.stat-prefetch: on
performance.cache-refresh-timeout: 60
performance.parallel-readdir: on
performance.write-behind-window-size: 128MB
cluster.self-heal-daemon: enable
features.inode-quota: on
features.quota: on
transport.address-family: inet
nfs.disable: on
performance.client-io-threads: off
client.event-threads: 1
features.scrub-throttle: normal
diagnostics.brick-log-level: ERROR
diagnostics.client-log-level: ERROR
config.brick-threads: 0
cluster.lookup-unhashed: on
config.client-threads: 1
cluster.use-anonymous-inode: off
diagnostics.brick-sys-log-level: CRITICAL
features.scrub-freq: monthly
cluster.data-self-heal: on
cluster.brick-multiplex: on
cluster.daemon-log-level: ERROR

Each node is a dual-Xeon 4210 (for a total of 20 cores, 40 threads)
equipped with 192GB RAM (that got exhausted quite often, before enabling
brick-multiplex).

Diego

Il 24/03/2023 19:21, Strahil Nikolov ha scritto:
> Try finding if any of them is missing on one of the systems.
>
> Best Regards,
> Strahil Nikolov
>
>    On Fri, Mar 24, 2023 at 15:59, Diego Zuccato
>    <diego.zuccato@xxxxxxxx> wrote:
>    There are 285 files in /var/lib/glusterd/vols/cluster_data ...
>    including
>    many files with names related to quorum bricks already moved to a
>    different path (like cluster_data.client.clustor02.srv-quorum-00-d.vol
>    that should already have been replaced by
>    cluster_data.clustor02.srv-bricks-00-q.vol -- and both vol files exist).
>
>    Is there something I should check inside the volfiles?
>
>    Diego
>
>    Il 24/03/2023 13:05, Strahil Nikolov ha scritto:
>      > Can you check your volume file contents?
>      > Maybe it really can't find (or access) a specific volfile ?
>      >
>      > Best Regards,
>      > Strahil Nikolov
>      >
>      >    On Fri, Mar 24, 2023 at 8:07, Diego Zuccato
>      >    <diego.zuccato@xxxxxxxx <mailto:diego.zuccato@xxxxxxxx>> wrote:
>      >    In glfsheal-Connection.log I see many lines like:
>      >    [2023-03-13 23:04:40.241481 +0000] E [MSGID: 104021]
>      >    [glfs-mgmt.c:586:glfs_mgmt_getspec_cbk] 0-gfapi: failed to get the
>      >    volume file [{from server}, {errno=2}, {error=File o directory non
>      >    esistente}]
>      >
>      >    And *lots* of gfid-mismatch errors in glustershd.log .
>      >
>      >    Couldn't find anything that would prevent heal to start. :(
>      >
>      >    Diego
>      >
>      >    Il 21/03/2023 20:39, Strahil Nikolov ha scritto:
>      >      > I have no clue. Have you checked for errors in the logs ?
>    Maybe you
>      >      > might find something useful.
>      >      >
>      >      > Best Regards,
>      >      > Strahil Nikolov
>      >      >
>      >      >    On Tue, Mar 21, 2023 at 9:56, Diego Zuccato
>      >      >    <diego.zuccato@xxxxxxxx <mailto:diego.zuccato@xxxxxxxx>
>    <mailto:diego.zuccato@xxxxxxxx>> wrote:
>      >      >    Killed glfsheal, after a day there were 218 processes, then
>      >    they got
>      >      >    killed by OOM during the weekend. Now there are no
>    processes
>      >    active.
>      >      >    Trying to run "heal info" reports lots of files quite
>    quickly
>      >    but does
>      >      >    not spawn any glfsheal process. And neither does restarting
>      >    glusterd.
>      >      >    Is there some way to selectively run glfsheal to fix
>    one brick
>      >    at a
>      >      >    time?
>      >      >
>      >      >    Diego
>      >      >
>      >      >    Il 21/03/2023 01:21, Strahil Nikolov ha scritto:
>      >      >      > Theoretically it might help.
>      >      >      > If possible, try to resolve any pending heals.
>      >      >      >
>      >      >      > Best Regards,
>      >      >      > Strahil Nikolov
>      >      >      >
>      >      >      >    On Thu, Mar 16, 2023 at 15:29, Diego Zuccato
>      >      >      >    <diego.zuccato@xxxxxxxx
>    <mailto:diego.zuccato@xxxxxxxx> <mailto:diego.zuccato@xxxxxxxx>
>      >    <mailto:diego.zuccato@xxxxxxxx>> wrote:
>      >      >      >    In Debian stopping glusterd does not stop brick
>      >    processes: to stop
>      >      >      >    everything (and free the memory) I have to
>      >      >      >    systemctl stop glusterd
>      >      >      >        killall glusterfs{,d}
>      >      >      >        killall glfsheal
>      >      >      >        systemctl start glusterd
>      >      >      >    [this behaviour hangs a simple reboot of a
>    machine running
>      >      >    glusterd...
>      >      >      >    not nice]
>      >      >      >
>      >      >      >    For now I just restarted glusterd w/o killing
>    the bricks:
>      >      >      >
>      >      >      >    root@str957-clustor00:~# ps aux|grep glfsheal|wc
>    -l ;
>      >      >    systemctl restart
>      >      >      >    glusterd ; ps aux|grep glfsheal|wc -l
>      >      >      >    618
>      >      >      >    618
>      >      >      >
>      >      >      >    No change neither in glfsheal processes nor in free
>      >    memory :(
>      >      >      >    Should I "killall glfsheal" before OOK kicks in?
>      >      >      >
>      >      >      >    Diego
>      >      >      >
>      >      >      >    Il 16/03/2023 12:37, Strahil Nikolov ha scritto:
>      >      >      >      > Can you restart glusterd service (first
>    check that
>      >    it was not
>      >      >      >    modified
>      >      >      >      > to kill the bricks)?
>      >      >      >      >
>      >      >      >      > Best Regards,
>      >      >      >      > Strahil Nikolov
>      >      >      >      >
>      >      >      >      >    On Thu, Mar 16, 2023 at 8:26, Diego Zuccato
>      >      >      >      >    <diego.zuccato@xxxxxxxx
>    <mailto:diego.zuccato@xxxxxxxx>
>      >    <mailto:diego.zuccato@xxxxxxxx> <mailto:diego.zuccato@xxxxxxxx>
>      >      >    <mailto:diego.zuccato@xxxxxxxx>> wrote:
>      >      >      >      >    OOM is just just a matter of time.
>      >      >      >      >
>      >      >      >      >    Today mem use is up to 177G/187 and:
>      >      >      >      >    # ps aux|grep glfsheal|wc -l
>      >      >      >      >    551
>      >      >      >      >
>      >      >      >      >    (well, one is actually the grep process, so
>      >    "only" 550
>      >      >    glfsheal
>      >      >      >      >    processes.
>      >      >      >      >
>      >      >      >      >    I'll take the last 5:
>      >      >      >      >    root    3266352  0.5  0.0 600292 93044 ?
>            Sl
>      >      >    06:55  0:07
>      >      >      >      >    /usr/libexec/glusterfs/glfsheal cluster_data
>      >      >    info-summary --xml
>      >      >      >      >    root    3267220  0.7  0.0 600292 91964 ?
>            Sl
>      >      >    07:00  0:07
>      >      >      >      >    /usr/libexec/glusterfs/glfsheal cluster_data
>      >      >    info-summary --xml
>      >      >      >      >    root    3268076  1.0  0.0 600160 88216 ?
>            Sl
>      >      >    07:05  0:08
>      >      >      >      >    /usr/libexec/glusterfs/glfsheal cluster_data
>      >      >    info-summary --xml
>      >      >      >      >    root    3269492  1.6  0.0 600292 91248 ?
>            Sl
>      >      >    07:10  0:07
>      >      >      >      >    /usr/libexec/glusterfs/glfsheal cluster_data
>      >      >    info-summary --xml
>      >      >      >      >    root    3270354  4.4  0.0 600292 93260 ?
>            Sl
>      >      >    07:15  0:07
>      >      >      >      >    /usr/libexec/glusterfs/glfsheal cluster_data
>      >      >    info-summary --xml
>      >      >      >      >
>      >      >      >      >    -8<--
>      >      >      >      >    root@str957-clustor00:~# ps -o ppid= 3266352
>      >      >      >      >    3266345
>      >      >      >      >    root@str957-clustor00:~# ps -o ppid= 3267220
>      >      >      >      >    3267213
>      >      >      >      >    root@str957-clustor00:~# ps -o ppid= 3268076
>      >      >      >      >    3268069
>      >      >      >      >    root@str957-clustor00:~# ps -o ppid= 3269492
>      >      >      >      >    3269485
>      >      >      >      >    root@str957-clustor00:~# ps -o ppid= 3270354
>      >      >      >      >    3270347
>      >      >      >      >    root@str957-clustor00:~# ps aux|grep 3266345
>      >      >      >      >    root    3266345  0.0  0.0 430536 10764 ?
>            Sl
>      >      >    06:55  0:00
>      >      >      >      >    gluster volume heal cluster_data info
>    summary --xml
>      >      >      >      >    root    3271532  0.0  0.0  6260  2500
>    pts/1    S+
>      >      >    07:21  0:00
>      >      >      >    grep
>      >      >      >      >    3266345
>      >      >      >      >    root@str957-clustor00:~# ps aux|grep 3267213
>      >      >      >      >    root    3267213  0.0  0.0 430536 10644 ?
>            Sl
>      >      >    07:00  0:00
>      >      >      >      >    gluster volume heal cluster_data info
>    summary --xml
>      >      >      >      >    root    3271599  0.0  0.0  6260  2480
>    pts/1    S+
>      >      >    07:22  0:00
>      >      >      >    grep
>      >      >      >      >    3267213
>      >      >      >      >    root@str957-clustor00:~# ps aux|grep 3268069
>      >      >      >      >    root    3268069  0.0  0.0 430536 10704 ?
>            Sl
>      >      >    07:05  0:00
>      >      >      >      >    gluster volume heal cluster_data info
>    summary --xml
>      >      >      >      >    root    3271626  0.0  0.0  6260  2516
>    pts/1    S+
>      >      >    07:22  0:00
>      >      >      >    grep
>      >      >      >      >    3268069
>      >      >      >      >    root@str957-clustor00:~# ps aux|grep 3269485
>      >      >      >      >    root    3269485  0.0  0.0 430536 10756 ?
>            Sl
>      >      >    07:10  0:00
>      >      >      >      >    gluster volume heal cluster_data info
>    summary --xml
>      >      >      >      >    root    3271647  0.0  0.0  6260  2480
>    pts/1    S+
>      >      >    07:22  0:00
>      >      >      >    grep
>      >      >      >      >    3269485
>      >      >      >      >    root@str957-clustor00:~# ps aux|grep 3270347
>      >      >      >      >    root    3270347  0.0  0.0 430536 10672 ?
>            Sl
>      >      >    07:15  0:00
>      >      >      >      >    gluster volume heal cluster_data info
>    summary --xml
>      >      >      >      >    root    3271666  0.0  0.0  6260  2568
>    pts/1    S+
>      >      >    07:22  0:00
>      >      >      >    grep
>      >      >      >      >    3270347
>      >      >      >      >    -8<--
>      >      >      >      >
>      >      >      >      >    Seems glfsheal is spawning more processes.
>      >      >      >      >    I can't rule out a metadata corruption (or at
>      >    least a
>      >      >    desync),
>      >      >      >    but it
>      >      >      >      >    shouldn't happen...
>      >      >      >      >
>      >      >      >      >    Diego
>      >      >      >      >
>      >      >      >      >    Il 15/03/2023 20:11, Strahil Nikolov ha
>    scritto:
>      >      >      >      >      > If you don't experience any OOM , you can
>      >    focus on
>      >      >    the heals.
>      >      >      >      >      >
>      >      >      >      >      > 284 processes of glfsheal seems odd.
>      >      >      >      >      >
>      >      >      >      >      > Can you check the ppid for 2-3
>    randomly picked ?
>      >      >      >      >      > ps -o ppid= <pid>
>      >      >      >      >      >
>      >      >      >      >      > Best Regards,
>      >      >      >      >      > Strahil Nikolov
>      >      >      >      >      >
>      >      >      >      >      >    On Wed, Mar 15, 2023 at 9:54,
>    Diego Zuccato
>      >      >      >      >      >    <diego.zuccato@xxxxxxxx
>    <mailto:diego.zuccato@xxxxxxxx>
>      >    <mailto:diego.zuccato@xxxxxxxx>
>      >      >    <mailto:diego.zuccato@xxxxxxxx>
>    <mailto:diego.zuccato@xxxxxxxx>
>      >      >      >    <mailto:diego.zuccato@xxxxxxxx>> wrote:
>      >      >      >      >      >    I enabled it yesterday and that
>    greatly
>      >    reduced
>      >      >    memory
>      >      >      >    pressure.
>      >      >      >      >      >    Current volume info:
>      >      >      >      >      >    -8<--
>      >      >      >      >      >    Volume Name: cluster_data
>      >      >      >      >      >    Type: Distributed-Replicate
>      >      >      >      >      >    Volume ID:
>      >    a8caaa90-d161-45bb-a68c-278263a8531a
>      >      >      >      >      >    Status: Started
>      >      >      >      >      >    Snapshot Count: 0
>      >      >      >      >      >    Number of Bricks: 45 x (2 + 1) = 135
>      >      >      >      >      >    Transport-type: tcp
>      >      >      >      >      >    Bricks:
>      >      >      >      >      >    Brick1: clustor00:/srv/bricks/00/d
>      >      >      >      >      >    Brick2: clustor01:/srv/bricks/00/d
>      >      >      >      >      >    Brick3: clustor02:/srv/bricks/00/q
>    (arbiter)
>      >      >      >      >      >    [...]
>      >      >      >      >      >    Brick133: clustor01:/srv/bricks/29/d
>      >      >      >      >      >    Brick134: clustor02:/srv/bricks/29/d
>      >      >      >      >      >    Brick135: clustor00:/srv/bricks/14/q
>      >    (arbiter)
>      >      >      >      >      >    Options Reconfigured:
>      >      >      >      >      >    performance.quick-read: off
>      >      >      >      >      >    cluster.entry-self-heal: on
>      >      >      >      >      >    cluster.data-self-heal-algorithm: full
>      >      >      >      >      >    cluster.metadata-self-heal: on
>      >      >      >      >      >    cluster.shd-max-threads: 2
>      >      >      >      >      >    network.inode-lru-limit: 500000
>      >      >      >      >      >    performance.md-cache-timeout: 600
>      >      >      >      >      >    performance.cache-invalidation: on
>      >      >      >      >      > 
>    features.cache-invalidation-timeout: 600
>      >      >      >      >      >    features.cache-invalidation: on
>      >      >      >      >      >    features.quota-deem-statfs: on
>      >      >      >      >      >    performance.readdir-ahead: on
>      >      >      >      >      >    cluster.granular-entry-heal: enable
>      >      >      >      >      >    features.scrub: Active
>      >      >      >      >      >    features.bitrot: on
>      >      >      >      >      >    cluster.lookup-optimize: on
>      >      >      >      >      >    performance.stat-prefetch: on
>      >      >      >      >      >    performance.cache-refresh-timeout: 60
>      >      >      >      >      >    performance.parallel-readdir: on
>      >      >      >      >      > 
>    performance.write-behind-window-size: 128MB
>      >      >      >      >      >    cluster.self-heal-daemon: enable
>      >      >      >      >      >    features.inode-quota: on
>      >      >      >      >      >    features.quota: on
>      >      >      >      >      >    transport.address-family: inet
>      >      >      >      >      >    nfs.disable: on
>      >      >      >      >      >    performance.client-io-threads: off
>      >      >      >      >      >    client.event-threads: 1
>      >      >      >      >      >    features.scrub-throttle: normal
>      >      >      >      >      >    diagnostics.brick-log-level: ERROR
>      >      >      >      >      >    diagnostics.client-log-level: ERROR
>      >      >      >      >      >    config.brick-threads: 0
>      >      >      >      >      >    cluster.lookup-unhashed: on
>      >      >      >      >      >    config.client-threads: 1
>      >      >      >      >      >    cluster.use-anonymous-inode: off
>      >      >      >      >      >    diagnostics.brick-sys-log-level:
>    CRITICAL
>      >      >      >      >      >    features.scrub-freq: monthly
>      >      >      >      >      >    cluster.data-self-heal: on
>      >      >      >      >      >    cluster.brick-multiplex: on
>      >      >      >      >      >    cluster.daemon-log-level: ERROR
>      >      >      >      >      >    -8<--
>      >      >      >      >      >
>      >      >      >      >      >    htop reports that memory usage is
>    up to 143G,
>      >      >    there are 602
>      >      >      >      >    tasks and
>      >      >      >      >      >    5232 threads (~20 running) on
>    clustor00,
>      >    117G/49
>      >      >    tasks/1565
>      >      >      >      >    threads on
>      >      >      >      >      >    clustor01 and 126G/45 tasks/1574
>    threads on
>      >      >    clustor02.
>      >      >      >      >      >    I see quite a lot (284!) of glfsheal
>      >    processes
>      >      >    running on
>      >      >      >      >    clustor00 (a
>      >      >      >      >      >    "gluster v heal cluster_data info
>    summary" is
>      >      >    running
>      >      >      >    on clustor02
>      >      >      >      >      >    since
>      >      >      >      >      >    yesterday, still no output).
>    Shouldn't be
>      >    just
>      >      >    one per
>      >      >      >    brick?
>      >      >      >      >      >
>      >      >      >      >      >    Diego
>      >      >      >      >      >
>      >      >      >      >      >    Il 15/03/2023 08:30, Strahil
>    Nikolov ha
>      >    scritto:
>      >      >      >      >      >      > Do you use brick multiplexing ?
>      >      >      >      >      >      >
>      >      >      >      >      >      > Best Regards,
>      >      >      >      >      >      > Strahil Nikolov
>      >      >      >      >      >      >
>      >      >      >      >      >      >    On Tue, Mar 14, 2023 at 16:44,
>      >    Diego Zuccato
>      >      >      >      >      >      >    <diego.zuccato@xxxxxxxx
>    <mailto:diego.zuccato@xxxxxxxx>
>      >    <mailto:diego.zuccato@xxxxxxxx>
>      >      >    <mailto:diego.zuccato@xxxxxxxx>
>      >      >      >    <mailto:diego.zuccato@xxxxxxxx>
>      >    <mailto:diego.zuccato@xxxxxxxx>
>      >      >      >      >    <mailto:diego.zuccato@xxxxxxxx>> wrote:
>      >      >      >      >      >      >    Hello all.
>      >      >      >      >      >      >
>      >      >      >      >      >      >    Our Gluster 9.6 cluster is
>    showing
>      >    increasing
>      >      >      >    problems.
>      >      >      >      >      >      >    Currently it's composed of 3
>      >    servers (2x
>      >      >    Intel Xeon
>      >      >      >      >    4210 [20
>      >      >      >      >      >    cores dual
>      >      >      >      >      >      >    thread, total 40 threads],
>    192GB
>      >    RAM, 30x
>      >      >    HGST
>      >      >      >      >    HUH721212AL5200
>      >      >      >      >      >    [12TB]),
>      >      >      >      >      >      >    configured in replica 3
>    arbiter 1.
>      >    Using
>      >      >    Debian
>      >      >      >      >    packages from
>      >      >      >      >      >    Gluster
>      >      >      >      >      >      >    9.x latest repository.
>      >      >      >      >      >      >
>      >      >      >      >      >      >    Seems 192G RAM are not
>    enough to
>      >    handle
>      >      >    30 data
>      >      >      >    bricks + 15
>      >      >      >      >      >    arbiters
>      >      >      >      >      >      >    and
>      >      >      >      >      >      >    I often had to reload
>    glusterfsd
>      >    because
>      >      >    glusterfs
>      >      >      >      >    processed
>      >      >      >      >      >    got killed
>      >      >      >      >      >      >    for OOM.
>      >      >      >      >      >      >    On top of that, performance
>    have been
>      >      >    quite bad,
>      >      >      >    especially
>      >      >      >      >      >    when we
>      >      >      >      >      >      >    reached about 20M files. On
>    top of
>      >    that,
>      >      >    one of
>      >      >      >    the servers
>      >      >      >      >      >    have had
>      >      >      >      >      >      >    mobo issues that resulted
>    in memory
>      >      >    errors that
>      >      >      >      >    corrupted some
>      >      >      >      >      >      >    bricks fs
>      >      >      >      >      >      >    (XFS, it required
>    "xfs_reparir -L"
>      >    to fix).
>      >      >      >      >      >      >    Now I'm getting lots of "stale
>      >    file handle"
>      >      >      >    errors and
>      >      >      >      >    other
>      >      >      >      >      >    errors
>      >      >      >      >      >      >    (like directories that seem
>    empty
>      >    from the
>      >      >      >    client but still
>      >      >      >      >      >    containing
>      >      >      >      >      >      >    files in some bricks) and auto
>      >    healing seems
>      >      >      >    unable to
>      >      >      >      >    complete.
>      >      >      >      >      >      >
>      >      >      >      >      >      >    Since I can't keep up
>    continuing to
>      >      >    manually fix
>      >      >      >    all the
>      >      >      >      >      >    issues, I'm
>      >      >      >      >      >      >    thinking about
>    backup+destroy+recreate
>      >      >    strategy.
>      >      >      >      >      >      >
>      >      >      >      >      >      >    I think that if I reduce
>    the number of
>      >      >    bricks per
>      >      >      >      >    server to just 5
>      >      >      >      >      >      >    (RAID1 of 6x12TB disks) I might
>      >    resolve RAM
>      >      >      >    issues - at the
>      >      >      >      >      >    cost of
>      >      >      >      >      >      >    longer heal times in case a
>    disk
>      >    fails. Am I
>      >      >      >    right or it's
>      >      >      >      >      >    useless?
>      >      >      >      >      >      >    Other recommendations?
>      >      >      >      >      >      >    Servers have space for
>    another 6
>      >    disks.
>      >      >    Maybe those
>      >      >      >      >    could be
>      >      >      >      >      >    used for
>      >      >      >      >      >      >    some SSDs to speed up access?
>      >      >      >      >      >      >
>      >      >      >      >      >      >    TIA.
>      >      >      >      >      >      >
>      >      >      >      >      >      >    --
>      >      >      >      >      >      >    Diego Zuccato
>      >      >      >      >      >      >    DIFA - Dip. di Fisica e
>    Astronomia
>      >      >      >      >      >      >    Servizi Informatici
>      >      >      >      >      >      >    Alma Mater Studiorum -
>    Università
>      >    di Bologna
>      >      >      >      >      >      >    V.le Berti-Pichat 6/2 - 40127
>      >    Bologna - Italy
>      >      >      >      >      >      >    tel.: +39 051 20 95786
>      >      >      >      >      >      >    ________
>      >      >      >      >      >      >
>      >      >      >      >      >      >
>      >      >      >      >      >      >
>      >      >      >      >      >      >    Community Meeting Calendar:
>      >      >      >      >      >      >
>      >      >      >      >      >      >    Schedule -
>      >      >      >      >      >      >    Every 2nd and 4th Tuesday
>    at 14:30
>      >    IST /
>      >      >    09:00 UTC
>      >      >      >      >      >      >    Bridge:
>      >      > https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>
>      >      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>>
>      >      >      >      >      > 
>    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>
>      >      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>>>
>      >      >      >      >      >      >
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>
>      >      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>>
>      >      >      >      >      > 
>    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>
>      >      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>>>>
>      >      >      >      >      >      >    Gluster-users mailing list
>      >      >      >      >      >      > Gluster-users@xxxxxxxxxxx
>    <mailto:Gluster-users@xxxxxxxxxxx>
>      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >      >      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >      >      >      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >      >      >      >      >
>      >      >      >
>    https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>
>      >      >      >      >
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>>
>      >      >      >      >      >
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>
>      >      >      >      >
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>>>
>      >      >      >      >      >      >
>      >      >      >      >
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>
>      >      >      >      >
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>>
>      >      >      >      >      >
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>
>      >      >      >      >
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>>>>
>      >      >
>      >      >      >
>      >      >      >      >
>      >      >      >      >      >
>      >      >      >      >      >      >
>      >      >      >      >      >
>      >      >      >      >      >    --
>      >      >      >      >      >    Diego Zuccato
>      >      >      >      >      >    DIFA - Dip. di Fisica e Astronomia
>      >      >      >      >      >    Servizi Informatici
>      >      >      >      >      >    Alma Mater Studiorum - Università
>    di Bologna
>      >      >      >      >      >    V.le Berti-Pichat 6/2 - 40127
>    Bologna - Italy
>      >      >      >      >      >    tel.: +39 051 20 95786
>      >      >      >      >      >    ________
>      >      >      >      >      >
>      >      >      >      >      >
>      >      >      >      >      >
>      >      >      >      >      >    Community Meeting Calendar:
>      >      >      >      >      >
>      >      >      >      >      >    Schedule -
>      >      >      >      >      >    Every 2nd and 4th Tuesday at 14:30
>    IST /
>      >    09:00 UTC
>      >      >      >      >      >    Bridge:
>      > https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>
>      >      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>>
>      >      >      >      >      > 
>    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>
>      >      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>
>      >      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>
>      >      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>
>      >    <https://meet.google.com/cpu-eiue-hvk
>    <https://meet.google.com/cpu-eiue-hvk>>>>>>
>      >      >      >      >      >    Gluster-users mailing list
>      >      >      >      >      > Gluster-users@xxxxxxxxxxx
>    <mailto:Gluster-users@xxxxxxxxxxx>
>      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >      >      >    <mailto:Gluster-users@xxxxxxxxxxx>
>      >      >      >      >      >
>      >      > https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>
>      >      >      >      >
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>>
>      >      >      >      >      >
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>
>      >      >      >      >
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>
>      >      >      >
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>
>      >      > 
>    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>
>      >    <https://lists.gluster.org/mailman/listinfo/gluster-users
>    <https://lists.gluster.org/mailman/listinfo/gluster-users>>>>>>
>      >      >      >      >      >
>      >      >      >      >
>      >      >      >      >    --
>      >      >      >      >    Diego Zuccato
>      >      >      >      >    DIFA - Dip. di Fisica e Astronomia
>      >      >      >      >    Servizi Informatici
>      >      >      >      >    Alma Mater Studiorum - Università di Bologna
>      >      >      >      >    V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
>      >      >      >      >    tel.: +39 051 20 95786
>      >      >      >      >
>      >      >      >
>      >      >      >    --
>      >      >      >    Diego Zuccato
>      >      >      >    DIFA - Dip. di Fisica e Astronomia
>      >      >      >    Servizi Informatici
>      >      >      >    Alma Mater Studiorum - Università di Bologna
>      >      >      >    V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
>      >      >      >    tel.: +39 051 20 95786
>      >      >      >
>      >      >
>      >      >    --
>      >      >    Diego Zuccato
>      >      >    DIFA - Dip. di Fisica e Astronomia
>      >      >    Servizi Informatici
>      >      >    Alma Mater Studiorum - Università di Bologna
>      >      >    V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
>      >      >    tel.: +39 051 20 95786
>      >      >
>      >
>      >    --
>      >    Diego Zuccato
>      >    DIFA - Dip. di Fisica e Astronomia
>      >    Servizi Informatici
>      >    Alma Mater Studiorum - Università di Bologna
>      >    V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
>      >    tel.: +39 051 20 95786
>      >
>
>    --
>    Diego Zuccato
>    DIFA - Dip. di Fisica e Astronomia
>    Servizi Informatici
>    Alma Mater Studiorum - Università di Bologna
>    V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
>    tel.: +39 051 20 95786
>

--
Diego Zuccato
DIFA - Dip. di Fisica e Astronomia
Servizi Informatici
Alma Mater Studiorum - Università di Bologna
V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
tel.: +39 051 20 95786
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux