Re: Healing issues

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Yep, you have to bring back hydra4:/gluster1/data

You can mount again /gluster1/data (don't forget SELINUX) and then use gluster's reset-brick to rebuild that brick.

Most probably you have an entry that was saved on hydra4:/gluster1/data and the arbiter but was not pushed to the surviving brick (based on the entry in the arbiter)

Usually, I use method 2 from https://docs.gluster.org/en/latest/Troubleshooting/gfid-to-path/ to identify the file and then you can check the status of that file on the bricks (and if necessary restore from backup). You should do that only after you have brought back hydra4:/gluster1/data to the volume.

You should also reconsider the brick layout, as in some cases a single hydra failure (whole server) will kill a subvolume and all VMs' disks on that subvolume will be unavailable.


Best Regards,
Strahil Nikolov

On Sat, Jul 31, 2021 at 2:18, Valerio Luccio
<valerio.luccio@xxxxxxx> wrote:

Strahil,

did some more digging into the heal info output.

Found also the following:

[...]
Brick hydra4:/gluster1/data                                                                                                                                                           
Status: Transport endpoint is not connected                                                                                                                                           
Number of entries: -                                                                                                                                                                  
                                                                                                                                                                                      
Brick hydra3:/arbiter/2                                                                                                                                                               
<gfid:2aa223b0-77f5-441e-bc76-34c8d459eeaa> 
[...]

So, to augment what I wrote before, the errors appear in "Brick hydra3:/gluster3/data" and "Brick hydra3:/arbiter/2", plus that "Transport endpoint is not connected" for "Brick hydra4:/gluster1/data".

I need to add that hydra4:/gluster1 is the RAID that had major hardware failure.


--
As a result of Coronavirus-related precautions, NYU and the Center for Brain Imaging operations will be managed remotely until further notice.
All telephone calls and e-mail correspondence are being monitored remotely during our normal business hours of 9am-5pm, Monday through Friday.
 
For MRI scanner-related emergency, please contact: Keith Sanzenbach at keith.sanzenbach@xxxxxxx and/or Pablo Velasco at pablo.velasco@xxxxxxx
For computer/hardware/software emergency, please contact: Valerio Luccio at valerio.luccio@xxxxxxx
For TMS/EEG-related emergency, please contact: Chrysa Papadaniil at chrysa@xxxxxxx
For CBI-related administrative emergency, please contact: Jennifer Mangan at jennifer.mangan@xxxxxxx

Valerio Luccio   (212) 998-8736
Center for Brain Imaging   4 Washington Place, Room 158
New York University   New York, NY 10003

"In an open world, who needs windows or gates ?"
________



Community Meeting Calendar:

Schedule -
Every 2nd and 4th Tuesday at 14:30 IST / 09:00 UTC
Bridge: https://meet.google.com/cpu-eiue-hvk
Gluster-users mailing list
Gluster-users@xxxxxxxxxxx
https://lists.gluster.org/mailman/listinfo/gluster-users

[Index of Archives]     [Gluster Development]     [Linux Filesytems Development]     [Linux ARM Kernel]     [Linux ARM]     [Linux Omap]     [Fedora ARM]     [IETF Annouce]     [Bugtraq]     [Linux OMAP]     [Linux MIPS]     [eCos]     [Asterisk Internet PBX]     [Linux API]

  Powered by Linux