Re: lvdisplay "open" field computing

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Veeambackup seems to access the volume (/var/log/veeam/veeamsvc.log):

[02.07.2019 00:01:18] <140503985145600> lpbcore| Found device: [/dev/dm-3]. Device number: [253:3]; Type: [dm]. [02.07.2019 00:01:18] <140503985145600> lpbcore|         Dm name: [vg_obm-var_spool_imap]. [02.07.2019 00:01:18] <140503985145600> lpbcore|         Dm UUID: [LVM-1i1v6pEjab2WslaDQRvkf8eLk6QfBW4J0oK7h0tDZHeUxEAyBwcK7xU9pSW7X4Uh].
[02.07.2019 00:01:18] <140503985145600> lpbcore| Slave of: [8:64].
[02.07.2019 00:01:18] <140503985145600> lpbcore| Link: [/dev/dm-3].
[02.07.2019 00:01:18] <140503985145600> lpbcore| Link: [/dev/vg_obm/var_spool_imap].
[02.07.2019 00:01:18] <140503985145600> lpbcore| Link: [/dev/block/253:3].
[02.07.2019 00:01:18] <140503985145600> lpbcore| Link: [/dev/mapper/vg_obm-var_spool_imap]. [02.07.2019 00:01:18] <140503985145600> lpbcore| Link: [/dev/disk/by-uuid/c9f9314e-4ce0-4138-b9b5-c745fdc22258]. [02.07.2019 00:01:18] <140503985145600> lpbcore| Link: [/dev/disk/by-id/dm-uuid-LVM-1i1v6pEjab2WslaDQRvkf8eLk6QfBW4J0oK7h0tDZHeUxEAyBwcK7xU9pSW7X4Uh]. [02.07.2019 00:01:18] <140503985145600> lpbcore| Link: [/dev/disk/by-id/dm-name-vg_obm-var_spool_imap]. [02.07.2019 00:01:18] <140503985145600> lpbcore| Filesystem UUID: [c9f9314e-4ce0-4138-b9b5-c745fdc22258]; Type: [ext4]; Mount points: [/var/spool/imap].

I wonder if veeam uses its own namespace.

Regards

On 10/07/2019 13:54, Simon ELBAZ wrote:
Hi Zdenek,

Thanks for your feedback.

The kernel version is:

[root@panoramix ~]# uname -a
Linux panoramix.ch-perrens.fr 2.6.32-573.12.1.el6.x86_64 #1 SMP Tue Dec 15 21:19:08 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux

This server is a corosync/pacemaker cluster node experiencing troubles at failover.

The error happening during the failover is:

Jul  2 03:08:09 panoramix pgsql(pri_postgres)[18708]: INFO: Changing pri_postgres-status on jolitorax.ch-perrens.fr : HS:alone->HS:async. Jul  2 03:08:10 panoramix LVM(pri_ISCSIVG0_vg_obm)[18618]: ERROR: Logical volume vg_obm/var_spool_imap in use. Can't deactivate volume group "vg_obm" with 1 open logical volume(s) Jul  2 03:08:10 panoramix LVM(pri_ISCSIVG0_vg_obm)[18618]: ERROR: LVM: vg_obm did not stop correctly Jul  2 03:08:10 panoramix LVM(pri_ISCSIVG0_vg_obm)[18618]: WARNING: vg_obm still Active Jul  2 03:08:10 panoramix LVM(pri_ISCSIVG0_vg_obm)[18618]: INFO: Retry deactivating volume group vg_obm

This is why I am trying to understand how the field is computed.

The grep in mountinfo is attached to the mail.

The volume is presented as a raw device by VMWare hypervisor and is backuped by Veeam:

[root@panoramix ~]# ps -ef | grep vee
root      2877     1  0 Jul04 ?        00:01:10 /usr/sbin/veeamservice --daemonize --pidfile=/var/run/veeamservice.pid
root      5207     2  0 Jul05 ?        00:00:00 [veeamsnap_log]
root     21475 21004  0 13:53 pts/2    00:00:00 grep vee

Regards


On 10/07/2019 12:51, Zdenek Kabelac wrote:
Dne 10. 07. 19 v 9:13 Simon ELBAZ napsal(a):
Hi,

This LV seems mounted on single mountpoint.

The lsof output is:


Hmm

Which kernel version is this - aren't we tracking issue for some ancient kernel ?

Why have you actually started to hunt the reason for opencount 2 ?

Has the machine experienced some troubles ?

Aren't there some namespaces in use - so i.e. volume is used in different namespace ??

grep 253:3  /proc/*/mountinfo  -  showing something unusual ?



Zdenek

--
Simon Elbaz
@Linagora
Mob: +33 (0) 6 38 99 18 34
Tour Franklin 31ème étage
100/101 Quartier Boieldieu
92042 La Défense
FRANCE

_______________________________________________
linux-lvm mailing list
linux-lvm@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/linux-lvm
read the LVM HOW-TO at http://tldp.org/HOWTO/LVM-HOWTO/




[Index of Archives]     [Gluster Users]     [Kernel Development]     [Linux Clusters]     [Device Mapper]     [Security]     [Bugtraq]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]

  Powered by Linux