this was trying to start up the vm
[25610] 2013-07-25 09:51:45.997895 lock: acquired /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
[25610] 2013-07-25 09:51:46.035698 Raising exception [46, The VDI is not available [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw not found]]
[25610] 2013-07-25 09:51:46.035831 lock: released /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
[25610] 2013-07-25 09:51:46.036917 ***** generic exception: vdi_attach: EXCEPTION SR.SROSError, The VDI is not available [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw not found]
File "/opt/xensource/sm/SRCommand.py", line 96, in run
return self._run_locked(sr)
File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked
target = sr.vdi(self.vdi_uuid)
File "/opt/xensource/sm/NFSSR", line 213, in vdi
return NFSFileVDI(self, uuid)
File "/opt/xensource/sm/VDI.py", line 102, in __init__
self.load(uuid)
File "/opt/xensource/sm/FileSR.py", line 370, in load
opterr="%s not found" % self.path)
File "/opt/xensource/sm/xs_errors.py", line 49, in __init__
raise SR.SROSError(errorcode, errormessage)
[25610] 2013-07-25 09:51:46.037204 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
and this is on a migrate(destination)
[29480] 2013-07-25 09:53:18.859918 lock: acquired /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
[29480] 2013-07-25 09:53:18.897479 Raising exception [46, The VDI is not available [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw not found]]
[29480] 2013-07-25 09:53:18.897609 lock: released /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
[29480] 2013-07-25 09:53:18.898701 ***** generic exception: vdi_attach: EXCEPTION SR.SROSError, The VDI is not available [opterr=/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.raw not found]
File "/opt/xensource/sm/SRCommand.py", line 96, in run
return self._run_locked(sr)
File "/opt/xensource/sm/SRCommand.py", line 137, in _run_locked
target = sr.vdi(self.vdi_uuid)
File "/opt/xensource/sm/NFSSR", line 213, in vdi
return NFSFileVDI(self, uuid)
File "/opt/xensource/sm/VDI.py", line 102, in __init__
self.load(uuid)
File "/opt/xensource/sm/FileSR.py", line 370, in load
opterr="%s not found" % self.path)
File "/opt/xensource/sm/xs_errors.py", line 49, in __init__
raise SR.SROSError(errorcode, errormessage)
[29480] 2013-07-25 09:53:18.898972 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
this is on migrate (source)
[16462] 2013-07-25 10:02:48.800862 blktap2.deactivate
[16462] 2013-07-25 10:02:48.800965 lock: acquired /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
[16462] 2013-07-25 10:02:48.819441 ['/usr/sbin/tap-ctl', 'close', '-p', '5578', '-m', '7']
[16462] 2013-07-25 10:02:49.295250 = 0
[16462] 2013-07-25 10:02:49.295467 ['/usr/sbin/tap-ctl', 'detach', '-p', '5578', '-m', '7']
[16462] 2013-07-25 10:02:49.299579 = 0
[16462] 2013-07-25 10:02:49.299794 ['/usr/sbin/tap-ctl', 'free', '-m', '7']
[16462] 2013-07-25 10:02:49.303645 = 0
[16462] 2013-07-25 10:02:49.303902 tap.deactivate: Shut down Tapdisk(vhd:/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd, pid=5578, minor=7, state=R)
[16462] 2013-07-25 10:02:49.485672 ['/usr/sbin/td-util', 'query', 'vhd', '-vpf', '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
[16462] 2013-07-25 10:02:49.510929 pread SUCCESS
[16462] 2013-07-25 10:02:49.537296 Removed host key host_OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080 for 72ad514a-f1f8-4a34-9907-9c6a3506520b
[16462] 2013-07-25 10:02:49.537451 lock: released /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
[16462] 2013-07-25 10:02:49.537540 lock: closed /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
[16462] 2013-07-25 10:02:49.537641 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
[16462] 2013-07-25 10:02:49.537862 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
[16636] 2013-07-25 10:02:50.103352 lock: acquired /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
[16636] 2013-07-25 10:02:50.117961 ['/usr/sbin/td-util', 'query', 'vhd', '-vpf', '/var/run/sr-mount/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/72ad514a-f1f8-4a34-9907-9c6a3506520b.vhd']
[16636] 2013-07-25 10:02:50.137963 pread SUCCESS
[16636] 2013-07-25 10:02:50.139106 vdi_detach {'sr_uuid': '9f9aa794-86c0-9c36-a99d-1e5fdc14a206', 'subtask_of': 'DummyRef:|ebe0d00f-b082-77ba-b209-095e71a0c1c7|VDI.detach', 'vdi_ref': 'OpaqueRef:31009428-3c98-c005-67ed-ddcc5e432e03', 'vdi_on_boot': 'persist', 'args': [], 'vdi_location': '72ad514a-f1f8-4a34-9907-9c6a3506520b', 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref': 'OpaqueRef:f4170801-402a-0935-a759-19a46e700a87', 'device_config': {'server': '10.254.253.9', 'SRmaster': 'true', 'serverpath': '/xen', 'options': ''}, 'command': 'vdi_detach', 'vdi_allow_caching': 'false', 'sr_ref': 'OpaqueRef:fefba283-7462-1f5a-b4e2-d58169c4b318', 'vdi_uuid': '72ad514a-f1f8-4a34-9907-9c6a3506520b'}
[16636] 2013-07-25 10:02:50.139415 lock: closed /var/lock/sm/72ad514a-f1f8-4a34-9907-9c6a3506520b/vdi
[16636] 2013-07-25 10:02:50.139520 lock: released /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
[16636] 2013-07-25 10:02:50.139779 lock: closed /var/lock/sm/9f9aa794-86c0-9c36-a99d-1e5fdc14a206/sr
[17886] 2013-07-25 10:03:16.326423 sr_scan {'sr_uuid': 'fc63fc27-89ca-dbc8-228d-27e3c74779bb', 'subtask_of': 'DummyRef:|2f34582a-2b3b-82be-b1f6-7f374565c8e8|SR.scan', 'args': [], 'host_ref': 'OpaqueRef:645996e3-d9cc-59e1-3842-65d679e9e080', 'session_ref': 'OpaqueRef:bfffd224-6edc-1cb4-9145-c0c95cbb063b', 'device_config': {'iso_path': '/iso', 'type': 'cifs', 'SRmaster': 'true', 'location': '//10.254.254.30/share'}, 'command': 'sr_scan', 'sr_ref': 'OpaqueRef:9c7f5cd0-fd88-16e2-2426-6e066a1183ab'}
----- Original Message -----
From: "Sébastien RICCIO" <sr@xxxxxxxxxxxxxxx>
To: "Andres E. Moya" <amoya@xxxxxxxxxxxxxxxxx>, "Alberto Castrillo" <castrillo@xxxxxxxxxx>
Cc: "xen-api" <xen-api@xxxxxxxxxxxxx>
Sent: Wednesday, July 24, 2013 10:55:40 PM
Subject: Re: [Xen-API] The vdi is not available
Hi,
When this happens, what does /var/log/SMlog says ?
Can you please tail -f /var/log/SMlog on both source and destination,
try to migrate the VM and paste the results?
Cheers,
Sébastien
On 24.07.2013 23:09, Andres E. Moya wrote:
I also just tried creating a new storage repository moving the vdi to the new storage repository is successful, i then try to migrate it to server C and still have the same issue
Moya Solutions, Inc.
amoya@xxxxxxxxxxxxxxxxx
0 | 646-918-5238 x 102
F | 646-390-1806
----- Original Message -----
From: "Alberto Castrillo" <castrillo@xxxxxxxxxx>
To: "xen-api" <xen-api@xxxxxxxxxxxxx>
Sent: Wednesday, July 24, 2013 4:12:13 PM
Subject: Re: [Xen-API] The vdi is not available
We use NFS as shared storage, and have faced some "VDI not available" issues with our VMs. I haven't been able to start a VM with the method of that URL in XCP 1.6 (in 1.1 and 1.5 beta worked). What worked for me:
- Detach the VDI from the VM
- Detach and forget the SR where the VDI is stored
- Reattach the forgotten SR (create new SR, give the same info that the detached SR, re-use the SR-UUID, ...)
- Reattach the VDI to the VM
El 24/07/2013, a las 21:10, hook escribió:
Past weekend (as usual O_o) we have experienced the issue in our XCP 1.6 production pool.
Shared iSCSI storage was shutted down due to misconfigured UPS settings while XCP servers continued to work.
When storage was returned to working state and reconnected to pool most VM did not boot with the same message - VDI is not available.
Googling give me mentioned above method - forgot and reconnect VDI.
Result was even worser - the whole SR become unusable.
Storage rescan gazered lot of errors like bad header on LVM and many other.
Finally i've disconnect failed SR from pool, connect it back and SR become healthy (it looks so). But anyone VM was not start with disk from this SR and freeze during startup.
I did not find solution and restored most VMs from backup (long live VMPP!)
So, i just wanna say - be highly careful with VDI on shared storage repository in production environment)
2013/7/24 Brian Menges < bmenges@xxxxxxxxxx >
Have you tried the following?: http://community.spiceworks.com/how_to/show/14199-xcp-xen-cloud-platform-xenserver-the-vdi-is-not-available
- Brian Menges
Principal Engineer, DevOps
GoGrid | ServePath | ColoServe | UpStream Networks
-----Original Message-----
From: xen-api-bounces@xxxxxxxxxxxxx [mailto: xen-api-bounces@xxxxxxxxxxxxx ] On Behalf Of Andres E. Moya
Sent: Wednesday, July 24, 2013 09:32
To: xen-api@xxxxxxxxxxxxx
Subject: [Xen-API] The vdi is not available
Guys need help trouble shooting this issue
I have an xcp 1.6 pool with 3 machines A,B, and C
I can migrate from A to B and B to A
WE cannot migrate from A or B to C, we also cannot shutdown a vm and start it up on C, when we do that we get the message The vdi is not available.
We have tried removing machine C from the pool and re joining and still have the issue.
when we first add host C to the pool it cannot load the nfs storage repository because we need to create a management interface from a bonded vlan that gets created after joining the pool. After we create the interface and run a re plug on the storage repository it says its connected / re plugged.
Thanks for any help in advance
_______________________________________________
Xen-api mailing list
Xen-api@xxxxxxxxxxxxx
http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api
________________________________
The information contained in this message, and any attachments, may contain confidential and legally privileged material. It is solely for the use of the person or entity to which it is addressed. Any review, retransmission, dissemination, or action taken in reliance upon this information by persons or entities other than the intended recipient is prohibited. If you receive this in error, please contact the sender and delete the material from any computer.
_______________________________________________
Xen-api mailing list
Xen-api@xxxxxxxxxxxxx
http://lists.xen.org/cgi-bin/mailman/listinfo/xen-api