Avatar of gmbaxter
gmbaxter
Flag for United Kingdom of Great Britain and Northern Ireland asked on

Xen Server Storage Issue

Hi,

I've just upgraded a XenServer host from 5.5.0 to 5.6.0 SP2, however on restart, I get the following message in XenCenter:

Failed to attached storage on server start

This refers to a 2 TB lun on a NetApp filer via Fibre Channel. Other hosts in the pool (not yet upgraded) can see the lun ok, yet it's status is broken as the host reports to be "unplugged" from the storage.

This is 1 of 8 luns accessible - other FC luns are ok.

No idea what to do now, so any help would be greatly appreciated!

Thanks.
CitrixVirtualizationStorage

Avatar of undefined
Last Comment
gmbaxter

8/22/2022 - Mon
ded_ch

I'm sure you tried right clicking the storage and elect "repair". What did that do?
Have you checked the log files on your upgraded host?
You'll find the important ones here:

/var/log/messages
/var/log/xensource.log

You can use a tool like WinSCP to connect to the Dom0 and look at the content.
Check for any errors, especially around boot time, and post back.
gmbaxter

ASKER
Yes, I tried repair - the error mentioned PBDs

/var/log/messages mentioned:
very big device

xensource.log extract attached, where it tries to attach the SR. PBD is mentioned in here also.

Thanks for the quick reply by the way.
xensource-extract.txt
ded_ch

Try to look up the messages in SMLog, right around the time the xensource log reported the error. SMlog is located in the same folder /var/log/
I forgot that storage operations are mostly logged in SMLog.
Experts Exchange is like having an extremely knowledgeable team sitting and waiting for your call. Couldn't do my job half as well as I do without it!
James Murphy
gmbaxter

ASKER
Extract below, but the full file is attached:

[14279] 2012-12-22 22:50:51.246524	SUCCESS
[14279] 2012-12-22 22:50:51.246669	Synching Metadata volume
[14279] 2012-12-22 22:50:51.246766	['/usr/sbin/lvchange', '-ay', '/dev/VG_XenStorage-25b74c46-b305-c164-1070-6bf5abb49c5b/MGT']
[14279] 2012-12-22 22:50:51.341808	SUCCESS
[14279] 2012-12-22 22:50:51.342376	lock: released /var/lock/sm/25b74c46-b305-c164-1070-6bf5abb49c5b/sr
[14279] 2012-12-22 22:50:51.351858	***** sr_attach: EXCEPTION exceptions.AssertionError, 
  File "/opt/xensource/sm/SRCommand.py", line 94, in run
    return self._run_locked(sr)
  File "/opt/xensource/sm/SRCommand.py", line 131, in _run_locked
    return self._run(sr, target)
  File "/opt/xensource/sm/SRCommand.py", line 223, in _run
    return sr.attach(self.params['sr_uuid'])
  File "/opt/xensource/sm/LVMoHBASR", line 124, in attach
    LVHDSR.LVHDSR.attach(self, sr_uuid)
  File "/opt/xensource/sm/LVHDSR.py", line 309, in attach
    self._checkMetadataVolume(self.sm_config)
  File "/opt/xensource/sm/LVHDSR.py", line 201, in _checkMetadataVolume
    self._synchMetaData(map)
  File "/opt/xensource/sm/LVHDSR.py", line 211, in _synchMetaData
    xml = metadata.retrieveXMLfromFile(self.mdpath)
  File "/opt/xensource/sm/metadata.py", line 103, in retrieveXMLfromFile
    _testHdr(hdr)
  File "/opt/xensource/sm/metadata.py", line 41, in _testHdr
    assert(hdr[2] <= MD_MAJOR)

[14279] 2012-12-22 22:50:51.357473	lock: closed /var/lock/sm/25b74c46-b305-c164-1070-6bf5abb49c5b/sr
[14451] 2012-12-22 22:50:51.373612	LVMCache created for VG_XenStorage-76c86946-ce7a-6690-ba60-2692a0b5e303
<14451> 2012-12-22 22:50:51.514441	Host f8b94db1-26eb-4bcb-96ac-293075c52204 not enabled
<14451> 2012-12-22 22:50:51.514677	Pool is not ready, exiting
<14451> 2012-12-22 22:50:51.514750	SR 76c8 ('NetApp 2.5TB 1st floor') (0 VDIs in 0 VHD trees): no changes
[14491] 2012-12-22 22:50:51.699425	LVMCache created for VG_XenStorage-c06d1738-f71a-4b4a-d7df-3685db30e2a8
[14491] 2012-12-22 22:50:51.708770	['/usr/sbin/vgs', 'VG_XenStorage-c06d1738-f71a-4b4a-d7df-3685db30e2a8']
[14491] 2012-12-22 22:50:51.841909	SUCCESS
[14491] 2012-12-22 22:50:51.842105	lock: acquired /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/sr
[14491] 2012-12-22 22:50:51.842172	LVMCache: will initialize now
[14491] 2012-12-22 22:50:51.842218	LVMCache: refreshing
[14491] 2012-12-22 22:50:51.842288	['/usr/sbin/lvs', '--noheadings', '--units', 'b', '-o', '+lv_tags', '/dev/VG_XenStorage-c06d1738-f71a-4b4a-d7df-3685db30e2a8']
[14491] 2012-12-22 22:50:51.921928	SUCCESS
[14491] 2012-12-22 22:50:51.922361	lock: released /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/sr
[14491] 2012-12-22 22:50:51.922611	lock: closed /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/sr
[14491] 2012-12-22 22:50:51.922684	LVMCache created for VG_XenStorage-c06d1738-f71a-4b4a-d7df-3685db30e2a8
[14491] 2012-12-22 22:50:51.932351	['/usr/sbin/vgs', 'VG_XenStorage-c06d1738-f71a-4b4a-d7df-3685db30e2a8']
[14491] 2012-12-22 22:50:52.008071	SUCCESS
[14491] 2012-12-22 22:50:52.008258	lock: acquired /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/sr
[14491] 2012-12-22 22:50:52.008326	LVMCache: will initialize now
[14491] 2012-12-22 22:50:52.008371	LVMCache: refreshing
[14491] 2012-12-22 22:50:52.008442	['/usr/sbin/lvs', '--noheadings', '--units', 'b', '-o', '+lv_tags', '/dev/VG_XenStorage-c06d1738-f71a-4b4a-d7df-3685db30e2a8']
[14491] 2012-12-22 22:50:52.147745	SUCCESS
[14491] 2012-12-22 22:50:52.148196	lock: released /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/sr
[14491] 2012-12-22 22:50:52.148327	lock: acquired /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/sr
[14491] 2012-12-22 22:50:52.148461	sr_attach {'sr_uuid': 'c06d1738-f71a-4b4a-d7df-3685db30e2a8', 'subtask_of': 'OpaqueRef:94fadc0d-5104-cea7-8f83-46c3f9ab1b5c', 'args': [], 'host_ref': 'OpaqueRef:a69330e3-757f-8ff5-6099-d397d54caf85', 'session_ref': 'OpaqueRef:e1791520-24a6-f91e-3670-373e4153111f', 'device_config': {'device': '/dev/disk/mpInuse/360a98000572d5038636f58795a6c3753', 'SCSIid': '360a98000572d5038636f58795a6c3753', 'SRmaster': 'true'}, 'command': 'sr_attach', 'sr_ref': 'OpaqueRef:ac0b5d46-1ac7-7051-83c0-d1e720b5779f'}
[14491] 2012-12-22 22:50:52.148766	lock: acquired /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/running
[14491] 2012-12-22 22:50:52.148830	lock: acquired /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/sr
[14491] 2012-12-22 22:50:52.158463	RESET for SR c06d1738-f71a-4b4a-d7df-3685db30e2a8 (master: True)
[14491] 2012-12-22 22:50:52.195078	lock: released /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/sr
[14491] 2012-12-22 22:50:52.195159	lock: released /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/running
[14491] 2012-12-22 22:50:52.195211	lock: closed /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/sr
[14491] 2012-12-22 22:50:52.195261	lock: closed /var/lock/sm/c06d1738-f71a-4b4a-d7df-3685db30e2a8/running
[14491] 2012-12-22 22:50:52.202113	set_dirty 'OpaqueRef:ac0b5d46-1ac7-7051-83c0-d1e720b5779f' succeeded
[14491] 2012-12-22 22:50:52.202177	MPATH: multipath activate called
[14491] 2012-12-22 22:50:52.202227	['/opt/xensource/sm/mpathHBA', 'enable']
[14491] 2012-12-22 22:50:52.221720	SUCCESS

Open in new window

SMlog.txt
gmbaxter

ASKER
From /var/log/messages

Dec 22 22:50:51 bcxenhostb3 fe: 14283 (/opt/xensource/sm/LVMoHBASR <methodCall><methodName>sr_scan</methodName><para...) exitted with code 0
Dec 22 22:50:51 bcxenhostb3 fe: 14279 (/opt/xensource/sm/LVMoHBASR <methodCall><methodName>sr_attach</methodName><pa...) exitted with code 1
Dec 22 22:50:51 bcxenhostb3 xapi: [error|bcxenhostb3|0 thread_zero|creating storage D:aeefcbfe0dce|xapi] Could not plug in pbd '8f3869e9-38c6-e041-da4e-5a21ae84dc61': Api_errors.Server_error("SR_BACKEND_FAILURE", _)
Dec 22 22:50:52 bcxenhostb3 multipathd: sdam: add path (operator)
Dec 22 22:50:52 bcxenhostb3 multipathd: sdam: spurious uevent, path already in pathvec
ded_ch

Unfortunately, there is not much info in the file that I could help you with. I would highly suggest you open an account with the official citrix forum and post your file outputs there if not already done so.
There are Citrix employees on these forums that might be able to help you further.

Also, I assume you disabled HA on your Server before doing the upgrade? How about Multipathing? If Multipathing is still active on your xenservers, try disabling that on all of them and then try to attach the storage again.
⚡ FREE TRIAL OFFER
Try out a week of full access for free.
Find out why thousands trust the EE community with their toughest problems.
gmbaxter

ASKER
Thanks, I have posted all of this information over there to see if anything comes out of it,

HA isn't licensed as it is the free version of xenserver, unsure on multipathing? This only affects the one I upgraded to 5.6 from 5.5, so I still have the option to rollback, and try again I suppose.
ded_ch

Yes, you can start over, especially if you have run a "Xe pool-dump-database" and saved the resulting file somewhere offsite.
Then you can simply reinstall the old version and run the restore of the PoolDB on that host.
gmbaxter

ASKER
Do i have to restore the DB? I have reverted the pool master back to 5.5 but still have the SR issue
Your help has saved me hundreds of hours of internet surfing.
fblack61
basraj

check if the Ip address of the storage or the account used is changed.
gmbaxter

ASKER
Its FC storage
ASKER CERTIFIED SOLUTION
gmbaxter

THIS SOLUTION ONLY AVAILABLE TO MEMBERS.
View this solution by signing up for a free trial.
Members can start a 7-Day free trial and enjoy unlimited access to the platform.
See Pricing Options
Start Free Trial
GET A PERSONALIZED SOLUTION
Ask your own question & get feedback from real experts
Find out why thousands trust the EE community with their toughest problems.
gmbaxter

ASKER
Correct steps to resolve the sr not appearing.
⚡ FREE TRIAL OFFER
Try out a week of full access for free.
Find out why thousands trust the EE community with their toughest problems.