[ClusterLabs] lvm on shared storage and a lot of...

Ken Gaillot kgaillot at redhat.com
Tue Apr 18 16:40:24 CEST 2017


On 04/18/2017 09:14 AM, lejeczek wrote:
> 
> 
> On 18/04/17 14:45, Digimer wrote:
>> On 18/04/17 07:31 AM, lejeczek wrote:
>>> .. device_block & device_unblock in dmesg.
>>>
>>> and I see that the LVM resource would fail.
>>> This to me seems to happen randomly, or I fail to spot a pattern.
>>>
>>> Shared storage is a sas3 enclosure.
>>> I believe I follow docs on LVM to the letter. I don't know what could be
>>> the problem.
>>>
>>> would you suggest ways to troubleshoot it? Is it faulty/failing hardware?
>>>
>>> many thanks,
>>> L.
>> LVM or clustered LVM?
>>
> no clvmd
> And inasmuch as the resource would start, fs would mount, if I start
> using it more intensely I'd get more of block/unblock and after a while
> mountpoint resource failes and then LVM resource too.
> It gets only worse after, even after I deleted resourced, I begin to
> see, eg.:
> 
> [ 6242.606870] sd 7:0:32:0: device_unblock and setting to running,
> handle(0x002c)
> [ 6334.248617] sd 7:0:18:0: [sdy] tag#0 FAILED Result: hostbyte=DID_OK
> driverbyte=DRIVER_SENSE
> [ 6334.248633] sd 7:0:18:0: [sdy] tag#0 Sense Key : Not Ready [current]
> [ 6334.248640] sd 7:0:18:0: [sdy] tag#0 Add. Sense: Logical unit is in
> process of becoming ready

This feels like a hardware issue to me. Have you checked the SMART data
on the drives?

> [ 6334.248647] sd 7:0:18:0: [sdy] tag#0 CDB: Read(10) 28 00 00 00 00 00
> 00 00 08 00
> [ 6334.248652] blk_update_request: I/O error, dev sdy, sector 0



More information about the Users mailing list