[gpfsug-discuss] File system recovery question

Craig Wilson craigawilson at gmail.com
Wed Feb 13 16:48:32 GMT 2013


Dealt with a similar issue a couple of months ago.

In that case the data was fine but two of the descriptors were over
written. You can use "mmfsadm test readdescraw /dev/$drive" to see the
descriptors, we managed to recover the disk but only after logging it to
IBM and manually rebuilding the descriptor.

-CW

On 13 February 2013 16:43, Jez Tucker <Jez.Tucker at rushes.co.uk> wrote:

> So, er. Fun.
>
> I checked our disks.
>
> 0000000 000000 000000 000000 000000 000000 000000 000000 000000
> *
> 0001000 <some data from hereon>
>
> Looks like you lost a fair bit.
>
>
> Presumably you don't have replication of 2?
> If so, I think you could just lose the NSD.
>
> Failing that:
>
> 1) Check your other disks and see if there's anything that you can figure
> out.  Though TBH, this may take forever.
> 2) Restore
> 3) Call IBM and log a SEV 1.
>
> 3) then 2) is probably the best course of action
>
> Jez
>
>
>
> > -----Original Message-----
> > From: gpfsug-discuss-bounces at gpfsug.org [mailto:gpfsug-discuss-
> > bounces at gpfsug.org] On Behalf Of Chad Kerner
> > Sent: 13 February 2013 16:29
> > To: gpfsug-discuss at gpfsug.org
> > Subject: [gpfsug-discuss] File system recovery question
> >
> > I have a file system, and it appears that someone dd'd over the first
> > part of one of the NSD's with zero's.  I see the device in multipath. I
> > can fdisk and dd the device out.
> >
> > Executing od shows it is zero's.
> > (! 21)-> od /dev/mapper/dh1_vd05_005 | head -n 5
> > 0000000 000000 000000 000000 000000 000000 000000 000000 000000
> > *
> > 0040000 120070 156006 120070 156006 120070 156006 120070 156006
> >
> > Dumping the header of one of the other disks shows read data for
> > the other NSD's in that file system.
> >
> > (! 25)-> mmlsnsd -m | grep dh1_vd05_005
> >  Disk name    NSD volume ID      Device         Node name
> > Remarks
> >
> ---------------------------------------------------------------------------------------
> >  dh1_vd05_005 8D8EEA98506C69CE   -              myhost (not found) server
> > node
> >
> > (! 27)-> mmnsddiscover -d dh1_vd05_005
> > mmnsddiscover:  Attempting to rediscover the disks.  This may take a
> > while ...
> > myhost:  Rediscovery failed for dh1_vd05_005.
> > mmnsddiscover:  Finished.
> >
> >
> > Wed Feb 13 09:14:03.694 2013: Command: mount desarchive Wed Feb
> > 13 09:14:07.101 2013: Disk failure.  Volume desarchive. rc = 19. Physical
> > volume dh1_vd05_005.
> > Wed Feb 13 09:14:07.102 2013: File System desarchive unmounted by
> > the system with return code 5 reason code 0 Wed Feb 13 09:14:07.103
> > 2013: Input/output error Wed Feb 13 09:14:07.102 2013: Failed to open
> > desarchive.
> > Wed Feb 13 09:14:07.103 2013: Input/output error Wed Feb 13
> > 09:14:07.102 2013: Command: err 666: mount desarchive Wed Feb 13
> > 09:14:07.104 2013: Input/output error Wed Feb 13 09:14:07 CST 2013:
> > mmcommon preunmount invoked.  File system: desarchive  Reason:
> > SGPanic
> >
> > Is there any way to repair the header on the NSD?
> >
> > Thanks for any ideas!
> > Chad
> >
> >
> > _______________________________________________
> > gpfsug-discuss mailing list
> > gpfsug-discuss at gpfsug.org
> > http://gpfsug.org/mailman/listinfo/gpfsug-discuss
>
>
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at gpfsug.org
> http://gpfsug.org/mailman/listinfo/gpfsug-discuss
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20130213/02615dcc/attachment-0003.htm>


More information about the gpfsug-discuss mailing list