[gpfsug-discuss] Trapped Inodes - releases, now count them properly!

Luke Raimbach Luke.Raimbach at crick.ac.uk
Mon Jul 4 10:44:02 BST 2016


Hi Marc,

Thanks again for the suggestions. An interesting report in the log while the another node took over managing the filesystem:


Mon Jul  4 10:24:08.616 2016: [W] Inode space 10 in file system gpfs is approaching the limit for the maximum number of inodes.


Inode space 10 was the independent fileset that the snapshot creation/deletion managed to remove. Still getting negative inode numbers reported after migrating manager functions and suspending/resuming the file system:

Inode Information
-----------------
Total number of used inodes in all Inode spaces:         -103900000
Total number of free inodes in all Inode spaces:          -24797856
Total number of allocated inodes in all Inode spaces:    -128697856
Total of Maximum number of inodes in all Inode spaces:   -103900000

I’ll have to wait until later today to try unmounting, daemon recycle or mmfsck.

Cheers,
Luke.

From: gpfsug-discuss-bounces at spectrumscale.org [mailto:gpfsug-discuss-bounces at spectrumscale.org] On Behalf Of Marc A Kaplan
Sent: 03 July 2016 19:43
To: gpfsug main discussion list <gpfsug-discuss at spectrumscale.org>
Subject: Re: [gpfsug-discuss] Trapped Inodes - releases, now count them properly!

mmdf statistics are not real-time accurate, there is a trade off in accuracy vs the cost of polling each node that might have the file system mounted.

That said, here are some possibilities, in increasing order of impact on users and your possible desperation ;-)

A1. Wait a while (at most a few minutes) and see if the mmdf stats are updated.

A2.
mmchmgr fs another-node

may force new stats to be sent to the new fs manager. (Not sure but I expect it will.)

B. Briefly quiesce the file system with:

mmfsctl fs suspend; mmfsctl fs resume;


C. If you have no users active ... I'm pretty sure

mmumount fs -a ; mmmount fs -a;

will  clear the problem ... but there's always

D.
mmshutdown -a ; mmstartup -a

E.
If none of those resolve the situation something is hosed --

F.
hope that

mmfsck

can fix it.


--marc



From:        Luke Raimbach <Luke.Raimbach at crick.ac.uk<mailto:Luke.Raimbach at crick.ac.uk>>
To:        gpfsug main discussion list <gpfsug-discuss at spectrumscale.org<mailto:gpfsug-discuss at spectrumscale.org>>
Date:        07/03/2016 10:55 AM
Subject:        Re: [gpfsug-discuss] Trapped Inodes
Sent by:        gpfsug-discuss-bounces at spectrumscale.org<mailto:gpfsug-discuss-bounces at spectrumscale.org>
________________________________



Hi Marc,

Thanks for that suggestions. This seems to have removed the NULL fileset from the list, however mmdf now shows even more strange statistics:

Inode Information
-----------------
Total number of used inodes in all Inode spaces:         -103900000
Total number of free inodes in all Inode spaces:          -24797856
Total number of allocated inodes in all Inode spaces:    -128697856
Total of Maximum number of inodes in all Inode spaces:   -103900000


Any ideas why these negative numbers are being reported?

Cheers,
Luke.

From: gpfsug-discuss-bounces at spectrumscale.org<mailto:gpfsug-discuss-bounces at spectrumscale.org> [mailto:gpfsug-discuss-bounces at spectrumscale.org] On Behalf Of Marc A Kaplan
Sent: 02 July 2016 20:17
To: gpfsug main discussion list <gpfsug-discuss at spectrumscale.org<mailto:gpfsug-discuss at spectrumscale.org>>
Subject: Re: [gpfsug-discuss] Trapped Inodes

I have been informed that it is possible that a glitch can occur (for example an abrupt shutdown) which can leave you in a situation where it looks like all snapshots are deleted, but there is still a hidden snapshot that must be cleaned up...

The workaround is to create a snapshot `mmcrsnapshot fs dummy`  and then delete it `mmdelsnapshot fs dummy` and see if that clears up the situation...

--marc



From:        Luke Raimbach <Luke.Raimbach at crick.ac.uk<mailto:Luke.Raimbach at crick.ac.uk>>
To:        gpfsug main discussion list <gpfsug-discuss at spectrumscale.org<mailto:gpfsug-discuss at spectrumscale.org>>
Date:        07/02/2016 06:05 AM
Subject:        Re: [gpfsug-discuss] Trapped Inodes
Sent by:        gpfsug-discuss-bounces at spectrumscale.org<mailto:gpfsug-discuss-bounces at spectrumscale.org>
________________________________




Hi Marc,

Thanks for the suggestion.

Snapshots were my first suspect but there are none anywhere on the filesystem.

Cheers,
Luke.

On 1 Jul 2016 5:30 pm, Marc A Kaplan <makaplan at us.ibm.com<mailto:makaplan at us.ibm.com>> wrote:
Question and Suggestion:  Do you have any snapshots that might include files that were in the fileset you are attempting to delete?  Deleting those snapshots will allow the fileset deletion to complete.  The snapshots are kinda intertwined with what was the "live" copy of the inodes. In the GPFS "ditto" implementation of snapshotting,  for a file that has not changed since the snapshot operation, the snapshot copy is not really a copy but just a pointer to the "live" file.   So even after you have logically deleted the "live" files, the snapshot still points to those inodes you thought you deleted.  Rather than invalidate the snapshot, (you wouldn't want that, would you?!) GPFS holds onto the inodes, until they are no longer referenced by any snapshot.

--marc



From:        Luke Raimbach <Luke.Raimbach at crick.ac.uk<mailto:Luke.Raimbach at crick.ac.uk>>
To:        gpfsug main discussion list <gpfsug-discuss at spectrumscale.org<mailto:gpfsug-discuss at spectrumscale.org>>
Date:        07/01/2016 06:32 AM
Subject:        [gpfsug-discuss] Trapped Inodes
Sent by:        gpfsug-discuss-bounces at spectrumscale.org<mailto:gpfsug-discuss-bounces at spectrumscale.org>

________________________________




Hi All,

I've run out of inodes on a relatively small filesystem. The total metadata capacity allows for a maximum of 188,743,680 inodes.

A fileset containing 158,000,000 inodes was force deleted and has gone into a bad state, where it is reported as (NULL) and has state "deleted":

Attributes for fileset (NULL):
===============================
Status                                  Deleted
Path                                    --
Id                                      15
Root inode                              latest:
Parent Id                               <none>
Created                                 Wed Jun 15 14:07:51 2016
Comment
Inode space                             8
Maximum number of inodes                158000000
Allocated inodes                        158000000
Permission change flag                  chmodAndSetacl
afm-associated                          No

Offline mmfsck fixed a few problems, but didn't free these poor, trapped inodes. Now I've run out and mmdf is telling me crazy things like this:

Inode Information
-----------------
Total number of used inodes in all Inode spaces:                  0
Total number of free inodes in all Inode spaces:           27895680
Total number of allocated inodes in all Inode spaces:      27895680
Total of Maximum number of inodes in all Inode spaces:     34100000


Current GPFS build: "4.2.0.3".

Who will help me rescue these inodes?

Cheers,
Luke.

Luke Raimbach​
Senior HPC Data and Storage Systems Engineer,
The Francis Crick Institute,
Gibbs Building,
215 Euston Road,
London NW1 2BE.

E: luke.raimbach at crick.ac.uk<mailto:luke.raimbach at crick.ac.uk>
W: www.crick.ac.uk

The Francis Crick Institute Limited is a registered charity in England and Wales no. 1140062 and a company registered in England and Wales no. 06885462, with its registered office at 215 Euston Road, London NW1 2BE.
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss

The Francis Crick Institute Limited is a registered charity in England and Wales no. 1140062 and a company registered in England and Wales no. 06885462, with its registered office at 215 Euston Road, London NW1 2BE. _______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss



The Francis Crick Institute Limited is a registered charity in England and Wales no. 1140062 and a company registered in England and Wales no. 06885462, with its registered office at 215 Euston Road, London NW1 2BE. _______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss



The Francis Crick Institute Limited is a registered charity in England and Wales no. 1140062 and a company registered in England and Wales no. 06885462, with its registered office at 215 Euston Road, London NW1 2BE.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20160704/1ddc84d1/attachment-0002.htm>


More information about the gpfsug-discuss mailing list