[gpfsug-discuss] Mmrestripefs -R --metadat-only - how to estimate remaining execution time
Billich Heinrich Rainer (ID SD)
heinrich.billich at id.ethz.ch
Fri May 28 09:55:53 BST 2021
Hello,
I just noticed:
Maybe mmrestripefs does some extra processing on snapshots? The output looks much more as expected on filesystems with no snapshots present, both number of inodes and MB data processed allow to estimate the remaining runtime. Unfortunately all our large production filesystems use snapshots …
Cheers,
Heiner
1.23 % complete on Tue May 18 21:07:10 2021 ( 37981765 inodes with total 328617 MB data processed)
1.25 % complete on Tue May 18 21:17:44 2021 ( 39439584 inodes with total 341032 MB data processed)
100.00 % complete on Tue May 18 21:22:44 2021 ( 41312000 inodes with total 356088 MB data processed). <<<< # of inodes matches # of allocated inodes, MB processed matches used metadata disk space
Scan completed successfully.
# mmdf fsxxx -F
Inode Information
-----------------
Total number of used inodes in all Inode spaces: 29007227
Total number of free inodes in all Inode spaces: 12304773
Total number of allocated inodes in all Inode spaces: 41312000 <<<<<<<<<< allocated inodes
Total of Maximum number of inodes in all Inode spaces: 87323648
]# mmdf fsxxx -m --block-size M
disk disk size failure holds holds free in MB free in MB
name in MB group metadata data in full blocks in fragments
--------------- ------------- -------- -------- ----- -------------------- -------------------
Disks in storage pool: system (Maximum disk size allowed is 4.13 TB)
------------- -------------------- -------------------
(pool total) 2425152 2067720 ( 85%) 1342 ( 0%). <<<<<< 356190MB used
From: <gpfsug-discuss-bounces at spectrumscale.org> on behalf of "Billich Heinrich Rainer (ID SD)" <heinrich.billich at id.ethz.ch>
Reply to: gpfsug main discussion list <gpfsug-discuss at spectrumscale.org>
Date: Friday, 28 May 2021 at 10:25
To: gpfsug main discussion list <gpfsug-discuss at spectrumscale.org>
Subject: [gpfsug-discuss] Mmrestripefs -R --metadat-only - how to estimate remaining execution time
Hello,
I want to estimate how much longer a running
mmrestripefs -R –metadata-only –qos maintenance
job will take to finish. We did switch from ‘-m 1’ to ‘-m 2 ’ and now run mmrestripefs to get the second copy of all metadata. I know that the % values in the output are useless, but now it looks like the number of inodes processed is of no use, too? The filesystem has about 1e9 allocated inodes, but the mmrestripefs reports up to 7e9 inodes processed? The number of inodes reported increases heavily since it passed the number of allocated inodes???
4.41 % complete on Fri May 28 09:22:05 2021 (7219160122 inodes with total 8985067 MB data processed)
Should I consider the ‘MB data processed’ instead – will the job finish when all used metadata disk space is processed?
I see little iops on the metadata disks and wonder if something is stuck, I’m not sure if it makes sense to wait any longer. But without an estimate on how much longer the job will take I hesitate to restart. A restart will need to scan all metadata again … Metadata is on SSD or NVMe, disks iops never were limiting. I turned qos off on the filesystem, previously I restricted the iops. But I don’t see any increase in iops.
Thank you. I know this is an old issue, but still I would strongly welcome any advice or comment.
Cheers,
Heiner
I’m aware of RFE ID:150409 mmrestripefs % complete metric is near useless. I’m not sure if it covers the metadata-only case, too.
Inode Information
-----------------
Total number of used inodes in all Inode spaces: 735005692
Total number of free inodes in all Inode spaces: 335462660
Total number of allocated inodes in all Inode spaces: 1070468352
Total of Maximum number of inodes in all Inode spaces: 1223782912
Mmrestripefs output:
START Mon May 24 20:27:25 CEST 2021
Scanning file system metadata, phase 1 ...
100 % complete on Mon May 24 21:49:27 2021
Scan completed successfully.
Scanning file system metadata, phase 2 ...
100 % complete on Mon May 24 21:49:28 2021
Scanning file system metadata for data storage pool
90 % complete on Mon May 24 21:49:34 2021
100 % complete on Mon May 24 21:49:35 2021
Scanning file system metadata for Capacity storage pool
49 % complete on Mon May 24 21:49:41 2021
100 % complete on Mon May 24 21:49:46 2021
Scan completed successfully.
Scanning file system metadata, phase 3 ...
Scan completed successfully.
Scanning file system metadata, phase 4 ...
100 % complete on Mon May 24 21:49:48 2021
Scan completed successfully.
Scanning file system metadata, phase 5 ...
100 % complete on Mon May 24 21:49:52 2021
Scan completed successfully.
Scanning user file metadata ...
0.01 % complete on Mon May 24 21:50:13 2021 ( 613108 inodes with total 76693 MB data processed)
0.01 % complete on Mon May 24 21:50:33 2021 ( 1090048 inodes with total 80495 MB data processed)
0.01 % complete on Mon May 24 21:50:53 2021 ( 1591808 inodes with total 84576 MB data processed)
…
3.97 % complete on Thu May 27 22:01:02 2021 (1048352497 inodes with total 8480467 MB data processed)
3.99 % complete on Thu May 27 22:30:18 2021 (1050254855 inodes with total 8495969 MB data processed)
4.01 % complete on Thu May 27 22:59:39 2021 (1052304294 inodes with total 8512683 MB data processed)
4.03 % complete on Thu May 27 23:29:11 2021 (1055390220 inodes with total 8537615 MB data processed)
4.05 % complete on Thu May 27 23:58:58 2021 (1059333989 inodes with total 8568871 MB data processed)
4.07 % complete on Fri May 28 00:28:48 2021 (1064728403 inodes with total 8611605 MB data processed)
4.09 % complete on Fri May 28 00:58:50 2021 (1067749260 inodes with total 8636120 MB data processed). <<< approximately number of allocated inodes
4.11 % complete on Fri May 28 01:29:00 2021 (1488665433 inodes with total 8661588 MB data processed) <<< whats going??
4.13 % complete on Fri May 28 01:59:23 2021 (1851124480 inodes with total 8682324 MB data processed)
4.15 % complete on Fri May 28 02:29:55 2021 (1885948840 inodes with total 8700082 MB data processed)
4.17 % complete on Fri May 28 03:00:38 2021 (2604503808 inodes with total 8724069 MB data processed)
4.19 % complete on Fri May 28 03:31:38 2021 (2877196260 inodes with total 8746504 MB data processed)
4.21 % complete on Fri May 28 04:02:38 2021 (2933166080 inodes with total 8762555 MB data processed)
4.23 % complete on Fri May 28 04:33:48 2021 (2956295936 inodes with total 8782298 MB data processed)
4.25 % complete on Fri May 28 05:05:09 2021 (3628799151 inodes with total 8802452 MB data processed)
4.27 % complete on Fri May 28 05:36:40 2021 (3970093965 inodes with total 8823885 MB data processed)
4.29 % complete on Fri May 28 06:08:20 2021 (4012553472 inodes with total 8841407 MB data processed)
4.31 % complete on Fri May 28 06:40:11 2021 (4029545087 inodes with total 8858676 MB data processed)
4.33 % complete on Fri May 28 07:12:11 2021 (6080613874 inodes with total 8889395 MB data processed)
4.35 % complete on Fri May 28 07:44:21 2021 (6146937531 inodes with total 8907253 MB data processed)
4.37 % complete on Fri May 28 08:16:45 2021 (6167408718 inodes with total 8925236 MB data processed)
4.39 % complete on Fri May 28 08:49:18 2021 (7151592448 inodes with total 8968126 MB data processed)
4.41 % complete on Fri May 28 09:22:05 2021 (7219160122 inodes with total 8985067 MB data processed)
# mmdf fsxxxx -m --block-size M
disk disk size failure holds holds free in MB free in MB
name in MB group metadata data in full blocks in fragments
--------------- ------------- -------- -------- ----- -------------------- -------------------
Disks in storage pool: system (Maximum disk size allowed is 32.20 TB)
fsxxxx_rg07b_1 2861295 3 yes no 1684452 ( 59%) 35886 ( 1%). << 1’140’957MB used – old nsd
fsxxxx_rg07a_1 2861295 3 yes no 1685002 ( 59%) 35883 ( 1%
fsxxxx_rg03b_1 2861295 3 yes no 1681515 ( 59%) 35627 ( 1%)
fsxxxx_rg03a_1 2861295 3 yes no 1680859 ( 59%) 35651 ( 1%)
RG003LG004VS015 2046239 12 yes no 904369 ( 44%) 114 ( 0%) << 1’141’756MB used – newly added nsd
RG003LG003VS015 2046239 12 yes no 904291 ( 44%) 118 ( 0%)
RG003LG002VS015 2046239 12 yes no 903632 ( 44%) 114 ( 0%)
RG003LG001VS015 2046239 12 yes no 903247 ( 44%) 115 ( 0%)
------------- -------------------- -------------------
(pool total) 19630136 10347367 ( 53%) 143503 ( 1%)
We run spectrum scale 5.0.5.4/5.0.5.5 on Power.
Filesystem:
-f 32768 Minimum fragment (subblock) size in bytes
-i 4096 Inode size in bytes
-I 32768 Indirect block size in bytes
-m 2 Default number of metadata replicas
-M 2 Maximum number of metadata replicas
-r 1 Default number of data replicas
-R 2 Maximum number of data replicas
-j scatter Block allocation type
-V 23.00 (5.0.5.0) Current file system version
15.01 (4.2.0.0) Original file system version
-L 33554432 Logfile size
--subblocks-per-full-block 32 Number of subblocks per full block
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20210528/ec0c5063/attachment.htm>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 5254 bytes
Desc: not available
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20210528/ec0c5063/attachment.bin>
More information about the gpfsug-discuss
mailing list