[gpfsug-discuss] Unbalanced pdisk free space

Sandeep Naik1 sannaik2 at in.ibm.com
Tue Feb 12 19:50:52 GMT 2019


Hi Alvise,

Here is response to your question in blue.

Q - Can anybody tell me if it is normal that all the pdisks of both my 
recovery groups, residing on the same physical enclosure have free space 
equal to (more or less) 1/3 of the free space of the pdisks residing on 
the other physical enclosure (see attached text files for the command line 
output) ?

Yes it is normal to see variation in free space between pdisks. The 
variation should be seen in the context of used space and not free space. 
GNR try to balance space equally across enclosures (failure groups). One 
enclosure has one SSD (per RG) so it has 41 disk in DA1 while the other 
one has 42. Enclosure with 42 disk show 360 GiB free space while one with 
41 disk show 120 GiB. If you look at used capacity and distribute it 
equally between two enclosures you will notice that used capacity is 
almost same between two enclosure.

42 * (10240 - 360) ≃ 41 * (10240 - 120)
 
I guess when the least free disks are fully occupied (while the others are 
still partially free) write performance will drop by a factor of two. 
Correct ? 
Is there a way (considering that the system is in production) to fix 
(rebalance) this free space among all pdisk of both enclosures ?

You should see in context of size of pdisk, which in your case in 10TB. 
The disk showing 120GB free is 98% full while the one showing 360GB free 
is 96% full. This free space is available for creating vdisks and should 
not be confused with free space available in filesystem. Your pdisk are by 
and large equally filled so there will be no impact on write performance 
because of small variation in free space. 

Hope this helps

Thanks,

Sandeep Naik
Elastic Storage server / GPFS Test 
ETZ-B, Hinjewadi Pune India
(+91) 8600994314



From:   "Dorigo Alvise (PSI)" <alvise.dorigo at psi.ch>
To:     gpfsug main discussion list <gpfsug-discuss at spectrumscale.org>
Date:   31/01/2019 04:07 PM
Subject:        Re: [gpfsug-discuss] Unbalanced pdisk free space
Sent by:        gpfsug-discuss-bounces at spectrumscale.org



They're attached.

Thanks!

   Alvise


From: gpfsug-discuss-bounces at spectrumscale.org 
[gpfsug-discuss-bounces at spectrumscale.org] on behalf of IBM Spectrum Scale 
[scale at us.ibm.com]
Sent: Wednesday, January 30, 2019 9:25 PM
To: gpfsug main discussion list
Subject: Re: [gpfsug-discuss] Unbalanced pdisk free space

Alvise,

Could you send us the output of the following commands from both server 
nodes. 
mmfsadm dump nspdclient > /tmp/dump_nspdclient.<nodeName> 
mmfsadm dump pdisk   > /tmp/dump_pdisk.<nodeName>
Regards, The Spectrum Scale (GPFS) team

------------------------------------------------------------------------------------------------------------------
If you feel that your question can benefit other users of  Spectrum Scale 
(GPFS), then please post it to the public IBM developerWroks Forum at 
https://www.ibm.com/developerworks/community/forums/html/forum?id=11111111-0000-0000-0000-000000000479
. 

If your query concerns a potential software error in Spectrum Scale (GPFS) 
and you have an IBM software maintenance contract please contact 
1-800-237-5511 in the United States or your local IBM Service Center in 
other countries. 

The forum is informally monitored as time permits and should not be used 
for priority messages to the Spectrum Scale (GPFS) team.



From:        "Dorigo Alvise (PSI)" <alvise.dorigo at psi.ch>
To:        "gpfsug-discuss at spectrumscale.org" 
<gpfsug-discuss at spectrumscale.org>
Date:        01/30/2019 08:24 AM
Subject:        [gpfsug-discuss] Unbalanced pdisk free space
Sent by:        gpfsug-discuss-bounces at spectrumscale.org



Hello,
I've a Lenovo Spectrum Scale system DSS-G220 (software dss-g-2.0a) 
composed of
2x x3560 M5 IO server nodes
1x x3550 M5 client/support node
2x disk enclosures D3284
GPFS/GNR 4.2.3-7

Can anybody tell me if it is normal that all the pdisks of both my 
recovery groups, residing on the same physical enclosure have free space 
equal to (more or less) 1/3 of the free space of the pdisks residing on 
the other physical enclosure (see attached text files for the command line 
output) ?

I guess when the least free disks are fully occupied (while the others are 
still partially free) write performance will drop by a factor of two. 
Correct ? 
Is there a way (considering that the system is in production) to fix 
(rebalance) this free space among all pdisk of both enclosures ?

Should I open a PMR to IBM ?

Many thanks,

   Alvise

[attachment "rg1" deleted by Brian Herr/Poughkeepsie/IBM] [attachment 
"rg2" deleted by Brian Herr/Poughkeepsie/IBM] 
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss


[attachment "dump_nspdclient.sf-dssio-1" deleted by Sandeep 
Naik1/India/IBM] [attachment "dump_nspdclient.sf-dssio-2" deleted by 
Sandeep Naik1/India/IBM] [attachment "dump_pdisk.sf-dssio-1" deleted by 
Sandeep Naik1/India/IBM] [attachment "dump_pdisk.sf-dssio-2" deleted by 
Sandeep Naik1/India/IBM] _______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
https://urldefense.proofpoint.com/v2/url?u=http-3A__gpfsug.org_mailman_listinfo_gpfsug-2Ddiscuss&d=DwICAg&c=jf_iaSHvJObTbx-siA1ZOg&r=DXkezTwrVXsEOfvoqY7_DLS86P5FtQszjm9zok6upRU&m=rrqeq4UVHOFW9aaiAj-N7Lu6Z7UKBo4-0e3yINS47W0&s=n2t4qaUh-0mamutSSx0E-5j09DbZImKsbDoiM0enBcg&e=





-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20190213/61f4037c/attachment-0001.htm>


More information about the gpfsug-discuss mailing list