[gpfsug-discuss] dssgmkfs.mmvdisk number of NSD's
Simon Thompson
S.J.Thompson at bham.ac.uk
Mon Mar 1 09:06:07 GMT 2021
Or for hedging your bets about how you might want to use it in future.
We are never quite sure if we want to do something different in the future with some of the storage, sure that might mean we want to steal some space from a file-system, but that is perfectly valid. And we have done this, both in temporary transient states (data migration between systems), or permanently (found we needed something on a separate file-system)
So yes whilst there might be no performance impact on doing this, we still do.
I vaguely recall some of the old reasoning was around IO queues in the OS, i.e. if you had 6 NSDs vs 16 NSDs attached to the NSD server, you have 16 IO queues passing to multipath, which can help keep the data pipes full. I suspect there was some optimal number of NSDs for different storage controllers, but I don't know if anyone ever benchmarked that.
Simon
On 01/03/2021, 08:16, "gpfsug-discuss-bounces at spectrumscale.org on behalf of Achim.Rehor at de.ibm.com" <gpfsug-discuss-bounces at spectrumscale.org on behalf of Achim.Rehor at de.ibm.com> wrote:
The reason for having multiple NSDs in legacy NSD (non-GNR) handling is
the increased parallelism, that gives you 'more spindles' and thus more
performance.
In GNR the drives are used in parallel anyway through the GNRstriping.
Therfore, you are using all drives of a ESS/GSS/DSS model under the hood
in the vdisks anyway.
The only reason for having more NSDs is for using them for different
filesystems.
Mit freundlichen Grüßen / Kind regards
Achim Rehor
IBM EMEA ESS/Spectrum Scale Support
gpfsug-discuss-bounces at spectrumscale.org wrote on 01/03/2021 08:58:43:
> From: Jonathan Buzzard <jonathan.buzzard at strath.ac.uk>
> To: gpfsug-discuss at spectrumscale.org
> Date: 01/03/2021 08:58
> Subject: [EXTERNAL] Re: [gpfsug-discuss] dssgmkfs.mmvdisk number of
NSD's
> Sent by: gpfsug-discuss-bounces at spectrumscale.org
>
> On 28/02/2021 09:31, Jan-Frode Myklebust wrote:
> >
> > I?ve tried benchmarking many vs. few vdisks per RG, and never could
see
> > any performance difference.
>
> That's encouraging.
>
> >
> > Usually we create 1 vdisk per enclosure per RG, thinking this will
> > allow us to grow with same size vdisks when adding additional
enclosures
> > in the future.
> >
> > Don?t think mmvdisk can be told to create multiple vdisks per RG
> > directly, so you have to manually create multiple vdisk sets each with
> > the apropriate size.
> >
>
> Thing is back in the day so GPFS v2.x/v3.x there where strict warnings
> that you needed a minimum of six NSD's for optimal performance. I have
> sat in presentations where IBM employees have said so. What we where
> told back then is that GPFS needs a minimum number of NSD's in order to
> be able to spread the I/O's out. So if an NSD is being pounded for reads
> and a write comes in it. can direct it to a less busy NSD.
>
> Now I can imagine that in a ESS/DSS-G that as it's being scattered to
> the winds under the hood this is no longer relevant. But some notes to
> the effect for us old timers would be nice if that is the case to put
> our minds to rest.
>
>
> JAB.
>
> --
> Jonathan A. Buzzard Tel: +44141-5483420
> HPC System Administrator, ARCHIE-WeSt.
> University of Strathclyde, John Anderson Building, Glasgow. G4 0NG
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at spectrumscale.org
> https://urldefense.proofpoint.com/v2/url?
>
u=http-3A__gpfsug.org_mailman_listinfo_gpfsug-2Ddiscuss&d=DwIGaQ&c=jf_iaSHvJObTbx-
> siA1ZOg&r=RGTETs2tk0Kz_VOpznDVDkqChhnfLapOTkxLvgmR2-
> M&m=Mr4A8ROO2t7qFYTfTRM_LoPLllETw72h51FK07dye7Q&s=z6yRHIKsH-
> IaOjtto4ZyUjFFe0vTGhqzYUiM23rEShg&e=
>
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
More information about the gpfsug-discuss
mailing list