[gpfsug-discuss] Protocol node recommendations

Sobey, Richard A r.sobey at imperial.ac.uk
Tue Apr 25 09:19:10 BST 2017


I tried it on one node but investing in what could be up to £5000 in SSDs when we don't know the gains isn't something I can argue. Not that LROC will hurt the environment but my users may not see any benefit. My cluster is the complete opposite of busy (relative to people saying they're seeing sustained 800MB/sec throughput), I just need it stable.

Richard
From: Frank Tower [mailto:frank.tower at outlook.com]
Sent: 24 April 2017 22:12
To: Sobey, Richard A <r.sobey at imperial.ac.uk>; gpfsug main discussion list <gpfsug-discuss at spectrumscale.org>; Jan-Frode Myklebust <janfrode at tanso.net>
Subject: Re: [gpfsug-discuss] Protocol node recommendations



>From what I've read from the Wiki:

'The NFS protocol performance is largely dependent on the base system performance of the protocol node hardware and network. This includes multiple factors the type and number of CPUs, the size of the main memory in the nodes, the type of disk drives used (HDD, SSD, etc.) and the disk configuration (RAID-level, replication etc.). In addition, NFS protocol performance can be impacted by the overall load of the node (such as number of clients accessing, snapshot creation/deletion and more) and administrative tasks (for example filesystem checks or online re-striping of disk arrays).'


Nowadays, SSD is worst to invest.
LROC could be an option in the future, but we need to quantify NFS/CIFS workload first.

Are you using LROC with your GPFS installation ?

Best,
Frank.
________________________________
From: Sobey, Richard A <r.sobey at imperial.ac.uk<mailto:r.sobey at imperial.ac.uk>>
Sent: Monday, April 24, 2017 11:11 AM
To: gpfsug main discussion list; Jan-Frode Myklebust
Subject: Re: [gpfsug-discuss] Protocol node recommendations

What's your SSD going to help with... will you implement it as a LROC device? Otherwise I can't see the benefit to using it to boot off.

From: gpfsug-discuss-bounces at spectrumscale.org<mailto:gpfsug-discuss-bounces at spectrumscale.org> [mailto:gpfsug-discuss-bounces at spectrumscale.org] On Behalf Of Frank Tower
Sent: 23 April 2017 22:28
To: Jan-Frode Myklebust <janfrode at tanso.net<mailto:janfrode at tanso.net>>; gpfsug-discuss at spectrumscale.org<mailto:gpfsug-discuss at spectrumscale.org>
Subject: Re: [gpfsug-discuss] Protocol node recommendations


Hi,



Nice ! didn't pay attention at the revision and the spreadsheet. If someone still have a copy somewhere it could be useful, Google didn't help :(



We will follow your advise and start with 3 protocol nodes equipped with 128GB memory, 2 x 12 cores  (maybe E5-2680 or E5-2670).



>From what I read, NFS-Ganesha mainly depend of the hardware, Linux on a SSD should be a big plus in our case.



Best,

Frank



________________________________
From: Jan-Frode Myklebust <janfrode at tanso.net<mailto:janfrode at tanso.net>>
Sent: Sunday, April 23, 2017 12:07:38 PM
To: Frank Tower; gpfsug-discuss at spectrumscale.org<mailto:gpfsug-discuss at spectrumscale.org>
Subject: Re: [gpfsug-discuss] Protocol node recommendations

The protocol sizing tool should be available from https://www.ibm.com/developerworks/community/wikis/home?lang=en#!/wiki/General%20Parallel%20File%20System%20(GPFS)/page/Sizing%20Guidance%20for%20Protocol%20Node/version/70a4c7c0-a5c6-4dde-b391-8f91c542dd7d , but I'm getting 404 now.

I think 128GB should be enough for both protocols on same nodes, and I think your 3 node suggestion is best. Better load sharing with not dedicating subset of nodes to each protocol.



-jf
lør. 22. apr. 2017 kl. 21.22 skrev Frank Tower <frank.tower at outlook.com<mailto:frank.tower at outlook.com>>:

Hi,



Thank for the recommendations.

Now we deal with the situation of:



- take 3 nodes with round robin DNS that handle both protocols

- take 4 nodes, split CIFS and NFS, still use round robin DNS for CIFS and NFS services.



Regarding your recommendations, 256GB memory node could be a plus if we mix both protocols for such case.



Is the spreadsheet publicly available or do we need to ask IBM ?



Thank for your help,

Frank.

________________________________
From: Jan-Frode Myklebust <janfrode at tanso.net<mailto:janfrode at tanso.net>>
Sent: Saturday, April 22, 2017 10:50 AM
To: gpfsug-discuss at spectrumscale.org<mailto:gpfsug-discuss at spectrumscale.org>
Subject: Re: [gpfsug-discuss] Protocol node recommendations

That's a tiny maxFilesToCache...

I would start by implementing the settings from /usr/lpp/mmfs/*/gpfsprotocolldefaul* plus a 64GB pagepool for your protocoll nodes, and leave further tuning to when you see you have issues.

Regarding sizing, we have a spreadsheet somewhere where you can input some workload parameters and get an idea for how many nodes you'll need. Your node config seems fine, but one node seems too few to serve 1000+ users. We support max 3000 SMB connections/node, and I believe the recommendation is 4000 NFS connections/node.


-jf
lør. 22. apr. 2017 kl. 08.34 skrev Frank Tower <frank.tower at outlook.com<mailto:frank.tower at outlook.com>>:
Hi,

We have here around 2PB GPFS (4.2.2) accessed through an HPC cluster with GPFS client on each node.

We will have to open GPFS to all our users over CIFS and kerberized NFS with ACL support for both protocol for around +1000 users

All users have different use case and needs:
- some will do random I/O through a large set of opened files (~5k files)
- some will do large write with 500GB-1TB files
- other will arrange sequential I/O with ~10k opened files

NFS and CIFS will share the same server, so I through to use SSD drive, at least 128GB memory with 2 sockets.

Regarding tuning parameters, I thought at:

maxFilesToCache 10000
syncIntervalStrict yes
workerThreads (8*core)
prefetchPct 40 (for now and update if needed)

I read the wiki 'Sizing Guidance for Protocol Node', but I was wondering if someone could share his experience/best practice regarding hardware sizing and/or tuning parameters.

Thank by advance,
Frank
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org<http://spectrumscale.org>
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20170425/c33b5b9c/attachment-0002.htm>


More information about the gpfsug-discuss mailing list