[gpfsug-discuss] Mounting GPFS data on OpenStack VM

Gaurang Tapase gaurang.tapase at in.ibm.com
Fri Jan 20 18:04:45 GMT 2017


Hi Brian,

For option #3, you can use GPFS Manila (OpenStack shared file system 
service) driver for exporting data from protocol servers to the OpenStack 
VMs.
It was updated to support CES in the Newton release.

A new feature of bringing existing filesets under Manila management has 
also been added recently.

Thanks,
Gaurang
------------------------------------------------------------------------
Gaurang S Tapase
Spectrum Scale & OpenStack
IBM India Storage Lab, Pune (India)
Email : gaurang.tapase at in.ibm.com
Phone : +91-20-42025699 (W), +91-9860082042(Cell)
-------------------------------------------------------------------------



From:   Brian Marshall <mimarsh2 at vt.edu>
To:     gpfsug main discussion list <gpfsug-discuss at spectrumscale.org>
Date:   01/18/2017 09:52 PM
Subject:        Re: [gpfsug-discuss] Mounting GPFS data on OpenStack VM
Sent by:        gpfsug-discuss-bounces at spectrumscale.org



To answer some more questions:

What sort of workload will your Nova VM's be running?
This is largely TBD but we anticipate webapps and other non-batch ways of 
interacting with and post processing data that has been computed on HPC 
batch systems.  For example a user might host a website that allows users 
to view pieces of a large data set and do some processing in private cloud 
or kick off larger jobs on HPC clusters

How many VM's are you running?
This work is still in the design / build phase.  We have 48 servers slated 
for the project.  At max maybe 500 VMs; again this is a pretty wild 
estimate.  This is a new service we are looking to provide

What is your Network interconnect between the Scale Storage cluster and 
the Nova Compute cluster
Each nova node has a dual 10gigE connection to switches that uplink to our 
core 40 gigE switches were NSD Servers are directly connectly.

The information so far has been awesome.  Thanks everyone.  I am 
definitely leaning towards option #3 of creating protocol servers.  Are 
there any design/build white papers targetting the virutalization use 
case?

Thanks,
Brian

On Tue, Jan 17, 2017 at 5:55 PM, Andrew Beattie <abeattie at au1.ibm.com> 
wrote:
HI Brian,
 
 
Couple of questions for you:
 
What sort of workload will your Nova VM's be running?
How many VM's are you running?
What is your Network interconnect between the Scale Storage cluster and 
the Nova Compute cluster
 
I have cc'd Jake Carrol from University of Queensland in on the email as I 
know they have done some basic performance testing using Scale to provide 
storage to Openstack.
One of the issues that they found was the Openstack network translation 
was a performance limiting factor.
 
I think from memory the best performance scenario they had was, when they 
installed the scale client locally into the virtual machines
 
 
Andrew Beattie
Software Defined Storage  - IT Specialist
Phone: 614-2133-7927
E-mail: abeattie at au1.ibm.com
 
 
----- Original message -----
From: Brian Marshall <mimarsh2 at vt.edu>
Sent by: gpfsug-discuss-bounces at spectrumscale.org
To: gpfsug main discussion list <gpfsug-discuss at spectrumscale.org>
Cc:
Subject: [gpfsug-discuss] Mounting GPFS data on OpenStack VM
Date: Wed, Jan 18, 2017 7:51 AM
  
UG, 
 
I have a GPFS filesystem.
 
I have a OpenStack private cloud.
 
What is the best way for Nova Compute VMs to have access to data inside 
the GPFS filesystem?
 
1)Should VMs mount GPFS directly with a GPFS client?
2) Should the hypervisor mount GPFS and share to nova computes?
3) Should I create GPFS protocol servers that allow nova computes to mount 
of NFS?
 
All advice is welcome.
 
 
Best,
Brian Marshall
Virginia Tech
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss
 

_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at spectrumscale.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss



-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20170120/30f7ce1b/attachment-0002.htm>


More information about the gpfsug-discuss mailing list