[gpfsug-discuss] add local nsd back to cluster?

shao feng shaof777 at gmail.com
Fri Jul 29 16:54:24 BST 2022


Thanks Olaf

I've setup the mmsdr backup as
https://www.ibm.com/docs/en/spectrum-scale/5.1.2?topic=exits-mmsdrbackup-user-exit,
since my cluster is CCR enabled, it generate a CCR backup file,
but when trying to restore from this file, it require quorum nodes to
shutdown? Is it possible to restore without touching quorum nodes?

[root at tofail ~]# mmsdrrestore -F
CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz
Restoring a CCR backup archive is a cluster-wide operation.
The -a flag is required.
mmsdrrestore: Command failed. Examine previous error messages to determine
cause.

[root at tofail ~]# mmsdrrestore -F
CCRBackup.986.2022.07.29.23.06.19.myquorum.tar.gz -a
Restoring CCR backup
Verifying that GPFS is inactive on quorum nodes
mmsdrrestore: GPFS is still active on myquorum
mmsdrrestore: Unexpected error from mmsdrrestore: CCR restore failed.
Return code: 192
mmsdrrestore: Command failed. Examine previous error messages to determine
cause.


On Thu, Jul 28, 2022 at 3:14 PM Olaf Weiser <olaf.weiser at de.ibm.com> wrote:

>
>
> Hi -
> assuming, you'll run it  withou ECE  ?!? ... just with replication on the
> file system level
> ba aware, every time a node goes offline, you 'll have to restart the
> disks in your filesystem .. This causes a complete scan of the meta data to
> detect files with missing updates / replication
>
>
> apart from that to your Q :
> you may consider to backup mmsdr
> additionally, take a look to   mmsdrrestore, in case you want to restore a
> nodes's SDR configuration
>
> quick and dirty..  save the content  of  /var/mmfs  may also help you
>
> during the node is "gone".. of course.. the disk is down , after restore
> of SDR / node's config .. it should be able to start ..
> the rest runs as usual
>
>
>
> ------------------------------
> *Von:* gpfsug-discuss <gpfsug-discuss-bounces at gpfsug.org> im Auftrag von
> shao feng <shaof777 at gmail.com>
> *Gesendet:* Donnerstag, 28. Juli 2022 09:02
> *An:* gpfsug main discussion list <gpfsug-discuss at gpfsug.org>
> *Betreff:* [EXTERNAL] [gpfsug-discuss] add local nsd back to cluster?
>
> Hi all, I am planning to implement  a cluster with a bunch of old x86
> machines, the disks are not connected to nodes via the SAN network, instead
> each x86 machine has some local attached disks. The question is regarding
> node failure, for example
> ZjQcmQRYFpfptBannerStart
> This Message Is From an External Sender
> This message came from outside your organization.
>
> ZjQcmQRYFpfptBannerEnd
> Hi all,
>
> I am planning to implement  a cluster with a bunch of old x86 machines,
> the disks are not connected to nodes via the SAN network, instead each x86
> machine has some local attached disks.
> The question is regarding node failure, for example only the operating
> system disk fails and the nsd disks are good. In that case I plan to
> replace the failing OS disk with a new one and install the OS on it and
> re-attach these nsd disks to that node, my question is: will this work? how
> can I add a nsd back to the cluster without restoring data from other
> replicas since the data/metadata is actually not corrupted on nsd.
>
> Best regards,
> _______________________________________________
> gpfsug-discuss mailing list
> gpfsug-discuss at gpfsug.org
> http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20220729/1fc0e167/attachment-0002.htm>


More information about the gpfsug-discuss mailing list