From ISHIMOTO at jp.ibm.com Mon Apr 4 13:13:28 2022 From: ISHIMOTO at jp.ibm.com (Takeshi Ishimoto) Date: Mon, 4 Apr 2022 12:13:28 +0000 Subject: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file In-Reply-To: <325BFDF6-2421-480E-8EE6-4EB8AE15E56F@us.ibm.com> References: <325BFDF6-2421-480E-8EE6-4EB8AE15E56F@us.ibm.com> Message-ID: Unfortunately, there is no mailing list available for Spectrum Archive, and the individuals Lyle listed are not the right contact. The closest option will be the IBM Tape Storage community at https://community.ibm.com/community/user/storage/communities/community-home/digestviewer?communitykey=85531a8a-8971-4c0e-8d2b-098ba927269e or one should open a support ticket for question. Shao, There is no notification mechanism at the completion of transparent recall nor other operation, but you should be able to monitor the recent operations by ?eeadm task list? periodically and find the detail of task with ?eeadm task show ?. Takeshi ISHIMOTO (?????) Spectrum Archive/LTFS Development - Product Architect and DCT Lead SNIA Linear Tape File System (LTFS) TWG chair IBM Systems ishimoto at jp.ibm.com From: Lyle Gayne > Date: Wednesday, March 30, 2022 at 7:10 AM To: gpfsug main discussion list >, Mark Hill >, Khanh V Ngo > Subject: Re: [EXTERNAL] Re: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file I guess this should start with Mark Hill (and Khanh), from Spectrum Archive team. Lyle ________________________________ From: gpfsug-discuss > on behalf of scale at us.ibm.com > Sent: Wednesday, March 30, 2022 10:02 AM To: gpfsug main discussion list > Cc: gpfsug-discuss > Subject: [EXTERNAL] Re: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file I think this can be answered by the IBM Spectrum Archive team and should go to a mailing list or discussion channel where they are involved. @Lyle: Do you know anyone who can answer this from Scale side. Regards, The Spectrum Scale (GPFS) team ------------------------------------------------------------------------------------------------------------------ If you feel that your question can benefit other users of Spectrum Scale (GPFS), then please post it to the public IBM developerWroks Forum at https://www.ibm.com/developerworks/community/forums/html/forum?id=11111111-0000-0000-0000-000000000479. If your query concerns a potential software error in Spectrum Scale (GPFS) and you have an IBM software maintenance contract please contact 1-800-237-5511 in the United States or your local IBM Service Center in other countries. The forum is informally monitored as time permits and should not be used for priority messages to the Spectrum Scale (GPFS) team. [Inactive hide details for "shao feng" ---29-03-2022 11.13.11 AM---Hello, Not sure if this is the right place to ask about Spect]"shao feng" ---29-03-2022 11.13.11 AM---Hello, Not sure if this is the right place to ask about Spectrum Archival EE. From: "shao feng" > To: "gpfsug main discussion list" > Date: 29-03-2022 11.13 AM Subject: [EXTERNAL] [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file Sent by: "gpfsug-discuss" > ________________________________ Hello, Not sure if this is the right place to ask about Spectrum Archival EE. In a GPFS+LTFS EE configuration, is there any mechanism to get notification from Spectrum Archival EE when it finishes a migrating/recall for a file, especially a transparent recall? Thank you!_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: image001.gif Type: image/gif Size: 106 bytes Desc: image001.gif URL: From info at odina.nl Mon Apr 4 13:39:15 2022 From: info at odina.nl (Jaap Jan Ouwehand) Date: Mon, 4 Apr 2022 14:39:15 +0200 Subject: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file In-Reply-To: References: <325BFDF6-2421-480E-8EE6-4EB8AE15E56F@us.ibm.com> Message-ID: <4ec4a1cc-7792-7d0e-30cf-a70ebf58aa59@odina.nl> Hello, In addition to the eeadm command, the REST API might also be a solution for job monitoring. https://www.ibm.com/docs/en/spectrum-archive-ee/1.3.0.0?topic=information-rest-api-task -- Kind regards Jaap Jan Ouwehand info at odina.nl Linux & Storage Solution expert Op 4-4-2022 om 14:13 schreef Takeshi Ishimoto: > > Unfortunately, there is no mailing list available for Spectrum > Archive, and the individuals Lyle listed are not the right contact. > > The closest option will be the IBM Tape Storage community at > https://community.ibm.com/community/user/storage/communities/community-home/digestviewer?communitykey=85531a8a-8971-4c0e-8d2b-098ba927269e > or > one should open a support ticket for question. > > Shao, > > There is no notification mechanism at the completion of transparent > recall nor other operation, but you should be able to monitor the > recent operations by ?eeadm task list? periodically and find the > detail of task with ?eeadm task show ?. > > Takeshi ISHIMOTO (?????) > Spectrum Archive/LTFS Development - Product Architect and DCT Lead > > SNIA Linear Tape File System (LTFS) TWG chair > IBM Systems > ishimoto at jp.ibm.com > > *From: *Lyle Gayne > > *Date: *Wednesday, March 30, 2022 at 7:10 AM > *To: *gpfsug main discussion list >, Mark Hill >, Khanh V Ngo > > *Subject: *Re: [EXTERNAL] Re: [gpfsug-discuss] notification when > Spectrum Archival EE finish migrating/recall for a file > > I guess this should start with Mark Hill (and Khanh), from Spectrum > Archive team. > > Lyle > > ------------------------------------------------------------------------ > > *From:*gpfsug-discuss on behalf of > scale at us.ibm.com > *Sent:* Wednesday, March 30, 2022 10:02 AM > *To:* gpfsug main discussion list > *Cc:* gpfsug-discuss > *Subject:* [EXTERNAL] Re: [gpfsug-discuss] notification when Spectrum > Archival EE finish migrating/recall for a file > > I think this can be answered by the IBM Spectrum Archive team and > should go to a mailing list or discussion channel where they are involved. > > @Lyle: Do you know anyone who can answer this from Scale side. > > > Regards, The Spectrum Scale (GPFS) team > > ------------------------------------------------------------------------------------------------------------------ > If you feel that your question can benefit other users of ?Spectrum > Scale (GPFS), then please post it to the public IBM developerWroks > Forum at > https://www.ibm.com/developerworks/community/forums/html/forum?id=11111111-0000-0000-0000-000000000479 > . > > > If your query concerns a potential software error in Spectrum Scale > (GPFS) and you have an IBM software maintenance contract please > contact ?1-800-237-5511 in the United States or your local IBM Service > Center in other countries. > > The forum is informally monitored as time permits and should not be > used for priority messages to the Spectrum Scale (GPFS) team. > > Inactive hide details for "shao feng" ---29-03-2022 11.13.11 > AM---Hello, Not sure if this is the right place to ask about > Spect"shao feng" ---29-03-2022 11.13.11 AM---Hello, Not sure if this > is the right place to ask about Spectrum Archival EE. > > From: "shao feng" > > To: "gpfsug main discussion list" > > Date: 29-03-2022 11.13 AM > Subject: [EXTERNAL] [gpfsug-discuss] notification when Spectrum > Archival EE finish migrating/recall for a file > Sent by: "gpfsug-discuss" > > > ------------------------------------------------------------------------ > > > > > Hello, > > Not sure if this is the right place to ask about Spectrum Archival EE. > In a GPFS+LTFS EE configuration, is there any mechanism to get > notification from Spectrum Archival EE when it finishes a > migrating/recall for a file, especially a transparent recall? > > Thank you!_______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org > > > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: image001.gif Type: image/gif Size: 106 bytes Desc: not available URL: From lgayne at us.ibm.com Tue Apr 19 13:33:00 2022 From: lgayne at us.ibm.com (Lyle Gayne) Date: Tue, 19 Apr 2022 12:33:00 +0000 Subject: [gpfsug-discuss] Mailing List Test In-Reply-To: <57a05639a9c416f4d6e44c09b79b9fbb@qsplace.co.uk> References: <57a05639a9c416f4d6e44c09b79b9fbb@qsplace.co.uk> Message-ID: Test of User Group to see if mailing list is still working... Apologies for this spam. Lyle ________________________________ From: gpfsug-discuss on behalf of laurence at qsplace.co.uk Sent: Sunday, March 13, 2022 6:43 PM To: gpfsug-discuss at gpfsug.org Subject: [EXTERNAL] [gpfsug-discuss] Mailing List Test Mailing List Test -------------- next part -------------- An HTML attachment was scrubbed... URL: From chair at gpfsug.org Thu Apr 21 08:10:11 2022 From: chair at gpfsug.org (chair at gpfsug.org) Date: Thu, 21 Apr 2022 08:10:11 +0100 Subject: [gpfsug-discuss] UK Spectrum Scale User Group meeting 30th June 2022 Message-ID: Hi all, Just a reminder that the next UK User Group meeting will be taking place in London (IBM York Road) on 30th June 2022. Registration is open at https://www.eventbrite.co.uk/e/spectrum-scale-user-group-registration-321290978967 Now is your opportunity to help shape the agenda, please feel free to send me ideas for talks we could ask IBM for (I don't promise we'll get them on the agenda mind!). Also if you would like to do a user talk, please get in touch with me and let me know. It could be a large scale deployment or even just a couple of nodes, it's your opportunity to showcase how you use Spectrum Scale and what for. Every year people tell us how valuable they find the user talks, but this needs YOU, so please do think about if you are able to offer a talk! As in the past we are looking for sponsorship to enable us to run an evening networking event. I've sent out details to those who have sponsored in the past and to those who have asked us directly about sponsorship opportunities. If you either haven't received this or are interested in becoming a sponsor, please email me directly. Thanks Paul From juergen.hannappel at desy.de Mon Apr 25 09:59:18 2022 From: juergen.hannappel at desy.de (Hannappel, Juergen) Date: Mon, 25 Apr 2022 10:59:18 +0200 (CEST) Subject: [gpfsug-discuss] 5.1.3.1 changes? Message-ID: <211580513.43933888.1650877158353.JavaMail.zimbra@desy.de> Hi, last Friday I got a notification that 5.1.3.1 is out, but https://public.dhe.ibm.com/storage/spectrumscale/spectrum_scale_apars_513x.html claims that it fixes no problems? -- Dr. J?rgen Hannappel DESY/IT Tel. : +49 40 8998-4616 From Renar.Grunenberg at huk-coburg.de Mon Apr 25 16:53:02 2022 From: Renar.Grunenberg at huk-coburg.de (Grunenberg, Renar) Date: Mon, 25 Apr 2022 15:53:02 +0000 Subject: [gpfsug-discuss] 5.1.3.1 changes? In-Reply-To: <211580513.43933888.1650877158353.JavaMail.zimbra@desy.de> References: <211580513.43933888.1650877158353.JavaMail.zimbra@desy.de> Message-ID: <1619424d058a489f96ca1baa63e3bff3@huk-coburg.de> Hallo J?rgen, i had the same problem, but since 1 hour the list is available. Renar Grunenberg Abteilung Informatik - Betrieb HUK-COBURG Bahnhofsplatz 96444 Coburg Telefon: 09561 96-44110 Telefax: 09561 96-44104 E-Mail: Renar.Grunenberg at huk-coburg.de Internet: www.huk.de ======================================================================= HUK-COBURG Haftpflicht-Unterst?tzungs-Kasse kraftfahrender Beamter Deutschlands a. G. in Coburg Reg.-Gericht Coburg HRB 100; St.-Nr. 9212/101/00021 Sitz der Gesellschaft: Bahnhofsplatz, 96444 Coburg Vorsitzender des Aufsichtsrats: Prof. Dr. Heinrich R. Schradin. Vorstand: Klaus-J?rgen Heitmann (Sprecher), Stefan Gronbach, Dr. Hans Olav Her?y, Dr. Helen Reck, Dr. J?rg Rheinl?nder, Thomas Sehn, Daniel Thomas. ======================================================================= Diese Nachricht enth?lt vertrauliche und/oder rechtlich gesch?tzte Informationen. Wenn Sie nicht der richtige Adressat sind oder diese Nachricht irrt?mlich erhalten haben, informieren Sie bitte sofort den Absender und vernichten Sie diese Nachricht. Das unerlaubte Kopieren sowie die unbefugte Weitergabe dieser Nachricht ist nicht gestattet. This information may contain confidential and/or privileged information. If you are not the intended recipient (or have received this information in error) please notify the sender immediately and destroy this information. Any unauthorized copying, disclosure or distribution of the material in this information is strictly forbidden. ======================================================================= -----Urspr?ngliche Nachricht----- Von: gpfsug-discuss Im Auftrag von Hannappel, Juergen Gesendet: Montag, 25. April 2022 10:59 An: gpfsug-discuss Betreff: [gpfsug-discuss] 5.1.3.1 changes? Hi, last Friday I got a notification that 5.1.3.1 is out, but https://public.dhe.ibm.com/storage/spectrumscale/spectrum_scale_apars_513x.html claims that it fixes no problems? -- Dr. J?rgen Hannappel DESY/IT Tel. : +49 40 8998-4616 _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org From juergen.hannappel at desy.de Wed Apr 27 14:19:02 2022 From: juergen.hannappel at desy.de (Hannappel, Juergen) Date: Wed, 27 Apr 2022 15:19:02 +0200 (CEST) Subject: [gpfsug-discuss] Updating a medium size cluster efficiently Message-ID: <733714875.45755494.1651065542428.JavaMail.zimbra@desy.de> Hi, we have a medium size gpfs client cluster (a few hundred nodes) and want to update the gpfs version in an efficient way in a rolling update, i.e.. update each node when it can be rebooted. Doing so via a slurm script when the node is drained jut before the reboot work only most of the time because in some cases even when the node is drained the file systems are still buly and can't be unmounted, so the update fails. Therefore I tried to trigger the update on the reboot, before gpfs starts. To do so I added a systemd service that is sheduled before the gpfs.service, which does a yum update (we run CentOs 7.9) but: In the postinstall script of gpfs.base the gpfs.service is disabled and re-enabled via systemctl, and systemd apparently get's that wrong, so that if the update really happens it afterwards will not start the gpfs.service. Does any one have a clever way how to do a rolling update that really works without maunually hunting after some per cent of machines that don't manage it on the first go? -- Dr. J?rgen Hannappel DESY/IT Tel. : +49 40 8998-4616 From jonathan.buzzard at strath.ac.uk Wed Apr 27 21:21:07 2022 From: jonathan.buzzard at strath.ac.uk (Jonathan Buzzard) Date: Wed, 27 Apr 2022 21:21:07 +0100 Subject: [gpfsug-discuss] Updating a medium size cluster efficiently In-Reply-To: <733714875.45755494.1651065542428.JavaMail.zimbra@desy.de> References: <733714875.45755494.1651065542428.JavaMail.zimbra@desy.de> Message-ID: On 27/04/2022 14:19, Hannappel, Juergen wrote: > > Hi, > we have a medium size gpfs client cluster (a few hundred nodes) > and want to update the gpfs version in an efficient way in a > rolling update, i.e.. update each node when it can be rebooted. > > Doing so via a slurm script when the node is drained jut before > the reboot work only most of the time because in some cases even > when the node is drained the file systems are still buly and can't be unmounted, > so the update fails. > I assume this is because you have dead jobs? The general trick is to submit a job as a special user that has sudo privileges, that runs as the next job on every node. That way you don't need to wait for the node to drain. Last "user" job on the node finishes and then the "special" job runs. It does it magic and reboots the node. Winner winner chicken dinner. > Therefore I tried to trigger the update on the reboot, before gpfs starts. > To do so I added a systemd service that is sheduled before the gpfs.service, > which does a yum update (we run CentOs 7.9) but: > > In the postinstall script of gpfs.base the gpfs.service is disabled and re-enabled > via systemctl, and systemd apparently get's that wrong, so that if > the update really happens it afterwards will not start the gpfs.service. > > Does any one have a clever way how to do a rolling update that really works > without maunually hunting after some per cent of machines that don't manage > it on the first go? > What you could do to find the nodes that don't work is have the upgrade script do an mmshutdown first before attempting the upgrade. Then check it actually managed to shutdown and if it didn't then send an email to an appropriate person saying there is an issue, before say putting the node in drain. The man page for mmshutdown says it has an exit code of zero on success, and none zero of failure so should be trivial to script. Being really clever I think you could then have the script submit a second copy of itself to the node that again will run as the next job and then reboot the node. That way when it comes back up it should be able to unmount GPFS and install the upgrade as the reboot will have cleared the issues that prevented the mmshutdown from working. You would obviously need to trial this out. If you are just looking to upgrade gpfs-gplbin and don't want to have it being recompiled on every node, then there is a trick with systemd. What you do is create /etc/systemd/system/gpfs.service.d/install-module.conf with the following contents [Service] ExecStartPre=-/usr/bin/yum --assumeyes install gpfs.gplbin-%v then everytime GPFS starts up it attempts to install the module for the currently running kernel (the special magic %v). This presumes you have a repository with the appropriate gpfs-gplbin RPM setup. Basically I take a node out build the RPM, test it is working and then deploy. I have a special RPM that installs the above local customization to the GPFS serivce unit file. JAB. -- Jonathan A. Buzzard Tel: +44141-5483420 HPC System Administrator, ARCHIE-WeSt. University of Strathclyde, John Anderson Building, Glasgow. G4 0NG From p.childs at qmul.ac.uk Thu Apr 28 15:42:38 2022 From: p.childs at qmul.ac.uk (Peter Childs) Date: Thu, 28 Apr 2022 14:42:38 +0000 Subject: [gpfsug-discuss] Job Opportunities In-Reply-To: References: Message-ID: We are recruiting for a couple of positions at Queen Mary University of London, within our Research group. Both these jobs will have some contact with Spectrum Scale and one of them is directly helping to support large Spectrum Scale cluster. If anyone is interested please apply. https://www.qmul.ac.uk/jobs/vacancies/items/6955.html https://www.qmul.ac.uk/jobs/vacancies/items/6949.html Peter Childs ITS Research Storage Queen Mary University of London. From ISHIMOTO at jp.ibm.com Mon Apr 4 13:13:28 2022 From: ISHIMOTO at jp.ibm.com (Takeshi Ishimoto) Date: Mon, 4 Apr 2022 12:13:28 +0000 Subject: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file In-Reply-To: <325BFDF6-2421-480E-8EE6-4EB8AE15E56F@us.ibm.com> References: <325BFDF6-2421-480E-8EE6-4EB8AE15E56F@us.ibm.com> Message-ID: Unfortunately, there is no mailing list available for Spectrum Archive, and the individuals Lyle listed are not the right contact. The closest option will be the IBM Tape Storage community at https://community.ibm.com/community/user/storage/communities/community-home/digestviewer?communitykey=85531a8a-8971-4c0e-8d2b-098ba927269e or one should open a support ticket for question. Shao, There is no notification mechanism at the completion of transparent recall nor other operation, but you should be able to monitor the recent operations by ?eeadm task list? periodically and find the detail of task with ?eeadm task show ?. Takeshi ISHIMOTO (?????) Spectrum Archive/LTFS Development - Product Architect and DCT Lead SNIA Linear Tape File System (LTFS) TWG chair IBM Systems ishimoto at jp.ibm.com From: Lyle Gayne > Date: Wednesday, March 30, 2022 at 7:10 AM To: gpfsug main discussion list >, Mark Hill >, Khanh V Ngo > Subject: Re: [EXTERNAL] Re: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file I guess this should start with Mark Hill (and Khanh), from Spectrum Archive team. Lyle ________________________________ From: gpfsug-discuss > on behalf of scale at us.ibm.com > Sent: Wednesday, March 30, 2022 10:02 AM To: gpfsug main discussion list > Cc: gpfsug-discuss > Subject: [EXTERNAL] Re: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file I think this can be answered by the IBM Spectrum Archive team and should go to a mailing list or discussion channel where they are involved. @Lyle: Do you know anyone who can answer this from Scale side. Regards, The Spectrum Scale (GPFS) team ------------------------------------------------------------------------------------------------------------------ If you feel that your question can benefit other users of Spectrum Scale (GPFS), then please post it to the public IBM developerWroks Forum at https://www.ibm.com/developerworks/community/forums/html/forum?id=11111111-0000-0000-0000-000000000479. If your query concerns a potential software error in Spectrum Scale (GPFS) and you have an IBM software maintenance contract please contact 1-800-237-5511 in the United States or your local IBM Service Center in other countries. The forum is informally monitored as time permits and should not be used for priority messages to the Spectrum Scale (GPFS) team. [Inactive hide details for "shao feng" ---29-03-2022 11.13.11 AM---Hello, Not sure if this is the right place to ask about Spect]"shao feng" ---29-03-2022 11.13.11 AM---Hello, Not sure if this is the right place to ask about Spectrum Archival EE. From: "shao feng" > To: "gpfsug main discussion list" > Date: 29-03-2022 11.13 AM Subject: [EXTERNAL] [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file Sent by: "gpfsug-discuss" > ________________________________ Hello, Not sure if this is the right place to ask about Spectrum Archival EE. In a GPFS+LTFS EE configuration, is there any mechanism to get notification from Spectrum Archival EE when it finishes a migrating/recall for a file, especially a transparent recall? Thank you!_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: image001.gif Type: image/gif Size: 106 bytes Desc: image001.gif URL: From info at odina.nl Mon Apr 4 13:39:15 2022 From: info at odina.nl (Jaap Jan Ouwehand) Date: Mon, 4 Apr 2022 14:39:15 +0200 Subject: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file In-Reply-To: References: <325BFDF6-2421-480E-8EE6-4EB8AE15E56F@us.ibm.com> Message-ID: <4ec4a1cc-7792-7d0e-30cf-a70ebf58aa59@odina.nl> Hello, In addition to the eeadm command, the REST API might also be a solution for job monitoring. https://www.ibm.com/docs/en/spectrum-archive-ee/1.3.0.0?topic=information-rest-api-task -- Kind regards Jaap Jan Ouwehand info at odina.nl Linux & Storage Solution expert Op 4-4-2022 om 14:13 schreef Takeshi Ishimoto: > > Unfortunately, there is no mailing list available for Spectrum > Archive, and the individuals Lyle listed are not the right contact. > > The closest option will be the IBM Tape Storage community at > https://community.ibm.com/community/user/storage/communities/community-home/digestviewer?communitykey=85531a8a-8971-4c0e-8d2b-098ba927269e > or > one should open a support ticket for question. > > Shao, > > There is no notification mechanism at the completion of transparent > recall nor other operation, but you should be able to monitor the > recent operations by ?eeadm task list? periodically and find the > detail of task with ?eeadm task show ?. > > Takeshi ISHIMOTO (?????) > Spectrum Archive/LTFS Development - Product Architect and DCT Lead > > SNIA Linear Tape File System (LTFS) TWG chair > IBM Systems > ishimoto at jp.ibm.com > > *From: *Lyle Gayne > > *Date: *Wednesday, March 30, 2022 at 7:10 AM > *To: *gpfsug main discussion list >, Mark Hill >, Khanh V Ngo > > *Subject: *Re: [EXTERNAL] Re: [gpfsug-discuss] notification when > Spectrum Archival EE finish migrating/recall for a file > > I guess this should start with Mark Hill (and Khanh), from Spectrum > Archive team. > > Lyle > > ------------------------------------------------------------------------ > > *From:*gpfsug-discuss on behalf of > scale at us.ibm.com > *Sent:* Wednesday, March 30, 2022 10:02 AM > *To:* gpfsug main discussion list > *Cc:* gpfsug-discuss > *Subject:* [EXTERNAL] Re: [gpfsug-discuss] notification when Spectrum > Archival EE finish migrating/recall for a file > > I think this can be answered by the IBM Spectrum Archive team and > should go to a mailing list or discussion channel where they are involved. > > @Lyle: Do you know anyone who can answer this from Scale side. > > > Regards, The Spectrum Scale (GPFS) team > > ------------------------------------------------------------------------------------------------------------------ > If you feel that your question can benefit other users of ?Spectrum > Scale (GPFS), then please post it to the public IBM developerWroks > Forum at > https://www.ibm.com/developerworks/community/forums/html/forum?id=11111111-0000-0000-0000-000000000479 > . > > > If your query concerns a potential software error in Spectrum Scale > (GPFS) and you have an IBM software maintenance contract please > contact ?1-800-237-5511 in the United States or your local IBM Service > Center in other countries. > > The forum is informally monitored as time permits and should not be > used for priority messages to the Spectrum Scale (GPFS) team. > > Inactive hide details for "shao feng" ---29-03-2022 11.13.11 > AM---Hello, Not sure if this is the right place to ask about > Spect"shao feng" ---29-03-2022 11.13.11 AM---Hello, Not sure if this > is the right place to ask about Spectrum Archival EE. > > From: "shao feng" > > To: "gpfsug main discussion list" > > Date: 29-03-2022 11.13 AM > Subject: [EXTERNAL] [gpfsug-discuss] notification when Spectrum > Archival EE finish migrating/recall for a file > Sent by: "gpfsug-discuss" > > > ------------------------------------------------------------------------ > > > > > Hello, > > Not sure if this is the right place to ask about Spectrum Archival EE. > In a GPFS+LTFS EE configuration, is there any mechanism to get > notification from Spectrum Archival EE when it finishes a > migrating/recall for a file, especially a transparent recall? > > Thank you!_______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org > > > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: image001.gif Type: image/gif Size: 106 bytes Desc: not available URL: From lgayne at us.ibm.com Tue Apr 19 13:33:00 2022 From: lgayne at us.ibm.com (Lyle Gayne) Date: Tue, 19 Apr 2022 12:33:00 +0000 Subject: [gpfsug-discuss] Mailing List Test In-Reply-To: <57a05639a9c416f4d6e44c09b79b9fbb@qsplace.co.uk> References: <57a05639a9c416f4d6e44c09b79b9fbb@qsplace.co.uk> Message-ID: Test of User Group to see if mailing list is still working... Apologies for this spam. Lyle ________________________________ From: gpfsug-discuss on behalf of laurence at qsplace.co.uk Sent: Sunday, March 13, 2022 6:43 PM To: gpfsug-discuss at gpfsug.org Subject: [EXTERNAL] [gpfsug-discuss] Mailing List Test Mailing List Test -------------- next part -------------- An HTML attachment was scrubbed... URL: From chair at gpfsug.org Thu Apr 21 08:10:11 2022 From: chair at gpfsug.org (chair at gpfsug.org) Date: Thu, 21 Apr 2022 08:10:11 +0100 Subject: [gpfsug-discuss] UK Spectrum Scale User Group meeting 30th June 2022 Message-ID: Hi all, Just a reminder that the next UK User Group meeting will be taking place in London (IBM York Road) on 30th June 2022. Registration is open at https://www.eventbrite.co.uk/e/spectrum-scale-user-group-registration-321290978967 Now is your opportunity to help shape the agenda, please feel free to send me ideas for talks we could ask IBM for (I don't promise we'll get them on the agenda mind!). Also if you would like to do a user talk, please get in touch with me and let me know. It could be a large scale deployment or even just a couple of nodes, it's your opportunity to showcase how you use Spectrum Scale and what for. Every year people tell us how valuable they find the user talks, but this needs YOU, so please do think about if you are able to offer a talk! As in the past we are looking for sponsorship to enable us to run an evening networking event. I've sent out details to those who have sponsored in the past and to those who have asked us directly about sponsorship opportunities. If you either haven't received this or are interested in becoming a sponsor, please email me directly. Thanks Paul From juergen.hannappel at desy.de Mon Apr 25 09:59:18 2022 From: juergen.hannappel at desy.de (Hannappel, Juergen) Date: Mon, 25 Apr 2022 10:59:18 +0200 (CEST) Subject: [gpfsug-discuss] 5.1.3.1 changes? Message-ID: <211580513.43933888.1650877158353.JavaMail.zimbra@desy.de> Hi, last Friday I got a notification that 5.1.3.1 is out, but https://public.dhe.ibm.com/storage/spectrumscale/spectrum_scale_apars_513x.html claims that it fixes no problems? -- Dr. J?rgen Hannappel DESY/IT Tel. : +49 40 8998-4616 From Renar.Grunenberg at huk-coburg.de Mon Apr 25 16:53:02 2022 From: Renar.Grunenberg at huk-coburg.de (Grunenberg, Renar) Date: Mon, 25 Apr 2022 15:53:02 +0000 Subject: [gpfsug-discuss] 5.1.3.1 changes? In-Reply-To: <211580513.43933888.1650877158353.JavaMail.zimbra@desy.de> References: <211580513.43933888.1650877158353.JavaMail.zimbra@desy.de> Message-ID: <1619424d058a489f96ca1baa63e3bff3@huk-coburg.de> Hallo J?rgen, i had the same problem, but since 1 hour the list is available. Renar Grunenberg Abteilung Informatik - Betrieb HUK-COBURG Bahnhofsplatz 96444 Coburg Telefon: 09561 96-44110 Telefax: 09561 96-44104 E-Mail: Renar.Grunenberg at huk-coburg.de Internet: www.huk.de ======================================================================= HUK-COBURG Haftpflicht-Unterst?tzungs-Kasse kraftfahrender Beamter Deutschlands a. G. in Coburg Reg.-Gericht Coburg HRB 100; St.-Nr. 9212/101/00021 Sitz der Gesellschaft: Bahnhofsplatz, 96444 Coburg Vorsitzender des Aufsichtsrats: Prof. Dr. Heinrich R. Schradin. Vorstand: Klaus-J?rgen Heitmann (Sprecher), Stefan Gronbach, Dr. Hans Olav Her?y, Dr. Helen Reck, Dr. J?rg Rheinl?nder, Thomas Sehn, Daniel Thomas. ======================================================================= Diese Nachricht enth?lt vertrauliche und/oder rechtlich gesch?tzte Informationen. Wenn Sie nicht der richtige Adressat sind oder diese Nachricht irrt?mlich erhalten haben, informieren Sie bitte sofort den Absender und vernichten Sie diese Nachricht. Das unerlaubte Kopieren sowie die unbefugte Weitergabe dieser Nachricht ist nicht gestattet. This information may contain confidential and/or privileged information. If you are not the intended recipient (or have received this information in error) please notify the sender immediately and destroy this information. Any unauthorized copying, disclosure or distribution of the material in this information is strictly forbidden. ======================================================================= -----Urspr?ngliche Nachricht----- Von: gpfsug-discuss Im Auftrag von Hannappel, Juergen Gesendet: Montag, 25. April 2022 10:59 An: gpfsug-discuss Betreff: [gpfsug-discuss] 5.1.3.1 changes? Hi, last Friday I got a notification that 5.1.3.1 is out, but https://public.dhe.ibm.com/storage/spectrumscale/spectrum_scale_apars_513x.html claims that it fixes no problems? -- Dr. J?rgen Hannappel DESY/IT Tel. : +49 40 8998-4616 _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org From juergen.hannappel at desy.de Wed Apr 27 14:19:02 2022 From: juergen.hannappel at desy.de (Hannappel, Juergen) Date: Wed, 27 Apr 2022 15:19:02 +0200 (CEST) Subject: [gpfsug-discuss] Updating a medium size cluster efficiently Message-ID: <733714875.45755494.1651065542428.JavaMail.zimbra@desy.de> Hi, we have a medium size gpfs client cluster (a few hundred nodes) and want to update the gpfs version in an efficient way in a rolling update, i.e.. update each node when it can be rebooted. Doing so via a slurm script when the node is drained jut before the reboot work only most of the time because in some cases even when the node is drained the file systems are still buly and can't be unmounted, so the update fails. Therefore I tried to trigger the update on the reboot, before gpfs starts. To do so I added a systemd service that is sheduled before the gpfs.service, which does a yum update (we run CentOs 7.9) but: In the postinstall script of gpfs.base the gpfs.service is disabled and re-enabled via systemctl, and systemd apparently get's that wrong, so that if the update really happens it afterwards will not start the gpfs.service. Does any one have a clever way how to do a rolling update that really works without maunually hunting after some per cent of machines that don't manage it on the first go? -- Dr. J?rgen Hannappel DESY/IT Tel. : +49 40 8998-4616 From jonathan.buzzard at strath.ac.uk Wed Apr 27 21:21:07 2022 From: jonathan.buzzard at strath.ac.uk (Jonathan Buzzard) Date: Wed, 27 Apr 2022 21:21:07 +0100 Subject: [gpfsug-discuss] Updating a medium size cluster efficiently In-Reply-To: <733714875.45755494.1651065542428.JavaMail.zimbra@desy.de> References: <733714875.45755494.1651065542428.JavaMail.zimbra@desy.de> Message-ID: On 27/04/2022 14:19, Hannappel, Juergen wrote: > > Hi, > we have a medium size gpfs client cluster (a few hundred nodes) > and want to update the gpfs version in an efficient way in a > rolling update, i.e.. update each node when it can be rebooted. > > Doing so via a slurm script when the node is drained jut before > the reboot work only most of the time because in some cases even > when the node is drained the file systems are still buly and can't be unmounted, > so the update fails. > I assume this is because you have dead jobs? The general trick is to submit a job as a special user that has sudo privileges, that runs as the next job on every node. That way you don't need to wait for the node to drain. Last "user" job on the node finishes and then the "special" job runs. It does it magic and reboots the node. Winner winner chicken dinner. > Therefore I tried to trigger the update on the reboot, before gpfs starts. > To do so I added a systemd service that is sheduled before the gpfs.service, > which does a yum update (we run CentOs 7.9) but: > > In the postinstall script of gpfs.base the gpfs.service is disabled and re-enabled > via systemctl, and systemd apparently get's that wrong, so that if > the update really happens it afterwards will not start the gpfs.service. > > Does any one have a clever way how to do a rolling update that really works > without maunually hunting after some per cent of machines that don't manage > it on the first go? > What you could do to find the nodes that don't work is have the upgrade script do an mmshutdown first before attempting the upgrade. Then check it actually managed to shutdown and if it didn't then send an email to an appropriate person saying there is an issue, before say putting the node in drain. The man page for mmshutdown says it has an exit code of zero on success, and none zero of failure so should be trivial to script. Being really clever I think you could then have the script submit a second copy of itself to the node that again will run as the next job and then reboot the node. That way when it comes back up it should be able to unmount GPFS and install the upgrade as the reboot will have cleared the issues that prevented the mmshutdown from working. You would obviously need to trial this out. If you are just looking to upgrade gpfs-gplbin and don't want to have it being recompiled on every node, then there is a trick with systemd. What you do is create /etc/systemd/system/gpfs.service.d/install-module.conf with the following contents [Service] ExecStartPre=-/usr/bin/yum --assumeyes install gpfs.gplbin-%v then everytime GPFS starts up it attempts to install the module for the currently running kernel (the special magic %v). This presumes you have a repository with the appropriate gpfs-gplbin RPM setup. Basically I take a node out build the RPM, test it is working and then deploy. I have a special RPM that installs the above local customization to the GPFS serivce unit file. JAB. -- Jonathan A. Buzzard Tel: +44141-5483420 HPC System Administrator, ARCHIE-WeSt. University of Strathclyde, John Anderson Building, Glasgow. G4 0NG From p.childs at qmul.ac.uk Thu Apr 28 15:42:38 2022 From: p.childs at qmul.ac.uk (Peter Childs) Date: Thu, 28 Apr 2022 14:42:38 +0000 Subject: [gpfsug-discuss] Job Opportunities In-Reply-To: References: Message-ID: We are recruiting for a couple of positions at Queen Mary University of London, within our Research group. Both these jobs will have some contact with Spectrum Scale and one of them is directly helping to support large Spectrum Scale cluster. If anyone is interested please apply. https://www.qmul.ac.uk/jobs/vacancies/items/6955.html https://www.qmul.ac.uk/jobs/vacancies/items/6949.html Peter Childs ITS Research Storage Queen Mary University of London. From ISHIMOTO at jp.ibm.com Mon Apr 4 13:13:28 2022 From: ISHIMOTO at jp.ibm.com (Takeshi Ishimoto) Date: Mon, 4 Apr 2022 12:13:28 +0000 Subject: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file In-Reply-To: <325BFDF6-2421-480E-8EE6-4EB8AE15E56F@us.ibm.com> References: <325BFDF6-2421-480E-8EE6-4EB8AE15E56F@us.ibm.com> Message-ID: Unfortunately, there is no mailing list available for Spectrum Archive, and the individuals Lyle listed are not the right contact. The closest option will be the IBM Tape Storage community at https://community.ibm.com/community/user/storage/communities/community-home/digestviewer?communitykey=85531a8a-8971-4c0e-8d2b-098ba927269e or one should open a support ticket for question. Shao, There is no notification mechanism at the completion of transparent recall nor other operation, but you should be able to monitor the recent operations by ?eeadm task list? periodically and find the detail of task with ?eeadm task show ?. Takeshi ISHIMOTO (?????) Spectrum Archive/LTFS Development - Product Architect and DCT Lead SNIA Linear Tape File System (LTFS) TWG chair IBM Systems ishimoto at jp.ibm.com From: Lyle Gayne > Date: Wednesday, March 30, 2022 at 7:10 AM To: gpfsug main discussion list >, Mark Hill >, Khanh V Ngo > Subject: Re: [EXTERNAL] Re: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file I guess this should start with Mark Hill (and Khanh), from Spectrum Archive team. Lyle ________________________________ From: gpfsug-discuss > on behalf of scale at us.ibm.com > Sent: Wednesday, March 30, 2022 10:02 AM To: gpfsug main discussion list > Cc: gpfsug-discuss > Subject: [EXTERNAL] Re: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file I think this can be answered by the IBM Spectrum Archive team and should go to a mailing list or discussion channel where they are involved. @Lyle: Do you know anyone who can answer this from Scale side. Regards, The Spectrum Scale (GPFS) team ------------------------------------------------------------------------------------------------------------------ If you feel that your question can benefit other users of Spectrum Scale (GPFS), then please post it to the public IBM developerWroks Forum at https://www.ibm.com/developerworks/community/forums/html/forum?id=11111111-0000-0000-0000-000000000479. If your query concerns a potential software error in Spectrum Scale (GPFS) and you have an IBM software maintenance contract please contact 1-800-237-5511 in the United States or your local IBM Service Center in other countries. The forum is informally monitored as time permits and should not be used for priority messages to the Spectrum Scale (GPFS) team. [Inactive hide details for "shao feng" ---29-03-2022 11.13.11 AM---Hello, Not sure if this is the right place to ask about Spect]"shao feng" ---29-03-2022 11.13.11 AM---Hello, Not sure if this is the right place to ask about Spectrum Archival EE. From: "shao feng" > To: "gpfsug main discussion list" > Date: 29-03-2022 11.13 AM Subject: [EXTERNAL] [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file Sent by: "gpfsug-discuss" > ________________________________ Hello, Not sure if this is the right place to ask about Spectrum Archival EE. In a GPFS+LTFS EE configuration, is there any mechanism to get notification from Spectrum Archival EE when it finishes a migrating/recall for a file, especially a transparent recall? Thank you!_______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: image001.gif Type: image/gif Size: 106 bytes Desc: image001.gif URL: From info at odina.nl Mon Apr 4 13:39:15 2022 From: info at odina.nl (Jaap Jan Ouwehand) Date: Mon, 4 Apr 2022 14:39:15 +0200 Subject: [gpfsug-discuss] notification when Spectrum Archival EE finish migrating/recall for a file In-Reply-To: References: <325BFDF6-2421-480E-8EE6-4EB8AE15E56F@us.ibm.com> Message-ID: <4ec4a1cc-7792-7d0e-30cf-a70ebf58aa59@odina.nl> Hello, In addition to the eeadm command, the REST API might also be a solution for job monitoring. https://www.ibm.com/docs/en/spectrum-archive-ee/1.3.0.0?topic=information-rest-api-task -- Kind regards Jaap Jan Ouwehand info at odina.nl Linux & Storage Solution expert Op 4-4-2022 om 14:13 schreef Takeshi Ishimoto: > > Unfortunately, there is no mailing list available for Spectrum > Archive, and the individuals Lyle listed are not the right contact. > > The closest option will be the IBM Tape Storage community at > https://community.ibm.com/community/user/storage/communities/community-home/digestviewer?communitykey=85531a8a-8971-4c0e-8d2b-098ba927269e > or > one should open a support ticket for question. > > Shao, > > There is no notification mechanism at the completion of transparent > recall nor other operation, but you should be able to monitor the > recent operations by ?eeadm task list? periodically and find the > detail of task with ?eeadm task show ?. > > Takeshi ISHIMOTO (?????) > Spectrum Archive/LTFS Development - Product Architect and DCT Lead > > SNIA Linear Tape File System (LTFS) TWG chair > IBM Systems > ishimoto at jp.ibm.com > > *From: *Lyle Gayne > > *Date: *Wednesday, March 30, 2022 at 7:10 AM > *To: *gpfsug main discussion list >, Mark Hill >, Khanh V Ngo > > *Subject: *Re: [EXTERNAL] Re: [gpfsug-discuss] notification when > Spectrum Archival EE finish migrating/recall for a file > > I guess this should start with Mark Hill (and Khanh), from Spectrum > Archive team. > > Lyle > > ------------------------------------------------------------------------ > > *From:*gpfsug-discuss on behalf of > scale at us.ibm.com > *Sent:* Wednesday, March 30, 2022 10:02 AM > *To:* gpfsug main discussion list > *Cc:* gpfsug-discuss > *Subject:* [EXTERNAL] Re: [gpfsug-discuss] notification when Spectrum > Archival EE finish migrating/recall for a file > > I think this can be answered by the IBM Spectrum Archive team and > should go to a mailing list or discussion channel where they are involved. > > @Lyle: Do you know anyone who can answer this from Scale side. > > > Regards, The Spectrum Scale (GPFS) team > > ------------------------------------------------------------------------------------------------------------------ > If you feel that your question can benefit other users of ?Spectrum > Scale (GPFS), then please post it to the public IBM developerWroks > Forum at > https://www.ibm.com/developerworks/community/forums/html/forum?id=11111111-0000-0000-0000-000000000479 > . > > > If your query concerns a potential software error in Spectrum Scale > (GPFS) and you have an IBM software maintenance contract please > contact ?1-800-237-5511 in the United States or your local IBM Service > Center in other countries. > > The forum is informally monitored as time permits and should not be > used for priority messages to the Spectrum Scale (GPFS) team. > > Inactive hide details for "shao feng" ---29-03-2022 11.13.11 > AM---Hello, Not sure if this is the right place to ask about > Spect"shao feng" ---29-03-2022 11.13.11 AM---Hello, Not sure if this > is the right place to ask about Spectrum Archival EE. > > From: "shao feng" > > To: "gpfsug main discussion list" > > Date: 29-03-2022 11.13 AM > Subject: [EXTERNAL] [gpfsug-discuss] notification when Spectrum > Archival EE finish migrating/recall for a file > Sent by: "gpfsug-discuss" > > > ------------------------------------------------------------------------ > > > > > Hello, > > Not sure if this is the right place to ask about Spectrum Archival EE. > In a GPFS+LTFS EE configuration, is there any mechanism to get > notification from Spectrum Archival EE when it finishes a > migrating/recall for a file, especially a transparent recall? > > Thank you!_______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org > > > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: image001.gif Type: image/gif Size: 106 bytes Desc: not available URL: From lgayne at us.ibm.com Tue Apr 19 13:33:00 2022 From: lgayne at us.ibm.com (Lyle Gayne) Date: Tue, 19 Apr 2022 12:33:00 +0000 Subject: [gpfsug-discuss] Mailing List Test In-Reply-To: <57a05639a9c416f4d6e44c09b79b9fbb@qsplace.co.uk> References: <57a05639a9c416f4d6e44c09b79b9fbb@qsplace.co.uk> Message-ID: Test of User Group to see if mailing list is still working... Apologies for this spam. Lyle ________________________________ From: gpfsug-discuss on behalf of laurence at qsplace.co.uk Sent: Sunday, March 13, 2022 6:43 PM To: gpfsug-discuss at gpfsug.org Subject: [EXTERNAL] [gpfsug-discuss] Mailing List Test Mailing List Test -------------- next part -------------- An HTML attachment was scrubbed... URL: From chair at gpfsug.org Thu Apr 21 08:10:11 2022 From: chair at gpfsug.org (chair at gpfsug.org) Date: Thu, 21 Apr 2022 08:10:11 +0100 Subject: [gpfsug-discuss] UK Spectrum Scale User Group meeting 30th June 2022 Message-ID: Hi all, Just a reminder that the next UK User Group meeting will be taking place in London (IBM York Road) on 30th June 2022. Registration is open at https://www.eventbrite.co.uk/e/spectrum-scale-user-group-registration-321290978967 Now is your opportunity to help shape the agenda, please feel free to send me ideas for talks we could ask IBM for (I don't promise we'll get them on the agenda mind!). Also if you would like to do a user talk, please get in touch with me and let me know. It could be a large scale deployment or even just a couple of nodes, it's your opportunity to showcase how you use Spectrum Scale and what for. Every year people tell us how valuable they find the user talks, but this needs YOU, so please do think about if you are able to offer a talk! As in the past we are looking for sponsorship to enable us to run an evening networking event. I've sent out details to those who have sponsored in the past and to those who have asked us directly about sponsorship opportunities. If you either haven't received this or are interested in becoming a sponsor, please email me directly. Thanks Paul From juergen.hannappel at desy.de Mon Apr 25 09:59:18 2022 From: juergen.hannappel at desy.de (Hannappel, Juergen) Date: Mon, 25 Apr 2022 10:59:18 +0200 (CEST) Subject: [gpfsug-discuss] 5.1.3.1 changes? Message-ID: <211580513.43933888.1650877158353.JavaMail.zimbra@desy.de> Hi, last Friday I got a notification that 5.1.3.1 is out, but https://public.dhe.ibm.com/storage/spectrumscale/spectrum_scale_apars_513x.html claims that it fixes no problems? -- Dr. J?rgen Hannappel DESY/IT Tel. : +49 40 8998-4616 From Renar.Grunenberg at huk-coburg.de Mon Apr 25 16:53:02 2022 From: Renar.Grunenberg at huk-coburg.de (Grunenberg, Renar) Date: Mon, 25 Apr 2022 15:53:02 +0000 Subject: [gpfsug-discuss] 5.1.3.1 changes? In-Reply-To: <211580513.43933888.1650877158353.JavaMail.zimbra@desy.de> References: <211580513.43933888.1650877158353.JavaMail.zimbra@desy.de> Message-ID: <1619424d058a489f96ca1baa63e3bff3@huk-coburg.de> Hallo J?rgen, i had the same problem, but since 1 hour the list is available. Renar Grunenberg Abteilung Informatik - Betrieb HUK-COBURG Bahnhofsplatz 96444 Coburg Telefon: 09561 96-44110 Telefax: 09561 96-44104 E-Mail: Renar.Grunenberg at huk-coburg.de Internet: www.huk.de ======================================================================= HUK-COBURG Haftpflicht-Unterst?tzungs-Kasse kraftfahrender Beamter Deutschlands a. G. in Coburg Reg.-Gericht Coburg HRB 100; St.-Nr. 9212/101/00021 Sitz der Gesellschaft: Bahnhofsplatz, 96444 Coburg Vorsitzender des Aufsichtsrats: Prof. Dr. Heinrich R. Schradin. Vorstand: Klaus-J?rgen Heitmann (Sprecher), Stefan Gronbach, Dr. Hans Olav Her?y, Dr. Helen Reck, Dr. J?rg Rheinl?nder, Thomas Sehn, Daniel Thomas. ======================================================================= Diese Nachricht enth?lt vertrauliche und/oder rechtlich gesch?tzte Informationen. Wenn Sie nicht der richtige Adressat sind oder diese Nachricht irrt?mlich erhalten haben, informieren Sie bitte sofort den Absender und vernichten Sie diese Nachricht. Das unerlaubte Kopieren sowie die unbefugte Weitergabe dieser Nachricht ist nicht gestattet. This information may contain confidential and/or privileged information. If you are not the intended recipient (or have received this information in error) please notify the sender immediately and destroy this information. Any unauthorized copying, disclosure or distribution of the material in this information is strictly forbidden. ======================================================================= -----Urspr?ngliche Nachricht----- Von: gpfsug-discuss Im Auftrag von Hannappel, Juergen Gesendet: Montag, 25. April 2022 10:59 An: gpfsug-discuss Betreff: [gpfsug-discuss] 5.1.3.1 changes? Hi, last Friday I got a notification that 5.1.3.1 is out, but https://public.dhe.ibm.com/storage/spectrumscale/spectrum_scale_apars_513x.html claims that it fixes no problems? -- Dr. J?rgen Hannappel DESY/IT Tel. : +49 40 8998-4616 _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org From juergen.hannappel at desy.de Wed Apr 27 14:19:02 2022 From: juergen.hannappel at desy.de (Hannappel, Juergen) Date: Wed, 27 Apr 2022 15:19:02 +0200 (CEST) Subject: [gpfsug-discuss] Updating a medium size cluster efficiently Message-ID: <733714875.45755494.1651065542428.JavaMail.zimbra@desy.de> Hi, we have a medium size gpfs client cluster (a few hundred nodes) and want to update the gpfs version in an efficient way in a rolling update, i.e.. update each node when it can be rebooted. Doing so via a slurm script when the node is drained jut before the reboot work only most of the time because in some cases even when the node is drained the file systems are still buly and can't be unmounted, so the update fails. Therefore I tried to trigger the update on the reboot, before gpfs starts. To do so I added a systemd service that is sheduled before the gpfs.service, which does a yum update (we run CentOs 7.9) but: In the postinstall script of gpfs.base the gpfs.service is disabled and re-enabled via systemctl, and systemd apparently get's that wrong, so that if the update really happens it afterwards will not start the gpfs.service. Does any one have a clever way how to do a rolling update that really works without maunually hunting after some per cent of machines that don't manage it on the first go? -- Dr. J?rgen Hannappel DESY/IT Tel. : +49 40 8998-4616 From jonathan.buzzard at strath.ac.uk Wed Apr 27 21:21:07 2022 From: jonathan.buzzard at strath.ac.uk (Jonathan Buzzard) Date: Wed, 27 Apr 2022 21:21:07 +0100 Subject: [gpfsug-discuss] Updating a medium size cluster efficiently In-Reply-To: <733714875.45755494.1651065542428.JavaMail.zimbra@desy.de> References: <733714875.45755494.1651065542428.JavaMail.zimbra@desy.de> Message-ID: On 27/04/2022 14:19, Hannappel, Juergen wrote: > > Hi, > we have a medium size gpfs client cluster (a few hundred nodes) > and want to update the gpfs version in an efficient way in a > rolling update, i.e.. update each node when it can be rebooted. > > Doing so via a slurm script when the node is drained jut before > the reboot work only most of the time because in some cases even > when the node is drained the file systems are still buly and can't be unmounted, > so the update fails. > I assume this is because you have dead jobs? The general trick is to submit a job as a special user that has sudo privileges, that runs as the next job on every node. That way you don't need to wait for the node to drain. Last "user" job on the node finishes and then the "special" job runs. It does it magic and reboots the node. Winner winner chicken dinner. > Therefore I tried to trigger the update on the reboot, before gpfs starts. > To do so I added a systemd service that is sheduled before the gpfs.service, > which does a yum update (we run CentOs 7.9) but: > > In the postinstall script of gpfs.base the gpfs.service is disabled and re-enabled > via systemctl, and systemd apparently get's that wrong, so that if > the update really happens it afterwards will not start the gpfs.service. > > Does any one have a clever way how to do a rolling update that really works > without maunually hunting after some per cent of machines that don't manage > it on the first go? > What you could do to find the nodes that don't work is have the upgrade script do an mmshutdown first before attempting the upgrade. Then check it actually managed to shutdown and if it didn't then send an email to an appropriate person saying there is an issue, before say putting the node in drain. The man page for mmshutdown says it has an exit code of zero on success, and none zero of failure so should be trivial to script. Being really clever I think you could then have the script submit a second copy of itself to the node that again will run as the next job and then reboot the node. That way when it comes back up it should be able to unmount GPFS and install the upgrade as the reboot will have cleared the issues that prevented the mmshutdown from working. You would obviously need to trial this out. If you are just looking to upgrade gpfs-gplbin and don't want to have it being recompiled on every node, then there is a trick with systemd. What you do is create /etc/systemd/system/gpfs.service.d/install-module.conf with the following contents [Service] ExecStartPre=-/usr/bin/yum --assumeyes install gpfs.gplbin-%v then everytime GPFS starts up it attempts to install the module for the currently running kernel (the special magic %v). This presumes you have a repository with the appropriate gpfs-gplbin RPM setup. Basically I take a node out build the RPM, test it is working and then deploy. I have a special RPM that installs the above local customization to the GPFS serivce unit file. JAB. -- Jonathan A. Buzzard Tel: +44141-5483420 HPC System Administrator, ARCHIE-WeSt. University of Strathclyde, John Anderson Building, Glasgow. G4 0NG From p.childs at qmul.ac.uk Thu Apr 28 15:42:38 2022 From: p.childs at qmul.ac.uk (Peter Childs) Date: Thu, 28 Apr 2022 14:42:38 +0000 Subject: [gpfsug-discuss] Job Opportunities In-Reply-To: References: Message-ID: We are recruiting for a couple of positions at Queen Mary University of London, within our Research group. Both these jobs will have some contact with Spectrum Scale and one of them is directly helping to support large Spectrum Scale cluster. If anyone is interested please apply. https://www.qmul.ac.uk/jobs/vacancies/items/6955.html https://www.qmul.ac.uk/jobs/vacancies/items/6949.html Peter Childs ITS Research Storage Queen Mary University of London.