[gpfsug-discuss] pagepool
scale
scale at us.ibm.com
Fri Mar 8 16:31:44 GMT 2024
To the original question about the size of the pagepool being 323908133683 bytes, I think that should only be applied to the IO nodes (the ones handling all the NSD IOs) not every node in the cluster. The large pagepool size is needed on the IO nodes for GNR to function properly.
From: gpfsug-discuss <gpfsug-discuss-bounces at gpfsug.org> on behalf of Jonathan Buzzard <jonathan.buzzard at strath.ac.uk>
Date: Friday, March 8, 2024 at 11:27 AM
To: gpfsug-discuss at gpfsug.org <gpfsug-discuss at gpfsug.org>
Subject: [EXTERNAL] Re: [gpfsug-discuss] pagepool
On 08/03/2024 14:50, Jarsulic, Michael [BSD] wrote:
> Ibán,
>
> What are you using for your scheduler?
>
> On my compute nodes, I am setting the pagepool to 16 GB and setting
> aside specialized memory for GPFS that will not be allocated to jobs.
>
What you would normally do is create a node class
mmcrnodeclass compute -N node001,node002,node003,node004,.....
then set the pagepool appropriately
mmchconfig pagepool=16G -i -N compute
We then use slurm to limit the maximum amount of RAM a job can have on a
node to be physical RAM minus the pagepool size minus a bit more for
good measure to allow for the OS.
If the OOM is kicking in then you need to reduce the RAM limit in slurm
some more till it stops.
Note we also twiddle with some other limits for compute nodes
mmchconfig maxFilesToCache=8000 -N compute
mmchconfig maxStatCache=16000 -N compute
We have a slew of node classes where these settings are tweaked to
account for their RAM and their role so dssg, compute,
gpu,protocol,teaching, and login. All nodes belong to one or more node
classes. Which reminds me I need a gui node class now.
JAB.
--
Jonathan A. Buzzard Tel: +44141-5483420
HPC System Administrator, ARCHIE-WeSt.
University of Strathclyde, John Anderson Building, Glasgow. G4 0NG
_______________________________________________
gpfsug-discuss mailing list
gpfsug-discuss at gpfsug.org
http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://gpfsug.org/pipermail/gpfsug-discuss_gpfsug.org/attachments/20240308/8df6a9e7/attachment.htm>
More information about the gpfsug-discuss
mailing list