From henrik.cednert at onepost.se Tue Sep 3 13:10:50 2024 From: henrik.cednert at onepost.se (Henrik Cednert) Date: Tue, 3 Sep 2024 12:10:50 +0000 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: Message-ID: Still no solution here regarding this. Have tested other cables. Have tested to change tcp window size, no change Played with numa in the bios, no change Played with hyperthreading in bios, no change Have anyone managed to get some speed out of windows 11 and gpfs? -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Friday, 9 August 2024 17:25 To: gpfsug-discuss at gpfsug.org Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. VARNING: DETTA ?R ETT EXTERNT MAIL. Klicka inte p? n?gra l?nkar oavsett hur legitima de verkar utan att verifiera. Hello I have some issues with write performance on a windows 11 pro system and I'm out of ideas here. Hopefully someone here have some bright ideas and/or experience of GPFS on Windows 11? The system is a: Windows 11 Pro 22H2 2 x Intel(R) Xeon(R) Gold 6418H 2.10 GHz 512 GB RAM GPFS 5.1.9.4 Mellanox ConnectX 6 Dx 100GbE connected to Mellanox Switch with 5m Mellanox DAC. Before deploying this workstation we had a single socket system as a test bench where we got 60 GbE in both directons with iPerf and around 6GB/sec write and 3GB/sec read from the system over GPFS (fio tests, same tests as furhter down here). With that system I had loads of issues before getting to that point though. MS Defender had to be forcefully disabled via regedit some other tweaks. All those tweaks have been performed in this new system as well, but I can't get the proper speed out of it. On this new system and with iPerf to the storage servers I get around 50-60GbE in both directions and send and receive. If I mount the storage over SMB and 100GbE via the storage gateway servers I get around 3GB/sec read and write with Blackmagics Disk speed test. I have not tweaked the system for samba performande, just a test to see what it would give and part of the troubleshooting. If I run Blackmagics diskspeed test to the GPFS mount I instead get around 700MB/sec write and 400MB/sec read. Starting to think that the Blackmagic test might not run properly on this machine with these CPUs though. Or it's related to the mmfsd process maybe, how that threads or not threads...? But if we instead look at fio. I have a bat script that loops through a bunch of FIO-tests. A test that I have been using over the years so that we easily can benchmark all deployed systems with the exakt same tests. The tests are named like: seqrw-gb-mb-t The result when I run this is like the below list. Number in parenthesis is the by fio reported latency. Job: seqrw-40gb-1mb-t1 ????????????Write: 162 MB/s (6 ms) ????????????Read: 1940 MB/s (1 ms) Job: seqrw-20gb-1mb-t2 ????????????Write: 286 MB/s (7 ms) ????????????Read: 3952 MB/s (1 ms) Job: seqrw-10gb-1mb-t4 ????????????Write: 549 MB/s (7 ms) ????????????Read: 6987 MB/s (1 ms) Job: seqrw-05gb-1mb-t8 ????????????Write: 989 MB/s (8 ms) ????????????Read: 7721 MB/s (1 ms) Job: seqrw-40gb-2mb-t1 ????????????Write: 161 MB/s (12 ms) ????????????Read: 2261 MB/s (0 ms) Job: seqrw-20gb-2mb-t2 ????????????Write: 348 MB/s (11 ms) ????????????Read: 4266 MB/s (1 ms) Job: seqrw-10gb-2mb-t4 ????????????Write: 626 MB/s (13 ms) ????????????Read: 4949 MB/s (1 ms) Job: seqrw-05gb-2mb-t8 ????????????Write: 1154 MB/s (14 ms) ????????????Read: 7007 MB/s (2 ms) Job: seqrw-40gb-4mb-t1 ????????????Write: 161 MB/s (25 ms) ????????????Read: 2083 MB/s (1 ms) Job: seqrw-20gb-4mb-t2 ????????????Write: 352 MB/s (23 ms) ????????????Read: 4317 MB/s (2 ms) Job: seqrw-10gb-4mb-t4 ????????????Write: 696 MB/s (23 ms) ????????????Read: 7358 MB/s (2 ms) Job: seqrw-05gb-4mb-t8 ????????????Write: 1251 MB/s (25 ms) ????????????Read: 6707 MB/s (5 ms) So with fio I get a very nice read speed, but the write is horrendous and I cannot find what causes it. I have looked at affinity settings for the mmfsd process but not sure I fully understand it. But no matter what I set it to, I see no difference. I have "played" with the bios and tried with/without hyperthreading, numa and so on. And nothing affects atleast the blackmagic disk speed test. the current settings for this host is like below. I write "current" because I have tested a few different settings here but nothing affects the write speed. maxTcpConnsPerNodeConn for sure bumped the read speed though. nsdMaxWorkerThreads 16 prefetchPct 60 maxTcpConnsPerNodeConn 8 maxMBpS 14000 Does anyone have any suggestions or ideas on how to troubleshoot this? Thanks -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. -------------- next part -------------- An HTML attachment was scrubbed... URL: From uwe.falke at kit.edu Tue Sep 3 16:35:48 2024 From: uwe.falke at kit.edu (Uwe Falke) Date: Tue, 3 Sep 2024 17:35:48 +0200 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: Message-ID: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> Hi, Henrik, while I am not using Windows I'd start investigating the usual things (see below). But first you should describe your set-up better. Where are the NSDs : locally attached to the Windows box? In some NSD servers? If the latter -- what is the link to the NSD servers? via your GbE link? FC? IB? separate Ethernet? What type of storage? Spinning Disks? Flash? How long are your I/Os waiting on the client (compare that to the waiting times on the NSD server if applicable)? not sure whether that is available on Windows, but mmdiag --waiters mmdiag --iohistory might be of use. Somewhere in the chain from your application to the storage backend there is a delay and you should first find out where that occurs I think. Bye Uwe On 03.09.24 14:10, Henrik Cednert wrote: > Still no solution here regarding this. > > Have tested other cables. > Have tested to change tcp window size, no change > Played with numa in the bios, no change > Played with hyperthreading in bios, no change > > > Have anyone managed to get some speed out of windows 11 and gpfs? > > > -- > > Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */? OnePost > *(formerly?Filmlance Post) > > ?? *OnePost*, formerly Filmlance's post-production, is now an > independent part of the Banijay Group. > New name, same team ? business as usual at OnePost. > > > > ------------------------------------------------------------------------ > *From:* gpfsug-discuss on behalf > of Henrik Cednert > *Sent:* Friday, 9 August 2024 17:25 > *To:* gpfsug-discuss at gpfsug.org > *Subject:* [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. > Performance issues, write. > > ***VARNING: DETTA ?R ETT EXTERNT MAIL. Klicka inte p? n?gra l?nkar > oavsett hur legitima de verkar utan att verifiera.* > > > Hello > > I have some issues with write performance on a windows 11 pro system > and I'm out of ideas here. Hopefully someone here have some bright > ideas and/or experience of GPFS on Windows 11? > > The system is a: > > Windows 11 Pro 22H2 > 2 x Intel(R) Xeon(R) Gold 6418H ? 2.10 GHz > 512 GB RAM > GPFS 5.1.9.4 > Mellanox ConnectX 6 Dx > 100GbE connected to Mellanox Switch with 5m Mellanox DAC. > > Before deploying this workstation we had a single socket system as a > test bench where we got 60 GbE in both directons with iPerf and around > 6GB/sec write and 3GB/sec read from the system over GPFS (fio tests, > same tests as furhter down here). > > With that system I had loads of issues before getting to that point > though. MS Defender had to be forcefully disabled via regedit some > other tweaks. All those tweaks have been performed in this new system > as well, but I can't get the proper speed out of it. > > > On this new system and with iPerf to the storage servers I get around > 50-60GbE in both directions and send and receive. > > If I mount the storage over SMB and 100GbE via the storage gateway > servers I get around 3GB/sec read and write with Blackmagics Disk > speed test. I have not tweaked the system for samba performande, just > a test to see what it would give and part of the troubleshooting. > > If I run Blackmagics diskspeed test to the GPFS mount I instead get > around 700MB/sec write and 400MB/sec read. > > Starting to think that the Blackmagic test might not run properly on > this machine with these CPUs though. Or it's related to the mmfsd > process maybe, how that threads or not threads...? > > But if we instead look at fio. I have a bat script that loops through > a bunch of FIO-tests. A test that I have been using over the years so > that we easily can benchmark all deployed systems with the exakt same > tests. The tests are named like: > > seqrw-gb-mb-t > > The result when I run this is like the below list. Number in > parenthesis is the by fio reported latency. > > Job: seqrw-40gb-1mb-t1 > ????????????Write: 162 MB/s (6 ms) > ????????????Read: 1940 MB/s (1 ms) > > Job: seqrw-20gb-1mb-t2 > ????????????Write: 286 MB/s (7 ms) > ????????????Read: 3952 MB/s (1 ms) > > Job: seqrw-10gb-1mb-t4 > ????????????Write: 549 MB/s (7 ms) > ????????????Read: 6987 MB/s (1 ms) > > Job: seqrw-05gb-1mb-t8 > ????????????Write: 989 MB/s (8 ms) > ????????????Read: 7721 MB/s (1 ms) > > Job: seqrw-40gb-2mb-t1 > ????????????Write: 161 MB/s (12 ms) > ????????????Read: 2261 MB/s (0 ms) > > Job: seqrw-20gb-2mb-t2 > ????????????Write: 348 MB/s (11 ms) > ????????????Read: 4266 MB/s (1 ms) > > Job: seqrw-10gb-2mb-t4 > ????????????Write: 626 MB/s (13 ms) > ????????????Read: 4949 MB/s (1 ms) > > Job: seqrw-05gb-2mb-t8 > ????????????Write: 1154 MB/s (14 ms) > ????????????Read: 7007 MB/s (2 ms) > > Job: seqrw-40gb-4mb-t1 > ????????????Write: 161 MB/s (25 ms) > ????????????Read: 2083 MB/s (1 ms) > > Job: seqrw-20gb-4mb-t2 > ????????????Write: 352 MB/s (23 ms) > ????????????Read: 4317 MB/s (2 ms) > > Job: seqrw-10gb-4mb-t4 > ????????????Write: 696 MB/s (23 ms) > ????????????Read: 7358 MB/s (2 ms) > > Job: seqrw-05gb-4mb-t8 > ????????????Write: 1251 MB/s (25 ms) > ????????????Read: 6707 MB/s (5 ms) > > > So with fio I get a very nice read speed, but the write is horrendous > and I cannot find what causes it. I have looked at affinity settings > for the mmfsd process but not sure I fully understand it. But no > matter what I set it to, I see no difference. > > I have "played" with the bios and tried with/without hyperthreading, > numa and so on. And nothing affects atleast the blackmagic disk speed > test. > > the current settings for this host is like below. I write "current" > because I have tested a few different settings here but nothing > affects the write speed. maxTcpConnsPerNodeConn for sure bumped the > read speed though. > > nsdMaxWorkerThreads 16 > prefetchPct 60 > maxTcpConnsPerNodeConn 8 > maxMBpS 14000 > > > Does anyone have any suggestions or ideas on how to troubleshoot this? > > Thanks > > > > -- > > Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */? OnePost > *(formerly?Filmlance Post) > > ?? *OnePost*, formerly Filmlance's post-production, is now an > independent part of the Banijay Group. > New name, same team ? business as usual at OnePost. > > > > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email:uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 5814 bytes Desc: S/MIME Cryptographic Signature URL: From jonathan.buzzard at strath.ac.uk Tue Sep 3 19:51:21 2024 From: jonathan.buzzard at strath.ac.uk (Jonathan Buzzard) Date: Tue, 3 Sep 2024 19:51:21 +0100 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: Message-ID: <1bc4f6a3-a3fc-4fea-bb0a-0bf726733499@strath.ac.uk> On 03/09/2024 13:10, Henrik Cednert wrote: > Still no solution here regarding this. > > Have tested other cables. > Have tested to change tcp window size, no change > Played with numa in the bios, no change > Played with hyperthreading in bios, no change > Have you tried disabling the second CPU in the BIOS? You say you have played with NUMA in the BIOS, but how about ruling it out completely by going to a single CPU? JAB. -- Jonathan A. Buzzard Tel: +44141-5483420 HPC System Administrator, ARCHIE-WeSt. University of Strathclyde, John Anderson Building, Glasgow. G4 0NG From henrik.cednert at onepost.se Wed Sep 4 10:08:46 2024 From: henrik.cednert at onepost.se (Henrik Cednert) Date: Wed, 4 Sep 2024 09:08:46 +0000 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> Message-ID: Hi Uwe Thanks. Worth noting is that we have win10 ltsc, win 2019 and have had a single CPU win 11 22h2 (as a test) clients that all perform as expected. Those machines is older though and connected with 10-40GbE, those client max out their NIC in read and write. Let me know if i missed something important here. Thanks again. Setups is Client: * Supermicro Workstation * Intel(R) Xeon(R) Gold 6418H 2.10 GHz (2 processors) * Mellanox ConnectX-6 Dx connected with 100GbE over dedicated vlan via mellanox sn2100. * Windows 11 Pro for Workstations, 22H2 Storage setup * 3 x 84 bay seagate chassis with spinning disks. * Storage connected with redundant 12Gb SAS to 2 x Storage node servers * 2 x mellanox sn2100 * The 2 storage node servers are for this vlan connected with 100GbE to each switch, so in total 4 x 100GbE. And the switches are connected with 2 * 100GbE I tested the commands you suggested. They are both new to me so not sure what the output is supposed to be, looks like -iohistory isn't available in windows. I ran --waiters a few times, as seen below. Not sure what the expected output is from that. mmdiag --waiters === mmdiag: waiters === Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 18616 MsgHandler at getData: for In function sendMessage Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 25084 WritebehindWorkerThread: on ThCond 0x31A7C360 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 16780 FsyncHandlerThread: on ThCond 0x37FFDAB0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 30308 MsgHandler at getData: for In function sendMessage C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0055 sec since 2024-09-04_10:05:21, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x37A25FF0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0029 sec since 2024-09-04_10:05:23, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x38281DE0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0019 sec since 2024-09-04_10:05:25, monitored, thread 11832 PrefetchWorkerThread: on ThCond 0x38278D20 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 16780 AcquireBRTHandlerThread: on ThCond 0x37A324E0 (MsgRecordCondvar), reason 'RPC wait' for tmMsgBRRevoke on node 192.168.45.161 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 2576 RangeRevokeWorkerThread: on ThCond 0x5419DAA0 (BrlObjCondvar), reason 'waiting because of local byte range lock conflict' C:\Users\m5-tkd01> C:\Users\m5-tkd01>mmdiag --iohistory Unrecognized option: --iohistory. Run mmdiag --help for the option list -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Tuesday, 3 September 2024 17:35 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, Henrik, while I am not using Windows I'd start investigating the usual things (see below). But first you should describe your set-up better. Where are the NSDs : locally attached to the Windows box? In some NSD servers? If the latter -- what is the link to the NSD servers? via your GbE link? FC? IB? separate Ethernet? What type of storage? Spinning Disks? Flash? How long are your I/Os waiting on the client (compare that to the waiting times on the NSD server if applicable)? not sure whether that is available on Windows, but mmdiag --waiters mmdiag --iohistory might be of use. Somewhere in the chain from your application to the storage backend there is a delay and you should first find out where that occurs I think. Bye Uwe On 03.09.24 14:10, Henrik Cednert wrote: Still no solution here regarding this. Have tested other cables. Have tested to change tcp window size, no change Played with numa in the bios, no change Played with hyperthreading in bios, no change Have anyone managed to get some speed out of windows 11 and gpfs? -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Friday, 9 August 2024 17:25 To: gpfsug-discuss at gpfsug.org Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. VARNING: DETTA ?R ETT EXTERNT MAIL. Klicka inte p? n?gra l?nkar oavsett hur legitima de verkar utan att verifiera. Hello I have some issues with write performance on a windows 11 pro system and I'm out of ideas here. Hopefully someone here have some bright ideas and/or experience of GPFS on Windows 11? The system is a: Windows 11 Pro 22H2 2 x Intel(R) Xeon(R) Gold 6418H 2.10 GHz 512 GB RAM GPFS 5.1.9.4 Mellanox ConnectX 6 Dx 100GbE connected to Mellanox Switch with 5m Mellanox DAC. Before deploying this workstation we had a single socket system as a test bench where we got 60 GbE in both directons with iPerf and around 6GB/sec write and 3GB/sec read from the system over GPFS (fio tests, same tests as furhter down here). With that system I had loads of issues before getting to that point though. MS Defender had to be forcefully disabled via regedit some other tweaks. All those tweaks have been performed in this new system as well, but I can't get the proper speed out of it. On this new system and with iPerf to the storage servers I get around 50-60GbE in both directions and send and receive. If I mount the storage over SMB and 100GbE via the storage gateway servers I get around 3GB/sec read and write with Blackmagics Disk speed test. I have not tweaked the system for samba performande, just a test to see what it would give and part of the troubleshooting. If I run Blackmagics diskspeed test to the GPFS mount I instead get around 700MB/sec write and 400MB/sec read. Starting to think that the Blackmagic test might not run properly on this machine with these CPUs though. Or it's related to the mmfsd process maybe, how that threads or not threads...? But if we instead look at fio. I have a bat script that loops through a bunch of FIO-tests. A test that I have been using over the years so that we easily can benchmark all deployed systems with the exakt same tests. The tests are named like: seqrw-gb-mb-t The result when I run this is like the below list. Number in parenthesis is the by fio reported latency. Job: seqrw-40gb-1mb-t1 ????????????Write: 162 MB/s (6 ms) ????????????Read: 1940 MB/s (1 ms) Job: seqrw-20gb-1mb-t2 ????????????Write: 286 MB/s (7 ms) ????????????Read: 3952 MB/s (1 ms) Job: seqrw-10gb-1mb-t4 ????????????Write: 549 MB/s (7 ms) ????????????Read: 6987 MB/s (1 ms) Job: seqrw-05gb-1mb-t8 ????????????Write: 989 MB/s (8 ms) ????????????Read: 7721 MB/s (1 ms) Job: seqrw-40gb-2mb-t1 ????????????Write: 161 MB/s (12 ms) ????????????Read: 2261 MB/s (0 ms) Job: seqrw-20gb-2mb-t2 ????????????Write: 348 MB/s (11 ms) ????????????Read: 4266 MB/s (1 ms) Job: seqrw-10gb-2mb-t4 ????????????Write: 626 MB/s (13 ms) ????????????Read: 4949 MB/s (1 ms) Job: seqrw-05gb-2mb-t8 ????????????Write: 1154 MB/s (14 ms) ????????????Read: 7007 MB/s (2 ms) Job: seqrw-40gb-4mb-t1 ????????????Write: 161 MB/s (25 ms) ????????????Read: 2083 MB/s (1 ms) Job: seqrw-20gb-4mb-t2 ????????????Write: 352 MB/s (23 ms) ????????????Read: 4317 MB/s (2 ms) Job: seqrw-10gb-4mb-t4 ????????????Write: 696 MB/s (23 ms) ????????????Read: 7358 MB/s (2 ms) Job: seqrw-05gb-4mb-t8 ????????????Write: 1251 MB/s (25 ms) ????????????Read: 6707 MB/s (5 ms) So with fio I get a very nice read speed, but the write is horrendous and I cannot find what causes it. I have looked at affinity settings for the mmfsd process but not sure I fully understand it. But no matter what I set it to, I see no difference. I have "played" with the bios and tried with/without hyperthreading, numa and so on. And nothing affects atleast the blackmagic disk speed test. the current settings for this host is like below. I write "current" because I have tested a few different settings here but nothing affects the write speed. maxTcpConnsPerNodeConn for sure bumped the read speed though. nsdMaxWorkerThreads 16 prefetchPct 60 maxTcpConnsPerNodeConn 8 maxMBpS 14000 Does anyone have any suggestions or ideas on how to troubleshoot this? Thanks -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: From henrik.cednert at onepost.se Wed Sep 4 10:16:23 2024 From: henrik.cednert at onepost.se (Henrik Cednert) Date: Wed, 4 Sep 2024 09:16:23 +0000 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: <1bc4f6a3-a3fc-4fea-bb0a-0bf726733499@strath.ac.uk> References: <1bc4f6a3-a3fc-4fea-bb0a-0bf726733499@strath.ac.uk> Message-ID: Hi Jonathan I looked in the supermicro bios. Sadly there are no options there for diabling a CPU. All you can do is disable CPU cores but from the looks of it you need a minimum of 1 available for a CPU. I will investigate that furhter and play with it. -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Jonathan Buzzard Sent: Tuesday, 3 September 2024 20:51 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. On 03/09/2024 13:10, Henrik Cednert wrote: > Still no solution here regarding this. > > Have tested other cables. > Have tested to change tcp window size, no change > Played with numa in the bios, no change > Played with hyperthreading in bios, no change > Have you tried disabling the second CPU in the BIOS? You say you have played with NUMA in the BIOS, but how about ruling it out completely by going to a single CPU? JAB. -- Jonathan A. Buzzard Tel: +44141-5483420 HPC System Administrator, ARCHIE-WeSt. University of Strathclyde, John Anderson Building, Glasgow. G4 0NG _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: From uwe.falke at kit.edu Wed Sep 4 11:59:03 2024 From: uwe.falke at kit.edu (Uwe Falke) Date: Wed, 4 Sep 2024 12:59:03 +0200 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> Message-ID: Hi, given you see read latencies of 1 ms, you do not get the data from disk but from some cache (on whatever level). From spinning disks, you can never expect such read latencies (mind that GPFS block reading, even if sequential from the application's PoV, typically translates to random I/O at the physical disk level). So, I do not know what the latencies on your other measurements (bypassing GPFS) were? but the numbers below do not represent sustained high-scale throughputs, apparently. It is nevertheless strange that your write rates are much below reads (and write latencies are that high) -- from my experience with different systems, when hammering GPFS with usual storage backends with both read and write requests, the writes tend to prevail. Your waiters indicate that the problem is above GPFS: GPFS is able to serve all I/O threads within a few ms, and there is not a long list of pending IOs. Sorry, the iohistory options is mmdiag --iohist But to me it looks like not GPFS is the culprit here. Uwe On 04.09.24 11:08, Henrik Cednert wrote: > Hi Uwe > > Thanks. > > Worth noting is that we have win10 ltsc, win 2019 and have had a > single CPU win 11 22h2 (as a test) clients that all perform as > expected. Those machines is older though and connected with 10-40GbE, > those client max out their NIC in read and write. > > Let me know if i missed something important here. Thanks again. > > Setups is > > Client: > > * Supermicro Workstation > * Intel(R) Xeon(R) Gold 6418H ? 2.10 GHz ?(2 processors) > * Mellanox ConnectX-6 Dx connected with 100GbE over dedicated vlan > via mellanox sn2100. > * Windows 11 Pro for Workstations, 22H2 > > > > Storage setup > > * > 3 x 84 bay seagate chassis with spinning disks. > * > Storage connected with redundant 12Gb SAS to 2 x Storage node servers > * > 2 x mellanox sn2100 > * > The 2 storage node servers are for this vlan connected with 100GbE > to each switch, so in total 4 x 100GbE.? And the switches are > connected with 2 * 100GbE > > > > I tested the commands you suggested. They are both new to me so not > sure what the output is supposed to be, looks like -iohistory isn't > available in windows. I ran --waiters a few times, as seen below. Not > sure what the expected output is from that. > > > mmdiag --waiters > > === mmdiag: waiters === > Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 18616 > MsgHandler at getData: for In function sendMessage > Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 25084 > WritebehindWorkerThread: on ThCond 0x31A7C360 (MsgRecordCondvar), > reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 > > C:\Users\m5-tkd01>mmdiag --waiters > > === mmdiag: waiters === > Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 16780 > FsyncHandlerThread: on ThCond 0x37FFDAB0 (MsgRecordCondvar), reason > 'RPC wait' for NSD I/O completion on node 192.168.45.214 > Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 30308 > MsgHandler at getData: for In function sendMessage > > C:\Users\m5-tkd01>mmdiag --waiters > > === mmdiag: waiters === > Waiting 0.0055 sec since 2024-09-04_10:05:21, monitored, thread 16780 > FileBlockReadFetchHandlerThread: on ThCond 0x37A25FF0 > (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node > 192.168.45.213 > > C:\Users\m5-tkd01>mmdiag --waiters > > === mmdiag: waiters === > Waiting 0.0029 sec since 2024-09-04_10:05:23, monitored, thread 16780 > FileBlockReadFetchHandlerThread: on ThCond 0x38281DE0 > (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node > 192.168.45.213 > > C:\Users\m5-tkd01>mmdiag --waiters > > === mmdiag: waiters === > Waiting 0.0019 sec since 2024-09-04_10:05:25, monitored, thread 11832 > PrefetchWorkerThread: on ThCond 0x38278D20 (MsgRecordCondvar), reason > 'RPC wait' for NSD I/O completion on node 192.168.45.214 > Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 16780 > AcquireBRTHandlerThread: on ThCond 0x37A324E0 (MsgRecordCondvar), > reason 'RPC wait' for tmMsgBRRevoke on node 192.168.45.161 > Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 2576 > RangeRevokeWorkerThread: on ThCond 0x5419DAA0 (BrlObjCondvar), reason > 'waiting because of local byte range lock conflict' > > C:\Users\m5-tkd01> > > > > > > > C:\Users\m5-tkd01>mmdiag --iohistory > Unrecognized option: --iohistory. > Run mmdiag --help for the option list > > > > > -- > > Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */? OnePost > *(formerly?Filmlance Post) > > ?? *OnePost*, formerly Filmlance's post-production, is now an > independent part of the Banijay Group. > New name, same team ? business as usual at OnePost. > > > > ------------------------------------------------------------------------ > *From:* gpfsug-discuss on behalf > of Uwe Falke > *Sent:* Tuesday, 3 September 2024 17:35 > *To:* gpfsug-discuss at gpfsug.org > *Subject:* Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. > Performance issues, write. > > Hi, Henrik, > > > while I am not using Windows I'd start investigating the usual things > (see below). > > > But first you should describe your set-up better. > > Where are the NSDs : locally attached to the Windows box? In some NSD > servers? > > If the latter -- what is the link to the NSD servers? via your GbE > link? FC? IB? separate Ethernet? > > What type of storage? Spinning Disks? Flash? > > > How long are your I/Os waiting on the client (compare that to the > waiting times on the NSD server if applicable)? > > not sure whether that is available on Windows, but > > mmdiag --waiters > > mmdiag --iohistory > > might be of use. > > > Somewhere in the chain from your application to the storage backend > there is a delay and you should first find out where that occurs I think. > > > Bye > > Uwe > > > > On 03.09.24 14:10, Henrik Cednert wrote: >> Still no solution here regarding this. >> >> Have tested other cables. >> Have tested to change tcp window size, no change >> Played with numa in the bios, no change >> Played with hyperthreading in bios, no change >> >> >> Have anyone managed to get some speed out of windows 11 and gpfs? >> >> >> -- >> >> Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */? OnePost >> *(formerly?Filmlance Post) >> >> ?? *OnePost*, formerly Filmlance's post-production, is now an >> independent part of the Banijay Group. >> New name, same team ? business as usual at OnePost. >> >> >> >> ------------------------------------------------------------------------ >> *From:* gpfsug-discuss >> on behalf of Henrik >> Cednert >> *Sent:* Friday, 9 August 2024 17:25 >> *To:* gpfsug-discuss at gpfsug.org >> >> *Subject:* [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. >> Performance issues, write. >> >> ***VARNING: DETTA ?R ETT EXTERNT MAIL. Klicka inte p? n?gra l?nkar >> oavsett hur legitima de verkar utan att verifiera.* >> >> >> Hello >> >> I have some issues with write performance on a windows 11 pro system >> and I'm out of ideas here. Hopefully someone here have some bright >> ideas and/or experience of GPFS on Windows 11? >> >> The system is a: >> >> Windows 11 Pro 22H2 >> 2 x Intel(R) Xeon(R) Gold 6418H ? 2.10 GHz >> 512 GB RAM >> GPFS 5.1.9.4 >> Mellanox ConnectX 6 Dx >> 100GbE connected to Mellanox Switch with 5m Mellanox DAC. >> >> Before deploying this workstation we had a single socket system as a >> test bench where we got 60 GbE in both directons with iPerf and >> around 6GB/sec write and 3GB/sec read from the system over GPFS (fio >> tests, same tests as furhter down here). >> >> With that system I had loads of issues before getting to that point >> though. MS Defender had to be forcefully disabled via regedit some >> other tweaks. All those tweaks have been performed in this new system >> as well, but I can't get the proper speed out of it. >> >> >> On this new system and with iPerf to the storage servers I get around >> 50-60GbE in both directions and send and receive. >> >> If I mount the storage over SMB and 100GbE via the storage gateway >> servers I get around 3GB/sec read and write with Blackmagics Disk >> speed test. I have not tweaked the system for samba performande, just >> a test to see what it would give and part of the troubleshooting. >> >> If I run Blackmagics diskspeed test to the GPFS mount I instead get >> around 700MB/sec write and 400MB/sec read. >> >> Starting to think that the Blackmagic test might not run properly on >> this machine with these CPUs though. Or it's related to the mmfsd >> process maybe, how that threads or not threads...? >> >> But if we instead look at fio. I have a bat script that loops through >> a bunch of FIO-tests. A test that I have been using over the years so >> that we easily can benchmark all deployed systems with the exakt same >> tests. The tests are named like: >> >> seqrw-gb-mb-t >> >> The result when I run this is like the below list. Number in >> parenthesis is the by fio reported latency. >> >> Job: seqrw-40gb-1mb-t1 >> ????????????Write: 162 MB/s (6 ms) >> ????????????Read: 1940 MB/s (1 ms) >> >> Job: seqrw-20gb-1mb-t2 >> ????????????Write: 286 MB/s (7 ms) >> ????????????Read: 3952 MB/s (1 ms) >> >> Job: seqrw-10gb-1mb-t4 >> ????????????Write: 549 MB/s (7 ms) >> ????????????Read: 6987 MB/s (1 ms) >> >> Job: seqrw-05gb-1mb-t8 >> ????????????Write: 989 MB/s (8 ms) >> ????????????Read: 7721 MB/s (1 ms) >> >> Job: seqrw-40gb-2mb-t1 >> ????????????Write: 161 MB/s (12 ms) >> ????????????Read: 2261 MB/s (0 ms) >> >> Job: seqrw-20gb-2mb-t2 >> ????????????Write: 348 MB/s (11 ms) >> ????????????Read: 4266 MB/s (1 ms) >> >> Job: seqrw-10gb-2mb-t4 >> ????????????Write: 626 MB/s (13 ms) >> ????????????Read: 4949 MB/s (1 ms) >> >> Job: seqrw-05gb-2mb-t8 >> ????????????Write: 1154 MB/s (14 ms) >> ????????????Read: 7007 MB/s (2 ms) >> >> Job: seqrw-40gb-4mb-t1 >> ????????????Write: 161 MB/s (25 ms) >> ????????????Read: 2083 MB/s (1 ms) >> >> Job: seqrw-20gb-4mb-t2 >> ????????????Write: 352 MB/s (23 ms) >> ????????????Read: 4317 MB/s (2 ms) >> >> Job: seqrw-10gb-4mb-t4 >> ????????????Write: 696 MB/s (23 ms) >> ????????????Read: 7358 MB/s (2 ms) >> >> Job: seqrw-05gb-4mb-t8 >> ????????????Write: 1251 MB/s (25 ms) >> ????????????Read: 6707 MB/s (5 ms) >> >> >> So with fio I get a very nice read speed, but the write is horrendous >> and I cannot find what causes it. I have looked at affinity settings >> for the mmfsd process but not sure I fully understand it. But no >> matter what I set it to, I see no difference. >> >> I have "played" with the bios and tried with/without hyperthreading, >> numa and so on. And nothing affects atleast the blackmagic disk speed >> test. >> >> the current settings for this host is like below. I write "current" >> because I have tested a few different settings here but nothing >> affects the write speed. maxTcpConnsPerNodeConn for sure bumped the >> read speed though. >> >> nsdMaxWorkerThreads 16 >> prefetchPct 60 >> maxTcpConnsPerNodeConn 8 >> maxMBpS 14000 >> >> >> Does anyone have any suggestions or ideas on how to troubleshoot this? >> >> Thanks >> >> >> >> -- >> >> Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */? OnePost >> *(formerly?Filmlance Post) >> >> ?? *OnePost*, formerly Filmlance's post-production, is now an >> independent part of the Banijay Group. >> New name, same team ? business as usual at OnePost. >> >> >> >> >> _______________________________________________ >> gpfsug-discuss mailing list >> gpfsug-discuss at gpfsug.org >> http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org > -- > Karlsruhe Institute of Technology (KIT) > Scientific Computing Centre (SCC) > Scientific Data Management (SDM) > > Uwe Falke > > Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 > D-76344 Eggenstein-Leopoldshafen > > Tel: +49 721 608 28024 > Email:uwe.falke at kit.edu > www.scc.kit.edu > > Registered office: > Kaiserstra?e 12, 76131 Karlsruhe, Germany > > KIT ? The Research University in the Helmholtz Association > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email:uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 5814 bytes Desc: S/MIME Cryptographic Signature URL: From henrik.cednert at onepost.se Wed Sep 4 12:24:55 2024 From: henrik.cednert at onepost.se (Henrik Cednert) Date: Wed, 4 Sep 2024 11:24:55 +0000 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> Message-ID: Hello My theory is that it's windows 11 related, or the combo windows 11 and this hardware. I guess the only thing to know for sure, is to boot it into *nix and install gpfs there and test. Which I guess isn't the worst of ideas at this stage. --iohist spits out a pretty hefty chuck of data. Below is a snippet from when I did a write test: mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 13:17:04.088621 W data 10:2178073088 10800 3,994 cli C0A82DD5:63877BD2 192.168.45.214 13:17:04.092614 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.092614 W inode 2:3040332814 1 0,999 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.093613 W data 10:2178083888 5072 2,995 cli C0A82DD5:63877BD2 192.168.45.214 13:17:04.094611 W data 18:2178072576 16384 6,502 cli C0A82DD5:63877BDB 192.168.45.214 13:17:04.101113 W logData 2:1869321537 2 0,998 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.101113 W logData 3:1877542209 2 0,998 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.103109 W data 18:2178078304 10656 3,994 cli C0A82DD5:63877BDB 192.168.45.214 13:17:04.103109 W data 26:2178072576 16384 6,529 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.109638 W logData 2:1869321538 2 0,994 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.109638 W logData 3:1877542210 2 0,994 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.111634 W data 26:2178072720 10800 3,992 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.115626 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.115626 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.116626 W data 26:2178083520 5440 2,995 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.117629 W data 11:2178072576 16384 4,987 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.122616 W logData 2:1869321539 2 0,999 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.122616 W logData 3:1877542211 2 0,999 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.124613 W data 11:2178077936 10800 3,999 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.128612 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.128612 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.129609 W data 11:2178088736 224 0,000 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.129609 W data 19:2178072576 16384 6,495 cli C0A82DD5:63877BDC 192.168.45.213 13:17:04.136104 W logData 2:1869321540 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.136104 W logData 3:1877542212 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.137103 W data 19:2178083152 5808 2,999 cli C0A82DD5:63877BDC 192.168.45.213 13:17:04.138100 W data 27:2178072576 16384 5,990 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.144091 W logData 2:1869321540 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.144091 W logData 3:1877542212 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.146088 W data 27:2178077568 10800 4,005 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.150092 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.150092 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.151092 W data 27:2178088368 592 0,000 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.151092 W data 12:2178072576 16384 4,995 cli C0A82DD5:63877BD4 192.168.45.214 13:17:04.156086 W logData 2:1869321541 2 0,996 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.156086 W logData 3:1877542213 2 0,996 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.157082 W data 12:2178082784 6176 2,994 cli C0A82DD5:63877BD4 192.168.45.214 13:17:04.158083 W data 20:2178072576 16384 7,498 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.165581 W logData 2:1869321542 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.165581 W logData 3:1877542214 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.167578 W data 20:2178077200 10800 2,994 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.170572 W inode 4:3060742158 1 0,996 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.171568 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.171568 W data 20:2178088000 960 1,001 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.172569 W data 28:2864381952 16384 5,988 cli C0A82DD5:63877BE5 192.168.45.214 13:17:04.178557 W logData 2:1869321543 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.178557 W logData 3:1877542215 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.179560 W data 28:2864391792 6544 2,995 cli C0A82DD5:63877BE5 192.168.45.214 13:17:04.179560 W data 5:2406842368 16384 5,987 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.185548 W logData 2:1869321544 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.185548 W logData 3:1877542216 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.186549 W data 5:2406846624 10800 4,993 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.191542 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.191542 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.192540 W data 5:2406857424 1328 0,995 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.193535 W data 13:2406842368 16384 6,019 cli C0A82DD5:63877BD5 192.168.45.213 13:17:04.199554 W logData 2:1869321544 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.199554 W logData 3:1877542216 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.200551 W data 13:2406851840 6912 2,997 cli C0A82DD5:63877BD5 192.168.45.213 13:17:04.201554 W data 21:2406842368 16384 5,912 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.207466 W logData 2:1869321545 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.207466 W logData 3:1877542217 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.208465 W data 21:2406846256 10800 3,990 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.212456 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.212456 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.213456 W data 21:2406857056 1696 1,998 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.214457 W data 6:2406842368 16384 5,015 cli C0A82DD5:63877BCE 192.168.45.214 13:17:04.219472 W logData 2:1869321546 2 1,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.219472 W logData 3:1877542218 2 1,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.221474 W data 6:2406851472 7280 1,994 cli C0A82DD5:63877BCE 192.168.45.214 13:17:04.221474 W data 14:2406842368 16384 7,502 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.228976 W logData 2:1869321547 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.228976 W logData 3:1877542219 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.229971 W data 14:2406845888 10800 3,994 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.233965 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.233965 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.234963 W data 14:2406856688 2064 1,999 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.235965 W data 22:2406842368 16384 4,993 cli C0A82DD5:63877BDF 192.168.45.214 13:17:04.240958 W logData 2:1869321547 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.240958 W logData 3:1877542219 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.241974 W data 22:2406851104 7648 2,531 cli C0A82DD5:63877BDF 192.168.45.214 13:17:04.242977 W data 7:2406842368 16384 5,526 cli C0A82DD5:63877BCF 192.168.45.213 13:17:04.248503 W logData 2:1869321548 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.248503 W logData 3:1877542220 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.249503 W data 7:2406845520 10800 3,994 cli C0A82DD5:63877BCF 192.168.45.213 13:17:04.253497 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Wednesday, 4 September 2024 12:59 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, given you see read latencies of 1 ms, you do not get the data from disk but from some cache (on whatever level). From spinning disks, you can never expect such read latencies (mind that GPFS block reading, even if sequential from the application's PoV, typically translates to random I/O at the physical disk level). So, I do not know what the latencies on your other measurements (bypassing GPFS) were but the numbers below do not represent sustained high-scale throughputs, apparently. It is nevertheless strange that your write rates are much below reads (and write latencies are that high) -- from my experience with different systems, when hammering GPFS with usual storage backends with both read and write requests, the writes tend to prevail. Your waiters indicate that the problem is above GPFS: GPFS is able to serve all I/O threads within a few ms, and there is not a long list of pending IOs. Sorry, the iohistory options is mmdiag --iohist But to me it looks like not GPFS is the culprit here. Uwe On 04.09.24 11:08, Henrik Cednert wrote: Hi Uwe Thanks. Worth noting is that we have win10 ltsc, win 2019 and have had a single CPU win 11 22h2 (as a test) clients that all perform as expected. Those machines is older though and connected with 10-40GbE, those client max out their NIC in read and write. Let me know if i missed something important here. Thanks again. Setups is Client: * Supermicro Workstation * Intel(R) Xeon(R) Gold 6418H 2.10 GHz (2 processors) * Mellanox ConnectX-6 Dx connected with 100GbE over dedicated vlan via mellanox sn2100. * Windows 11 Pro for Workstations, 22H2 Storage setup * 3 x 84 bay seagate chassis with spinning disks. * Storage connected with redundant 12Gb SAS to 2 x Storage node servers * 2 x mellanox sn2100 * The 2 storage node servers are for this vlan connected with 100GbE to each switch, so in total 4 x 100GbE. And the switches are connected with 2 * 100GbE I tested the commands you suggested. They are both new to me so not sure what the output is supposed to be, looks like -iohistory isn't available in windows. I ran --waiters a few times, as seen below. Not sure what the expected output is from that. mmdiag --waiters === mmdiag: waiters === Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 18616 MsgHandler at getData: for In function sendMessage Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 25084 WritebehindWorkerThread: on ThCond 0x31A7C360 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 16780 FsyncHandlerThread: on ThCond 0x37FFDAB0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 30308 MsgHandler at getData: for In function sendMessage C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0055 sec since 2024-09-04_10:05:21, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x37A25FF0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0029 sec since 2024-09-04_10:05:23, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x38281DE0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0019 sec since 2024-09-04_10:05:25, monitored, thread 11832 PrefetchWorkerThread: on ThCond 0x38278D20 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 16780 AcquireBRTHandlerThread: on ThCond 0x37A324E0 (MsgRecordCondvar), reason 'RPC wait' for tmMsgBRRevoke on node 192.168.45.161 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 2576 RangeRevokeWorkerThread: on ThCond 0x5419DAA0 (BrlObjCondvar), reason 'waiting because of local byte range lock conflict' C:\Users\m5-tkd01> C:\Users\m5-tkd01>mmdiag --iohistory Unrecognized option: --iohistory. Run mmdiag --help for the option list -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Tuesday, 3 September 2024 17:35 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, Henrik, while I am not using Windows I'd start investigating the usual things (see below). But first you should describe your set-up better. Where are the NSDs : locally attached to the Windows box? In some NSD servers? If the latter -- what is the link to the NSD servers? via your GbE link? FC? IB? separate Ethernet? What type of storage? Spinning Disks? Flash? How long are your I/Os waiting on the client (compare that to the waiting times on the NSD server if applicable)? not sure whether that is available on Windows, but mmdiag --waiters mmdiag --iohistory might be of use. Somewhere in the chain from your application to the storage backend there is a delay and you should first find out where that occurs I think. Bye Uwe On 03.09.24 14:10, Henrik Cednert wrote: Still no solution here regarding this. Have tested other cables. Have tested to change tcp window size, no change Played with numa in the bios, no change Played with hyperthreading in bios, no change Have anyone managed to get some speed out of windows 11 and gpfs? -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Friday, 9 August 2024 17:25 To: gpfsug-discuss at gpfsug.org Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. VARNING: DETTA ?R ETT EXTERNT MAIL. Klicka inte p? n?gra l?nkar oavsett hur legitima de verkar utan att verifiera. Hello I have some issues with write performance on a windows 11 pro system and I'm out of ideas here. Hopefully someone here have some bright ideas and/or experience of GPFS on Windows 11? The system is a: Windows 11 Pro 22H2 2 x Intel(R) Xeon(R) Gold 6418H 2.10 GHz 512 GB RAM GPFS 5.1.9.4 Mellanox ConnectX 6 Dx 100GbE connected to Mellanox Switch with 5m Mellanox DAC. Before deploying this workstation we had a single socket system as a test bench where we got 60 GbE in both directons with iPerf and around 6GB/sec write and 3GB/sec read from the system over GPFS (fio tests, same tests as furhter down here). With that system I had loads of issues before getting to that point though. MS Defender had to be forcefully disabled via regedit some other tweaks. All those tweaks have been performed in this new system as well, but I can't get the proper speed out of it. On this new system and with iPerf to the storage servers I get around 50-60GbE in both directions and send and receive. If I mount the storage over SMB and 100GbE via the storage gateway servers I get around 3GB/sec read and write with Blackmagics Disk speed test. I have not tweaked the system for samba performande, just a test to see what it would give and part of the troubleshooting. If I run Blackmagics diskspeed test to the GPFS mount I instead get around 700MB/sec write and 400MB/sec read. Starting to think that the Blackmagic test might not run properly on this machine with these CPUs though. Or it's related to the mmfsd process maybe, how that threads or not threads...? But if we instead look at fio. I have a bat script that loops through a bunch of FIO-tests. A test that I have been using over the years so that we easily can benchmark all deployed systems with the exakt same tests. The tests are named like: seqrw-gb-mb-t The result when I run this is like the below list. Number in parenthesis is the by fio reported latency. Job: seqrw-40gb-1mb-t1 ????????????Write: 162 MB/s (6 ms) ????????????Read: 1940 MB/s (1 ms) Job: seqrw-20gb-1mb-t2 ????????????Write: 286 MB/s (7 ms) ????????????Read: 3952 MB/s (1 ms) Job: seqrw-10gb-1mb-t4 ????????????Write: 549 MB/s (7 ms) ????????????Read: 6987 MB/s (1 ms) Job: seqrw-05gb-1mb-t8 ????????????Write: 989 MB/s (8 ms) ????????????Read: 7721 MB/s (1 ms) Job: seqrw-40gb-2mb-t1 ????????????Write: 161 MB/s (12 ms) ????????????Read: 2261 MB/s (0 ms) Job: seqrw-20gb-2mb-t2 ????????????Write: 348 MB/s (11 ms) ????????????Read: 4266 MB/s (1 ms) Job: seqrw-10gb-2mb-t4 ????????????Write: 626 MB/s (13 ms) ????????????Read: 4949 MB/s (1 ms) Job: seqrw-05gb-2mb-t8 ????????????Write: 1154 MB/s (14 ms) ????????????Read: 7007 MB/s (2 ms) Job: seqrw-40gb-4mb-t1 ????????????Write: 161 MB/s (25 ms) ????????????Read: 2083 MB/s (1 ms) Job: seqrw-20gb-4mb-t2 ????????????Write: 352 MB/s (23 ms) ????????????Read: 4317 MB/s (2 ms) Job: seqrw-10gb-4mb-t4 ????????????Write: 696 MB/s (23 ms) ????????????Read: 7358 MB/s (2 ms) Job: seqrw-05gb-4mb-t8 ????????????Write: 1251 MB/s (25 ms) ????????????Read: 6707 MB/s (5 ms) So with fio I get a very nice read speed, but the write is horrendous and I cannot find what causes it. I have looked at affinity settings for the mmfsd process but not sure I fully understand it. But no matter what I set it to, I see no difference. I have "played" with the bios and tried with/without hyperthreading, numa and so on. And nothing affects atleast the blackmagic disk speed test. the current settings for this host is like below. I write "current" because I have tested a few different settings here but nothing affects the write speed. maxTcpConnsPerNodeConn for sure bumped the read speed though. nsdMaxWorkerThreads 16 prefetchPct 60 maxTcpConnsPerNodeConn 8 maxMBpS 14000 Does anyone have any suggestions or ideas on how to troubleshoot this? Thanks -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: From henrik.cednert at onepost.se Wed Sep 4 12:28:37 2024 From: henrik.cednert at onepost.se (Henrik Cednert) Date: Wed, 4 Sep 2024 11:28:37 +0000 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> Message-ID: Adding a snippet from iohist when reading as well. >mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 13:24:59.180277 R data 19:1491763200 16384 32,433 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.212710 R data 8:805453824 16384 5,987 cli C0A82DD5:63877BD0 192.168.45.214 13:24:59.220698 R data 8:805453824 16384 6,012 cli C0A82DD5:63877BD0 192.168.45.214 13:24:59.226710 R data 16:805453824 16384 9,129 cli C0A82DD5:63877BD9 192.168.45.214 13:24:59.226710 R data 12:1491763200 16384 18,526 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.226710 R data 20:2178072576 16384 23,520 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.251229 R data 16:805453824 16384 5,497 cli C0A82DD5:63877BD9 192.168.45.214 13:24:59.257730 R data 24:805453824 16384 5,990 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.257730 R data 5:1720532992 16384 33,458 cli C0A82DD5:63877BCD 192.168.45.213 13:24:59.257730 R data 28:1720532992 16384 33,458 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.292189 R data 24:805453824 16384 4,992 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.300176 R data 24:805453824 16384 4,991 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.306169 R data 9:805453824 16384 7,987 cli C0A82DD5:63877BD1 192.168.45.213 13:24:59.306169 R data 13:1720532992 16384 25,471 cli C0A82DD5:63877BD5 192.168.45.213 13:24:59.306169 R data 21:1720532992 16384 28,466 cli C0A82DD5:63877BDE 192.168.45.213 13:24:59.335637 R data 9:805453824 16384 4,994 cli C0A82DD5:63877BD1 192.168.45.213 13:24:59.341629 R data 17:805453824 16384 3,993 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.341629 R data 6:1720532992 16384 27,471 cli C0A82DD5:63877BCE 192.168.45.214 13:24:59.341629 R data 14:1720532992 16384 44,953 cli C0A82DD5:63877BD6 192.168.45.214 13:24:59.387584 R data 17:805453824 16384 5,990 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.394572 R data 17:805453824 16384 7,993 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.402565 R data 25:805453824 16384 4,991 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.402565 R data 22:1720532992 16384 26,309 cli C0A82DD5:63877BDF 192.168.45.214 13:24:59.402565 R data 7:1720532992 16384 142,856 cli C0A82DD5:63877BCF 192.168.45.213 13:24:59.546419 R data 25:805453824 16384 7,992 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.556408 R data 25:805453824 16384 7,987 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.564395 R data 10:805453824 16384 5,505 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.564395 R data 23:1720532992 16384 28,053 cli C0A82DD5:63877BE0 192.168.45.213 13:24:59.564395 R data 15:1720532992 16384 33,044 cli C0A82DD5:63877BD8 192.168.45.213 13:24:59.598437 R data 10:805453824 16384 5,504 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.604939 R data 18:805453824 16384 4,993 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.604939 R data 8:1720532992 16384 36,015 cli C0A82DD5:63877BD0 192.168.45.214 13:24:59.609932 R data 16:1720532992 16384 42,010 cli C0A82DD5:63877BD9 192.168.45.214 13:24:59.652940 R data 18:805453824 16384 7,989 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.662434 R data 18:805453824 16384 6,994 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.669428 R data 26:805453824 16384 8,986 cli C0A82DD5:63877BE3 192.168.45.214 13:24:59.669428 R data 24:1720532992 16384 20,308 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.669428 R data 9:1720532992 16384 25,812 cli C0A82DD5:63877BD1 192.168.45.213 13:24:59.696239 R data 26:805453824 16384 6,989 cli C0A82DD5:63877BE3 192.168.45.214 13:24:59.703228 R data 11:805453824 16384 4,992 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.703228 R data 25:1720532992 16384 17,976 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.703228 R data 17:1720532992 16384 22,481 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.725709 R data 11:805453824 16384 4,992 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.731699 R data 11:805453824 16384 8,986 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.740685 R data 19:805453824 16384 4,997 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.740685 R data 18:1720532992 16384 19,486 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.740685 R data 10:1720532992 16384 27,474 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.769157 R data 19:805453824 16384 4,997 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.774154 R data 27:805453824 16384 4,992 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.774154 R data 11:1720532992 16384 22,476 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.774154 R data 26:1720532992 16384 29,464 cli C0A82DD5:63877BE3 192.168.45.214 13:24:59.803618 R data 27:805453824 16384 4,997 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.809614 R data 27:805453824 16384 7,987 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.817601 R data 12:805453824 16384 8,500 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.817601 R data 27:1720532992 16384 9,499 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.817601 R data 19:1720532992 16384 74,075 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.892674 R data 12:805453824 16384 4,997 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.897671 R data 20:1491763200 16384 4,992 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.898670 R data 12:1720532992 16384 24,479 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.898670 R data 20:2406842368 16384 26,476 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.925146 R data 20:1491763200 16384 8,990 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.935430 R data 20:1491763200 16384 4,691 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.940121 R data 28:1034223616 16384 4,312 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.940121 R data 28:1949302784 16384 26,664 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.940121 R data 5:1949302784 16384 34,831 cli C0A82DD5:63877BCD 192.168.45.213 13:24:59.975955 R data 28:1034223616 16384 5,583 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.981539 R data 5:1034223616 16384 5,168 cli C0A82DD5:63877BCD 192.168.45.213 13:24:59.981539 R data 21:1949302784 16384 25,392 cli C0A82DD5:63877BDE 192.168.45.213 13:24:59.981539 R data 13:1949302784 16384 34,412 cli C0A82DD5:63877BD5 192.168.45.213 13:25:00.016950 R data 5:1034223616 16384 7,991 cli C0A82DD5:63877BCD 192.168.45.213 13:25:00.026938 R data 5:1034223616 16384 7,750 cli C0A82DD5:63877BCD 192.168.45.213 13:25:00.034688 R data 13:1034223616 16384 5,498 cli C0A82DD5:63877BD5 192.168.45.213 13:25:00.034688 R data 6:1949302784 16384 27,661 cli C0A82DD5:63877BCE 192.168.45.214 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Wednesday, 4 September 2024 13:24 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hello My theory is that it's windows 11 related, or the combo windows 11 and this hardware. I guess the only thing to know for sure, is to boot it into *nix and install gpfs there and test. Which I guess isn't the worst of ideas at this stage. --iohist spits out a pretty hefty chuck of data. Below is a snippet from when I did a write test: mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 13:17:04.088621 W data 10:2178073088 10800 3,994 cli C0A82DD5:63877BD2 192.168.45.214 13:17:04.092614 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.092614 W inode 2:3040332814 1 0,999 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.093613 W data 10:2178083888 5072 2,995 cli C0A82DD5:63877BD2 192.168.45.214 13:17:04.094611 W data 18:2178072576 16384 6,502 cli C0A82DD5:63877BDB 192.168.45.214 13:17:04.101113 W logData 2:1869321537 2 0,998 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.101113 W logData 3:1877542209 2 0,998 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.103109 W data 18:2178078304 10656 3,994 cli C0A82DD5:63877BDB 192.168.45.214 13:17:04.103109 W data 26:2178072576 16384 6,529 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.109638 W logData 2:1869321538 2 0,994 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.109638 W logData 3:1877542210 2 0,994 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.111634 W data 26:2178072720 10800 3,992 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.115626 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.115626 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.116626 W data 26:2178083520 5440 2,995 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.117629 W data 11:2178072576 16384 4,987 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.122616 W logData 2:1869321539 2 0,999 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.122616 W logData 3:1877542211 2 0,999 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.124613 W data 11:2178077936 10800 3,999 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.128612 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.128612 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.129609 W data 11:2178088736 224 0,000 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.129609 W data 19:2178072576 16384 6,495 cli C0A82DD5:63877BDC 192.168.45.213 13:17:04.136104 W logData 2:1869321540 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.136104 W logData 3:1877542212 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.137103 W data 19:2178083152 5808 2,999 cli C0A82DD5:63877BDC 192.168.45.213 13:17:04.138100 W data 27:2178072576 16384 5,990 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.144091 W logData 2:1869321540 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.144091 W logData 3:1877542212 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.146088 W data 27:2178077568 10800 4,005 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.150092 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.150092 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.151092 W data 27:2178088368 592 0,000 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.151092 W data 12:2178072576 16384 4,995 cli C0A82DD5:63877BD4 192.168.45.214 13:17:04.156086 W logData 2:1869321541 2 0,996 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.156086 W logData 3:1877542213 2 0,996 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.157082 W data 12:2178082784 6176 2,994 cli C0A82DD5:63877BD4 192.168.45.214 13:17:04.158083 W data 20:2178072576 16384 7,498 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.165581 W logData 2:1869321542 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.165581 W logData 3:1877542214 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.167578 W data 20:2178077200 10800 2,994 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.170572 W inode 4:3060742158 1 0,996 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.171568 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.171568 W data 20:2178088000 960 1,001 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.172569 W data 28:2864381952 16384 5,988 cli C0A82DD5:63877BE5 192.168.45.214 13:17:04.178557 W logData 2:1869321543 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.178557 W logData 3:1877542215 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.179560 W data 28:2864391792 6544 2,995 cli C0A82DD5:63877BE5 192.168.45.214 13:17:04.179560 W data 5:2406842368 16384 5,987 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.185548 W logData 2:1869321544 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.185548 W logData 3:1877542216 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.186549 W data 5:2406846624 10800 4,993 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.191542 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.191542 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.192540 W data 5:2406857424 1328 0,995 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.193535 W data 13:2406842368 16384 6,019 cli C0A82DD5:63877BD5 192.168.45.213 13:17:04.199554 W logData 2:1869321544 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.199554 W logData 3:1877542216 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.200551 W data 13:2406851840 6912 2,997 cli C0A82DD5:63877BD5 192.168.45.213 13:17:04.201554 W data 21:2406842368 16384 5,912 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.207466 W logData 2:1869321545 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.207466 W logData 3:1877542217 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.208465 W data 21:2406846256 10800 3,990 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.212456 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.212456 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.213456 W data 21:2406857056 1696 1,998 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.214457 W data 6:2406842368 16384 5,015 cli C0A82DD5:63877BCE 192.168.45.214 13:17:04.219472 W logData 2:1869321546 2 1,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.219472 W logData 3:1877542218 2 1,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.221474 W data 6:2406851472 7280 1,994 cli C0A82DD5:63877BCE 192.168.45.214 13:17:04.221474 W data 14:2406842368 16384 7,502 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.228976 W logData 2:1869321547 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.228976 W logData 3:1877542219 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.229971 W data 14:2406845888 10800 3,994 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.233965 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.233965 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.234963 W data 14:2406856688 2064 1,999 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.235965 W data 22:2406842368 16384 4,993 cli C0A82DD5:63877BDF 192.168.45.214 13:17:04.240958 W logData 2:1869321547 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.240958 W logData 3:1877542219 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.241974 W data 22:2406851104 7648 2,531 cli C0A82DD5:63877BDF 192.168.45.214 13:17:04.242977 W data 7:2406842368 16384 5,526 cli C0A82DD5:63877BCF 192.168.45.213 13:17:04.248503 W logData 2:1869321548 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.248503 W logData 3:1877542220 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.249503 W data 7:2406845520 10800 3,994 cli C0A82DD5:63877BCF 192.168.45.213 13:17:04.253497 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Wednesday, 4 September 2024 12:59 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, given you see read latencies of 1 ms, you do not get the data from disk but from some cache (on whatever level). From spinning disks, you can never expect such read latencies (mind that GPFS block reading, even if sequential from the application's PoV, typically translates to random I/O at the physical disk level). So, I do not know what the latencies on your other measurements (bypassing GPFS) were but the numbers below do not represent sustained high-scale throughputs, apparently. It is nevertheless strange that your write rates are much below reads (and write latencies are that high) -- from my experience with different systems, when hammering GPFS with usual storage backends with both read and write requests, the writes tend to prevail. Your waiters indicate that the problem is above GPFS: GPFS is able to serve all I/O threads within a few ms, and there is not a long list of pending IOs. Sorry, the iohistory options is mmdiag --iohist But to me it looks like not GPFS is the culprit here. Uwe On 04.09.24 11:08, Henrik Cednert wrote: Hi Uwe Thanks. Worth noting is that we have win10 ltsc, win 2019 and have had a single CPU win 11 22h2 (as a test) clients that all perform as expected. Those machines is older though and connected with 10-40GbE, those client max out their NIC in read and write. Let me know if i missed something important here. Thanks again. Setups is Client: * Supermicro Workstation * Intel(R) Xeon(R) Gold 6418H 2.10 GHz (2 processors) * Mellanox ConnectX-6 Dx connected with 100GbE over dedicated vlan via mellanox sn2100. * Windows 11 Pro for Workstations, 22H2 Storage setup * 3 x 84 bay seagate chassis with spinning disks. * Storage connected with redundant 12Gb SAS to 2 x Storage node servers * 2 x mellanox sn2100 * The 2 storage node servers are for this vlan connected with 100GbE to each switch, so in total 4 x 100GbE. And the switches are connected with 2 * 100GbE I tested the commands you suggested. They are both new to me so not sure what the output is supposed to be, looks like -iohistory isn't available in windows. I ran --waiters a few times, as seen below. Not sure what the expected output is from that. mmdiag --waiters === mmdiag: waiters === Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 18616 MsgHandler at getData: for In function sendMessage Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 25084 WritebehindWorkerThread: on ThCond 0x31A7C360 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 16780 FsyncHandlerThread: on ThCond 0x37FFDAB0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 30308 MsgHandler at getData: for In function sendMessage C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0055 sec since 2024-09-04_10:05:21, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x37A25FF0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0029 sec since 2024-09-04_10:05:23, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x38281DE0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0019 sec since 2024-09-04_10:05:25, monitored, thread 11832 PrefetchWorkerThread: on ThCond 0x38278D20 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 16780 AcquireBRTHandlerThread: on ThCond 0x37A324E0 (MsgRecordCondvar), reason 'RPC wait' for tmMsgBRRevoke on node 192.168.45.161 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 2576 RangeRevokeWorkerThread: on ThCond 0x5419DAA0 (BrlObjCondvar), reason 'waiting because of local byte range lock conflict' C:\Users\m5-tkd01> C:\Users\m5-tkd01>mmdiag --iohistory Unrecognized option: --iohistory. Run mmdiag --help for the option list -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Tuesday, 3 September 2024 17:35 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, Henrik, while I am not using Windows I'd start investigating the usual things (see below). But first you should describe your set-up better. Where are the NSDs : locally attached to the Windows box? In some NSD servers? If the latter -- what is the link to the NSD servers? via your GbE link? FC? IB? separate Ethernet? What type of storage? Spinning Disks? Flash? How long are your I/Os waiting on the client (compare that to the waiting times on the NSD server if applicable)? not sure whether that is available on Windows, but mmdiag --waiters mmdiag --iohistory might be of use. Somewhere in the chain from your application to the storage backend there is a delay and you should first find out where that occurs I think. Bye Uwe On 03.09.24 14:10, Henrik Cednert wrote: Still no solution here regarding this. Have tested other cables. Have tested to change tcp window size, no change Played with numa in the bios, no change Played with hyperthreading in bios, no change Have anyone managed to get some speed out of windows 11 and gpfs? -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Friday, 9 August 2024 17:25 To: gpfsug-discuss at gpfsug.org Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. VARNING: DETTA ?R ETT EXTERNT MAIL. Klicka inte p? n?gra l?nkar oavsett hur legitima de verkar utan att verifiera. Hello I have some issues with write performance on a windows 11 pro system and I'm out of ideas here. Hopefully someone here have some bright ideas and/or experience of GPFS on Windows 11? The system is a: Windows 11 Pro 22H2 2 x Intel(R) Xeon(R) Gold 6418H 2.10 GHz 512 GB RAM GPFS 5.1.9.4 Mellanox ConnectX 6 Dx 100GbE connected to Mellanox Switch with 5m Mellanox DAC. Before deploying this workstation we had a single socket system as a test bench where we got 60 GbE in both directons with iPerf and around 6GB/sec write and 3GB/sec read from the system over GPFS (fio tests, same tests as furhter down here). With that system I had loads of issues before getting to that point though. MS Defender had to be forcefully disabled via regedit some other tweaks. All those tweaks have been performed in this new system as well, but I can't get the proper speed out of it. On this new system and with iPerf to the storage servers I get around 50-60GbE in both directions and send and receive. If I mount the storage over SMB and 100GbE via the storage gateway servers I get around 3GB/sec read and write with Blackmagics Disk speed test. I have not tweaked the system for samba performande, just a test to see what it would give and part of the troubleshooting. If I run Blackmagics diskspeed test to the GPFS mount I instead get around 700MB/sec write and 400MB/sec read. Starting to think that the Blackmagic test might not run properly on this machine with these CPUs though. Or it's related to the mmfsd process maybe, how that threads or not threads...? But if we instead look at fio. I have a bat script that loops through a bunch of FIO-tests. A test that I have been using over the years so that we easily can benchmark all deployed systems with the exakt same tests. The tests are named like: seqrw-gb-mb-t The result when I run this is like the below list. Number in parenthesis is the by fio reported latency. Job: seqrw-40gb-1mb-t1 ????????????Write: 162 MB/s (6 ms) ????????????Read: 1940 MB/s (1 ms) Job: seqrw-20gb-1mb-t2 ????????????Write: 286 MB/s (7 ms) ????????????Read: 3952 MB/s (1 ms) Job: seqrw-10gb-1mb-t4 ????????????Write: 549 MB/s (7 ms) ????????????Read: 6987 MB/s (1 ms) Job: seqrw-05gb-1mb-t8 ????????????Write: 989 MB/s (8 ms) ????????????Read: 7721 MB/s (1 ms) Job: seqrw-40gb-2mb-t1 ????????????Write: 161 MB/s (12 ms) ????????????Read: 2261 MB/s (0 ms) Job: seqrw-20gb-2mb-t2 ????????????Write: 348 MB/s (11 ms) ????????????Read: 4266 MB/s (1 ms) Job: seqrw-10gb-2mb-t4 ????????????Write: 626 MB/s (13 ms) ????????????Read: 4949 MB/s (1 ms) Job: seqrw-05gb-2mb-t8 ????????????Write: 1154 MB/s (14 ms) ????????????Read: 7007 MB/s (2 ms) Job: seqrw-40gb-4mb-t1 ????????????Write: 161 MB/s (25 ms) ????????????Read: 2083 MB/s (1 ms) Job: seqrw-20gb-4mb-t2 ????????????Write: 352 MB/s (23 ms) ????????????Read: 4317 MB/s (2 ms) Job: seqrw-10gb-4mb-t4 ????????????Write: 696 MB/s (23 ms) ????????????Read: 7358 MB/s (2 ms) Job: seqrw-05gb-4mb-t8 ????????????Write: 1251 MB/s (25 ms) ????????????Read: 6707 MB/s (5 ms) So with fio I get a very nice read speed, but the write is horrendous and I cannot find what causes it. I have looked at affinity settings for the mmfsd process but not sure I fully understand it. But no matter what I set it to, I see no difference. I have "played" with the bios and tried with/without hyperthreading, numa and so on. And nothing affects atleast the blackmagic disk speed test. the current settings for this host is like below. I write "current" because I have tested a few different settings here but nothing affects the write speed. maxTcpConnsPerNodeConn for sure bumped the read speed though. nsdMaxWorkerThreads 16 prefetchPct 60 maxTcpConnsPerNodeConn 8 maxMBpS 14000 Does anyone have any suggestions or ideas on how to troubleshoot this? Thanks -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: From uwe.falke at kit.edu Wed Sep 4 13:39:21 2024 From: uwe.falke at kit.edu (Uwe Falke) Date: Wed, 4 Sep 2024 14:39:21 +0200 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> Message-ID: <33582c29-da8f-4cc1-a8c2-3debb8373a71@kit.edu> Hi, the writes look strange: you seem to use a blocksize of 8MiB on your file system. The reads are entirely full block reads (16ki sectors a 0.5kiB => 8MiB), but of the writes less than half are comprising full blocks. The service time for the write I/O seems to correlate prtetty well with I/O size, i.e. is rate-bound, not IOps-bound (which again could be rooted in the data link to your NSD server): 8MiB transfers take about 6.0...6.5ms, related to approx 1.3GiB/s, a bunch of 10800-sector transfers take 4.0ms on rough average, this relates to 1.3GiB/s as well. As for the read data: none of the seen service times comes close to the 1ms you've reported below, confirming those numbers have nothing to do with your real data traffic. The service times go down to 5ms which is rather low, in particular compared to the somewhat higher service times for writes. This could mean those read I/Os with times of 6ms and below are served from the storage systems cache (NSD servers do buffer but not cache). Then, the read IOs with substantially higher service times are the ones really served from disk which again would mean your storage is the culprit . But then the question arises wy do your other machines behave well? But given the caching issues indicated above - are you sure the measurements at your other machines are trustworthy? The iohist snippet for reads comprises 74 IOs in about 0.854s, this relates to roughly 690MiB/s, far from the 10-fold value you reported. So, while reason is to assume some IOs are served from the storage cache, much more of those seems to come from the client's cache Begfore that hasn't been clarified I think you do not need to play with the BIOS and stuff :-) Uwe On 04.09.24 13:28, Henrik Cednert wrote: > Adding a snippet from iohist when reading as well. > > >mmdiag --iohist > > === mmdiag: iohist === > > I/O history: > > ?I/O start time RW ? ?Buf type disk:sectorNum ? ? nSec ?time ms ?Type > ? ? ?Device/NSD ID ? ? ? ?NSD node > --------------- -- ----------- ----------------- ----- ?------- ?---- > ------------------ --------------- > 13:24:59.180277 ?R ? ? ? ?data ? 19:1491763200 ? 16384 ? 32,433 ? cli > C0A82DD5:63877BDC ? 192.168.45.213 > 13:24:59.212710 ?R ? ? ? ?data ? ?8:805453824 ? ?16384 ? ?5,987 ? cli > C0A82DD5:63877BD0 ? 192.168.45.214 > 13:24:59.220698 ?R ? ? ? ?data ? ?8:805453824 ? ?16384 ? ?6,012 ? cli > C0A82DD5:63877BD0 ? 192.168.45.214 > 13:24:59.226710 ?R ? ? ? ?data ? 16:805453824 ? ?16384 ? ?9,129 ? cli > C0A82DD5:63877BD9 ? 192.168.45.214 > 13:24:59.226710 ?R ? ? ? ?data ? 12:1491763200 ? 16384 ? 18,526 ? cli > C0A82DD5:63877BD4 ? 192.168.45.214 > 13:24:59.226710 ?R ? ? ? ?data ? 20:2178072576 ? 16384 ? 23,520 ? cli > C0A82DD5:63877BDD ? 192.168.45.214 > 13:24:59.251229 ?R ? ? ? ?data ? 16:805453824 ? ?16384 ? ?5,497 ? cli > C0A82DD5:63877BD9 ? 192.168.45.214 > 13:24:59.257730 ?R ? ? ? ?data ? 24:805453824 ? ?16384 ? ?5,990 ? cli > C0A82DD5:63877BE1 ? 192.168.45.214 > 13:24:59.257730 ?R ? ? ? ?data ? ?5:1720532992 ? 16384 ? 33,458 ? cli > C0A82DD5:63877BCD ? 192.168.45.213 > 13:24:59.257730 ?R ? ? ? ?data ? 28:1720532992 ? 16384 ? 33,458 ? cli > C0A82DD5:63877BE5 ? 192.168.45.214 > 13:24:59.292189 ?R ? ? ? ?data ? 24:805453824 ? ?16384 ? ?4,992 ? cli > C0A82DD5:63877BE1 ? 192.168.45.214 > 13:24:59.300176 ?R ? ? ? ?data ? 24:805453824 ? ?16384 ? ?4,991 ? cli > C0A82DD5:63877BE1 ? 192.168.45.214 > 13:24:59.306169 ?R ? ? ? ?data ? ?9:805453824 ? ?16384 ? ?7,987 ? cli > C0A82DD5:63877BD1 ? 192.168.45.213 > 13:24:59.306169 ?R ? ? ? ?data ? 13:1720532992 ? 16384 ? 25,471 ? cli > C0A82DD5:63877BD5 ? 192.168.45.213 > 13:24:59.306169 ?R ? ? ? ?data ? 21:1720532992 ? 16384 ? 28,466 ? cli > C0A82DD5:63877BDE ? 192.168.45.213 > 13:24:59.335637 ?R ? ? ? ?data ? ?9:805453824 ? ?16384 ? ?4,994 ? cli > C0A82DD5:63877BD1 ? 192.168.45.213 > 13:24:59.341629 ?R ? ? ? ?data ? 17:805453824 ? ?16384 ? ?3,993 ? cli > C0A82DD5:63877BDA ? 192.168.45.213 > 13:24:59.341629 ?R ? ? ? ?data ? ?6:1720532992 ? 16384 ? 27,471 ? cli > C0A82DD5:63877BCE ? 192.168.45.214 > 13:24:59.341629 ?R ? ? ? ?data ? 14:1720532992 ? 16384 ? 44,953 ? cli > C0A82DD5:63877BD6 ? 192.168.45.214 > 13:24:59.387584 ?R ? ? ? ?data ? 17:805453824 ? ?16384 ? ?5,990 ? cli > C0A82DD5:63877BDA ? 192.168.45.213 > 13:24:59.394572 ?R ? ? ? ?data ? 17:805453824 ? ?16384 ? ?7,993 ? cli > C0A82DD5:63877BDA ? 192.168.45.213 > 13:24:59.402565 ?R ? ? ? ?data ? 25:805453824 ? ?16384 ? ?4,991 ? cli > C0A82DD5:63877BE2 ? 192.168.45.213 > 13:24:59.402565 ?R ? ? ? ?data ? 22:1720532992 ? 16384 ? 26,309 ? cli > C0A82DD5:63877BDF ? 192.168.45.214 > 13:24:59.402565 ?R ? ? ? ?data ? ?7:1720532992 ? 16384 ?142,856 ? cli > C0A82DD5:63877BCF ? 192.168.45.213 > 13:24:59.546419 ?R ? ? ? ?data ? 25:805453824 ? ?16384 ? ?7,992 ? cli > C0A82DD5:63877BE2 ? 192.168.45.213 > 13:24:59.556408 ?R ? ? ? ?data ? 25:805453824 ? ?16384 ? ?7,987 ? cli > C0A82DD5:63877BE2 ? 192.168.45.213 > 13:24:59.564395 ?R ? ? ? ?data ? 10:805453824 ? ?16384 ? ?5,505 ? cli > C0A82DD5:63877BD2 ? 192.168.45.214 > 13:24:59.564395 ?R ? ? ? ?data ? 23:1720532992 ? 16384 ? 28,053 ? cli > C0A82DD5:63877BE0 ? 192.168.45.213 > 13:24:59.564395 ?R ? ? ? ?data ? 15:1720532992 ? 16384 ? 33,044 ? cli > C0A82DD5:63877BD8 ? 192.168.45.213 > 13:24:59.598437 ?R ? ? ? ?data ? 10:805453824 ? ?16384 ? ?5,504 ? cli > C0A82DD5:63877BD2 ? 192.168.45.214 > 13:24:59.604939 ?R ? ? ? ?data ? 18:805453824 ? ?16384 ? ?4,993 ? cli > C0A82DD5:63877BDB ? 192.168.45.214 > 13:24:59.604939 ?R ? ? ? ?data ? ?8:1720532992 ? 16384 ? 36,015 ? cli > C0A82DD5:63877BD0 ? 192.168.45.214 > 13:24:59.609932 ?R ? ? ? ?data ? 16:1720532992 ? 16384 ? 42,010 ? cli > C0A82DD5:63877BD9 ? 192.168.45.214 > 13:24:59.652940 ?R ? ? ? ?data ? 18:805453824 ? ?16384 ? ?7,989 ? cli > C0A82DD5:63877BDB ? 192.168.45.214 > 13:24:59.662434 ?R ? ? ? ?data ? 18:805453824 ? ?16384 ? ?6,994 ? cli > C0A82DD5:63877BDB ? 192.168.45.214 > 13:24:59.669428 ?R ? ? ? ?data ? 26:805453824 ? ?16384 ? ?8,986 ? cli > C0A82DD5:63877BE3 ? 192.168.45.214 > 13:24:59.669428 ?R ? ? ? ?data ? 24:1720532992 ? 16384 ? 20,308 ? cli > C0A82DD5:63877BE1 ? 192.168.45.214 > 13:24:59.669428 ?R ? ? ? ?data ? ?9:1720532992 ? 16384 ? 25,812 ? cli > C0A82DD5:63877BD1 ? 192.168.45.213 > 13:24:59.696239 ?R ? ? ? ?data ? 26:805453824 ? ?16384 ? ?6,989 ? cli > C0A82DD5:63877BE3 ? 192.168.45.214 > 13:24:59.703228 ?R ? ? ? ?data ? 11:805453824 ? ?16384 ? ?4,992 ? cli > C0A82DD5:63877BD3 ? 192.168.45.213 > 13:24:59.703228 ?R ? ? ? ?data ? 25:1720532992 ? 16384 ? 17,976 ? cli > C0A82DD5:63877BE2 ? 192.168.45.213 > 13:24:59.703228 ?R ? ? ? ?data ? 17:1720532992 ? 16384 ? 22,481 ? cli > C0A82DD5:63877BDA ? 192.168.45.213 > 13:24:59.725709 ?R ? ? ? ?data ? 11:805453824 ? ?16384 ? ?4,992 ? cli > C0A82DD5:63877BD3 ? 192.168.45.213 > 13:24:59.731699 ?R ? ? ? ?data ? 11:805453824 ? ?16384 ? ?8,986 ? cli > C0A82DD5:63877BD3 ? 192.168.45.213 > 13:24:59.740685 ?R ? ? ? ?data ? 19:805453824 ? ?16384 ? ?4,997 ? cli > C0A82DD5:63877BDC ? 192.168.45.213 > 13:24:59.740685 ?R ? ? ? ?data ? 18:1720532992 ? 16384 ? 19,486 ? cli > C0A82DD5:63877BDB ? 192.168.45.214 > 13:24:59.740685 ?R ? ? ? ?data ? 10:1720532992 ? 16384 ? 27,474 ? cli > C0A82DD5:63877BD2 ? 192.168.45.214 > 13:24:59.769157 ?R ? ? ? ?data ? 19:805453824 ? ?16384 ? ?4,997 ? cli > C0A82DD5:63877BDC ? 192.168.45.213 > 13:24:59.774154 ?R ? ? ? ?data ? 27:805453824 ? ?16384 ? ?4,992 ? cli > C0A82DD5:63877BE4 ? 192.168.45.213 > 13:24:59.774154 ?R ? ? ? ?data ? 11:1720532992 ? 16384 ? 22,476 ? cli > C0A82DD5:63877BD3 ? 192.168.45.213 > 13:24:59.774154 ?R ? ? ? ?data ? 26:1720532992 ? 16384 ? 29,464 ? cli > C0A82DD5:63877BE3 ? 192.168.45.214 > 13:24:59.803618 ?R ? ? ? ?data ? 27:805453824 ? ?16384 ? ?4,997 ? cli > C0A82DD5:63877BE4 ? 192.168.45.213 > 13:24:59.809614 ?R ? ? ? ?data ? 27:805453824 ? ?16384 ? ?7,987 ? cli > C0A82DD5:63877BE4 ? 192.168.45.213 > 13:24:59.817601 ?R ? ? ? ?data ? 12:805453824 ? ?16384 ? ?8,500 ? cli > C0A82DD5:63877BD4 ? 192.168.45.214 > 13:24:59.817601 ?R ? ? ? ?data ? 27:1720532992 ? 16384 ? ?9,499 ? cli > C0A82DD5:63877BE4 ? 192.168.45.213 > 13:24:59.817601 ?R ? ? ? ?data ? 19:1720532992 ? 16384 ? 74,075 ? cli > C0A82DD5:63877BDC ? 192.168.45.213 > 13:24:59.892674 ?R ? ? ? ?data ? 12:805453824 ? ?16384 ? ?4,997 ? cli > C0A82DD5:63877BD4 ? 192.168.45.214 > 13:24:59.897671 ?R ? ? ? ?data ? 20:1491763200 ? 16384 ? ?4,992 ? cli > C0A82DD5:63877BDD ? 192.168.45.214 > 13:24:59.898670 ?R ? ? ? ?data ? 12:1720532992 ? 16384 ? 24,479 ? cli > C0A82DD5:63877BD4 ? 192.168.45.214 > 13:24:59.898670 ?R ? ? ? ?data ? 20:2406842368 ? 16384 ? 26,476 ? cli > C0A82DD5:63877BDD ? 192.168.45.214 > 13:24:59.925146 ?R ? ? ? ?data ? 20:1491763200 ? 16384 ? ?8,990 ? cli > C0A82DD5:63877BDD ? 192.168.45.214 > 13:24:59.935430 ?R ? ? ? ?data ? 20:1491763200 ? 16384 ? ?4,691 ? cli > C0A82DD5:63877BDD ? 192.168.45.214 > 13:24:59.940121 ?R ? ? ? ?data ? 28:1034223616 ? 16384 ? ?4,312 ? cli > C0A82DD5:63877BE5 ? 192.168.45.214 > 13:24:59.940121 ?R ? ? ? ?data ? 28:1949302784 ? 16384 ? 26,664 ? cli > C0A82DD5:63877BE5 ? 192.168.45.214 > 13:24:59.940121 ?R ? ? ? ?data ? ?5:1949302784 ? 16384 ? 34,831 ? cli > C0A82DD5:63877BCD ? 192.168.45.213 > 13:24:59.975955 ?R ? ? ? ?data ? 28:1034223616 ? 16384 ? ?5,583 ? cli > C0A82DD5:63877BE5 ? 192.168.45.214 > 13:24:59.981539 ?R ? ? ? ?data ? ?5:1034223616 ? 16384 ? ?5,168 ? cli > C0A82DD5:63877BCD ? 192.168.45.213 > 13:24:59.981539 ?R ? ? ? ?data ? 21:1949302784 ? 16384 ? 25,392 ? cli > C0A82DD5:63877BDE ? 192.168.45.213 > 13:24:59.981539 ?R ? ? ? ?data ? 13:1949302784 ? 16384 ? 34,412 ? cli > C0A82DD5:63877BD5 ? 192.168.45.213 > 13:25:00.016950 ?R ? ? ? ?data ? ?5:1034223616 ? 16384 ? ?7,991 ? cli > C0A82DD5:63877BCD ? 192.168.45.213 > 13:25:00.026938 ?R ? ? ? ?data ? ?5:1034223616 ? 16384 ? ?7,750 ? cli > C0A82DD5:63877BCD ? 192.168.45.213 > 13:25:00.034688 ?R ? ? ? ?data ? 13:1034223616 ? 16384 ? ?5,498 ? cli > C0A82DD5:63877BD5 ? 192.168.45.213 > 13:25:00.034688 ?R ? ? ? ?data ? ?6:1949302784 ? 16384 ? 27,661 ? cli > C0A82DD5:63877BCE ? 192.168.45.214 > > > -- > > Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */? OnePost > *(formerly?Filmlance Post) > > ?? *OnePost*, formerly Filmlance's post-production, is now an > independent part of the Banijay Group. > New name, same team ? business as usual at OnePost. > > > > ------------------------------------------------------------------------ > *From:* gpfsug-discuss on behalf > of Henrik Cednert > *Sent:* Wednesday, 4 September 2024 13:24 > *To:* gpfsug-discuss at gpfsug.org > *Subject:* Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. > Performance issues, write. > Hello > > My theory is that it's windows 11 related, or the combo windows 11 and > this hardware. I guess the only thing to know for sure, is to boot it > into *nix and install gpfs there and test. Which I guess isn't the > worst of ideas at this stage. > > --iohist spits out a pretty hefty chuck of data. Below is a snippet > from when I did a write test: > > > > mmdiag --iohist > > === mmdiag: iohist === > > I/O history: > > ?I/O start time RW ? ?Buf type disk:sectorNum ? ? nSec ?time ms ?Type > ? ? ?Device/NSD ID ? ? ? ?NSD node > --------------- -- ----------- ----------------- ----- ?------- ?---- > ------------------ --------------- > 13:17:04.088621 ?W ? ? ? ?data ? 10:2178073088 ? 10800 ?3,994 ? cli > C0A82DD5:63877BD2 ? 192.168.45.214 > 13:17:04.092614 ?W ? ? ? inode ? ?4:3060742158 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCC ? 192.168.45.214 > 13:17:04.092614 ?W ? ? ? inode ? ?2:3040332814 ? ? ? 1 ?0,999 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.093613 ?W ? ? ? ?data ? 10:2178083888 ? ?5072 ?2,995 ? cli > C0A82DD5:63877BD2 ? 192.168.45.214 > 13:17:04.094611 ?W ? ? ? ?data ? 18:2178072576 ? 16384 ?6,502 ? cli > C0A82DD5:63877BDB ? 192.168.45.214 > 13:17:04.101113 ?W ? ? logData ? ?2:1869321537 ? ? ? 2 ?0,998 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.101113 ?W ? ? logData ? ?3:1877542209 ? ? ? 2 ?0,998 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.103109 ?W ? ? ? ?data ? 18:2178078304 ? 10656 ?3,994 ? cli > C0A82DD5:63877BDB ? 192.168.45.214 > 13:17:04.103109 ?W ? ? ? ?data ? 26:2178072576 ? 16384 ?6,529 ? cli > C0A82DD5:63877BE3 ? 192.168.45.214 > 13:17:04.109638 ?W ? ? logData ? ?2:1869321538 ? ? ? 2 ?0,994 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.109638 ?W ? ? logData ? ?3:1877542210 ? ? ? 2 ?0,994 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.111634 ?W ? ? ? ?data ? 26:2178072720 ? 10800 ?3,992 ? cli > C0A82DD5:63877BE3 ? 192.168.45.214 > 13:17:04.115626 ?W ? ? ? inode ? ?4:3060742158 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCC ? 192.168.45.214 > 13:17:04.115626 ?W ? ? ? inode ? ?2:3040332814 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.116626 ?W ? ? ? ?data ? 26:2178083520 ? ?5440 ?2,995 ? cli > C0A82DD5:63877BE3 ? 192.168.45.214 > 13:17:04.117629 ?W ? ? ? ?data ? 11:2178072576 ? 16384 ?4,987 ? cli > C0A82DD5:63877BD3 ? 192.168.45.213 > 13:17:04.122616 ?W ? ? logData ? ?2:1869321539 ? ? ? 2 ?0,999 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.122616 ?W ? ? logData ? ?3:1877542211 ? ? ? 2 ?0,999 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.124613 ?W ? ? ? ?data ? 11:2178077936 ? 10800 ?3,999 ? cli > C0A82DD5:63877BD3 ? 192.168.45.213 > 13:17:04.128612 ?W ? ? ? inode ? ?4:3060742158 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCC ? 192.168.45.214 > 13:17:04.128612 ?W ? ? ? inode ? ?2:3040332814 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.129609 ?W ? ? ? ?data ? 11:2178088736 ? ? 224 ?0,000 ? cli > C0A82DD5:63877BD3 ? 192.168.45.213 > 13:17:04.129609 ?W ? ? ? ?data ? 19:2178072576 ? 16384 ?6,495 ? cli > C0A82DD5:63877BDC ? 192.168.45.213 > 13:17:04.136104 ?W ? ? logData ? ?2:1869321540 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.136104 ?W ? ? logData ? ?3:1877542212 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.137103 ?W ? ? ? ?data ? 19:2178083152 ? ?5808 ?2,999 ? cli > C0A82DD5:63877BDC ? 192.168.45.213 > 13:17:04.138100 ?W ? ? ? ?data ? 27:2178072576 ? 16384 ?5,990 ? cli > C0A82DD5:63877BE4 ? 192.168.45.213 > 13:17:04.144091 ?W ? ? logData ? ?2:1869321540 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.144091 ?W ? ? logData ? ?3:1877542212 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.146088 ?W ? ? ? ?data ? 27:2178077568 ? 10800 ?4,005 ? cli > C0A82DD5:63877BE4 ? 192.168.45.213 > 13:17:04.150092 ?W ? ? ? inode ? ?4:3060742158 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCC ? 192.168.45.214 > 13:17:04.150092 ?W ? ? ? inode ? ?2:3040332814 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.151092 ?W ? ? ? ?data ? 27:2178088368 ? ? 592 ?0,000 ? cli > C0A82DD5:63877BE4 ? 192.168.45.213 > 13:17:04.151092 ?W ? ? ? ?data ? 12:2178072576 ? 16384 ?4,995 ? cli > C0A82DD5:63877BD4 ? 192.168.45.214 > 13:17:04.156086 ?W ? ? logData ? ?2:1869321541 ? ? ? 2 ?0,996 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.156086 ?W ? ? logData ? ?3:1877542213 ? ? ? 2 ?0,996 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.157082 ?W ? ? ? ?data ? 12:2178082784 ? ?6176 ?2,994 ? cli > C0A82DD5:63877BD4 ? 192.168.45.214 > 13:17:04.158083 ?W ? ? ? ?data ? 20:2178072576 ? 16384 ?7,498 ? cli > C0A82DD5:63877BDD ? 192.168.45.214 > 13:17:04.165581 ?W ? ? logData ? ?2:1869321542 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.165581 ?W ? ? logData ? ?3:1877542214 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.167578 ?W ? ? ? ?data ? 20:2178077200 ? 10800 ?2,994 ? cli > C0A82DD5:63877BDD ? 192.168.45.214 > 13:17:04.170572 ?W ? ? ? inode ? ?4:3060742158 ? ? ? 1 ?0,996 ? cli > C0A82DD5:63877BCC ? 192.168.45.214 > 13:17:04.171568 ?W ? ? ? inode ? ?2:3040332814 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.171568 ?W ? ? ? ?data ? 20:2178088000 ? ? 960 ?1,001 ? cli > C0A82DD5:63877BDD ? 192.168.45.214 > 13:17:04.172569 ?W ? ? ? ?data ? 28:2864381952 ? 16384 ?5,988 ? cli > C0A82DD5:63877BE5 ? 192.168.45.214 > 13:17:04.178557 ?W ? ? logData ? ?2:1869321543 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.178557 ?W ? ? logData ? ?3:1877542215 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.179560 ?W ? ? ? ?data ? 28:2864391792 ? ?6544 ?2,995 ? cli > C0A82DD5:63877BE5 ? 192.168.45.214 > 13:17:04.179560 ?W ? ? ? ?data ? ?5:2406842368 ? 16384 ?5,987 ? cli > C0A82DD5:63877BCD ? 192.168.45.213 > 13:17:04.185548 ?W ? ? logData ? ?2:1869321544 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.185548 ?W ? ? logData ? ?3:1877542216 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.186549 ?W ? ? ? ?data ? ?5:2406846624 ? 10800 ?4,993 ? cli > C0A82DD5:63877BCD ? 192.168.45.213 > 13:17:04.191542 ?W ? ? ? inode ? ?4:3060742158 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCC ? 192.168.45.214 > 13:17:04.191542 ?W ? ? ? inode ? ?2:3040332814 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.192540 ?W ? ? ? ?data ? ?5:2406857424 ? ?1328 ?0,995 ? cli > C0A82DD5:63877BCD ? 192.168.45.213 > 13:17:04.193535 ?W ? ? ? ?data ? 13:2406842368 ? 16384 ?6,019 ? cli > C0A82DD5:63877BD5 ? 192.168.45.213 > 13:17:04.199554 ?W ? ? logData ? ?2:1869321544 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.199554 ?W ? ? logData ? ?3:1877542216 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.200551 ?W ? ? ? ?data ? 13:2406851840 ? ?6912 ?2,997 ? cli > C0A82DD5:63877BD5 ? 192.168.45.213 > 13:17:04.201554 ?W ? ? ? ?data ? 21:2406842368 ? 16384 ?5,912 ? cli > C0A82DD5:63877BDE ? 192.168.45.213 > 13:17:04.207466 ?W ? ? logData ? ?2:1869321545 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.207466 ?W ? ? logData ? ?3:1877542217 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.208465 ?W ? ? ? ?data ? 21:2406846256 ? 10800 ?3,990 ? cli > C0A82DD5:63877BDE ? 192.168.45.213 > 13:17:04.212456 ?W ? ? ? inode ? ?4:3060742158 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCC ? 192.168.45.214 > 13:17:04.212456 ?W ? ? ? inode ? ?2:3040332814 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.213456 ?W ? ? ? ?data ? 21:2406857056 ? ?1696 ?1,998 ? cli > C0A82DD5:63877BDE ? 192.168.45.213 > 13:17:04.214457 ?W ? ? ? ?data ? ?6:2406842368 ? 16384 ?5,015 ? cli > C0A82DD5:63877BCE ? 192.168.45.214 > 13:17:04.219472 ?W ? ? logData ? ?2:1869321546 ? ? ? 2 ?1,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.219472 ?W ? ? logData ? ?3:1877542218 ? ? ? 2 ?1,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.221474 ?W ? ? ? ?data ? ?6:2406851472 ? ?7280 ?1,994 ? cli > C0A82DD5:63877BCE ? 192.168.45.214 > 13:17:04.221474 ?W ? ? ? ?data ? 14:2406842368 ? 16384 ?7,502 ? cli > C0A82DD5:63877BD6 ? 192.168.45.214 > 13:17:04.228976 ?W ? ? logData ? ?2:1869321547 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.228976 ?W ? ? logData ? ?3:1877542219 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.229971 ?W ? ? ? ?data ? 14:2406845888 ? 10800 ?3,994 ? cli > C0A82DD5:63877BD6 ? 192.168.45.214 > 13:17:04.233965 ?W ? ? ? inode ? ?4:3060742158 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCC ? 192.168.45.214 > 13:17:04.233965 ?W ? ? ? inode ? ?2:3040332814 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.234963 ?W ? ? ? ?data ? 14:2406856688 ? ?2064 ?1,999 ? cli > C0A82DD5:63877BD6 ? 192.168.45.214 > 13:17:04.235965 ?W ? ? ? ?data ? 22:2406842368 ? 16384 ?4,993 ? cli > C0A82DD5:63877BDF ? 192.168.45.214 > 13:17:04.240958 ?W ? ? logData ? ?2:1869321547 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.240958 ?W ? ? logData ? ?3:1877542219 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.241974 ?W ? ? ? ?data ? 22:2406851104 ? ?7648 ?2,531 ? cli > C0A82DD5:63877BDF ? 192.168.45.214 > 13:17:04.242977 ?W ? ? ? ?data ? ?7:2406842368 ? 16384 ?5,526 ? cli > C0A82DD5:63877BCF ? 192.168.45.213 > 13:17:04.248503 ?W ? ? logData ? ?2:1869321548 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCA ? 192.168.45.214 > 13:17:04.248503 ?W ? ? logData ? ?3:1877542220 ? ? ? 2 ?0,000 ? cli > C0A82DD5:63877BCB ? 192.168.45.213 > 13:17:04.249503 ?W ? ? ? ?data ? ?7:2406845520 ? 10800 ?3,994 ? cli > C0A82DD5:63877BCF ? 192.168.45.213 > 13:17:04.253497 ?W ? ? ? inode ? ?4:3060742158 ? ? ? 1 ?0,000 ? cli > C0A82DD5:63877BCC ? 192.168.45.214 > > > -- > > Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */? OnePost > *(formerly?Filmlance Post) > > ?? *OnePost*, formerly Filmlance's post-production, is now an > independent part of the Banijay Group. > New name, same team ? business as usual at OnePost. > > > > ------------------------------------------------------------------------ > *From:* gpfsug-discuss on behalf > of Uwe Falke > *Sent:* Wednesday, 4 September 2024 12:59 > *To:* gpfsug-discuss at gpfsug.org > *Subject:* Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. > Performance issues, write. > > Hi, given you see read latencies of 1 ms, you do not get the data from > disk but from some cache (on whatever level). From spinning disks, you > can never expect such read latencies (mind that GPFS block reading, > even if sequential from the application's PoV, typically translates to > random I/O at the physical disk level). > > So, I do not know what the latencies on your other measurements > (bypassing GPFS) were? but the numbers below do not represent > sustained high-scale throughputs, apparently. > > > It is nevertheless strange that your write rates are much below reads > (and write latencies are that high) -- from my experience with > different systems, when hammering GPFS with usual storage backends > with both read and write requests, the writes tend to prevail. > > > > > Your waiters indicate that the problem is above GPFS: > > GPFS is able to serve all I/O threads within a few ms, and there is > not a long list of pending IOs. > > > Sorry, the iohistory options is > > mmdiag --iohist > > > But to me it looks like not GPFS is the culprit here. > > > Uwe > > On 04.09.24 11:08, Henrik Cednert wrote: >> Hi Uwe >> >> Thanks. >> >> Worth noting is that we have win10 ltsc, win 2019 and have had a >> single CPU win 11 22h2 (as a test) clients that all perform as >> expected. Those machines is older though and connected with 10-40GbE, >> those client max out their NIC in read and write. >> >> Let me know if i missed something important here. Thanks again. >> >> Setups is >> >> Client: >> >> * Supermicro Workstation >> * Intel(R) Xeon(R) Gold 6418H ? 2.10 GHz ?(2 processors) >> * Mellanox ConnectX-6 Dx connected with 100GbE over dedicated vlan >> via mellanox sn2100. >> * Windows 11 Pro for Workstations, 22H2 >> >> >> >> Storage setup >> >> * >> 3 x 84 bay seagate chassis with spinning disks. >> * >> Storage connected with redundant 12Gb SAS to 2 x Storage node servers >> * >> 2 x mellanox sn2100 >> * >> The 2 storage node servers are for this vlan connected with >> 100GbE to each switch, so in total 4 x 100GbE.? And the switches >> are connected with 2 * 100GbE >> >> >> >> I tested the commands you suggested. They are both new to me so not >> sure what the output is supposed to be, looks like -iohistory isn't >> available in windows. I ran --waiters a few times, as seen below. Not >> sure what the expected output is from that. >> >> >> mmdiag --waiters >> >> === mmdiag: waiters === >> Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 18616 >> MsgHandler at getData: for In function sendMessage >> Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 25084 >> WritebehindWorkerThread: on ThCond 0x31A7C360 (MsgRecordCondvar), >> reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 >> >> C:\Users\m5-tkd01>mmdiag --waiters >> >> === mmdiag: waiters === >> Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 16780 >> FsyncHandlerThread: on ThCond 0x37FFDAB0 (MsgRecordCondvar), reason >> 'RPC wait' for NSD I/O completion on node 192.168.45.214 >> Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 30308 >> MsgHandler at getData: for In function sendMessage >> >> C:\Users\m5-tkd01>mmdiag --waiters >> >> === mmdiag: waiters === >> Waiting 0.0055 sec since 2024-09-04_10:05:21, monitored, thread 16780 >> FileBlockReadFetchHandlerThread: on ThCond 0x37A25FF0 >> (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node >> 192.168.45.213 >> >> C:\Users\m5-tkd01>mmdiag --waiters >> >> === mmdiag: waiters === >> Waiting 0.0029 sec since 2024-09-04_10:05:23, monitored, thread 16780 >> FileBlockReadFetchHandlerThread: on ThCond 0x38281DE0 >> (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node >> 192.168.45.213 >> >> C:\Users\m5-tkd01>mmdiag --waiters >> >> === mmdiag: waiters === >> Waiting 0.0019 sec since 2024-09-04_10:05:25, monitored, thread 11832 >> PrefetchWorkerThread: on ThCond 0x38278D20 (MsgRecordCondvar), reason >> 'RPC wait' for NSD I/O completion on node 192.168.45.214 >> Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 16780 >> AcquireBRTHandlerThread: on ThCond 0x37A324E0 (MsgRecordCondvar), >> reason 'RPC wait' for tmMsgBRRevoke on node 192.168.45.161 >> Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 2576 >> RangeRevokeWorkerThread: on ThCond 0x5419DAA0 (BrlObjCondvar), reason >> 'waiting because of local byte range lock conflict' >> >> C:\Users\m5-tkd01> >> >> >> >> >> >> >> C:\Users\m5-tkd01>mmdiag --iohistory >> Unrecognized option: --iohistory. >> Run mmdiag --help for the option list >> >> >> >> >> -- >> >> Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */? OnePost >> *(formerly?Filmlance Post) >> >> ?? *OnePost*, formerly Filmlance's post-production, is now an >> independent part of the Banijay Group. >> New name, same team ? business as usual at OnePost. >> >> >> >> ------------------------------------------------------------------------ >> *From:* gpfsug-discuss >> on behalf of Uwe Falke >> >> *Sent:* Tuesday, 3 September 2024 17:35 >> *To:* gpfsug-discuss at gpfsug.org >> >> *Subject:* Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. >> Performance issues, write. >> >> Hi, Henrik, >> >> >> while I am not using Windows I'd start investigating the usual things >> (see below). >> >> >> But first you should describe your set-up better. >> >> Where are the NSDs : locally attached to the Windows box? In some NSD >> servers? >> >> If the latter -- what is the link to the NSD servers? via your GbE >> link? FC? IB? separate Ethernet? >> >> What type of storage? Spinning Disks? Flash? >> >> >> How long are your I/Os waiting on the client (compare that to the >> waiting times on the NSD server if applicable)? >> >> not sure whether that is available on Windows, but >> >> mmdiag --waiters >> >> mmdiag --iohistory >> >> might be of use. >> >> >> Somewhere in the chain from your application to the storage backend >> there is a delay and you should first find out where that occurs I >> think. >> >> >> Bye >> >> Uwe >> >> >> >> On 03.09.24 14:10, Henrik Cednert wrote: >>> Still no solution here regarding this. >>> >>> Have tested other cables. >>> Have tested to change tcp window size, no change >>> Played with numa in the bios, no change >>> Played with hyperthreading in bios, no change >>> >>> >>> Have anyone managed to get some speed out of windows 11 and gpfs? >>> >>> >>> -- >>> >>> Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */ OnePost >>> *(formerly?Filmlance Post) >>> >>> ?? *OnePost*, formerly Filmlance's post-production, is now an >>> independent part of the Banijay Group. >>> New name, same team ? business as usual at OnePost. >>> >>> >>> >>> ------------------------------------------------------------------------ >>> *From:* gpfsug-discuss >>> on behalf of Henrik >>> Cednert >>> *Sent:* Friday, 9 August 2024 17:25 >>> *To:* gpfsug-discuss at gpfsug.org >>> >>> *Subject:* [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. >>> Performance issues, write. >>> >>> ***VARNING: DETTA ?R ETT EXTERNT MAIL. Klicka inte p? n?gra l?nkar >>> oavsett hur legitima de verkar utan att verifiera.* >>> >>> >>> Hello >>> >>> I have some issues with write performance on a windows 11 pro system >>> and I'm out of ideas here. Hopefully someone here have some bright >>> ideas and/or experience of GPFS on Windows 11? >>> >>> The system is a: >>> >>> Windows 11 Pro 22H2 >>> 2 x Intel(R) Xeon(R) Gold 6418H ? 2.10 GHz >>> 512 GB RAM >>> GPFS 5.1.9.4 >>> Mellanox ConnectX 6 Dx >>> 100GbE connected to Mellanox Switch with 5m Mellanox DAC. >>> >>> Before deploying this workstation we had a single socket system as a >>> test bench where we got 60 GbE in both directons with iPerf and >>> around 6GB/sec write and 3GB/sec read from the system over GPFS (fio >>> tests, same tests as furhter down here). >>> >>> With that system I had loads of issues before getting to that point >>> though. MS Defender had to be forcefully disabled via regedit some >>> other tweaks. All those tweaks have been performed in this new >>> system as well, but I can't get the proper speed out of it. >>> >>> >>> On this new system and with iPerf to the storage servers I get >>> around 50-60GbE in both directions and send and receive. >>> >>> If I mount the storage over SMB and 100GbE via the storage gateway >>> servers I get around 3GB/sec read and write with Blackmagics Disk >>> speed test. I have not tweaked the system for samba performande, >>> just a test to see what it would give and part of the troubleshooting. >>> >>> If I run Blackmagics diskspeed test to the GPFS mount I instead get >>> around 700MB/sec write and 400MB/sec read. >>> >>> Starting to think that the Blackmagic test might not run properly on >>> this machine with these CPUs though. Or it's related to the mmfsd >>> process maybe, how that threads or not threads...? >>> >>> But if we instead look at fio. I have a bat script that loops >>> through a bunch of FIO-tests. A test that I have been using over the >>> years so that we easily can benchmark all deployed systems with the >>> exakt same tests. The tests are named like: >>> >>> seqrw-gb-mb-t >>> >>> The result when I run this is like the below list. Number in >>> parenthesis is the by fio reported latency. >>> >>> Job: seqrw-40gb-1mb-t1 >>> ????????????Write: 162 MB/s (6 ms) >>> ????????????Read: 1940 MB/s (1 ms) >>> >>> Job: seqrw-20gb-1mb-t2 >>> ????????????Write: 286 MB/s (7 ms) >>> ????????????Read: 3952 MB/s (1 ms) >>> >>> Job: seqrw-10gb-1mb-t4 >>> ????????????Write: 549 MB/s (7 ms) >>> ????????????Read: 6987 MB/s (1 ms) >>> >>> Job: seqrw-05gb-1mb-t8 >>> ????????????Write: 989 MB/s (8 ms) >>> ????????????Read: 7721 MB/s (1 ms) >>> >>> Job: seqrw-40gb-2mb-t1 >>> ????????????Write: 161 MB/s (12 ms) >>> ????????????Read: 2261 MB/s (0 ms) >>> >>> Job: seqrw-20gb-2mb-t2 >>> ????????????Write: 348 MB/s (11 ms) >>> ????????????Read: 4266 MB/s (1 ms) >>> >>> Job: seqrw-10gb-2mb-t4 >>> ????????????Write: 626 MB/s (13 ms) >>> ????????????Read: 4949 MB/s (1 ms) >>> >>> Job: seqrw-05gb-2mb-t8 >>> ????????????Write: 1154 MB/s (14 ms) >>> ????????????Read: 7007 MB/s (2 ms) >>> >>> Job: seqrw-40gb-4mb-t1 >>> ????????????Write: 161 MB/s (25 ms) >>> ????????????Read: 2083 MB/s (1 ms) >>> >>> Job: seqrw-20gb-4mb-t2 >>> ????????????Write: 352 MB/s (23 ms) >>> ????????????Read: 4317 MB/s (2 ms) >>> >>> Job: seqrw-10gb-4mb-t4 >>> ????????????Write: 696 MB/s (23 ms) >>> ????????????Read: 7358 MB/s (2 ms) >>> >>> Job: seqrw-05gb-4mb-t8 >>> ????????????Write: 1251 MB/s (25 ms) >>> ????????????Read: 6707 MB/s (5 ms) >>> >>> >>> So with fio I get a very nice read speed, but the write is >>> horrendous and I cannot find what causes it. I have looked at >>> affinity settings for the mmfsd process but not sure I fully >>> understand it. But no matter what I set it to, I see no difference. >>> >>> I have "played" with the bios and tried with/without hyperthreading, >>> numa and so on. And nothing affects atleast the blackmagic disk >>> speed test. >>> >>> the current settings for this host is like below. I write "current" >>> because I have tested a few different settings here but nothing >>> affects the write speed. maxTcpConnsPerNodeConn for sure bumped the >>> read speed though. >>> >>> nsdMaxWorkerThreads 16 >>> prefetchPct 60 >>> maxTcpConnsPerNodeConn 8 >>> maxMBpS 14000 >>> >>> >>> Does anyone have any suggestions or ideas on how to troubleshoot this? >>> >>> Thanks >>> >>> >>> >>> -- >>> >>> Henrik Cednert */ *?+ 46 704 71 89 54 */*? CTO */ OnePost >>> *(formerly?Filmlance Post) >>> >>> ?? *OnePost*, formerly Filmlance's post-production, is now an >>> independent part of the Banijay Group. >>> New name, same team ? business as usual at OnePost. >>> >>> >>> >>> >>> _______________________________________________ >>> gpfsug-discuss mailing list >>> gpfsug-discuss at gpfsug.org >>> http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org >> -- >> Karlsruhe Institute of Technology (KIT) >> Scientific Computing Centre (SCC) >> Scientific Data Management (SDM) >> >> Uwe Falke >> >> Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 >> D-76344 Eggenstein-Leopoldshafen >> >> Tel: +49 721 608 28024 >> Email:uwe.falke at kit.edu >> www.scc.kit.edu >> >> Registered office: >> Kaiserstra?e 12, 76131 Karlsruhe, Germany >> >> KIT ? The Research University in the Helmholtz Association >> >> _______________________________________________ >> gpfsug-discuss mailing list >> gpfsug-discuss at gpfsug.org >> http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org > -- > Karlsruhe Institute of Technology (KIT) > Scientific Computing Centre (SCC) > Scientific Data Management (SDM) > > Uwe Falke > > Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 > D-76344 Eggenstein-Leopoldshafen > > Tel: +49 721 608 28024 > Email:uwe.falke at kit.edu > www.scc.kit.edu > > Registered office: > Kaiserstra?e 12, 76131 Karlsruhe, Germany > > KIT ? The Research University in the Helmholtz Association > > _______________________________________________ > gpfsug-discuss mailing list > gpfsug-discuss at gpfsug.org > http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email:uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 5814 bytes Desc: S/MIME Cryptographic Signature URL: From henrik.cednert at onepost.se Wed Sep 4 14:59:10 2024 From: henrik.cednert at onepost.se (Henrik Cednert) Date: Wed, 4 Sep 2024 13:59:10 +0000 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: <33582c29-da8f-4cc1-a8c2-3debb8373a71@kit.edu> References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> <33582c29-da8f-4cc1-a8c2-3debb8373a71@kit.edu> Message-ID: Thanks Uwe I will have to digest what you write for a while, at this time of day some of it flies over my head (...and some probably will no matter the time of day). =) But to add some info about the file system. File system attributes for /dev/mmfs1: ====================================== flag value description ------------------- ------------------------ ----------------------------------- -f 8192 Minimum fragment (subblock) size in bytes (system pool) 131072 Minimum fragment (subblock) size in bytes (other pools) -i 512 Inode size in bytes -I 32768 Indirect block size in bytes -m 2 Default number of metadata replicas -M 2 Maximum number of metadata replicas -r 1 Default number of data replicas -R 2 Maximum number of data replicas -j scatter Block allocation type -D nfs4 File locking semantics in effect -k nfs4 ACL semantics in effect -n 32 Estimated number of nodes that will mount file system -B 524288 Block size (system pool) 8388608 Block size (other pools) Regarding, >The iohist snippet for reads comprises 74 IOs in about 0.854s, this relates to roughly 690MiB/s, far from the 10-fold value you reported. I'm not sure about that 10-fold value you refer to here. 690MiB is pretty much exactly what I saw reported in the disk speed test I was running when extracting that data. I will re-run my fio-tests on the other systems so that I have fresh values. If I'm sure that they are trustworthy...? Can one ever be? Network graphs and reported fio results are all I have to lean against. Attaching a few lines of --iohist for one of those 10GbE clients that currently is running my batch fio test. mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 15:46:18.576674 W data 6:161531625472 16384 51,417 cli C0A82DD5:63877BCE 192.168.45.214 15:46:18.628591 W logData 3:1877521729 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.628591 W logData 1:1869336897 2 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.630088 W data 6:12630425600 16384 13,978 cli C0A82DD5:63877BCE 192.168.45.214 15:46:18.630088 R data 14:12401655808 16384 46,425 cli C0A82DD5:63877BD6 192.168.45.214 15:46:18.676514 W data 14:166107021312 16384 12,979 cli C0A82DD5:63877BD6 192.168.45.214 15:46:18.689493 W logData 3:1877521730 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.689493 W logData 1:1869336898 1 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.690991 W data 14:12401655808 16384 13,977 cli C0A82DD5:63877BD6 192.168.45.214 15:46:18.690991 R data 22:11486576640 16384 32,947 cli C0A82DD5:63877BDF 192.168.45.214 15:46:18.723938 W data 22:168394719232 16384 14,476 cli C0A82DD5:63877BDF 192.168.45.214 15:46:18.738414 W logData 3:1877521730 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.738414 W logData 1:1869336898 1 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.739912 W data 22:11486576640 16384 14,977 cli C0A82DD5:63877BDF 192.168.45.214 15:46:18.740411 R data 7:13087965184 16384 38,438 cli C0A82DD5:63877BCF 192.168.45.213 15:46:18.778849 W data 7:170224877568 16384 46,925 cli C0A82DD5:63877BCF 192.168.45.213 15:46:18.825774 W logData 3:1877521730 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.825774 W logData 1:1869336898 2 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.827271 W data 7:13087965184 16384 11,482 cli C0A82DD5:63877BCF 192.168.45.213 15:46:18.827271 R data 15:12172886016 16384 35,443 cli C0A82DD5:63877BD8 192.168.45.213 15:46:18.862714 W data 15:167937179648 16384 11,482 cli C0A82DD5:63877BD8 192.168.45.213 15:46:18.874696 W logData 3:1877521731 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.874696 W logData 1:1869336899 1 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.876194 W data 15:12172886016 16384 12,479 cli C0A82DD5:63877BD8 192.168.45.213 15:46:18.876194 R data 23:12401655808 16384 30,949 cli C0A82DD5:63877BE0 192.168.45.213 15:46:18.907143 W data 23:164276862976 16384 13,978 cli C0A82DD5:63877BE0 192.168.45.213 15:46:18.921121 W logData 3:1877521731 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.921121 W logData 1:1869336899 2 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.922618 W data 23:12401655808 16384 13,479 cli C0A82DD5:63877BE0 192.168.45.213 15:46:18.922618 R data 8:13774274560 16384 32,947 cli C0A82DD5:63877BD0 192.168.45.214 15:46:18.955565 W data 8:163133014016 16384 53,415 cli C0A82DD5:63877BD0 192.168.45.214 15:46:19.008980 W logData 3:1877521732 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:19.008980 W logData 1:1869336900 1 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:19.010477 W data 8:13774274560 16384 13,478 cli C0A82DD5:63877BD0 192.168.45.214 15:46:19.010976 R data 16:13087965184 16384 37,440 cli C0A82DD5:63877BD9 192.168.45.214 15:46:19.048416 W data 16:154439761920 16384 13,978 cli C0A82DD5:63877BD9 192.168.45.214 15:46:19.062394 W logData 3:1877521732 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:19.062394 W logData 1:1869336900 2 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:19.063891 W data 16:13087965184 16384 13,478 cli C0A82DD5:63877BD9 192.168.45.214 15:46:19.063891 R data 24:13087965184 16384 34,445 cli C0A82DD5:63877BE1 192.168.45.214 15:46:19.098336 W data 24:150779445248 16384 13,479 cli C0A82DD5:63877BE1 192.168.45.214 15:46:19.111814 W logData 3:1877521733 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Wednesday, 4 September 2024 14:39 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, the writes look strange: you seem to use a blocksize of 8MiB on your file system. The reads are entirely full block reads (16ki sectors a 0.5kiB => 8MiB), but of the writes less than half are comprising full blocks. The service time for the write I/O seems to correlate prtetty well with I/O size, i.e. is rate-bound, not IOps-bound (which again could be rooted in the data link to your NSD server): 8MiB transfers take about 6.0...6.5ms, related to approx 1.3GiB/s, a bunch of 10800-sector transfers take 4.0ms on rough average, this relates to 1.3GiB/s as well. As for the read data: none of the seen service times comes close to the 1ms you've reported below, confirming those numbers have nothing to do with your real data traffic. The service times go down to 5ms which is rather low, in particular compared to the somewhat higher service times for writes. This could mean those read I/Os with times of 6ms and below are served from the storage systems cache (NSD servers do buffer but not cache). Then, the read IOs with substantially higher service times are the ones really served from disk which again would mean your storage is the culprit . But then the question arises wy do your other machines behave well? But given the caching issues indicated above - are you sure the measurements at your other machines are trustworthy? The iohist snippet for reads comprises 74 IOs in about 0.854s, this relates to roughly 690MiB/s, far from the 10-fold value you reported. So, while reason is to assume some IOs are served from the storage cache, much more of those seems to come from the client's cache Begfore that hasn't been clarified I think you do not need to play with the BIOS and stuff :-) Uwe On 04.09.24 13:28, Henrik Cednert wrote: Adding a snippet from iohist when reading as well. >mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 13:24:59.180277 R data 19:1491763200 16384 32,433 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.212710 R data 8:805453824 16384 5,987 cli C0A82DD5:63877BD0 192.168.45.214 13:24:59.220698 R data 8:805453824 16384 6,012 cli C0A82DD5:63877BD0 192.168.45.214 13:24:59.226710 R data 16:805453824 16384 9,129 cli C0A82DD5:63877BD9 192.168.45.214 13:24:59.226710 R data 12:1491763200 16384 18,526 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.226710 R data 20:2178072576 16384 23,520 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.251229 R data 16:805453824 16384 5,497 cli C0A82DD5:63877BD9 192.168.45.214 13:24:59.257730 R data 24:805453824 16384 5,990 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.257730 R data 5:1720532992 16384 33,458 cli C0A82DD5:63877BCD 192.168.45.213 13:24:59.257730 R data 28:1720532992 16384 33,458 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.292189 R data 24:805453824 16384 4,992 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.300176 R data 24:805453824 16384 4,991 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.306169 R data 9:805453824 16384 7,987 cli C0A82DD5:63877BD1 192.168.45.213 13:24:59.306169 R data 13:1720532992 16384 25,471 cli C0A82DD5:63877BD5 192.168.45.213 13:24:59.306169 R data 21:1720532992 16384 28,466 cli C0A82DD5:63877BDE 192.168.45.213 13:24:59.335637 R data 9:805453824 16384 4,994 cli C0A82DD5:63877BD1 192.168.45.213 13:24:59.341629 R data 17:805453824 16384 3,993 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.341629 R data 6:1720532992 16384 27,471 cli C0A82DD5:63877BCE 192.168.45.214 13:24:59.341629 R data 14:1720532992 16384 44,953 cli C0A82DD5:63877BD6 192.168.45.214 13:24:59.387584 R data 17:805453824 16384 5,990 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.394572 R data 17:805453824 16384 7,993 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.402565 R data 25:805453824 16384 4,991 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.402565 R data 22:1720532992 16384 26,309 cli C0A82DD5:63877BDF 192.168.45.214 13:24:59.402565 R data 7:1720532992 16384 142,856 cli C0A82DD5:63877BCF 192.168.45.213 13:24:59.546419 R data 25:805453824 16384 7,992 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.556408 R data 25:805453824 16384 7,987 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.564395 R data 10:805453824 16384 5,505 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.564395 R data 23:1720532992 16384 28,053 cli C0A82DD5:63877BE0 192.168.45.213 13:24:59.564395 R data 15:1720532992 16384 33,044 cli C0A82DD5:63877BD8 192.168.45.213 13:24:59.598437 R data 10:805453824 16384 5,504 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.604939 R data 18:805453824 16384 4,993 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.604939 R data 8:1720532992 16384 36,015 cli C0A82DD5:63877BD0 192.168.45.214 13:24:59.609932 R data 16:1720532992 16384 42,010 cli C0A82DD5:63877BD9 192.168.45.214 13:24:59.652940 R data 18:805453824 16384 7,989 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.662434 R data 18:805453824 16384 6,994 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.669428 R data 26:805453824 16384 8,986 cli C0A82DD5:63877BE3 192.168.45.214 13:24:59.669428 R data 24:1720532992 16384 20,308 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.669428 R data 9:1720532992 16384 25,812 cli C0A82DD5:63877BD1 192.168.45.213 13:24:59.696239 R data 26:805453824 16384 6,989 cli C0A82DD5:63877BE3 192.168.45.214 13:24:59.703228 R data 11:805453824 16384 4,992 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.703228 R data 25:1720532992 16384 17,976 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.703228 R data 17:1720532992 16384 22,481 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.725709 R data 11:805453824 16384 4,992 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.731699 R data 11:805453824 16384 8,986 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.740685 R data 19:805453824 16384 4,997 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.740685 R data 18:1720532992 16384 19,486 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.740685 R data 10:1720532992 16384 27,474 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.769157 R data 19:805453824 16384 4,997 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.774154 R data 27:805453824 16384 4,992 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.774154 R data 11:1720532992 16384 22,476 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.774154 R data 26:1720532992 16384 29,464 cli C0A82DD5:63877BE3 192.168.45.214 13:24:59.803618 R data 27:805453824 16384 4,997 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.809614 R data 27:805453824 16384 7,987 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.817601 R data 12:805453824 16384 8,500 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.817601 R data 27:1720532992 16384 9,499 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.817601 R data 19:1720532992 16384 74,075 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.892674 R data 12:805453824 16384 4,997 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.897671 R data 20:1491763200 16384 4,992 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.898670 R data 12:1720532992 16384 24,479 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.898670 R data 20:2406842368 16384 26,476 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.925146 R data 20:1491763200 16384 8,990 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.935430 R data 20:1491763200 16384 4,691 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.940121 R data 28:1034223616 16384 4,312 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.940121 R data 28:1949302784 16384 26,664 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.940121 R data 5:1949302784 16384 34,831 cli C0A82DD5:63877BCD 192.168.45.213 13:24:59.975955 R data 28:1034223616 16384 5,583 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.981539 R data 5:1034223616 16384 5,168 cli C0A82DD5:63877BCD 192.168.45.213 13:24:59.981539 R data 21:1949302784 16384 25,392 cli C0A82DD5:63877BDE 192.168.45.213 13:24:59.981539 R data 13:1949302784 16384 34,412 cli C0A82DD5:63877BD5 192.168.45.213 13:25:00.016950 R data 5:1034223616 16384 7,991 cli C0A82DD5:63877BCD 192.168.45.213 13:25:00.026938 R data 5:1034223616 16384 7,750 cli C0A82DD5:63877BCD 192.168.45.213 13:25:00.034688 R data 13:1034223616 16384 5,498 cli C0A82DD5:63877BD5 192.168.45.213 13:25:00.034688 R data 6:1949302784 16384 27,661 cli C0A82DD5:63877BCE 192.168.45.214 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Wednesday, 4 September 2024 13:24 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hello My theory is that it's windows 11 related, or the combo windows 11 and this hardware. I guess the only thing to know for sure, is to boot it into *nix and install gpfs there and test. Which I guess isn't the worst of ideas at this stage. --iohist spits out a pretty hefty chuck of data. Below is a snippet from when I did a write test: mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 13:17:04.088621 W data 10:2178073088 10800 3,994 cli C0A82DD5:63877BD2 192.168.45.214 13:17:04.092614 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.092614 W inode 2:3040332814 1 0,999 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.093613 W data 10:2178083888 5072 2,995 cli C0A82DD5:63877BD2 192.168.45.214 13:17:04.094611 W data 18:2178072576 16384 6,502 cli C0A82DD5:63877BDB 192.168.45.214 13:17:04.101113 W logData 2:1869321537 2 0,998 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.101113 W logData 3:1877542209 2 0,998 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.103109 W data 18:2178078304 10656 3,994 cli C0A82DD5:63877BDB 192.168.45.214 13:17:04.103109 W data 26:2178072576 16384 6,529 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.109638 W logData 2:1869321538 2 0,994 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.109638 W logData 3:1877542210 2 0,994 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.111634 W data 26:2178072720 10800 3,992 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.115626 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.115626 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.116626 W data 26:2178083520 5440 2,995 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.117629 W data 11:2178072576 16384 4,987 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.122616 W logData 2:1869321539 2 0,999 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.122616 W logData 3:1877542211 2 0,999 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.124613 W data 11:2178077936 10800 3,999 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.128612 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.128612 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.129609 W data 11:2178088736 224 0,000 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.129609 W data 19:2178072576 16384 6,495 cli C0A82DD5:63877BDC 192.168.45.213 13:17:04.136104 W logData 2:1869321540 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.136104 W logData 3:1877542212 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.137103 W data 19:2178083152 5808 2,999 cli C0A82DD5:63877BDC 192.168.45.213 13:17:04.138100 W data 27:2178072576 16384 5,990 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.144091 W logData 2:1869321540 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.144091 W logData 3:1877542212 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.146088 W data 27:2178077568 10800 4,005 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.150092 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.150092 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.151092 W data 27:2178088368 592 0,000 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.151092 W data 12:2178072576 16384 4,995 cli C0A82DD5:63877BD4 192.168.45.214 13:17:04.156086 W logData 2:1869321541 2 0,996 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.156086 W logData 3:1877542213 2 0,996 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.157082 W data 12:2178082784 6176 2,994 cli C0A82DD5:63877BD4 192.168.45.214 13:17:04.158083 W data 20:2178072576 16384 7,498 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.165581 W logData 2:1869321542 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.165581 W logData 3:1877542214 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.167578 W data 20:2178077200 10800 2,994 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.170572 W inode 4:3060742158 1 0,996 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.171568 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.171568 W data 20:2178088000 960 1,001 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.172569 W data 28:2864381952 16384 5,988 cli C0A82DD5:63877BE5 192.168.45.214 13:17:04.178557 W logData 2:1869321543 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.178557 W logData 3:1877542215 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.179560 W data 28:2864391792 6544 2,995 cli C0A82DD5:63877BE5 192.168.45.214 13:17:04.179560 W data 5:2406842368 16384 5,987 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.185548 W logData 2:1869321544 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.185548 W logData 3:1877542216 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.186549 W data 5:2406846624 10800 4,993 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.191542 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.191542 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.192540 W data 5:2406857424 1328 0,995 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.193535 W data 13:2406842368 16384 6,019 cli C0A82DD5:63877BD5 192.168.45.213 13:17:04.199554 W logData 2:1869321544 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.199554 W logData 3:1877542216 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.200551 W data 13:2406851840 6912 2,997 cli C0A82DD5:63877BD5 192.168.45.213 13:17:04.201554 W data 21:2406842368 16384 5,912 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.207466 W logData 2:1869321545 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.207466 W logData 3:1877542217 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.208465 W data 21:2406846256 10800 3,990 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.212456 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.212456 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.213456 W data 21:2406857056 1696 1,998 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.214457 W data 6:2406842368 16384 5,015 cli C0A82DD5:63877BCE 192.168.45.214 13:17:04.219472 W logData 2:1869321546 2 1,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.219472 W logData 3:1877542218 2 1,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.221474 W data 6:2406851472 7280 1,994 cli C0A82DD5:63877BCE 192.168.45.214 13:17:04.221474 W data 14:2406842368 16384 7,502 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.228976 W logData 2:1869321547 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.228976 W logData 3:1877542219 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.229971 W data 14:2406845888 10800 3,994 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.233965 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.233965 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.234963 W data 14:2406856688 2064 1,999 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.235965 W data 22:2406842368 16384 4,993 cli C0A82DD5:63877BDF 192.168.45.214 13:17:04.240958 W logData 2:1869321547 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.240958 W logData 3:1877542219 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.241974 W data 22:2406851104 7648 2,531 cli C0A82DD5:63877BDF 192.168.45.214 13:17:04.242977 W data 7:2406842368 16384 5,526 cli C0A82DD5:63877BCF 192.168.45.213 13:17:04.248503 W logData 2:1869321548 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.248503 W logData 3:1877542220 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.249503 W data 7:2406845520 10800 3,994 cli C0A82DD5:63877BCF 192.168.45.213 13:17:04.253497 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Wednesday, 4 September 2024 12:59 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, given you see read latencies of 1 ms, you do not get the data from disk but from some cache (on whatever level). From spinning disks, you can never expect such read latencies (mind that GPFS block reading, even if sequential from the application's PoV, typically translates to random I/O at the physical disk level). So, I do not know what the latencies on your other measurements (bypassing GPFS) were but the numbers below do not represent sustained high-scale throughputs, apparently. It is nevertheless strange that your write rates are much below reads (and write latencies are that high) -- from my experience with different systems, when hammering GPFS with usual storage backends with both read and write requests, the writes tend to prevail. Your waiters indicate that the problem is above GPFS: GPFS is able to serve all I/O threads within a few ms, and there is not a long list of pending IOs. Sorry, the iohistory options is mmdiag --iohist But to me it looks like not GPFS is the culprit here. Uwe On 04.09.24 11:08, Henrik Cednert wrote: Hi Uwe Thanks. Worth noting is that we have win10 ltsc, win 2019 and have had a single CPU win 11 22h2 (as a test) clients that all perform as expected. Those machines is older though and connected with 10-40GbE, those client max out their NIC in read and write. Let me know if i missed something important here. Thanks again. Setups is Client: * Supermicro Workstation * Intel(R) Xeon(R) Gold 6418H 2.10 GHz (2 processors) * Mellanox ConnectX-6 Dx connected with 100GbE over dedicated vlan via mellanox sn2100. * Windows 11 Pro for Workstations, 22H2 Storage setup * 3 x 84 bay seagate chassis with spinning disks. * Storage connected with redundant 12Gb SAS to 2 x Storage node servers * 2 x mellanox sn2100 * The 2 storage node servers are for this vlan connected with 100GbE to each switch, so in total 4 x 100GbE. And the switches are connected with 2 * 100GbE I tested the commands you suggested. They are both new to me so not sure what the output is supposed to be, looks like -iohistory isn't available in windows. I ran --waiters a few times, as seen below. Not sure what the expected output is from that. mmdiag --waiters === mmdiag: waiters === Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 18616 MsgHandler at getData: for In function sendMessage Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 25084 WritebehindWorkerThread: on ThCond 0x31A7C360 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 16780 FsyncHandlerThread: on ThCond 0x37FFDAB0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 30308 MsgHandler at getData: for In function sendMessage C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0055 sec since 2024-09-04_10:05:21, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x37A25FF0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0029 sec since 2024-09-04_10:05:23, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x38281DE0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0019 sec since 2024-09-04_10:05:25, monitored, thread 11832 PrefetchWorkerThread: on ThCond 0x38278D20 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 16780 AcquireBRTHandlerThread: on ThCond 0x37A324E0 (MsgRecordCondvar), reason 'RPC wait' for tmMsgBRRevoke on node 192.168.45.161 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 2576 RangeRevokeWorkerThread: on ThCond 0x5419DAA0 (BrlObjCondvar), reason 'waiting because of local byte range lock conflict' C:\Users\m5-tkd01> C:\Users\m5-tkd01>mmdiag --iohistory Unrecognized option: --iohistory. Run mmdiag --help for the option list -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Tuesday, 3 September 2024 17:35 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, Henrik, while I am not using Windows I'd start investigating the usual things (see below). But first you should describe your set-up better. Where are the NSDs : locally attached to the Windows box? In some NSD servers? If the latter -- what is the link to the NSD servers? via your GbE link? FC? IB? separate Ethernet? What type of storage? Spinning Disks? Flash? How long are your I/Os waiting on the client (compare that to the waiting times on the NSD server if applicable)? not sure whether that is available on Windows, but mmdiag --waiters mmdiag --iohistory might be of use. Somewhere in the chain from your application to the storage backend there is a delay and you should first find out where that occurs I think. Bye Uwe On 03.09.24 14:10, Henrik Cednert wrote: Still no solution here regarding this. Have tested other cables. Have tested to change tcp window size, no change Played with numa in the bios, no change Played with hyperthreading in bios, no change Have anyone managed to get some speed out of windows 11 and gpfs? -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Friday, 9 August 2024 17:25 To: gpfsug-discuss at gpfsug.org Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. VARNING: DETTA ?R ETT EXTERNT MAIL. Klicka inte p? n?gra l?nkar oavsett hur legitima de verkar utan att verifiera. Hello I have some issues with write performance on a windows 11 pro system and I'm out of ideas here. Hopefully someone here have some bright ideas and/or experience of GPFS on Windows 11? The system is a: Windows 11 Pro 22H2 2 x Intel(R) Xeon(R) Gold 6418H 2.10 GHz 512 GB RAM GPFS 5.1.9.4 Mellanox ConnectX 6 Dx 100GbE connected to Mellanox Switch with 5m Mellanox DAC. Before deploying this workstation we had a single socket system as a test bench where we got 60 GbE in both directons with iPerf and around 6GB/sec write and 3GB/sec read from the system over GPFS (fio tests, same tests as furhter down here). With that system I had loads of issues before getting to that point though. MS Defender had to be forcefully disabled via regedit some other tweaks. All those tweaks have been performed in this new system as well, but I can't get the proper speed out of it. On this new system and with iPerf to the storage servers I get around 50-60GbE in both directions and send and receive. If I mount the storage over SMB and 100GbE via the storage gateway servers I get around 3GB/sec read and write with Blackmagics Disk speed test. I have not tweaked the system for samba performande, just a test to see what it would give and part of the troubleshooting. If I run Blackmagics diskspeed test to the GPFS mount I instead get around 700MB/sec write and 400MB/sec read. Starting to think that the Blackmagic test might not run properly on this machine with these CPUs though. Or it's related to the mmfsd process maybe, how that threads or not threads...? But if we instead look at fio. I have a bat script that loops through a bunch of FIO-tests. A test that I have been using over the years so that we easily can benchmark all deployed systems with the exakt same tests. The tests are named like: seqrw-gb-mb-t The result when I run this is like the below list. Number in parenthesis is the by fio reported latency. Job: seqrw-40gb-1mb-t1 ????????????Write: 162 MB/s (6 ms) ????????????Read: 1940 MB/s (1 ms) Job: seqrw-20gb-1mb-t2 ????????????Write: 286 MB/s (7 ms) ????????????Read: 3952 MB/s (1 ms) Job: seqrw-10gb-1mb-t4 ????????????Write: 549 MB/s (7 ms) ????????????Read: 6987 MB/s (1 ms) Job: seqrw-05gb-1mb-t8 ????????????Write: 989 MB/s (8 ms) ????????????Read: 7721 MB/s (1 ms) Job: seqrw-40gb-2mb-t1 ????????????Write: 161 MB/s (12 ms) ????????????Read: 2261 MB/s (0 ms) Job: seqrw-20gb-2mb-t2 ????????????Write: 348 MB/s (11 ms) ????????????Read: 4266 MB/s (1 ms) Job: seqrw-10gb-2mb-t4 ????????????Write: 626 MB/s (13 ms) ????????????Read: 4949 MB/s (1 ms) Job: seqrw-05gb-2mb-t8 ????????????Write: 1154 MB/s (14 ms) ????????????Read: 7007 MB/s (2 ms) Job: seqrw-40gb-4mb-t1 ????????????Write: 161 MB/s (25 ms) ????????????Read: 2083 MB/s (1 ms) Job: seqrw-20gb-4mb-t2 ????????????Write: 352 MB/s (23 ms) ????????????Read: 4317 MB/s (2 ms) Job: seqrw-10gb-4mb-t4 ????????????Write: 696 MB/s (23 ms) ????????????Read: 7358 MB/s (2 ms) Job: seqrw-05gb-4mb-t8 ????????????Write: 1251 MB/s (25 ms) ????????????Read: 6707 MB/s (5 ms) So with fio I get a very nice read speed, but the write is horrendous and I cannot find what causes it. I have looked at affinity settings for the mmfsd process but not sure I fully understand it. But no matter what I set it to, I see no difference. I have "played" with the bios and tried with/without hyperthreading, numa and so on. And nothing affects atleast the blackmagic disk speed test. the current settings for this host is like below. I write "current" because I have tested a few different settings here but nothing affects the write speed. maxTcpConnsPerNodeConn for sure bumped the read speed though. nsdMaxWorkerThreads 16 prefetchPct 60 maxTcpConnsPerNodeConn 8 maxMBpS 14000 Does anyone have any suggestions or ideas on how to troubleshoot this? Thanks -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: From henrik.cednert at onepost.se Wed Sep 4 15:08:26 2024 From: henrik.cednert at onepost.se (Henrik Cednert) Date: Wed, 4 Sep 2024 14:08:26 +0000 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> <33582c29-da8f-4cc1-a8c2-3debb8373a71@kit.edu> Message-ID: Apologies, the machine I sent the additional --iohist from did not perform as expected. Looks like it needs a reboot. This is from one of the other that is dunning fio atm and from the looks of network graphs is maxing out its 10GbE interface. 130|Administrator at m5-blade01-c:~ $ mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 16:04:32.318864 R data 22:27958001664 16384 1465,152 cli C0A82DD5:63877BDF 192.168.45.214 16:04:32.722717 R data 19:30245699584 16384 1069,286 cli C0A82DD5:63877BDC 192.168.45.213 16:04:32.720721 R data 11:29559390208 16384 1077,772 cli C0A82DD5:63877BD3 192.168.45.213 16:04:32.314870 R data 6:28873080832 16384 1491,610 cli C0A82DD5:63877BCE 192.168.45.214 16:04:32.316867 R data 14:28644311040 16384 1495,603 cli C0A82DD5:63877BD6 192.168.45.214 16:04:32.724714 R data 27:30016929792 16384 1093,247 cli C0A82DD5:63877BE4 192.168.45.213 16:04:32.766646 R data 5:30932008960 16384 1056,806 cli C0A82DD5:63877BCD 192.168.45.213 16:04:32.334339 R data 16:29559390208 16384 1493,606 cli C0A82DD5:63877BD9 192.168.45.214 16:04:32.770640 R data 21:29101850624 16384 1071,782 cli C0A82DD5:63877BDE 192.168.45.213 16:04:32.768643 R data 13:28415541248 16384 1080,768 cli C0A82DD5:63877BD5 192.168.45.213 16:04:32.344323 R data 10:29330620416 16384 1506,585 cli C0A82DD5:63877BD2 192.168.45.214 16:04:32.328349 R data 8:29559390208 16384 1525,056 cli C0A82DD5:63877BD0 192.168.45.214 16:04:32.857001 R data 15:29330620416 16384 1009,881 cli C0A82DD5:63877BD8 192.168.45.213 16:04:32.377270 R data 18:28873080832 16384 1496,601 cli C0A82DD5:63877BDB 192.168.45.214 16:04:32.858998 R data 23:30016929792 16384 1018,867 cli C0A82DD5:63877BE0 192.168.45.213 16:04:32.855004 R data 7:29788160000 16384 1037,837 cli C0A82DD5:63877BCF 192.168.45.213 16:04:32.379267 R data 26:30703239168 16384 1516,569 cli C0A82DD5:63877BE3 192.168.45.214 16:04:32.336336 R data 24:29330620416 16384 1560,499 cli C0A82DD5:63877BE1 192.168.45.214 16:04:32.866985 R data 9:31389548544 16384 1038,336 cli C0A82DD5:63877BD1 192.168.45.213 16:04:32.387254 R data 12:29559390208 16384 1530,547 cli C0A82DD5:63877BD4 192.168.45.214 16:04:32.868982 R data 17:30245699584 16384 1050,317 cli C0A82DD5:63877BDA 192.168.45.213 16:04:32.870979 R data 25:31160778752 16384 1060,301 cli C0A82DD5:63877BE2 192.168.45.213 16:04:32.389251 R data 20:29330620416 16384 1551,513 cli C0A82DD5:63877BDD 192.168.45.214 16:04:32.431683 R data 6:29101850624 16384 1510,080 cli C0A82DD5:63877BCE 192.168.45.214 16:04:32.898934 R data 11:29788160000 16384 1043,827 cli C0A82DD5:63877BD3 192.168.45.213 16:04:32.900931 R data 19:30703239168 16384 1056,806 cli C0A82DD5:63877BDC 192.168.45.213 16:04:32.433680 R data 14:28873080832 16384 1531,545 cli C0A82DD5:63877BD6 192.168.45.214 16:04:32.902928 R data 27:30245699584 16384 1066,291 cli C0A82DD5:63877BE4 192.168.45.213 16:04:32.988291 R data 5:31160778752 16384 997,401 cli C0A82DD5:63877BCD 192.168.45.213 16:04:32.423696 R data 28:29330620416 16384 1562,496 cli C0A82DD5:63877BE5 192.168.45.214 16:04:32.444163 R data 8:29788160000 16384 1544,025 cli C0A82DD5:63877BD0 192.168.45.214 16:04:32.992284 R data 21:29330620416 16384 1001,894 cli C0A82DD5:63877BDE 192.168.45.213 16:04:32.468125 R data 16:29788160000 16384 1544,524 cli C0A82DD5:63877BD9 192.168.45.214 16:04:32.990287 R data 13:28644311040 16384 1023,859 cli C0A82DD5:63877BD5 192.168.45.213 16:04:33.036713 R data 15:29559390208 16384 985,421 cli C0A82DD5:63877BD8 192.168.45.213 16:04:32.435676 R data 22:28186771456 16384 1598,937 cli C0A82DD5:63877BDF 192.168.45.214 16:04:32.502070 R data 10:29559390208 16384 1534,541 cli C0A82DD5:63877BD2 192.168.45.214 16:04:33.038710 R data 23:30245699584 16384 1004,390 cli C0A82DD5:63877BE0 192.168.45.213 16:04:33.034217 R data 7:30016929792 16384 1015,372 cli C0A82DD5:63877BCF 192.168.45.213 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Wednesday, 4 September 2024 15:59 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Thanks Uwe I will have to digest what you write for a while, at this time of day some of it flies over my head (...and some probably will no matter the time of day). =) But to add some info about the file system. File system attributes for /dev/mmfs1: ====================================== flag value description ------------------- ------------------------ ----------------------------------- -f 8192 Minimum fragment (subblock) size in bytes (system pool) 131072 Minimum fragment (subblock) size in bytes (other pools) -i 512 Inode size in bytes -I 32768 Indirect block size in bytes -m 2 Default number of metadata replicas -M 2 Maximum number of metadata replicas -r 1 Default number of data replicas -R 2 Maximum number of data replicas -j scatter Block allocation type -D nfs4 File locking semantics in effect -k nfs4 ACL semantics in effect -n 32 Estimated number of nodes that will mount file system -B 524288 Block size (system pool) 8388608 Block size (other pools) Regarding, >The iohist snippet for reads comprises 74 IOs in about 0.854s, this relates to roughly 690MiB/s, far from the 10-fold value you reported. I'm not sure about that 10-fold value you refer to here. 690MiB is pretty much exactly what I saw reported in the disk speed test I was running when extracting that data. I will re-run my fio-tests on the other systems so that I have fresh values. If I'm sure that they are trustworthy...? Can one ever be? Network graphs and reported fio results are all I have to lean against. Attaching a few lines of --iohist for one of those 10GbE clients that currently is running my batch fio test. mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 15:46:18.576674 W data 6:161531625472 16384 51,417 cli C0A82DD5:63877BCE 192.168.45.214 15:46:18.628591 W logData 3:1877521729 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.628591 W logData 1:1869336897 2 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.630088 W data 6:12630425600 16384 13,978 cli C0A82DD5:63877BCE 192.168.45.214 15:46:18.630088 R data 14:12401655808 16384 46,425 cli C0A82DD5:63877BD6 192.168.45.214 15:46:18.676514 W data 14:166107021312 16384 12,979 cli C0A82DD5:63877BD6 192.168.45.214 15:46:18.689493 W logData 3:1877521730 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.689493 W logData 1:1869336898 1 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.690991 W data 14:12401655808 16384 13,977 cli C0A82DD5:63877BD6 192.168.45.214 15:46:18.690991 R data 22:11486576640 16384 32,947 cli C0A82DD5:63877BDF 192.168.45.214 15:46:18.723938 W data 22:168394719232 16384 14,476 cli C0A82DD5:63877BDF 192.168.45.214 15:46:18.738414 W logData 3:1877521730 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.738414 W logData 1:1869336898 1 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.739912 W data 22:11486576640 16384 14,977 cli C0A82DD5:63877BDF 192.168.45.214 15:46:18.740411 R data 7:13087965184 16384 38,438 cli C0A82DD5:63877BCF 192.168.45.213 15:46:18.778849 W data 7:170224877568 16384 46,925 cli C0A82DD5:63877BCF 192.168.45.213 15:46:18.825774 W logData 3:1877521730 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.825774 W logData 1:1869336898 2 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.827271 W data 7:13087965184 16384 11,482 cli C0A82DD5:63877BCF 192.168.45.213 15:46:18.827271 R data 15:12172886016 16384 35,443 cli C0A82DD5:63877BD8 192.168.45.213 15:46:18.862714 W data 15:167937179648 16384 11,482 cli C0A82DD5:63877BD8 192.168.45.213 15:46:18.874696 W logData 3:1877521731 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.874696 W logData 1:1869336899 1 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.876194 W data 15:12172886016 16384 12,479 cli C0A82DD5:63877BD8 192.168.45.213 15:46:18.876194 R data 23:12401655808 16384 30,949 cli C0A82DD5:63877BE0 192.168.45.213 15:46:18.907143 W data 23:164276862976 16384 13,978 cli C0A82DD5:63877BE0 192.168.45.213 15:46:18.921121 W logData 3:1877521731 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:18.921121 W logData 1:1869336899 2 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:18.922618 W data 23:12401655808 16384 13,479 cli C0A82DD5:63877BE0 192.168.45.213 15:46:18.922618 R data 8:13774274560 16384 32,947 cli C0A82DD5:63877BD0 192.168.45.214 15:46:18.955565 W data 8:163133014016 16384 53,415 cli C0A82DD5:63877BD0 192.168.45.214 15:46:19.008980 W logData 3:1877521732 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:19.008980 W logData 1:1869336900 1 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:19.010477 W data 8:13774274560 16384 13,478 cli C0A82DD5:63877BD0 192.168.45.214 15:46:19.010976 R data 16:13087965184 16384 37,440 cli C0A82DD5:63877BD9 192.168.45.214 15:46:19.048416 W data 16:154439761920 16384 13,978 cli C0A82DD5:63877BD9 192.168.45.214 15:46:19.062394 W logData 3:1877521732 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 15:46:19.062394 W logData 1:1869336900 2 0,000 cli C0A82DD5:63877BC9 192.168.45.213 15:46:19.063891 W data 16:13087965184 16384 13,478 cli C0A82DD5:63877BD9 192.168.45.214 15:46:19.063891 R data 24:13087965184 16384 34,445 cli C0A82DD5:63877BE1 192.168.45.214 15:46:19.098336 W data 24:150779445248 16384 13,479 cli C0A82DD5:63877BE1 192.168.45.214 15:46:19.111814 W logData 3:1877521733 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Wednesday, 4 September 2024 14:39 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, the writes look strange: you seem to use a blocksize of 8MiB on your file system. The reads are entirely full block reads (16ki sectors a 0.5kiB => 8MiB), but of the writes less than half are comprising full blocks. The service time for the write I/O seems to correlate prtetty well with I/O size, i.e. is rate-bound, not IOps-bound (which again could be rooted in the data link to your NSD server): 8MiB transfers take about 6.0...6.5ms, related to approx 1.3GiB/s, a bunch of 10800-sector transfers take 4.0ms on rough average, this relates to 1.3GiB/s as well. As for the read data: none of the seen service times comes close to the 1ms you've reported below, confirming those numbers have nothing to do with your real data traffic. The service times go down to 5ms which is rather low, in particular compared to the somewhat higher service times for writes. This could mean those read I/Os with times of 6ms and below are served from the storage systems cache (NSD servers do buffer but not cache). Then, the read IOs with substantially higher service times are the ones really served from disk which again would mean your storage is the culprit . But then the question arises wy do your other machines behave well? But given the caching issues indicated above - are you sure the measurements at your other machines are trustworthy? The iohist snippet for reads comprises 74 IOs in about 0.854s, this relates to roughly 690MiB/s, far from the 10-fold value you reported. So, while reason is to assume some IOs are served from the storage cache, much more of those seems to come from the client's cache Begfore that hasn't been clarified I think you do not need to play with the BIOS and stuff :-) Uwe On 04.09.24 13:28, Henrik Cednert wrote: Adding a snippet from iohist when reading as well. >mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 13:24:59.180277 R data 19:1491763200 16384 32,433 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.212710 R data 8:805453824 16384 5,987 cli C0A82DD5:63877BD0 192.168.45.214 13:24:59.220698 R data 8:805453824 16384 6,012 cli C0A82DD5:63877BD0 192.168.45.214 13:24:59.226710 R data 16:805453824 16384 9,129 cli C0A82DD5:63877BD9 192.168.45.214 13:24:59.226710 R data 12:1491763200 16384 18,526 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.226710 R data 20:2178072576 16384 23,520 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.251229 R data 16:805453824 16384 5,497 cli C0A82DD5:63877BD9 192.168.45.214 13:24:59.257730 R data 24:805453824 16384 5,990 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.257730 R data 5:1720532992 16384 33,458 cli C0A82DD5:63877BCD 192.168.45.213 13:24:59.257730 R data 28:1720532992 16384 33,458 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.292189 R data 24:805453824 16384 4,992 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.300176 R data 24:805453824 16384 4,991 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.306169 R data 9:805453824 16384 7,987 cli C0A82DD5:63877BD1 192.168.45.213 13:24:59.306169 R data 13:1720532992 16384 25,471 cli C0A82DD5:63877BD5 192.168.45.213 13:24:59.306169 R data 21:1720532992 16384 28,466 cli C0A82DD5:63877BDE 192.168.45.213 13:24:59.335637 R data 9:805453824 16384 4,994 cli C0A82DD5:63877BD1 192.168.45.213 13:24:59.341629 R data 17:805453824 16384 3,993 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.341629 R data 6:1720532992 16384 27,471 cli C0A82DD5:63877BCE 192.168.45.214 13:24:59.341629 R data 14:1720532992 16384 44,953 cli C0A82DD5:63877BD6 192.168.45.214 13:24:59.387584 R data 17:805453824 16384 5,990 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.394572 R data 17:805453824 16384 7,993 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.402565 R data 25:805453824 16384 4,991 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.402565 R data 22:1720532992 16384 26,309 cli C0A82DD5:63877BDF 192.168.45.214 13:24:59.402565 R data 7:1720532992 16384 142,856 cli C0A82DD5:63877BCF 192.168.45.213 13:24:59.546419 R data 25:805453824 16384 7,992 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.556408 R data 25:805453824 16384 7,987 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.564395 R data 10:805453824 16384 5,505 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.564395 R data 23:1720532992 16384 28,053 cli C0A82DD5:63877BE0 192.168.45.213 13:24:59.564395 R data 15:1720532992 16384 33,044 cli C0A82DD5:63877BD8 192.168.45.213 13:24:59.598437 R data 10:805453824 16384 5,504 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.604939 R data 18:805453824 16384 4,993 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.604939 R data 8:1720532992 16384 36,015 cli C0A82DD5:63877BD0 192.168.45.214 13:24:59.609932 R data 16:1720532992 16384 42,010 cli C0A82DD5:63877BD9 192.168.45.214 13:24:59.652940 R data 18:805453824 16384 7,989 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.662434 R data 18:805453824 16384 6,994 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.669428 R data 26:805453824 16384 8,986 cli C0A82DD5:63877BE3 192.168.45.214 13:24:59.669428 R data 24:1720532992 16384 20,308 cli C0A82DD5:63877BE1 192.168.45.214 13:24:59.669428 R data 9:1720532992 16384 25,812 cli C0A82DD5:63877BD1 192.168.45.213 13:24:59.696239 R data 26:805453824 16384 6,989 cli C0A82DD5:63877BE3 192.168.45.214 13:24:59.703228 R data 11:805453824 16384 4,992 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.703228 R data 25:1720532992 16384 17,976 cli C0A82DD5:63877BE2 192.168.45.213 13:24:59.703228 R data 17:1720532992 16384 22,481 cli C0A82DD5:63877BDA 192.168.45.213 13:24:59.725709 R data 11:805453824 16384 4,992 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.731699 R data 11:805453824 16384 8,986 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.740685 R data 19:805453824 16384 4,997 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.740685 R data 18:1720532992 16384 19,486 cli C0A82DD5:63877BDB 192.168.45.214 13:24:59.740685 R data 10:1720532992 16384 27,474 cli C0A82DD5:63877BD2 192.168.45.214 13:24:59.769157 R data 19:805453824 16384 4,997 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.774154 R data 27:805453824 16384 4,992 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.774154 R data 11:1720532992 16384 22,476 cli C0A82DD5:63877BD3 192.168.45.213 13:24:59.774154 R data 26:1720532992 16384 29,464 cli C0A82DD5:63877BE3 192.168.45.214 13:24:59.803618 R data 27:805453824 16384 4,997 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.809614 R data 27:805453824 16384 7,987 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.817601 R data 12:805453824 16384 8,500 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.817601 R data 27:1720532992 16384 9,499 cli C0A82DD5:63877BE4 192.168.45.213 13:24:59.817601 R data 19:1720532992 16384 74,075 cli C0A82DD5:63877BDC 192.168.45.213 13:24:59.892674 R data 12:805453824 16384 4,997 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.897671 R data 20:1491763200 16384 4,992 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.898670 R data 12:1720532992 16384 24,479 cli C0A82DD5:63877BD4 192.168.45.214 13:24:59.898670 R data 20:2406842368 16384 26,476 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.925146 R data 20:1491763200 16384 8,990 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.935430 R data 20:1491763200 16384 4,691 cli C0A82DD5:63877BDD 192.168.45.214 13:24:59.940121 R data 28:1034223616 16384 4,312 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.940121 R data 28:1949302784 16384 26,664 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.940121 R data 5:1949302784 16384 34,831 cli C0A82DD5:63877BCD 192.168.45.213 13:24:59.975955 R data 28:1034223616 16384 5,583 cli C0A82DD5:63877BE5 192.168.45.214 13:24:59.981539 R data 5:1034223616 16384 5,168 cli C0A82DD5:63877BCD 192.168.45.213 13:24:59.981539 R data 21:1949302784 16384 25,392 cli C0A82DD5:63877BDE 192.168.45.213 13:24:59.981539 R data 13:1949302784 16384 34,412 cli C0A82DD5:63877BD5 192.168.45.213 13:25:00.016950 R data 5:1034223616 16384 7,991 cli C0A82DD5:63877BCD 192.168.45.213 13:25:00.026938 R data 5:1034223616 16384 7,750 cli C0A82DD5:63877BCD 192.168.45.213 13:25:00.034688 R data 13:1034223616 16384 5,498 cli C0A82DD5:63877BD5 192.168.45.213 13:25:00.034688 R data 6:1949302784 16384 27,661 cli C0A82DD5:63877BCE 192.168.45.214 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Wednesday, 4 September 2024 13:24 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hello My theory is that it's windows 11 related, or the combo windows 11 and this hardware. I guess the only thing to know for sure, is to boot it into *nix and install gpfs there and test. Which I guess isn't the worst of ideas at this stage. --iohist spits out a pretty hefty chuck of data. Below is a snippet from when I did a write test: mmdiag --iohist === mmdiag: iohist === I/O history: I/O start time RW Buf type disk:sectorNum nSec time ms Type Device/NSD ID NSD node --------------- -- ----------- ----------------- ----- ------- ---- ------------------ --------------- 13:17:04.088621 W data 10:2178073088 10800 3,994 cli C0A82DD5:63877BD2 192.168.45.214 13:17:04.092614 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.092614 W inode 2:3040332814 1 0,999 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.093613 W data 10:2178083888 5072 2,995 cli C0A82DD5:63877BD2 192.168.45.214 13:17:04.094611 W data 18:2178072576 16384 6,502 cli C0A82DD5:63877BDB 192.168.45.214 13:17:04.101113 W logData 2:1869321537 2 0,998 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.101113 W logData 3:1877542209 2 0,998 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.103109 W data 18:2178078304 10656 3,994 cli C0A82DD5:63877BDB 192.168.45.214 13:17:04.103109 W data 26:2178072576 16384 6,529 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.109638 W logData 2:1869321538 2 0,994 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.109638 W logData 3:1877542210 2 0,994 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.111634 W data 26:2178072720 10800 3,992 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.115626 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.115626 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.116626 W data 26:2178083520 5440 2,995 cli C0A82DD5:63877BE3 192.168.45.214 13:17:04.117629 W data 11:2178072576 16384 4,987 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.122616 W logData 2:1869321539 2 0,999 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.122616 W logData 3:1877542211 2 0,999 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.124613 W data 11:2178077936 10800 3,999 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.128612 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.128612 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.129609 W data 11:2178088736 224 0,000 cli C0A82DD5:63877BD3 192.168.45.213 13:17:04.129609 W data 19:2178072576 16384 6,495 cli C0A82DD5:63877BDC 192.168.45.213 13:17:04.136104 W logData 2:1869321540 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.136104 W logData 3:1877542212 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.137103 W data 19:2178083152 5808 2,999 cli C0A82DD5:63877BDC 192.168.45.213 13:17:04.138100 W data 27:2178072576 16384 5,990 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.144091 W logData 2:1869321540 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.144091 W logData 3:1877542212 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.146088 W data 27:2178077568 10800 4,005 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.150092 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.150092 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.151092 W data 27:2178088368 592 0,000 cli C0A82DD5:63877BE4 192.168.45.213 13:17:04.151092 W data 12:2178072576 16384 4,995 cli C0A82DD5:63877BD4 192.168.45.214 13:17:04.156086 W logData 2:1869321541 2 0,996 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.156086 W logData 3:1877542213 2 0,996 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.157082 W data 12:2178082784 6176 2,994 cli C0A82DD5:63877BD4 192.168.45.214 13:17:04.158083 W data 20:2178072576 16384 7,498 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.165581 W logData 2:1869321542 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.165581 W logData 3:1877542214 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.167578 W data 20:2178077200 10800 2,994 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.170572 W inode 4:3060742158 1 0,996 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.171568 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.171568 W data 20:2178088000 960 1,001 cli C0A82DD5:63877BDD 192.168.45.214 13:17:04.172569 W data 28:2864381952 16384 5,988 cli C0A82DD5:63877BE5 192.168.45.214 13:17:04.178557 W logData 2:1869321543 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.178557 W logData 3:1877542215 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.179560 W data 28:2864391792 6544 2,995 cli C0A82DD5:63877BE5 192.168.45.214 13:17:04.179560 W data 5:2406842368 16384 5,987 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.185548 W logData 2:1869321544 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.185548 W logData 3:1877542216 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.186549 W data 5:2406846624 10800 4,993 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.191542 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.191542 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.192540 W data 5:2406857424 1328 0,995 cli C0A82DD5:63877BCD 192.168.45.213 13:17:04.193535 W data 13:2406842368 16384 6,019 cli C0A82DD5:63877BD5 192.168.45.213 13:17:04.199554 W logData 2:1869321544 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.199554 W logData 3:1877542216 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.200551 W data 13:2406851840 6912 2,997 cli C0A82DD5:63877BD5 192.168.45.213 13:17:04.201554 W data 21:2406842368 16384 5,912 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.207466 W logData 2:1869321545 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.207466 W logData 3:1877542217 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.208465 W data 21:2406846256 10800 3,990 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.212456 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.212456 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.213456 W data 21:2406857056 1696 1,998 cli C0A82DD5:63877BDE 192.168.45.213 13:17:04.214457 W data 6:2406842368 16384 5,015 cli C0A82DD5:63877BCE 192.168.45.214 13:17:04.219472 W logData 2:1869321546 2 1,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.219472 W logData 3:1877542218 2 1,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.221474 W data 6:2406851472 7280 1,994 cli C0A82DD5:63877BCE 192.168.45.214 13:17:04.221474 W data 14:2406842368 16384 7,502 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.228976 W logData 2:1869321547 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.228976 W logData 3:1877542219 1 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.229971 W data 14:2406845888 10800 3,994 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.233965 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 13:17:04.233965 W inode 2:3040332814 1 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.234963 W data 14:2406856688 2064 1,999 cli C0A82DD5:63877BD6 192.168.45.214 13:17:04.235965 W data 22:2406842368 16384 4,993 cli C0A82DD5:63877BDF 192.168.45.214 13:17:04.240958 W logData 2:1869321547 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.240958 W logData 3:1877542219 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.241974 W data 22:2406851104 7648 2,531 cli C0A82DD5:63877BDF 192.168.45.214 13:17:04.242977 W data 7:2406842368 16384 5,526 cli C0A82DD5:63877BCF 192.168.45.213 13:17:04.248503 W logData 2:1869321548 2 0,000 cli C0A82DD5:63877BCA 192.168.45.214 13:17:04.248503 W logData 3:1877542220 2 0,000 cli C0A82DD5:63877BCB 192.168.45.213 13:17:04.249503 W data 7:2406845520 10800 3,994 cli C0A82DD5:63877BCF 192.168.45.213 13:17:04.253497 W inode 4:3060742158 1 0,000 cli C0A82DD5:63877BCC 192.168.45.214 -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Wednesday, 4 September 2024 12:59 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, given you see read latencies of 1 ms, you do not get the data from disk but from some cache (on whatever level). From spinning disks, you can never expect such read latencies (mind that GPFS block reading, even if sequential from the application's PoV, typically translates to random I/O at the physical disk level). So, I do not know what the latencies on your other measurements (bypassing GPFS) were but the numbers below do not represent sustained high-scale throughputs, apparently. It is nevertheless strange that your write rates are much below reads (and write latencies are that high) -- from my experience with different systems, when hammering GPFS with usual storage backends with both read and write requests, the writes tend to prevail. Your waiters indicate that the problem is above GPFS: GPFS is able to serve all I/O threads within a few ms, and there is not a long list of pending IOs. Sorry, the iohistory options is mmdiag --iohist But to me it looks like not GPFS is the culprit here. Uwe On 04.09.24 11:08, Henrik Cednert wrote: Hi Uwe Thanks. Worth noting is that we have win10 ltsc, win 2019 and have had a single CPU win 11 22h2 (as a test) clients that all perform as expected. Those machines is older though and connected with 10-40GbE, those client max out their NIC in read and write. Let me know if i missed something important here. Thanks again. Setups is Client: * Supermicro Workstation * Intel(R) Xeon(R) Gold 6418H 2.10 GHz (2 processors) * Mellanox ConnectX-6 Dx connected with 100GbE over dedicated vlan via mellanox sn2100. * Windows 11 Pro for Workstations, 22H2 Storage setup * 3 x 84 bay seagate chassis with spinning disks. * Storage connected with redundant 12Gb SAS to 2 x Storage node servers * 2 x mellanox sn2100 * The 2 storage node servers are for this vlan connected with 100GbE to each switch, so in total 4 x 100GbE. And the switches are connected with 2 * 100GbE I tested the commands you suggested. They are both new to me so not sure what the output is supposed to be, looks like -iohistory isn't available in windows. I ran --waiters a few times, as seen below. Not sure what the expected output is from that. mmdiag --waiters === mmdiag: waiters === Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 18616 MsgHandler at getData: for In function sendMessage Waiting 0.0000 sec since 2024-09-04_10:05:14, monitored, thread 25084 WritebehindWorkerThread: on ThCond 0x31A7C360 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 16780 FsyncHandlerThread: on ThCond 0x37FFDAB0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:17, monitored, thread 30308 MsgHandler at getData: for In function sendMessage C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0055 sec since 2024-09-04_10:05:21, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x37A25FF0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0029 sec since 2024-09-04_10:05:23, monitored, thread 16780 FileBlockReadFetchHandlerThread: on ThCond 0x38281DE0 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.213 C:\Users\m5-tkd01>mmdiag --waiters === mmdiag: waiters === Waiting 0.0019 sec since 2024-09-04_10:05:25, monitored, thread 11832 PrefetchWorkerThread: on ThCond 0x38278D20 (MsgRecordCondvar), reason 'RPC wait' for NSD I/O completion on node 192.168.45.214 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 16780 AcquireBRTHandlerThread: on ThCond 0x37A324E0 (MsgRecordCondvar), reason 'RPC wait' for tmMsgBRRevoke on node 192.168.45.161 Waiting 0.0009 sec since 2024-09-04_10:05:25, monitored, thread 2576 RangeRevokeWorkerThread: on ThCond 0x5419DAA0 (BrlObjCondvar), reason 'waiting because of local byte range lock conflict' C:\Users\m5-tkd01> C:\Users\m5-tkd01>mmdiag --iohistory Unrecognized option: --iohistory. Run mmdiag --help for the option list -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Uwe Falke Sent: Tuesday, 3 September 2024 17:35 To: gpfsug-discuss at gpfsug.org Subject: Re: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. Hi, Henrik, while I am not using Windows I'd start investigating the usual things (see below). But first you should describe your set-up better. Where are the NSDs : locally attached to the Windows box? In some NSD servers? If the latter -- what is the link to the NSD servers? via your GbE link? FC? IB? separate Ethernet? What type of storage? Spinning Disks? Flash? How long are your I/Os waiting on the client (compare that to the waiting times on the NSD server if applicable)? not sure whether that is available on Windows, but mmdiag --waiters mmdiag --iohistory might be of use. Somewhere in the chain from your application to the storage backend there is a delay and you should first find out where that occurs I think. Bye Uwe On 03.09.24 14:10, Henrik Cednert wrote: Still no solution here regarding this. Have tested other cables. Have tested to change tcp window size, no change Played with numa in the bios, no change Played with hyperthreading in bios, no change Have anyone managed to get some speed out of windows 11 and gpfs? -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. ________________________________ From: gpfsug-discuss on behalf of Henrik Cednert Sent: Friday, 9 August 2024 17:25 To: gpfsug-discuss at gpfsug.org Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. VARNING: DETTA ?R ETT EXTERNT MAIL. Klicka inte p? n?gra l?nkar oavsett hur legitima de verkar utan att verifiera. Hello I have some issues with write performance on a windows 11 pro system and I'm out of ideas here. Hopefully someone here have some bright ideas and/or experience of GPFS on Windows 11? The system is a: Windows 11 Pro 22H2 2 x Intel(R) Xeon(R) Gold 6418H 2.10 GHz 512 GB RAM GPFS 5.1.9.4 Mellanox ConnectX 6 Dx 100GbE connected to Mellanox Switch with 5m Mellanox DAC. Before deploying this workstation we had a single socket system as a test bench where we got 60 GbE in both directons with iPerf and around 6GB/sec write and 3GB/sec read from the system over GPFS (fio tests, same tests as furhter down here). With that system I had loads of issues before getting to that point though. MS Defender had to be forcefully disabled via regedit some other tweaks. All those tweaks have been performed in this new system as well, but I can't get the proper speed out of it. On this new system and with iPerf to the storage servers I get around 50-60GbE in both directions and send and receive. If I mount the storage over SMB and 100GbE via the storage gateway servers I get around 3GB/sec read and write with Blackmagics Disk speed test. I have not tweaked the system for samba performande, just a test to see what it would give and part of the troubleshooting. If I run Blackmagics diskspeed test to the GPFS mount I instead get around 700MB/sec write and 400MB/sec read. Starting to think that the Blackmagic test might not run properly on this machine with these CPUs though. Or it's related to the mmfsd process maybe, how that threads or not threads...? But if we instead look at fio. I have a bat script that loops through a bunch of FIO-tests. A test that I have been using over the years so that we easily can benchmark all deployed systems with the exakt same tests. The tests are named like: seqrw-gb-mb-t The result when I run this is like the below list. Number in parenthesis is the by fio reported latency. Job: seqrw-40gb-1mb-t1 ????????????Write: 162 MB/s (6 ms) ????????????Read: 1940 MB/s (1 ms) Job: seqrw-20gb-1mb-t2 ????????????Write: 286 MB/s (7 ms) ????????????Read: 3952 MB/s (1 ms) Job: seqrw-10gb-1mb-t4 ????????????Write: 549 MB/s (7 ms) ????????????Read: 6987 MB/s (1 ms) Job: seqrw-05gb-1mb-t8 ????????????Write: 989 MB/s (8 ms) ????????????Read: 7721 MB/s (1 ms) Job: seqrw-40gb-2mb-t1 ????????????Write: 161 MB/s (12 ms) ????????????Read: 2261 MB/s (0 ms) Job: seqrw-20gb-2mb-t2 ????????????Write: 348 MB/s (11 ms) ????????????Read: 4266 MB/s (1 ms) Job: seqrw-10gb-2mb-t4 ????????????Write: 626 MB/s (13 ms) ????????????Read: 4949 MB/s (1 ms) Job: seqrw-05gb-2mb-t8 ????????????Write: 1154 MB/s (14 ms) ????????????Read: 7007 MB/s (2 ms) Job: seqrw-40gb-4mb-t1 ????????????Write: 161 MB/s (25 ms) ????????????Read: 2083 MB/s (1 ms) Job: seqrw-20gb-4mb-t2 ????????????Write: 352 MB/s (23 ms) ????????????Read: 4317 MB/s (2 ms) Job: seqrw-10gb-4mb-t4 ????????????Write: 696 MB/s (23 ms) ????????????Read: 7358 MB/s (2 ms) Job: seqrw-05gb-4mb-t8 ????????????Write: 1251 MB/s (25 ms) ????????????Read: 6707 MB/s (5 ms) So with fio I get a very nice read speed, but the write is horrendous and I cannot find what causes it. I have looked at affinity settings for the mmfsd process but not sure I fully understand it. But no matter what I set it to, I see no difference. I have "played" with the bios and tried with/without hyperthreading, numa and so on. And nothing affects atleast the blackmagic disk speed test. the current settings for this host is like below. I write "current" because I have tested a few different settings here but nothing affects the write speed. maxTcpConnsPerNodeConn for sure bumped the read speed though. nsdMaxWorkerThreads 16 prefetchPct 60 maxTcpConnsPerNodeConn 8 maxMBpS 14000 Does anyone have any suggestions or ideas on how to troubleshoot this? Thanks -- Henrik Cednert / + 46 704 71 89 54 / CTO / OnePost (formerly Filmlance Post) ?? OnePost, formerly Filmlance's post-production, is now an independent part of the Banijay Group. New name, same team ? business as usual at OnePost. _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email: uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: From uwe.falke at kit.edu Wed Sep 4 15:46:32 2024 From: uwe.falke at kit.edu (Uwe Falke) Date: Wed, 4 Sep 2024 16:46:32 +0200 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> <33582c29-da8f-4cc1-a8c2-3debb8373a71@kit.edu> Message-ID: <0bc9b196-c7a4-4b43-afda-fe50c6f47a6a@kit.edu> On 04.09.24 15:59, Henrik Cednert wrote: > Thanks Uwe > > I will have to digest what you write for a while, at this time of day > some of it flies over my head (...and some probably will no matter the > time of day). =) > > But to add some info about the file system. > > File system attributes for /dev/mmfs1: > ====================================== > flag ? ? ? ? ? ? ? ?value ? ? ? ? ? ? ? ? ? ?description > ------------------- ------------------------ > ----------------------------------- > ?-f ? ? ? ? ? ? ? ? 8192 ? ? ? ? ? ? ? ? ? ? Minimum fragment > (subblock) size in bytes (system pool) > ? ? ? ? ? ? ? ? ? ? 131072 ? ? ? ? ? ? ? ? ? Minimum fragment > (subblock) size in bytes (other pools) > ?-i ? ? ? ? ? ? ? ? 512 ? ? ? ? ? ? ? ? ? ? ?Inode size in bytes > ?-I ? ? ? ? ? ? ? ? 32768 ? ? ? ? ? ? ? ? ? ?Indirect block size in bytes > ?-m ? ? ? ? ? ? ? ? 2 ? ? ? ? ? ? ? ? ? ? ? ?Default number of > metadata replicas > ?-M ? ? ? ? ? ? ? ? 2 ? ? ? ? ? ? ? ? ? ? ? ?Maximum number of > metadata replicas > ?-r ? ? ? ? ? ? ? ? 1 ? ? ? ? ? ? ? ? ? ? ? ?Default number of data > replicas > ?-R ? ? ? ? ? ? ? ? 2 ? ? ? ? ? ? ? ? ? ? ? ?Maximum number of data > replicas > ?-j ? ? ? ? ? ? ? ? scatter ? ? ? ? ? ? ? ? ?Block allocation type > ?-D ? ? ? ? ? ? ? ? nfs4 ? ? ? ? ? ? ? ? ? ? File locking semantics in > effect > ?-k ? ? ? ? ? ? ? ? nfs4 ? ? ? ? ? ? ? ? ? ? ACL semantics in effect > ?-n ? ? ? ? ? ? ? ? 32 ? ? ? ? ? ? ? ? ? ? ? Estimated number of nodes > that will mount file system > ?-B ? ? ? ? ? ? ? ? 524288 ? ? ? ? ? ? ? ? ? Block size (system pool) > ? ? ? ? ? ? ? ? ? ? 8388608 ? ? ? ? ? ? ? ? ?Block size (other pools) > > > Regarding, > >The iohist snippet for reads comprises 74 IOs in about 0.854s, this > relates to roughly 690MiB/s, far from the 10-fold value you reported. > > I'm not sure about that 10-fold value you refer to here. 690MiB is > pretty much exactly what I saw reported in the disk speed test I was > running when extracting that data. I referred to numbers like in > Job: seqrw-10gb-1mb-t4 > ????????????Write: 549 MB/s (7 ms) > ????????????Read: 6987 MB/s (1 ms) was I mistaken ? > > > I will re-run my fio-tests on the other systems so that I have fresh > values. If I'm sure that they are trustworthy...? Can one ever be? > Network graphs and reported fio results are all I have to lean > against. Attaching a few lines of --iohist for one of those 10GbE > clients that currently is running my batch fio test. > > <--------------------------------------snip------------------------------------> There is always bit of uncertainty but one could try to minimize that and also look for signs of it. Caching effects are prone to be a vivid source of errors in benchmarks. I am not sure we have it here but thought there are serious indicators.? but i might have mixed up your tests. also I see you do not run concurring reads and writes , but just read or write tests at a time. but it remains from your waiters list: GPFS doesn't seem to be (held) busy (by the test app). If there are 74 IO requests in 0.85secs each one taking less than 6ms that accounts for less than just 0.444s -- about half of the time GPFS is idling. with full IO queues you should see at least about 1.3GiB/s (which is still not 3, but would be better than what you had now). But the next thing to explore is: while many read IOs take about 5..6ms (mind: that is supsiciously low!) what is causing the others to take 20..40ms (or even more). You could also check the iohistory on the NSD servers where you'd see the times it takes the IO requests issued by the NSD server against the storage backend. But you should not post the output. You may send the output as an attachment (but maybe not to the entire mailing list :-). Can you start your tests with multiple threads (or maybe multiple tests in parallel)? Uwe -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email:uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 5814 bytes Desc: S/MIME Cryptographic Signature URL: From uwe.falke at kit.edu Wed Sep 4 16:04:51 2024 From: uwe.falke at kit.edu (Uwe Falke) Date: Wed, 4 Sep 2024 17:04:51 +0200 Subject: [gpfsug-discuss] GPFS 5.1.9.4 on Windows 11 Pro. Performance issues, write. In-Reply-To: References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> <33582c29-da8f-4cc1-a8c2-3debb8373a71@kit.edu> Message-ID: On 04.09.24 15:59, Henrik Cednert wrote: > Thanks Uwe Just saw: the app seems to have issued IO requests not sequentially but in bunches. what is apparent (but I have not recognized it before): after such a bunch of IOs, the next bunch is typically issued when the longest-taking IO from the prious bunch was completed. For example: 13:24:59.341629 ?R ? ? ? ?data ? 17:805453824 ? ?16384 ? ?3,993 cli C0A82DD5:63877BDA ? 192.168.45.213 13:24:59.341629 ?R ? ? ? ?data ? ?6:1720532992 ? 16384 ? 27,471 cli C0A82DD5:63877BCE ? 192.168.45.214 13:24:59.341629 ?R ? ? ? ?data ? 14:1720532992 ? 16384 ? 44,953 cli C0A82DD5:63877BD6 ? 192.168.45.214 3 read IOs issued at 13:24:59.341629, longest taking 44.953ms. 59.341629+0.044953=59.386582, which is 1ms before the next IO req is issued at 13:24:59.387584 13:24:59.387584 ?R ? ? ? ?data ? 17:805453824 ? ?16384 ? ?5,990 ? cli C0A82DD5:63877BDA ? 192.168.45.213 that one takes just 5.990ms ,? the next bunch of IO reqs is seen 6.988 ms later (i.e. again the service time plus about 1 ms): 13:24:59.394572 ?R ? ? ? ?data ? 17:805453824 ? ?16384 ? ?7,993 cli C0A82DD5:63877BDA ? 192.168.45.213 now a triplet of IOs again, but one of them taking very long: 13:24:59.402565 ?R ? ? ? ?data ? 25:805453824 ? ?16384 ? ?4,991 cli C0A82DD5:63877BE2 ? 192.168.45.213 13:24:59.402565 ?R ? ? ? ?data ? 22:1720532992 ? 16384 ? 26,309 cli C0A82DD5:63877BDF ? 192.168.45.214 13:24:59.402565 ?R ? ? ? ?data ? ?7:1720532992 ? 16384 ?142,856 cli C0A82DD5:63877BCF ? 192.168.45.213 That incurs the expected delay of 143.854ms : 13:24:59.546419 ?R ? ? ? ?data ? 25:805453824 ? ?16384 ? ?7,992 ? cli C0A82DD5:63877BE2 ? 192.168.45.213 etc. pp. 13:24:59.556408 ?R ? ? ? ?data ? 25:805453824 ? ?16384 ? ?7,987 cli C0A82DD5:63877BE2 ? 192.168.45.213 13:24:59.564395 ?R ? ? ? ?data ? 10:805453824 ? ?16384 ? ?5,505 cli C0A82DD5:63877BD2 ? 192.168.45.214 13:24:59.564395 ?R ? ? ? ?data ? 23:1720532992 ? 16384 ? 28,053 cli C0A82DD5:63877BE0 ? 192.168.45.213 13:24:59.564395 ?R ? ? ? ?data ? 15:1720532992 ? 16384 ? 33,044 cli C0A82DD5:63877BD8 ? 192.168.45.213 13:24:59.598437 ?R ? ? ? ?data ? 10:805453824 ? ?16384 ? ?5,504 cli C0A82DD5:63877BD2 ? 192.168.45.214 13:24:59.604939 ?R ? ? ? ?data ? 18:805453824 ? ?16384 ? ?4,993 cli C0A82DD5:63877BDB ? 192.168.45.214 13:24:59.604939 ?R ? ? ? ?data ? ?8:1720532992 ? 16384 ? 36,015 cli C0A82DD5:63877BD0 ? 192.168.45.214 ... so there is some parallelization in IO, but using multiple threads should improve things as the app apparently still waits for IOs being performed. and of course with an otherwise unloaded storage and just one client, service times >25ms should not be seen and most should be 10...20ms. Uwe -- Karlsruhe Institute of Technology (KIT) Scientific Computing Centre (SCC) Scientific Data Management (SDM) Uwe Falke Hermann-von-Helmholtz-Platz 1, Building 442, Room 187 D-76344 Eggenstein-Leopoldshafen Tel: +49 721 608 28024 Email:uwe.falke at kit.edu www.scc.kit.edu Registered office: Kaiserstra?e 12, 76131 Karlsruhe, Germany KIT ? The Research University in the Helmholtz Association -------------- next part -------------- An HTML attachment was scrubbed... URL: -------------- next part -------------- A non-text attachment was scrubbed... Name: smime.p7s Type: application/pkcs7-signature Size: 5814 bytes Desc: S/MIME Cryptographic Signature URL: From scottg at emailhosting.com Thu Sep 5 17:31:52 2024 From: scottg at emailhosting.com (scottg at emailhosting.com) Date: Thu, 5 Sep 2024 12:31:52 -0400 Subject: [gpfsug-discuss] Possible dates for User Group Meetings - NYC? and MIami In-Reply-To: <0bc9b196-c7a4-4b43-afda-fe50c6f47a6a@kit.edu> References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> <33582c29-da8f-4cc1-a8c2-3debb8373a71@kit.edu> <0bc9b196-c7a4-4b43-afda-fe50c6f47a6a@kit.edu> Message-ID: <048301daffb1$1da3a890$58eaf9b0$@emailhosting.com> Hi All, Did I miss the notice for a Storage Scale Users group meeting in NYC? I seem to remember there is usually one at the end of September. And will the Miami meeting be rescheduled? https://www.spectrumscaleug.org/registration-open-for-2024-ibm-storage-scale-user-group-miami-meeting/ Thanks, Scott G. -------------- next part -------------- An HTML attachment was scrubbed... URL: From TROPPENS at de.ibm.com Fri Sep 6 13:37:33 2024 From: TROPPENS at de.ibm.com (Ulf Troppens) Date: Fri, 6 Sep 2024 12:37:33 +0000 Subject: [gpfsug-discuss] Possible dates for User Group Meetings - NYC? and MIami In-Reply-To: <048301daffb1$1da3a890$58eaf9b0$@emailhosting.com> References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> <33582c29-da8f-4cc1-a8c2-3debb8373a71@kit.edu> <0bc9b196-c7a4-4b43-afda-fe50c6f47a6a@kit.edu> <048301daffb1$1da3a890$58eaf9b0$@emailhosting.com> Message-ID: Hi Scott, the NYC meeting is planned for October first, and for SC24 we plan a meeting for November 17, the Sunday before SC. More details will follow soon. Best, Ulf Ulf Troppens Product Manager - IBM Storage for Data and AI, Data-Intensive Workflows IBM Deutschland Research & Development GmbH Vorsitzender des Aufsichtsrats: Wolfgang Wendt / Gesch?ftsf?hrung: David Faller Sitz der Gesellschaft: B?blingen / Registergericht: Amtsgericht Stuttgart, HRB 243294 ________________________________ From: gpfsug-discuss on behalf of scottg at emailhosting.com Sent: Thursday, September 5, 2024 6:31 PM To: 'gpfsug main discussion list' Subject: [EXTERNAL] [gpfsug-discuss] Possible dates for User Group Meetings - NYC? and MIami Hi All, Did I miss the notice for a Storage Scale Users group meeting in NYC? I seem to remember there is usually one at the end of September. And will the Miami meeting be rescheduled? https:?//www.?spectrumscaleug.?org/registration-open-for-2024-ibm-storage-scale-user-group-miami-meeting/ Hi All, Did I miss the notice for a Storage Scale Users group meeting in NYC? I seem to remember there is usually one at the end of September. And will the Miami meeting be rescheduled? https://www.spectrumscaleug.org/registration-open-for-2024-ibm-storage-scale-user-group-miami-meeting/ Thanks, Scott G. -------------- next part -------------- An HTML attachment was scrubbed... URL: From Luis.Bolinches at worldquant.com Fri Sep 6 14:28:10 2024 From: Luis.Bolinches at worldquant.com (Bolinches, Luis (WorldQuant)) Date: Fri, 6 Sep 2024 13:28:10 +0000 Subject: [gpfsug-discuss] [EXTERNAL] Re: Possible dates for User Group Meetings - NYC? and MIami In-Reply-To: References: <9a895ef6-4d43-48bf-bf17-88b3941a4191@kit.edu> <33582c29-da8f-4cc1-a8c2-3debb8373a71@kit.edu> <0bc9b196-c7a4-4b43-afda-fe50c6f47a6a@kit.edu> <048301daffb1$1da3a890$58eaf9b0$@emailhosting.com> Message-ID: Hi Can you share the links for those? Thanks -- Yst?v?llisin terveisin/Regards/Saludos/Salutations/Salutacions Luis Bolinches WQ Aligned Infrastructure "If you always give you will always have" -- Anonymous https://www.credly.com/users/luis-bolinches/badges From: gpfsug-discuss On Behalf Of Ulf Troppens Sent: Friday, 6 September 2024 15.38 To: 'gpfsug main discussion list' Subject: [EXTERNAL] Re: [gpfsug-discuss] Possible dates for User Group Meetings - NYC? and MIami Hi Scott, the NYC meeting is planned for October first, and for SC24 we plan a meeting for November 17, the Sunday before SC. More details will follow soon. Best, Ulf Ulf Troppens Product Manager - IBM Storage for Data and AI, Data-Intensive Workflows IBM Deutschland Research & Development GmbH Vorsitzender des Aufsichtsrats: Wolfgang Wendt / Gesch?ftsf?hrung: David Faller Sitz der Gesellschaft: B?blingen / Registergericht: Amtsgericht Stuttgart, HRB 243294 ________________________________ From: gpfsug-discuss > on behalf of scottg at emailhosting.com > Sent: Thursday, September 5, 2024 6:31 PM To: 'gpfsug main discussion list' > Subject: [EXTERNAL] [gpfsug-discuss] Possible dates for User Group Meetings - NYC? and MIami Hi All, Did I miss the notice for a Storage Scale Users group meeting in NYC? I seem to remember there is usually one at the end of September. And will the Miami meeting be rescheduled? https:?//www.?spectrumscaleug.?org/registration-open-for-2024-ibm-storage-scale-user-group-miami-meeting/ Hi All, Did I miss the notice for a Storage Scale Users group meeting in NYC? I seem to remember there is usually one at the end of September. And will the Miami meeting be rescheduled? https://www.spectrumscaleug.org/registration-open-for-2024-ibm-storage-scale-user-group-miami-meeting/ Thanks, Scott G. ################################################################################### The information contained in this communication is confidential, may be subject to legal privilege, and is intended only for the individual named. If you are not the named addressee, please notify the sender immediately and delete this email from your system. The views expressed in this email are the views of the sender only. Outgoing and incoming electronic communications to this address are electronically archived and subject to review and/or disclosure to someone other than the recipient. ################################################################################### -------------- next part -------------- An HTML attachment was scrubbed... URL: From smcgrat at tcd.ie Fri Sep 6 16:52:55 2024 From: smcgrat at tcd.ie (Sean Mc Grath) Date: Fri, 6 Sep 2024 15:52:55 +0000 Subject: [gpfsug-discuss] InfiniBand and OmiPath NSD servers Message-ID: Hi, We have a GPFS cluster where the NSD servers mount the storage over fibre channel and export the file system over InfiniBand for clients. We will be getting some used equipment that uses OmniPath. As per the "IBM Storage Scale Frequently Asked Questions and Answers" it states [1]: > RDMA is not supported on a node when both Mellanox HCAs and Cornelis Networks Omni-Path HFIs are enabled for RDMA. Does this mean that we wouldn't be able to consolidate both IB & OPA HCA's in NSD servers and would have to have 2 types of NSD servers? 1) InfiniBand exporting and 2) OmniPath exporting? If so, is it then a matter of using the Multi-Rail over TCP "subnets =" setting in mmchonfig to distinguish which nsd server the clients should connect to? [2]. Or am I completely miss understanding all this? Many thanks in advance. Sean [1] https://www.ibm.com/docs/en/STXKQY/pdf/gpfsclustersfaq.pdf [2] https://www.ibm.com/docs/en/storage-scale/5.1.6?topic=configuring-multi-rail-over-tcp-mrot --- Sean McGrath smcgrat at tcd.ie Senior Systems Administrator Research IT, IT Services, Trinity College Dublin https://www.tcd.ie/itservices/ https://www.tchpc.tcd.ie/ From cdmaestas at us.ibm.com Mon Sep 9 18:44:35 2024 From: cdmaestas at us.ibm.com (CHRIS MAESTAS) Date: Mon, 9 Sep 2024 17:44:35 +0000 Subject: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Message-ID: Hello everyone, There will be a Scale User Group event in New York City on October 1st, 2024. We have an exciting agenda covering user stories, roadmap update, the latest insights from the user community, development, research, NVIDIA, SuperMicro and DST, plus access to IBM experts and your peers. We look forward to welcoming you to this event. Register here - http://ibm.biz/ScaleNY ps It may say it?s online, but it?s in person! * https://storagescale.org/nyc-2024-scale-day/ * https://storagescale.org/event/nyc2024-ibm-storage-scale-users-group/ -- Chief Troublemaking Officer Chris -------------- next part -------------- An HTML attachment was scrubbed... URL: From cdmaestas at us.ibm.com Tue Sep 10 15:16:53 2024 From: cdmaestas at us.ibm.com (CHRIS MAESTAS) Date: Tue, 10 Sep 2024 14:16:53 +0000 Subject: [gpfsug-discuss] Save the date - Scale User Group @ SC2024! Message-ID: There will be a Scale User Group event in Atlanta, Georgia, USA on the afternoon of November 17th, 2024. We will have an exciting agenda and access to experts and your peers. We look forward to welcoming you to this event. The agenda will cover a variety of topics including: * Updates from the user community * IBM Storage Scale Development updates ? strategy, roadmap, performance updates * Updates from NVIDIA * Partner updates It will be held at: The Westin Peachtree Plaza, Atlanta 210 Peachtree St. NW, Atlanta, GA 30303 Registration link is coming! -cdm -------------- next part -------------- An HTML attachment was scrubbed... URL: From novosirj at rutgers.edu Tue Sep 10 16:27:31 2024 From: novosirj at rutgers.edu (Ryan Novosielski) Date: Tue, 10 Sep 2024 15:27:31 +0000 Subject: [gpfsug-discuss] Scale Day in NYC October 1, 2024 In-Reply-To: References: Message-ID: Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' On Sep 9, 2024, at 13:44, CHRIS MAESTAS wrote: Hello everyone, There will be a Scale User Group event in New York City on October 1st, 2024. We have an exciting agenda covering user stories, roadmap update, the latest insights from the user community, development, research, NVIDIA, SuperMicro and DST, plus access to IBM experts and your peers. We look forward to welcoming you to this event. Register here - http://ibm.biz/ScaleNY ps It may say it?s online, but it?s in person! * https://storagescale.org/nyc-2024-scale-day/ * https://storagescale.org/event/nyc2024-ibm-storage-scale-users-group/ -- Chief Troublemaking Officer Chris _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: From cdmaestas at us.ibm.com Tue Sep 10 18:21:36 2024 From: cdmaestas at us.ibm.com (CHRIS MAESTAS) Date: Tue, 10 Sep 2024 17:21:36 +0000 Subject: [gpfsug-discuss] Scale Day in NYC October 1, 2024 In-Reply-To: References: Message-ID: Hi Ryan, We have been talking about re-starting with the digital series of talks that are online. Stay tuned and let me see if we can kick this off starting with the latest what?s new talk in Scale 5.2.1! The previous series is here: https://storagescale.org/experttalks/ -cdm From: gpfsug-discuss on behalf of Ryan Novosielski Date: Tuesday, September 10, 2024 at 9:29?AM To: gpfsug main discussion list Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' On Sep 9, 2024, at 13:44, CHRIS MAESTAS wrote: Hello everyone, There will be a Scale User Group event in New York City on October 1st, 2024. We have an exciting agenda covering user stories, roadmap update, the latest insights from the user community, development, research, NVIDIA, SuperMicro and DST, plus access to IBM experts and your peers. We look forward to welcoming you to this event. Register here - http://ibm.biz/ScaleNY ps It may say it?s online, but it?s in person! * https://storagescale.org/nyc-2024-scale-day/ * https://storagescale.org/event/nyc2024-ibm-storage-scale-users-group/ -- Chief Troublemaking Officer Chris _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: From novosirj at rutgers.edu Wed Sep 11 00:12:19 2024 From: novosirj at rutgers.edu (Ryan Novosielski) Date: Tue, 10 Sep 2024 23:12:19 +0000 Subject: [gpfsug-discuss] Scale Day in NYC October 1, 2024 In-Reply-To: References: Message-ID: Thanks, great news! We?ll have a few people who will want to come to that. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' On Sep 10, 2024, at 13:21, CHRIS MAESTAS wrote: Hi Ryan, We have been talking about re-starting with the digital series of talks that are online. Stay tuned and let me see if we can kick this off starting with the latest what?s new talk in Scale 5.2.1! The previous series is here: https://storagescale.org/experttalks/ -cdm From: gpfsug-discuss > on behalf of Ryan Novosielski > Date: Tuesday, September 10, 2024 at 9:29?AM To: gpfsug main discussion list > Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' On Sep 9, 2024, at 13:44, CHRIS MAESTAS wrote: Hello everyone, There will be a Scale User Group event in New York City on October 1st, 2024. We have an exciting agenda covering user stories, roadmap update, the latest insights from the user community, development, research, NVIDIA, SuperMicro and DST, plus access to IBM experts and your peers. We look forward to welcoming you to this event. Register here - http://ibm.biz/ScaleNY ps It may say it?s online, but it?s in person! * https://storagescale.org/nyc-2024-scale-day/ * https://storagescale.org/event/nyc2024-ibm-storage-scale-users-group/ -- Chief Troublemaking Officer Chris _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: From sadaniel at us.ibm.com Thu Sep 12 05:01:06 2024 From: sadaniel at us.ibm.com (Steve Daniels) Date: Thu, 12 Sep 2024 04:01:06 +0000 Subject: [gpfsug-discuss] Scale Day in NYC October 1, 2024 In-Reply-To: References: Message-ID: Chris that link appears to be broken It has Database Error (spectrumscale.org) Steven A. Daniels Fax and Voice: 303-810-1229 From: gpfsug-discuss On Behalf Of CHRIS MAESTAS Sent: Tuesday, September 10, 2024 11:22 AM To: gpfsug main discussion list Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Hi Ryan, We have been talking about re-starting with the digital series of talks that are online. Stay tuned and let me see if we can kick this off starting with the latest what?s new talk in Scale 5.?2.?1! The previous series is here: https:?//storagescale.?org/experttalks/ Hi Ryan, We have been talking about re-starting with the digital series of talks that are online. Stay tuned and let me see if we can kick this off starting with the latest what?s new talk in Scale 5.2.1! The previous series is here: https://storagescale.org/experttalks/ -cdm From: gpfsug-discuss > on behalf of Ryan Novosielski > Date: Tuesday, September 10, 2024 at 9:29?AM To: gpfsug main discussion list > Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' On Sep 9, 2024, at 13:44, CHRIS MAESTAS > wrote: Hello everyone, There will be a Scale User Group event in New York City on October 1st, 2024. We have an exciting agenda covering user stories, roadmap update, the latest insights from the user community, development, research, NVIDIA, SuperMicro and DST, plus access to IBM experts and your peers. We look forward to welcoming you to this event. Register here - http://ibm.biz/ScaleNY ps It may say it?s online, but it?s in person! * https://storagescale.org/nyc-2024-scale-day/ * https://storagescale.org/event/nyc2024-ibm-storage-scale-users-group/ -- Chief Troublemaking Officer Chris _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: From Luis.Bolinches at worldquant.com Fri Sep 13 06:45:26 2024 From: Luis.Bolinches at worldquant.com (Bolinches, Luis (WorldQuant)) Date: Fri, 13 Sep 2024 05:45:26 +0000 Subject: [gpfsug-discuss] [EXTERNAL] Re: Scale Day in NYC October 1, 2024 In-Reply-To: References: Message-ID: Hi Is the whole https://www.spectrumscaleug.org/ that has the DB issue it seems I know Ulf had some hand on that before but is a non IBM system or org so not sure how is managing that these days. * https://storagescale.org/nyc-2024-scale-day/ * https://storagescale.org/event/nyc2024-ibm-storage-scale-users-group/ Are linking to spectrumscaleug.org so cant register neither ? -- Yst?v?llisin terveisin/Regards/Saludos/Salutations/Salutacions Luis Bolinches WQ Aligned Infrastructure "If you always give you will always have" -- Anonymous https://www.credly.com/users/luis-bolinches/badges From: gpfsug-discuss On Behalf Of Steve Daniels Sent: Thursday, 12 September 2024 7.01 To: gpfsug main discussion list Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Chris that link appears to be broken It has Database Error (spectrumscale.org) Steven A. Daniels Fax and Voice: 303-810-1229 From: gpfsug-discuss > On Behalf Of CHRIS MAESTAS Sent: Tuesday, September 10, 2024 11:22 AM To: gpfsug main discussion list > Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Hi Ryan, We have been talking about re-starting with the digital series of talks that are online. Stay tuned and let me see if we can kick this off starting with the latest what?s new talk in Scale 5.?2.?1! The previous series is here: https:?//storagescale.?org/experttalks/ Hi Ryan, We have been talking about re-starting with the digital series of talks that are online. Stay tuned and let me see if we can kick this off starting with the latest what?s new talk in Scale 5.2.1! The previous series is here: https://storagescale.org/experttalks/ -cdm From: gpfsug-discuss > on behalf of Ryan Novosielski > Date: Tuesday, September 10, 2024 at 9:29?AM To: gpfsug main discussion list > Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' On Sep 9, 2024, at 13:44, CHRIS MAESTAS > wrote: Hello everyone, There will be a Scale User Group event in New York City on October 1st, 2024. We have an exciting agenda covering user stories, roadmap update, the latest insights from the user community, development, research, NVIDIA, SuperMicro and DST, plus access to IBM experts and your peers. We look forward to welcoming you to this event. Register here - http://ibm.biz/ScaleNY ps It may say it?s online, but it?s in person! * https://storagescale.org/nyc-2024-scale-day/ * https://storagescale.org/event/nyc2024-ibm-storage-scale-users-group/ -- Chief Troublemaking Officer Chris _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org ################################################################################### The information contained in this communication is confidential, may be subject to legal privilege, and is intended only for the individual named. If you are not the named addressee, please notify the sender immediately and delete this email from your system. The views expressed in this email are the views of the sender only. Outgoing and incoming electronic communications to this address are electronically archived and subject to review and/or disclosure to someone other than the recipient. ################################################################################### -------------- next part -------------- An HTML attachment was scrubbed... URL: From TROPPENS at de.ibm.com Fri Sep 13 08:06:30 2024 From: TROPPENS at de.ibm.com (Ulf Troppens) Date: Fri, 13 Sep 2024 07:06:30 +0000 Subject: [gpfsug-discuss] Scale Day in NYC October 1, 2024 In-Reply-To: References: Message-ID: I have informed the admins team to look in the web page issue. Here is the link to the IBM event page: http://ibm.biz/ScaleNY Ulf Troppens Product Manager - IBM Storage for Data and AI, Data-Intensive Workflows IBM Deutschland Research & Development GmbH Vorsitzender des Aufsichtsrats: Wolfgang Wendt / Gesch?ftsf?hrung: David Faller Sitz der Gesellschaft: B?blingen / Registergericht: Amtsgericht Stuttgart, HRB 243294 ________________________________ From: gpfsug-discuss on behalf of Bolinches, Luis (WorldQuant) Sent: Friday, September 13, 2024 7:45 AM To: gpfsug main discussion list Subject: Re: [gpfsug-discuss] [EXTERNAL] Re: Scale Day in NYC October 1, 2024 Hi Is the whole https:?//www.?spectrumscaleug.?org/ that has the DB issue it seems I know Ulf had some hand on that before but is a non IBM system or org so not sure how is managing that these days. https:?//storagescale.?org/nyc-2024-scale-day/ Hi Is the whole https://www.spectrumscaleug.org/ that has the DB issue it seems I know Ulf had some hand on that before but is a non IBM system or org so not sure how is managing that these days. * https://storagescale.org/nyc-2024-scale-day/ * https://storagescale.org/event/nyc2024-ibm-storage-scale-users-group/ Are linking to spectrumscaleug.org so cant register neither ? -- Yst?v?llisin terveisin/Regards/Saludos/Salutations/Salutacions Luis Bolinches WQ Aligned Infrastructure "If you always give you will always have" -- Anonymous https://www.credly.com/users/luis-bolinches/badges From: gpfsug-discuss On Behalf Of Steve Daniels Sent: Thursday, 12 September 2024 7.01 To: gpfsug main discussion list Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Chris that link appears to be broken It has Database Error (spectrumscale.org) Steven A. Daniels Fax and Voice: 303-810-1229 From: gpfsug-discuss > On Behalf Of CHRIS MAESTAS Sent: Tuesday, September 10, 2024 11:22 AM To: gpfsug main discussion list > Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Hi Ryan, We have been talking about re-starting with the digital series of talks that are online. Stay tuned and let me see if we can kick this off starting with the latest what?s new talk in Scale 5.?2.?1! The previous series is here: https:?//storagescale.?org/experttalks/ Hi Ryan, We have been talking about re-starting with the digital series of talks that are online. Stay tuned and let me see if we can kick this off starting with the latest what?s new talk in Scale 5.2.1! The previous series is here: https://storagescale.org/experttalks/ -cdm From: gpfsug-discuss > on behalf of Ryan Novosielski > Date: Tuesday, September 10, 2024 at 9:29?AM To: gpfsug main discussion list > Subject: [EXTERNAL] Re: [gpfsug-discuss] Scale Day in NYC October 1, 2024 Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- Any online option? It would be appreciated by me personally (and compatible with the format, if previous sessions were any indicator), and in 2024 is an accessbility issue. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' On Sep 9, 2024, at 13:44, CHRIS MAESTAS > wrote: Hello everyone, There will be a Scale User Group event in New York City on October 1st, 2024. We have an exciting agenda covering user stories, roadmap update, the latest insights from the user community, development, research, NVIDIA, SuperMicro and DST, plus access to IBM experts and your peers. We look forward to welcoming you to this event. Register here - http://ibm.biz/ScaleNY ps It may say it?s online, but it?s in person! * https://storagescale.org/nyc-2024-scale-day/ * https://storagescale.org/event/nyc2024-ibm-storage-scale-users-group/ -- Chief Troublemaking Officer Chris _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org ################################################################################### The information contained in this communication is confidential, may be subject to legal privilege, and is intended only for the individual named. If you are not the named addressee, please notify the sender immediately and delete this email from your system. The views expressed in this email are the views of the sender only. Outgoing and incoming electronic communications to this address are electronically archived and subject to review and/or disclosure to someone other than the recipient. ################################################################################### -------------- next part -------------- An HTML attachment was scrubbed... URL: From p.childs at qmul.ac.uk Fri Sep 13 09:25:47 2024 From: p.childs at qmul.ac.uk (Peter Childs) Date: Fri, 13 Sep 2024 08:25:47 +0000 Subject: [gpfsug-discuss] mmhealth alerts too quickly Message-ID: We have a nagios alert that watches the output of mmhealth and alerts us if Scale is unhappy on a node. its fairly simple and straight forward, and is very good at letting us know quickly to simple issues. Since we upgraded to 5.1.9-5 we're getting random nodes moaning about lost connections when ever a another machine is rebooted or stops working. This is great, however there does not seam to be any great way to acknowledge the alerts, or close the connections gracefully if the machine is turned off rather than actually failing. I'm aware of the "mmhealth --refresh" method but I've not actually seen this achieve anything and normally endup running "mmsysmoncontrol restart" to get the message to reset. The problem is we don't exactly want to lose the alerts, they are useful if there is a problem, but it would be nice if they where a little more helpful and could be acknowledged. Maybe mmshutdown just needs to close all the cluster connections gracefully so that other nodes don't moan. I've always found it a little abrupt. The other main issue we have is the old memory leak in our ESS5000 https://www.ibm.com/support/pages/node/7027786 We have been working with IBM over the last 18 months on this issue, but still no resolution is in sight and I'm not sure the workaround is relevant any more. Also we found going straight from 5.1.2 to 5.2.0 (to 5.2.1) is a stable upgrade path, and its best to pass though 5.1.9 first. I'm not sure if there is an issue here and something needs adding to the release notes, but that is certainly what we discovered. I hope our findings help others. Peter Childs From stefan.dietrich at desy.de Fri Sep 13 13:15:12 2024 From: stefan.dietrich at desy.de (Dietrich, Stefan) Date: Fri, 13 Sep 2024 14:15:12 +0200 (CEST) Subject: [gpfsug-discuss] mmhealth alerts too quickly In-Reply-To: References: Message-ID: <1555998480.67550472.1726229712150.JavaMail.zimbra@desy.de> Hello Peter, > Since we upgraded to 5.1.9-5 we're getting random nodes moaning about lost > connections when ever a another machine is rebooted or stops working. This is > great, however there does not seam to be any great way to acknowledge the > alerts, or close the connections gracefully if the machine is turned off rather > than actually failing. it's possible to resolve event in mmhealth: # mmhealth event resolve Missing arguments. Usage: mmhealth event resolve {EventName} [Identifier] -> `mmhealth event resolve cluster_connections_down AFFECTED_IP` should do the trick. In our clusters, a regular reboot doesn't seem to trigger this event. All our nodes are running Scale >= 5.2.0 Regards, Stefan -- ------------------------------------------------------------------------ Stefan Dietrich Deutsches Elektronen-Synchrotron (IT-Systems) Ein Forschungszentrum der Helmholtz-Gemeinschaft Notkestr. 85 phone: +49-40-8998-4696 22607 Hamburg e-mail: stefan.dietrich at desy.de Germany ------------------------------------------------------------------------ From ralph.wuerthner at de.ibm.com Fri Sep 13 13:59:51 2024 From: ralph.wuerthner at de.ibm.com (Ralph Wurthner) Date: Fri, 13 Sep 2024 12:59:51 +0000 Subject: [gpfsug-discuss] InfiniBand and OmiPath NSD servers In-Reply-To: References: Message-ID: <9e5d6f5e2e5887847fb8ea98d08076d2d8b43826.camel@de.ibm.com> Hello Sean, you wrote: > Does this mean that we wouldn't be able to consolidate both IB & OPA > HCA's in NSD servers and would have to have 2 types of NSD servers? > 1) InfiniBand exporting and 2) OmniPath exporting? Eventually there is a solution to your problem. We will have to do some internal research and will come back to you later. Mit freundlichen Gr??en / Kind regards Ralph W?rthner IBM Storage Scale Development IBM Systems & Technology Group, Systems Software Development ________________________________ IBM Data Privacy Statement IBM Deutschland Research & Development GmbH / Vorsitzende des Aufsichtsrats: Wolfgang Wendt Gesch?ftsf?hrung: David Faller Sitz der Gesellschaft: B?blingen / Registergericht: Amtsgericht Stuttgart, HRB 243294 -----Original Message----- From: Sean Mc Grath > Reply-To: gpfsug main discussion list > To: gpfsug-discuss at gpfsug.org > Subject: [EXTERNAL] [gpfsug-discuss] InfiniBand and OmiPath NSD servers Date: 09/06/2024 05:52:55 PM Hi, We have a GPFS cluster where the NSD servers mount the storage over fibre channel and export the file system over InfiniBand for clients. We will be getting some used equipment that uses OmniPath. As per the "IBM Storage Scale Frequently Asked Questions and Answers" it states [1]: RDMA is not supported on a node when both Mellanox HCAs and Cornelis Networks Omni-Path HFIs are enabled for RDMA. Does this mean that we wouldn't be able to consolidate both IB & OPA HCA's in NSD servers and would have to have 2 types of NSD servers? 1) InfiniBand exporting and 2) OmniPath exporting? If so, is it then a matter of using the Multi-Rail over TCP "subnets =" setting in mmchonfig to distinguish which nsd server the clients should connect to? [2]. Or am I completely miss understanding all this? Many thanks in advance. Sean [1] https://www.ibm.com/docs/en/STXKQY/pdf/gpfsclustersfaq.pdf [2] https://www.ibm.com/docs/en/storage-scale/5.1.6?topic=configuring-multi-rail-over-tcp-mrot --- Sean McGrath smcgrat at tcd.ie Senior Systems Administrator Research IT, IT Services, Trinity College Dublin https://www.tcd.ie/itservices/ https://www.tchpc.tcd.ie/ _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: From novosirj at rutgers.edu Fri Sep 13 15:33:44 2024 From: novosirj at rutgers.edu (Ryan Novosielski) Date: Fri, 13 Sep 2024 14:33:44 +0000 Subject: [gpfsug-discuss] mmhealth alerts too quickly In-Reply-To: <1555998480.67550472.1726229712150.JavaMail.zimbra@desy.de> References: <1555998480.67550472.1726229712150.JavaMail.zimbra@desy.de> Message-ID: On Sep 13, 2024, at 08:15, Dietrich, Stefan wrote: Hello Peter, Since we upgraded to 5.1.9-5 we're getting random nodes moaning about lost connections when ever a another machine is rebooted or stops working. This is great, however there does not seam to be any great way to acknowledge the alerts, or close the connections gracefully if the machine is turned off rather than actually failing. it's possible to resolve event in mmhealth: # mmhealth event resolve Missing arguments. Usage: mmhealth event resolve {EventName} [Identifier] -> `mmhealth event resolve cluster_connections_down AFFECTED_IP` should do the trick. In our clusters, a regular reboot doesn't seem to trigger this event. All our nodes are running Scale >= 5.2.0 Our clusters (5.1.9-3 on the client side, and either 5.1.5-1 or 5.1.9-2 on the storage side) also show downed connections, but I wish this were somehow tunable. A single downed client that?s not even part of the same cluster is not a reason to alert us on our storage cluster. We monitor MMHEALTH via Nagios, and so we?re occasionally getting messages about a single client. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' -------------- next part -------------- An HTML attachment was scrubbed... URL: From NSCHULD at de.ibm.com Fri Sep 13 15:58:56 2024 From: NSCHULD at de.ibm.com (Norbert Schuld) Date: Fri, 13 Sep 2024 14:58:56 +0000 Subject: [gpfsug-discuss] mmhealth alerts too quickly In-Reply-To: References: <1555998480.67550472.1726229712150.JavaMail.zimbra@desy.de> Message-ID: Good news on that matter, 5.2.2.0 will have an option to ignore those events when they come from a client cluster. Mit freundlichen Gr??en / Kind regards Norbert Schuld Software Engineer, Release Architect IBM Storage Scale IBM Systems / 00E636 Br?sseler Stra?e 1-3 60327 Frankfurt Phone: +49-160-7070335 E-Mail: nschuld at de.ibm.com IBM Data Privacy Statement IBM Deutschland Research & Development GmbH Vorsitzender des Aufsichtsrats: Wolfgang Wendt / Gesch?ftsf?hrung: David Faller Sitz der Gesellschaft: B?blingen / Registergericht: Amtsgericht Stuttgart, HRB 243294 From: gpfsug-discuss On Behalf Of Ryan Novosielski Sent: Friday, September 13, 2024 4:34 PM To: gpfsug main discussion list Subject: [EXTERNAL] Re: [gpfsug-discuss] mmhealth alerts too quickly On Sep 13, 2024, at 08:?15, Dietrich, Stefan wrote: Hello Peter, Since we upgraded to 5.?1.?9-5 we're getting random nodes moaning about lost connections when ever a another machine is rebooted or stops working.? On Sep 13, 2024, at 08:15, Dietrich, Stefan > wrote: Hello Peter, Since we upgraded to 5.1.9-5 we're getting random nodes moaning about lost connections when ever a another machine is rebooted or stops working. This is great, however there does not seam to be any great way to acknowledge the alerts, or close the connections gracefully if the machine is turned off rather than actually failing. it's possible to resolve event in mmhealth: # mmhealth event resolve Missing arguments. Usage: mmhealth event resolve {EventName} [Identifier] -> `mmhealth event resolve cluster_connections_down AFFECTED_IP` should do the trick. In our clusters, a regular reboot doesn't seem to trigger this event. All our nodes are running Scale >= 5.2.0 Our clusters (5.1.9-3 on the client side, and either 5.1.5-1 or 5.1.9-2 on the storage side) also show downed connections, but I wish this were somehow tunable. A single downed client that?s not even part of the same cluster is not a reason to alert us on our storage cluster. We monitor MMHEALTH via Nagios, and so we?re occasionally getting messages about a single client. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' -------------- next part -------------- An HTML attachment was scrubbed... URL: From novosirj at rutgers.edu Fri Sep 13 16:17:24 2024 From: novosirj at rutgers.edu (Ryan Novosielski) Date: Fri, 13 Sep 2024 15:17:24 +0000 Subject: [gpfsug-discuss] InfiniBand and OmiPath NSD servers In-Reply-To: References: Message-ID: We had an environment like this and had gateway nodes between the OPA and IB networks that had both cards. The DDN NetScaler in this case was IB only. Doesn?t answer your question, but that does work. -- #BlackLivesMatter ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novosirj at rutgers.edu || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB A555B, Newark `' On Sep 6, 2024, at 11:52, Sean Mc Grath wrote: Hi, We have a GPFS cluster where the NSD servers mount the storage over fibre channel and export the file system over InfiniBand for clients. We will be getting some used equipment that uses OmniPath. As per the "IBM Storage Scale Frequently Asked Questions and Answers" it states [1]: RDMA is not supported on a node when both Mellanox HCAs and Cornelis Networks Omni-Path HFIs are enabled for RDMA. Does this mean that we wouldn't be able to consolidate both IB & OPA HCA's in NSD servers and would have to have 2 types of NSD servers? 1) InfiniBand exporting and 2) OmniPath exporting? If so, is it then a matter of using the Multi-Rail over TCP "subnets =" setting in mmchonfig to distinguish which nsd server the clients should connect to? [2]. Or am I completely miss understanding all this? Many thanks in advance. Sean [1] https://www.ibm.com/docs/en/STXKQY/pdf/gpfsclustersfaq.pdf [2] https://www.ibm.com/docs/en/storage-scale/5.1.6?topic=configuring-multi-rail-over-tcp-mrot --- Sean McGrath smcgrat at tcd.ie Senior Systems Administrator Research IT, IT Services, Trinity College Dublin https://www.tcd.ie/itservices/ https://www.tchpc.tcd.ie/ _______________________________________________ gpfsug-discuss mailing list gpfsug-discuss at gpfsug.org http://gpfsug.org/mailman/listinfo/gpfsug-discuss_gpfsug.org -------------- next part -------------- An HTML attachment was scrubbed... URL: From Marcy.D.Cortes at wellsfargo.com Mon Sep 16 21:30:04 2024 From: Marcy.D.Cortes at wellsfargo.com (Cortes, Marcy D.) Date: Mon, 16 Sep 2024 20:30:04 +0000 Subject: [gpfsug-discuss] mmhealth alerts too quickly In-Reply-To: References: <1555998480.67550472.1726229712150.JavaMail.zimbra@desy.de> Message-ID: Stefan, do you have a case open? I think this is what we are seeing too. Not worried about the alerts, but MQ qmgrs decide the fs is too slow and it dies. Marcy Since we upgraded to 5.1.9-5 we?re getting random nodes moaning about lost connections when ever a another machine is rebooted or stops working. This is great, however there does not seam to be any great way to acknowledge the alerts, or close the connections gracefully if the machine is turned off rather than actually failing. -------------- next part -------------- An HTML attachment was scrubbed... URL: From snowdonj at us.ibm.com Tue Sep 17 21:45:21 2024 From: snowdonj at us.ibm.com (Jane Snowdon) Date: Tue, 17 Sep 2024 20:45:21 +0000 Subject: [gpfsug-discuss] New Member Message-ID: Hi, My name is Jane Snowdon with IBM Research in Yorktown Heights, NY. I am a new member of the Spectrum Scale User Group. I am collaborating on a project using Fusion HCI with a university. I look forward to attending the October 1 Spectrum Scale User Group meeting in New York City. Regards, Jane Snowdon, PhD, FAMIA IBM Research and IBM Corporate Technical Strategy SPEED Innovation Engagement Leader IBM Industry Diamond Phone: 1-914-945-2422 E-mail: snowdonj at us.ibm.com IBM T. J. Watson Research Center 1101 Kitchawan Road, Office 33-104 Yorktown Heights, NY 10598 @SnowdonJane https://www.linkedin.com/in/snowdonjane -------------- next part -------------- An HTML attachment was scrubbed... URL: