|
[Sponsors] |
April 19, 2023, 13:21 |
|
#701 |
Senior Member
Will Kernkamp
Join Date: Jun 2014
Posts: 365
Rep Power: 14 |
||
April 28, 2023, 13:49 |
|
#702 |
New Member
Joost
Join Date: Mar 2023
Posts: 3
Rep Power: 3 |
Little update on the hobo-cluster (8 x dual socket e5-2670v1 + 1333ddr3) now with infiniband (40GBit QDR) as it was meant to be. I'm basically seeing linear scaling with the GAMG solver and super linear scaling with PCG solver as advised by the AWS team. Super-linear scaling is due to small size of the benchmark relative to the cpu cache according to the theory. Biggest learning has been that the 1us mpi latency offered by infiniband RMDA is really a must for scaling OpenFOAM in a multi-node setup. I first had the mellanox cards running in 10Gbit ethernet mode using the classic tcp stack and scaling was just awful.
GAMG solver: 16 130,1 32 65,2 48 43,1 64 31,5 80 26,5 96 21,5 112 18,3 128 16,2 PGJ solver: Flow Calculation: 16 130,0 32 65,0 48 41,9 64 28,9 80 22,4 96 17,8 112 15,1 128 13,4 I'm pretty happy with the results as-is. I might be able to get another 10-25% increase by overclocking the memory to 1666 and some bios/infiniband tuning. Not sure if it's worth the stability trade-off though, rather hook-up an extra 4 nodes if the need arises. Last edited by Lavos; April 28, 2023 at 16:40. |
|
April 28, 2023, 21:17 |
|
#703 | |
Senior Member
Will Kernkamp
Join Date: Jun 2014
Posts: 365
Rep Power: 14 |
Quote:
|
||
May 2, 2023, 06:15 |
|
#704 | |
New Member
Joost
Join Date: Mar 2023
Posts: 3
Rep Power: 3 |
Quote:
Last edited by Lavos; May 2, 2023 at 07:48. |
||
May 2, 2023, 11:52 |
|
#705 | |
Senior Member
Will Kernkamp
Join Date: Jun 2014
Posts: 365
Rep Power: 14 |
Quote:
Go for it! |
||
May 24, 2023, 14:42 |
Ryzen 7700X
|
#706 |
New Member
Andrew
Join Date: Apr 2012
Posts: 15
Rep Power: 14 |
Bench results for modern workstation/desktop when on a budget
HW: AMD Ryzen 7700X (8-core Zen4), MSI MAG B650, 2*16Gb DDR5 (XMP 6200MHz C40, Hynix M-die based) HW tuning: SMT off, PBO on, Custom optimizer to reduce core voltage by 30 mW, timings, subtimings of memory carefully optimized to 6200Mhz 30-37...etc, FCLK 2133MHz SW: Win: Win 10 Pro 22H2, WSL2, OF10 on Ubuntu 22.04.2Results (average on 3 runs for benchv02 from thread head): Win10 + WSL2 Cores | Wall (flow calculation) time, s -- Meshing time, s 1 | 312.3 -- 636.5 2 | 189.2 -- 430.6 4 | 130.2 -- 243.3 6 | 112.5 -- 202.3 8 | 109.9 -- 184.2 Linux native Cores | Wall (flow calculation) time, s -- Meshing time, s 1 | 331.5 -- 567.0 2 | 192.9 -- 399.4 4 | 126.2 -- 241.0 6 | 110.3 -- 209.4 8 | 105.9 -- 162.9 Conclusions: decent machine for pre- post-processing (see respectable meshing times of 160ish seconds); even capable of light calculations, but .. IMHO this particular model (and may be the latest Ryzen consumer line as a whole except for *X3D models) does not use full capability of fast DDR5 modules. 6200-6400Mhz is typically the highest sustainable frequency, memory bandwidth is still gimmicky with Infinity Fabric. This particular CPU has 1 CCX module, which (reportedly and consistent with my observations) likely makes FCLK another bottleneck in memory read tasks. All-in-all, performance-wise in CFD workloads it is more like 5800X3D, and considerably lags behind rivals from latest 13*00k Intel line that are capable of achieving higher memory bandwidth. Still, a decent upgrade for 2-3 year consumer-class hardware as a relatively quiet desktop workstation. |
|
May 31, 2023, 01:54 |
|
#707 | |
Senior Member
Dongyue Li
Join Date: Jun 2012
Location: Beijing, China
Posts: 844
Rep Power: 18 |
Quote:
Alright, attachment does not work, please download this one: https://www.cfd-china.com/assets/upl...9-2000w.tar.xz
__________________
My OpenFOAM algorithm website: http://dyfluid.com By far the largest Chinese CFD-based forum: http://www.cfd-china.com/category/6/openfoam We provide lots of clusters to Chinese customers, and we are considering to do business overseas: http://dyfluid.com/DMCmodel.html Last edited by sharonyue; May 31, 2023 at 07:54. |
||
July 11, 2023, 23:19 |
|
#708 | |
Member
|
Just wondering if ddr5 6000MHz would be faster than this.
Quote:
|
||
July 11, 2023, 23:41 |
|
#709 | |
Member
|
Those numbers are just too great!
However would you mind checking if the cases with 48+ threads actually finished normally? I am curious just because the improvement is huge, compared with normal L3 7003. Quote:
|
||
July 24, 2023, 03:42 |
|
#710 |
Member
Join Date: Sep 2010
Location: Leipzig, Germany
Posts: 95
Rep Power: 16 |
Hi Yan,
I checked it and all runs finished as intended. |
|
July 24, 2023, 05:56 |
|
#711 |
Super Moderator
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,426
Rep Power: 49 |
The results look pretty tame compared to "normal" Epyc Milan without 3D-Vcache.
Comparing to my results with two 7543: OpenFOAM benchmarks on various hardware Code:
#threads | 7543 | 7573X =========|========|======= 01 | 471.92 | 492.5 02 | 227.14 | --- 04 | 108.51 | 113.53 08 | 52.11 | 57.91 16 | 28.81 | 31.88 32 | 18.11 | 22.34 48 | 15.46 | 17.17 64 | 13.81 | 11.55 |
|
July 25, 2023, 11:06 |
Genoa X OpenFOAM performance information released
|
#712 |
Member
dab bence
Join Date: Mar 2013
Posts: 47
Rep Power: 13 |
||
July 25, 2023, 12:34 |
|
#713 | |
Senior Member
Will Kernkamp
Join Date: Jun 2014
Posts: 365
Rep Power: 14 |
Quote:
That is based on a 100x40x40 grid, which is really a small problem. The benefit of the L3 cache reduces the larger the problem. |
||
July 26, 2023, 14:45 |
|
#714 |
New Member
Join Date: Aug 2022
Posts: 8
Rep Power: 4 |
Just take note that the link shows results for the new Genoa-X which has a revised microarchitecture and increased L1 and L2 cache per core, so it more likely scales differently than Milan-X.
|
|
July 26, 2023, 21:21 |
|
#715 | |
Senior Member
Will Kernkamp
Join Date: Jun 2014
Posts: 365
Rep Power: 14 |
Quote:
I ran a dual xeon v2 system on the phoronix 30M and 60M OpenFOAM test. I compared 2xE5-4627 v2 (16 cores) to 2xE5-2697 v2 (24 cores). (The additional cores beyond 16 don't add much). The difference was quite large on the 30M problem in favor of the E5-2697v2, However, the difference decreased for the 60M problem. I attributed the difference to the 50% larger cache of this processor. The larger the problem gets, the more the equal bandwidth to memory equalizes the run time. On the 2M OpenFOAM Benchmark, the 2xE5-4627v2 completes in 100 seconds and the 2xE5-2697v2 in 86 seconds. I tried to look up the openbenchmark.org results, but gave up. That website is badly in need of a usable interface. |
||
July 27, 2023, 05:22 |
|
#716 | |
Member
|
Your 7543 is fast, for sure.
But I think and fair comparison would be runing the case without mpi tuning. (or with the same level of tuning, but this is sometimes difficult..) Quote:
|
||
July 27, 2023, 14:00 |
|
#717 | |
New Member
Join Date: Aug 2022
Posts: 8
Rep Power: 4 |
Quote:
|
||
July 28, 2023, 14:26 |
Slow 5800x3d
|
#718 |
New Member
Johannes
Join Date: Sep 2022
Posts: 2
Rep Power: 0 |
Hi everyone,
I'm having problems replicating the good results seen from the 5800x3d. I'm using 2 sticks of I think single rank 3600MHz memory. DOCP is on in the bios, and I have populated the memory slots in accordance with the manual. I have tried both benchmarks linked in this thread, and openfoam2306 and openfoam11 both packaged and compiled myself. I have also tried turning SMT off in bios, but that does not seem to make a major difference. My OS is Ubuntu 22.04.2 LTS The best results I have gotten are: # cores Wall time (s): ------------------------ 8 171.62 6 167.26 4 174.13 2 235.1 1 409.13 Quite far from what I have seen others report. Under windows the system performs fine in Cinebench so I don't think it a temperature problem. The PC uses custom loop water cooling. Any help getting the performance up would be greatly appreciated. |
|
July 28, 2023, 15:15 |
|
#719 |
Senior Member
Will Kernkamp
Join Date: Jun 2014
Posts: 365
Rep Power: 14 |
I have not run this cpu myself. However, I seem to remember that the RAM was run at 4800 MT/s. That would be 33% faster. If you can run your memory at that frequency you might reduce your run time by 20-25%.
|
|
July 28, 2023, 15:57 |
|
#720 | |
New Member
Johannes
Join Date: Sep 2022
Posts: 2
Rep Power: 0 |
Quote:
|
||
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
How to contribute to the community of OpenFOAM users and to the OpenFOAM technology | wyldckat | OpenFOAM | 17 | November 10, 2017 15:54 |
UNIGE February 13th-17th - 2107. OpenFOAM advaced training days | joegi.geo | OpenFOAM Announcements from Other Sources | 0 | October 1, 2016 19:20 |
OpenFOAM Training Beijing 22-26 Aug 2016 | cfd.direct | OpenFOAM Announcements from Other Sources | 0 | May 3, 2016 04:57 |
New OpenFOAM Forum Structure | jola | OpenFOAM | 2 | October 19, 2011 06:55 |
Hardware for OpenFOAM LES | LijieNPIC | Hardware | 0 | November 8, 2010 09:54 |