|
[Sponsors] |
CFD workstation configuration calling for help |
|
LinkBack | Thread Tools | Search this Thread | Display Modes |
July 2, 2020, 11:28 |
CFD workstation configuration calling for help
|
#1 |
New Member
Join Date: Aug 2014
Posts: 18
Rep Power: 12 |
Hi all,
I need a machine for OpenFOAM modeling with budget of $15000~$20000. My typical number of cells is ~125M (500x500x500). I inclined to choose AMD's EPYC 7002 CPU because of their
Here are the reference prices for key components from NEWEGG: CPU: 1xEPYC 7742: $8,513x1 = $8,513 2xEPYC 7542: $4,122x2 = $8,244 ECC RAM (512GB = 32GBx16 optimized for the 8-channel RAM support of CPU): ~$170x16 = $2,710 Thanks in advance. Last edited by Freewill1; July 5, 2020 at 00:02. |
|
July 2, 2020, 11:35 |
|
#2 |
Super Moderator
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,427
Rep Power: 49 |
||
July 2, 2020, 11:36 |
|
#3 |
New Member
Join Date: Aug 2014
Posts: 18
Rep Power: 12 |
||
July 2, 2020, 12:08 |
|
#4 | |||
Super Moderator
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,427
Rep Power: 49 |
Quote:
So if you can squeeze it into your budget, two dual-socket nodes would be ideal. If you want everything inside one workstation-style box, then dual-socket is the way to go. Of course, that option will be slower than two nodes, even if they have lower core count CPUs. Avoid a single CPU with extremely high core count, that's just a waste of money for CFD. Quote:
Sticking to your example: 2x Epyc 7542 will be about twice as fast for running simulations in OpenFOAM, compared to a single Epyc 7742. The latter will be severely limited by a lack of memory bandwidth, and stop scaling somewhere between 24-32 cores. Quote:
With all that out of the way: don't buy the Epyc 7542 either. It costs almost twice as much as an Epyc 7452 (same core count and chip layout, but lower TDP), and the performance difference running OpenFOAM will be negligible. If you feel up to the task of connecting two machines and running the case distributed, two machines with 2x16 or 2x24 cores and 256GB RAM (16x16GB) each would be even better. For memory, make sure to get DDR4-3200 reg ECC. Epyc can't handle unbuffered memory. |
||||
July 2, 2020, 12:57 |
|
#5 | |
New Member
Join Date: Aug 2014
Posts: 18
Rep Power: 12 |
Quote:
I feel surprised by what you mentioned about the bottleneck problem of the memory bandwidth for CFD (maybe other types of numerical computations too), as well as the low efficiency of budget if choosing EPYC 7742/7542, which I was not acquainted with. |
||
July 2, 2020, 15:04 |
|
#6 |
New Member
Join Date: Aug 2014
Posts: 18
Rep Power: 12 |
Hi Alex,
After careful thinking, new questions arise for me. If I manage to set up two nodes working as a cluster, as mentioned above as the best option among others, additional interconnection between nodes will be needed, e.g., Infiniband now at node-to-node speed of ~100Gbps at most. Will this low-bandwidth network counteract the benifits of the much more rapid eight-channel Epyc-RAM exchange and bring in new bottleneck? Also, Infiniband network system seems quite expensive and requires extra budget? I have on idea if I understand these issues correctly. |
|
July 2, 2020, 17:23 |
|
#7 | ||
Super Moderator
Alex
Join Date: Jun 2012
Location: Germany
Posts: 3,427
Rep Power: 49 |
Quote:
Most FV CFD codes have a rather high code balance. Meaning a high amount of data traffic per floating point operation. Combined with machine balance getting smaller and smaller thanks to ever increasing core counts and IPC, with memory bandwidth barely keeping up, memory bandwidth bottlenecks have long been a thing in CFD, and will become more frequent in other applications in the future. Here is what that looks like in OpenFOAM: scaling.png Results were taken from the pinned thread in this sub-forum. The 7551 result is mine. No, this is not just the test case being too small for high core counts. This case can scale to even higher core counts, albeit on several machines. Quote:
Parallelization in OF is implemented via MPI+domain decomposition. The only data that has to be transferred between cores are values on the domain boundaries. This pales in comparison to the amount of data that has to be accessed for the algorithm itself. As a matter of fact, you could probably get away with 10Gig Ethernet as an interconnect for only two nodes. If you want Infiniband, all you need are two cards and a cable for two nodes. No expensive switch required. The main benefit of Infiniband over Ethernet as node interconnect is not bandwidth, but latency. So you don't have to go all in with 100G or even faster Infiniband cards. 40G or 56G would be plenty. |
|||
July 2, 2020, 20:44 |
|
#8 |
New Member
Join Date: Aug 2014
Posts: 18
Rep Power: 12 |
Many thanks, I have learned quite a lot from your expertise.
|
|
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
Workstation fot research position in CFD | Laerte | Hardware | 9 | June 21, 2020 16:17 |
Home workstation for large memory demand CFD | yutsumi | Hardware | 17 | May 5, 2020 10:13 |
Buying refurbished workstation for CFD | fbelga | Hardware | 10 | November 10, 2019 15:12 |
Alienware Area 51 R5 as a CFD workstation | fusij | Hardware | 1 | June 13, 2019 11:15 |
CFD Online Celebrates 20 Years Online | jola | Site News & Announcements | 22 | January 31, 2015 01:30 |