|
[Sponsors] |
February 16, 2015, 10:19 |
|
#21 |
Senior Member
Andrea Pasquali
Join Date: Sep 2009
Location: Germany
Posts: 142
Rep Power: 17 |
Hi,
I did not see different running times when running on nodes on a single switch. My test was with mesh generation whit the refinement stage in snappyHexMesh. As I said, I did not investigate it in detail yet. I only tried once recompiling mpi and openfoam with intel12 but having still the same (bad) performance with infiniband... Andrea
__________________
Andrea Pasquali |
|
February 18, 2015, 06:09 |
|
#22 |
New Member
Join Date: May 2013
Posts: 23
Rep Power: 13 |
Hello,
So I have tried the rebumberMesh before solving and it looks like it has improved a bit the performances of both single and multiple switches, reducing the running time by ~10%. But I still can't see why the running times are so slow across multiple switches. RenumberMesh or not, we should get roughly the same running time whatever the nodes selected, right ? |
|
February 22, 2015, 15:12 |
|
#23 | |
Retired Super Moderator
Bruno Santos
Join Date: Mar 2009
Location: Lisbon, Portugal
Posts: 10,981
Blog Entries: 45
Rep Power: 128 |
Greetings to all!
@arnaud6: Quote:
Problem is that when 3 switches are used, the tree becomes a lot larger and is sectioned in 3 parts, making it a bit harder to map out communications. Commercial CFD software might already have these kinds of configurations taken into account, by either asking the InfiniBand controls to adjust accordingly, or the CFD software itself tries to balance this out on its own, by placing sub-domains closer to each other on the same machines that share a switch and keeping communication to a minimum when communicating with machines that are connected on other switches. But when you use OpenFOAM, you're probably not taking this into account. I haven't had to deal with this myself, so I have no idea how this is properly configured, but there are at least a few things I can imagine that could work:
Bruno |
||
February 26, 2015, 06:15 |
|
#24 |
New Member
Join Date: May 2013
Posts: 23
Rep Power: 13 |
Hi Bruno,
Thanks for your ideas ! I am looking at the PCG solvers. Would you advice to use the combination PCG for p and PBiCG for other variables or using PCG for p and keep other variables with a smopothSolver/Gauss Seidel ? In my cases it looks like p is the most difficult to solve (at least it is the variable that takes the longest time to be solved at each iteration). The difficulty is that I don't have much control on the nodes thus the switches that will be selected when I submit my parallel job ... I will see what I can do with the IB support. |
|
October 24, 2015, 16:37 |
|
#25 | |
Retired Super Moderator
Bruno Santos
Join Date: Mar 2009
Location: Lisbon, Portugal
Posts: 10,981
Blog Entries: 45
Rep Power: 128 |
Hi arnaud6,
Quote:
The best I could tell you back then and now is that you try running for a few iterations yourself with each configuration. Even the GAMG matrix solver can sometimes be improved if you fine tune the parameters and do some trial and error sessions with your case, because these parameters depend on the case size and how the sub-domains in the case are structured. Either way, I hope you managed to figure this out on your own. Best regards, Bruno |
||
November 4, 2015, 11:49 |
|
#26 | |
New Member
Join Date: Nov 2012
Posts: 27
Rep Power: 14 |
Hi Bruno,
indeed. In my expericence, how the subdomain is structured has strong impact on the performance. So I choose to decompose manually. My problem now is as following: I am running a DNS case (22 Mio. cells) using buoyantPimpleFoam (OF V2.4). The case is a long pipe with an inlet and outlet. The fluid is air. Inlet Re is about 5400. For getting better scalability, I use PCG for pressure equation. If I use perfect gas equation of state, the number of iterations will be around 100, which is acceptable. If I use icopolynom or rhoConst to describe the density, the number of iterations will be around 4000! If I use GAMG for p equation, number of iteration will be under 5, but the scalability is poor with above 500 cores. Does anyone has any opinion? How can I improve PCG solver to decrease the number of iterations? Thank you. Quote:
|
||
November 7, 2015, 12:52 |
|
#27 | |
Retired Super Moderator
Bruno Santos
Join Date: Mar 2009
Location: Lisbon, Portugal
Posts: 10,981
Blog Entries: 45
Rep Power: 128 |
Quote:
__________________
|
||
March 25, 2016, 06:03 |
|
#28 |
New Member
Join Date: May 2013
Posts: 23
Rep Power: 13 |
Sorry for getting back so late on this one. The problem was mpirun 1.6.5. As soon as I switched to mpirun 1.8.3, the slowness disappeared !
|
|
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
Large test case for running OpenFoam in parallel | fhy | OpenFOAM Running, Solving & CFD | 23 | April 6, 2019 10:55 |
Running AMI case in parallel | Kaskade | OpenFOAM Running, Solving & CFD | 3 | March 14, 2016 16:58 |
Parallel Moving Mesh Bug for Multi-patch Case | albcem | OpenFOAM | 0 | May 21, 2009 01:23 |
Parallel Performance of Fluent | Soheyl | FLUENT | 2 | October 30, 2005 07:11 |
PC vs. Workstation | Tim Franke | Main CFD Forum | 5 | September 29, 1999 16:01 |