|
[Sponsors] |
June 3, 2010, 14:30 |
DamBreak tuorial
|
#21 | ||
Member
Join Date: Dec 2009
Posts: 39
Rep Power: 17 |
Hi all!
I'm running through the tutorials and have problems with parallel running in the dam break tutorial. This is the error I get; Quote:
Quote:
Regards Marco |
|||
June 3, 2010, 15:55 |
|
#22 | |
Retired Super Moderator
Bruno Santos
Join Date: Mar 2009
Location: Lisbon, Portugal
Posts: 10,981
Blog Entries: 45
Rep Power: 128 |
Hi Marco!
This has happened to me more then once Quote:
Code:
cd .. By the way, personally I've grown use to using the script foamJob, so in your case, I would use: Code:
foamJob -s -p interFoam Best regards, Bruno
__________________
|
||
June 4, 2010, 17:24 |
|
#23 |
Member
Join Date: Mar 2010
Posts: 31
Rep Power: 16 |
Ok, I'm back. After having installed openfoam 1.6.x, I'm having exactly the same problem running in parallel as I was before. It runs fine on a single processor.
I've tried to attach the output from the screen, which is what I posted above. I'll try to post the details of the case in another message. |
|
June 4, 2010, 17:41 |
gtz with the file stuff
|
#24 |
Member
Join Date: Mar 2010
Posts: 31
Rep Power: 16 |
Here should be the data to recreate the case. You'll need to run blockMesh on it, but hopefully the rest of the files are there. I've saved it as quart.tgz.gz because the uploader would not take quart.tgz. Therefore :
step 1 $ mv quart.tgz.gz quart.tgz step 2 $ tar xvfz quart.tgz and, should be created a directory tree called qcyl. You can descend into this to run blockMesh, etc. It has been running for days with 1 proc, but crashes immediately with 2 or more. I've got a simple decomposition through the z-plane with 2 procs in the decomposeParDict. Anyway, thanks for any ideas. |
|
June 6, 2010, 07:56 |
|
#25 |
Retired Super Moderator
Bruno Santos
Join Date: Mar 2009
Location: Lisbon, Portugal
Posts: 10,981
Blog Entries: 45
Rep Power: 128 |
Hi Bunni,
Well, the same things that happen with you, have happened with me too. I've confirmed that my OpenMPI is working with OpenFOAM, by testing parallelTest and the interFoam/laminar/damBreak case with dual core parallel execution. edit: I forgot to mentioned that I used Ubuntu 8.04 i686 and OpenFOAM 1.6.x I've managed to solve in part the error you get. Just edit the file "OpenFOAM-1.6.x/etc/settings.sh", find the variable minBufferSize and increase the buffer size: Code:
# Set the minimum MPI buffer size (used by all platforms except SGI MPI) # ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ minBufferSize=150000000 But by the tests I've made of increasing the buffer size, it only seems to postpone the crash farther in time. interFoam seems to always crashes during the preparation part of the case. Now I know why other users report that it freezes... in fact, it just takes reaaally long playing around with meshes and memory and MPI messages, at least AFAIK, and sooner or later it will just crash without doing anything useful edit2: yep, 400MB of buffer and it still crashes... So, Bunni, I suggest that you try increasing that buffer variable, in an attempt to avoid the crashing. But my best bet is what I've said to you previously about OpenFOAM 1.6: this seems to be a bug in OpenFOAM, which apparently is yet to be fixed! So please post a bug report on the Bug report part of the OpenFOAM forum: http://www.cfd-online.com/Forums/openfoam-bugs/ If you want to save some time, you can refer to your post #23 from here and onward! Best regards, Bruno
__________________
Last edited by wyldckat; June 6, 2010 at 09:27. Reason: missed a few words... and forgot to mention Linux version |
|
June 6, 2010, 21:56 |
thanks
|
#26 |
Member
Join Date: Mar 2010
Posts: 31
Rep Power: 16 |
Thanks for checking that. I will post a bug report. I'm running on fedora and centos. I'll check out that variable change. As for right now, it's been running on a single processor without crashing for 4 days, so the run itself is stable.
|
|
September 20, 2011, 13:35 |
|
#27 |
New Member
Perry L. Johnson
Join Date: Feb 2011
Location: Orlando, FL, USA
Posts: 17
Rep Power: 15 |
Hello,
I have recently encountered the same problem as Nolwenn and Gonzalo regarding the stopping of a solver at the first time loop without any error message or the job exiting the queue (procs still occupied at 100%). I am in OF1.7.1 (with gcc 4.5.1) on a cluster with RHEL 5.4. The issue only occurs when running large cases, smaller cases work perfectly fine; but, there is plenty of memory per node even for the large cases (~50GB). The parallelTest utility reports fine as suggested above. Is there any knowledge to fixing this issue besides switching compilers? If not, which compilers should I switch to for OF1.7.1, since there is no default compiler? Thanks in advance for any helpful advice, Perry |
|
September 20, 2011, 17:36 |
|
#28 |
Retired Super Moderator
Bruno Santos
Join Date: Mar 2009
Location: Lisbon, Portugal
Posts: 10,981
Blog Entries: 45
Rep Power: 128 |
Greetings Perry,
Before I answer you, I just want to wrap up the solution to bunni's predicament - the thread where the solution is, is this one: http://www.cfd-online.com/Forums/ope...-parallel.html Now back to you Perry: OK, when it comes to the issue of compiler version, there are two/three other libraries whose versions are also important, namely: GMP, MPFR and MPC. For example, from my experience, MPFR 3.0.0 doesn't work very well, so I still hang on to the older 2.4.2 version. As for Gcc 4.5.1, it should work just fine with OpenFOAM 1.7.1. I might on the other hand, be triggering a couple of old bugs that have been solved since then. As I vaguely remember, they were related to some issues with cyclic or wedge or some other special type of patch, that would crash the solver when used in parallel. Aside from such old bugs, one still needs to use (if I'm not mistaken) the "preservePatches" parameter in decomposeParDict. Either way, I've got a blog post where I'm gathering information on how to run in parallel with OpenFOAM (it's accessible from the link on my signature): Notes about running OpenFOAM in parallel The ones that might interest you:
Best regards, Bruno
__________________
|
|
September 20, 2011, 18:56 |
|
#29 | |||||
New Member
Perry L. Johnson
Join Date: Feb 2011
Location: Orlando, FL, USA
Posts: 17
Rep Power: 15 |
Quote:
Quote:
Quote:
Quote:
2) I'm using a custom solver based on simpleFoam (with an extra equation for passive scalar transport), but have also tested on simpleFoam itself with no difference. 3) I have one cyclic patch, 3 directMapped patches, and a number of inlets, outlets, and walls. 4) Right now, I'm using RAS, k-w SST. 5) I have not tried scaling the geometry, but I have run the same geometry on a coarser mesh successfully with the same boundary conditions. I only experience this problem on my fine mesh. Quote:
Thanks very much for your help, Perry |
||||||
September 21, 2011, 16:11 |
|
#30 | |||
Retired Super Moderator
Bruno Santos
Join Date: Mar 2009
Location: Lisbon, Portugal
Posts: 10,981
Blog Entries: 45
Rep Power: 128 |
Hi Perry,
Quote:
Oh, here is the link to the makeGcc file on ThirdParty 2.0.x: https://github.com/OpenFOAM/ThirdPar...master/makeGcc - as you can see, Gcc 4.5.x needs MPFR, GMP and MPC to build properly. Quote:
Quote:
2) OK, then the problem must be elsewhere... 3) Are the directMapped patches also protected by the preserve patches parameter? 4) OK, seems pretty standard... 5) Have you tried visualizing the sub-domains in ParaView, to check where things are being split? Have you executed checkMesh on the fine resolution mesh before decomposing, to verify if the mesh is OK? There is an environment variable that OpenMPI uses that is defined in settings.sh... ah, line 347: https://github.com/OpenCFD/OpenFOAM-...ttings.sh#L347 - try increasing that value, perhaps 10x. Although this is only a valid solution in some cases. And I know I've seen more reports like this before... and if I'm not mistaken, most were related to the patches being split between sub-domains, but my memory hasn't been very trustworthy lately If my memory gets better, I'll search for what I've read in the past and post here. ...Wait... maybe it's the nonBlocking flag: https://github.com/OpenCFD/OpenFOAM-...ntrolDict#L875 - have you tried with other possibilities for the parameter commsType? I know there was a bug report a while back that was fixed in 2.0.x... here we go, it's in fact related to "directMappedPatch", although it might not affect your case: http://www.openfoam.com/mantisbt/view.php?id=280 Best regard and good luck! Bruno
__________________
|
||||
September 21, 2011, 21:05 |
|
#31 | |||||||
New Member
Perry L. Johnson
Join Date: Feb 2011
Location: Orlando, FL, USA
Posts: 17
Rep Power: 15 |
Quote:
Quote:
simple stops after building the mesh and fields, while starting the first time loop: "Time = 1", as if it is taking hours to complete the U-eqn; it also stops here when running serially (just tested) Quote:
Preserving patches for directMapped does not make sense to me, since the directMapped patch is not a shared boundary situation, but rather a case where the inlet looks to the nearest interior cell to a given offset location and finds the value there. Is there a good way to visualize all of the sub-domains in one paraview session? Quote:
1) Two regions not connected by any faces (which is purposeful for my simulation, e.g. one region feeds the other via directMappedPatch). 2) 156 Non-orthogonal faces, but still says OK (max 86.4). 3) 3 skew faces, says that mesh fails, but this has not stopped me in the past. Would you think this problem could be related to 3 skewed faces (max skewness 4.66)? Doesn't seem like skew cells could prevent the solver from running, but I could be wrong...? Quote:
Quote:
Quote:
Thanks for your continued ideas, Perry |
||||||||
September 22, 2011, 15:41 |
|
#32 | |||||||
Retired Super Moderator
Bruno Santos
Join Date: Mar 2009
Location: Lisbon, Portugal
Posts: 10,981
Blog Entries: 45
Rep Power: 128 |
Hi Perry,
Quote:
Quote:
Quote:
Quote:
Quote:
Quote:
Quote:
Best regards and good luck! Bruno
__________________
|
||||||||
September 22, 2011, 23:40 |
|
#33 | |||
New Member
Perry L. Johnson
Join Date: Feb 2011
Location: Orlando, FL, USA
Posts: 17
Rep Power: 15 |
Bruno,
Quote:
Quote:
I have narrowed it down to one of the directMapped B.C.'s (the other two are fine). When I switch it to fixedValue, everything works fine. Switch it back to direct mapped, and it stalls at the first Time loop. The checkMesh utility gives me a cellToRegions file, suggesting that I should use the splitMeshRegions utility and use two different regions. The problematic directMapped boundary pulls its data from a separate domain of the flow that is not connected geometrically. Could this explain the problems I am having? (I am completely unfamiliar with multiple regions in OF, other that the little reading I have done today.) Quote:
Regards, Perry |
||||
September 23, 2011, 15:36 |
|
#34 |
Retired Super Moderator
Bruno Santos
Join Date: Mar 2009
Location: Lisbon, Portugal
Posts: 10,981
Blog Entries: 45
Rep Power: 128 |
Hi Perry,
Mmm... well, at least MPFR doesn't seem to be the one to blame... for now And this is getting further into details that I'm not very familiar with either. Did checkMesh on the coarser mesh give you the same information, that it should divide the mesh into two separate regions? For multiple regions, I only know about two solvers that should support this (I don't know if all other solvers support this or not); they are (and respective tutorials):
Best regards and Good luck! Bruno
__________________
|
|
September 24, 2011, 15:51 |
|
#35 |
New Member
Perry L. Johnson
Join Date: Feb 2011
Location: Orlando, FL, USA
Posts: 17
Rep Power: 15 |
Bruno,
I really appreciate all your help with this issue, and the side-tips along the way. I narrowed it down to the influence of the directMapped boundary condition with the fine mesh I was using, so I played around with meshing until I got one to work. I seem to have resolved the issue just by using a different mesh. Sincere regards, Perry |
|
May 26, 2017, 18:36 |
non-interactive parallel run
|
#36 |
New Member
elham usefi
Join Date: Apr 2016
Location: tabriz,iran
Posts: 13
Rep Power: 10 |
greetings all!
I have installed OF-2.4.0(with gcc-4.8.1 , gmp-5.1.2 , mpc-1.0.1 , mpfr-3.1.2 ) on a cluster with CentOS 6.5 with this instructions HTML Code:
https://openfoamwiki.net/index.php/Installation/Linux/OpenFOAM-2.3.0/CentOS_SL_RHEL I use this command Code:
$ nohup foamJob -s -p simpleFoam & nohup.out is like Code:
3 total processes killed (some possibly by mpirun during cleanup)". |
|
March 21, 2021, 05:56 |
|
#37 | |
New Member
Febriyan Prayoga
Join Date: Apr 2016
Location: Seoul
Posts: 21
Rep Power: 10 |
Quote:
Hi Bruno, I am using OF v2012 on Ubuntu 20.04 Focal Fossa. I could not find file "setting.sh" in "OpenFoam-v2012/etc" folder. I found the file "settings.sh" inside "OpenFoam-v2012/etc/config.csh" and "OpenFoam-v2012/etc/config.csh" folders. Unfortunately I could not find the string " minBufferSize". I wonder whether my OpenFoam installation is correct or not. I followed this instruction: http://openfoamwiki.net/index.php/In...M-v1806/Ubuntu I change the version string v1806 to v2012. I attach the settings files. my big appreciation if you could give me some hints. Thank you in advance.. |
||
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
Unable to run OF in parallel on a multiple-node cluster | quartzian | OpenFOAM | 3 | November 24, 2009 14:37 |
Swap usage on parallel run | nikhilesh | OpenFOAM Bugs | 1 | April 30, 2009 05:42 |
Problem on Parallel Run Setup | Hamidur Rahman | CFX | 0 | September 23, 2007 18:11 |
Windows 64-bit, Distributed Parallel Run Issues... | Erich | CFX | 3 | March 28, 2006 17:36 |
Serial run OK parallel one fails | r2d2 | OpenFOAM Running, Solving & CFD | 2 | November 16, 2005 13:44 |