|
[Sponsors] |
March 5, 2020, 08:21 |
SU2 v 7.0.2 not writing solution files
|
#1 |
New Member
Join Date: Feb 2018
Posts: 27
Rep Power: 8 |
Hi,
I've just switched to v7 (.0.2) and I am having trouble with the solution files. I am running the naca0012 laminar test case (with the .cfg file in TestCases\cont_adj_navierstokes\naca0012_sub) in 1 processor with the mpi binary executable for windows and the code only writes the restart (.csv) and history (.dat) files, but not the solution files (flow and surface_flow). I execute SU2_SOL and it does nothing. Any advice? |
|
March 5, 2020, 09:39 |
|
#2 |
Senior Member
Pedro Gomes
Join Date: Dec 2017
Posts: 466
Rep Power: 14 |
The behaviour of the "OUTPUT_FILES" option is documented here: https://su2code.github.io/docs_v7/Custom-Output/
|
|
March 5, 2020, 10:16 |
|
#3 | |
New Member
Join Date: Feb 2018
Posts: 27
Rep Power: 8 |
Quote:
|
||
March 5, 2020, 10:26 |
|
#4 |
Senior Member
Pedro Gomes
Join Date: Dec 2017
Posts: 466
Rep Power: 14 |
Yup, whenever the major version number increases, you can expect significant changes and loss of backwards compatibility.
|
|
March 6, 2020, 11:41 |
|
#5 |
New Member
cfdjetman
Join Date: Mar 2019
Posts: 25
Rep Power: 7 |
Has anyone had problems writing the surface and flow files when they use more than 16 processors using SU2 v7.0.0?
When I use more than 16 processors, the surface and flow files write incomplete data. I had the same issue when I use the shape_optimization.py script. In this case the deformed mesh file is incomplete and therefore my optimization stops. |
|
March 11, 2020, 05:06 |
|
#6 |
Senior Member
Pedro Gomes
Join Date: Dec 2017
Posts: 466
Rep Power: 14 |
That issue kind of rings a bell. Try 7.0.2 we have monthly releases now so that we can distribute small fixes as early as possible.
|
|
March 14, 2020, 04:24 |
|
#7 |
New Member
cfdjetman
Join Date: Mar 2019
Posts: 25
Rep Power: 7 |
Unfortunately, I have the same problem after installing 7.0.2.
|
|
March 14, 2020, 06:56 |
|
#8 |
Senior Member
Pedro Gomes
Join Date: Dec 2017
Posts: 466
Rep Power: 14 |
Can you give a concrete example? I never had that problem before and I just tried to run the quick start case on 64 cores and the files seem ok.
|
|
March 15, 2020, 16:18 |
|
#9 |
New Member
cfdjetman
Join Date: Mar 2019
Posts: 25
Rep Power: 7 |
I use a cluster to run my simulations. Each node on the cluster has 16 processors. When I use more than 1 node, that’s when the problem begins.
So when I run an airfoil cfd case using more than 1 node, the surface file only writes out values from 0 to 0.3 of the chord length, not all the way till chord length of 1. This does not happen when I use 1 node. When I try to open the flow.dat file in tecplot, I cannot open it as it has some data missing. When I run the airfoil optimization case, it runs the first design iteration, then it deforms the mesh and prints out the mesh_deform.su2 file. This file does not contain the NPOIN part of the mesh file, hence su2 cannot run the cfd simulation for the next design iteration. |
|
March 16, 2020, 04:19 |
|
#10 |
Senior Member
Pedro Gomes
Join Date: Dec 2017
Posts: 466
Rep Power: 14 |
Is this only a problem when you use the python scripts or also if you directly launch SU2_CFD/DEF?
Conceptually there is no difference between what the code does when running on one or multiple nodes, mpi makes all that opaque. That being said I never tried the python scripts on a multi-node environment (I always assumed they would not work as they are not "mpi-ready"). |
|
March 16, 2020, 12:29 |
|
#11 |
New Member
cfdjetman
Join Date: Mar 2019
Posts: 25
Rep Power: 7 |
I have the same problem when I run SU2_CFD/DEF using more than one node.
|
|
April 1, 2020, 21:48 |
|
#12 |
New Member
cfdjetman
Join Date: Mar 2019
Posts: 25
Rep Power: 7 |
Pedro Gomes,
Do you see any reason why this could be happening? |
|
April 3, 2020, 05:10 |
|
#13 |
Senior Member
Pedro Gomes
Join Date: Dec 2017
Posts: 466
Rep Power: 14 |
If no output format is working, I would guess you have some issue with the file system, or the way you prepare the working directory is not adequate (maybe some nodes do not have write permissions).
But I've never setup distributed systems... If I have issues I go to the people that maintain the system. If that is not an option, and you are comfortable programming, edit SU2_CFD.cpp and add instructions to print the mpi rank to screen and create a file with the mpi rank as its name. On screen you should see numbers 0 to n-1, if numbers repeat you are launching two simulations. On disk you should see files 0 to n-1, if some are missing it is either because they failed to open (which you can detect when you open the file in c++) or got lost. If some output formats work and others don't, please open an issue on GitHub. |
|
April 18, 2020, 11:06 |
|
#14 |
Senior Member
Zach Davis
Join Date: Jan 2010
Location: Los Angeles, CA
Posts: 101
Rep Power: 16 |
I don't mean to hijack this thread. I do have a similar issue, but it seems isolated with the Tecplot binary writer in SU2 v7.0.3. I can write the volume solution and surface solution files out just fine if I have OUTPUT_FILES= ( RESTART, PARAVIEW, SURFACE_PARAVIEW). However, if I use the Tecplot writer as OUTPUT_FILES= ( RESTART, TECPLOT, SURFACE_TECPLOT), then SU2_CFD will quit with a cryptic error: Error 137.
I have been able to run the 2D Quick Start example case with the Tecplot writer, and I do get a flow.szplt volume solution in that case. I'm not sure whether the parallel Tecplot writer is running out of memory in my much larger 3-D parallel case, or if there is something else going on. I would have expected to it to error with a Segmentation Fault if memory were an issue, but I don't exactly see that here. If anyone else has experienced anything similar with the Tecplot binary writer in recent releases of SU2 with large grids run in parallel, then I would be interested in knowing that. For now, I'll just continue with the ParaView file format until I have a system with a lot more memory to test on. |
|
June 13, 2020, 04:41 |
|
#15 |
New Member
cfdjetman
Join Date: Mar 2019
Posts: 25
Rep Power: 7 |
Hi Zach Davis,
Have you been able to fix your issue? I changed my OUTPUT_FILES to ( RESTART, PARAVIEW, SURFACE_PARAVIEW) to see if this would resolve my problem, but I have the same issue as before. I am using SU2 v7.0.2. |
|
June 13, 2020, 20:37 |
|
#16 |
New Member
Hernán David Cogollo
Join Date: Jun 2020
Location: Bogotá, Colombia
Posts: 6
Rep Power: 6 |
Hi everyone.
I'm trying to run SU2 on Windows but the output files should be ".vtk" and the files that I obtain are ".vtu", I don´t know the reason of this and I would like what can I do for resolve it. |
|
July 17, 2021, 02:27 |
|
#17 |
Senior Member
Arijit Saha
Join Date: Feb 2019
Location: Singapore
Posts: 132
Rep Power: 7 |
Make the setup in your cfg file as OUTPUT_FILES = ( RESTART, PARAVIEW_ASCII, SURFACE_PARAVIEW_ASCII)
|
|
August 3, 2021, 12:01 |
|
#18 |
Senior Member
|
Hi guys,
I have been experiencing some similar issues referred down here by people. I have 24 cores on a node and when I was going to launch an optimization problem on more than on node let's say 72 cores, I was getting some error. I also had experience what Zach Davis was experiencing with writing tecplot binaries for a medium-size mesh. I had to switch to ASCII OUTPUT_FILES = (RESTART_ASCII, TECPLOT_ASCII, SURFACE_TECPLOT_ASCII) But now I am randomly getting this: MPI_ABORT was invoked on rank 23 in communicator MPI_COMM_WORLD with errorcode 1. NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes. You may or may not see output from other processes, depending on exactly when Open MPI kills them. -------------------------------------------------------------------------- [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] PMIX ERROR: UNREACHABLE in file server/pmix_server.c at line 2193 [node229:197058] 71 more processes have sent help message help-mpi-btl-openib.txt / ib port not selected [node229:197058] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages [node229:197058] 71 more processes have sent help message help-mpi-btl-openib.txt / error in device init [node229:197058] 71 more processes have sent help message help-mpi-api.txt / mpi-abort This happens during the optimization cycle randomly. In case I navigate to the same directory and launch the CFD job manually instead of using FADO to do that everything goes fine. This happens to happen out of blue. It was during the primal solution of the second design iteration, I turned off symlink in FADO and then disappeared and appeared in the fifth design iteration, again primal solution. Any idea? Best, Pay p.s. this is SU2 v 7.1.1 Last edited by pdp.aero; August 3, 2021 at 12:13. Reason: version |
|
August 19, 2021, 10:43 |
|
#19 | |
Senior Member
|
Quote:
Hi there, I just wanna point out the above error had something to do with MPI and rather not SU2. It has been solved by: export OMPI_MCA_btl_openib_allow_ib=1 export OMPI_MCA_btl_openib_if_include="mlx5_0:1" export OMPI_MCA_btl=self,tcp Now everything running fine even with 300 cores. Cheers, Pay |
||
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
[Other] Contribution a new utility: refine wall layer mesh based on yPlus field | lakeat | OpenFOAM Community Contributions | 58 | December 23, 2021 03:36 |
[Other] refineWallLayer Error | Yuby | OpenFOAM Meshing & Mesh Conversion | 2 | November 11, 2021 12:04 |
Writing report files and creating report definitions. | cfd_worker99 | FLUENT | 3 | June 12, 2020 13:55 |
Forcing the creation of solution files | OVS | SU2 | 3 | April 15, 2016 05:13 |
OpenFOAM15 paraFoam bug | koen | OpenFOAM Bugs | 19 | June 30, 2009 11:46 |