Parallel Meshing and Remeshing | Efficient Model Optimization
In the latest release, several advancements were added to speed up the generation of meshes in Workbench. For those users that build large models, one new feature will be especially helpful in saving time.
For the first time, Ansys is able to use multiple CPUs to generate a mesh WITHOUT ANY ADDITIONAL LICENSES. Ansys will attempt to divvy the geometry up between the cores so that more of the computer resources is used during this part of the analysis process. The speed-up is especially dramatic for models with lots of bodies connected together with contact regions.
Breaking up the Model Increases Speed
Since the mesh on individual bodies do affect each other, this allows Ansys to break the model into many pieces thus utilizing all cores. If the model is made up of a single multi-body part, the entire part will need to be sent to one CPU so the speed increases may not be fully realized. The image below shows a sampling of the speed increases that were witnessed by Ansys using several customer models.
By default, the program will attempt to use all cores on the machine. One requirement is that each core will need to have at least 2 GB of RAM available for its use. So, if the computer has 4 cores, a minimum of 8 GB of RAM will need to be on the computer.
As an added bonus, other mesh methods (MultiZone Quad/Tri, Patch Independent Tetra, and MultiZone) will be able to use multiple cores as well.
The Number of CPUs that are available to Ansys can be specified in two places in Tool>>Options>>Meshing>>Meshing. The picture below shows the location in the Options menu.
One setting provides the number that can be used for the various mesh methods and the other will set the number that can be used for parallel part meshes. By default, both of these are set to 0 which will permit Ansys to use all cores on the computer. If you would actually like to keep working, a specific number can be entered instead.
We hope you take advantage of this new advancement. Happy Meshing, Everyone!