Issue in parallelization of mesh file in liggghts

error:

Primary job terminated normally, but 1 process returned
a non-zero exit code. Per user-direction, the job has been aborted.
mpirun detected that one or more processes exited with non-zero status, thus causing
the job to be terminated. The first process to do so was:

Process name: [[5963,1],4]
Exit code: 1

This is a forum for LAMMPS and not LIGGGHTS.
Please post your question to the corresponding LIGGGHTS forum.

Please consider my apology; if possible, try to answer. In my understanding liggghts and lammps use similar domain-based decomposition for parallelizing. In my case, some of the same set of elements are processed by different processors. Please suggest some way to avoid that.
And I tried on liggghts forum but didn’t get anything.
Thanks a lot

This is a generic MPI error with no useful information–we cannot give any further help other to say you need to figure where/why it is crashing in LIGGGHTS. Typically the code should give an error message.

A heartfelt thanks for your reply.
Sir, I tried a couples of related to remeshing and mesh quality check but it doesn’t help. please give a look on error.

LIGGGHTS (Version LIGGGHTS-PUBLIC 3.8.0, compiled 2023-03-06-18:23:02 by root, git commit unknown)
Created orthogonal box = (-0.02 0 -0.02) to (0.02 0.04 0.08)
3 by 4 by 8 MPI processor grid

Reading STL file ‘barrel12.stl’ (mesh processing step 1/3)

Reading STL file ‘screw.stl’ (mesh processing step 1/3)
Fix particledistribution/discrete (id pdd1): distribution based on mass%:
pts1: d=9.000000e-04 (max. bounding sphere) mass%=30.000000%
pts2: d=1.000000e-03 (max. bounding sphere) mass%=40.000000%
pts3: d=1.100000e-03 (max. bounding sphere) mass%=30.000000%
Fix particledistribution/discrete (id pdd1): distribution based on number%:
pts1: d=9.000000e-04 (max. bounding sphere) number%=39.687131%
pts2: d=1.000000e-03 (max. bounding sphere) number%=38.575891%
pts3: d=1.100000e-03 (max. bounding sphere) number%=21.736978%
Setting up run at Sun May 14 08:06:25 2023

Import and parallelization of mesh cad1 containing 2183 triangle(s) successful
ERROR: Mesh cad2: elements 4756 and 4792 (lines 0 and 0) are duplicate
ERROR on proc 90: Fix mesh: Bad mesh, cannot continue. You can try re-running with ‘heal auto_remove_duplicates’ (…/surface_mesh_I.h:619)
ERROR: Mesh cad2: elements 4768 and 4804 (lines 0 and 0) are duplicate
ERROR on proc 26: Fix mesh: Bad mesh, cannot continue. You can try re-running with ‘heal auto_remove_duplicates’ (…/surface_mesh_I.h:619)
ERROR: Mesh cad2: elements 4768 and 4804 (lines 0 and 0) are duplicate
ERROR on proc 58: Fix mesh: Bad mesh, cannot continue. You can try re-running with ‘heal auto_remove_duplicates’ (…/surface_mesh_I.h:619)
ERROR: Mesh cad2: elements 4780 and 4816 (lines 0 and 0) are duplicate
ERROR on proc 56: Fix mesh: Bad mesh, cannot continue. You can try re-running with ‘heal auto_remove_duplicates’ (…/surface_mesh_I.h:619)
ERROR: Mesh cad2: elements 4853 and 4883 (lines 0 and 34183) are duplicate
ERROR on proc 89: Fix mesh: Bad mesh, cannot continue. You can try re-running with ‘heal auto_remove_duplicates’ (…/surface_mesh_I.h:619)
ERROR: Mesh cad2: elements 4804 and 4768 (lines 0 and 0) are duplicate
ERROR on proc 2: Fix mesh: Bad mesh, cannot continue. You can try re-running with ‘heal auto_remove_duplicates’ (…/surface_mesh_I.h:619)
ERROR: Mesh cad2: elements 4827 and 4857 (lines 0 and 34001) are duplicate
ERROR on proc 57: Fix mesh: Bad mesh, cannot continue. You can try re-running with ‘heal auto_remove_duplicates’ (…/surface_mesh_I.h:619)
ERROR: Mesh cad2: elements 4816 and 4780 (lines 0 and 0) are duplicate
ERROR on proc 32: Fix mesh: Bad mesh, cannot continue. You can try re-running with ‘heal auto_remove_duplicates’ (…/surface_mesh_I.h:619)

SIGINT/SIGTERM caught - Writing restart on next occasion and quitting after that.

Apparently your mesh file has duplicate elements and is thus corrupted. You will need to contact the LIGGGHTS developers for more help, this is not related to anything in LAMMPS.

sure, sir
I will try to figure it out. And thanks a lot for your replies.