Creating explicit walls from complex regions—computational efficiency and wall integrity

Following up on a previous thread (Regions NOT forming single region in region union command - #14 by mmoosa92), I first intended to create a “complex” wall using the union of sub-regions. This region comprises of a cylinder connecting two boxes, similar to the one discussed in the linked thread above. The goal was to use fix wall/region to apply a “soft” repulsion of the atoms and keep them contained in the region. However, it was advisable that such complex region can suffer from some ambiguities, especially when an atom crosses the planes between the sub-regions, and this is exactly what I observed in my simulations. Hence, I am creating the wall with explicit atoms with the following commands:

lattice bcc 1
region simbox block -50 50 -50 50 -55 64.6 units box
create_box 3 simbox

region pore cylinder z 0.0 0.0 11.1 0 14 units box 
region boxOne block INF INF INF INF -50.6 0.5
region boxTwo block INF INF INF INF 13 EDGE
region freeSpace union 3 pore boxOne boxTwo
region carveBoxOne block INF INF INF INF -49.9 -0.1
region carveBoxTwo block INF INF INF INF 13.6 63.6
region carvePore cylinder z 0 0 10.1 -0.1 13.5 units box

create_atoms 3 region freeSpace
delete_atoms region carveBoxOne
delete_atoms region carveBoxTwo
delete_atoms region carvePore

With the above commands, I get the following walls (my simulation box is periodic in x and y, that is why I “carved” the atoms all the way along in those directions):

where I made all the walls double layered (and staggered using the BCC lattice). Below is a side view of the cylindrical region showing the side walls with two layers.

Just to test the integrity of this wall, I simulated some atoms (starting from one box) through Langevin Dynamics and NVE integrator. I am also adding a force from a table through fix wall/table simulating the effect of the solvent and pore implicitly. I am facing two issues:

(1) By simulating the wall using explicit atoms, the computational time increases significantly*. I used fix setforce to set all the forces to zero for the wall atoms. Is there a more efficient way to do this without compromising the computational efficiency significantly? When using multiple processors, I used the following commands:

neighbor 2.0 bin
neigh_modify exclude type 3 3

where type 3 is for the wall atoms. This speeds up the calculations. However, I want to know what are some suggestions to maximize the computational efficiency even when using a single processor.

* I have less than a 100 mobile atoms. For 1 ns, simulation time increases from a couple of seconds to a couple of hours!

(2) Even with a double-layered wall as described above, my simulation often fails due to an atom penetrating the wall (specifically the bottom wall). I am using lj96/cut 1.0 1.0 for the interactions between the wall atoms and all other atoms in my simulation box. Perhaps increasing the \sigma will result in a more packed wall and that can prevent any penetration, however that did not work (for \sigma up to 3.5). Is it advisable to make the wall thicker? However, in doing so, I would have more atoms and this brings me back to the first inquiry. Also, is there a better pair style to use instead of lj96? I intended to use lj93 but it is not available as a pair style but rather in fix wall. I really just want to “softly” bounce the atoms of the walls (simulation box edges and the pore wall). I used the regular lj/cut pair style and I do not see any penetration for at least 10 ns of simulation time, but I was seeking to implement a “softer” repulsion. Any suggestions?

PS: I am using LAMMPS 02Aug2023 version.

Any input is appreciated. Thanks!

For the computational efficiency issue – you may get some mileage out of using atom_modify first to keep the mobile atoms first-in-memory (and omit unintegrated atoms). You may also get some mileage out of simply omitting the wall atoms from your integrator fix.

Beyond that you would need to verify:
a) whether your simulations with and without walls are truly exactly the same other than having the wall particles added
b) if they are, then what portion of the performance timing is showing the most time used? (See the CPU time breakdown at the end of each log file.)

It would be easier, if you showed your entire input script and log file (especially, the last part with timings), so some of the advice below may not be applicable.

About your first question:

  1. Use the group with your mobile atoms in all fixes and computes.
  2. Don’t use fix setforce, the atoms are immobile already because they won’t be integrated anyway.
  3. Dump only positions of mobile atoms
  4. Create compute temp and compute press and use them instead of the standard ones (see thermo_modify manual page).
  5. Add every 1 delay yes to your neigh_modify command.

About the second question, make your wall denser. The distance between its atoms doesn’t need to be tied to \sigma, because the atoms are immobile anyway.

1 Like

Thank you @srtee and @mkanski for the suggestions.

Thank you for suggestion. I will look into how much this can speed up the calculation.

They are truly the same in the sense that only the way I am implementing the containment is different (i.e., virtual wall verses explicit ones). As such, the extra computational time is attributed to the explicit interactions between the wall atoms and mobile atoms in the simulations box, which is provably more expensive than using wall/region for instance.

The log file shows the following breakdown (using 4 processors):

Loop time of 5981.65 on 4 procs for 10000000 steps with 80618 atoms

Performance: 144.442 ns/day, 0.166 hours/ns, 1671.779 timesteps/s, 134.775 Matom-step/s
99.7% CPU use with 4 MPI tasks x no OpenMP threads

MPI task timing breakdown:
Section |  min time  |  avg time  |  max time  |%varavg| %total
---------------------------------------------------------------
Pair    | 336.55     | 356.6      | 379.77     |  92.9 |  5.96
Neigh   | 998.9      | 1047.7     | 1185.2     | 245.4 | 17.52
Comm    | 1361.8     | 1488.7     | 1563.9     | 197.0 | 24.89
Output  | 23.542     | 23.858     | 24.213     |   5.0 |  0.40
Modify  | 2330.7     | 2342.9     | 2354.5     |  17.5 | 39.17
Other   |            | 721.9      |            |       | 12.07

Nlocal:        20154.5 ave       20158 max       20151 min
Histogram: 1 0 1 0 0 0 0 1 0 1
Nghost:        8429.75 ave        8433 max        8427 min
Histogram: 1 0 0 1 0 1 0 0 0 1
Neighs:              0 ave           0 max           0 min
Histogram: 4 0 0 0 0 0 0 0 0 0

where 39% of the total time is consumed in Modify. I am using 5 fix commands in my script, namely 2 fix wall/table, fix setforce for the atom walls, fix langevin, and fix nvz for the integrator.

fix 1 cationInRange wall/table spline 200 zlo -50.2 forces.txt MYPMF 73.9
fix 2 sodiumInRange wall/table spline 200 zlo -50.2 forces.txt NaPMF 73.9
fix freezeWall wallAtoms setforce 0.0 0.0 0.0
fix langevin allIons langevin 300.0 300.0 100.0 3157
fix mobileNVE allIons nve

I am currently running this same simulation with the atom_modify first and every 1 delay yes commands, and I will see how much time does that save.

Your log shows the run finishing in under two hours which is frankly not bad and within line of typical LAMMPS speeds.

The other thing to consider is as @mkanski said – you do not need to include the fixed atoms in the integrator fix. That is, assuming your mobile atoms are only types 1 and 2, the following input script snippet:

group mobile type 1 2
...
fix nvt mobile nvt temp ...

will move only group mobile and automatically leave all other atoms stationary.

(I am assuming fix nvz is a typo – if this is a customized integration fix, then any code inefficiency in it is not the LAMMPS dev team’s fault. :slight_smile: )

The large amount of time spent on Comm and Modify compared to the very small amount of time spent on Pair is very suspicious. How many of those 80618 atoms are wall atoms?

As others have mentioned before, it would be best if you posted a complete input deck, so it can be independently confirmed using advanced profiling techniques where the time is spent.