Increasing GPU Memory Usage in LAMMPS for M3GNet

Hello Everyone
I am trying to install LAMMPS from Conda-Forge to perform simulations using M3GNet. However, as you may know, the version of LAMMPS available on Conda-Forge doesn’t support parallelization via OpenMPI. Therefore, I would like to utilize the GPU for my simulations. I have already installed the necessary GPU drivers on my system.

Here is the information for the CUDA binary:
nvcc: NVIDIA (R) Cuda compiler driver
Copyright (c) 2005-2023 NVIDIA Corporation
Built on Tue_Aug_15_22:02:13_PDT_2023
Cuda compilation tools, release 12.2, V12.2.140
Build cuda_12.2.r12.2/compiler.33191640_0

The steps I followed for installing LAMMPS from Conda-Forge, as well as M3GNet, MatGl, DGL, and PyTorch, are as follows:

Create and activate a new Conda environment:
conda create -n lammps_m3gnet python=3.11
conda activate lammps_m3gnet

Install LAMMPS from Conda-Forge:
conda install -c conda-forge lammps

Install M3GNet and MatGl:
pip install m3gnet
pip install matgl

Install CUDA Toolkit (v11.8.0):
conda install -c conda-forge cudatoolkit=11.8.0

Install PyTorch compatible with matgl:
pip install torch==2.2.0 torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118

Install DGL:
pip install dgl -f https://data.dgl.ai/wheels/cu118/repo.html
pip install dglgo -f https://data.dgl.ai/wheels-test/repo.html

After completing the installation, when I run a job using the LAMMPS binary (lmp), I get the following output:

PID USER DEV TYPE GPU GPU MEM CPU HOST MEM Command
3923427 ss 0 Compute 0% 386MiB 2% 100% 6742MiB lmp -in input.txt
3928976 ss 0 Compute 0% 386MiB 2% 100% 4725MiB lmp -in input.txt

As you can see, only 2% of the GPU memory is being utilized for each job. My question is: How can I increase the percentage of GPU memory used by each job?

I appreciate your assistance in advance.