-
Notifications
You must be signed in to change notification settings - Fork 32
Open
Description
Hi,
I was running the example scripts for HOOMD-Blue umbrella sampling and I am getting the following error.
******ubuntu-20.04.4.sif I have no name!@cs001 HOOMD]$ mpiexec --mca btl self -np 12 /path-to-SSAGES/ssages multiwalker_umbrella_input.json
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
Installing SSAGES HOOMD-blue hook.
HOOMD-blue is running on the CPU
HOOMD-blue is running on the CPU
Installing SSAGES HOOMD-blue hook.
HOOMD-blue v2.9.3 SINGLE MPI SSE SSE2
Compiled: 09/27/22
Copyright (c) 2009-2019 The Regents of the University of Michigan.
You are using HOOMD-blue. Please cite the following:
- J A Anderson, J Glaser, and S C Glotzer. "HOOMD-blue: A Python package for
high-performance molecular dynamics and hard particle Monte Carlo
simulations", Computational Materials Science 173 (2020) 109363
Ranks 0-11: HOOMD-blue is running on the CPU
notice(2): -- Neighborlist exclusion statistics -- :
notice(2): Particles with 7 exclusions : 6
notice(2): Particles with 10 exclusions : 6
notice(2): Particles with 13 exclusions : 2
notice(2): Neighbors included by diameter : no
notice(2): Neighbors excluded when in the same body: no
MPI_ABORT was invoked on rank 6 in communicator MPI_COMM_WORLD
with errorcode -1.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.******
Can I please know how I can resolve this? Please let me know if I need to provide anything else regarding this.
Thanking you,
Sanjib
Reactions are currently unavailable
Metadata
Metadata
Assignees
Labels
No labels