<< Nelson 1.10.0.0
Message Passing Interface
MPI_Allreduce >>
Message Passing Interface
In the world of parallel computing the Message Passing Interface (MPI) is the de facto standard for implementing programs on multiple processors.
MPI_Allreduce
—
Combines values from all processes and distributes the result back to all processes.
MPI_Barrier
—
Blocks until all processes in the communicator have reached this routine.
MPI_Bcast
—
Broadcasts a message from the process with rank "root" to all other processes of the communicator
MPI_Comm_delete
—
Removes MPI_Comm object.
MPI_Comm_get_name
—
Return the print name from the communicator.
MPI_Comm_object
—
Creates MPI_Comm object.
MPI_Comm_rank
—
Determines the rank of the calling process in the communicator.
MPI_Comm_size
—
Determines the size of the group associated with a communicator.
MPI_Comm_split
—
Partitions the group that is associated with the specified communicator into a specified number of disjoint subgroups.
MPI_Comm_used
—
Returns list of current used MPI_Comm handle.
MPI_Finalize
—
Terminate the MPI execution environment.
MPI_Get_library_version
—
Return the version number of MPI library.
MPI_Get_processor_name
—
Gets the name of the processor.
MPI_Get_version
—
Return the version number of MPI.
MPI_Init
—
Initialize the MPI execution environment.
MPI_Initialized
—
Indicates whether MPI_Init has been called.
MPI_Iprobe
—
Nonblocking test for a message.
MPI_Probe
—
Blocking test for a message.
MPI_Recv
—
Blocking receive for a message.
MPI_Reduce
—
Reduces values on all processes to a single value.
MPI_Send
—
Performs a blocking send.
MPI examples
—
Some Nelson MPI examples.
MPI overview
—
Access to MPI features from Nelson.
mpiexec
—
Run an MPI script.
<< Nelson 1.10.0.0
Message Passing Interface
MPI_Allreduce >>