Message Passing Interface
Distributed Computing with MPI:
- mpiexec : Run an MPI script.
- MPI_Allreduce : Combines values from all processes and distributes the result back to all processes.
- MPI_Barrier : Blocks until all processes in the communicator have reached this routine.
- MPI_Bcast : Broadcasts a message from the process with rank “root” to all other processes of the communicator
- MPI_Comm_delete : Removes MPI_Comm object.
- MPI_Comm_get_name : Return the print name from the communicator.
- MPI_Comm_object : Creates MPI_Comm object.
- MPI_Comm_rank : Determines the rank of the calling process in the communicator.
- MPI_Comm_size : Determines the size of the group associated with a communicator.
- MPI_Comm_split : Partitions the group that is associated with the specified communicator into a specified number of disjoint subgroups.
- MPI_Comm_used : Returns list of current used MPI_Comm handle.
- MPI examples : Some Nelson MPI examples.
- MPI_Finalize : Terminate the MPI execution environment.
- MPI_Get_library_version : Return the version number of MPI library.
- MPI_Get_processor_name : Gets the name of the processor.
- MPI_Get_version : Return the version number of MPI.
- MPI_Init : Initialize the MPI execution environment.
- MPI_Initialized : Indicates whether MPI_Init has been called.
- MPI_Iprobe : Nonblocking test for a message.
- MPI overview : Access to MPI features from Nelson.
- MPI_Probe : Blocking test for a message.
- MPI_Recv : Blocking receive for a message.
- MPI_Reduce : Reduces values on all processes to a single value.
- MPI_Send : Performs a blocking send.
Example: Sum on distributed computing: