int MPIDI_Win_free(MPID_Win **win_ptr) { int mpi_errno=MPI_SUCCESS, total_pt_rma_puts_accs; int in_use; MPID_Comm *comm_ptr; int errflag = FALSE; MPIDI_STATE_DECL(MPID_STATE_MPIDI_WIN_FREE); MPIDI_RMA_FUNC_ENTER(MPID_STATE_MPIDI_WIN_FREE); comm_ptr = (*win_ptr)->comm_ptr; MPIU_INSTR_DURATION_START(winfree_rs); mpi_errno = MPIR_Reduce_scatter_block_impl((*win_ptr)->pt_rma_puts_accs, &total_pt_rma_puts_accs, 1, MPI_INT, MPI_SUM, comm_ptr, &errflag); if (mpi_errno) { MPIU_ERR_POP(mpi_errno); } MPIU_ERR_CHKANDJUMP(errflag, mpi_errno, MPI_ERR_OTHER, "**coll_fail"); MPIU_INSTR_DURATION_END(winfree_rs); if (total_pt_rma_puts_accs != (*win_ptr)->my_pt_rma_puts_accs) { MPID_Progress_state progress_state; /* poke the progress engine until the two are equal */ MPIU_INSTR_DURATION_START(winfree_complete); MPID_Progress_start(&progress_state); while (total_pt_rma_puts_accs != (*win_ptr)->my_pt_rma_puts_accs) { mpi_errno = MPID_Progress_wait(&progress_state); /* --BEGIN ERROR HANDLING-- */ if (mpi_errno != MPI_SUCCESS) { MPID_Progress_end(&progress_state); MPIU_ERR_SETANDJUMP(mpi_errno,MPI_ERR_OTHER,"**winnoprogress"); } /* --END ERROR HANDLING-- */ } MPID_Progress_end(&progress_state); MPIU_INSTR_DURATION_END(winfree_complete); } mpi_errno = MPIR_Comm_free_impl(comm_ptr); if (mpi_errno) MPIU_ERR_POP(mpi_errno); MPIU_Free((*win_ptr)->base_addrs); MPIU_Free((*win_ptr)->disp_units); MPIU_Free((*win_ptr)->all_win_handles); MPIU_Free((*win_ptr)->pt_rma_puts_accs); MPIU_Object_release_ref(*win_ptr, &in_use); /* MPI windows don't have reference count semantics, so this should always be true */ MPIU_Assert(!in_use); MPIU_Handle_obj_free( &MPID_Win_mem, *win_ptr ); fn_exit: MPIDI_RMA_FUNC_EXIT(MPID_STATE_MPIDI_WIN_FREE); return mpi_errno; fn_fail: goto fn_exit; }
/*@ MPI_Dist_graph_create - MPI_DIST_GRAPH_CREATE returns a handle to a new communicator to which the distributed graph topology information is attached. Input Parameters: + comm_old - input communicator (handle) . n - number of source nodes for which this process specifies edges (non-negative integer) . sources - array containing the n source nodes for which this process specifies edges (array of non-negative integers) . degrees - array specifying the number of destinations for each source node in the source node array (array of non-negative integers) . destinations - destination nodes for the source nodes in the source node array (array of non-negative integers) . weights - weights for source to destination edges (array of non-negative integers or MPI_UNWEIGHTED) . info - hints on optimization and interpretation of weights (handle) - reorder - the process may be reordered (true) or not (false) (logical) Output Parameters: . comm_dist_graph - communicator with distributed graph topology added (handle) .N ThreadSafe .N Fortran .N Errors .N MPI_SUCCESS .N MPI_ERR_ARG .N MPI_ERR_OTHER @*/ int MPI_Dist_graph_create(MPI_Comm comm_old, int n, const int sources[], const int degrees[], const int destinations[], const int weights[], MPI_Info info, int reorder, MPI_Comm *comm_dist_graph) { int mpi_errno = MPI_SUCCESS; MPID_Comm *comm_ptr = NULL; MPID_Comm *comm_dist_graph_ptr = NULL; MPI_Request *reqs = NULL; MPIR_Topology *topo_ptr = NULL; MPIR_Dist_graph_topology *dist_graph_ptr = NULL; int i; int j; int idx; int comm_size = 0; int in_capacity; int out_capacity; int **rout = NULL; int **rin = NULL; int *rin_sizes; int *rout_sizes; int *rin_idx; int *rout_idx; int *rs; int in_out_peers[2] = {-1, -1}; int errflag = FALSE; MPIU_CHKLMEM_DECL(9); MPIU_CHKPMEM_DECL(1); MPID_MPI_STATE_DECL(MPID_STATE_MPI_DIST_GRAPH_CREATE); MPIR_ERRTEST_INITIALIZED_ORDIE(); MPIU_THREAD_CS_ENTER(ALLFUNC,); MPID_MPI_FUNC_ENTER(MPID_STATE_MPI_DIST_GRAPH_CREATE); /* Validate parameters, especially handles needing to be converted */ # ifdef HAVE_ERROR_CHECKING { MPID_BEGIN_ERROR_CHECKS; { MPIR_ERRTEST_COMM(comm_old, mpi_errno); MPIR_ERRTEST_INFO_OR_NULL(info, mpi_errno); if (mpi_errno != MPI_SUCCESS) goto fn_fail; } MPID_END_ERROR_CHECKS; } # endif /* Convert MPI object handles to object pointers */ MPID_Comm_get_ptr(comm_old, comm_ptr); /* Validate parameters and objects (post conversion) */ # ifdef HAVE_ERROR_CHECKING { MPID_BEGIN_ERROR_CHECKS; { /* Validate comm_ptr */ MPID_Comm_valid_ptr(comm_ptr, mpi_errno); /* If comm_ptr is not valid, it will be reset to null */ if (comm_ptr) { MPIR_ERRTEST_COMM_INTRA(comm_ptr, mpi_errno); } MPIR_ERRTEST_ARGNEG(n, "n", mpi_errno); if (n > 0) { int have_degrees = 0; MPIR_ERRTEST_ARGNULL(sources, "sources", mpi_errno); MPIR_ERRTEST_ARGNULL(degrees, "degrees", mpi_errno); for (i = 0; i < n; ++i) { if (degrees[i]) { have_degrees = 1; break; } } if (have_degrees) { MPIR_ERRTEST_ARGNULL(destinations, "destinations", mpi_errno); if (weights != MPI_UNWEIGHTED) MPIR_ERRTEST_ARGNULL(weights, "weights", mpi_errno); } } if (mpi_errno != MPI_SUCCESS) goto fn_fail; } MPID_END_ERROR_CHECKS; } # endif /* HAVE_ERROR_CHECKING */ /* ... body of routine ... */ /* Implementation based on Torsten Hoefler's reference implementation * attached to MPI-2.2 ticket #33. */ *comm_dist_graph = MPI_COMM_NULL; comm_size = comm_ptr->local_size; /* following the spirit of the old topo interface, attributes do not * propagate to the new communicator (see MPI-2.1 pp. 243 line 11) */ mpi_errno = MPIR_Comm_copy(comm_ptr, comm_size, &comm_dist_graph_ptr); if (mpi_errno) MPIU_ERR_POP(mpi_errno); MPIU_Assert(comm_dist_graph_ptr != NULL); /* rin is an array of size comm_size containing pointers to arrays of * rin_sizes[x]. rin[x] is locally known number of edges into this process * from rank x. * * rout is an array of comm_size containing pointers to arrays of * rout_sizes[x]. rout[x] is the locally known number of edges out of this * process to rank x. */ MPIU_CHKLMEM_MALLOC(rout, int **, comm_size*sizeof(int*), mpi_errno, "rout"); MPIU_CHKLMEM_MALLOC(rin, int **, comm_size*sizeof(int*), mpi_errno, "rin"); MPIU_CHKLMEM_MALLOC(rin_sizes, int *, comm_size*sizeof(int), mpi_errno, "rin_sizes"); MPIU_CHKLMEM_MALLOC(rout_sizes, int *, comm_size*sizeof(int), mpi_errno, "rout_sizes"); MPIU_CHKLMEM_MALLOC(rin_idx, int *, comm_size*sizeof(int), mpi_errno, "rin_idx"); MPIU_CHKLMEM_MALLOC(rout_idx, int *, comm_size*sizeof(int), mpi_errno, "rout_idx"); memset(rout, 0, comm_size*sizeof(int*)); memset(rin, 0, comm_size*sizeof(int*)); memset(rin_sizes, 0, comm_size*sizeof(int)); memset(rout_sizes, 0, comm_size*sizeof(int)); memset(rin_idx, 0, comm_size*sizeof(int)); memset(rout_idx, 0, comm_size*sizeof(int)); /* compute array sizes */ idx = 0; for (i = 0; i < n; ++i) { MPIU_Assert(sources[i] < comm_size); for (j = 0; j < degrees[i]; ++j) { MPIU_Assert(destinations[idx] < comm_size); /* rout_sizes[i] is twice as long as the number of edges to be * sent to rank i by this process */ rout_sizes[sources[i]] += 2; rin_sizes[destinations[idx]] += 2; ++idx; } } /* allocate arrays */ for (i = 0; i < comm_size; ++i) { /* can't use CHKLMEM macros b/c we are in a loop */ if (rin_sizes[i]) { rin[i] = MPIU_Malloc(rin_sizes[i] * sizeof(int)); } if (rout_sizes[i]) { rout[i] = MPIU_Malloc(rout_sizes[i] * sizeof(int)); } } /* populate arrays */ idx = 0; for (i = 0; i < n; ++i) { /* TODO add this assert as proper error checking above */ int s_rank = sources[i]; MPIU_Assert(s_rank < comm_size); MPIU_Assert(s_rank >= 0); for (j = 0; j < degrees[i]; ++j) { int d_rank = destinations[idx]; int weight = (weights == MPI_UNWEIGHTED ? 0 : weights[idx]); /* TODO add this assert as proper error checking above */ MPIU_Assert(d_rank < comm_size); MPIU_Assert(d_rank >= 0); /* XXX DJG what about self-edges? do we need to drop one of these * cases when there is a self-edge to avoid double-counting? */ /* rout[s][2*x] is the value of d for the j'th edge between (s,d) * with weight rout[s][2*x+1], where x is the current end of the * outgoing edge list for s. x==(rout_idx[s]/2) */ rout[s_rank][rout_idx[s_rank]++] = d_rank; rout[s_rank][rout_idx[s_rank]++] = weight; /* rin[d][2*x] is the value of s for the j'th edge between (s,d) * with weight rout[d][2*x+1], where x is the current end of the * incoming edge list for d. x==(rin_idx[d]/2) */ rin[d_rank][rin_idx[d_rank]++] = s_rank; rin[d_rank][rin_idx[d_rank]++] = weight; ++idx; } } for (i = 0; i < comm_size; ++i) { /* sanity check that all arrays are fully populated*/ MPIU_Assert(rin_idx[i] == rin_sizes[i]); MPIU_Assert(rout_idx[i] == rout_sizes[i]); } MPIU_CHKLMEM_MALLOC(rs, int *, 2*comm_size*sizeof(int), mpi_errno, "red-scat source buffer"); for (i = 0; i < comm_size; ++i) { rs[2*i] = (rin_sizes[i] ? 1 : 0); rs[2*i+1] = (rout_sizes[i] ? 1 : 0); } /* compute the number of peers I will recv from */ mpi_errno = MPIR_Reduce_scatter_block_impl(rs, in_out_peers, 2, MPI_INT, MPI_SUM, comm_ptr, &errflag); if (mpi_errno) MPIU_ERR_POP(mpi_errno); MPIU_ERR_CHKANDJUMP(errflag, mpi_errno, MPI_ERR_OTHER, "**coll_fail"); MPIU_Assert(in_out_peers[0] <= comm_size && in_out_peers[0] >= 0); MPIU_Assert(in_out_peers[1] <= comm_size && in_out_peers[1] >= 0); idx = 0; /* must be 2*comm_size requests because we will possibly send inbound and * outbound edges to everyone in our communicator */ MPIU_CHKLMEM_MALLOC(reqs, MPI_Request *, 2*comm_size*sizeof(MPI_Request), mpi_errno, "temp request array"); for (i = 0; i < comm_size; ++i) { if (rin_sizes[i]) { /* send edges where i is a destination to process i */ mpi_errno = MPIC_Isend(&rin[i][0], rin_sizes[i], MPI_INT, i, MPIR_TOPO_A_TAG, comm_old, &reqs[idx++]); if (mpi_errno) MPIU_ERR_POP(mpi_errno); } if (rout_sizes[i]) { /* send edges where i is a source to process i */ mpi_errno = MPIC_Isend(&rout[i][0], rout_sizes[i], MPI_INT, i, MPIR_TOPO_B_TAG, comm_old, &reqs[idx++]); if (mpi_errno) MPIU_ERR_POP(mpi_errno); } } MPIU_Assert(idx <= (2 * comm_size)); /* Create the topology structure */ MPIU_CHKPMEM_MALLOC(topo_ptr, MPIR_Topology *, sizeof(MPIR_Topology), mpi_errno, "topo_ptr"); topo_ptr->kind = MPI_DIST_GRAPH; dist_graph_ptr = &topo_ptr->topo.dist_graph; dist_graph_ptr->indegree = 0; dist_graph_ptr->in = NULL; dist_graph_ptr->in_weights = NULL; dist_graph_ptr->outdegree = 0; dist_graph_ptr->out = NULL; dist_graph_ptr->out_weights = NULL; dist_graph_ptr->is_weighted = (weights != MPI_UNWEIGHTED); /* can't use CHKPMEM macros for this b/c we need to realloc */ in_capacity = 10; /* arbitrary */ dist_graph_ptr->in = MPIU_Malloc(in_capacity*sizeof(int)); if (dist_graph_ptr->is_weighted) dist_graph_ptr->in_weights = MPIU_Malloc(in_capacity*sizeof(int)); out_capacity = 10; /* arbitrary */ dist_graph_ptr->out = MPIU_Malloc(out_capacity*sizeof(int)); if (dist_graph_ptr->is_weighted) dist_graph_ptr->out_weights = MPIU_Malloc(out_capacity*sizeof(int)); for (i = 0; i < in_out_peers[0]; ++i) { MPI_Status status; int count; int *buf; /* receive inbound edges */ mpi_errno = MPIC_Probe(MPI_ANY_SOURCE, MPIR_TOPO_A_TAG, comm_old, &status); if (mpi_errno) MPIU_ERR_POP(mpi_errno); MPIR_Get_count_impl(&status, MPI_INT, &count); /* can't use CHKLMEM macros b/c we are in a loop */ buf = MPIU_Malloc(count*sizeof(int)); MPIU_ERR_CHKANDJUMP(!buf, mpi_errno, MPIR_ERR_RECOVERABLE, "**nomem"); mpi_errno = MPIC_Recv(buf, count, MPI_INT, MPI_ANY_SOURCE, MPIR_TOPO_A_TAG, comm_old, MPI_STATUS_IGNORE); if (mpi_errno) MPIU_ERR_POP(mpi_errno); for (j = 0; j < count/2; ++j) { int deg = dist_graph_ptr->indegree++; if (deg >= in_capacity) { in_capacity *= 2; MPIU_REALLOC_ORJUMP(dist_graph_ptr->in, in_capacity*sizeof(int), mpi_errno); if (dist_graph_ptr->is_weighted) MPIU_REALLOC_ORJUMP(dist_graph_ptr->in_weights, in_capacity*sizeof(int), mpi_errno); } dist_graph_ptr->in[deg] = buf[2*j]; if (dist_graph_ptr->is_weighted) dist_graph_ptr->in_weights[deg] = buf[2*j+1]; } MPIU_Free(buf); } for (i = 0; i < in_out_peers[1]; ++i) { MPI_Status status; int count; int *buf; /* receive outbound edges */ mpi_errno = MPIC_Probe(MPI_ANY_SOURCE, MPIR_TOPO_B_TAG, comm_old, &status); if (mpi_errno) MPIU_ERR_POP(mpi_errno); MPIR_Get_count_impl(&status, MPI_INT, &count); /* can't use CHKLMEM macros b/c we are in a loop */ buf = MPIU_Malloc(count*sizeof(int)); MPIU_ERR_CHKANDJUMP(!buf, mpi_errno, MPIR_ERR_RECOVERABLE, "**nomem"); mpi_errno = MPIC_Recv(buf, count, MPI_INT, MPI_ANY_SOURCE, MPIR_TOPO_B_TAG, comm_old, MPI_STATUS_IGNORE); if (mpi_errno) MPIU_ERR_POP(mpi_errno); for (j = 0; j < count/2; ++j) { int deg = dist_graph_ptr->outdegree++; if (deg >= out_capacity) { out_capacity *= 2; MPIU_REALLOC_ORJUMP(dist_graph_ptr->out, out_capacity*sizeof(int), mpi_errno); if (dist_graph_ptr->is_weighted) MPIU_REALLOC_ORJUMP(dist_graph_ptr->out_weights, out_capacity*sizeof(int), mpi_errno); } dist_graph_ptr->out[deg] = buf[2*j]; if (dist_graph_ptr->is_weighted) dist_graph_ptr->out_weights[deg] = buf[2*j+1]; } MPIU_Free(buf); } mpi_errno = MPIR_Waitall_impl(idx, reqs, MPI_STATUSES_IGNORE); if (mpi_errno) MPIU_ERR_POP(mpi_errno); /* remove any excess memory allocation */ MPIU_REALLOC_ORJUMP(dist_graph_ptr->in, dist_graph_ptr->indegree*sizeof(int), mpi_errno); MPIU_REALLOC_ORJUMP(dist_graph_ptr->out, dist_graph_ptr->outdegree*sizeof(int), mpi_errno); if (dist_graph_ptr->is_weighted) { MPIU_REALLOC_ORJUMP(dist_graph_ptr->in_weights, dist_graph_ptr->indegree*sizeof(int), mpi_errno); MPIU_REALLOC_ORJUMP(dist_graph_ptr->out_weights, dist_graph_ptr->outdegree*sizeof(int), mpi_errno); } mpi_errno = MPIR_Topology_put(comm_dist_graph_ptr, topo_ptr); if (mpi_errno) MPIU_ERR_POP(mpi_errno); MPIU_CHKPMEM_COMMIT(); MPIU_OBJ_PUBLISH_HANDLE(*comm_dist_graph, comm_dist_graph_ptr->handle); /* ... end of body of routine ... */ fn_exit: for (i = 0; i < comm_size; ++i) { if (rin[i]) MPIU_Free(rin[i]); if (rout[i]) MPIU_Free(rout[i]); } MPIU_CHKLMEM_FREEALL(); MPID_MPI_FUNC_EXIT(MPID_STATE_MPI_DIST_GRAPH_CREATE); MPIU_THREAD_CS_EXIT(ALLFUNC,); return mpi_errno; /* --BEGIN ERROR HANDLING-- */ fn_fail: if (dist_graph_ptr && dist_graph_ptr->in) MPIU_Free(dist_graph_ptr->in); if (dist_graph_ptr && dist_graph_ptr->in_weights) MPIU_Free(dist_graph_ptr->in_weights); if (dist_graph_ptr && dist_graph_ptr->out) MPIU_Free(dist_graph_ptr->out); if (dist_graph_ptr && dist_graph_ptr->out_weights) MPIU_Free(dist_graph_ptr->out_weights); MPIU_CHKPMEM_REAP(); #ifdef HAVE_ERROR_CHECKING mpi_errno = MPIR_Err_create_code( mpi_errno, MPIR_ERR_RECOVERABLE, FCNAME, __LINE__, MPI_ERR_OTHER, "**mpi_dist_graph_create", "**mpi_dist_graph_create %C %d %p %p %p %p %I %d %p", comm_old, n, sources, degrees, destinations, weights, info, reorder, comm_dist_graph); #endif mpi_errno = MPIR_Err_return_comm(comm_ptr, FCNAME, mpi_errno); goto fn_exit; /* --END ERROR HANDLING-- */ }