/* * This is the function that iterates through the sparse groups to the dense group * to reach the process pointer */ ompi_proc_t* ompi_group_get_proc_ptr (ompi_group_t* group , int rank) { int ranks1,ranks2; do { if(OMPI_GROUP_IS_DENSE(group)) { return group->grp_proc_pointers[rank]; } ranks1 = rank; ompi_group_translate_ranks( group, 1, &ranks1, group->grp_parent_group_ptr,&ranks2); rank = ranks2; group = group->grp_parent_group_ptr; } while (1); }
ompi_proc_t *ompi_group_get_proc_ptr_raw (ompi_group_t *group, int rank) { #if OMPI_GROUP_SPARSE do { if (OMPI_GROUP_IS_DENSE(group)) { return ompi_group_dense_lookup_raw (group, rank); } int ranks1 = rank; ompi_group_translate_ranks (group, 1, &ranks1, group->grp_parent_group_ptr, &rank); group = group->grp_parent_group_ptr; } while (1); #else return ompi_group_dense_lookup_raw (group, rank); #endif }
int mca_topo_base_cart_create(mca_topo_base_module_t *topo, ompi_communicator_t* old_comm, int ndims, const int *dims, const int *periods, bool reorder, ompi_communicator_t** comm_topo) { int nprocs = 1, i, new_rank, num_procs, ret; ompi_communicator_t *new_comm; ompi_proc_t **topo_procs = NULL; mca_topo_base_comm_cart_2_2_0_t* cart; num_procs = old_comm->c_local_group->grp_proc_count; new_rank = old_comm->c_local_group->grp_my_rank; assert(topo->type == OMPI_COMM_CART); /* Calculate the number of processes in this grid */ for (i = 0; i < ndims; ++i) { if(dims[i] <= 0) { return OMPI_ERROR; } nprocs *= dims[i]; } /* check for the error condition */ if (num_procs < nprocs) { return MPI_ERR_DIMS; } /* check if we have to trim the list of processes */ if (nprocs < num_procs) { num_procs = nprocs; } if (new_rank > (nprocs-1)) { ndims = 0; new_rank = MPI_UNDEFINED; num_procs = 0; } cart = OBJ_NEW(mca_topo_base_comm_cart_2_2_0_t); if( NULL == cart ) { return OMPI_ERR_OUT_OF_RESOURCE; } cart->ndims = ndims; /* MPI-2.1 allows 0-dimension cartesian communicators, so prevent a 0-byte malloc -- leave dims as NULL */ if( ndims > 0 ) { cart->dims = (int*)malloc(sizeof(int) * ndims); if (NULL == cart->dims) { OBJ_RELEASE(cart); return OMPI_ERROR; } memcpy(cart->dims, dims, ndims * sizeof(int)); /* Cartesian communicator; copy the right data to the common information */ cart->periods = (int*)malloc(sizeof(int) * ndims); if (NULL == cart->periods) { OBJ_RELEASE(cart); return OMPI_ERR_OUT_OF_RESOURCE; } memcpy(cart->periods, periods, ndims * sizeof(int)); cart->coords = (int*)malloc(sizeof(int) * ndims); if (NULL == cart->coords) { OBJ_RELEASE(cart); return OMPI_ERR_OUT_OF_RESOURCE; } { /* setup the cartesian topology */ int nprocs = num_procs, rank = new_rank; for (i = 0; i < ndims; ++i) { nprocs /= cart->dims[i]; cart->coords[i] = rank / nprocs; rank %= nprocs; } } } /* JMS: This should really be refactored to use comm_create_group(), because ompi_comm_allocate() still complains about 0-byte mallocs in debug builds for 0-member groups. */ if (num_procs > 0) { /* Copy the proc structure from the previous communicator over to the new one. The topology module is then able to work on this copy and rearrange it as it deems fit. */ topo_procs = (ompi_proc_t**)malloc(num_procs * sizeof(ompi_proc_t *)); if (NULL == topo_procs) { OBJ_RELEASE(cart); return OMPI_ERR_OUT_OF_RESOURCE; } if(OMPI_GROUP_IS_DENSE(old_comm->c_local_group)) { memcpy(topo_procs, old_comm->c_local_group->grp_proc_pointers, num_procs * sizeof(ompi_proc_t *)); } else { for(i = 0 ; i < num_procs; i++) { topo_procs[i] = ompi_group_peer_lookup(old_comm->c_local_group,i); } } } /* allocate a new communicator */ new_comm = ompi_comm_allocate(num_procs, 0); if (NULL == new_comm) { free(topo_procs); OBJ_RELEASE(cart); return MPI_ERR_INTERN; } assert(NULL == new_comm->c_topo); assert(!(new_comm->c_flags & OMPI_COMM_CART)); new_comm->c_topo = topo; new_comm->c_topo->mtc.cart = cart; new_comm->c_topo->reorder = reorder; new_comm->c_flags |= OMPI_COMM_CART; ret = ompi_comm_enable(old_comm, new_comm, new_rank, num_procs, topo_procs); if (OMPI_SUCCESS != ret) { /* something wrong happened during setting the communicator */ free(topo_procs); OBJ_RELEASE(cart); if (MPI_COMM_NULL != new_comm) { new_comm->c_topo = NULL; new_comm->c_flags &= ~OMPI_COMM_CART; ompi_comm_free (&new_comm); } return ret; } *comm_topo = new_comm; if( MPI_UNDEFINED == new_rank ) { ompi_comm_free(&new_comm); *comm_topo = MPI_COMM_NULL; } /* end here */ return OMPI_SUCCESS; }
int mca_topo_base_graph_create(mca_topo_base_module_t *topo, ompi_communicator_t* old_comm, int nnodes, int *index, int *edges, bool reorder, ompi_communicator_t** comm_topo) { ompi_communicator_t *new_comm; int new_rank, num_procs, ret, i; ompi_proc_t **topo_procs = NULL; mca_topo_base_comm_graph_2_2_0_t* graph; num_procs = old_comm->c_local_group->grp_proc_count; new_rank = old_comm->c_local_group->grp_my_rank; assert(topo->type == OMPI_COMM_GRAPH); if( num_procs < nnodes ) { return MPI_ERR_DIMS; } if( num_procs > nnodes ) { num_procs = nnodes; } if( new_rank > (nnodes - 1) ) { new_rank = MPI_UNDEFINED; num_procs = 0; nnodes = 0; } graph = OBJ_NEW(mca_topo_base_comm_graph_2_2_0_t); if( NULL == graph ) { return OMPI_ERR_OUT_OF_RESOURCE; } graph->nnodes = nnodes; /* Don't do any of the other initialization if we're not supposed to be part of the new communicator (because nnodes has been reset to 0, making things like index[nnodes-1] be junk). JMS: This should really be refactored to use comm_create_group(), because ompi_comm_allocate() still complains about 0-byte mallocs in debug builds for 0-member groups. */ if (MPI_UNDEFINED != new_rank) { graph->index = (int*)malloc(sizeof(int) * nnodes); if (NULL == graph->index) { OBJ_RELEASE(graph); return OMPI_ERR_OUT_OF_RESOURCE; } memcpy(graph->index, index, nnodes * sizeof(int)); /* Graph communicator; copy the right data to the common information */ graph->edges = (int*)malloc(sizeof(int) * index[nnodes-1]); if (NULL == graph->edges) { OBJ_RELEASE(graph); return OMPI_ERR_OUT_OF_RESOURCE; } memcpy(graph->edges, edges, index[nnodes-1] * sizeof(int)); topo_procs = (ompi_proc_t**)malloc(num_procs * sizeof(ompi_proc_t *)); if (NULL == topo_procs) { OBJ_RELEASE(graph); return OMPI_ERR_OUT_OF_RESOURCE; } if(OMPI_GROUP_IS_DENSE(old_comm->c_local_group)) { memcpy(topo_procs, old_comm->c_local_group->grp_proc_pointers, num_procs * sizeof(ompi_proc_t *)); } else { for(i = 0 ; i < num_procs; i++) { topo_procs[i] = ompi_group_peer_lookup(old_comm->c_local_group,i); } } } /* allocate a new communicator */ new_comm = ompi_comm_allocate(nnodes, 0); if (NULL == new_comm) { free(topo_procs); OBJ_RELEASE(graph); return OMPI_ERR_OUT_OF_RESOURCE; } ret = ompi_comm_enable(old_comm, new_comm, new_rank, num_procs, topo_procs); if (OMPI_SUCCESS != ret) { free(topo_procs); OBJ_RELEASE(graph); ompi_comm_free (&new_comm); return ret; } new_comm->c_topo = topo; new_comm->c_topo->mtc.graph = graph; new_comm->c_flags |= OMPI_COMM_GRAPH; new_comm->c_topo->reorder = reorder; *comm_topo = new_comm; if( MPI_UNDEFINED == new_rank ) { ompi_comm_free(&new_comm); *comm_topo = MPI_COMM_NULL; } return OMPI_SUCCESS; }