/** * Pack data * * @param btl (IN) BTL module */ static struct mca_btl_base_descriptor_t *vader_prepare_src (struct mca_btl_base_module_t *btl, struct mca_btl_base_endpoint_t *endpoint, mca_mpool_base_registration_t *registration, struct opal_convertor_t *convertor, uint8_t order, size_t reserve, size_t *size, uint32_t flags) { const size_t total_size = reserve + *size; struct iovec iov; mca_btl_vader_frag_t *frag; uint32_t iov_count = 1; void *data_ptr, *fbox_ptr; int rc; opal_convertor_get_current_pointer (convertor, &data_ptr); if (OPAL_LIKELY(reserve)) { /* in place send fragment */ if (OPAL_UNLIKELY(opal_convertor_need_buffers(convertor))) { /* non-contiguous data requires the convertor */ (void) MCA_BTL_VADER_FRAG_ALLOC_EAGER(frag); if (OPAL_UNLIKELY(NULL == frag)) { return NULL; } iov.iov_len = *size; iov.iov_base = (IOVBASE_TYPE *)(((uintptr_t)(frag->segments[0].seg_addr.pval)) + reserve); rc = opal_convertor_pack (convertor, &iov, &iov_count, size); if (OPAL_UNLIKELY(rc < 0)) { MCA_BTL_VADER_FRAG_RETURN(frag); return NULL; } frag->segments[0].seg_len = total_size; } else { (void) MCA_BTL_VADER_FRAG_ALLOC_USER(frag); if (OPAL_UNLIKELY(NULL == frag)) { return NULL; } if (total_size > (size_t) mca_btl_vader_component.max_inline_send) { /* single copy send */ frag->hdr->flags = MCA_BTL_VADER_FLAG_SINGLE_COPY; /* set up single copy io vector */ frag->hdr->sc_iov.iov_base = data_ptr; frag->hdr->sc_iov.iov_len = *size; frag->segments[0].seg_len = reserve; frag->segments[1].seg_len = *size; frag->segments[1].seg_addr.pval = data_ptr; frag->base.des_src_cnt = 2; } else { /* inline send */ /* try to reserve a fast box for this transfer */ fbox_ptr = mca_btl_vader_reserve_fbox (endpoint, total_size); if (fbox_ptr) { frag->hdr->flags |= MCA_BTL_VADER_FLAG_FBOX; frag->segments[0].seg_addr.pval = fbox_ptr; } /* NTH: the covertor adds some latency so we bypass it here */ vader_memmove ((void *)((uintptr_t)frag->segments[0].seg_addr.pval + reserve), data_ptr, *size); frag->segments[0].seg_len = total_size; } } } else { /* put/get fragment */ (void) MCA_BTL_VADER_FRAG_ALLOC_USER(frag); if (OPAL_UNLIKELY(NULL == frag)) { return NULL; } frag->segments[0].seg_addr.lval = (uint64_t)(uintptr_t) data_ptr; frag->segments[0].seg_len = total_size; } frag->base.order = order; frag->base.des_flags = flags; frag->endpoint = endpoint; return &frag->base; }
/** * Pack data * * @param btl (IN) BTL module */ static struct mca_btl_base_descriptor_t *vader_prepare_src (struct mca_btl_base_module_t *btl, struct mca_btl_base_endpoint_t *endpoint, struct opal_convertor_t *convertor, uint8_t order, size_t reserve, size_t *size, uint32_t flags) { const size_t total_size = reserve + *size; mca_btl_vader_frag_t *frag; unsigned char *fbox; void *data_ptr; int rc; opal_convertor_get_current_pointer (convertor, &data_ptr); /* in place send fragment */ if (OPAL_UNLIKELY(opal_convertor_need_buffers(convertor))) { uint32_t iov_count = 1; struct iovec iov; /* non-contiguous data requires the convertor */ if (MCA_BTL_VADER_XPMEM != mca_btl_vader_component.single_copy_mechanism && total_size > mca_btl_vader.super.btl_eager_limit) { (void) MCA_BTL_VADER_FRAG_ALLOC_MAX(frag, endpoint); } else (void) MCA_BTL_VADER_FRAG_ALLOC_EAGER(frag, endpoint); if (OPAL_UNLIKELY(NULL == frag)) { return NULL; } iov.iov_len = *size; iov.iov_base = (IOVBASE_TYPE *)(((uintptr_t)(frag->segments[0].seg_addr.pval)) + reserve); rc = opal_convertor_pack (convertor, &iov, &iov_count, size); if (OPAL_UNLIKELY(rc < 0)) { MCA_BTL_VADER_FRAG_RETURN(frag); return NULL; } frag->segments[0].seg_len = *size + reserve; } else { if (MCA_BTL_VADER_XPMEM != mca_btl_vader_component.single_copy_mechanism) { if (OPAL_LIKELY(total_size <= mca_btl_vader.super.btl_eager_limit)) { (void) MCA_BTL_VADER_FRAG_ALLOC_EAGER(frag, endpoint); } else { (void) MCA_BTL_VADER_FRAG_ALLOC_MAX(frag, endpoint); } } else (void) MCA_BTL_VADER_FRAG_ALLOC_USER(frag, endpoint); if (OPAL_UNLIKELY(NULL == frag)) { return NULL; } #if OPAL_BTL_VADER_HAVE_XPMEM /* use xpmem to send this segment if it is above the max inline send size */ if (OPAL_UNLIKELY(MCA_BTL_VADER_XPMEM == mca_btl_vader_component.single_copy_mechanism && total_size > (size_t) mca_btl_vader_component.max_inline_send)) { /* single copy send */ frag->hdr->flags = MCA_BTL_VADER_FLAG_SINGLE_COPY; /* set up single copy io vector */ frag->hdr->sc_iov.iov_base = data_ptr; frag->hdr->sc_iov.iov_len = *size; frag->segments[0].seg_len = reserve; frag->segments[1].seg_len = *size; frag->segments[1].seg_addr.pval = data_ptr; frag->base.des_segment_count = 2; } else { #endif /* inline send */ if (OPAL_LIKELY(MCA_BTL_DES_FLAGS_BTL_OWNERSHIP & flags)) { /* try to reserve a fast box for this transfer only if the * fragment does not belong to the caller */ fbox = mca_btl_vader_reserve_fbox (endpoint, total_size); if (OPAL_LIKELY(fbox)) { frag->segments[0].seg_addr.pval = fbox; } frag->fbox = fbox; } /* NTH: the covertor adds some latency so we bypass it here */ memcpy ((void *)((uintptr_t)frag->segments[0].seg_addr.pval + reserve), data_ptr, *size); frag->segments[0].seg_len = total_size; #if OPAL_BTL_VADER_HAVE_XPMEM } #endif } frag->base.order = order; frag->base.des_flags = flags; return &frag->base; }