static int fmr_op_init_mr(struct rpcrdma_ia *ia, struct rpcrdma_mw *mw) { static struct ib_fmr_attr fmr_attr = { .max_pages = RPCRDMA_MAX_FMR_SGES, .max_maps = 1, .page_shift = PAGE_SHIFT }; mw->fmr.fm_physaddrs = kcalloc(RPCRDMA_MAX_FMR_SGES, sizeof(u64), GFP_KERNEL); if (!mw->fmr.fm_physaddrs) goto out_free; mw->mw_sg = kcalloc(RPCRDMA_MAX_FMR_SGES, sizeof(*mw->mw_sg), GFP_KERNEL); if (!mw->mw_sg) goto out_free; sg_init_table(mw->mw_sg, RPCRDMA_MAX_FMR_SGES); mw->fmr.fm_mr = ib_alloc_fmr(ia->ri_pd, RPCRDMA_FMR_ACCESS_FLAGS, &fmr_attr); if (IS_ERR(mw->fmr.fm_mr)) goto out_fmr_err; return 0; out_fmr_err: dprintk("RPC: %s: ib_alloc_fmr returned %ld\n", __func__, PTR_ERR(mw->fmr.fm_mr)); out_free: kfree(mw->mw_sg); kfree(mw->fmr.fm_physaddrs); return -ENOMEM; } static int __fmr_unmap(struct rpcrdma_mw *mw) { LIST_HEAD(l); int rc; list_add(&mw->fmr.fm_mr->list, &l); rc = ib_unmap_fmr(&l); list_del_init(&mw->fmr.fm_mr->list); return rc; }
void rds_ib_unreg_fmr(struct list_head *list, unsigned int *nfreed, unsigned long *unpinned, unsigned int goal) { struct rds_ib_mr *ibmr, *next; struct rds_ib_fmr *fmr; LIST_HEAD(fmr_list); int ret = 0; unsigned int freed = *nfreed; /* String all ib_mr's onto one list and hand them to ib_unmap_fmr */ list_for_each_entry(ibmr, list, unmap_list) { fmr = &ibmr->u.fmr; list_add(&fmr->fmr->list, &fmr_list); } ret = ib_unmap_fmr(&fmr_list); if (ret) pr_warn("RDS/IB: FMR invalidation failed (err=%d)\n", ret); /* Now we can destroy the DMA mapping and unpin any pages */ list_for_each_entry_safe(ibmr, next, list, unmap_list) { fmr = &ibmr->u.fmr; *unpinned += ibmr->sg_len; __rds_ib_teardown_mr(ibmr); if (freed < goal || ibmr->remap_count >= ibmr->pool->fmr_attr.max_maps) { if (ibmr->pool->pool_type == RDS_IB_MR_8K_POOL) rds_ib_stats_inc(s_ib_rdma_mr_8k_free); else rds_ib_stats_inc(s_ib_rdma_mr_1m_free); list_del(&ibmr->unmap_list);
static int fmr_op_init(struct rpcrdma_xprt *r_xprt) { struct rpcrdma_buffer *buf = &r_xprt->rx_buf; int mr_access_flags = IB_ACCESS_REMOTE_WRITE | IB_ACCESS_REMOTE_READ; struct ib_fmr_attr fmr_attr = { .max_pages = RPCRDMA_MAX_FMR_SGES, .max_maps = 1, .page_shift = PAGE_SHIFT }; struct ib_pd *pd = r_xprt->rx_ia.ri_pd; struct rpcrdma_mw *r; int i, rc; spin_lock_init(&buf->rb_mwlock); INIT_LIST_HEAD(&buf->rb_mws); INIT_LIST_HEAD(&buf->rb_all); i = max_t(int, RPCRDMA_MAX_DATA_SEGS / RPCRDMA_MAX_FMR_SGES, 1); i += 2; /* head + tail */ i *= buf->rb_max_requests; /* one set for each RPC slot */ dprintk("RPC: %s: initalizing %d FMRs\n", __func__, i); rc = -ENOMEM; while (i--) { r = kzalloc(sizeof(*r), GFP_KERNEL); if (!r) goto out; r->r.fmr.physaddrs = kmalloc(RPCRDMA_MAX_FMR_SGES * sizeof(u64), GFP_KERNEL); if (!r->r.fmr.physaddrs) goto out_free; r->r.fmr.fmr = ib_alloc_fmr(pd, mr_access_flags, &fmr_attr); if (IS_ERR(r->r.fmr.fmr)) goto out_fmr_err; list_add(&r->mw_list, &buf->rb_mws); list_add(&r->mw_all, &buf->rb_all); } return 0; out_fmr_err: rc = PTR_ERR(r->r.fmr.fmr); dprintk("RPC: %s: ib_alloc_fmr status %i\n", __func__, rc); kfree(r->r.fmr.physaddrs); out_free: kfree(r); out: return rc; } static int __fmr_unmap(struct rpcrdma_mw *r) { LIST_HEAD(l); list_add(&r->r.fmr.fmr->list, &l); return ib_unmap_fmr(&l); }