void rxe_av_fill_ip_info(struct rxe_av *av, struct rdma_ah_attr *attr, struct ib_gid_attr *sgid_attr, union ib_gid *sgid) { rdma_gid2ip(&av->sgid_addr._sockaddr, sgid); rdma_gid2ip(&av->dgid_addr._sockaddr, &rdma_ah_read_grh(attr)->dgid); av->network_type = ib_gid_to_network_type(sgid_attr->gid_type, sgid); }
/* called by the modify qp verb */ int rxe_qp_from_attr(struct rxe_qp *qp, struct ib_qp_attr *attr, int mask, struct ib_udata *udata) { int err; struct rxe_dev *rxe = to_rdev(qp->ibqp.device); union ib_gid sgid; struct ib_gid_attr sgid_attr; /* TODO should handle error by leaving old resources intact */ if (mask & IB_QP_MAX_QP_RD_ATOMIC) { int max_rd_atomic = __roundup_pow_of_two(attr->max_rd_atomic); free_rd_atomic_resources(qp); err = alloc_rd_atomic_resources(qp, max_rd_atomic); if (err) return err; qp->attr.max_rd_atomic = max_rd_atomic; atomic_set(&qp->req.rd_atomic, max_rd_atomic); } if (mask & IB_QP_CUR_STATE) qp->attr.cur_qp_state = attr->qp_state; if (mask & IB_QP_EN_SQD_ASYNC_NOTIFY) qp->attr.en_sqd_async_notify = attr->en_sqd_async_notify; if (mask & IB_QP_ACCESS_FLAGS) qp->attr.qp_access_flags = attr->qp_access_flags; if (mask & IB_QP_PKEY_INDEX) qp->attr.pkey_index = attr->pkey_index; if (mask & IB_QP_PORT) qp->attr.port_num = attr->port_num; if (mask & IB_QP_QKEY) qp->attr.qkey = attr->qkey; if (mask & IB_QP_AV) { rcu_read_lock(); ib_get_cached_gid(&rxe->ib_dev, 1, attr->ah_attr.grh.sgid_index, &sgid, &sgid_attr); rcu_read_unlock(); rxe_av_from_attr(rxe, attr->port_num, &qp->pri_av, &attr->ah_attr); qp->pri_av.network_type = ib_gid_to_network_type(sgid_attr.gid_type, &sgid); rxe_av_fill_ip_info(rxe, &qp->pri_av, &attr->ah_attr, &sgid); } if (mask & IB_QP_ALT_PATH) { rcu_read_lock(); ib_get_cached_gid(&rxe->ib_dev, 1, attr->alt_ah_attr.grh.sgid_index, &sgid, &sgid_attr); rcu_read_unlock(); rxe_av_from_attr(rxe, attr->alt_port_num, &qp->alt_av, &attr->alt_ah_attr); qp->alt_av.network_type = ib_gid_to_network_type(sgid_attr.gid_type, &sgid); rxe_av_fill_ip_info(rxe, &qp->alt_av, &attr->alt_ah_attr, &sgid); qp->attr.alt_port_num = attr->alt_port_num; qp->attr.alt_pkey_index = attr->alt_pkey_index; qp->attr.alt_timeout = attr->alt_timeout; } if (mask & IB_QP_PATH_MTU) { qp->attr.path_mtu = attr->path_mtu; qp->mtu = rxe_mtu_enum_to_int((enum rxe_mtu)attr->path_mtu); } if (mask & IB_QP_TIMEOUT) { qp->attr.timeout = attr->timeout; if (attr->timeout == 0) { qp->qp_timeout_jiffies = 0; } else { int j = usecs_to_jiffies(4ULL << attr->timeout); qp->qp_timeout_jiffies = j ? j : 1; } } if (mask & IB_QP_RETRY_CNT) { qp->attr.retry_cnt = attr->retry_cnt; qp->comp.retry_cnt = attr->retry_cnt; pr_debug("set retry count = %d\n", attr->retry_cnt); } if (mask & IB_QP_RNR_RETRY) { qp->attr.rnr_retry = attr->rnr_retry; qp->comp.rnr_retry = attr->rnr_retry; pr_debug("set rnr retry count = %d\n", attr->rnr_retry); } if (mask & IB_QP_RQ_PSN) { qp->attr.rq_psn = (attr->rq_psn & BTH_PSN_MASK); qp->resp.psn = qp->attr.rq_psn; pr_debug("set resp psn = 0x%x\n", qp->resp.psn); } if (mask & IB_QP_MIN_RNR_TIMER) { qp->attr.min_rnr_timer = attr->min_rnr_timer; pr_debug("set min rnr timer = 0x%x\n", attr->min_rnr_timer); } if (mask & IB_QP_SQ_PSN) { qp->attr.sq_psn = (attr->sq_psn & BTH_PSN_MASK); qp->req.psn = qp->attr.sq_psn; qp->comp.psn = qp->attr.sq_psn; pr_debug("set req psn = 0x%x\n", qp->req.psn); } if (mask & IB_QP_MAX_DEST_RD_ATOMIC) { qp->attr.max_dest_rd_atomic = __roundup_pow_of_two(attr->max_dest_rd_atomic); } if (mask & IB_QP_PATH_MIG_STATE) qp->attr.path_mig_state = attr->path_mig_state; if (mask & IB_QP_DEST_QPN) qp->attr.dest_qp_num = attr->dest_qp_num; if (mask & IB_QP_STATE) { qp->attr.qp_state = attr->qp_state; switch (attr->qp_state) { case IB_QPS_RESET: pr_debug("qp state -> RESET\n"); rxe_qp_reset(qp); break; case IB_QPS_INIT: pr_debug("qp state -> INIT\n"); qp->req.state = QP_STATE_INIT; qp->resp.state = QP_STATE_INIT; break; case IB_QPS_RTR: pr_debug("qp state -> RTR\n"); qp->resp.state = QP_STATE_READY; break; case IB_QPS_RTS: pr_debug("qp state -> RTS\n"); qp->req.state = QP_STATE_READY; break; case IB_QPS_SQD: pr_debug("qp state -> SQD\n"); rxe_qp_drain(qp); break; case IB_QPS_SQE: pr_warn("qp state -> SQE !!?\n"); /* Not possible from modify_qp. */ break; case IB_QPS_ERR: pr_debug("qp state -> ERR\n"); rxe_qp_error(qp); break; } } return 0; }
struct ib_ah *ocrdma_create_ah(struct ib_pd *ibpd, struct ib_ah_attr *attr, struct ib_udata *udata) { u32 *ahid_addr; int status; struct ocrdma_ah *ah; bool isvlan = false; u16 vlan_tag = 0xffff; struct ib_gid_attr sgid_attr; struct ocrdma_pd *pd = get_ocrdma_pd(ibpd); struct ocrdma_dev *dev = get_ocrdma_dev(ibpd->device); union ib_gid sgid; if (!(attr->ah_flags & IB_AH_GRH)) return ERR_PTR(-EINVAL); if (atomic_cmpxchg(&dev->update_sl, 1, 0)) ocrdma_init_service_level(dev); ah = kzalloc(sizeof(*ah), GFP_ATOMIC); if (!ah) return ERR_PTR(-ENOMEM); status = ocrdma_alloc_av(dev, ah); if (status) goto av_err; status = ib_get_cached_gid(&dev->ibdev, 1, attr->grh.sgid_index, &sgid, &sgid_attr); if (status) { pr_err("%s(): Failed to query sgid, status = %d\n", __func__, status); goto av_conf_err; } if (sgid_attr.ndev) { if (is_vlan_dev(sgid_attr.ndev)) vlan_tag = vlan_dev_vlan_id(sgid_attr.ndev); dev_put(sgid_attr.ndev); } /* Get network header type for this GID */ ah->hdr_type = ib_gid_to_network_type(sgid_attr.gid_type, &sgid); if ((pd->uctx) && (!rdma_is_multicast_addr((struct in6_addr *)attr->grh.dgid.raw)) && (!rdma_link_local_addr((struct in6_addr *)attr->grh.dgid.raw))) { status = rdma_addr_find_l2_eth_by_grh(&sgid, &attr->grh.dgid, attr->dmac, &vlan_tag, &sgid_attr.ndev->ifindex, NULL); if (status) { pr_err("%s(): Failed to resolve dmac from gid." "status = %d\n", __func__, status); goto av_conf_err; } } status = set_av_attr(dev, ah, attr, &sgid, pd->id, &isvlan, vlan_tag); if (status) goto av_conf_err; /* if pd is for the user process, pass the ah_id to user space */ if ((pd->uctx) && (pd->uctx->ah_tbl.va)) { ahid_addr = pd->uctx->ah_tbl.va + attr->dlid; *ahid_addr = 0; *ahid_addr |= ah->id & OCRDMA_AH_ID_MASK; if (ocrdma_is_udp_encap_supported(dev)) { *ahid_addr |= ((u32)ah->hdr_type & OCRDMA_AH_L3_TYPE_MASK) << OCRDMA_AH_L3_TYPE_SHIFT; } if (isvlan) *ahid_addr |= (OCRDMA_AH_VLAN_VALID_MASK << OCRDMA_AH_VLAN_VALID_SHIFT); } return &ah->ibah; av_conf_err: ocrdma_free_av(dev, ah); av_err: kfree(ah); return ERR_PTR(status); }