ipath_qp.c

来自「linux 内核源代码」· C语言 代码 · 共 1,082 行 · 第 1/2 页

C
1,082
字号
		break;	default:		break;	}	if (attr_mask & IB_QP_PKEY_INDEX)		qp->s_pkey_index = attr->pkey_index;	if (attr_mask & IB_QP_DEST_QPN)		qp->remote_qpn = attr->dest_qp_num;	if (attr_mask & IB_QP_SQ_PSN) {		qp->s_psn = qp->s_next_psn = attr->sq_psn;		qp->s_last_psn = qp->s_next_psn - 1;	}	if (attr_mask & IB_QP_RQ_PSN)		qp->r_psn = attr->rq_psn;	if (attr_mask & IB_QP_ACCESS_FLAGS)		qp->qp_access_flags = attr->qp_access_flags;	if (attr_mask & IB_QP_AV)		qp->remote_ah_attr = attr->ah_attr;	if (attr_mask & IB_QP_PATH_MTU)		qp->path_mtu = attr->path_mtu;	if (attr_mask & IB_QP_RETRY_CNT)		qp->s_retry = qp->s_retry_cnt = attr->retry_cnt;	if (attr_mask & IB_QP_RNR_RETRY) {		qp->s_rnr_retry = attr->rnr_retry;		if (qp->s_rnr_retry > 7)			qp->s_rnr_retry = 7;		qp->s_rnr_retry_cnt = qp->s_rnr_retry;	}	if (attr_mask & IB_QP_MIN_RNR_TIMER)		qp->r_min_rnr_timer = attr->min_rnr_timer;	if (attr_mask & IB_QP_TIMEOUT)		qp->timeout = attr->timeout;	if (attr_mask & IB_QP_QKEY)		qp->qkey = attr->qkey;	if (attr_mask & IB_QP_MAX_DEST_RD_ATOMIC)		qp->r_max_rd_atomic = attr->max_dest_rd_atomic;	if (attr_mask & IB_QP_MAX_QP_RD_ATOMIC)		qp->s_max_rd_atomic = attr->max_rd_atomic;	qp->state = new_state;	spin_unlock_irqrestore(&qp->s_lock, flags);	if (lastwqe) {		struct ib_event ev;		ev.device = qp->ibqp.device;		ev.element.qp = &qp->ibqp;		ev.event = IB_EVENT_QP_LAST_WQE_REACHED;		qp->ibqp.event_handler(&ev, qp->ibqp.qp_context);	}	ret = 0;	goto bail;inval:	spin_unlock_irqrestore(&qp->s_lock, flags);	ret = -EINVAL;bail:	return ret;}int ipath_query_qp(struct ib_qp *ibqp, struct ib_qp_attr *attr,		   int attr_mask, struct ib_qp_init_attr *init_attr){	struct ipath_qp *qp = to_iqp(ibqp);	attr->qp_state = qp->state;	attr->cur_qp_state = attr->qp_state;	attr->path_mtu = qp->path_mtu;	attr->path_mig_state = 0;	attr->qkey = qp->qkey;	attr->rq_psn = qp->r_psn;	attr->sq_psn = qp->s_next_psn;	attr->dest_qp_num = qp->remote_qpn;	attr->qp_access_flags = qp->qp_access_flags;	attr->cap.max_send_wr = qp->s_size - 1;	attr->cap.max_recv_wr = qp->ibqp.srq ? 0 : qp->r_rq.size - 1;	attr->cap.max_send_sge = qp->s_max_sge;	attr->cap.max_recv_sge = qp->r_rq.max_sge;	attr->cap.max_inline_data = 0;	attr->ah_attr = qp->remote_ah_attr;	memset(&attr->alt_ah_attr, 0, sizeof(attr->alt_ah_attr));	attr->pkey_index = qp->s_pkey_index;	attr->alt_pkey_index = 0;	attr->en_sqd_async_notify = 0;	attr->sq_draining = 0;	attr->max_rd_atomic = qp->s_max_rd_atomic;	attr->max_dest_rd_atomic = qp->r_max_rd_atomic;	attr->min_rnr_timer = qp->r_min_rnr_timer;	attr->port_num = 1;	attr->timeout = qp->timeout;	attr->retry_cnt = qp->s_retry_cnt;	attr->rnr_retry = qp->s_rnr_retry;	attr->alt_port_num = 0;	attr->alt_timeout = 0;	init_attr->event_handler = qp->ibqp.event_handler;	init_attr->qp_context = qp->ibqp.qp_context;	init_attr->send_cq = qp->ibqp.send_cq;	init_attr->recv_cq = qp->ibqp.recv_cq;	init_attr->srq = qp->ibqp.srq;	init_attr->cap = attr->cap;	if (qp->s_flags & IPATH_S_SIGNAL_REQ_WR)		init_attr->sq_sig_type = IB_SIGNAL_REQ_WR;	else		init_attr->sq_sig_type = IB_SIGNAL_ALL_WR;	init_attr->qp_type = qp->ibqp.qp_type;	init_attr->port_num = 1;	return 0;}/** * ipath_compute_aeth - compute the AETH (syndrome + MSN) * @qp: the queue pair to compute the AETH for * * Returns the AETH. */__be32 ipath_compute_aeth(struct ipath_qp *qp){	u32 aeth = qp->r_msn & IPATH_MSN_MASK;	if (qp->ibqp.srq) {		/*		 * Shared receive queues don't generate credits.		 * Set the credit field to the invalid value.		 */		aeth |= IPATH_AETH_CREDIT_INVAL << IPATH_AETH_CREDIT_SHIFT;	} else {		u32 min, max, x;		u32 credits;		struct ipath_rwq *wq = qp->r_rq.wq;		u32 head;		u32 tail;		/* sanity check pointers before trusting them */		head = wq->head;		if (head >= qp->r_rq.size)			head = 0;		tail = wq->tail;		if (tail >= qp->r_rq.size)			tail = 0;		/*		 * Compute the number of credits available (RWQEs).		 * XXX Not holding the r_rq.lock here so there is a small		 * chance that the pair of reads are not atomic.		 */		credits = head - tail;		if ((int)credits < 0)			credits += qp->r_rq.size;		/*		 * Binary search the credit table to find the code to		 * use.		 */		min = 0;		max = 31;		for (;;) {			x = (min + max) / 2;			if (credit_table[x] == credits)				break;			if (credit_table[x] > credits)				max = x;			else if (min == x)				break;			else				min = x;		}		aeth |= x << IPATH_AETH_CREDIT_SHIFT;	}	return cpu_to_be32(aeth);}/** * ipath_create_qp - create a queue pair for a device * @ibpd: the protection domain who's device we create the queue pair for * @init_attr: the attributes of the queue pair * @udata: unused by InfiniPath * * Returns the queue pair on success, otherwise returns an errno. * * Called by the ib_create_qp() core verbs function. */struct ib_qp *ipath_create_qp(struct ib_pd *ibpd,			      struct ib_qp_init_attr *init_attr,			      struct ib_udata *udata){	struct ipath_qp *qp;	int err;	struct ipath_swqe *swq = NULL;	struct ipath_ibdev *dev;	size_t sz;	struct ib_qp *ret;	if (init_attr->cap.max_send_sge > ib_ipath_max_sges ||	    init_attr->cap.max_recv_sge > ib_ipath_max_sges ||	    init_attr->cap.max_send_wr > ib_ipath_max_qp_wrs ||	    init_attr->cap.max_recv_wr > ib_ipath_max_qp_wrs) {		ret = ERR_PTR(-ENOMEM);		goto bail;	}	if (init_attr->cap.max_send_sge +	    init_attr->cap.max_recv_sge +	    init_attr->cap.max_send_wr +	    init_attr->cap.max_recv_wr == 0) {		ret = ERR_PTR(-EINVAL);		goto bail;	}	switch (init_attr->qp_type) {	case IB_QPT_UC:	case IB_QPT_RC:	case IB_QPT_UD:	case IB_QPT_SMI:	case IB_QPT_GSI:		sz = sizeof(struct ipath_sge) *			init_attr->cap.max_send_sge +			sizeof(struct ipath_swqe);		swq = vmalloc((init_attr->cap.max_send_wr + 1) * sz);		if (swq == NULL) {			ret = ERR_PTR(-ENOMEM);			goto bail;		}		sz = sizeof(*qp);		if (init_attr->srq) {			struct ipath_srq *srq = to_isrq(init_attr->srq);			sz += sizeof(*qp->r_sg_list) *				srq->rq.max_sge;		} else			sz += sizeof(*qp->r_sg_list) *				init_attr->cap.max_recv_sge;		qp = kmalloc(sz, GFP_KERNEL);		if (!qp) {			ret = ERR_PTR(-ENOMEM);			goto bail_swq;		}		if (init_attr->srq) {			sz = 0;			qp->r_rq.size = 0;			qp->r_rq.max_sge = 0;			qp->r_rq.wq = NULL;			init_attr->cap.max_recv_wr = 0;			init_attr->cap.max_recv_sge = 0;		} else {			qp->r_rq.size = init_attr->cap.max_recv_wr + 1;			qp->r_rq.max_sge = init_attr->cap.max_recv_sge;			sz = (sizeof(struct ib_sge) * qp->r_rq.max_sge) +				sizeof(struct ipath_rwqe);			qp->r_rq.wq = vmalloc_user(sizeof(struct ipath_rwq) +					      qp->r_rq.size * sz);			if (!qp->r_rq.wq) {				ret = ERR_PTR(-ENOMEM);				goto bail_qp;			}		}		/*		 * ib_create_qp() will initialize qp->ibqp		 * except for qp->ibqp.qp_num.		 */		spin_lock_init(&qp->s_lock);		spin_lock_init(&qp->r_rq.lock);		atomic_set(&qp->refcount, 0);		init_waitqueue_head(&qp->wait);		tasklet_init(&qp->s_task, ipath_do_send, (unsigned long)qp);		INIT_LIST_HEAD(&qp->piowait);		INIT_LIST_HEAD(&qp->timerwait);		qp->state = IB_QPS_RESET;		qp->s_wq = swq;		qp->s_size = init_attr->cap.max_send_wr + 1;		qp->s_max_sge = init_attr->cap.max_send_sge;		if (init_attr->sq_sig_type == IB_SIGNAL_REQ_WR)			qp->s_flags = IPATH_S_SIGNAL_REQ_WR;		else			qp->s_flags = 0;		dev = to_idev(ibpd->device);		err = ipath_alloc_qpn(&dev->qp_table, qp,				      init_attr->qp_type);		if (err) {			ret = ERR_PTR(err);			vfree(qp->r_rq.wq);			goto bail_qp;		}		qp->ip = NULL;		ipath_reset_qp(qp);		break;	default:		/* Don't support raw QPs */		ret = ERR_PTR(-ENOSYS);		goto bail;	}	init_attr->cap.max_inline_data = 0;	/*	 * Return the address of the RWQ as the offset to mmap.	 * See ipath_mmap() for details.	 */	if (udata && udata->outlen >= sizeof(__u64)) {		int err;		if (!qp->r_rq.wq) {			__u64 offset = 0;			err = ib_copy_to_udata(udata, &offset,					       sizeof(offset));			if (err) {				ret = ERR_PTR(err);				goto bail_ip;			}		} else {			u32 s = sizeof(struct ipath_rwq) +				qp->r_rq.size * sz;			qp->ip =			    ipath_create_mmap_info(dev, s,						   ibpd->uobject->context,						   qp->r_rq.wq);			if (!qp->ip) {				ret = ERR_PTR(-ENOMEM);				goto bail_ip;			}			err = ib_copy_to_udata(udata, &(qp->ip->offset),					       sizeof(qp->ip->offset));			if (err) {				ret = ERR_PTR(err);				goto bail_ip;			}		}	}	spin_lock(&dev->n_qps_lock);	if (dev->n_qps_allocated == ib_ipath_max_qps) {		spin_unlock(&dev->n_qps_lock);		ret = ERR_PTR(-ENOMEM);		goto bail_ip;	}	dev->n_qps_allocated++;	spin_unlock(&dev->n_qps_lock);	if (qp->ip) {		spin_lock_irq(&dev->pending_lock);		list_add(&qp->ip->pending_mmaps, &dev->pending_mmaps);		spin_unlock_irq(&dev->pending_lock);	}	ret = &qp->ibqp;	goto bail;bail_ip:	if (qp->ip)		kref_put(&qp->ip->ref, ipath_release_mmap_info);	else		vfree(qp->r_rq.wq);	ipath_free_qp(&dev->qp_table, qp);bail_qp:	kfree(qp);bail_swq:	vfree(swq);bail:	return ret;}/** * ipath_destroy_qp - destroy a queue pair * @ibqp: the queue pair to destroy * * Returns 0 on success. * * Note that this can be called while the QP is actively sending or * receiving! */int ipath_destroy_qp(struct ib_qp *ibqp){	struct ipath_qp *qp = to_iqp(ibqp);	struct ipath_ibdev *dev = to_idev(ibqp->device);	unsigned long flags;	spin_lock_irqsave(&qp->s_lock, flags);	qp->state = IB_QPS_ERR;	spin_unlock_irqrestore(&qp->s_lock, flags);	spin_lock(&dev->n_qps_lock);	dev->n_qps_allocated--;	spin_unlock(&dev->n_qps_lock);	/* Stop the sending tasklet. */	tasklet_kill(&qp->s_task);	/* Make sure the QP isn't on the timeout list. */	spin_lock_irqsave(&dev->pending_lock, flags);	if (!list_empty(&qp->timerwait))		list_del_init(&qp->timerwait);	if (!list_empty(&qp->piowait))		list_del_init(&qp->piowait);	spin_unlock_irqrestore(&dev->pending_lock, flags);	/*	 * Make sure that the QP is not in the QPN table so receive	 * interrupts will discard packets for this QP.  XXX Also remove QP	 * from multicast table.	 */	if (atomic_read(&qp->refcount) != 0)		ipath_free_qp(&dev->qp_table, qp);	if (qp->ip)		kref_put(&qp->ip->ref, ipath_release_mmap_info);	else		vfree(qp->r_rq.wq);	vfree(qp->s_wq);	kfree(qp);	return 0;}/** * ipath_init_qp_table - initialize the QP table for a device * @idev: the device who's QP table we're initializing * @size: the size of the QP table * * Returns 0 on success, otherwise returns an errno. */int ipath_init_qp_table(struct ipath_ibdev *idev, int size){	int i;	int ret;	idev->qp_table.last = 1;	/* QPN 0 and 1 are special. */	idev->qp_table.max = size;	idev->qp_table.nmaps = 1;	idev->qp_table.table = kzalloc(size * sizeof(*idev->qp_table.table),				       GFP_KERNEL);	if (idev->qp_table.table == NULL) {		ret = -ENOMEM;		goto bail;	}	for (i = 0; i < ARRAY_SIZE(idev->qp_table.map); i++) {		atomic_set(&idev->qp_table.map[i].n_free, BITS_PER_PAGE);		idev->qp_table.map[i].page = NULL;	}	ret = 0;bail:	return ret;}/** * ipath_sqerror_qp - put a QP's send queue into an error state * @qp: QP who's send queue will be put into an error state * @wc: the WC responsible for putting the QP in this state * * Flushes the send work queue. * The QP s_lock should be held and interrupts disabled. */void ipath_sqerror_qp(struct ipath_qp *qp, struct ib_wc *wc){	struct ipath_ibdev *dev = to_idev(qp->ibqp.device);	struct ipath_swqe *wqe = get_swqe_ptr(qp, qp->s_last);	ipath_dbg("Send queue error on QP%d/%d: err: %d\n",		  qp->ibqp.qp_num, qp->remote_qpn, wc->status);	spin_lock(&dev->pending_lock);	/* XXX What if its already removed by the timeout code? */	if (!list_empty(&qp->timerwait))		list_del_init(&qp->timerwait);	if (!list_empty(&qp->piowait))		list_del_init(&qp->piowait);	spin_unlock(&dev->pending_lock);	ipath_cq_enter(to_icq(qp->ibqp.send_cq), wc, 1);	if (++qp->s_last >= qp->s_size)		qp->s_last = 0;	wc->status = IB_WC_WR_FLUSH_ERR;	while (qp->s_last != qp->s_head) {		wqe = get_swqe_ptr(qp, qp->s_last);		wc->wr_id = wqe->wr.wr_id;		wc->opcode = ib_ipath_wc_opcode[wqe->wr.opcode];		ipath_cq_enter(to_icq(qp->ibqp.send_cq), wc, 1);		if (++qp->s_last >= qp->s_size)			qp->s_last = 0;	}	qp->s_cur = qp->s_tail = qp->s_head;	qp->state = IB_QPS_SQE;}/** * ipath_get_credit - flush the send work queue of a QP * @qp: the qp who's send work queue to flush * @aeth: the Acknowledge Extended Transport Header * * The QP s_lock should be held. */void ipath_get_credit(struct ipath_qp *qp, u32 aeth){	u32 credit = (aeth >> IPATH_AETH_CREDIT_SHIFT) & IPATH_AETH_CREDIT_MASK;	/*	 * If the credit is invalid, we can send	 * as many packets as we like.  Otherwise, we have to	 * honor the credit field.	 */	if (credit == IPATH_AETH_CREDIT_INVAL)		qp->s_lsn = (u32) -1;	else if (qp->s_lsn != (u32) -1) {		/* Compute new LSN (i.e., MSN + credit) */		credit = (aeth + credit_table[credit]) & IPATH_MSN_MASK;		if (ipath_cmp24(credit, qp->s_lsn) > 0)			qp->s_lsn = credit;	}	/* Restart sending if it was blocked due to lack of credits. */	if (qp->s_cur != qp->s_head &&	    (qp->s_lsn == (u32) -1 ||	     ipath_cmp24(get_swqe_ptr(qp, qp->s_cur)->ssn,			 qp->s_lsn + 1) <= 0))		tasklet_hi_schedule(&qp->s_task);}

⌨️ 快捷键说明

复制代码Ctrl + C
搜索代码Ctrl + F
全屏模式F11
增大字号Ctrl + =
减小字号Ctrl + -
显示快捷键?