⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 ipoib_ib.c

📁 linux内核源码
💻 C
📖 第 1 页 / 共 2 页
字号:
/* * Copyright (c) 2004, 2005 Topspin Communications.  All rights reserved. * Copyright (c) 2005 Sun Microsystems, Inc. All rights reserved. * Copyright (c) 2005 Mellanox Technologies. All rights reserved. * Copyright (c) 2004, 2005 Voltaire, Inc. All rights reserved. * * This software is available to you under a choice of one of two * licenses.  You may choose to be licensed under the terms of the GNU * General Public License (GPL) Version 2, available from the file * COPYING in the main directory of this source tree, or the * OpenIB.org BSD license below: * *     Redistribution and use in source and binary forms, with or *     without modification, are permitted provided that the following *     conditions are met: * *      - Redistributions of source code must retain the above *        copyright notice, this list of conditions and the following *        disclaimer. * *      - Redistributions in binary form must reproduce the above *        copyright notice, this list of conditions and the following *        disclaimer in the documentation and/or other materials *        provided with the distribution. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. * * $Id: ipoib_ib.c 1386 2004-12-27 16:23:17Z roland $ */#include <linux/delay.h>#include <linux/dma-mapping.h>#include <rdma/ib_cache.h>#include "ipoib.h"#ifdef CONFIG_INFINIBAND_IPOIB_DEBUG_DATAstatic int data_debug_level;module_param(data_debug_level, int, 0644);MODULE_PARM_DESC(data_debug_level,		 "Enable data path debug tracing if > 0");#endifstatic DEFINE_MUTEX(pkey_mutex);struct ipoib_ah *ipoib_create_ah(struct net_device *dev,				 struct ib_pd *pd, struct ib_ah_attr *attr){	struct ipoib_ah *ah;	ah = kmalloc(sizeof *ah, GFP_KERNEL);	if (!ah)		return NULL;	ah->dev       = dev;	ah->last_send = 0;	kref_init(&ah->ref);	ah->ah = ib_create_ah(pd, attr);	if (IS_ERR(ah->ah)) {		kfree(ah);		ah = NULL;	} else		ipoib_dbg(netdev_priv(dev), "Created ah %p\n", ah->ah);	return ah;}void ipoib_free_ah(struct kref *kref){	struct ipoib_ah *ah = container_of(kref, struct ipoib_ah, ref);	struct ipoib_dev_priv *priv = netdev_priv(ah->dev);	unsigned long flags;	spin_lock_irqsave(&priv->lock, flags);	list_add_tail(&ah->list, &priv->dead_ahs);	spin_unlock_irqrestore(&priv->lock, flags);}static int ipoib_ib_post_receive(struct net_device *dev, int id){	struct ipoib_dev_priv *priv = netdev_priv(dev);	struct ib_sge list;	struct ib_recv_wr param;	struct ib_recv_wr *bad_wr;	int ret;	list.addr     = priv->rx_ring[id].mapping;	list.length   = IPOIB_BUF_SIZE;	list.lkey     = priv->mr->lkey;	param.next    = NULL;	param.wr_id   = id | IPOIB_OP_RECV;	param.sg_list = &list;	param.num_sge = 1;	ret = ib_post_recv(priv->qp, &param, &bad_wr);	if (unlikely(ret)) {		ipoib_warn(priv, "receive failed for buf %d (%d)\n", id, ret);		ib_dma_unmap_single(priv->ca, priv->rx_ring[id].mapping,				    IPOIB_BUF_SIZE, DMA_FROM_DEVICE);		dev_kfree_skb_any(priv->rx_ring[id].skb);		priv->rx_ring[id].skb = NULL;	}	return ret;}static int ipoib_alloc_rx_skb(struct net_device *dev, int id){	struct ipoib_dev_priv *priv = netdev_priv(dev);	struct sk_buff *skb;	u64 addr;	skb = dev_alloc_skb(IPOIB_BUF_SIZE + 4);	if (!skb)		return -ENOMEM;	/*	 * IB will leave a 40 byte gap for a GRH and IPoIB adds a 4 byte	 * header.  So we need 4 more bytes to get to 48 and align the	 * IP header to a multiple of 16.	 */	skb_reserve(skb, 4);	addr = ib_dma_map_single(priv->ca, skb->data, IPOIB_BUF_SIZE,				 DMA_FROM_DEVICE);	if (unlikely(ib_dma_mapping_error(priv->ca, addr))) {		dev_kfree_skb_any(skb);		return -EIO;	}	priv->rx_ring[id].skb     = skb;	priv->rx_ring[id].mapping = addr;	return 0;}static int ipoib_ib_post_receives(struct net_device *dev){	struct ipoib_dev_priv *priv = netdev_priv(dev);	int i;	for (i = 0; i < ipoib_recvq_size; ++i) {		if (ipoib_alloc_rx_skb(dev, i)) {			ipoib_warn(priv, "failed to allocate receive buffer %d\n", i);			return -ENOMEM;		}		if (ipoib_ib_post_receive(dev, i)) {			ipoib_warn(priv, "ipoib_ib_post_receive failed for buf %d\n", i);			return -EIO;		}	}	return 0;}static void ipoib_ib_handle_rx_wc(struct net_device *dev, struct ib_wc *wc){	struct ipoib_dev_priv *priv = netdev_priv(dev);	unsigned int wr_id = wc->wr_id & ~IPOIB_OP_RECV;	struct sk_buff *skb;	u64 addr;	ipoib_dbg_data(priv, "recv completion: id %d, status: %d\n",		       wr_id, wc->status);	if (unlikely(wr_id >= ipoib_recvq_size)) {		ipoib_warn(priv, "recv completion event with wrid %d (> %d)\n",			   wr_id, ipoib_recvq_size);		return;	}	skb  = priv->rx_ring[wr_id].skb;	addr = priv->rx_ring[wr_id].mapping;	if (unlikely(wc->status != IB_WC_SUCCESS)) {		if (wc->status != IB_WC_WR_FLUSH_ERR)			ipoib_warn(priv, "failed recv event "				   "(status=%d, wrid=%d vend_err %x)\n",				   wc->status, wr_id, wc->vendor_err);		ib_dma_unmap_single(priv->ca, addr,				    IPOIB_BUF_SIZE, DMA_FROM_DEVICE);		dev_kfree_skb_any(skb);		priv->rx_ring[wr_id].skb = NULL;		return;	}	/*	 * Drop packets that this interface sent, ie multicast packets	 * that the HCA has replicated.	 */	if (wc->slid == priv->local_lid && wc->src_qp == priv->qp->qp_num)		goto repost;	/*	 * If we can't allocate a new RX buffer, dump	 * this packet and reuse the old buffer.	 */	if (unlikely(ipoib_alloc_rx_skb(dev, wr_id))) {		++dev->stats.rx_dropped;		goto repost;	}	ipoib_dbg_data(priv, "received %d bytes, SLID 0x%04x\n",		       wc->byte_len, wc->slid);	ib_dma_unmap_single(priv->ca, addr, IPOIB_BUF_SIZE, DMA_FROM_DEVICE);	skb_put(skb, wc->byte_len);	skb_pull(skb, IB_GRH_BYTES);	skb->protocol = ((struct ipoib_header *) skb->data)->proto;	skb_reset_mac_header(skb);	skb_pull(skb, IPOIB_ENCAP_LEN);	dev->last_rx = jiffies;	++dev->stats.rx_packets;	dev->stats.rx_bytes += skb->len;	skb->dev = dev;	/* XXX get correct PACKET_ type here */	skb->pkt_type = PACKET_HOST;	netif_receive_skb(skb);repost:	if (unlikely(ipoib_ib_post_receive(dev, wr_id)))		ipoib_warn(priv, "ipoib_ib_post_receive failed "			   "for buf %d\n", wr_id);}static void ipoib_ib_handle_tx_wc(struct net_device *dev, struct ib_wc *wc){	struct ipoib_dev_priv *priv = netdev_priv(dev);	unsigned int wr_id = wc->wr_id;	struct ipoib_tx_buf *tx_req;	unsigned long flags;	ipoib_dbg_data(priv, "send completion: id %d, status: %d\n",		       wr_id, wc->status);	if (unlikely(wr_id >= ipoib_sendq_size)) {		ipoib_warn(priv, "send completion event with wrid %d (> %d)\n",			   wr_id, ipoib_sendq_size);		return;	}	tx_req = &priv->tx_ring[wr_id];	ib_dma_unmap_single(priv->ca, tx_req->mapping,			    tx_req->skb->len, DMA_TO_DEVICE);	++dev->stats.tx_packets;	dev->stats.tx_bytes += tx_req->skb->len;	dev_kfree_skb_any(tx_req->skb);	spin_lock_irqsave(&priv->tx_lock, flags);	++priv->tx_tail;	if (unlikely(--priv->tx_outstanding == ipoib_sendq_size >> 1) &&	    netif_queue_stopped(dev) &&	    test_bit(IPOIB_FLAG_ADMIN_UP, &priv->flags))		netif_wake_queue(dev);	spin_unlock_irqrestore(&priv->tx_lock, flags);	if (wc->status != IB_WC_SUCCESS &&	    wc->status != IB_WC_WR_FLUSH_ERR)		ipoib_warn(priv, "failed send event "			   "(status=%d, wrid=%d vend_err %x)\n",			   wc->status, wr_id, wc->vendor_err);}int ipoib_poll(struct napi_struct *napi, int budget){	struct ipoib_dev_priv *priv = container_of(napi, struct ipoib_dev_priv, napi);	struct net_device *dev = priv->dev;	int done;	int t;	int n, i;	done  = 0;poll_more:	while (done < budget) {		int max = (budget - done);		t = min(IPOIB_NUM_WC, max);		n = ib_poll_cq(priv->cq, t, priv->ibwc);		for (i = 0; i < n; i++) {			struct ib_wc *wc = priv->ibwc + i;			if (wc->wr_id & IPOIB_OP_RECV) {				++done;				if (wc->wr_id & IPOIB_OP_CM)					ipoib_cm_handle_rx_wc(dev, wc);				else					ipoib_ib_handle_rx_wc(dev, wc);			} else {				if (wc->wr_id & IPOIB_OP_CM)					ipoib_cm_handle_tx_wc(dev, wc);				else					ipoib_ib_handle_tx_wc(dev, wc);			}		}		if (n != t)			break;	}	if (done < budget) {		netif_rx_complete(dev, napi);		if (unlikely(ib_req_notify_cq(priv->cq,					      IB_CQ_NEXT_COMP |					      IB_CQ_REPORT_MISSED_EVENTS)) &&		    netif_rx_reschedule(dev, napi))			goto poll_more;	}	return done;}void ipoib_ib_completion(struct ib_cq *cq, void *dev_ptr){	struct net_device *dev = dev_ptr;	struct ipoib_dev_priv *priv = netdev_priv(dev);	netif_rx_schedule(dev, &priv->napi);}static inline int post_send(struct ipoib_dev_priv *priv,			    unsigned int wr_id,			    struct ib_ah *address, u32 qpn,			    u64 addr, int len){	struct ib_send_wr *bad_wr;	priv->tx_sge.addr             = addr;	priv->tx_sge.length           = len;	priv->tx_wr.wr_id 	      = wr_id;	priv->tx_wr.wr.ud.remote_qpn  = qpn;	priv->tx_wr.wr.ud.ah 	      = address;	return ib_post_send(priv->qp, &priv->tx_wr, &bad_wr);}void ipoib_send(struct net_device *dev, struct sk_buff *skb,		struct ipoib_ah *address, u32 qpn){	struct ipoib_dev_priv *priv = netdev_priv(dev);	struct ipoib_tx_buf *tx_req;	u64 addr;	if (unlikely(skb->len > priv->mcast_mtu + IPOIB_ENCAP_LEN)) {		ipoib_warn(priv, "packet len %d (> %d) too long to send, dropping\n",			   skb->len, priv->mcast_mtu + IPOIB_ENCAP_LEN);		++dev->stats.tx_dropped;		++dev->stats.tx_errors;		ipoib_cm_skb_too_long(dev, skb, priv->mcast_mtu);		return;	}	ipoib_dbg_data(priv, "sending packet, length=%d address=%p qpn=0x%06x\n",		       skb->len, address, qpn);	/*	 * We put the skb into the tx_ring _before_ we call post_send()	 * because it's entirely possible that the completion handler will	 * run before we execute anything after the post_send().  That	 * means we have to make sure everything is properly recorded and	 * our state is consistent before we call post_send().	 */	tx_req = &priv->tx_ring[priv->tx_head & (ipoib_sendq_size - 1)];	tx_req->skb = skb;	addr = ib_dma_map_single(priv->ca, skb->data, skb->len,				 DMA_TO_DEVICE);	if (unlikely(ib_dma_mapping_error(priv->ca, addr))) {		++dev->stats.tx_errors;		dev_kfree_skb_any(skb);		return;	}	tx_req->mapping = addr;	if (unlikely(post_send(priv, priv->tx_head & (ipoib_sendq_size - 1),			       address->ah, qpn, addr, skb->len))) {		ipoib_warn(priv, "post_send failed\n");		++dev->stats.tx_errors;		ib_dma_unmap_single(priv->ca, addr, skb->len, DMA_TO_DEVICE);		dev_kfree_skb_any(skb);	} else {		dev->trans_start = jiffies;		address->last_send = priv->tx_head;		++priv->tx_head;		if (++priv->tx_outstanding == ipoib_sendq_size) {			ipoib_dbg(priv, "TX ring full, stopping kernel net queue\n");			netif_stop_queue(dev);		}	}}static void __ipoib_reap_ah(struct net_device *dev){	struct ipoib_dev_priv *priv = netdev_priv(dev);	struct ipoib_ah *ah, *tah;	LIST_HEAD(remove_list);	spin_lock_irq(&priv->tx_lock);	spin_lock(&priv->lock);	list_for_each_entry_safe(ah, tah, &priv->dead_ahs, list)		if ((int) priv->tx_tail - (int) ah->last_send >= 0) {			list_del(&ah->list);

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -