From: Felix Fietkau <nbd@openwrt.org>
Date: Wed, 19 Sep 2012 23:50:18 +0000 (+0000)
Subject: cns3xxx: fix ethernet jumbo-frame support, use single-page rx buffers and skb fragmen... 
X-Git-Url: http://git.cdn.openwrt.org/?a=commitdiff_plain;h=34d30654ae750e984cfd7897c22d683f8db3fed0;p=openwrt%2Fstaging%2Fdedeckeh.git

cns3xxx: fix ethernet jumbo-frame support, use single-page rx buffers and skb fragment lists

SVN-Revision: 33491
---

diff --git a/target/linux/cns3xxx/patches-3.3/410-ethernet_fix_jumbo_frame.patch b/target/linux/cns3xxx/patches-3.3/410-ethernet_fix_jumbo_frame.patch
new file mode 100644
index 0000000000..268770ffb5
--- /dev/null
+++ b/target/linux/cns3xxx/patches-3.3/410-ethernet_fix_jumbo_frame.patch
@@ -0,0 +1,317 @@
+--- a/drivers/net/ethernet/cavium/cns3xxx_eth.c
++++ b/drivers/net/ethernet/cavium/cns3xxx_eth.c
+@@ -26,15 +26,18 @@
+ 
+ #define DRV_NAME "cns3xxx_eth"
+ 
+-#define RX_DESCS 512
+-#define TX_DESCS 512
+-#define SKB_DMA_REALIGN ((PAGE_SIZE - NET_SKB_PAD) % SMP_CACHE_BYTES)
++#define RX_DESCS 128
++#define TX_DESCS 128
+ 
+ #define RX_POOL_ALLOC_SIZE (sizeof(struct rx_desc) * RX_DESCS)
+ #define TX_POOL_ALLOC_SIZE (sizeof(struct tx_desc) * TX_DESCS)
+ #define REGS_SIZE 336
+-#define MAX_MRU (1536 + SKB_DMA_REALIGN)
+-#define CNS3XXX_MAX_MTU (1536)
++
++#define SKB_HEAD_ALIGN (((PAGE_SIZE - NET_SKB_PAD) % SMP_CACHE_BYTES) + NET_SKB_PAD + NET_IP_ALIGN)
++#define RX_SEGMENT_ALLOC_SIZE 4096
++#define RX_SEGMENT_BUFSIZE (SKB_WITH_OVERHEAD(RX_SEGMENT_ALLOC_SIZE))
++#define RX_SEGMENT_MRU ((RX_SEGMENT_BUFSIZE - SKB_HEAD_ALIGN))
++#define MAX_MTU	9500
+ 
+ #define NAPI_WEIGHT 64
+ 
+@@ -266,7 +269,7 @@ struct _rx_ring {
+ 	struct rx_desc *desc;
+ 	dma_addr_t phys_addr;
+ 	struct rx_desc *cur_addr;
+-	struct sk_buff *buff_tab[RX_DESCS];
++	void *buff_tab[RX_DESCS];
+ 	unsigned int phys_tab[RX_DESCS];
+ 	u32 cur_index;
+ 	u32 alloc_index;
+@@ -280,6 +283,8 @@ struct sw {
+ 	struct cns3xxx_plat_info *plat;
+ 	struct _tx_ring *tx_ring;
+ 	struct _rx_ring *rx_ring;
++	struct sk_buff *frag_first;
++	struct sk_buff *frag_last;
+ };
+ 
+ struct port {
+@@ -500,37 +505,35 @@ static void cns3xxx_alloc_rx_buf(struct 
+ 	struct _rx_ring *rx_ring = sw->rx_ring;
+ 	unsigned int i = rx_ring->alloc_index;
+ 	struct rx_desc *desc = &(rx_ring)->desc[i];
+-	struct sk_buff *skb;
++	void *buf;
+ 	unsigned int phys;
+ 
+ 	for (received += rx_ring->alloc_count; received > 0; received--) {
+-		if ((skb = dev_alloc_skb(MAX_MRU))) {
+-			if (SKB_DMA_REALIGN)
+-				skb_reserve(skb, SKB_DMA_REALIGN);
+-			skb_reserve(skb, NET_IP_ALIGN);
+-			phys = dma_map_single(NULL, skb->data,
+-				    CNS3XXX_MAX_MTU, DMA_FROM_DEVICE);
+-			if (dma_mapping_error(NULL, phys)) {
+-				dev_kfree_skb(skb);
+-				/* Failed to map, better luck next time */
+-				goto out;;
+-			}
+-			desc->sdp = phys;
+-		} else {
+-			/* Failed to allocate skb, try again next time */
++		buf = kzalloc(RX_SEGMENT_ALLOC_SIZE, GFP_ATOMIC);
++		if (!buf)
++			goto out;
++
++		phys = dma_map_single(NULL, buf + SKB_HEAD_ALIGN,
++				      RX_SEGMENT_MRU, DMA_FROM_DEVICE);
++		if (dma_mapping_error(NULL, phys)) {
++			kfree(buf);
+ 			goto out;
+ 		}
+ 
++		desc->sdl = RX_SEGMENT_MRU;
++		desc->sdp = phys;
++
+ 		/* put the new buffer on RX-free queue */
+-		rx_ring->buff_tab[i] = skb;
++		rx_ring->buff_tab[i] = buf;
+ 		rx_ring->phys_tab[i] = phys;
+ 		if (i == RX_DESCS - 1) {
+ 			i = 0;
+ 			desc->config0 = END_OF_RING | FIRST_SEGMENT |
+-					LAST_SEGMENT | CNS3XXX_MAX_MTU;
++					LAST_SEGMENT | RX_SEGMENT_MRU;
+ 			desc = &(rx_ring)->desc[i];
+ 		} else {
+-			desc->config0 = FIRST_SEGMENT | LAST_SEGMENT | CNS3XXX_MAX_MTU;
++			desc->config0 = FIRST_SEGMENT | LAST_SEGMENT |
++					RX_SEGMENT_MRU;
+ 			i++;
+ 			desc++;
+ 		}
+@@ -588,49 +591,78 @@ static int eth_poll(struct napi_struct *
+ 
+ 	while (desc->cown) {
+ 		struct sk_buff *skb;
++		int reserve = SKB_HEAD_ALIGN;
+ 
+ 		if (received >= budget)
+ 			break;
+ 
+-		skb = rx_ring->buff_tab[i];
++		/* process received frame */
++		dma_unmap_single(&dev->dev, rx_ring->phys_tab[i],
++				 RX_SEGMENT_MRU, DMA_FROM_DEVICE);
++
++		skb = build_skb(rx_ring->buff_tab[i]);
++		if (!skb)
++			break;
+ 
+ 		dev = switch_port_tab[desc->sp]->netdev;
+ 
+ 		length = desc->sdl;
+-		/* process received frame */
+-		dma_unmap_single(&dev->dev, rx_ring->phys_tab[i],
+-				 length, DMA_FROM_DEVICE);
++		if (desc->fsd && !desc->lsd)
++			length = RX_SEGMENT_MRU;
+ 
++		if (!desc->fsd) {
++			reserve -= NET_IP_ALIGN;
++			if (!desc->lsd)
++				length += NET_IP_ALIGN;
++		}
++
++		skb_reserve(skb, reserve);
+ 		skb_put(skb, length);
+ 
+-		skb->dev = dev;
+-		skb->protocol = eth_type_trans(skb, dev);
++		if (!sw->frag_first)
++			sw->frag_first = skb;
++		else {
++			if (sw->frag_first == sw->frag_last)
++				skb_frag_add_head(sw->frag_first, skb);
++			sw->frag_last->next = skb;
++			sw->frag_first->len += skb->len;
++			sw->frag_first->data_len += skb->len;
++			sw->frag_first->truesize += skb->truesize;
++		}
++		sw->frag_last = skb;
+ 
+-		dev->stats.rx_packets++;
+-		dev->stats.rx_bytes += length;
++		if (desc->lsd) {
++			skb = sw->frag_first;
++			skb->protocol = eth_type_trans(skb, dev);
++
++			dev->stats.rx_packets++;
++			dev->stats.rx_bytes += skb->len;
++
++			/* RX Hardware checksum offload */
++			skb->ip_summed = CHECKSUM_NONE;
++			switch (desc->prot) {
++				case 1:
++				case 2:
++				case 5:
++				case 6:
++				case 13:
++				case 14:
++					if (desc->l4f)
++						break;
+ 
+-		/* RX Hardware checksum offload */
+-		switch (desc->prot) {
+-			case 1:
+-			case 2:
+-			case 5:
+-			case 6:
+-			case 13:
+-			case 14:
+-				if (desc->l4f)
+-					skb->ip_summed = CHECKSUM_NONE;
+-				else
+ 					skb->ip_summed = CHECKSUM_UNNECESSARY;
+-			break;
+-			default:
+-				skb->ip_summed = CHECKSUM_NONE;
+-			break;
+-		}
++					break;
++				default:
++				break;
++			}
+ 
+-		napi_gro_receive(napi, skb);
++			napi_gro_receive(napi, skb);
+ 
+-		received++;
++			sw->frag_first = NULL;
++			sw->frag_last = NULL;
++		}
+ 
++		received++;
+ 		if (++i == RX_DESCS) {
+ 			i = 0;
+ 			desc = &(rx_ring)->desc[i];
+@@ -671,12 +703,6 @@ static int eth_xmit(struct sk_buff *skb,
+ 	if (pmap == 8)
+ 		pmap = (1 << 4);
+ 
+-	if (skb->len > CNS3XXX_MAX_MTU) {
+-		dev_kfree_skb(skb);
+-		dev->stats.tx_errors++;
+-		return NETDEV_TX_OK;
+-	}
+-
+ 	spin_lock(&tx_lock);
+ 
+ 	if ((tx_ring->num_used + nr_frags) >= TX_DESCS) {
+@@ -701,8 +727,7 @@ static int eth_xmit(struct sk_buff *skb,
+ 
+ 		len = skb->len;
+ 
+-		phys = dma_map_single(NULL, skb->data, len,
+-				      	DMA_TO_DEVICE);
++		phys = dma_map_single(NULL, skb->data, len, DMA_TO_DEVICE);
+ 
+ 		tx_desc->sdp = phys;
+ 		tx_desc->pmap = pmap;
+@@ -849,24 +874,24 @@ static int init_rings(struct sw *sw)
+ 	/* Setup RX buffers */
+ 	for (i = 0; i < RX_DESCS; i++) {
+ 		struct rx_desc *desc = &(rx_ring)->desc[i];
+-		struct sk_buff *skb;
+-		if (!(skb = dev_alloc_skb(MAX_MRU)))
++		void *buf;
++
++		buf = kzalloc(RX_SEGMENT_ALLOC_SIZE, GFP_KERNEL);
++		if (!buf)
+ 			return -ENOMEM;
+-		if (SKB_DMA_REALIGN)
+-			skb_reserve(skb, SKB_DMA_REALIGN);
+-		skb_reserve(skb, NET_IP_ALIGN);
+-		desc->sdl = CNS3XXX_MAX_MTU;
++
++		desc->sdl = RX_SEGMENT_MRU;
+ 		if (i == (RX_DESCS - 1))
+ 			desc->eor = 1;
+ 		desc->fsd = 1;
+ 		desc->lsd = 1;
+ 
+-		desc->sdp = dma_map_single(NULL, skb->data,
+-					    CNS3XXX_MAX_MTU, DMA_FROM_DEVICE);
+-		if (dma_mapping_error(NULL, desc->sdp)) {
++		desc->sdp = dma_map_single(NULL, buf + SKB_HEAD_ALIGN,
++					   RX_SEGMENT_MRU, DMA_FROM_DEVICE);
++		if (dma_mapping_error(NULL, desc->sdp))
+ 			return -EIO;
+-		}
+-		rx_ring->buff_tab[i] = skb;
++
++		rx_ring->buff_tab[i] = buf;
+ 		rx_ring->phys_tab[i] = desc->sdp;
+ 		desc->cown = 0;
+ 	}
+@@ -905,12 +930,13 @@ static void destroy_rings(struct sw *sw)
+ 			struct _rx_ring *rx_ring = sw->rx_ring;
+ 			struct rx_desc *desc = &(rx_ring)->desc[i];
+ 			struct sk_buff *skb = sw->rx_ring->buff_tab[i];
+-			if (skb) {
+-				dma_unmap_single(NULL,
+-						 desc->sdp,
+-						 CNS3XXX_MAX_MTU, DMA_FROM_DEVICE);
+-				dev_kfree_skb(skb);
+-			}
++
++			if (!skb)
++				continue;
++
++			dma_unmap_single(NULL, desc->sdp, RX_SEGMENT_MRU,
++					 DMA_FROM_DEVICE);
++			dev_kfree_skb(skb);
+ 		}
+ 		dma_pool_free(rx_dma_pool, sw->rx_ring->desc, sw->rx_ring->phys_addr);
+ 		dma_pool_destroy(rx_dma_pool);
+@@ -1085,13 +1111,22 @@ static int eth_set_mac(struct net_device
+ 	return 0;
+ }
+ 
++static int cns3xxx_change_mtu(struct net_device *dev, int new_mtu)
++{
++	if (new_mtu > MAX_MTU)
++		return -EINVAL;
++
++	dev->mtu = new_mtu;
++	return 0;
++}
++
+ static const struct net_device_ops cns3xxx_netdev_ops = {
+ 	.ndo_open = eth_open,
+ 	.ndo_stop = eth_close,
+ 	.ndo_start_xmit = eth_xmit,
+ 	.ndo_set_rx_mode = eth_rx_mode,
+ 	.ndo_do_ioctl = eth_ioctl,
+-	.ndo_change_mtu = eth_change_mtu,
++	.ndo_change_mtu = cns3xxx_change_mtu,
+ 	.ndo_set_mac_address = eth_set_mac,
+ 	.ndo_validate_addr = eth_validate_addr,
+ };
+@@ -1124,6 +1159,10 @@ static int __devinit eth_init_one(struct
+ 		goto err_free;
+ 	}
+ 
++	temp = __raw_readl(&sw->regs->phy_auto_addr);
++	temp |= (3 << 30); /* maximum frame length: 9600 bytes */
++	__raw_writel(temp, &sw->regs->phy_auto_addr);
++
+ 	for (i = 0; i < 4; i++) {
+ 		temp = __raw_readl(&sw->regs->mac_cfg[i]);
+ 		temp |= (PORT_DISABLE);