1 --- a/drivers/net/ethernet/cavium/cns3xxx_eth.c
2 +++ b/drivers/net/ethernet/cavium/cns3xxx_eth.c
5 #define DRV_NAME "cns3xxx_eth"
9 -#define SKB_DMA_REALIGN ((PAGE_SIZE - NET_SKB_PAD) % SMP_CACHE_BYTES)
13 #define RX_POOL_ALLOC_SIZE (sizeof(struct rx_desc) * RX_DESCS)
14 #define TX_POOL_ALLOC_SIZE (sizeof(struct tx_desc) * TX_DESCS)
16 -#define MAX_MRU (1536 + SKB_DMA_REALIGN)
17 -#define CNS3XXX_MAX_MTU (1536)
19 +#define RX_BUFFER_ALIGN 64
20 +#define RX_BUFFER_ALIGN_MASK (~(RX_BUFFER_ALIGN - 1))
22 +#define SKB_HEAD_ALIGN (((PAGE_SIZE - NET_SKB_PAD) % RX_BUFFER_ALIGN) + NET_SKB_PAD + NET_IP_ALIGN)
23 +#define RX_SEGMENT_ALLOC_SIZE 4096
24 +#define RX_SEGMENT_BUFSIZE (SKB_WITH_OVERHEAD(RX_SEGMENT_ALLOC_SIZE))
25 +#define RX_SEGMENT_MRU (((RX_SEGMENT_BUFSIZE - SKB_HEAD_ALIGN) & RX_BUFFER_ALIGN_MASK) - NET_IP_ALIGN)
28 #define NAPI_WEIGHT 64
30 @@ -266,7 +272,7 @@ struct _rx_ring {
33 struct rx_desc *cur_addr;
34 - struct sk_buff *buff_tab[RX_DESCS];
35 + void *buff_tab[RX_DESCS];
36 unsigned int phys_tab[RX_DESCS];
39 @@ -280,6 +286,8 @@ struct sw {
40 struct cns3xxx_plat_info *plat;
41 struct _tx_ring *tx_ring;
42 struct _rx_ring *rx_ring;
43 + struct sk_buff *frag_first;
44 + struct sk_buff *frag_last;
48 @@ -500,37 +508,35 @@ static void cns3xxx_alloc_rx_buf(struct
49 struct _rx_ring *rx_ring = sw->rx_ring;
50 unsigned int i = rx_ring->alloc_index;
51 struct rx_desc *desc = &(rx_ring)->desc[i];
52 - struct sk_buff *skb;
56 for (received += rx_ring->alloc_count; received > 0; received--) {
57 - if ((skb = dev_alloc_skb(MAX_MRU))) {
58 - if (SKB_DMA_REALIGN)
59 - skb_reserve(skb, SKB_DMA_REALIGN);
60 - skb_reserve(skb, NET_IP_ALIGN);
61 - phys = dma_map_single(NULL, skb->data,
62 - CNS3XXX_MAX_MTU, DMA_FROM_DEVICE);
63 - if (dma_mapping_error(NULL, phys)) {
65 - /* Failed to map, better luck next time */
70 - /* Failed to allocate skb, try again next time */
71 + buf = kzalloc(RX_SEGMENT_ALLOC_SIZE, GFP_ATOMIC);
75 + phys = dma_map_single(NULL, buf + SKB_HEAD_ALIGN,
76 + RX_SEGMENT_MRU, DMA_FROM_DEVICE);
77 + if (dma_mapping_error(NULL, phys)) {
82 + desc->sdl = RX_SEGMENT_MRU;
85 /* put the new buffer on RX-free queue */
86 - rx_ring->buff_tab[i] = skb;
87 + rx_ring->buff_tab[i] = buf;
88 rx_ring->phys_tab[i] = phys;
89 if (i == RX_DESCS - 1) {
91 desc->config0 = END_OF_RING | FIRST_SEGMENT |
92 - LAST_SEGMENT | CNS3XXX_MAX_MTU;
93 + LAST_SEGMENT | RX_SEGMENT_MRU;
94 desc = &(rx_ring)->desc[i];
96 - desc->config0 = FIRST_SEGMENT | LAST_SEGMENT | CNS3XXX_MAX_MTU;
97 + desc->config0 = FIRST_SEGMENT | LAST_SEGMENT |
102 @@ -588,49 +594,79 @@ static int eth_poll(struct napi_struct *
106 + int reserve = SKB_HEAD_ALIGN;
108 if (received >= budget)
111 - skb = rx_ring->buff_tab[i];
112 + /* process received frame */
113 + dma_unmap_single(&dev->dev, rx_ring->phys_tab[i],
114 + RX_SEGMENT_MRU, DMA_FROM_DEVICE);
116 + skb = build_skb(rx_ring->buff_tab[i]);
120 dev = switch_port_tab[desc->sp]->netdev;
123 - /* process received frame */
124 - dma_unmap_single(&dev->dev, rx_ring->phys_tab[i],
125 - length, DMA_FROM_DEVICE);
126 + if (desc->fsd && !desc->lsd)
127 + length = RX_SEGMENT_MRU;
130 + reserve -= NET_IP_ALIGN;
132 + length += NET_IP_ALIGN;
135 + skb_reserve(skb, reserve);
136 skb_put(skb, length);
139 - skb->protocol = eth_type_trans(skb, dev);
140 + if (!sw->frag_first)
141 + sw->frag_first = skb;
143 + if (sw->frag_first == sw->frag_last)
144 + skb_frag_add_head(sw->frag_first, skb);
146 + sw->frag_last->next = skb;
147 + sw->frag_first->len += skb->len;
148 + sw->frag_first->data_len += skb->len;
149 + sw->frag_first->truesize += skb->truesize;
151 + sw->frag_last = skb;
153 - dev->stats.rx_packets++;
154 - dev->stats.rx_bytes += length;
156 + skb = sw->frag_first;
157 + skb->protocol = eth_type_trans(skb, dev);
159 + dev->stats.rx_packets++;
160 + dev->stats.rx_bytes += skb->len;
162 + /* RX Hardware checksum offload */
163 + skb->ip_summed = CHECKSUM_NONE;
164 + switch (desc->prot) {
174 - /* RX Hardware checksum offload */
175 - switch (desc->prot) {
183 - skb->ip_summed = CHECKSUM_NONE;
185 skb->ip_summed = CHECKSUM_UNNECESSARY;
188 - skb->ip_summed = CHECKSUM_NONE;
196 - napi_gro_receive(napi, skb);
197 + napi_gro_receive(napi, skb);
200 + sw->frag_first = NULL;
201 + sw->frag_last = NULL;
205 if (++i == RX_DESCS) {
207 desc = &(rx_ring)->desc[i];
208 @@ -671,12 +707,6 @@ static int eth_xmit(struct sk_buff *skb,
212 - if (skb->len > CNS3XXX_MAX_MTU) {
213 - dev_kfree_skb(skb);
214 - dev->stats.tx_errors++;
215 - return NETDEV_TX_OK;
220 if ((tx_ring->num_used + nr_frags) >= TX_DESCS) {
221 @@ -701,8 +731,7 @@ static int eth_xmit(struct sk_buff *skb,
225 - phys = dma_map_single(NULL, skb->data, len,
227 + phys = dma_map_single(NULL, skb->data, len, DMA_TO_DEVICE);
230 tx_desc->pmap = pmap;
231 @@ -849,24 +878,24 @@ static int init_rings(struct sw *sw)
232 /* Setup RX buffers */
233 for (i = 0; i < RX_DESCS; i++) {
234 struct rx_desc *desc = &(rx_ring)->desc[i];
235 - struct sk_buff *skb;
236 - if (!(skb = dev_alloc_skb(MAX_MRU)))
239 + buf = kzalloc(RX_SEGMENT_ALLOC_SIZE, GFP_KERNEL);
242 - if (SKB_DMA_REALIGN)
243 - skb_reserve(skb, SKB_DMA_REALIGN);
244 - skb_reserve(skb, NET_IP_ALIGN);
245 - desc->sdl = CNS3XXX_MAX_MTU;
247 + desc->sdl = RX_SEGMENT_MRU;
248 if (i == (RX_DESCS - 1))
253 - desc->sdp = dma_map_single(NULL, skb->data,
254 - CNS3XXX_MAX_MTU, DMA_FROM_DEVICE);
255 - if (dma_mapping_error(NULL, desc->sdp)) {
256 + desc->sdp = dma_map_single(NULL, buf + SKB_HEAD_ALIGN,
257 + RX_SEGMENT_MRU, DMA_FROM_DEVICE);
258 + if (dma_mapping_error(NULL, desc->sdp))
261 - rx_ring->buff_tab[i] = skb;
263 + rx_ring->buff_tab[i] = buf;
264 rx_ring->phys_tab[i] = desc->sdp;
267 @@ -905,12 +934,13 @@ static void destroy_rings(struct sw *sw)
268 struct _rx_ring *rx_ring = sw->rx_ring;
269 struct rx_desc *desc = &(rx_ring)->desc[i];
270 struct sk_buff *skb = sw->rx_ring->buff_tab[i];
272 - dma_unmap_single(NULL,
274 - CNS3XXX_MAX_MTU, DMA_FROM_DEVICE);
275 - dev_kfree_skb(skb);
281 + dma_unmap_single(NULL, desc->sdp, RX_SEGMENT_MRU,
283 + dev_kfree_skb(skb);
285 dma_pool_free(rx_dma_pool, sw->rx_ring->desc, sw->rx_ring->phys_addr);
286 dma_pool_destroy(rx_dma_pool);
287 @@ -1085,13 +1115,22 @@ static int eth_set_mac(struct net_device
291 +static int cns3xxx_change_mtu(struct net_device *dev, int new_mtu)
293 + if (new_mtu > MAX_MTU)
296 + dev->mtu = new_mtu;
300 static const struct net_device_ops cns3xxx_netdev_ops = {
301 .ndo_open = eth_open,
302 .ndo_stop = eth_close,
303 .ndo_start_xmit = eth_xmit,
304 .ndo_set_rx_mode = eth_rx_mode,
305 .ndo_do_ioctl = eth_ioctl,
306 - .ndo_change_mtu = eth_change_mtu,
307 + .ndo_change_mtu = cns3xxx_change_mtu,
308 .ndo_set_mac_address = eth_set_mac,
309 .ndo_validate_addr = eth_validate_addr,
311 @@ -1124,6 +1163,10 @@ static int __devinit eth_init_one(struct
315 + temp = __raw_readl(&sw->regs->phy_auto_addr);
316 + temp |= (3 << 30); /* maximum frame length: 9600 bytes */
317 + __raw_writel(temp, &sw->regs->phy_auto_addr);
319 for (i = 0; i < 4; i++) {
320 temp = __raw_readl(&sw->regs->mac_cfg[i]);
321 temp |= (PORT_DISABLE);