1 // SPDX-License-Identifier: (GPL-2.0 OR MIT)
3 * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates.
4 * stmmac XGMAC support.
7 #include <linux/iopoll.h>
11 static int dwxgmac2_dma_reset(void __iomem *ioaddr)
13 u32 value = readl(ioaddr + XGMAC_DMA_MODE);
16 writel(value | XGMAC_SWR, ioaddr + XGMAC_DMA_MODE);
18 return readl_poll_timeout(ioaddr + XGMAC_DMA_MODE, value,
19 !(value & XGMAC_SWR), 0, 100000);
22 static void dwxgmac2_dma_init(void __iomem *ioaddr,
23 struct stmmac_dma_cfg *dma_cfg, int atds)
25 u32 value = readl(ioaddr + XGMAC_DMA_SYSBUS_MODE);
30 writel(value | XGMAC_EAME, ioaddr + XGMAC_DMA_SYSBUS_MODE);
33 static void dwxgmac2_dma_init_chan(void __iomem *ioaddr,
34 struct stmmac_dma_cfg *dma_cfg, u32 chan)
36 u32 value = readl(ioaddr + XGMAC_DMA_CH_CONTROL(chan));
41 writel(value, ioaddr + XGMAC_DMA_CH_CONTROL(chan));
42 writel(XGMAC_DMA_INT_DEFAULT_EN, ioaddr + XGMAC_DMA_CH_INT_EN(chan));
45 static void dwxgmac2_dma_init_rx_chan(void __iomem *ioaddr,
46 struct stmmac_dma_cfg *dma_cfg,
47 dma_addr_t phy, u32 chan)
49 u32 rxpbl = dma_cfg->rxpbl ?: dma_cfg->pbl;
52 value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
53 value &= ~XGMAC_RxPBL;
54 value |= (rxpbl << XGMAC_RxPBL_SHIFT) & XGMAC_RxPBL;
55 writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
57 writel(upper_32_bits(phy), ioaddr + XGMAC_DMA_CH_RxDESC_HADDR(chan));
58 writel(lower_32_bits(phy), ioaddr + XGMAC_DMA_CH_RxDESC_LADDR(chan));
61 static void dwxgmac2_dma_init_tx_chan(void __iomem *ioaddr,
62 struct stmmac_dma_cfg *dma_cfg,
63 dma_addr_t phy, u32 chan)
65 u32 txpbl = dma_cfg->txpbl ?: dma_cfg->pbl;
68 value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
69 value &= ~XGMAC_TxPBL;
70 value |= (txpbl << XGMAC_TxPBL_SHIFT) & XGMAC_TxPBL;
72 writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
74 writel(upper_32_bits(phy), ioaddr + XGMAC_DMA_CH_TxDESC_HADDR(chan));
75 writel(lower_32_bits(phy), ioaddr + XGMAC_DMA_CH_TxDESC_LADDR(chan));
78 static void dwxgmac2_dma_axi(void __iomem *ioaddr, struct stmmac_axi *axi)
80 u32 value = readl(ioaddr + XGMAC_DMA_SYSBUS_MODE);
84 value |= XGMAC_EN_LPI;
86 value |= XGMAC_LPI_XIT_PKT;
88 value &= ~XGMAC_WR_OSR_LMT;
89 value |= (axi->axi_wr_osr_lmt << XGMAC_WR_OSR_LMT_SHIFT) &
92 value &= ~XGMAC_RD_OSR_LMT;
93 value |= (axi->axi_rd_osr_lmt << XGMAC_RD_OSR_LMT_SHIFT) &
100 for (i = 0; i < AXI_BLEN; i++) {
101 switch (axi->axi_blen[i]) {
103 value |= XGMAC_BLEN256;
106 value |= XGMAC_BLEN128;
109 value |= XGMAC_BLEN64;
112 value |= XGMAC_BLEN32;
115 value |= XGMAC_BLEN16;
118 value |= XGMAC_BLEN8;
121 value |= XGMAC_BLEN4;
126 writel(value, ioaddr + XGMAC_DMA_SYSBUS_MODE);
127 writel(XGMAC_TDPS, ioaddr + XGMAC_TX_EDMA_CTRL);
128 writel(XGMAC_RDPS, ioaddr + XGMAC_RX_EDMA_CTRL);
131 static void dwxgmac2_dma_rx_mode(void __iomem *ioaddr, int mode,
132 u32 channel, int fifosz, u8 qmode)
134 u32 value = readl(ioaddr + XGMAC_MTL_RXQ_OPMODE(channel));
135 unsigned int rqs = fifosz / 256 - 1;
137 if (mode == SF_DMA_MODE) {
144 value |= 0x0 << XGMAC_RTC_SHIFT;
146 value |= 0x2 << XGMAC_RTC_SHIFT;
148 value |= 0x3 << XGMAC_RTC_SHIFT;
152 value |= (rqs << XGMAC_RQS_SHIFT) & XGMAC_RQS;
154 if ((fifosz >= 4096) && (qmode != MTL_QUEUE_AVB)) {
155 u32 flow = readl(ioaddr + XGMAC_MTL_RXQ_FLOW_CONTROL(channel));
156 unsigned int rfd, rfa;
160 /* Set Threshold for Activating Flow Control to min 2 frames,
161 * i.e. 1500 * 2 = 3000 bytes.
163 * Set Threshold for Deactivating Flow Control to min 1 frame,
168 /* This violates the above formula because of FIFO size
169 * limit therefore overflow may occur in spite of this.
171 rfd = 0x03; /* Full-2.5K */
172 rfa = 0x01; /* Full-1.5K */
176 rfd = 0x06; /* Full-4K */
177 rfa = 0x0a; /* Full-6K */
181 rfd = 0x06; /* Full-4K */
182 rfa = 0x12; /* Full-10K */
186 rfd = 0x06; /* Full-4K */
187 rfa = 0x1e; /* Full-16K */
192 flow |= rfd << XGMAC_RFD_SHIFT;
195 flow |= rfa << XGMAC_RFA_SHIFT;
197 writel(flow, ioaddr + XGMAC_MTL_RXQ_FLOW_CONTROL(channel));
200 writel(value, ioaddr + XGMAC_MTL_RXQ_OPMODE(channel));
202 /* Enable MTL RX overflow */
203 value = readl(ioaddr + XGMAC_MTL_QINTEN(channel));
204 writel(value | XGMAC_RXOIE, ioaddr + XGMAC_MTL_QINTEN(channel));
207 static void dwxgmac2_dma_tx_mode(void __iomem *ioaddr, int mode,
208 u32 channel, int fifosz, u8 qmode)
210 u32 value = readl(ioaddr + XGMAC_MTL_TXQ_OPMODE(channel));
211 unsigned int tqs = fifosz / 256 - 1;
213 if (mode == SF_DMA_MODE) {
220 value |= 0x0 << XGMAC_TTC_SHIFT;
222 value |= 0x2 << XGMAC_TTC_SHIFT;
223 else if (mode <= 128)
224 value |= 0x3 << XGMAC_TTC_SHIFT;
225 else if (mode <= 192)
226 value |= 0x4 << XGMAC_TTC_SHIFT;
227 else if (mode <= 256)
228 value |= 0x5 << XGMAC_TTC_SHIFT;
229 else if (mode <= 384)
230 value |= 0x6 << XGMAC_TTC_SHIFT;
232 value |= 0x7 << XGMAC_TTC_SHIFT;
235 /* Use static TC to Queue mapping */
236 value |= (channel << XGMAC_Q2TCMAP_SHIFT) & XGMAC_Q2TCMAP;
238 value &= ~XGMAC_TXQEN;
239 if (qmode != MTL_QUEUE_AVB)
240 value |= 0x2 << XGMAC_TXQEN_SHIFT;
242 value |= 0x1 << XGMAC_TXQEN_SHIFT;
245 value |= (tqs << XGMAC_TQS_SHIFT) & XGMAC_TQS;
247 writel(value, ioaddr + XGMAC_MTL_TXQ_OPMODE(channel));
250 static void dwxgmac2_enable_dma_irq(void __iomem *ioaddr, u32 chan)
252 writel(XGMAC_DMA_INT_DEFAULT_EN, ioaddr + XGMAC_DMA_CH_INT_EN(chan));
255 static void dwxgmac2_disable_dma_irq(void __iomem *ioaddr, u32 chan)
257 writel(0, ioaddr + XGMAC_DMA_CH_INT_EN(chan));
260 static void dwxgmac2_dma_start_tx(void __iomem *ioaddr, u32 chan)
264 value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
266 writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
268 value = readl(ioaddr + XGMAC_TX_CONFIG);
269 value |= XGMAC_CONFIG_TE;
270 writel(value, ioaddr + XGMAC_TX_CONFIG);
273 static void dwxgmac2_dma_stop_tx(void __iomem *ioaddr, u32 chan)
277 value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
278 value &= ~XGMAC_TXST;
279 writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
281 value = readl(ioaddr + XGMAC_TX_CONFIG);
282 value &= ~XGMAC_CONFIG_TE;
283 writel(value, ioaddr + XGMAC_TX_CONFIG);
286 static void dwxgmac2_dma_start_rx(void __iomem *ioaddr, u32 chan)
290 value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
292 writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
294 value = readl(ioaddr + XGMAC_RX_CONFIG);
295 value |= XGMAC_CONFIG_RE;
296 writel(value, ioaddr + XGMAC_RX_CONFIG);
299 static void dwxgmac2_dma_stop_rx(void __iomem *ioaddr, u32 chan)
303 value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
304 value &= ~XGMAC_RXST;
305 writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
308 static int dwxgmac2_dma_interrupt(void __iomem *ioaddr,
309 struct stmmac_extra_stats *x, u32 chan)
311 u32 intr_status = readl(ioaddr + XGMAC_DMA_CH_STATUS(chan));
312 u32 intr_en = readl(ioaddr + XGMAC_DMA_CH_INT_EN(chan));
315 /* ABNORMAL interrupts */
316 if (unlikely(intr_status & XGMAC_AIS)) {
317 if (unlikely(intr_status & XGMAC_TPS)) {
318 x->tx_process_stopped_irq++;
319 ret |= tx_hard_error;
321 if (unlikely(intr_status & XGMAC_FBE)) {
322 x->fatal_bus_error_irq++;
323 ret |= tx_hard_error;
327 /* TX/RX NORMAL interrupts */
328 if (likely(intr_status & XGMAC_NIS)) {
331 if (likely(intr_status & XGMAC_RI)) {
332 x->rx_normal_irq_n++;
335 if (likely(intr_status & (XGMAC_TI | XGMAC_TBU))) {
336 x->tx_normal_irq_n++;
341 /* Clear interrupts */
342 writel(intr_en & intr_status, ioaddr + XGMAC_DMA_CH_STATUS(chan));
347 static void dwxgmac2_get_hw_feature(void __iomem *ioaddr,
348 struct dma_features *dma_cap)
352 /* MAC HW feature 0 */
353 hw_cap = readl(ioaddr + XGMAC_HW_FEATURE0);
354 dma_cap->rx_coe = (hw_cap & XGMAC_HWFEAT_RXCOESEL) >> 16;
355 dma_cap->tx_coe = (hw_cap & XGMAC_HWFEAT_TXCOESEL) >> 14;
356 dma_cap->atime_stamp = (hw_cap & XGMAC_HWFEAT_TSSEL) >> 12;
357 dma_cap->av = (hw_cap & XGMAC_HWFEAT_AVSEL) >> 11;
358 dma_cap->av &= (hw_cap & XGMAC_HWFEAT_RAVSEL) >> 10;
359 dma_cap->pmt_magic_frame = (hw_cap & XGMAC_HWFEAT_MGKSEL) >> 7;
360 dma_cap->pmt_remote_wake_up = (hw_cap & XGMAC_HWFEAT_RWKSEL) >> 6;
361 dma_cap->mbps_1000 = (hw_cap & XGMAC_HWFEAT_GMIISEL) >> 1;
363 /* MAC HW feature 1 */
364 hw_cap = readl(ioaddr + XGMAC_HW_FEATURE1);
365 dma_cap->tsoen = (hw_cap & XGMAC_HWFEAT_TSOEN) >> 18;
367 dma_cap->addr64 = (hw_cap & XGMAC_HWFEAT_ADDR64) >> 14;
368 switch (dma_cap->addr64) {
370 dma_cap->addr64 = 32;
373 dma_cap->addr64 = 40;
376 dma_cap->addr64 = 48;
379 dma_cap->addr64 = 32;
383 dma_cap->tx_fifo_size =
384 128 << ((hw_cap & XGMAC_HWFEAT_TXFIFOSIZE) >> 6);
385 dma_cap->rx_fifo_size =
386 128 << ((hw_cap & XGMAC_HWFEAT_RXFIFOSIZE) >> 0);
388 /* MAC HW feature 2 */
389 hw_cap = readl(ioaddr + XGMAC_HW_FEATURE2);
390 dma_cap->pps_out_num = (hw_cap & XGMAC_HWFEAT_PPSOUTNUM) >> 24;
391 dma_cap->number_tx_channel =
392 ((hw_cap & XGMAC_HWFEAT_TXCHCNT) >> 18) + 1;
393 dma_cap->number_rx_channel =
394 ((hw_cap & XGMAC_HWFEAT_RXCHCNT) >> 12) + 1;
395 dma_cap->number_tx_queues =
396 ((hw_cap & XGMAC_HWFEAT_TXQCNT) >> 6) + 1;
397 dma_cap->number_rx_queues =
398 ((hw_cap & XGMAC_HWFEAT_RXQCNT) >> 0) + 1;
401 static void dwxgmac2_rx_watchdog(void __iomem *ioaddr, u32 riwt, u32 nchan)
405 for (i = 0; i < nchan; i++)
406 writel(riwt & XGMAC_RWT, ioaddr + XGMAC_DMA_CH_Rx_WATCHDOG(i));
409 static void dwxgmac2_set_rx_ring_len(void __iomem *ioaddr, u32 len, u32 chan)
411 writel(len, ioaddr + XGMAC_DMA_CH_RxDESC_RING_LEN(chan));
414 static void dwxgmac2_set_tx_ring_len(void __iomem *ioaddr, u32 len, u32 chan)
416 writel(len, ioaddr + XGMAC_DMA_CH_TxDESC_RING_LEN(chan));
419 static void dwxgmac2_set_rx_tail_ptr(void __iomem *ioaddr, u32 ptr, u32 chan)
421 writel(ptr, ioaddr + XGMAC_DMA_CH_RxDESC_TAIL_LPTR(chan));
424 static void dwxgmac2_set_tx_tail_ptr(void __iomem *ioaddr, u32 ptr, u32 chan)
426 writel(ptr, ioaddr + XGMAC_DMA_CH_TxDESC_TAIL_LPTR(chan));
429 static void dwxgmac2_enable_tso(void __iomem *ioaddr, bool en, u32 chan)
431 u32 value = readl(ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
438 writel(value, ioaddr + XGMAC_DMA_CH_TX_CONTROL(chan));
441 static void dwxgmac2_qmode(void __iomem *ioaddr, u32 channel, u8 qmode)
443 u32 value = readl(ioaddr + XGMAC_MTL_TXQ_OPMODE(channel));
445 value &= ~XGMAC_TXQEN;
446 if (qmode != MTL_QUEUE_AVB) {
447 value |= 0x2 << XGMAC_TXQEN_SHIFT;
448 writel(0, ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(channel));
450 value |= 0x1 << XGMAC_TXQEN_SHIFT;
453 writel(value, ioaddr + XGMAC_MTL_TXQ_OPMODE(channel));
456 static void dwxgmac2_set_bfsize(void __iomem *ioaddr, int bfsize, u32 chan)
460 value = readl(ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
461 value |= bfsize << 1;
462 writel(value, ioaddr + XGMAC_DMA_CH_RX_CONTROL(chan));
465 const struct stmmac_dma_ops dwxgmac210_dma_ops = {
466 .reset = dwxgmac2_dma_reset,
467 .init = dwxgmac2_dma_init,
468 .init_chan = dwxgmac2_dma_init_chan,
469 .init_rx_chan = dwxgmac2_dma_init_rx_chan,
470 .init_tx_chan = dwxgmac2_dma_init_tx_chan,
471 .axi = dwxgmac2_dma_axi,
473 .dma_rx_mode = dwxgmac2_dma_rx_mode,
474 .dma_tx_mode = dwxgmac2_dma_tx_mode,
475 .enable_dma_irq = dwxgmac2_enable_dma_irq,
476 .disable_dma_irq = dwxgmac2_disable_dma_irq,
477 .start_tx = dwxgmac2_dma_start_tx,
478 .stop_tx = dwxgmac2_dma_stop_tx,
479 .start_rx = dwxgmac2_dma_start_rx,
480 .stop_rx = dwxgmac2_dma_stop_rx,
481 .dma_interrupt = dwxgmac2_dma_interrupt,
482 .get_hw_feature = dwxgmac2_get_hw_feature,
483 .rx_watchdog = dwxgmac2_rx_watchdog,
484 .set_rx_ring_len = dwxgmac2_set_rx_ring_len,
485 .set_tx_ring_len = dwxgmac2_set_tx_ring_len,
486 .set_rx_tail_ptr = dwxgmac2_set_rx_tail_ptr,
487 .set_tx_tail_ptr = dwxgmac2_set_tx_tail_ptr,
488 .enable_tso = dwxgmac2_enable_tso,
489 .qmode = dwxgmac2_qmode,
490 .set_bfsize = dwxgmac2_set_bfsize,