mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-28 11:18:45 +07:00
3c32be635c
The patch splits core and dma parts for the mac10/100 device. This was already done for the GMAC device. It should make more flexible the driver to support other chips. Signed-off-by: Giuseppe Cavallaro <peppe.cavallaro@st.com> Signed-off-by: David S. Miller <davem@davemloft.net>
354 lines
9.5 KiB
C
354 lines
9.5 KiB
C
/*******************************************************************************
|
|
This is the driver for the MAC 10/100 on-chip Ethernet controller
|
|
currently tested on all the ST boards based on STb7109 and stx7200 SoCs.
|
|
|
|
DWC Ether MAC 10/100 Universal version 4.0 has been used for developing
|
|
this code.
|
|
|
|
This contains the functions to handle the dma and descriptors.
|
|
|
|
Copyright (C) 2007-2009 STMicroelectronics Ltd
|
|
|
|
This program is free software; you can redistribute it and/or modify it
|
|
under the terms and conditions of the GNU General Public License,
|
|
version 2, as published by the Free Software Foundation.
|
|
|
|
This program is distributed in the hope it will be useful, but WITHOUT
|
|
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
|
|
FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
|
|
more details.
|
|
|
|
You should have received a copy of the GNU General Public License along with
|
|
this program; if not, write to the Free Software Foundation, Inc.,
|
|
51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
|
|
|
|
The full GNU General Public License is included in this distribution in
|
|
the file called "COPYING".
|
|
|
|
Author: Giuseppe Cavallaro <peppe.cavallaro@st.com>
|
|
*******************************************************************************/
|
|
|
|
#include "dwmac100.h"
|
|
#include "dwmac_dma.h"
|
|
|
|
static int dwmac100_dma_init(unsigned long ioaddr, int pbl, u32 dma_tx,
|
|
u32 dma_rx)
|
|
{
|
|
u32 value = readl(ioaddr + DMA_BUS_MODE);
|
|
/* DMA SW reset */
|
|
value |= DMA_BUS_MODE_SFT_RESET;
|
|
writel(value, ioaddr + DMA_BUS_MODE);
|
|
do {} while ((readl(ioaddr + DMA_BUS_MODE) & DMA_BUS_MODE_SFT_RESET));
|
|
|
|
/* Enable Application Access by writing to DMA CSR0 */
|
|
writel(DMA_BUS_MODE_DEFAULT | (pbl << DMA_BUS_MODE_PBL_SHIFT),
|
|
ioaddr + DMA_BUS_MODE);
|
|
|
|
/* Mask interrupts by writing to CSR7 */
|
|
writel(DMA_INTR_DEFAULT_MASK, ioaddr + DMA_INTR_ENA);
|
|
|
|
/* The base address of the RX/TX descriptor lists must be written into
|
|
* DMA CSR3 and CSR4, respectively. */
|
|
writel(dma_tx, ioaddr + DMA_TX_BASE_ADDR);
|
|
writel(dma_rx, ioaddr + DMA_RCV_BASE_ADDR);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Store and Forward capability is not used at all..
|
|
* The transmit threshold can be programmed by
|
|
* setting the TTC bits in the DMA control register.*/
|
|
static void dwmac100_dma_operation_mode(unsigned long ioaddr, int txmode,
|
|
int rxmode)
|
|
{
|
|
u32 csr6 = readl(ioaddr + DMA_CONTROL);
|
|
|
|
if (txmode <= 32)
|
|
csr6 |= DMA_CONTROL_TTC_32;
|
|
else if (txmode <= 64)
|
|
csr6 |= DMA_CONTROL_TTC_64;
|
|
else
|
|
csr6 |= DMA_CONTROL_TTC_128;
|
|
|
|
writel(csr6, ioaddr + DMA_CONTROL);
|
|
|
|
return;
|
|
}
|
|
|
|
static void dwmac100_dump_dma_regs(unsigned long ioaddr)
|
|
{
|
|
int i;
|
|
|
|
DBG(KERN_DEBUG "DWMAC 100 DMA CSR\n");
|
|
for (i = 0; i < 9; i++)
|
|
pr_debug("\t CSR%d (offset 0x%x): 0x%08x\n", i,
|
|
(DMA_BUS_MODE + i * 4),
|
|
readl(ioaddr + DMA_BUS_MODE + i * 4));
|
|
DBG(KERN_DEBUG "\t CSR20 (offset 0x%x): 0x%08x\n",
|
|
DMA_CUR_TX_BUF_ADDR, readl(ioaddr + DMA_CUR_TX_BUF_ADDR));
|
|
DBG(KERN_DEBUG "\t CSR21 (offset 0x%x): 0x%08x\n",
|
|
DMA_CUR_RX_BUF_ADDR, readl(ioaddr + DMA_CUR_RX_BUF_ADDR));
|
|
return;
|
|
}
|
|
|
|
/* DMA controller has two counters to track the number of
|
|
* the receive missed frames. */
|
|
static void dwmac100_dma_diagnostic_fr(void *data, struct stmmac_extra_stats *x,
|
|
unsigned long ioaddr)
|
|
{
|
|
struct net_device_stats *stats = (struct net_device_stats *)data;
|
|
u32 csr8 = readl(ioaddr + DMA_MISSED_FRAME_CTR);
|
|
|
|
if (unlikely(csr8)) {
|
|
if (csr8 & DMA_MISSED_FRAME_OVE) {
|
|
stats->rx_over_errors += 0x800;
|
|
x->rx_overflow_cntr += 0x800;
|
|
} else {
|
|
unsigned int ove_cntr;
|
|
ove_cntr = ((csr8 & DMA_MISSED_FRAME_OVE_CNTR) >> 17);
|
|
stats->rx_over_errors += ove_cntr;
|
|
x->rx_overflow_cntr += ove_cntr;
|
|
}
|
|
|
|
if (csr8 & DMA_MISSED_FRAME_OVE_M) {
|
|
stats->rx_missed_errors += 0xffff;
|
|
x->rx_missed_cntr += 0xffff;
|
|
} else {
|
|
unsigned int miss_f = (csr8 & DMA_MISSED_FRAME_M_CNTR);
|
|
stats->rx_missed_errors += miss_f;
|
|
x->rx_missed_cntr += miss_f;
|
|
}
|
|
}
|
|
return;
|
|
}
|
|
|
|
static int dwmac100_get_tx_status(void *data, struct stmmac_extra_stats *x,
|
|
struct dma_desc *p, unsigned long ioaddr)
|
|
{
|
|
int ret = 0;
|
|
struct net_device_stats *stats = (struct net_device_stats *)data;
|
|
|
|
if (unlikely(p->des01.tx.error_summary)) {
|
|
if (unlikely(p->des01.tx.underflow_error)) {
|
|
x->tx_underflow++;
|
|
stats->tx_fifo_errors++;
|
|
}
|
|
if (unlikely(p->des01.tx.no_carrier)) {
|
|
x->tx_carrier++;
|
|
stats->tx_carrier_errors++;
|
|
}
|
|
if (unlikely(p->des01.tx.loss_carrier)) {
|
|
x->tx_losscarrier++;
|
|
stats->tx_carrier_errors++;
|
|
}
|
|
if (unlikely((p->des01.tx.excessive_deferral) ||
|
|
(p->des01.tx.excessive_collisions) ||
|
|
(p->des01.tx.late_collision)))
|
|
stats->collisions += p->des01.tx.collision_count;
|
|
ret = -1;
|
|
}
|
|
if (unlikely(p->des01.tx.heartbeat_fail)) {
|
|
x->tx_heartbeat++;
|
|
stats->tx_heartbeat_errors++;
|
|
ret = -1;
|
|
}
|
|
if (unlikely(p->des01.tx.deferred))
|
|
x->tx_deferred++;
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int dwmac100_get_tx_len(struct dma_desc *p)
|
|
{
|
|
return p->des01.tx.buffer1_size;
|
|
}
|
|
|
|
/* This function verifies if each incoming frame has some errors
|
|
* and, if required, updates the multicast statistics.
|
|
* In case of success, it returns csum_none becasue the device
|
|
* is not able to compute the csum in HW. */
|
|
static int dwmac100_get_rx_status(void *data, struct stmmac_extra_stats *x,
|
|
struct dma_desc *p)
|
|
{
|
|
int ret = csum_none;
|
|
struct net_device_stats *stats = (struct net_device_stats *)data;
|
|
|
|
if (unlikely(p->des01.rx.last_descriptor == 0)) {
|
|
pr_warning("dwmac100 Error: Oversized Ethernet "
|
|
"frame spanned multiple buffers\n");
|
|
stats->rx_length_errors++;
|
|
return discard_frame;
|
|
}
|
|
|
|
if (unlikely(p->des01.rx.error_summary)) {
|
|
if (unlikely(p->des01.rx.descriptor_error))
|
|
x->rx_desc++;
|
|
if (unlikely(p->des01.rx.partial_frame_error))
|
|
x->rx_partial++;
|
|
if (unlikely(p->des01.rx.run_frame))
|
|
x->rx_runt++;
|
|
if (unlikely(p->des01.rx.frame_too_long))
|
|
x->rx_toolong++;
|
|
if (unlikely(p->des01.rx.collision)) {
|
|
x->rx_collision++;
|
|
stats->collisions++;
|
|
}
|
|
if (unlikely(p->des01.rx.crc_error)) {
|
|
x->rx_crc++;
|
|
stats->rx_crc_errors++;
|
|
}
|
|
ret = discard_frame;
|
|
}
|
|
if (unlikely(p->des01.rx.dribbling))
|
|
ret = discard_frame;
|
|
|
|
if (unlikely(p->des01.rx.length_error)) {
|
|
x->rx_length++;
|
|
ret = discard_frame;
|
|
}
|
|
if (unlikely(p->des01.rx.mii_error)) {
|
|
x->rx_mii++;
|
|
ret = discard_frame;
|
|
}
|
|
if (p->des01.rx.multicast_frame) {
|
|
x->rx_multicast++;
|
|
stats->multicast++;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static void dwmac100_init_rx_desc(struct dma_desc *p, unsigned int ring_size,
|
|
int disable_rx_ic)
|
|
{
|
|
int i;
|
|
for (i = 0; i < ring_size; i++) {
|
|
p->des01.rx.own = 1;
|
|
p->des01.rx.buffer1_size = BUF_SIZE_2KiB - 1;
|
|
if (i == ring_size - 1)
|
|
p->des01.rx.end_ring = 1;
|
|
if (disable_rx_ic)
|
|
p->des01.rx.disable_ic = 1;
|
|
p++;
|
|
}
|
|
return;
|
|
}
|
|
|
|
static void dwmac100_init_tx_desc(struct dma_desc *p, unsigned int ring_size)
|
|
{
|
|
int i;
|
|
for (i = 0; i < ring_size; i++) {
|
|
p->des01.tx.own = 0;
|
|
if (i == ring_size - 1)
|
|
p->des01.tx.end_ring = 1;
|
|
p++;
|
|
}
|
|
return;
|
|
}
|
|
|
|
static int dwmac100_get_tx_owner(struct dma_desc *p)
|
|
{
|
|
return p->des01.tx.own;
|
|
}
|
|
|
|
static int dwmac100_get_rx_owner(struct dma_desc *p)
|
|
{
|
|
return p->des01.rx.own;
|
|
}
|
|
|
|
static void dwmac100_set_tx_owner(struct dma_desc *p)
|
|
{
|
|
p->des01.tx.own = 1;
|
|
}
|
|
|
|
static void dwmac100_set_rx_owner(struct dma_desc *p)
|
|
{
|
|
p->des01.rx.own = 1;
|
|
}
|
|
|
|
static int dwmac100_get_tx_ls(struct dma_desc *p)
|
|
{
|
|
return p->des01.tx.last_segment;
|
|
}
|
|
|
|
static void dwmac100_release_tx_desc(struct dma_desc *p)
|
|
{
|
|
int ter = p->des01.tx.end_ring;
|
|
|
|
/* clean field used within the xmit */
|
|
p->des01.tx.first_segment = 0;
|
|
p->des01.tx.last_segment = 0;
|
|
p->des01.tx.buffer1_size = 0;
|
|
|
|
/* clean status reported */
|
|
p->des01.tx.error_summary = 0;
|
|
p->des01.tx.underflow_error = 0;
|
|
p->des01.tx.no_carrier = 0;
|
|
p->des01.tx.loss_carrier = 0;
|
|
p->des01.tx.excessive_deferral = 0;
|
|
p->des01.tx.excessive_collisions = 0;
|
|
p->des01.tx.late_collision = 0;
|
|
p->des01.tx.heartbeat_fail = 0;
|
|
p->des01.tx.deferred = 0;
|
|
|
|
/* set termination field */
|
|
p->des01.tx.end_ring = ter;
|
|
|
|
return;
|
|
}
|
|
|
|
static void dwmac100_prepare_tx_desc(struct dma_desc *p, int is_fs, int len,
|
|
int csum_flag)
|
|
{
|
|
p->des01.tx.first_segment = is_fs;
|
|
p->des01.tx.buffer1_size = len;
|
|
}
|
|
|
|
static void dwmac100_clear_tx_ic(struct dma_desc *p)
|
|
{
|
|
p->des01.tx.interrupt = 0;
|
|
}
|
|
|
|
static void dwmac100_close_tx_desc(struct dma_desc *p)
|
|
{
|
|
p->des01.tx.last_segment = 1;
|
|
p->des01.tx.interrupt = 1;
|
|
}
|
|
|
|
static int dwmac100_get_rx_frame_len(struct dma_desc *p)
|
|
{
|
|
return p->des01.rx.frame_length;
|
|
}
|
|
|
|
struct stmmac_dma_ops dwmac100_dma_ops = {
|
|
.init = dwmac100_dma_init,
|
|
.dump_regs = dwmac100_dump_dma_regs,
|
|
.dma_mode = dwmac100_dma_operation_mode,
|
|
.dma_diagnostic_fr = dwmac100_dma_diagnostic_fr,
|
|
.enable_dma_transmission = dwmac_enable_dma_transmission,
|
|
.enable_dma_irq = dwmac_enable_dma_irq,
|
|
.disable_dma_irq = dwmac_disable_dma_irq,
|
|
.start_tx = dwmac_dma_start_tx,
|
|
.stop_tx = dwmac_dma_stop_tx,
|
|
.start_rx = dwmac_dma_start_rx,
|
|
.stop_rx = dwmac_dma_stop_rx,
|
|
.dma_interrupt = dwmac_dma_interrupt,
|
|
};
|
|
|
|
struct stmmac_desc_ops dwmac100_desc_ops = {
|
|
.tx_status = dwmac100_get_tx_status,
|
|
.rx_status = dwmac100_get_rx_status,
|
|
.get_tx_len = dwmac100_get_tx_len,
|
|
.init_rx_desc = dwmac100_init_rx_desc,
|
|
.init_tx_desc = dwmac100_init_tx_desc,
|
|
.get_tx_owner = dwmac100_get_tx_owner,
|
|
.get_rx_owner = dwmac100_get_rx_owner,
|
|
.release_tx_desc = dwmac100_release_tx_desc,
|
|
.prepare_tx_desc = dwmac100_prepare_tx_desc,
|
|
.clear_tx_ic = dwmac100_clear_tx_ic,
|
|
.close_tx_desc = dwmac100_close_tx_desc,
|
|
.get_tx_ls = dwmac100_get_tx_ls,
|
|
.set_tx_owner = dwmac100_set_tx_owner,
|
|
.set_rx_owner = dwmac100_set_rx_owner,
|
|
.get_rx_frame_len = dwmac100_get_rx_frame_len,
|
|
};
|