/branches/gl-inet/target/linux/ar71xx/files/drivers/net/ethernet/atheros/ag71xx/ag71xx.h |
@@ -0,0 +1,506 @@ |
/* |
* Atheros AR71xx built-in ethernet mac driver |
* |
* Copyright (C) 2008-2010 Gabor Juhos <juhosg@openwrt.org> |
* Copyright (C) 2008 Imre Kaloz <kaloz@openwrt.org> |
* |
* Based on Atheros' AG7100 driver |
* |
* This program is free software; you can redistribute it and/or modify it |
* under the terms of the GNU General Public License version 2 as published |
* by the Free Software Foundation. |
*/ |
|
#ifndef __AG71XX_H |
#define __AG71XX_H |
|
#include <linux/kernel.h> |
#include <linux/version.h> |
#include <linux/module.h> |
#include <linux/init.h> |
#include <linux/types.h> |
#include <linux/random.h> |
#include <linux/spinlock.h> |
#include <linux/interrupt.h> |
#include <linux/platform_device.h> |
#include <linux/ethtool.h> |
#include <linux/etherdevice.h> |
#include <linux/if_vlan.h> |
#include <linux/phy.h> |
#include <linux/skbuff.h> |
#include <linux/dma-mapping.h> |
#include <linux/workqueue.h> |
|
#include <linux/bitops.h> |
|
#include <asm/mach-ath79/ar71xx_regs.h> |
#include <asm/mach-ath79/ath79.h> |
#include <asm/mach-ath79/ag71xx_platform.h> |
|
#define AG71XX_DRV_NAME "ag71xx" |
#define AG71XX_DRV_VERSION "0.5.35" |
|
/* |
* For our NAPI weight bigger does *NOT* mean better - it means more |
* D-cache misses and lots more wasted cycles than we'll ever |
* possibly gain from saving instructions. |
*/ |
#define AG71XX_NAPI_WEIGHT 32 |
#define AG71XX_OOM_REFILL (1 + HZ/10) |
|
#define AG71XX_INT_ERR (AG71XX_INT_RX_BE | AG71XX_INT_TX_BE) |
#define AG71XX_INT_TX (AG71XX_INT_TX_PS) |
#define AG71XX_INT_RX (AG71XX_INT_RX_PR | AG71XX_INT_RX_OF) |
|
#define AG71XX_INT_POLL (AG71XX_INT_RX | AG71XX_INT_TX) |
#define AG71XX_INT_INIT (AG71XX_INT_ERR | AG71XX_INT_POLL) |
|
#define AG71XX_TX_MTU_LEN 1540 |
|
#define AG71XX_TX_RING_SPLIT 512 |
#define AG71XX_TX_RING_DS_PER_PKT DIV_ROUND_UP(AG71XX_TX_MTU_LEN, \ |
AG71XX_TX_RING_SPLIT) |
#define AG71XX_TX_RING_SIZE_DEFAULT 128 |
#define AG71XX_RX_RING_SIZE_DEFAULT 256 |
|
#define AG71XX_TX_RING_SIZE_MAX 128 |
#define AG71XX_RX_RING_SIZE_MAX 256 |
|
#ifdef CONFIG_AG71XX_DEBUG |
#define DBG(fmt, args...) pr_debug(fmt, ## args) |
#else |
#define DBG(fmt, args...) do {} while (0) |
#endif |
|
#define ag71xx_assert(_cond) \ |
do { \ |
if (_cond) \ |
break; \ |
printk("%s,%d: assertion failed\n", __FILE__, __LINE__); \ |
BUG(); \ |
} while (0) |
|
struct ag71xx_desc { |
u32 data; |
u32 ctrl; |
#define DESC_EMPTY BIT(31) |
#define DESC_MORE BIT(24) |
#define DESC_PKTLEN_M 0xfff |
u32 next; |
u32 pad; |
} __attribute__((aligned(4))); |
|
#define AG71XX_DESC_SIZE roundup(sizeof(struct ag71xx_desc), \ |
L1_CACHE_BYTES) |
|
struct ag71xx_buf { |
union { |
struct sk_buff *skb; |
void *rx_buf; |
}; |
union { |
dma_addr_t dma_addr; |
unsigned int len; |
}; |
}; |
|
struct ag71xx_ring { |
struct ag71xx_buf *buf; |
u8 *descs_cpu; |
dma_addr_t descs_dma; |
u16 desc_split; |
u16 order; |
unsigned int curr; |
unsigned int dirty; |
}; |
|
struct ag71xx_mdio { |
struct mii_bus *mii_bus; |
#if LINUX_VERSION_CODE < KERNEL_VERSION(4,5,0) |
int mii_irq[PHY_MAX_ADDR]; |
#endif |
void __iomem *mdio_base; |
struct ag71xx_mdio_platform_data *pdata; |
}; |
|
struct ag71xx_int_stats { |
unsigned long rx_pr; |
unsigned long rx_be; |
unsigned long rx_of; |
unsigned long tx_ps; |
unsigned long tx_be; |
unsigned long tx_ur; |
unsigned long total; |
}; |
|
struct ag71xx_napi_stats { |
unsigned long napi_calls; |
unsigned long rx_count; |
unsigned long rx_packets; |
unsigned long rx_packets_max; |
unsigned long tx_count; |
unsigned long tx_packets; |
unsigned long tx_packets_max; |
|
unsigned long rx[AG71XX_NAPI_WEIGHT + 1]; |
unsigned long tx[AG71XX_NAPI_WEIGHT + 1]; |
}; |
|
struct ag71xx_debug { |
struct dentry *debugfs_dir; |
|
struct ag71xx_int_stats int_stats; |
struct ag71xx_napi_stats napi_stats; |
}; |
|
struct ag71xx { |
/* |
* Critical data related to the per-packet data path are clustered |
* early in this structure to help improve the D-cache footprint. |
*/ |
struct ag71xx_ring rx_ring ____cacheline_aligned; |
struct ag71xx_ring tx_ring ____cacheline_aligned; |
|
unsigned int max_frame_len; |
unsigned int desc_pktlen_mask; |
unsigned int rx_buf_size; |
|
struct net_device *dev; |
struct platform_device *pdev; |
spinlock_t lock; |
struct napi_struct napi; |
u32 msg_enable; |
|
/* |
* From this point onwards we're not looking at per-packet fields. |
*/ |
void __iomem *mac_base; |
|
struct ag71xx_desc *stop_desc; |
dma_addr_t stop_desc_dma; |
|
struct mii_bus *mii_bus; |
struct phy_device *phy_dev; |
void *phy_priv; |
|
unsigned int link; |
unsigned int speed; |
int duplex; |
|
struct delayed_work restart_work; |
struct delayed_work link_work; |
struct timer_list oom_timer; |
|
#ifdef CONFIG_AG71XX_DEBUG_FS |
struct ag71xx_debug debug; |
#endif |
}; |
|
extern struct ethtool_ops ag71xx_ethtool_ops; |
void ag71xx_link_adjust(struct ag71xx *ag); |
|
int ag71xx_mdio_driver_init(void) __init; |
void ag71xx_mdio_driver_exit(void); |
|
int ag71xx_phy_connect(struct ag71xx *ag); |
void ag71xx_phy_disconnect(struct ag71xx *ag); |
void ag71xx_phy_start(struct ag71xx *ag); |
void ag71xx_phy_stop(struct ag71xx *ag); |
|
static inline struct ag71xx_platform_data *ag71xx_get_pdata(struct ag71xx *ag) |
{ |
return ag->pdev->dev.platform_data; |
} |
|
static inline int ag71xx_desc_empty(struct ag71xx_desc *desc) |
{ |
return (desc->ctrl & DESC_EMPTY) != 0; |
} |
|
static inline struct ag71xx_desc * |
ag71xx_ring_desc(struct ag71xx_ring *ring, int idx) |
{ |
return (struct ag71xx_desc *) &ring->descs_cpu[idx * AG71XX_DESC_SIZE]; |
} |
|
static inline int |
ag71xx_ring_size_order(int size) |
{ |
return fls(size - 1); |
} |
|
/* Register offsets */ |
#define AG71XX_REG_MAC_CFG1 0x0000 |
#define AG71XX_REG_MAC_CFG2 0x0004 |
#define AG71XX_REG_MAC_IPG 0x0008 |
#define AG71XX_REG_MAC_HDX 0x000c |
#define AG71XX_REG_MAC_MFL 0x0010 |
#define AG71XX_REG_MII_CFG 0x0020 |
#define AG71XX_REG_MII_CMD 0x0024 |
#define AG71XX_REG_MII_ADDR 0x0028 |
#define AG71XX_REG_MII_CTRL 0x002c |
#define AG71XX_REG_MII_STATUS 0x0030 |
#define AG71XX_REG_MII_IND 0x0034 |
#define AG71XX_REG_MAC_IFCTL 0x0038 |
#define AG71XX_REG_MAC_ADDR1 0x0040 |
#define AG71XX_REG_MAC_ADDR2 0x0044 |
#define AG71XX_REG_FIFO_CFG0 0x0048 |
#define AG71XX_REG_FIFO_CFG1 0x004c |
#define AG71XX_REG_FIFO_CFG2 0x0050 |
#define AG71XX_REG_FIFO_CFG3 0x0054 |
#define AG71XX_REG_FIFO_CFG4 0x0058 |
#define AG71XX_REG_FIFO_CFG5 0x005c |
#define AG71XX_REG_FIFO_RAM0 0x0060 |
#define AG71XX_REG_FIFO_RAM1 0x0064 |
#define AG71XX_REG_FIFO_RAM2 0x0068 |
#define AG71XX_REG_FIFO_RAM3 0x006c |
#define AG71XX_REG_FIFO_RAM4 0x0070 |
#define AG71XX_REG_FIFO_RAM5 0x0074 |
#define AG71XX_REG_FIFO_RAM6 0x0078 |
#define AG71XX_REG_FIFO_RAM7 0x007c |
|
#define AG71XX_REG_TX_CTRL 0x0180 |
#define AG71XX_REG_TX_DESC 0x0184 |
#define AG71XX_REG_TX_STATUS 0x0188 |
#define AG71XX_REG_RX_CTRL 0x018c |
#define AG71XX_REG_RX_DESC 0x0190 |
#define AG71XX_REG_RX_STATUS 0x0194 |
#define AG71XX_REG_INT_ENABLE 0x0198 |
#define AG71XX_REG_INT_STATUS 0x019c |
|
#define AG71XX_REG_FIFO_DEPTH 0x01a8 |
#define AG71XX_REG_RX_SM 0x01b0 |
#define AG71XX_REG_TX_SM 0x01b4 |
|
#define MAC_CFG1_TXE BIT(0) /* Tx Enable */ |
#define MAC_CFG1_STX BIT(1) /* Synchronize Tx Enable */ |
#define MAC_CFG1_RXE BIT(2) /* Rx Enable */ |
#define MAC_CFG1_SRX BIT(3) /* Synchronize Rx Enable */ |
#define MAC_CFG1_TFC BIT(4) /* Tx Flow Control Enable */ |
#define MAC_CFG1_RFC BIT(5) /* Rx Flow Control Enable */ |
#define MAC_CFG1_LB BIT(8) /* Loopback mode */ |
#define MAC_CFG1_SR BIT(31) /* Soft Reset */ |
|
#define MAC_CFG2_FDX BIT(0) |
#define MAC_CFG2_CRC_EN BIT(1) |
#define MAC_CFG2_PAD_CRC_EN BIT(2) |
#define MAC_CFG2_LEN_CHECK BIT(4) |
#define MAC_CFG2_HUGE_FRAME_EN BIT(5) |
#define MAC_CFG2_IF_1000 BIT(9) |
#define MAC_CFG2_IF_10_100 BIT(8) |
|
#define FIFO_CFG0_WTM BIT(0) /* Watermark Module */ |
#define FIFO_CFG0_RXS BIT(1) /* Rx System Module */ |
#define FIFO_CFG0_RXF BIT(2) /* Rx Fabric Module */ |
#define FIFO_CFG0_TXS BIT(3) /* Tx System Module */ |
#define FIFO_CFG0_TXF BIT(4) /* Tx Fabric Module */ |
#define FIFO_CFG0_ALL (FIFO_CFG0_WTM | FIFO_CFG0_RXS | FIFO_CFG0_RXF \ |
| FIFO_CFG0_TXS | FIFO_CFG0_TXF) |
|
#define FIFO_CFG0_ENABLE_SHIFT 8 |
|
#define FIFO_CFG4_DE BIT(0) /* Drop Event */ |
#define FIFO_CFG4_DV BIT(1) /* RX_DV Event */ |
#define FIFO_CFG4_FC BIT(2) /* False Carrier */ |
#define FIFO_CFG4_CE BIT(3) /* Code Error */ |
#define FIFO_CFG4_CR BIT(4) /* CRC error */ |
#define FIFO_CFG4_LM BIT(5) /* Length Mismatch */ |
#define FIFO_CFG4_LO BIT(6) /* Length out of range */ |
#define FIFO_CFG4_OK BIT(7) /* Packet is OK */ |
#define FIFO_CFG4_MC BIT(8) /* Multicast Packet */ |
#define FIFO_CFG4_BC BIT(9) /* Broadcast Packet */ |
#define FIFO_CFG4_DR BIT(10) /* Dribble */ |
#define FIFO_CFG4_LE BIT(11) /* Long Event */ |
#define FIFO_CFG4_CF BIT(12) /* Control Frame */ |
#define FIFO_CFG4_PF BIT(13) /* Pause Frame */ |
#define FIFO_CFG4_UO BIT(14) /* Unsupported Opcode */ |
#define FIFO_CFG4_VT BIT(15) /* VLAN tag detected */ |
#define FIFO_CFG4_FT BIT(16) /* Frame Truncated */ |
#define FIFO_CFG4_UC BIT(17) /* Unicast Packet */ |
|
#define FIFO_CFG5_DE BIT(0) /* Drop Event */ |
#define FIFO_CFG5_DV BIT(1) /* RX_DV Event */ |
#define FIFO_CFG5_FC BIT(2) /* False Carrier */ |
#define FIFO_CFG5_CE BIT(3) /* Code Error */ |
#define FIFO_CFG5_LM BIT(4) /* Length Mismatch */ |
#define FIFO_CFG5_LO BIT(5) /* Length Out of Range */ |
#define FIFO_CFG5_OK BIT(6) /* Packet is OK */ |
#define FIFO_CFG5_MC BIT(7) /* Multicast Packet */ |
#define FIFO_CFG5_BC BIT(8) /* Broadcast Packet */ |
#define FIFO_CFG5_DR BIT(9) /* Dribble */ |
#define FIFO_CFG5_CF BIT(10) /* Control Frame */ |
#define FIFO_CFG5_PF BIT(11) /* Pause Frame */ |
#define FIFO_CFG5_UO BIT(12) /* Unsupported Opcode */ |
#define FIFO_CFG5_VT BIT(13) /* VLAN tag detected */ |
#define FIFO_CFG5_LE BIT(14) /* Long Event */ |
#define FIFO_CFG5_FT BIT(15) /* Frame Truncated */ |
#define FIFO_CFG5_16 BIT(16) /* unknown */ |
#define FIFO_CFG5_17 BIT(17) /* unknown */ |
#define FIFO_CFG5_SF BIT(18) /* Short Frame */ |
#define FIFO_CFG5_BM BIT(19) /* Byte Mode */ |
|
#define AG71XX_INT_TX_PS BIT(0) |
#define AG71XX_INT_TX_UR BIT(1) |
#define AG71XX_INT_TX_BE BIT(3) |
#define AG71XX_INT_RX_PR BIT(4) |
#define AG71XX_INT_RX_OF BIT(6) |
#define AG71XX_INT_RX_BE BIT(7) |
|
#define MAC_IFCTL_SPEED BIT(16) |
|
#define MII_CFG_CLK_DIV_4 0 |
#define MII_CFG_CLK_DIV_6 2 |
#define MII_CFG_CLK_DIV_8 3 |
#define MII_CFG_CLK_DIV_10 4 |
#define MII_CFG_CLK_DIV_14 5 |
#define MII_CFG_CLK_DIV_20 6 |
#define MII_CFG_CLK_DIV_28 7 |
#define MII_CFG_CLK_DIV_34 8 |
#define MII_CFG_CLK_DIV_42 9 |
#define MII_CFG_CLK_DIV_50 10 |
#define MII_CFG_CLK_DIV_58 11 |
#define MII_CFG_CLK_DIV_66 12 |
#define MII_CFG_CLK_DIV_74 13 |
#define MII_CFG_CLK_DIV_82 14 |
#define MII_CFG_CLK_DIV_98 15 |
#define MII_CFG_RESET BIT(31) |
|
#define MII_CMD_WRITE 0x0 |
#define MII_CMD_READ 0x1 |
#define MII_ADDR_SHIFT 8 |
#define MII_IND_BUSY BIT(0) |
#define MII_IND_INVALID BIT(2) |
|
#define TX_CTRL_TXE BIT(0) /* Tx Enable */ |
|
#define TX_STATUS_PS BIT(0) /* Packet Sent */ |
#define TX_STATUS_UR BIT(1) /* Tx Underrun */ |
#define TX_STATUS_BE BIT(3) /* Bus Error */ |
|
#define RX_CTRL_RXE BIT(0) /* Rx Enable */ |
|
#define RX_STATUS_PR BIT(0) /* Packet Received */ |
#define RX_STATUS_OF BIT(2) /* Rx Overflow */ |
#define RX_STATUS_BE BIT(3) /* Bus Error */ |
|
static inline void ag71xx_check_reg_offset(struct ag71xx *ag, unsigned reg) |
{ |
switch (reg) { |
case AG71XX_REG_MAC_CFG1 ... AG71XX_REG_MAC_MFL: |
case AG71XX_REG_MAC_IFCTL ... AG71XX_REG_TX_SM: |
case AG71XX_REG_MII_CFG: |
break; |
|
default: |
BUG(); |
} |
} |
|
static inline void ag71xx_wr(struct ag71xx *ag, unsigned reg, u32 value) |
{ |
ag71xx_check_reg_offset(ag, reg); |
|
__raw_writel(value, ag->mac_base + reg); |
/* flush write */ |
(void) __raw_readl(ag->mac_base + reg); |
} |
|
static inline u32 ag71xx_rr(struct ag71xx *ag, unsigned reg) |
{ |
ag71xx_check_reg_offset(ag, reg); |
|
return __raw_readl(ag->mac_base + reg); |
} |
|
static inline void ag71xx_sb(struct ag71xx *ag, unsigned reg, u32 mask) |
{ |
void __iomem *r; |
|
ag71xx_check_reg_offset(ag, reg); |
|
r = ag->mac_base + reg; |
__raw_writel(__raw_readl(r) | mask, r); |
/* flush write */ |
(void)__raw_readl(r); |
} |
|
static inline void ag71xx_cb(struct ag71xx *ag, unsigned reg, u32 mask) |
{ |
void __iomem *r; |
|
ag71xx_check_reg_offset(ag, reg); |
|
r = ag->mac_base + reg; |
__raw_writel(__raw_readl(r) & ~mask, r); |
/* flush write */ |
(void) __raw_readl(r); |
} |
|
static inline void ag71xx_int_enable(struct ag71xx *ag, u32 ints) |
{ |
ag71xx_sb(ag, AG71XX_REG_INT_ENABLE, ints); |
} |
|
static inline void ag71xx_int_disable(struct ag71xx *ag, u32 ints) |
{ |
ag71xx_cb(ag, AG71XX_REG_INT_ENABLE, ints); |
} |
|
#ifdef CONFIG_AG71XX_AR8216_SUPPORT |
void ag71xx_add_ar8216_header(struct ag71xx *ag, struct sk_buff *skb); |
int ag71xx_remove_ar8216_header(struct ag71xx *ag, struct sk_buff *skb, |
int pktlen); |
static inline int ag71xx_has_ar8216(struct ag71xx *ag) |
{ |
return ag71xx_get_pdata(ag)->has_ar8216; |
} |
#else |
static inline void ag71xx_add_ar8216_header(struct ag71xx *ag, |
struct sk_buff *skb) |
{ |
} |
|
static inline int ag71xx_remove_ar8216_header(struct ag71xx *ag, |
struct sk_buff *skb, |
int pktlen) |
{ |
return 0; |
} |
static inline int ag71xx_has_ar8216(struct ag71xx *ag) |
{ |
return 0; |
} |
#endif |
|
#ifdef CONFIG_AG71XX_DEBUG_FS |
int ag71xx_debugfs_root_init(void); |
void ag71xx_debugfs_root_exit(void); |
int ag71xx_debugfs_init(struct ag71xx *ag); |
void ag71xx_debugfs_exit(struct ag71xx *ag); |
void ag71xx_debugfs_update_int_stats(struct ag71xx *ag, u32 status); |
void ag71xx_debugfs_update_napi_stats(struct ag71xx *ag, int rx, int tx); |
#else |
static inline int ag71xx_debugfs_root_init(void) { return 0; } |
static inline void ag71xx_debugfs_root_exit(void) {} |
static inline int ag71xx_debugfs_init(struct ag71xx *ag) { return 0; } |
static inline void ag71xx_debugfs_exit(struct ag71xx *ag) {} |
static inline void ag71xx_debugfs_update_int_stats(struct ag71xx *ag, |
u32 status) {} |
static inline void ag71xx_debugfs_update_napi_stats(struct ag71xx *ag, |
int rx, int tx) {} |
#endif /* CONFIG_AG71XX_DEBUG_FS */ |
|
void ag71xx_ar7240_start(struct ag71xx *ag); |
void ag71xx_ar7240_stop(struct ag71xx *ag); |
int ag71xx_ar7240_init(struct ag71xx *ag); |
void ag71xx_ar7240_cleanup(struct ag71xx *ag); |
|
int ag71xx_mdio_mii_read(struct ag71xx_mdio *am, int addr, int reg); |
void ag71xx_mdio_mii_write(struct ag71xx_mdio *am, int addr, int reg, u16 val); |
|
u16 ar7240sw_phy_read(struct mii_bus *mii, unsigned phy_addr, |
unsigned reg_addr); |
int ar7240sw_phy_write(struct mii_bus *mii, unsigned phy_addr, |
unsigned reg_addr, u16 reg_val); |
|
#endif /* _AG71XX_H */ |
/branches/gl-inet/target/linux/ar71xx/files/drivers/net/ethernet/atheros/ag71xx/ag71xx_ar7240.c |
@@ -0,0 +1,1382 @@ |
/* |
* Driver for the built-in ethernet switch of the Atheros AR7240 SoC |
* Copyright (c) 2010 Gabor Juhos <juhosg@openwrt.org> |
* Copyright (c) 2010 Felix Fietkau <nbd@nbd.name> |
* |
* This program is free software; you can redistribute it and/or modify it |
* under the terms of the GNU General Public License version 2 as published |
* by the Free Software Foundation. |
* |
*/ |
|
#include <linux/etherdevice.h> |
#include <linux/list.h> |
#include <linux/netdevice.h> |
#include <linux/phy.h> |
#include <linux/mii.h> |
#include <linux/bitops.h> |
#include <linux/switch.h> |
#include "ag71xx.h" |
|
#define BITM(_count) (BIT(_count) - 1) |
#define BITS(_shift, _count) (BITM(_count) << _shift) |
|
#define AR7240_REG_MASK_CTRL 0x00 |
#define AR7240_MASK_CTRL_REVISION_M BITM(8) |
#define AR7240_MASK_CTRL_VERSION_M BITM(8) |
#define AR7240_MASK_CTRL_VERSION_S 8 |
#define AR7240_MASK_CTRL_VERSION_AR7240 0x01 |
#define AR7240_MASK_CTRL_VERSION_AR934X 0x02 |
#define AR7240_MASK_CTRL_SOFT_RESET BIT(31) |
|
#define AR7240_REG_MAC_ADDR0 0x20 |
#define AR7240_REG_MAC_ADDR1 0x24 |
|
#define AR7240_REG_FLOOD_MASK 0x2c |
#define AR7240_FLOOD_MASK_BROAD_TO_CPU BIT(26) |
|
#define AR7240_REG_GLOBAL_CTRL 0x30 |
#define AR7240_GLOBAL_CTRL_MTU_M BITM(11) |
#define AR9340_GLOBAL_CTRL_MTU_M BITM(14) |
|
#define AR7240_REG_VTU 0x0040 |
#define AR7240_VTU_OP BITM(3) |
#define AR7240_VTU_OP_NOOP 0x0 |
#define AR7240_VTU_OP_FLUSH 0x1 |
#define AR7240_VTU_OP_LOAD 0x2 |
#define AR7240_VTU_OP_PURGE 0x3 |
#define AR7240_VTU_OP_REMOVE_PORT 0x4 |
#define AR7240_VTU_ACTIVE BIT(3) |
#define AR7240_VTU_FULL BIT(4) |
#define AR7240_VTU_PORT BITS(8, 4) |
#define AR7240_VTU_PORT_S 8 |
#define AR7240_VTU_VID BITS(16, 12) |
#define AR7240_VTU_VID_S 16 |
#define AR7240_VTU_PRIO BITS(28, 3) |
#define AR7240_VTU_PRIO_S 28 |
#define AR7240_VTU_PRIO_EN BIT(31) |
|
#define AR7240_REG_VTU_DATA 0x0044 |
#define AR7240_VTUDATA_MEMBER BITS(0, 10) |
#define AR7240_VTUDATA_VALID BIT(11) |
|
#define AR7240_REG_ATU 0x50 |
#define AR7240_ATU_FLUSH_ALL 0x1 |
|
#define AR7240_REG_AT_CTRL 0x5c |
#define AR7240_AT_CTRL_AGE_TIME BITS(0, 15) |
#define AR7240_AT_CTRL_AGE_EN BIT(17) |
#define AR7240_AT_CTRL_LEARN_CHANGE BIT(18) |
#define AR7240_AT_CTRL_RESERVED BIT(19) |
#define AR7240_AT_CTRL_ARP_EN BIT(20) |
|
#define AR7240_REG_TAG_PRIORITY 0x70 |
|
#define AR7240_REG_SERVICE_TAG 0x74 |
#define AR7240_SERVICE_TAG_M BITM(16) |
|
#define AR7240_REG_CPU_PORT 0x78 |
#define AR7240_MIRROR_PORT_S 4 |
#define AR7240_MIRROR_PORT_M BITM(4) |
#define AR7240_CPU_PORT_EN BIT(8) |
|
#define AR7240_REG_MIB_FUNCTION0 0x80 |
#define AR7240_MIB_TIMER_M BITM(16) |
#define AR7240_MIB_AT_HALF_EN BIT(16) |
#define AR7240_MIB_BUSY BIT(17) |
#define AR7240_MIB_FUNC_S 24 |
#define AR7240_MIB_FUNC_M BITM(3) |
#define AR7240_MIB_FUNC_NO_OP 0x0 |
#define AR7240_MIB_FUNC_FLUSH 0x1 |
#define AR7240_MIB_FUNC_CAPTURE 0x3 |
|
#define AR7240_REG_MDIO_CTRL 0x98 |
#define AR7240_MDIO_CTRL_DATA_M BITM(16) |
#define AR7240_MDIO_CTRL_REG_ADDR_S 16 |
#define AR7240_MDIO_CTRL_PHY_ADDR_S 21 |
#define AR7240_MDIO_CTRL_CMD_WRITE 0 |
#define AR7240_MDIO_CTRL_CMD_READ BIT(27) |
#define AR7240_MDIO_CTRL_MASTER_EN BIT(30) |
#define AR7240_MDIO_CTRL_BUSY BIT(31) |
|
#define AR7240_REG_PORT_BASE(_port) (0x100 + (_port) * 0x100) |
|
#define AR7240_REG_PORT_STATUS(_port) (AR7240_REG_PORT_BASE((_port)) + 0x00) |
#define AR7240_PORT_STATUS_SPEED_S 0 |
#define AR7240_PORT_STATUS_SPEED_M BITM(2) |
#define AR7240_PORT_STATUS_SPEED_10 0 |
#define AR7240_PORT_STATUS_SPEED_100 1 |
#define AR7240_PORT_STATUS_SPEED_1000 2 |
#define AR7240_PORT_STATUS_TXMAC BIT(2) |
#define AR7240_PORT_STATUS_RXMAC BIT(3) |
#define AR7240_PORT_STATUS_TXFLOW BIT(4) |
#define AR7240_PORT_STATUS_RXFLOW BIT(5) |
#define AR7240_PORT_STATUS_DUPLEX BIT(6) |
#define AR7240_PORT_STATUS_LINK_UP BIT(8) |
#define AR7240_PORT_STATUS_LINK_AUTO BIT(9) |
#define AR7240_PORT_STATUS_LINK_PAUSE BIT(10) |
|
#define AR7240_REG_PORT_CTRL(_port) (AR7240_REG_PORT_BASE((_port)) + 0x04) |
#define AR7240_PORT_CTRL_STATE_M BITM(3) |
#define AR7240_PORT_CTRL_STATE_DISABLED 0 |
#define AR7240_PORT_CTRL_STATE_BLOCK 1 |
#define AR7240_PORT_CTRL_STATE_LISTEN 2 |
#define AR7240_PORT_CTRL_STATE_LEARN 3 |
#define AR7240_PORT_CTRL_STATE_FORWARD 4 |
#define AR7240_PORT_CTRL_LEARN_LOCK BIT(7) |
#define AR7240_PORT_CTRL_VLAN_MODE_S 8 |
#define AR7240_PORT_CTRL_VLAN_MODE_KEEP 0 |
#define AR7240_PORT_CTRL_VLAN_MODE_STRIP 1 |
#define AR7240_PORT_CTRL_VLAN_MODE_ADD 2 |
#define AR7240_PORT_CTRL_VLAN_MODE_DOUBLE_TAG 3 |
#define AR7240_PORT_CTRL_IGMP_SNOOP BIT(10) |
#define AR7240_PORT_CTRL_HEADER BIT(11) |
#define AR7240_PORT_CTRL_MAC_LOOP BIT(12) |
#define AR7240_PORT_CTRL_SINGLE_VLAN BIT(13) |
#define AR7240_PORT_CTRL_LEARN BIT(14) |
#define AR7240_PORT_CTRL_DOUBLE_TAG BIT(15) |
#define AR7240_PORT_CTRL_MIRROR_TX BIT(16) |
#define AR7240_PORT_CTRL_MIRROR_RX BIT(17) |
|
#define AR7240_REG_PORT_VLAN(_port) (AR7240_REG_PORT_BASE((_port)) + 0x08) |
|
#define AR7240_PORT_VLAN_DEFAULT_ID_S 0 |
#define AR7240_PORT_VLAN_DEST_PORTS_S 16 |
#define AR7240_PORT_VLAN_MODE_S 30 |
#define AR7240_PORT_VLAN_MODE_PORT_ONLY 0 |
#define AR7240_PORT_VLAN_MODE_PORT_FALLBACK 1 |
#define AR7240_PORT_VLAN_MODE_VLAN_ONLY 2 |
#define AR7240_PORT_VLAN_MODE_SECURE 3 |
|
|
#define AR7240_REG_STATS_BASE(_port) (0x20000 + (_port) * 0x100) |
|
#define AR7240_STATS_RXBROAD 0x00 |
#define AR7240_STATS_RXPAUSE 0x04 |
#define AR7240_STATS_RXMULTI 0x08 |
#define AR7240_STATS_RXFCSERR 0x0c |
#define AR7240_STATS_RXALIGNERR 0x10 |
#define AR7240_STATS_RXRUNT 0x14 |
#define AR7240_STATS_RXFRAGMENT 0x18 |
#define AR7240_STATS_RX64BYTE 0x1c |
#define AR7240_STATS_RX128BYTE 0x20 |
#define AR7240_STATS_RX256BYTE 0x24 |
#define AR7240_STATS_RX512BYTE 0x28 |
#define AR7240_STATS_RX1024BYTE 0x2c |
#define AR7240_STATS_RX1518BYTE 0x30 |
#define AR7240_STATS_RXMAXBYTE 0x34 |
#define AR7240_STATS_RXTOOLONG 0x38 |
#define AR7240_STATS_RXGOODBYTE 0x3c |
#define AR7240_STATS_RXBADBYTE 0x44 |
#define AR7240_STATS_RXOVERFLOW 0x4c |
#define AR7240_STATS_FILTERED 0x50 |
#define AR7240_STATS_TXBROAD 0x54 |
#define AR7240_STATS_TXPAUSE 0x58 |
#define AR7240_STATS_TXMULTI 0x5c |
#define AR7240_STATS_TXUNDERRUN 0x60 |
#define AR7240_STATS_TX64BYTE 0x64 |
#define AR7240_STATS_TX128BYTE 0x68 |
#define AR7240_STATS_TX256BYTE 0x6c |
#define AR7240_STATS_TX512BYTE 0x70 |
#define AR7240_STATS_TX1024BYTE 0x74 |
#define AR7240_STATS_TX1518BYTE 0x78 |
#define AR7240_STATS_TXMAXBYTE 0x7c |
#define AR7240_STATS_TXOVERSIZE 0x80 |
#define AR7240_STATS_TXBYTE 0x84 |
#define AR7240_STATS_TXCOLLISION 0x8c |
#define AR7240_STATS_TXABORTCOL 0x90 |
#define AR7240_STATS_TXMULTICOL 0x94 |
#define AR7240_STATS_TXSINGLECOL 0x98 |
#define AR7240_STATS_TXEXCDEFER 0x9c |
#define AR7240_STATS_TXDEFER 0xa0 |
#define AR7240_STATS_TXLATECOL 0xa4 |
|
#define AR7240_PORT_CPU 0 |
#define AR7240_NUM_PORTS 6 |
#define AR7240_NUM_PHYS 5 |
|
#define AR7240_PHY_ID1 0x004d |
#define AR7240_PHY_ID2 0xd041 |
|
#define AR934X_PHY_ID1 0x004d |
#define AR934X_PHY_ID2 0xd042 |
|
#define AR7240_MAX_VLANS 16 |
|
#define AR934X_REG_OPER_MODE0 0x04 |
#define AR934X_OPER_MODE0_MAC_GMII_EN BIT(6) |
#define AR934X_OPER_MODE0_PHY_MII_EN BIT(10) |
|
#define AR934X_REG_OPER_MODE1 0x08 |
#define AR934X_REG_OPER_MODE1_PHY4_MII_EN BIT(28) |
|
#define AR934X_REG_FLOOD_MASK 0x2c |
#define AR934X_FLOOD_MASK_MC_DP(_p) BIT(16 + (_p)) |
#define AR934X_FLOOD_MASK_BC_DP(_p) BIT(25 + (_p)) |
|
#define AR934X_REG_QM_CTRL 0x3c |
#define AR934X_QM_CTRL_ARP_EN BIT(15) |
|
#define AR934X_REG_AT_CTRL 0x5c |
#define AR934X_AT_CTRL_AGE_TIME BITS(0, 15) |
#define AR934X_AT_CTRL_AGE_EN BIT(17) |
#define AR934X_AT_CTRL_LEARN_CHANGE BIT(18) |
|
#define AR934X_MIB_ENABLE BIT(30) |
|
#define AR934X_REG_PORT_BASE(_port) (0x100 + (_port) * 0x100) |
|
#define AR934X_REG_PORT_VLAN1(_port) (AR934X_REG_PORT_BASE((_port)) + 0x08) |
#define AR934X_PORT_VLAN1_DEFAULT_SVID_S 0 |
#define AR934X_PORT_VLAN1_FORCE_DEFAULT_VID_EN BIT(12) |
#define AR934X_PORT_VLAN1_PORT_TLS_MODE BIT(13) |
#define AR934X_PORT_VLAN1_PORT_VLAN_PROP_EN BIT(14) |
#define AR934X_PORT_VLAN1_PORT_CLONE_EN BIT(15) |
#define AR934X_PORT_VLAN1_DEFAULT_CVID_S 16 |
#define AR934X_PORT_VLAN1_FORCE_PORT_VLAN_EN BIT(28) |
#define AR934X_PORT_VLAN1_ING_PORT_PRI_S 29 |
|
#define AR934X_REG_PORT_VLAN2(_port) (AR934X_REG_PORT_BASE((_port)) + 0x0c) |
#define AR934X_PORT_VLAN2_PORT_VID_MEM_S 16 |
#define AR934X_PORT_VLAN2_8021Q_MODE_S 30 |
#define AR934X_PORT_VLAN2_8021Q_MODE_PORT_ONLY 0 |
#define AR934X_PORT_VLAN2_8021Q_MODE_PORT_FALLBACK 1 |
#define AR934X_PORT_VLAN2_8021Q_MODE_VLAN_ONLY 2 |
#define AR934X_PORT_VLAN2_8021Q_MODE_SECURE 3 |
|
#define sw_to_ar7240(_dev) container_of(_dev, struct ar7240sw, swdev) |
|
struct ar7240sw_port_stat { |
unsigned long rx_broadcast; |
unsigned long rx_pause; |
unsigned long rx_multicast; |
unsigned long rx_fcs_error; |
unsigned long rx_align_error; |
unsigned long rx_runt; |
unsigned long rx_fragments; |
unsigned long rx_64byte; |
unsigned long rx_128byte; |
unsigned long rx_256byte; |
unsigned long rx_512byte; |
unsigned long rx_1024byte; |
unsigned long rx_1518byte; |
unsigned long rx_maxbyte; |
unsigned long rx_toolong; |
unsigned long rx_good_byte; |
unsigned long rx_bad_byte; |
unsigned long rx_overflow; |
unsigned long filtered; |
|
unsigned long tx_broadcast; |
unsigned long tx_pause; |
unsigned long tx_multicast; |
unsigned long tx_underrun; |
unsigned long tx_64byte; |
unsigned long tx_128byte; |
unsigned long tx_256byte; |
unsigned long tx_512byte; |
unsigned long tx_1024byte; |
unsigned long tx_1518byte; |
unsigned long tx_maxbyte; |
unsigned long tx_oversize; |
unsigned long tx_byte; |
unsigned long tx_collision; |
unsigned long tx_abortcol; |
unsigned long tx_multicol; |
unsigned long tx_singlecol; |
unsigned long tx_excdefer; |
unsigned long tx_defer; |
unsigned long tx_xlatecol; |
}; |
|
struct ar7240sw { |
struct mii_bus *mii_bus; |
struct ag71xx_switch_platform_data *swdata; |
struct switch_dev swdev; |
int num_ports; |
u8 ver; |
bool vlan; |
u16 vlan_id[AR7240_MAX_VLANS]; |
u8 vlan_table[AR7240_MAX_VLANS]; |
u8 vlan_tagged; |
u16 pvid[AR7240_NUM_PORTS]; |
char buf[80]; |
|
rwlock_t stats_lock; |
struct ar7240sw_port_stat port_stats[AR7240_NUM_PORTS]; |
}; |
|
struct ar7240sw_hw_stat { |
char string[ETH_GSTRING_LEN]; |
int sizeof_stat; |
int reg; |
}; |
|
static DEFINE_MUTEX(reg_mutex); |
|
static inline int sw_is_ar7240(struct ar7240sw *as) |
{ |
return as->ver == AR7240_MASK_CTRL_VERSION_AR7240; |
} |
|
static inline int sw_is_ar934x(struct ar7240sw *as) |
{ |
return as->ver == AR7240_MASK_CTRL_VERSION_AR934X; |
} |
|
static inline u32 ar7240sw_port_mask(struct ar7240sw *as, int port) |
{ |
return BIT(port); |
} |
|
static inline u32 ar7240sw_port_mask_all(struct ar7240sw *as) |
{ |
return BIT(as->swdev.ports) - 1; |
} |
|
static inline u32 ar7240sw_port_mask_but(struct ar7240sw *as, int port) |
{ |
return ar7240sw_port_mask_all(as) & ~BIT(port); |
} |
|
static inline u16 mk_phy_addr(u32 reg) |
{ |
return 0x17 & ((reg >> 4) | 0x10); |
} |
|
static inline u16 mk_phy_reg(u32 reg) |
{ |
return (reg << 1) & 0x1e; |
} |
|
static inline u16 mk_high_addr(u32 reg) |
{ |
return (reg >> 7) & 0x1ff; |
} |
|
static u32 __ar7240sw_reg_read(struct mii_bus *mii, u32 reg) |
{ |
unsigned long flags; |
u16 phy_addr; |
u16 phy_reg; |
u32 hi, lo; |
|
reg = (reg & 0xfffffffc) >> 2; |
phy_addr = mk_phy_addr(reg); |
phy_reg = mk_phy_reg(reg); |
|
local_irq_save(flags); |
ag71xx_mdio_mii_write(mii->priv, 0x1f, 0x10, mk_high_addr(reg)); |
lo = (u32) ag71xx_mdio_mii_read(mii->priv, phy_addr, phy_reg); |
hi = (u32) ag71xx_mdio_mii_read(mii->priv, phy_addr, phy_reg + 1); |
local_irq_restore(flags); |
|
return (hi << 16) | lo; |
} |
|
static void __ar7240sw_reg_write(struct mii_bus *mii, u32 reg, u32 val) |
{ |
unsigned long flags; |
u16 phy_addr; |
u16 phy_reg; |
|
reg = (reg & 0xfffffffc) >> 2; |
phy_addr = mk_phy_addr(reg); |
phy_reg = mk_phy_reg(reg); |
|
local_irq_save(flags); |
ag71xx_mdio_mii_write(mii->priv, 0x1f, 0x10, mk_high_addr(reg)); |
ag71xx_mdio_mii_write(mii->priv, phy_addr, phy_reg + 1, (val >> 16)); |
ag71xx_mdio_mii_write(mii->priv, phy_addr, phy_reg, (val & 0xffff)); |
local_irq_restore(flags); |
} |
|
static u32 ar7240sw_reg_read(struct mii_bus *mii, u32 reg_addr) |
{ |
u32 ret; |
|
mutex_lock(®_mutex); |
ret = __ar7240sw_reg_read(mii, reg_addr); |
mutex_unlock(®_mutex); |
|
return ret; |
} |
|
static void ar7240sw_reg_write(struct mii_bus *mii, u32 reg_addr, u32 reg_val) |
{ |
mutex_lock(®_mutex); |
__ar7240sw_reg_write(mii, reg_addr, reg_val); |
mutex_unlock(®_mutex); |
} |
|
static u32 ar7240sw_reg_rmw(struct mii_bus *mii, u32 reg, u32 mask, u32 val) |
{ |
u32 t; |
|
mutex_lock(®_mutex); |
t = __ar7240sw_reg_read(mii, reg); |
t &= ~mask; |
t |= val; |
__ar7240sw_reg_write(mii, reg, t); |
mutex_unlock(®_mutex); |
|
return t; |
} |
|
static void ar7240sw_reg_set(struct mii_bus *mii, u32 reg, u32 val) |
{ |
u32 t; |
|
mutex_lock(®_mutex); |
t = __ar7240sw_reg_read(mii, reg); |
t |= val; |
__ar7240sw_reg_write(mii, reg, t); |
mutex_unlock(®_mutex); |
} |
|
static int __ar7240sw_reg_wait(struct mii_bus *mii, u32 reg, u32 mask, u32 val, |
unsigned timeout) |
{ |
int i; |
|
for (i = 0; i < timeout; i++) { |
u32 t; |
|
t = __ar7240sw_reg_read(mii, reg); |
if ((t & mask) == val) |
return 0; |
|
usleep_range(1000, 2000); |
} |
|
return -ETIMEDOUT; |
} |
|
static int ar7240sw_reg_wait(struct mii_bus *mii, u32 reg, u32 mask, u32 val, |
unsigned timeout) |
{ |
int ret; |
|
mutex_lock(®_mutex); |
ret = __ar7240sw_reg_wait(mii, reg, mask, val, timeout); |
mutex_unlock(®_mutex); |
return ret; |
} |
|
u16 ar7240sw_phy_read(struct mii_bus *mii, unsigned phy_addr, |
unsigned reg_addr) |
{ |
u32 t, val = 0xffff; |
int err; |
|
if (phy_addr >= AR7240_NUM_PHYS) |
return 0xffff; |
|
mutex_lock(®_mutex); |
t = (reg_addr << AR7240_MDIO_CTRL_REG_ADDR_S) | |
(phy_addr << AR7240_MDIO_CTRL_PHY_ADDR_S) | |
AR7240_MDIO_CTRL_MASTER_EN | |
AR7240_MDIO_CTRL_BUSY | |
AR7240_MDIO_CTRL_CMD_READ; |
|
__ar7240sw_reg_write(mii, AR7240_REG_MDIO_CTRL, t); |
err = __ar7240sw_reg_wait(mii, AR7240_REG_MDIO_CTRL, |
AR7240_MDIO_CTRL_BUSY, 0, 5); |
if (!err) |
val = __ar7240sw_reg_read(mii, AR7240_REG_MDIO_CTRL); |
mutex_unlock(®_mutex); |
|
return val & AR7240_MDIO_CTRL_DATA_M; |
} |
|
int ar7240sw_phy_write(struct mii_bus *mii, unsigned phy_addr, |
unsigned reg_addr, u16 reg_val) |
{ |
u32 t; |
int ret; |
|
if (phy_addr >= AR7240_NUM_PHYS) |
return -EINVAL; |
|
mutex_lock(®_mutex); |
t = (phy_addr << AR7240_MDIO_CTRL_PHY_ADDR_S) | |
(reg_addr << AR7240_MDIO_CTRL_REG_ADDR_S) | |
AR7240_MDIO_CTRL_MASTER_EN | |
AR7240_MDIO_CTRL_BUSY | |
AR7240_MDIO_CTRL_CMD_WRITE | |
reg_val; |
|
__ar7240sw_reg_write(mii, AR7240_REG_MDIO_CTRL, t); |
ret = __ar7240sw_reg_wait(mii, AR7240_REG_MDIO_CTRL, |
AR7240_MDIO_CTRL_BUSY, 0, 5); |
mutex_unlock(®_mutex); |
|
return ret; |
} |
|
static int ar7240sw_capture_stats(struct ar7240sw *as) |
{ |
struct mii_bus *mii = as->mii_bus; |
int port; |
int ret; |
|
write_lock(&as->stats_lock); |
|
/* Capture the hardware statistics for all ports */ |
ar7240sw_reg_rmw(mii, AR7240_REG_MIB_FUNCTION0, |
(AR7240_MIB_FUNC_M << AR7240_MIB_FUNC_S), |
(AR7240_MIB_FUNC_CAPTURE << AR7240_MIB_FUNC_S)); |
|
/* Wait for the capturing to complete. */ |
ret = ar7240sw_reg_wait(mii, AR7240_REG_MIB_FUNCTION0, |
AR7240_MIB_BUSY, 0, 10); |
|
if (ret) |
goto unlock; |
|
for (port = 0; port < AR7240_NUM_PORTS; port++) { |
unsigned int base; |
struct ar7240sw_port_stat *stats; |
|
base = AR7240_REG_STATS_BASE(port); |
stats = &as->port_stats[port]; |
|
#define READ_STAT(_r) ar7240sw_reg_read(mii, base + AR7240_STATS_ ## _r) |
|
stats->rx_good_byte += READ_STAT(RXGOODBYTE); |
stats->tx_byte += READ_STAT(TXBYTE); |
|
#undef READ_STAT |
} |
|
ret = 0; |
|
unlock: |
write_unlock(&as->stats_lock); |
return ret; |
} |
|
static void ar7240sw_disable_port(struct ar7240sw *as, unsigned port) |
{ |
ar7240sw_reg_write(as->mii_bus, AR7240_REG_PORT_CTRL(port), |
AR7240_PORT_CTRL_STATE_DISABLED); |
} |
|
static void ar7240sw_setup(struct ar7240sw *as) |
{ |
struct mii_bus *mii = as->mii_bus; |
|
/* Enable CPU port, and disable mirror port */ |
ar7240sw_reg_write(mii, AR7240_REG_CPU_PORT, |
AR7240_CPU_PORT_EN | |
(15 << AR7240_MIRROR_PORT_S)); |
|
/* Setup TAG priority mapping */ |
ar7240sw_reg_write(mii, AR7240_REG_TAG_PRIORITY, 0xfa50); |
|
if (sw_is_ar934x(as)) { |
/* Enable aging, MAC replacing */ |
ar7240sw_reg_write(mii, AR934X_REG_AT_CTRL, |
0x2b /* 5 min age time */ | |
AR934X_AT_CTRL_AGE_EN | |
AR934X_AT_CTRL_LEARN_CHANGE); |
/* Enable ARP frame acknowledge */ |
ar7240sw_reg_set(mii, AR934X_REG_QM_CTRL, |
AR934X_QM_CTRL_ARP_EN); |
/* Enable Broadcast/Multicast frames transmitted to the CPU */ |
ar7240sw_reg_set(mii, AR934X_REG_FLOOD_MASK, |
AR934X_FLOOD_MASK_BC_DP(0) | |
AR934X_FLOOD_MASK_MC_DP(0)); |
|
/* setup MTU */ |
ar7240sw_reg_rmw(mii, AR7240_REG_GLOBAL_CTRL, |
AR9340_GLOBAL_CTRL_MTU_M, |
AR9340_GLOBAL_CTRL_MTU_M); |
|
/* Enable MIB counters */ |
ar7240sw_reg_set(mii, AR7240_REG_MIB_FUNCTION0, |
AR934X_MIB_ENABLE); |
|
} else { |
/* Enable ARP frame acknowledge, aging, MAC replacing */ |
ar7240sw_reg_write(mii, AR7240_REG_AT_CTRL, |
AR7240_AT_CTRL_RESERVED | |
0x2b /* 5 min age time */ | |
AR7240_AT_CTRL_AGE_EN | |
AR7240_AT_CTRL_ARP_EN | |
AR7240_AT_CTRL_LEARN_CHANGE); |
/* Enable Broadcast frames transmitted to the CPU */ |
ar7240sw_reg_set(mii, AR7240_REG_FLOOD_MASK, |
AR7240_FLOOD_MASK_BROAD_TO_CPU); |
|
/* setup MTU */ |
ar7240sw_reg_rmw(mii, AR7240_REG_GLOBAL_CTRL, |
AR7240_GLOBAL_CTRL_MTU_M, |
AR7240_GLOBAL_CTRL_MTU_M); |
} |
|
/* setup Service TAG */ |
ar7240sw_reg_rmw(mii, AR7240_REG_SERVICE_TAG, AR7240_SERVICE_TAG_M, 0); |
} |
|
/* inspired by phy_poll_reset in drivers/net/phy/phy_device.c */ |
static int |
ar7240sw_phy_poll_reset(struct mii_bus *bus) |
{ |
const unsigned int sleep_msecs = 20; |
int ret, elapsed, i; |
|
for (elapsed = sleep_msecs; elapsed <= 600; |
elapsed += sleep_msecs) { |
msleep(sleep_msecs); |
for (i = 0; i < AR7240_NUM_PHYS; i++) { |
ret = ar7240sw_phy_read(bus, i, MII_BMCR); |
if (ret < 0) |
return ret; |
if (ret & BMCR_RESET) |
break; |
if (i == AR7240_NUM_PHYS - 1) { |
usleep_range(1000, 2000); |
return 0; |
} |
} |
} |
return -ETIMEDOUT; |
} |
|
static int ar7240sw_reset(struct ar7240sw *as) |
{ |
struct mii_bus *mii = as->mii_bus; |
int ret; |
int i; |
|
/* Set all ports to disabled state. */ |
for (i = 0; i < AR7240_NUM_PORTS; i++) |
ar7240sw_disable_port(as, i); |
|
/* Wait for transmit queues to drain. */ |
usleep_range(2000, 3000); |
|
/* Reset the switch. */ |
ar7240sw_reg_write(mii, AR7240_REG_MASK_CTRL, |
AR7240_MASK_CTRL_SOFT_RESET); |
|
ret = ar7240sw_reg_wait(mii, AR7240_REG_MASK_CTRL, |
AR7240_MASK_CTRL_SOFT_RESET, 0, 1000); |
|
/* setup PHYs */ |
for (i = 0; i < AR7240_NUM_PHYS; i++) { |
ar7240sw_phy_write(mii, i, MII_ADVERTISE, |
ADVERTISE_ALL | ADVERTISE_PAUSE_CAP | |
ADVERTISE_PAUSE_ASYM); |
ar7240sw_phy_write(mii, i, MII_BMCR, |
BMCR_RESET | BMCR_ANENABLE); |
} |
ret = ar7240sw_phy_poll_reset(mii); |
if (ret) |
return ret; |
|
ar7240sw_setup(as); |
return ret; |
} |
|
static void ar7240sw_setup_port(struct ar7240sw *as, unsigned port, u8 portmask) |
{ |
struct mii_bus *mii = as->mii_bus; |
u32 ctrl; |
u32 vid, mode; |
|
ctrl = AR7240_PORT_CTRL_STATE_FORWARD | AR7240_PORT_CTRL_LEARN | |
AR7240_PORT_CTRL_SINGLE_VLAN; |
|
if (port == AR7240_PORT_CPU) { |
ar7240sw_reg_write(mii, AR7240_REG_PORT_STATUS(port), |
AR7240_PORT_STATUS_SPEED_1000 | |
AR7240_PORT_STATUS_TXFLOW | |
AR7240_PORT_STATUS_RXFLOW | |
AR7240_PORT_STATUS_TXMAC | |
AR7240_PORT_STATUS_RXMAC | |
AR7240_PORT_STATUS_DUPLEX); |
} else { |
ar7240sw_reg_write(mii, AR7240_REG_PORT_STATUS(port), |
AR7240_PORT_STATUS_LINK_AUTO); |
} |
|
/* Set the default VID for this port */ |
if (as->vlan) { |
vid = as->vlan_id[as->pvid[port]]; |
mode = AR7240_PORT_VLAN_MODE_SECURE; |
} else { |
vid = port; |
mode = AR7240_PORT_VLAN_MODE_PORT_ONLY; |
} |
|
if (as->vlan) { |
if (as->vlan_tagged & BIT(port)) |
ctrl |= AR7240_PORT_CTRL_VLAN_MODE_ADD << |
AR7240_PORT_CTRL_VLAN_MODE_S; |
else |
ctrl |= AR7240_PORT_CTRL_VLAN_MODE_STRIP << |
AR7240_PORT_CTRL_VLAN_MODE_S; |
} else { |
ctrl |= AR7240_PORT_CTRL_VLAN_MODE_KEEP << |
AR7240_PORT_CTRL_VLAN_MODE_S; |
} |
|
if (!portmask) { |
if (port == AR7240_PORT_CPU) |
portmask = ar7240sw_port_mask_but(as, AR7240_PORT_CPU); |
else |
portmask = ar7240sw_port_mask(as, AR7240_PORT_CPU); |
} |
|
/* allow the port to talk to all other ports, but exclude its |
* own ID to prevent frames from being reflected back to the |
* port that they came from */ |
portmask &= ar7240sw_port_mask_but(as, port); |
|
ar7240sw_reg_write(mii, AR7240_REG_PORT_CTRL(port), ctrl); |
if (sw_is_ar934x(as)) { |
u32 vlan1, vlan2; |
|
vlan1 = (vid << AR934X_PORT_VLAN1_DEFAULT_CVID_S); |
vlan2 = (portmask << AR934X_PORT_VLAN2_PORT_VID_MEM_S) | |
(mode << AR934X_PORT_VLAN2_8021Q_MODE_S); |
ar7240sw_reg_write(mii, AR934X_REG_PORT_VLAN1(port), vlan1); |
ar7240sw_reg_write(mii, AR934X_REG_PORT_VLAN2(port), vlan2); |
} else { |
u32 vlan; |
|
vlan = vid | (mode << AR7240_PORT_VLAN_MODE_S) | |
(portmask << AR7240_PORT_VLAN_DEST_PORTS_S); |
|
ar7240sw_reg_write(mii, AR7240_REG_PORT_VLAN(port), vlan); |
} |
} |
|
static int ar7240_set_addr(struct ar7240sw *as, u8 *addr) |
{ |
struct mii_bus *mii = as->mii_bus; |
u32 t; |
|
t = (addr[4] << 8) | addr[5]; |
ar7240sw_reg_write(mii, AR7240_REG_MAC_ADDR0, t); |
|
t = (addr[0] << 24) | (addr[1] << 16) | (addr[2] << 8) | addr[3]; |
ar7240sw_reg_write(mii, AR7240_REG_MAC_ADDR1, t); |
|
return 0; |
} |
|
static int |
ar7240_set_vid(struct switch_dev *dev, const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
as->vlan_id[val->port_vlan] = val->value.i; |
return 0; |
} |
|
static int |
ar7240_get_vid(struct switch_dev *dev, const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
val->value.i = as->vlan_id[val->port_vlan]; |
return 0; |
} |
|
static int |
ar7240_set_pvid(struct switch_dev *dev, int port, int vlan) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
|
/* make sure no invalid PVIDs get set */ |
|
if (vlan >= dev->vlans) |
return -EINVAL; |
|
as->pvid[port] = vlan; |
return 0; |
} |
|
static int |
ar7240_get_pvid(struct switch_dev *dev, int port, int *vlan) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
*vlan = as->pvid[port]; |
return 0; |
} |
|
static int |
ar7240_get_ports(struct switch_dev *dev, struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
u8 ports = as->vlan_table[val->port_vlan]; |
int i; |
|
val->len = 0; |
for (i = 0; i < as->swdev.ports; i++) { |
struct switch_port *p; |
|
if (!(ports & (1 << i))) |
continue; |
|
p = &val->value.ports[val->len++]; |
p->id = i; |
if (as->vlan_tagged & (1 << i)) |
p->flags = (1 << SWITCH_PORT_FLAG_TAGGED); |
else |
p->flags = 0; |
} |
return 0; |
} |
|
static int |
ar7240_set_ports(struct switch_dev *dev, struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
u8 *vt = &as->vlan_table[val->port_vlan]; |
int i, j; |
|
*vt = 0; |
for (i = 0; i < val->len; i++) { |
struct switch_port *p = &val->value.ports[i]; |
|
if (p->flags & (1 << SWITCH_PORT_FLAG_TAGGED)) |
as->vlan_tagged |= (1 << p->id); |
else { |
as->vlan_tagged &= ~(1 << p->id); |
as->pvid[p->id] = val->port_vlan; |
|
/* make sure that an untagged port does not |
* appear in other vlans */ |
for (j = 0; j < AR7240_MAX_VLANS; j++) { |
if (j == val->port_vlan) |
continue; |
as->vlan_table[j] &= ~(1 << p->id); |
} |
} |
|
*vt |= 1 << p->id; |
} |
return 0; |
} |
|
static int |
ar7240_set_vlan(struct switch_dev *dev, const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
as->vlan = !!val->value.i; |
return 0; |
} |
|
static int |
ar7240_get_vlan(struct switch_dev *dev, const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
val->value.i = as->vlan; |
return 0; |
} |
|
static void |
ar7240_vtu_op(struct ar7240sw *as, u32 op, u32 val) |
{ |
struct mii_bus *mii = as->mii_bus; |
|
if (ar7240sw_reg_wait(mii, AR7240_REG_VTU, AR7240_VTU_ACTIVE, 0, 5)) |
return; |
|
if ((op & AR7240_VTU_OP) == AR7240_VTU_OP_LOAD) { |
val &= AR7240_VTUDATA_MEMBER; |
val |= AR7240_VTUDATA_VALID; |
ar7240sw_reg_write(mii, AR7240_REG_VTU_DATA, val); |
} |
op |= AR7240_VTU_ACTIVE; |
ar7240sw_reg_write(mii, AR7240_REG_VTU, op); |
} |
|
static int |
ar7240_hw_apply(struct switch_dev *dev) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
u8 portmask[AR7240_NUM_PORTS]; |
int i, j; |
|
/* flush all vlan translation unit entries */ |
ar7240_vtu_op(as, AR7240_VTU_OP_FLUSH, 0); |
|
memset(portmask, 0, sizeof(portmask)); |
if (as->vlan) { |
/* calculate the port destination masks and load vlans |
* into the vlan translation unit */ |
for (j = 0; j < AR7240_MAX_VLANS; j++) { |
u8 vp = as->vlan_table[j]; |
|
if (!vp) |
continue; |
|
for (i = 0; i < as->swdev.ports; i++) { |
u8 mask = (1 << i); |
if (vp & mask) |
portmask[i] |= vp & ~mask; |
} |
|
ar7240_vtu_op(as, |
AR7240_VTU_OP_LOAD | |
(as->vlan_id[j] << AR7240_VTU_VID_S), |
as->vlan_table[j]); |
} |
} else { |
/* vlan disabled: |
* isolate all ports, but connect them to the cpu port */ |
for (i = 0; i < as->swdev.ports; i++) { |
if (i == AR7240_PORT_CPU) |
continue; |
|
portmask[i] = 1 << AR7240_PORT_CPU; |
portmask[AR7240_PORT_CPU] |= (1 << i); |
} |
} |
|
/* update the port destination mask registers and tag settings */ |
for (i = 0; i < as->swdev.ports; i++) |
ar7240sw_setup_port(as, i, portmask[i]); |
|
return 0; |
} |
|
static int |
ar7240_reset_switch(struct switch_dev *dev) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
ar7240sw_reset(as); |
return 0; |
} |
|
static int |
ar7240_get_port_link(struct switch_dev *dev, int port, |
struct switch_port_link *link) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
struct mii_bus *mii = as->mii_bus; |
u32 status; |
|
if (port >= AR7240_NUM_PORTS) |
return -EINVAL; |
|
status = ar7240sw_reg_read(mii, AR7240_REG_PORT_STATUS(port)); |
link->aneg = !!(status & AR7240_PORT_STATUS_LINK_AUTO); |
if (link->aneg) { |
link->link = !!(status & AR7240_PORT_STATUS_LINK_UP); |
if (!link->link) |
return 0; |
} else { |
link->link = true; |
} |
|
link->duplex = !!(status & AR7240_PORT_STATUS_DUPLEX); |
link->tx_flow = !!(status & AR7240_PORT_STATUS_TXFLOW); |
link->rx_flow = !!(status & AR7240_PORT_STATUS_RXFLOW); |
switch (status & AR7240_PORT_STATUS_SPEED_M) { |
case AR7240_PORT_STATUS_SPEED_10: |
link->speed = SWITCH_PORT_SPEED_10; |
break; |
case AR7240_PORT_STATUS_SPEED_100: |
link->speed = SWITCH_PORT_SPEED_100; |
break; |
case AR7240_PORT_STATUS_SPEED_1000: |
link->speed = SWITCH_PORT_SPEED_1000; |
break; |
} |
|
return 0; |
} |
|
static int |
ar7240_get_port_stats(struct switch_dev *dev, int port, |
struct switch_port_stats *stats) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
|
if (port >= AR7240_NUM_PORTS) |
return -EINVAL; |
|
ar7240sw_capture_stats(as); |
|
read_lock(&as->stats_lock); |
stats->rx_bytes = as->port_stats[port].rx_good_byte; |
stats->tx_bytes = as->port_stats[port].tx_byte; |
read_unlock(&as->stats_lock); |
|
return 0; |
} |
|
static int |
ar7240_set_mirror_monitor_port(struct switch_dev *dev, |
const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
struct mii_bus *mii = as->mii_bus; |
|
int port = val->value.i; |
|
if (port > 15) |
return -EINVAL; |
|
ar7240sw_reg_rmw(mii, AR7240_REG_CPU_PORT, |
AR7240_MIRROR_PORT_M << AR7240_MIRROR_PORT_S, |
port << AR7240_MIRROR_PORT_S); |
|
return 0; |
} |
|
static int |
ar7240_get_mirror_monitor_port(struct switch_dev *dev, |
const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
struct mii_bus *mii = as->mii_bus; |
|
u32 ret; |
|
ret = ar7240sw_reg_read(mii, AR7240_REG_CPU_PORT); |
val->value.i = (ret >> AR7240_MIRROR_PORT_S) & AR7240_MIRROR_PORT_M; |
|
return 0; |
} |
|
static int |
ar7240_set_mirror_rx(struct switch_dev *dev, const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
struct mii_bus *mii = as->mii_bus; |
|
int port = val->port_vlan; |
|
if (port >= dev->ports) |
return -EINVAL; |
|
if (val && val->value.i == 1) |
ar7240sw_reg_set(mii, AR7240_REG_PORT_CTRL(port), |
AR7240_PORT_CTRL_MIRROR_RX); |
else |
ar7240sw_reg_rmw(mii, AR7240_REG_PORT_CTRL(port), |
AR7240_PORT_CTRL_MIRROR_RX, 0); |
|
return 0; |
} |
|
static int |
ar7240_get_mirror_rx(struct switch_dev *dev, const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
struct mii_bus *mii = as->mii_bus; |
|
u32 ctrl; |
|
int port = val->port_vlan; |
|
if (port >= dev->ports) |
return -EINVAL; |
|
ctrl = ar7240sw_reg_read(mii, AR7240_REG_PORT_CTRL(port)); |
|
if ((ctrl & AR7240_PORT_CTRL_MIRROR_RX) == AR7240_PORT_CTRL_MIRROR_RX) |
val->value.i = 1; |
else |
val->value.i = 0; |
|
return 0; |
} |
|
static int |
ar7240_set_mirror_tx(struct switch_dev *dev, const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
struct mii_bus *mii = as->mii_bus; |
|
int port = val->port_vlan; |
|
if (port >= dev->ports) |
return -EINVAL; |
|
if (val && val->value.i == 1) |
ar7240sw_reg_set(mii, AR7240_REG_PORT_CTRL(port), |
AR7240_PORT_CTRL_MIRROR_TX); |
else |
ar7240sw_reg_rmw(mii, AR7240_REG_PORT_CTRL(port), |
AR7240_PORT_CTRL_MIRROR_TX, 0); |
|
return 0; |
} |
|
static int |
ar7240_get_mirror_tx(struct switch_dev *dev, const struct switch_attr *attr, |
struct switch_val *val) |
{ |
struct ar7240sw *as = sw_to_ar7240(dev); |
struct mii_bus *mii = as->mii_bus; |
|
u32 ctrl; |
|
int port = val->port_vlan; |
|
if (port >= dev->ports) |
return -EINVAL; |
|
ctrl = ar7240sw_reg_read(mii, AR7240_REG_PORT_CTRL(port)); |
|
if ((ctrl & AR7240_PORT_CTRL_MIRROR_TX) == AR7240_PORT_CTRL_MIRROR_TX) |
val->value.i = 1; |
else |
val->value.i = 0; |
|
return 0; |
} |
|
static struct switch_attr ar7240_globals[] = { |
{ |
.type = SWITCH_TYPE_INT, |
.name = "enable_vlan", |
.description = "Enable VLAN mode", |
.set = ar7240_set_vlan, |
.get = ar7240_get_vlan, |
.max = 1 |
}, |
{ |
.type = SWITCH_TYPE_INT, |
.name = "mirror_monitor_port", |
.description = "Mirror monitor port", |
.set = ar7240_set_mirror_monitor_port, |
.get = ar7240_get_mirror_monitor_port, |
.max = 15 |
}, |
}; |
|
static struct switch_attr ar7240_port[] = { |
{ |
.type = SWITCH_TYPE_INT, |
.name = "enable_mirror_rx", |
.description = "Enable mirroring of RX packets", |
.set = ar7240_set_mirror_rx, |
.get = ar7240_get_mirror_rx, |
.max = 1 |
}, |
{ |
.type = SWITCH_TYPE_INT, |
.name = "enable_mirror_tx", |
.description = "Enable mirroring of TX packets", |
.set = ar7240_set_mirror_tx, |
.get = ar7240_get_mirror_tx, |
.max = 1 |
}, |
}; |
|
static struct switch_attr ar7240_vlan[] = { |
{ |
.type = SWITCH_TYPE_INT, |
.name = "vid", |
.description = "VLAN ID", |
.set = ar7240_set_vid, |
.get = ar7240_get_vid, |
.max = 4094, |
}, |
}; |
|
static const struct switch_dev_ops ar7240_ops = { |
.attr_global = { |
.attr = ar7240_globals, |
.n_attr = ARRAY_SIZE(ar7240_globals), |
}, |
.attr_port = { |
.attr = ar7240_port, |
.n_attr = ARRAY_SIZE(ar7240_port), |
}, |
.attr_vlan = { |
.attr = ar7240_vlan, |
.n_attr = ARRAY_SIZE(ar7240_vlan), |
}, |
.get_port_pvid = ar7240_get_pvid, |
.set_port_pvid = ar7240_set_pvid, |
.get_vlan_ports = ar7240_get_ports, |
.set_vlan_ports = ar7240_set_ports, |
.apply_config = ar7240_hw_apply, |
.reset_switch = ar7240_reset_switch, |
.get_port_link = ar7240_get_port_link, |
.get_port_stats = ar7240_get_port_stats, |
}; |
|
static struct ar7240sw *ar7240_probe(struct ag71xx *ag) |
{ |
struct ag71xx_platform_data *pdata = ag71xx_get_pdata(ag); |
struct mii_bus *mii = ag->mii_bus; |
struct ar7240sw *as; |
struct switch_dev *swdev; |
u32 ctrl; |
u16 phy_id1; |
u16 phy_id2; |
int i; |
|
phy_id1 = ar7240sw_phy_read(mii, 0, MII_PHYSID1); |
phy_id2 = ar7240sw_phy_read(mii, 0, MII_PHYSID2); |
if ((phy_id1 != AR7240_PHY_ID1 || phy_id2 != AR7240_PHY_ID2) && |
(phy_id1 != AR934X_PHY_ID1 || phy_id2 != AR934X_PHY_ID2)) { |
pr_err("%s: unknown phy id '%04x:%04x'\n", |
dev_name(&mii->dev), phy_id1, phy_id2); |
return NULL; |
} |
|
as = kzalloc(sizeof(*as), GFP_KERNEL); |
if (!as) |
return NULL; |
|
as->mii_bus = mii; |
as->swdata = pdata->switch_data; |
|
swdev = &as->swdev; |
|
ctrl = ar7240sw_reg_read(mii, AR7240_REG_MASK_CTRL); |
as->ver = (ctrl >> AR7240_MASK_CTRL_VERSION_S) & |
AR7240_MASK_CTRL_VERSION_M; |
|
if (sw_is_ar7240(as)) { |
swdev->name = "AR7240/AR9330 built-in switch"; |
swdev->ports = AR7240_NUM_PORTS - 1; |
} else if (sw_is_ar934x(as)) { |
swdev->name = "AR934X built-in switch"; |
|
if (pdata->phy_if_mode == PHY_INTERFACE_MODE_GMII) { |
ar7240sw_reg_set(mii, AR934X_REG_OPER_MODE0, |
AR934X_OPER_MODE0_MAC_GMII_EN); |
} else if (pdata->phy_if_mode == PHY_INTERFACE_MODE_MII) { |
ar7240sw_reg_set(mii, AR934X_REG_OPER_MODE0, |
AR934X_OPER_MODE0_PHY_MII_EN); |
} else { |
pr_err("%s: invalid PHY interface mode\n", |
dev_name(&mii->dev)); |
goto err_free; |
} |
|
if (as->swdata->phy4_mii_en) { |
ar7240sw_reg_set(mii, AR934X_REG_OPER_MODE1, |
AR934X_REG_OPER_MODE1_PHY4_MII_EN); |
swdev->ports = AR7240_NUM_PORTS - 1; |
} else { |
swdev->ports = AR7240_NUM_PORTS; |
} |
} else { |
pr_err("%s: unsupported chip, ctrl=%08x\n", |
dev_name(&mii->dev), ctrl); |
goto err_free; |
} |
|
swdev->cpu_port = AR7240_PORT_CPU; |
swdev->vlans = AR7240_MAX_VLANS; |
swdev->ops = &ar7240_ops; |
|
if (register_switch(&as->swdev, ag->dev) < 0) |
goto err_free; |
|
pr_info("%s: Found an %s\n", dev_name(&mii->dev), swdev->name); |
|
/* initialize defaults */ |
for (i = 0; i < AR7240_MAX_VLANS; i++) |
as->vlan_id[i] = i; |
|
as->vlan_table[0] = ar7240sw_port_mask_all(as); |
|
return as; |
|
err_free: |
kfree(as); |
return NULL; |
} |
|
static void link_function(struct work_struct *work) { |
struct ag71xx *ag = container_of(work, struct ag71xx, link_work.work); |
struct ar7240sw *as = ag->phy_priv; |
unsigned long flags; |
u8 mask; |
int i; |
int status = 0; |
|
mask = ~as->swdata->phy_poll_mask; |
for (i = 0; i < AR7240_NUM_PHYS; i++) { |
int link; |
|
if (!(mask & BIT(i))) |
continue; |
|
link = ar7240sw_phy_read(ag->mii_bus, i, MII_BMSR); |
if (link & BMSR_LSTATUS) { |
status = 1; |
break; |
} |
} |
|
spin_lock_irqsave(&ag->lock, flags); |
if (status != ag->link) { |
ag->link = status; |
ag71xx_link_adjust(ag); |
} |
spin_unlock_irqrestore(&ag->lock, flags); |
|
schedule_delayed_work(&ag->link_work, HZ / 2); |
} |
|
void ag71xx_ar7240_start(struct ag71xx *ag) |
{ |
struct ar7240sw *as = ag->phy_priv; |
|
ar7240sw_reset(as); |
|
ag->speed = SPEED_1000; |
ag->duplex = 1; |
|
ar7240_set_addr(as, ag->dev->dev_addr); |
ar7240_hw_apply(&as->swdev); |
|
schedule_delayed_work(&ag->link_work, HZ / 10); |
} |
|
void ag71xx_ar7240_stop(struct ag71xx *ag) |
{ |
cancel_delayed_work_sync(&ag->link_work); |
} |
|
int ag71xx_ar7240_init(struct ag71xx *ag) |
{ |
struct ar7240sw *as; |
|
as = ar7240_probe(ag); |
if (!as) |
return -ENODEV; |
|
ag->phy_priv = as; |
ar7240sw_reset(as); |
|
rwlock_init(&as->stats_lock); |
INIT_DELAYED_WORK(&ag->link_work, link_function); |
|
return 0; |
} |
|
void ag71xx_ar7240_cleanup(struct ag71xx *ag) |
{ |
struct ar7240sw *as = ag->phy_priv; |
|
if (!as) |
return; |
|
unregister_switch(&as->swdev); |
kfree(as); |
ag->phy_priv = NULL; |
} |
/branches/gl-inet/target/linux/ar71xx/files/drivers/net/ethernet/atheros/ag71xx/ag71xx_debugfs.c |
@@ -0,0 +1,285 @@ |
/* |
* Atheros AR71xx built-in ethernet mac driver |
* |
* Copyright (C) 2008-2010 Gabor Juhos <juhosg@openwrt.org> |
* Copyright (C) 2008 Imre Kaloz <kaloz@openwrt.org> |
* |
* Based on Atheros' AG7100 driver |
* |
* This program is free software; you can redistribute it and/or modify it |
* under the terms of the GNU General Public License version 2 as published |
* by the Free Software Foundation. |
*/ |
|
#include <linux/debugfs.h> |
|
#include "ag71xx.h" |
|
static struct dentry *ag71xx_debugfs_root; |
|
static int ag71xx_debugfs_generic_open(struct inode *inode, struct file *file) |
{ |
file->private_data = inode->i_private; |
return 0; |
} |
|
void ag71xx_debugfs_update_int_stats(struct ag71xx *ag, u32 status) |
{ |
if (status) |
ag->debug.int_stats.total++; |
if (status & AG71XX_INT_TX_PS) |
ag->debug.int_stats.tx_ps++; |
if (status & AG71XX_INT_TX_UR) |
ag->debug.int_stats.tx_ur++; |
if (status & AG71XX_INT_TX_BE) |
ag->debug.int_stats.tx_be++; |
if (status & AG71XX_INT_RX_PR) |
ag->debug.int_stats.rx_pr++; |
if (status & AG71XX_INT_RX_OF) |
ag->debug.int_stats.rx_of++; |
if (status & AG71XX_INT_RX_BE) |
ag->debug.int_stats.rx_be++; |
} |
|
static ssize_t read_file_int_stats(struct file *file, char __user *user_buf, |
size_t count, loff_t *ppos) |
{ |
#define PR_INT_STAT(_label, _field) \ |
len += snprintf(buf + len, sizeof(buf) - len, \ |
"%20s: %10lu\n", _label, ag->debug.int_stats._field); |
|
struct ag71xx *ag = file->private_data; |
char buf[256]; |
unsigned int len = 0; |
|
PR_INT_STAT("TX Packet Sent", tx_ps); |
PR_INT_STAT("TX Underrun", tx_ur); |
PR_INT_STAT("TX Bus Error", tx_be); |
PR_INT_STAT("RX Packet Received", rx_pr); |
PR_INT_STAT("RX Overflow", rx_of); |
PR_INT_STAT("RX Bus Error", rx_be); |
len += snprintf(buf + len, sizeof(buf) - len, "\n"); |
PR_INT_STAT("Total", total); |
|
return simple_read_from_buffer(user_buf, count, ppos, buf, len); |
#undef PR_INT_STAT |
} |
|
static const struct file_operations ag71xx_fops_int_stats = { |
.open = ag71xx_debugfs_generic_open, |
.read = read_file_int_stats, |
.owner = THIS_MODULE |
}; |
|
void ag71xx_debugfs_update_napi_stats(struct ag71xx *ag, int rx, int tx) |
{ |
struct ag71xx_napi_stats *stats = &ag->debug.napi_stats; |
|
if (rx) { |
stats->rx_count++; |
stats->rx_packets += rx; |
if (rx <= AG71XX_NAPI_WEIGHT) |
stats->rx[rx]++; |
if (rx > stats->rx_packets_max) |
stats->rx_packets_max = rx; |
} |
|
if (tx) { |
stats->tx_count++; |
stats->tx_packets += tx; |
if (tx <= AG71XX_NAPI_WEIGHT) |
stats->tx[tx]++; |
if (tx > stats->tx_packets_max) |
stats->tx_packets_max = tx; |
} |
} |
|
static ssize_t read_file_napi_stats(struct file *file, char __user *user_buf, |
size_t count, loff_t *ppos) |
{ |
struct ag71xx *ag = file->private_data; |
struct ag71xx_napi_stats *stats = &ag->debug.napi_stats; |
char *buf; |
unsigned int buflen; |
unsigned int len = 0; |
unsigned long rx_avg = 0; |
unsigned long tx_avg = 0; |
int ret; |
int i; |
|
buflen = 2048; |
buf = kmalloc(buflen, GFP_KERNEL); |
if (!buf) |
return -ENOMEM; |
|
if (stats->rx_count) |
rx_avg = stats->rx_packets / stats->rx_count; |
|
if (stats->tx_count) |
tx_avg = stats->tx_packets / stats->tx_count; |
|
len += snprintf(buf + len, buflen - len, "%3s %10s %10s\n", |
"len", "rx", "tx"); |
|
for (i = 1; i <= AG71XX_NAPI_WEIGHT; i++) |
len += snprintf(buf + len, buflen - len, |
"%3d: %10lu %10lu\n", |
i, stats->rx[i], stats->tx[i]); |
|
len += snprintf(buf + len, buflen - len, "\n"); |
|
len += snprintf(buf + len, buflen - len, "%3s: %10lu %10lu\n", |
"sum", stats->rx_count, stats->tx_count); |
len += snprintf(buf + len, buflen - len, "%3s: %10lu %10lu\n", |
"avg", rx_avg, tx_avg); |
len += snprintf(buf + len, buflen - len, "%3s: %10lu %10lu\n", |
"max", stats->rx_packets_max, stats->tx_packets_max); |
len += snprintf(buf + len, buflen - len, "%3s: %10lu %10lu\n", |
"pkt", stats->rx_packets, stats->tx_packets); |
|
ret = simple_read_from_buffer(user_buf, count, ppos, buf, len); |
kfree(buf); |
|
return ret; |
} |
|
static const struct file_operations ag71xx_fops_napi_stats = { |
.open = ag71xx_debugfs_generic_open, |
.read = read_file_napi_stats, |
.owner = THIS_MODULE |
}; |
|
#define DESC_PRINT_LEN 64 |
|
static ssize_t read_file_ring(struct file *file, char __user *user_buf, |
size_t count, loff_t *ppos, |
struct ag71xx *ag, |
struct ag71xx_ring *ring, |
unsigned desc_reg) |
{ |
int ring_size = BIT(ring->order); |
int ring_mask = ring_size - 1; |
char *buf; |
unsigned int buflen; |
unsigned int len = 0; |
unsigned long flags; |
ssize_t ret; |
int curr; |
int dirty; |
u32 desc_hw; |
int i; |
|
buflen = (ring_size * DESC_PRINT_LEN); |
buf = kmalloc(buflen, GFP_KERNEL); |
if (!buf) |
return -ENOMEM; |
|
len += snprintf(buf + len, buflen - len, |
"Idx ... %-8s %-8s %-8s %-8s .\n", |
"desc", "next", "data", "ctrl"); |
|
spin_lock_irqsave(&ag->lock, flags); |
|
curr = (ring->curr & ring_mask); |
dirty = (ring->dirty & ring_mask); |
desc_hw = ag71xx_rr(ag, desc_reg); |
for (i = 0; i < ring_size; i++) { |
struct ag71xx_desc *desc = ag71xx_ring_desc(ring, i); |
u32 desc_dma = ((u32) ring->descs_dma) + i * AG71XX_DESC_SIZE; |
|
len += snprintf(buf + len, buflen - len, |
"%3d %c%c%c %08x %08x %08x %08x %c\n", |
i, |
(i == curr) ? 'C' : ' ', |
(i == dirty) ? 'D' : ' ', |
(desc_hw == desc_dma) ? 'H' : ' ', |
desc_dma, |
desc->next, |
desc->data, |
desc->ctrl, |
(desc->ctrl & DESC_EMPTY) ? 'E' : '*'); |
} |
|
spin_unlock_irqrestore(&ag->lock, flags); |
|
ret = simple_read_from_buffer(user_buf, count, ppos, buf, len); |
kfree(buf); |
|
return ret; |
} |
|
static ssize_t read_file_tx_ring(struct file *file, char __user *user_buf, |
size_t count, loff_t *ppos) |
{ |
struct ag71xx *ag = file->private_data; |
|
return read_file_ring(file, user_buf, count, ppos, ag, &ag->tx_ring, |
AG71XX_REG_TX_DESC); |
} |
|
static const struct file_operations ag71xx_fops_tx_ring = { |
.open = ag71xx_debugfs_generic_open, |
.read = read_file_tx_ring, |
.owner = THIS_MODULE |
}; |
|
static ssize_t read_file_rx_ring(struct file *file, char __user *user_buf, |
size_t count, loff_t *ppos) |
{ |
struct ag71xx *ag = file->private_data; |
|
return read_file_ring(file, user_buf, count, ppos, ag, &ag->rx_ring, |
AG71XX_REG_RX_DESC); |
} |
|
static const struct file_operations ag71xx_fops_rx_ring = { |
.open = ag71xx_debugfs_generic_open, |
.read = read_file_rx_ring, |
.owner = THIS_MODULE |
}; |
|
void ag71xx_debugfs_exit(struct ag71xx *ag) |
{ |
debugfs_remove_recursive(ag->debug.debugfs_dir); |
} |
|
int ag71xx_debugfs_init(struct ag71xx *ag) |
{ |
struct device *dev = &ag->pdev->dev; |
|
ag->debug.debugfs_dir = debugfs_create_dir(dev_name(dev), |
ag71xx_debugfs_root); |
if (!ag->debug.debugfs_dir) { |
dev_err(dev, "unable to create debugfs directory\n"); |
return -ENOENT; |
} |
|
debugfs_create_file("int_stats", S_IRUGO, ag->debug.debugfs_dir, |
ag, &ag71xx_fops_int_stats); |
debugfs_create_file("napi_stats", S_IRUGO, ag->debug.debugfs_dir, |
ag, &ag71xx_fops_napi_stats); |
debugfs_create_file("tx_ring", S_IRUGO, ag->debug.debugfs_dir, |
ag, &ag71xx_fops_tx_ring); |
debugfs_create_file("rx_ring", S_IRUGO, ag->debug.debugfs_dir, |
ag, &ag71xx_fops_rx_ring); |
|
return 0; |
} |
|
int ag71xx_debugfs_root_init(void) |
{ |
if (ag71xx_debugfs_root) |
return -EBUSY; |
|
ag71xx_debugfs_root = debugfs_create_dir(KBUILD_MODNAME, NULL); |
if (!ag71xx_debugfs_root) |
return -ENOENT; |
|
return 0; |
} |
|
void ag71xx_debugfs_root_exit(void) |
{ |
debugfs_remove(ag71xx_debugfs_root); |
ag71xx_debugfs_root = NULL; |
} |
/branches/gl-inet/target/linux/ar71xx/files/drivers/net/ethernet/atheros/ag71xx/ag71xx_main.c |
@@ -0,0 +1,1479 @@ |
/* |
* Atheros AR71xx built-in ethernet mac driver |
* |
* Copyright (C) 2008-2010 Gabor Juhos <juhosg@openwrt.org> |
* Copyright (C) 2008 Imre Kaloz <kaloz@openwrt.org> |
* |
* Based on Atheros' AG7100 driver |
* |
* This program is free software; you can redistribute it and/or modify it |
* under the terms of the GNU General Public License version 2 as published |
* by the Free Software Foundation. |
*/ |
|
#include "ag71xx.h" |
|
#define AG71XX_DEFAULT_MSG_ENABLE \ |
(NETIF_MSG_DRV \ |
| NETIF_MSG_PROBE \ |
| NETIF_MSG_LINK \ |
| NETIF_MSG_TIMER \ |
| NETIF_MSG_IFDOWN \ |
| NETIF_MSG_IFUP \ |
| NETIF_MSG_RX_ERR \ |
| NETIF_MSG_TX_ERR) |
|
static int ag71xx_msg_level = -1; |
|
module_param_named(msg_level, ag71xx_msg_level, int, 0); |
MODULE_PARM_DESC(msg_level, "Message level (-1=defaults,0=none,...,16=all)"); |
|
#define ETH_SWITCH_HEADER_LEN 2 |
|
static int ag71xx_tx_packets(struct ag71xx *ag, bool flush); |
|
static inline unsigned int ag71xx_max_frame_len(unsigned int mtu) |
{ |
return ETH_SWITCH_HEADER_LEN + ETH_HLEN + VLAN_HLEN + mtu + ETH_FCS_LEN; |
} |
|
static void ag71xx_dump_dma_regs(struct ag71xx *ag) |
{ |
DBG("%s: dma_tx_ctrl=%08x, dma_tx_desc=%08x, dma_tx_status=%08x\n", |
ag->dev->name, |
ag71xx_rr(ag, AG71XX_REG_TX_CTRL), |
ag71xx_rr(ag, AG71XX_REG_TX_DESC), |
ag71xx_rr(ag, AG71XX_REG_TX_STATUS)); |
|
DBG("%s: dma_rx_ctrl=%08x, dma_rx_desc=%08x, dma_rx_status=%08x\n", |
ag->dev->name, |
ag71xx_rr(ag, AG71XX_REG_RX_CTRL), |
ag71xx_rr(ag, AG71XX_REG_RX_DESC), |
ag71xx_rr(ag, AG71XX_REG_RX_STATUS)); |
} |
|
static void ag71xx_dump_regs(struct ag71xx *ag) |
{ |
DBG("%s: mac_cfg1=%08x, mac_cfg2=%08x, ipg=%08x, hdx=%08x, mfl=%08x\n", |
ag->dev->name, |
ag71xx_rr(ag, AG71XX_REG_MAC_CFG1), |
ag71xx_rr(ag, AG71XX_REG_MAC_CFG2), |
ag71xx_rr(ag, AG71XX_REG_MAC_IPG), |
ag71xx_rr(ag, AG71XX_REG_MAC_HDX), |
ag71xx_rr(ag, AG71XX_REG_MAC_MFL)); |
DBG("%s: mac_ifctl=%08x, mac_addr1=%08x, mac_addr2=%08x\n", |
ag->dev->name, |
ag71xx_rr(ag, AG71XX_REG_MAC_IFCTL), |
ag71xx_rr(ag, AG71XX_REG_MAC_ADDR1), |
ag71xx_rr(ag, AG71XX_REG_MAC_ADDR2)); |
DBG("%s: fifo_cfg0=%08x, fifo_cfg1=%08x, fifo_cfg2=%08x\n", |
ag->dev->name, |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG0), |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG1), |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG2)); |
DBG("%s: fifo_cfg3=%08x, fifo_cfg4=%08x, fifo_cfg5=%08x\n", |
ag->dev->name, |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG3), |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG4), |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG5)); |
} |
|
static inline void ag71xx_dump_intr(struct ag71xx *ag, char *label, u32 intr) |
{ |
DBG("%s: %s intr=%08x %s%s%s%s%s%s\n", |
ag->dev->name, label, intr, |
(intr & AG71XX_INT_TX_PS) ? "TXPS " : "", |
(intr & AG71XX_INT_TX_UR) ? "TXUR " : "", |
(intr & AG71XX_INT_TX_BE) ? "TXBE " : "", |
(intr & AG71XX_INT_RX_PR) ? "RXPR " : "", |
(intr & AG71XX_INT_RX_OF) ? "RXOF " : "", |
(intr & AG71XX_INT_RX_BE) ? "RXBE " : ""); |
} |
|
static void ag71xx_ring_tx_clean(struct ag71xx *ag) |
{ |
struct ag71xx_ring *ring = &ag->tx_ring; |
struct net_device *dev = ag->dev; |
int ring_mask = BIT(ring->order) - 1; |
u32 bytes_compl = 0, pkts_compl = 0; |
|
while (ring->curr != ring->dirty) { |
struct ag71xx_desc *desc; |
u32 i = ring->dirty & ring_mask; |
|
desc = ag71xx_ring_desc(ring, i); |
if (!ag71xx_desc_empty(desc)) { |
desc->ctrl = 0; |
dev->stats.tx_errors++; |
} |
|
if (ring->buf[i].skb) { |
bytes_compl += ring->buf[i].len; |
pkts_compl++; |
dev_kfree_skb_any(ring->buf[i].skb); |
} |
ring->buf[i].skb = NULL; |
ring->dirty++; |
} |
|
/* flush descriptors */ |
wmb(); |
|
netdev_completed_queue(dev, pkts_compl, bytes_compl); |
} |
|
static void ag71xx_ring_tx_init(struct ag71xx *ag) |
{ |
struct ag71xx_ring *ring = &ag->tx_ring; |
int ring_size = BIT(ring->order); |
int ring_mask = ring_size - 1; |
int i; |
|
for (i = 0; i < ring_size; i++) { |
struct ag71xx_desc *desc = ag71xx_ring_desc(ring, i); |
|
desc->next = (u32) (ring->descs_dma + |
AG71XX_DESC_SIZE * ((i + 1) & ring_mask)); |
|
desc->ctrl = DESC_EMPTY; |
ring->buf[i].skb = NULL; |
} |
|
/* flush descriptors */ |
wmb(); |
|
ring->curr = 0; |
ring->dirty = 0; |
netdev_reset_queue(ag->dev); |
} |
|
static void ag71xx_ring_rx_clean(struct ag71xx *ag) |
{ |
struct ag71xx_ring *ring = &ag->rx_ring; |
int ring_size = BIT(ring->order); |
int i; |
|
if (!ring->buf) |
return; |
|
for (i = 0; i < ring_size; i++) |
if (ring->buf[i].rx_buf) { |
dma_unmap_single(&ag->dev->dev, ring->buf[i].dma_addr, |
ag->rx_buf_size, DMA_FROM_DEVICE); |
skb_free_frag(ring->buf[i].rx_buf); |
} |
} |
|
static int ag71xx_buffer_offset(struct ag71xx *ag) |
{ |
int offset = NET_SKB_PAD; |
|
/* |
* On AR71xx/AR91xx packets must be 4-byte aligned. |
* |
* When using builtin AR8216 support, hardware adds a 2-byte header, |
* so we don't need any extra alignment in that case. |
*/ |
if (!ag71xx_get_pdata(ag)->is_ar724x || ag71xx_has_ar8216(ag)) |
return offset; |
|
return offset + NET_IP_ALIGN; |
} |
|
static int ag71xx_buffer_size(struct ag71xx *ag) |
{ |
return ag->rx_buf_size + |
SKB_DATA_ALIGN(sizeof(struct skb_shared_info)); |
} |
|
static bool ag71xx_fill_rx_buf(struct ag71xx *ag, struct ag71xx_buf *buf, |
int offset, |
void *(*alloc)(unsigned int size)) |
{ |
struct ag71xx_ring *ring = &ag->rx_ring; |
struct ag71xx_desc *desc = ag71xx_ring_desc(ring, buf - &ring->buf[0]); |
void *data; |
|
data = alloc(ag71xx_buffer_size(ag)); |
if (!data) |
return false; |
|
buf->rx_buf = data; |
buf->dma_addr = dma_map_single(&ag->dev->dev, data, ag->rx_buf_size, |
DMA_FROM_DEVICE); |
desc->data = (u32) buf->dma_addr + offset; |
return true; |
} |
|
static int ag71xx_ring_rx_init(struct ag71xx *ag) |
{ |
struct ag71xx_ring *ring = &ag->rx_ring; |
int ring_size = BIT(ring->order); |
int ring_mask = BIT(ring->order) - 1; |
unsigned int i; |
int ret; |
int offset = ag71xx_buffer_offset(ag); |
|
ret = 0; |
for (i = 0; i < ring_size; i++) { |
struct ag71xx_desc *desc = ag71xx_ring_desc(ring, i); |
|
desc->next = (u32) (ring->descs_dma + |
AG71XX_DESC_SIZE * ((i + 1) & ring_mask)); |
|
DBG("ag71xx: RX desc at %p, next is %08x\n", |
desc, desc->next); |
} |
|
for (i = 0; i < ring_size; i++) { |
struct ag71xx_desc *desc = ag71xx_ring_desc(ring, i); |
|
if (!ag71xx_fill_rx_buf(ag, &ring->buf[i], offset, |
netdev_alloc_frag)) { |
ret = -ENOMEM; |
break; |
} |
|
desc->ctrl = DESC_EMPTY; |
} |
|
/* flush descriptors */ |
wmb(); |
|
ring->curr = 0; |
ring->dirty = 0; |
|
return ret; |
} |
|
static int ag71xx_ring_rx_refill(struct ag71xx *ag) |
{ |
struct ag71xx_ring *ring = &ag->rx_ring; |
int ring_mask = BIT(ring->order) - 1; |
unsigned int count; |
int offset = ag71xx_buffer_offset(ag); |
|
count = 0; |
for (; ring->curr - ring->dirty > 0; ring->dirty++) { |
struct ag71xx_desc *desc; |
unsigned int i; |
|
i = ring->dirty & ring_mask; |
desc = ag71xx_ring_desc(ring, i); |
|
if (!ring->buf[i].rx_buf && |
!ag71xx_fill_rx_buf(ag, &ring->buf[i], offset, |
napi_alloc_frag)) |
break; |
|
desc->ctrl = DESC_EMPTY; |
count++; |
} |
|
/* flush descriptors */ |
wmb(); |
|
DBG("%s: %u rx descriptors refilled\n", ag->dev->name, count); |
|
return count; |
} |
|
static int ag71xx_rings_init(struct ag71xx *ag) |
{ |
struct ag71xx_ring *tx = &ag->tx_ring; |
struct ag71xx_ring *rx = &ag->rx_ring; |
int ring_size = BIT(tx->order) + BIT(rx->order); |
int tx_size = BIT(tx->order); |
|
tx->buf = kzalloc(ring_size * sizeof(*tx->buf), GFP_KERNEL); |
if (!tx->buf) |
return -ENOMEM; |
|
tx->descs_cpu = dma_alloc_coherent(NULL, ring_size * AG71XX_DESC_SIZE, |
&tx->descs_dma, GFP_ATOMIC); |
if (!tx->descs_cpu) { |
kfree(tx->buf); |
tx->buf = NULL; |
return -ENOMEM; |
} |
|
rx->buf = &tx->buf[BIT(tx->order)]; |
rx->descs_cpu = ((void *)tx->descs_cpu) + tx_size * AG71XX_DESC_SIZE; |
rx->descs_dma = tx->descs_dma + tx_size * AG71XX_DESC_SIZE; |
|
ag71xx_ring_tx_init(ag); |
return ag71xx_ring_rx_init(ag); |
} |
|
static void ag71xx_rings_free(struct ag71xx *ag) |
{ |
struct ag71xx_ring *tx = &ag->tx_ring; |
struct ag71xx_ring *rx = &ag->rx_ring; |
int ring_size = BIT(tx->order) + BIT(rx->order); |
|
if (tx->descs_cpu) |
dma_free_coherent(NULL, ring_size * AG71XX_DESC_SIZE, |
tx->descs_cpu, tx->descs_dma); |
|
kfree(tx->buf); |
|
tx->descs_cpu = NULL; |
rx->descs_cpu = NULL; |
tx->buf = NULL; |
rx->buf = NULL; |
} |
|
static void ag71xx_rings_cleanup(struct ag71xx *ag) |
{ |
ag71xx_ring_rx_clean(ag); |
ag71xx_ring_tx_clean(ag); |
ag71xx_rings_free(ag); |
|
netdev_reset_queue(ag->dev); |
} |
|
static unsigned char *ag71xx_speed_str(struct ag71xx *ag) |
{ |
switch (ag->speed) { |
case SPEED_1000: |
return "1000"; |
case SPEED_100: |
return "100"; |
case SPEED_10: |
return "10"; |
} |
|
return "?"; |
} |
|
static void ag71xx_hw_set_macaddr(struct ag71xx *ag, unsigned char *mac) |
{ |
u32 t; |
|
t = (((u32) mac[5]) << 24) | (((u32) mac[4]) << 16) |
| (((u32) mac[3]) << 8) | ((u32) mac[2]); |
|
ag71xx_wr(ag, AG71XX_REG_MAC_ADDR1, t); |
|
t = (((u32) mac[1]) << 24) | (((u32) mac[0]) << 16); |
ag71xx_wr(ag, AG71XX_REG_MAC_ADDR2, t); |
} |
|
static void ag71xx_dma_reset(struct ag71xx *ag) |
{ |
u32 val; |
int i; |
|
ag71xx_dump_dma_regs(ag); |
|
/* stop RX and TX */ |
ag71xx_wr(ag, AG71XX_REG_RX_CTRL, 0); |
ag71xx_wr(ag, AG71XX_REG_TX_CTRL, 0); |
|
/* |
* give the hardware some time to really stop all rx/tx activity |
* clearing the descriptors too early causes random memory corruption |
*/ |
mdelay(1); |
|
/* clear descriptor addresses */ |
ag71xx_wr(ag, AG71XX_REG_TX_DESC, ag->stop_desc_dma); |
ag71xx_wr(ag, AG71XX_REG_RX_DESC, ag->stop_desc_dma); |
|
/* clear pending RX/TX interrupts */ |
for (i = 0; i < 256; i++) { |
ag71xx_wr(ag, AG71XX_REG_RX_STATUS, RX_STATUS_PR); |
ag71xx_wr(ag, AG71XX_REG_TX_STATUS, TX_STATUS_PS); |
} |
|
/* clear pending errors */ |
ag71xx_wr(ag, AG71XX_REG_RX_STATUS, RX_STATUS_BE | RX_STATUS_OF); |
ag71xx_wr(ag, AG71XX_REG_TX_STATUS, TX_STATUS_BE | TX_STATUS_UR); |
|
val = ag71xx_rr(ag, AG71XX_REG_RX_STATUS); |
if (val) |
pr_alert("%s: unable to clear DMA Rx status: %08x\n", |
ag->dev->name, val); |
|
val = ag71xx_rr(ag, AG71XX_REG_TX_STATUS); |
|
/* mask out reserved bits */ |
val &= ~0xff000000; |
|
if (val) |
pr_alert("%s: unable to clear DMA Tx status: %08x\n", |
ag->dev->name, val); |
|
ag71xx_dump_dma_regs(ag); |
} |
|
#define MAC_CFG1_INIT (MAC_CFG1_RXE | MAC_CFG1_TXE | \ |
MAC_CFG1_SRX | MAC_CFG1_STX) |
|
#define FIFO_CFG0_INIT (FIFO_CFG0_ALL << FIFO_CFG0_ENABLE_SHIFT) |
|
#define FIFO_CFG4_INIT (FIFO_CFG4_DE | FIFO_CFG4_DV | FIFO_CFG4_FC | \ |
FIFO_CFG4_CE | FIFO_CFG4_CR | FIFO_CFG4_LM | \ |
FIFO_CFG4_LO | FIFO_CFG4_OK | FIFO_CFG4_MC | \ |
FIFO_CFG4_BC | FIFO_CFG4_DR | FIFO_CFG4_LE | \ |
FIFO_CFG4_CF | FIFO_CFG4_PF | FIFO_CFG4_UO | \ |
FIFO_CFG4_VT) |
|
#define FIFO_CFG5_INIT (FIFO_CFG5_DE | FIFO_CFG5_DV | FIFO_CFG5_FC | \ |
FIFO_CFG5_CE | FIFO_CFG5_LO | FIFO_CFG5_OK | \ |
FIFO_CFG5_MC | FIFO_CFG5_BC | FIFO_CFG5_DR | \ |
FIFO_CFG5_CF | FIFO_CFG5_PF | FIFO_CFG5_VT | \ |
FIFO_CFG5_LE | FIFO_CFG5_FT | FIFO_CFG5_16 | \ |
FIFO_CFG5_17 | FIFO_CFG5_SF) |
|
static void ag71xx_hw_stop(struct ag71xx *ag) |
{ |
/* disable all interrupts and stop the rx/tx engine */ |
ag71xx_wr(ag, AG71XX_REG_INT_ENABLE, 0); |
ag71xx_wr(ag, AG71XX_REG_RX_CTRL, 0); |
ag71xx_wr(ag, AG71XX_REG_TX_CTRL, 0); |
} |
|
static void ag71xx_hw_setup(struct ag71xx *ag) |
{ |
struct ag71xx_platform_data *pdata = ag71xx_get_pdata(ag); |
u32 init = MAC_CFG1_INIT; |
|
/* setup MAC configuration registers */ |
if (pdata->use_flow_control) |
init |= MAC_CFG1_TFC | MAC_CFG1_RFC; |
ag71xx_wr(ag, AG71XX_REG_MAC_CFG1, init); |
|
ag71xx_sb(ag, AG71XX_REG_MAC_CFG2, |
MAC_CFG2_PAD_CRC_EN | MAC_CFG2_LEN_CHECK); |
|
/* setup max frame length to zero */ |
ag71xx_wr(ag, AG71XX_REG_MAC_MFL, 0); |
|
/* setup FIFO configuration registers */ |
ag71xx_wr(ag, AG71XX_REG_FIFO_CFG0, FIFO_CFG0_INIT); |
if (pdata->is_ar724x) { |
ag71xx_wr(ag, AG71XX_REG_FIFO_CFG1, 0x0010ffff); |
ag71xx_wr(ag, AG71XX_REG_FIFO_CFG2, 0x015500aa); |
} else { |
ag71xx_wr(ag, AG71XX_REG_FIFO_CFG1, 0x0fff0000); |
ag71xx_wr(ag, AG71XX_REG_FIFO_CFG2, 0x00001fff); |
} |
ag71xx_wr(ag, AG71XX_REG_FIFO_CFG4, FIFO_CFG4_INIT); |
ag71xx_wr(ag, AG71XX_REG_FIFO_CFG5, FIFO_CFG5_INIT); |
} |
|
static void ag71xx_hw_init(struct ag71xx *ag) |
{ |
struct ag71xx_platform_data *pdata = ag71xx_get_pdata(ag); |
u32 reset_mask = pdata->reset_bit; |
|
ag71xx_hw_stop(ag); |
|
if (pdata->is_ar724x) { |
u32 reset_phy = reset_mask; |
|
reset_phy &= AR71XX_RESET_GE0_PHY | AR71XX_RESET_GE1_PHY; |
reset_mask &= ~(AR71XX_RESET_GE0_PHY | AR71XX_RESET_GE1_PHY); |
|
ath79_device_reset_set(reset_phy); |
msleep(50); |
ath79_device_reset_clear(reset_phy); |
msleep(200); |
} |
|
ag71xx_sb(ag, AG71XX_REG_MAC_CFG1, MAC_CFG1_SR); |
udelay(20); |
|
ath79_device_reset_set(reset_mask); |
msleep(100); |
ath79_device_reset_clear(reset_mask); |
msleep(200); |
|
ag71xx_hw_setup(ag); |
|
ag71xx_dma_reset(ag); |
} |
|
static void ag71xx_fast_reset(struct ag71xx *ag) |
{ |
struct ag71xx_platform_data *pdata = ag71xx_get_pdata(ag); |
struct net_device *dev = ag->dev; |
u32 reset_mask = pdata->reset_bit; |
u32 rx_ds; |
u32 mii_reg; |
|
reset_mask &= AR71XX_RESET_GE0_MAC | AR71XX_RESET_GE1_MAC; |
|
ag71xx_hw_stop(ag); |
wmb(); |
|
mii_reg = ag71xx_rr(ag, AG71XX_REG_MII_CFG); |
rx_ds = ag71xx_rr(ag, AG71XX_REG_RX_DESC); |
|
ag71xx_tx_packets(ag, true); |
|
ath79_device_reset_set(reset_mask); |
udelay(10); |
ath79_device_reset_clear(reset_mask); |
udelay(10); |
|
ag71xx_dma_reset(ag); |
ag71xx_hw_setup(ag); |
ag->tx_ring.curr = 0; |
ag->tx_ring.dirty = 0; |
netdev_reset_queue(ag->dev); |
|
/* setup max frame length */ |
ag71xx_wr(ag, AG71XX_REG_MAC_MFL, |
ag71xx_max_frame_len(ag->dev->mtu)); |
|
ag71xx_wr(ag, AG71XX_REG_RX_DESC, rx_ds); |
ag71xx_wr(ag, AG71XX_REG_TX_DESC, ag->tx_ring.descs_dma); |
ag71xx_wr(ag, AG71XX_REG_MII_CFG, mii_reg); |
|
ag71xx_hw_set_macaddr(ag, dev->dev_addr); |
} |
|
static void ag71xx_hw_start(struct ag71xx *ag) |
{ |
/* start RX engine */ |
ag71xx_wr(ag, AG71XX_REG_RX_CTRL, RX_CTRL_RXE); |
|
/* enable interrupts */ |
ag71xx_wr(ag, AG71XX_REG_INT_ENABLE, AG71XX_INT_INIT); |
|
netif_wake_queue(ag->dev); |
} |
|
static void |
__ag71xx_link_adjust(struct ag71xx *ag, bool update) |
{ |
struct ag71xx_platform_data *pdata = ag71xx_get_pdata(ag); |
u32 cfg2; |
u32 ifctl; |
u32 fifo5; |
u32 fifo3; |
|
if (!ag->link && update) { |
ag71xx_hw_stop(ag); |
netif_carrier_off(ag->dev); |
if (netif_msg_link(ag)) |
pr_info("%s: link down\n", ag->dev->name); |
return; |
} |
|
if (pdata->is_ar724x) |
ag71xx_fast_reset(ag); |
|
cfg2 = ag71xx_rr(ag, AG71XX_REG_MAC_CFG2); |
cfg2 &= ~(MAC_CFG2_IF_1000 | MAC_CFG2_IF_10_100 | MAC_CFG2_FDX); |
cfg2 |= (ag->duplex) ? MAC_CFG2_FDX : 0; |
|
ifctl = ag71xx_rr(ag, AG71XX_REG_MAC_IFCTL); |
ifctl &= ~(MAC_IFCTL_SPEED); |
|
fifo5 = ag71xx_rr(ag, AG71XX_REG_FIFO_CFG5); |
fifo5 &= ~FIFO_CFG5_BM; |
|
switch (ag->speed) { |
case SPEED_1000: |
cfg2 |= MAC_CFG2_IF_1000; |
fifo5 |= FIFO_CFG5_BM; |
break; |
case SPEED_100: |
cfg2 |= MAC_CFG2_IF_10_100; |
ifctl |= MAC_IFCTL_SPEED; |
break; |
case SPEED_10: |
cfg2 |= MAC_CFG2_IF_10_100; |
break; |
default: |
BUG(); |
return; |
} |
|
if (pdata->is_ar91xx) |
fifo3 = 0x00780fff; |
else if (pdata->is_ar724x) |
fifo3 = 0x01f00140; |
else |
fifo3 = 0x008001ff; |
|
if (ag->tx_ring.desc_split) { |
fifo3 &= 0xffff; |
fifo3 |= ((2048 - ag->tx_ring.desc_split) / 4) << 16; |
} |
|
ag71xx_wr(ag, AG71XX_REG_FIFO_CFG3, fifo3); |
|
if (update && pdata->set_speed) |
pdata->set_speed(ag->speed); |
|
ag71xx_wr(ag, AG71XX_REG_MAC_CFG2, cfg2); |
ag71xx_wr(ag, AG71XX_REG_FIFO_CFG5, fifo5); |
ag71xx_wr(ag, AG71XX_REG_MAC_IFCTL, ifctl); |
|
if (pdata->disable_inline_checksum_engine) { |
/* |
* The rx ring buffer can stall on small packets on QCA953x and |
* QCA956x. Disabling the inline checksum engine fixes the stall. |
* The wr, rr functions cannot be used since this hidden register |
* is outside of the normal ag71xx register block. |
*/ |
void __iomem *dam = ioremap_nocache(0xb90001bc, 0x4); |
if (dam) { |
__raw_writel(__raw_readl(dam) & ~BIT(27), dam); |
(void)__raw_readl(dam); |
iounmap(dam); |
} |
} |
|
ag71xx_hw_start(ag); |
|
netif_carrier_on(ag->dev); |
if (update && netif_msg_link(ag)) |
pr_info("%s: link up (%sMbps/%s duplex)\n", |
ag->dev->name, |
ag71xx_speed_str(ag), |
(DUPLEX_FULL == ag->duplex) ? "Full" : "Half"); |
|
DBG("%s: fifo_cfg0=%#x, fifo_cfg1=%#x, fifo_cfg2=%#x\n", |
ag->dev->name, |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG0), |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG1), |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG2)); |
|
DBG("%s: fifo_cfg3=%#x, fifo_cfg4=%#x, fifo_cfg5=%#x\n", |
ag->dev->name, |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG3), |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG4), |
ag71xx_rr(ag, AG71XX_REG_FIFO_CFG5)); |
|
DBG("%s: mac_cfg2=%#x, mac_ifctl=%#x\n", |
ag->dev->name, |
ag71xx_rr(ag, AG71XX_REG_MAC_CFG2), |
ag71xx_rr(ag, AG71XX_REG_MAC_IFCTL)); |
} |
|
void ag71xx_link_adjust(struct ag71xx *ag) |
{ |
__ag71xx_link_adjust(ag, true); |
} |
|
static int ag71xx_hw_enable(struct ag71xx *ag) |
{ |
int ret; |
|
ret = ag71xx_rings_init(ag); |
if (ret) |
return ret; |
|
napi_enable(&ag->napi); |
ag71xx_wr(ag, AG71XX_REG_TX_DESC, ag->tx_ring.descs_dma); |
ag71xx_wr(ag, AG71XX_REG_RX_DESC, ag->rx_ring.descs_dma); |
netif_start_queue(ag->dev); |
|
return 0; |
} |
|
static void ag71xx_hw_disable(struct ag71xx *ag) |
{ |
unsigned long flags; |
|
spin_lock_irqsave(&ag->lock, flags); |
|
netif_stop_queue(ag->dev); |
|
ag71xx_hw_stop(ag); |
ag71xx_dma_reset(ag); |
|
napi_disable(&ag->napi); |
del_timer_sync(&ag->oom_timer); |
|
spin_unlock_irqrestore(&ag->lock, flags); |
|
ag71xx_rings_cleanup(ag); |
} |
|
static int ag71xx_open(struct net_device *dev) |
{ |
struct ag71xx *ag = netdev_priv(dev); |
unsigned int max_frame_len; |
int ret; |
|
netif_carrier_off(dev); |
max_frame_len = ag71xx_max_frame_len(dev->mtu); |
ag->rx_buf_size = SKB_DATA_ALIGN(max_frame_len + NET_SKB_PAD + NET_IP_ALIGN); |
|
/* setup max frame length */ |
ag71xx_wr(ag, AG71XX_REG_MAC_MFL, max_frame_len); |
ag71xx_hw_set_macaddr(ag, dev->dev_addr); |
|
ret = ag71xx_hw_enable(ag); |
if (ret) |
goto err; |
|
ag71xx_phy_start(ag); |
|
return 0; |
|
err: |
ag71xx_rings_cleanup(ag); |
return ret; |
} |
|
static int ag71xx_stop(struct net_device *dev) |
{ |
struct ag71xx *ag = netdev_priv(dev); |
|
netif_carrier_off(dev); |
ag71xx_phy_stop(ag); |
ag71xx_hw_disable(ag); |
|
return 0; |
} |
|
static int ag71xx_fill_dma_desc(struct ag71xx_ring *ring, u32 addr, int len) |
{ |
int i; |
struct ag71xx_desc *desc; |
int ring_mask = BIT(ring->order) - 1; |
int ndesc = 0; |
int split = ring->desc_split; |
|
if (!split) |
split = len; |
|
while (len > 0) { |
unsigned int cur_len = len; |
|
i = (ring->curr + ndesc) & ring_mask; |
desc = ag71xx_ring_desc(ring, i); |
|
if (!ag71xx_desc_empty(desc)) |
return -1; |
|
if (cur_len > split) { |
cur_len = split; |
|
/* |
* TX will hang if DMA transfers <= 4 bytes, |
* make sure next segment is more than 4 bytes long. |
*/ |
if (len <= split + 4) |
cur_len -= 4; |
} |
|
desc->data = addr; |
addr += cur_len; |
len -= cur_len; |
|
if (len > 0) |
cur_len |= DESC_MORE; |
|
/* prevent early tx attempt of this descriptor */ |
if (!ndesc) |
cur_len |= DESC_EMPTY; |
|
desc->ctrl = cur_len; |
ndesc++; |
} |
|
return ndesc; |
} |
|
static netdev_tx_t ag71xx_hard_start_xmit(struct sk_buff *skb, |
struct net_device *dev) |
{ |
struct ag71xx *ag = netdev_priv(dev); |
struct ag71xx_ring *ring = &ag->tx_ring; |
int ring_mask = BIT(ring->order) - 1; |
int ring_size = BIT(ring->order); |
struct ag71xx_desc *desc; |
dma_addr_t dma_addr; |
int i, n, ring_min; |
|
if (ag71xx_has_ar8216(ag)) |
ag71xx_add_ar8216_header(ag, skb); |
|
if (skb->len <= 4) { |
DBG("%s: packet len is too small\n", ag->dev->name); |
goto err_drop; |
} |
|
dma_addr = dma_map_single(&dev->dev, skb->data, skb->len, |
DMA_TO_DEVICE); |
|
i = ring->curr & ring_mask; |
desc = ag71xx_ring_desc(ring, i); |
|
/* setup descriptor fields */ |
n = ag71xx_fill_dma_desc(ring, (u32) dma_addr, skb->len & ag->desc_pktlen_mask); |
if (n < 0) |
goto err_drop_unmap; |
|
i = (ring->curr + n - 1) & ring_mask; |
ring->buf[i].len = skb->len; |
ring->buf[i].skb = skb; |
|
netdev_sent_queue(dev, skb->len); |
|
skb_tx_timestamp(skb); |
|
desc->ctrl &= ~DESC_EMPTY; |
ring->curr += n; |
|
/* flush descriptor */ |
wmb(); |
|
ring_min = 2; |
if (ring->desc_split) |
ring_min *= AG71XX_TX_RING_DS_PER_PKT; |
|
if (ring->curr - ring->dirty >= ring_size - ring_min) { |
DBG("%s: tx queue full\n", dev->name); |
netif_stop_queue(dev); |
} |
|
DBG("%s: packet injected into TX queue\n", ag->dev->name); |
|
/* enable TX engine */ |
ag71xx_wr(ag, AG71XX_REG_TX_CTRL, TX_CTRL_TXE); |
|
return NETDEV_TX_OK; |
|
err_drop_unmap: |
dma_unmap_single(&dev->dev, dma_addr, skb->len, DMA_TO_DEVICE); |
|
err_drop: |
dev->stats.tx_dropped++; |
|
dev_kfree_skb(skb); |
return NETDEV_TX_OK; |
} |
|
static int ag71xx_do_ioctl(struct net_device *dev, struct ifreq *ifr, int cmd) |
{ |
struct ag71xx *ag = netdev_priv(dev); |
int ret; |
|
switch (cmd) { |
case SIOCETHTOOL: |
if (ag->phy_dev == NULL) |
break; |
|
spin_lock_irq(&ag->lock); |
ret = phy_ethtool_ioctl(ag->phy_dev, (void *) ifr->ifr_data); |
spin_unlock_irq(&ag->lock); |
return ret; |
|
case SIOCSIFHWADDR: |
if (copy_from_user |
(dev->dev_addr, ifr->ifr_data, sizeof(dev->dev_addr))) |
return -EFAULT; |
return 0; |
|
case SIOCGIFHWADDR: |
if (copy_to_user |
(ifr->ifr_data, dev->dev_addr, sizeof(dev->dev_addr))) |
return -EFAULT; |
return 0; |
|
case SIOCGMIIPHY: |
case SIOCGMIIREG: |
case SIOCSMIIREG: |
if (ag->phy_dev == NULL) |
break; |
|
return phy_mii_ioctl(ag->phy_dev, ifr, cmd); |
|
default: |
break; |
} |
|
return -EOPNOTSUPP; |
} |
|
static void ag71xx_oom_timer_handler(unsigned long data) |
{ |
struct net_device *dev = (struct net_device *) data; |
struct ag71xx *ag = netdev_priv(dev); |
|
napi_schedule(&ag->napi); |
} |
|
static void ag71xx_tx_timeout(struct net_device *dev) |
{ |
struct ag71xx *ag = netdev_priv(dev); |
|
if (netif_msg_tx_err(ag)) |
pr_info("%s: tx timeout\n", ag->dev->name); |
|
schedule_delayed_work(&ag->restart_work, 1); |
} |
|
static void ag71xx_restart_work_func(struct work_struct *work) |
{ |
struct ag71xx *ag = container_of(work, struct ag71xx, restart_work.work); |
|
rtnl_lock(); |
ag71xx_hw_disable(ag); |
ag71xx_hw_enable(ag); |
if (ag->link) |
__ag71xx_link_adjust(ag, false); |
rtnl_unlock(); |
} |
|
static bool ag71xx_check_dma_stuck(struct ag71xx *ag) |
{ |
unsigned long timestamp; |
u32 rx_sm, tx_sm, rx_fd; |
|
timestamp = netdev_get_tx_queue(ag->dev, 0)->trans_start; |
if (likely(time_before(jiffies, timestamp + HZ/10))) |
return false; |
|
if (!netif_carrier_ok(ag->dev)) |
return false; |
|
rx_sm = ag71xx_rr(ag, AG71XX_REG_RX_SM); |
if ((rx_sm & 0x7) == 0x3 && ((rx_sm >> 4) & 0x7) == 0x6) |
return true; |
|
tx_sm = ag71xx_rr(ag, AG71XX_REG_TX_SM); |
rx_fd = ag71xx_rr(ag, AG71XX_REG_FIFO_DEPTH); |
if (((tx_sm >> 4) & 0x7) == 0 && ((rx_sm & 0x7) == 0) && |
((rx_sm >> 4) & 0x7) == 0 && rx_fd == 0) |
return true; |
|
return false; |
} |
|
static int ag71xx_tx_packets(struct ag71xx *ag, bool flush) |
{ |
struct ag71xx_ring *ring = &ag->tx_ring; |
struct ag71xx_platform_data *pdata = ag71xx_get_pdata(ag); |
bool dma_stuck = false; |
int ring_mask = BIT(ring->order) - 1; |
int ring_size = BIT(ring->order); |
int sent = 0; |
int bytes_compl = 0; |
int n = 0; |
|
DBG("%s: processing TX ring\n", ag->dev->name); |
|
while (ring->dirty + n != ring->curr) { |
unsigned int i = (ring->dirty + n) & ring_mask; |
struct ag71xx_desc *desc = ag71xx_ring_desc(ring, i); |
struct sk_buff *skb = ring->buf[i].skb; |
|
if (!flush && !ag71xx_desc_empty(desc)) { |
if (pdata->is_ar724x && |
ag71xx_check_dma_stuck(ag)) { |
schedule_delayed_work(&ag->restart_work, HZ / 2); |
dma_stuck = true; |
} |
break; |
} |
|
if (flush) |
desc->ctrl |= DESC_EMPTY; |
|
n++; |
if (!skb) |
continue; |
|
dev_kfree_skb_any(skb); |
ring->buf[i].skb = NULL; |
|
bytes_compl += ring->buf[i].len; |
|
sent++; |
ring->dirty += n; |
|
while (n > 0) { |
ag71xx_wr(ag, AG71XX_REG_TX_STATUS, TX_STATUS_PS); |
n--; |
} |
} |
|
DBG("%s: %d packets sent out\n", ag->dev->name, sent); |
|
if (!sent) |
return 0; |
|
ag->dev->stats.tx_bytes += bytes_compl; |
ag->dev->stats.tx_packets += sent; |
|
netdev_completed_queue(ag->dev, sent, bytes_compl); |
if ((ring->curr - ring->dirty) < (ring_size * 3) / 4) |
netif_wake_queue(ag->dev); |
|
if (!dma_stuck) |
cancel_delayed_work(&ag->restart_work); |
|
return sent; |
} |
|
static int ag71xx_rx_packets(struct ag71xx *ag, int limit) |
{ |
struct net_device *dev = ag->dev; |
struct ag71xx_ring *ring = &ag->rx_ring; |
int offset = ag71xx_buffer_offset(ag); |
unsigned int pktlen_mask = ag->desc_pktlen_mask; |
int ring_mask = BIT(ring->order) - 1; |
int ring_size = BIT(ring->order); |
struct sk_buff_head queue; |
struct sk_buff *skb; |
int done = 0; |
|
DBG("%s: rx packets, limit=%d, curr=%u, dirty=%u\n", |
dev->name, limit, ring->curr, ring->dirty); |
|
skb_queue_head_init(&queue); |
|
while (done < limit) { |
unsigned int i = ring->curr & ring_mask; |
struct ag71xx_desc *desc = ag71xx_ring_desc(ring, i); |
int pktlen; |
int err = 0; |
|
if (ag71xx_desc_empty(desc)) |
break; |
|
if ((ring->dirty + ring_size) == ring->curr) { |
ag71xx_assert(0); |
break; |
} |
|
ag71xx_wr(ag, AG71XX_REG_RX_STATUS, RX_STATUS_PR); |
|
pktlen = desc->ctrl & pktlen_mask; |
pktlen -= ETH_FCS_LEN; |
|
dma_unmap_single(&dev->dev, ring->buf[i].dma_addr, |
ag->rx_buf_size, DMA_FROM_DEVICE); |
|
dev->stats.rx_packets++; |
dev->stats.rx_bytes += pktlen; |
|
skb = build_skb(ring->buf[i].rx_buf, ag71xx_buffer_size(ag)); |
if (!skb) { |
skb_free_frag(ring->buf[i].rx_buf); |
goto next; |
} |
|
skb_reserve(skb, offset); |
skb_put(skb, pktlen); |
|
if (ag71xx_has_ar8216(ag)) |
err = ag71xx_remove_ar8216_header(ag, skb, pktlen); |
|
if (err) { |
dev->stats.rx_dropped++; |
kfree_skb(skb); |
} else { |
skb->dev = dev; |
skb->ip_summed = CHECKSUM_NONE; |
__skb_queue_tail(&queue, skb); |
} |
|
next: |
ring->buf[i].rx_buf = NULL; |
done++; |
|
ring->curr++; |
} |
|
ag71xx_ring_rx_refill(ag); |
|
while ((skb = __skb_dequeue(&queue)) != NULL) { |
skb->protocol = eth_type_trans(skb, dev); |
netif_receive_skb(skb); |
} |
|
DBG("%s: rx finish, curr=%u, dirty=%u, done=%d\n", |
dev->name, ring->curr, ring->dirty, done); |
|
return done; |
} |
|
static int ag71xx_poll(struct napi_struct *napi, int limit) |
{ |
struct ag71xx *ag = container_of(napi, struct ag71xx, napi); |
struct ag71xx_platform_data *pdata = ag71xx_get_pdata(ag); |
struct net_device *dev = ag->dev; |
struct ag71xx_ring *rx_ring = &ag->rx_ring; |
int rx_ring_size = BIT(rx_ring->order); |
unsigned long flags; |
u32 status; |
int tx_done; |
int rx_done; |
|
pdata->ddr_flush(); |
tx_done = ag71xx_tx_packets(ag, false); |
|
DBG("%s: processing RX ring\n", dev->name); |
rx_done = ag71xx_rx_packets(ag, limit); |
|
ag71xx_debugfs_update_napi_stats(ag, rx_done, tx_done); |
|
if (rx_ring->buf[rx_ring->dirty % rx_ring_size].rx_buf == NULL) |
goto oom; |
|
status = ag71xx_rr(ag, AG71XX_REG_RX_STATUS); |
if (unlikely(status & RX_STATUS_OF)) { |
ag71xx_wr(ag, AG71XX_REG_RX_STATUS, RX_STATUS_OF); |
dev->stats.rx_fifo_errors++; |
|
/* restart RX */ |
ag71xx_wr(ag, AG71XX_REG_RX_CTRL, RX_CTRL_RXE); |
} |
|
if (rx_done < limit) { |
if (status & RX_STATUS_PR) |
goto more; |
|
status = ag71xx_rr(ag, AG71XX_REG_TX_STATUS); |
if (status & TX_STATUS_PS) |
goto more; |
|
DBG("%s: disable polling mode, rx=%d, tx=%d,limit=%d\n", |
dev->name, rx_done, tx_done, limit); |
|
napi_complete(napi); |
|
/* enable interrupts */ |
spin_lock_irqsave(&ag->lock, flags); |
ag71xx_int_enable(ag, AG71XX_INT_POLL); |
spin_unlock_irqrestore(&ag->lock, flags); |
return rx_done; |
} |
|
more: |
DBG("%s: stay in polling mode, rx=%d, tx=%d, limit=%d\n", |
dev->name, rx_done, tx_done, limit); |
return limit; |
|
oom: |
if (netif_msg_rx_err(ag)) |
pr_info("%s: out of memory\n", dev->name); |
|
mod_timer(&ag->oom_timer, jiffies + AG71XX_OOM_REFILL); |
napi_complete(napi); |
return 0; |
} |
|
static irqreturn_t ag71xx_interrupt(int irq, void *dev_id) |
{ |
struct net_device *dev = dev_id; |
struct ag71xx *ag = netdev_priv(dev); |
u32 status; |
|
status = ag71xx_rr(ag, AG71XX_REG_INT_STATUS); |
ag71xx_dump_intr(ag, "raw", status); |
|
if (unlikely(!status)) |
return IRQ_NONE; |
|
if (unlikely(status & AG71XX_INT_ERR)) { |
if (status & AG71XX_INT_TX_BE) { |
ag71xx_wr(ag, AG71XX_REG_TX_STATUS, TX_STATUS_BE); |
dev_err(&dev->dev, "TX BUS error\n"); |
} |
if (status & AG71XX_INT_RX_BE) { |
ag71xx_wr(ag, AG71XX_REG_RX_STATUS, RX_STATUS_BE); |
dev_err(&dev->dev, "RX BUS error\n"); |
} |
} |
|
if (likely(status & AG71XX_INT_POLL)) { |
ag71xx_int_disable(ag, AG71XX_INT_POLL); |
DBG("%s: enable polling mode\n", dev->name); |
napi_schedule(&ag->napi); |
} |
|
ag71xx_debugfs_update_int_stats(ag, status); |
|
return IRQ_HANDLED; |
} |
|
#ifdef CONFIG_NET_POLL_CONTROLLER |
/* |
* Polling 'interrupt' - used by things like netconsole to send skbs |
* without having to re-enable interrupts. It's not called while |
* the interrupt routine is executing. |
*/ |
static void ag71xx_netpoll(struct net_device *dev) |
{ |
disable_irq(dev->irq); |
ag71xx_interrupt(dev->irq, dev); |
enable_irq(dev->irq); |
} |
#endif |
|
static int ag71xx_change_mtu(struct net_device *dev, int new_mtu) |
{ |
struct ag71xx *ag = netdev_priv(dev); |
unsigned int max_frame_len; |
|
max_frame_len = ag71xx_max_frame_len(new_mtu); |
if (new_mtu < 68 || max_frame_len > ag->max_frame_len) |
return -EINVAL; |
|
if (netif_running(dev)) |
return -EBUSY; |
|
dev->mtu = new_mtu; |
return 0; |
} |
|
static const struct net_device_ops ag71xx_netdev_ops = { |
.ndo_open = ag71xx_open, |
.ndo_stop = ag71xx_stop, |
.ndo_start_xmit = ag71xx_hard_start_xmit, |
.ndo_do_ioctl = ag71xx_do_ioctl, |
.ndo_tx_timeout = ag71xx_tx_timeout, |
.ndo_change_mtu = ag71xx_change_mtu, |
.ndo_set_mac_address = eth_mac_addr, |
.ndo_validate_addr = eth_validate_addr, |
#ifdef CONFIG_NET_POLL_CONTROLLER |
.ndo_poll_controller = ag71xx_netpoll, |
#endif |
}; |
|
static const char *ag71xx_get_phy_if_mode_name(phy_interface_t mode) |
{ |
switch (mode) { |
case PHY_INTERFACE_MODE_MII: |
return "MII"; |
case PHY_INTERFACE_MODE_GMII: |
return "GMII"; |
case PHY_INTERFACE_MODE_RMII: |
return "RMII"; |
case PHY_INTERFACE_MODE_RGMII: |
return "RGMII"; |
case PHY_INTERFACE_MODE_SGMII: |
return "SGMII"; |
default: |
break; |
} |
|
return "unknown"; |
} |
|
|
static int ag71xx_probe(struct platform_device *pdev) |
{ |
struct net_device *dev; |
struct resource *res; |
struct ag71xx *ag; |
struct ag71xx_platform_data *pdata; |
int tx_size, err; |
|
pdata = pdev->dev.platform_data; |
if (!pdata) { |
dev_err(&pdev->dev, "no platform data specified\n"); |
err = -ENXIO; |
goto err_out; |
} |
|
if (pdata->mii_bus_dev == NULL && pdata->phy_mask) { |
dev_err(&pdev->dev, "no MII bus device specified\n"); |
err = -EINVAL; |
goto err_out; |
} |
|
dev = alloc_etherdev(sizeof(*ag)); |
if (!dev) { |
dev_err(&pdev->dev, "alloc_etherdev failed\n"); |
err = -ENOMEM; |
goto err_out; |
} |
|
if (!pdata->max_frame_len || !pdata->desc_pktlen_mask) |
return -EINVAL; |
|
SET_NETDEV_DEV(dev, &pdev->dev); |
|
ag = netdev_priv(dev); |
ag->pdev = pdev; |
ag->dev = dev; |
ag->msg_enable = netif_msg_init(ag71xx_msg_level, |
AG71XX_DEFAULT_MSG_ENABLE); |
spin_lock_init(&ag->lock); |
|
res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "mac_base"); |
if (!res) { |
dev_err(&pdev->dev, "no mac_base resource found\n"); |
err = -ENXIO; |
goto err_out; |
} |
|
ag->mac_base = ioremap_nocache(res->start, res->end - res->start + 1); |
if (!ag->mac_base) { |
dev_err(&pdev->dev, "unable to ioremap mac_base\n"); |
err = -ENOMEM; |
goto err_free_dev; |
} |
|
dev->irq = platform_get_irq(pdev, 0); |
err = request_irq(dev->irq, ag71xx_interrupt, |
0x0, |
dev->name, dev); |
if (err) { |
dev_err(&pdev->dev, "unable to request IRQ %d\n", dev->irq); |
goto err_unmap_base; |
} |
|
dev->base_addr = (unsigned long)ag->mac_base; |
dev->netdev_ops = &ag71xx_netdev_ops; |
dev->ethtool_ops = &ag71xx_ethtool_ops; |
|
INIT_DELAYED_WORK(&ag->restart_work, ag71xx_restart_work_func); |
|
init_timer(&ag->oom_timer); |
ag->oom_timer.data = (unsigned long) dev; |
ag->oom_timer.function = ag71xx_oom_timer_handler; |
|
tx_size = AG71XX_TX_RING_SIZE_DEFAULT; |
ag->rx_ring.order = ag71xx_ring_size_order(AG71XX_RX_RING_SIZE_DEFAULT); |
|
ag->max_frame_len = pdata->max_frame_len; |
ag->desc_pktlen_mask = pdata->desc_pktlen_mask; |
|
if (!pdata->is_ar724x && !pdata->is_ar91xx) { |
ag->tx_ring.desc_split = AG71XX_TX_RING_SPLIT; |
tx_size *= AG71XX_TX_RING_DS_PER_PKT; |
} |
ag->tx_ring.order = ag71xx_ring_size_order(tx_size); |
|
ag->stop_desc = dma_alloc_coherent(NULL, |
sizeof(struct ag71xx_desc), &ag->stop_desc_dma, GFP_KERNEL); |
|
if (!ag->stop_desc) |
goto err_free_irq; |
|
ag->stop_desc->data = 0; |
ag->stop_desc->ctrl = 0; |
ag->stop_desc->next = (u32) ag->stop_desc_dma; |
|
memcpy(dev->dev_addr, pdata->mac_addr, ETH_ALEN); |
|
netif_napi_add(dev, &ag->napi, ag71xx_poll, AG71XX_NAPI_WEIGHT); |
|
ag71xx_dump_regs(ag); |
|
ag71xx_hw_init(ag); |
|
ag71xx_dump_regs(ag); |
|
err = ag71xx_phy_connect(ag); |
if (err) |
goto err_free_desc; |
|
err = ag71xx_debugfs_init(ag); |
if (err) |
goto err_phy_disconnect; |
|
platform_set_drvdata(pdev, dev); |
|
err = register_netdev(dev); |
if (err) { |
dev_err(&pdev->dev, "unable to register net device\n"); |
goto err_debugfs_exit; |
} |
|
pr_info("%s: Atheros AG71xx at 0x%08lx, irq %d, mode:%s\n", |
dev->name, dev->base_addr, dev->irq, |
ag71xx_get_phy_if_mode_name(pdata->phy_if_mode)); |
|
return 0; |
|
err_debugfs_exit: |
ag71xx_debugfs_exit(ag); |
err_phy_disconnect: |
ag71xx_phy_disconnect(ag); |
err_free_desc: |
dma_free_coherent(NULL, sizeof(struct ag71xx_desc), ag->stop_desc, |
ag->stop_desc_dma); |
err_free_irq: |
free_irq(dev->irq, dev); |
err_unmap_base: |
iounmap(ag->mac_base); |
err_free_dev: |
kfree(dev); |
err_out: |
platform_set_drvdata(pdev, NULL); |
return err; |
} |
|
static int ag71xx_remove(struct platform_device *pdev) |
{ |
struct net_device *dev = platform_get_drvdata(pdev); |
|
if (dev) { |
struct ag71xx *ag = netdev_priv(dev); |
|
ag71xx_debugfs_exit(ag); |
ag71xx_phy_disconnect(ag); |
unregister_netdev(dev); |
free_irq(dev->irq, dev); |
iounmap(ag->mac_base); |
kfree(dev); |
platform_set_drvdata(pdev, NULL); |
} |
|
return 0; |
} |
|
static struct platform_driver ag71xx_driver = { |
.probe = ag71xx_probe, |
.remove = ag71xx_remove, |
.driver = { |
.name = AG71XX_DRV_NAME, |
} |
}; |
|
static int __init ag71xx_module_init(void) |
{ |
int ret; |
|
ret = ag71xx_debugfs_root_init(); |
if (ret) |
goto err_out; |
|
ret = ag71xx_mdio_driver_init(); |
if (ret) |
goto err_debugfs_exit; |
|
ret = platform_driver_register(&ag71xx_driver); |
if (ret) |
goto err_mdio_exit; |
|
return 0; |
|
err_mdio_exit: |
ag71xx_mdio_driver_exit(); |
err_debugfs_exit: |
ag71xx_debugfs_root_exit(); |
err_out: |
return ret; |
} |
|
static void __exit ag71xx_module_exit(void) |
{ |
platform_driver_unregister(&ag71xx_driver); |
ag71xx_mdio_driver_exit(); |
ag71xx_debugfs_root_exit(); |
} |
|
module_init(ag71xx_module_init); |
module_exit(ag71xx_module_exit); |
|
MODULE_VERSION(AG71XX_DRV_VERSION); |
MODULE_AUTHOR("Gabor Juhos <juhosg@openwrt.org>"); |
MODULE_AUTHOR("Imre Kaloz <kaloz@openwrt.org>"); |
MODULE_LICENSE("GPL v2"); |
MODULE_ALIAS("platform:" AG71XX_DRV_NAME); |
/branches/gl-inet/target/linux/ar71xx/files/drivers/net/ethernet/atheros/ag71xx/ag71xx_mdio.c |
@@ -0,0 +1,320 @@ |
/* |
* Atheros AR71xx built-in ethernet mac driver |
* |
* Copyright (C) 2008-2010 Gabor Juhos <juhosg@openwrt.org> |
* Copyright (C) 2008 Imre Kaloz <kaloz@openwrt.org> |
* |
* Based on Atheros' AG7100 driver |
* |
* This program is free software; you can redistribute it and/or modify it |
* under the terms of the GNU General Public License version 2 as published |
* by the Free Software Foundation. |
*/ |
|
#include "ag71xx.h" |
|
#define AG71XX_MDIO_RETRY 1000 |
#define AG71XX_MDIO_DELAY 5 |
|
static inline void ag71xx_mdio_wr(struct ag71xx_mdio *am, unsigned reg, |
u32 value) |
{ |
void __iomem *r; |
|
r = am->mdio_base + reg; |
__raw_writel(value, r); |
|
/* flush write */ |
(void) __raw_readl(r); |
} |
|
static inline u32 ag71xx_mdio_rr(struct ag71xx_mdio *am, unsigned reg) |
{ |
return __raw_readl(am->mdio_base + reg); |
} |
|
static void ag71xx_mdio_dump_regs(struct ag71xx_mdio *am) |
{ |
DBG("%s: mii_cfg=%08x, mii_cmd=%08x, mii_addr=%08x\n", |
am->mii_bus->name, |
ag71xx_mdio_rr(am, AG71XX_REG_MII_CFG), |
ag71xx_mdio_rr(am, AG71XX_REG_MII_CMD), |
ag71xx_mdio_rr(am, AG71XX_REG_MII_ADDR)); |
DBG("%s: mii_ctrl=%08x, mii_status=%08x, mii_ind=%08x\n", |
am->mii_bus->name, |
ag71xx_mdio_rr(am, AG71XX_REG_MII_CTRL), |
ag71xx_mdio_rr(am, AG71XX_REG_MII_STATUS), |
ag71xx_mdio_rr(am, AG71XX_REG_MII_IND)); |
} |
|
static int ag71xx_mdio_wait_busy(struct ag71xx_mdio *am) |
{ |
int i; |
|
for (i = 0; i < AG71XX_MDIO_RETRY; i++) { |
u32 busy; |
|
udelay(AG71XX_MDIO_DELAY); |
|
busy = ag71xx_mdio_rr(am, AG71XX_REG_MII_IND); |
if (!busy) |
return 0; |
|
udelay(AG71XX_MDIO_DELAY); |
} |
|
pr_err("%s: MDIO operation timed out\n", am->mii_bus->name); |
|
return -ETIMEDOUT; |
} |
|
int ag71xx_mdio_mii_read(struct ag71xx_mdio *am, int addr, int reg) |
{ |
int err; |
int ret; |
|
err = ag71xx_mdio_wait_busy(am); |
if (err) |
return 0xffff; |
|
ag71xx_mdio_wr(am, AG71XX_REG_MII_CMD, MII_CMD_WRITE); |
ag71xx_mdio_wr(am, AG71XX_REG_MII_ADDR, |
((addr & 0xff) << MII_ADDR_SHIFT) | (reg & 0xff)); |
ag71xx_mdio_wr(am, AG71XX_REG_MII_CMD, MII_CMD_READ); |
|
err = ag71xx_mdio_wait_busy(am); |
if (err) |
return 0xffff; |
|
ret = ag71xx_mdio_rr(am, AG71XX_REG_MII_STATUS) & 0xffff; |
ag71xx_mdio_wr(am, AG71XX_REG_MII_CMD, MII_CMD_WRITE); |
|
DBG("mii_read: addr=%04x, reg=%04x, value=%04x\n", addr, reg, ret); |
|
return ret; |
} |
|
void ag71xx_mdio_mii_write(struct ag71xx_mdio *am, int addr, int reg, u16 val) |
{ |
DBG("mii_write: addr=%04x, reg=%04x, value=%04x\n", addr, reg, val); |
|
ag71xx_mdio_wr(am, AG71XX_REG_MII_ADDR, |
((addr & 0xff) << MII_ADDR_SHIFT) | (reg & 0xff)); |
ag71xx_mdio_wr(am, AG71XX_REG_MII_CTRL, val); |
|
ag71xx_mdio_wait_busy(am); |
} |
|
static const u32 ar71xx_mdio_div_table[] = { |
4, 4, 6, 8, 10, 14, 20, 28, |
}; |
|
static const u32 ar7240_mdio_div_table[] = { |
2, 2, 4, 6, 8, 12, 18, 26, 32, 40, 48, 56, 62, 70, 78, 96, |
}; |
|
static const u32 ar933x_mdio_div_table[] = { |
4, 4, 6, 8, 10, 14, 20, 28, 34, 42, 50, 58, 66, 74, 82, 98, |
}; |
|
static int ag71xx_mdio_get_divider(struct ag71xx_mdio *am, u32 *div) |
{ |
unsigned long ref_clock, mdio_clock; |
const u32 *table; |
int ndivs; |
int i; |
|
ref_clock = am->pdata->ref_clock; |
mdio_clock = am->pdata->mdio_clock; |
|
if (!ref_clock || !mdio_clock) |
return -EINVAL; |
|
if (am->pdata->is_ar9330 || am->pdata->is_ar934x) { |
table = ar933x_mdio_div_table; |
ndivs = ARRAY_SIZE(ar933x_mdio_div_table); |
} else if (am->pdata->is_ar7240) { |
table = ar7240_mdio_div_table; |
ndivs = ARRAY_SIZE(ar7240_mdio_div_table); |
} else { |
table = ar71xx_mdio_div_table; |
ndivs = ARRAY_SIZE(ar71xx_mdio_div_table); |
} |
|
for (i = 0; i < ndivs; i++) { |
unsigned long t; |
|
t = ref_clock / table[i]; |
if (t <= mdio_clock) { |
*div = i; |
return 0; |
} |
} |
|
dev_err(&am->mii_bus->dev, "no divider found for %lu/%lu\n", |
ref_clock, mdio_clock); |
return -ENOENT; |
} |
|
static int ag71xx_mdio_reset(struct mii_bus *bus) |
{ |
struct ag71xx_mdio *am = bus->priv; |
u32 t; |
int err; |
|
err = ag71xx_mdio_get_divider(am, &t); |
if (err) { |
/* fallback */ |
if (am->pdata->is_ar7240) |
t = MII_CFG_CLK_DIV_6; |
else if (am->pdata->builtin_switch && !am->pdata->is_ar934x) |
t = MII_CFG_CLK_DIV_10; |
else if (!am->pdata->builtin_switch && am->pdata->is_ar934x) |
t = MII_CFG_CLK_DIV_58; |
else |
t = MII_CFG_CLK_DIV_28; |
} |
|
ag71xx_mdio_wr(am, AG71XX_REG_MII_CFG, t | MII_CFG_RESET); |
udelay(100); |
|
ag71xx_mdio_wr(am, AG71XX_REG_MII_CFG, t); |
udelay(100); |
|
if (am->pdata->reset) |
am->pdata->reset(bus); |
|
return 0; |
} |
|
static int ag71xx_mdio_read(struct mii_bus *bus, int addr, int reg) |
{ |
struct ag71xx_mdio *am = bus->priv; |
|
if (am->pdata->builtin_switch) |
return ar7240sw_phy_read(bus, addr, reg); |
else |
return ag71xx_mdio_mii_read(am, addr, reg); |
} |
|
static int ag71xx_mdio_write(struct mii_bus *bus, int addr, int reg, u16 val) |
{ |
struct ag71xx_mdio *am = bus->priv; |
|
if (am->pdata->builtin_switch) |
ar7240sw_phy_write(bus, addr, reg, val); |
else |
ag71xx_mdio_mii_write(am, addr, reg, val); |
return 0; |
} |
|
static int ag71xx_mdio_probe(struct platform_device *pdev) |
{ |
struct ag71xx_mdio_platform_data *pdata; |
struct ag71xx_mdio *am; |
struct resource *res; |
int i; |
int err; |
|
pdata = pdev->dev.platform_data; |
if (!pdata) { |
dev_err(&pdev->dev, "no platform data specified\n"); |
return -EINVAL; |
} |
|
am = kzalloc(sizeof(*am), GFP_KERNEL); |
if (!am) { |
err = -ENOMEM; |
goto err_out; |
} |
|
am->pdata = pdata; |
|
res = platform_get_resource(pdev, IORESOURCE_MEM, 0); |
if (!res) { |
dev_err(&pdev->dev, "no iomem resource found\n"); |
err = -ENXIO; |
goto err_free_mdio; |
} |
|
am->mdio_base = ioremap_nocache(res->start, res->end - res->start + 1); |
if (!am->mdio_base) { |
dev_err(&pdev->dev, "unable to ioremap registers\n"); |
err = -ENOMEM; |
goto err_free_mdio; |
} |
|
am->mii_bus = mdiobus_alloc(); |
if (am->mii_bus == NULL) { |
err = -ENOMEM; |
goto err_iounmap; |
} |
|
am->mii_bus->name = "ag71xx_mdio"; |
am->mii_bus->read = ag71xx_mdio_read; |
am->mii_bus->write = ag71xx_mdio_write; |
am->mii_bus->reset = ag71xx_mdio_reset; |
#if LINUX_VERSION_CODE < KERNEL_VERSION(4,5,0) |
am->mii_bus->irq = am->mii_irq; |
#endif |
am->mii_bus->priv = am; |
am->mii_bus->parent = &pdev->dev; |
snprintf(am->mii_bus->id, MII_BUS_ID_SIZE, "%s", dev_name(&pdev->dev)); |
am->mii_bus->phy_mask = pdata->phy_mask; |
|
for (i = 0; i < PHY_MAX_ADDR; i++) |
am->mii_bus->irq[i] = PHY_POLL; |
|
ag71xx_mdio_wr(am, AG71XX_REG_MAC_CFG1, 0); |
|
err = mdiobus_register(am->mii_bus); |
if (err) |
goto err_free_bus; |
|
ag71xx_mdio_dump_regs(am); |
|
platform_set_drvdata(pdev, am); |
return 0; |
|
err_free_bus: |
mdiobus_free(am->mii_bus); |
err_iounmap: |
iounmap(am->mdio_base); |
err_free_mdio: |
kfree(am); |
err_out: |
return err; |
} |
|
static int ag71xx_mdio_remove(struct platform_device *pdev) |
{ |
struct ag71xx_mdio *am = platform_get_drvdata(pdev); |
|
if (am) { |
mdiobus_unregister(am->mii_bus); |
mdiobus_free(am->mii_bus); |
iounmap(am->mdio_base); |
kfree(am); |
platform_set_drvdata(pdev, NULL); |
} |
|
return 0; |
} |
|
static struct platform_driver ag71xx_mdio_driver = { |
.probe = ag71xx_mdio_probe, |
.remove = ag71xx_mdio_remove, |
.driver = { |
.name = "ag71xx-mdio", |
} |
}; |
|
int __init ag71xx_mdio_driver_init(void) |
{ |
return platform_driver_register(&ag71xx_mdio_driver); |
} |
|
void ag71xx_mdio_driver_exit(void) |
{ |
platform_driver_unregister(&ag71xx_mdio_driver); |
} |