| /* Synopsys DesignWare Core Enterprise Ethernet (XLGMAC) Driver | 
 |  * | 
 |  * Copyright (c) 2017 Synopsys, Inc. (www.synopsys.com) | 
 |  * | 
 |  * This program is dual-licensed; you may select either version 2 of | 
 |  * the GNU General Public License ("GPL") or BSD license ("BSD"). | 
 |  * | 
 |  * This Synopsys DWC XLGMAC software driver and associated documentation | 
 |  * (hereinafter the "Software") is an unsupported proprietary work of | 
 |  * Synopsys, Inc. unless otherwise expressly agreed to in writing between | 
 |  * Synopsys and you. The Software IS NOT an item of Licensed Software or a | 
 |  * Licensed Product under any End User Software License Agreement or | 
 |  * Agreement for Licensed Products with Synopsys or any supplement thereto. | 
 |  * Synopsys is a registered trademark of Synopsys, Inc. Other names included | 
 |  * in the SOFTWARE may be the trademarks of their respective owners. | 
 |  */ | 
 |  | 
 | #include <linux/phy.h> | 
 | #include <linux/mdio.h> | 
 | #include <linux/clk.h> | 
 | #include <linux/bitrev.h> | 
 | #include <linux/crc32.h> | 
 | #include <linux/crc32poly.h> | 
 | #include <linux/dcbnl.h> | 
 |  | 
 | #include "dwc-xlgmac.h" | 
 | #include "dwc-xlgmac-reg.h" | 
 |  | 
 | static int xlgmac_tx_complete(struct xlgmac_dma_desc *dma_desc) | 
 | { | 
 | 	return !XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				TX_NORMAL_DESC3_OWN_POS, | 
 | 				TX_NORMAL_DESC3_OWN_LEN); | 
 | } | 
 |  | 
 | static int xlgmac_disable_rx_csum(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_RCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_IPC_POS, | 
 | 				     MAC_RCR_IPC_LEN, 0); | 
 | 	writel(regval, pdata->mac_regs + MAC_RCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_enable_rx_csum(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_RCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_IPC_POS, | 
 | 				     MAC_RCR_IPC_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MAC_RCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_set_mac_address(struct xlgmac_pdata *pdata, u8 *addr) | 
 | { | 
 | 	unsigned int mac_addr_hi, mac_addr_lo; | 
 |  | 
 | 	mac_addr_hi = (addr[5] <<  8) | (addr[4] <<  0); | 
 | 	mac_addr_lo = (addr[3] << 24) | (addr[2] << 16) | | 
 | 		      (addr[1] <<  8) | (addr[0] <<  0); | 
 |  | 
 | 	writel(mac_addr_hi, pdata->mac_regs + MAC_MACA0HR); | 
 | 	writel(mac_addr_lo, pdata->mac_regs + MAC_MACA0LR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_set_mac_reg(struct xlgmac_pdata *pdata, | 
 | 			       struct netdev_hw_addr *ha, | 
 | 			       unsigned int *mac_reg) | 
 | { | 
 | 	unsigned int mac_addr_hi, mac_addr_lo; | 
 | 	u8 *mac_addr; | 
 |  | 
 | 	mac_addr_lo = 0; | 
 | 	mac_addr_hi = 0; | 
 |  | 
 | 	if (ha) { | 
 | 		mac_addr = (u8 *)&mac_addr_lo; | 
 | 		mac_addr[0] = ha->addr[0]; | 
 | 		mac_addr[1] = ha->addr[1]; | 
 | 		mac_addr[2] = ha->addr[2]; | 
 | 		mac_addr[3] = ha->addr[3]; | 
 | 		mac_addr = (u8 *)&mac_addr_hi; | 
 | 		mac_addr[0] = ha->addr[4]; | 
 | 		mac_addr[1] = ha->addr[5]; | 
 |  | 
 | 		netif_dbg(pdata, drv, pdata->netdev, | 
 | 			  "adding mac address %pM at %#x\n", | 
 | 			  ha->addr, *mac_reg); | 
 |  | 
 | 		mac_addr_hi = XLGMAC_SET_REG_BITS(mac_addr_hi, | 
 | 						  MAC_MACA1HR_AE_POS, | 
 | 						MAC_MACA1HR_AE_LEN, | 
 | 						1); | 
 | 	} | 
 |  | 
 | 	writel(mac_addr_hi, pdata->mac_regs + *mac_reg); | 
 | 	*mac_reg += MAC_MACA_INC; | 
 | 	writel(mac_addr_lo, pdata->mac_regs + *mac_reg); | 
 | 	*mac_reg += MAC_MACA_INC; | 
 | } | 
 |  | 
 | static int xlgmac_enable_rx_vlan_stripping(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_VLANTR); | 
 | 	/* Put the VLAN tag in the Rx descriptor */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_EVLRXS_POS, | 
 | 				     MAC_VLANTR_EVLRXS_LEN, 1); | 
 | 	/* Don't check the VLAN type */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_DOVLTC_POS, | 
 | 				     MAC_VLANTR_DOVLTC_LEN, 1); | 
 | 	/* Check only C-TAG (0x8100) packets */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_ERSVLM_POS, | 
 | 				     MAC_VLANTR_ERSVLM_LEN, 0); | 
 | 	/* Don't consider an S-TAG (0x88A8) packet as a VLAN packet */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_ESVL_POS, | 
 | 				     MAC_VLANTR_ESVL_LEN, 0); | 
 | 	/* Enable VLAN tag stripping */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_EVLS_POS, | 
 | 				     MAC_VLANTR_EVLS_LEN, 0x3); | 
 | 	writel(regval, pdata->mac_regs + MAC_VLANTR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_disable_rx_vlan_stripping(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_VLANTR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_EVLS_POS, | 
 | 				     MAC_VLANTR_EVLS_LEN, 0); | 
 | 	writel(regval, pdata->mac_regs + MAC_VLANTR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_enable_rx_vlan_filtering(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_PFR); | 
 | 	/* Enable VLAN filtering */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_VTFE_POS, | 
 | 				     MAC_PFR_VTFE_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MAC_PFR); | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_VLANTR); | 
 | 	/* Enable VLAN Hash Table filtering */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_VTHM_POS, | 
 | 				     MAC_VLANTR_VTHM_LEN, 1); | 
 | 	/* Disable VLAN tag inverse matching */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_VTIM_POS, | 
 | 				     MAC_VLANTR_VTIM_LEN, 0); | 
 | 	/* Only filter on the lower 12-bits of the VLAN tag */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_ETV_POS, | 
 | 				     MAC_VLANTR_ETV_LEN, 1); | 
 | 	/* In order for the VLAN Hash Table filtering to be effective, | 
 | 	 * the VLAN tag identifier in the VLAN Tag Register must not | 
 | 	 * be zero.  Set the VLAN tag identifier to "1" to enable the | 
 | 	 * VLAN Hash Table filtering.  This implies that a VLAN tag of | 
 | 	 * 1 will always pass filtering. | 
 | 	 */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANTR_VL_POS, | 
 | 				     MAC_VLANTR_VL_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MAC_VLANTR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_disable_rx_vlan_filtering(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_PFR); | 
 | 	/* Disable VLAN filtering */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_VTFE_POS, | 
 | 				     MAC_PFR_VTFE_LEN, 0); | 
 | 	writel(regval, pdata->mac_regs + MAC_PFR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static u32 xlgmac_vid_crc32_le(__le16 vid_le) | 
 | { | 
 | 	unsigned char *data = (unsigned char *)&vid_le; | 
 | 	unsigned char data_byte = 0; | 
 | 	u32 crc = ~0; | 
 | 	u32 temp = 0; | 
 | 	int i, bits; | 
 |  | 
 | 	bits = get_bitmask_order(VLAN_VID_MASK); | 
 | 	for (i = 0; i < bits; i++) { | 
 | 		if ((i % 8) == 0) | 
 | 			data_byte = data[i / 8]; | 
 |  | 
 | 		temp = ((crc & 1) ^ data_byte) & 1; | 
 | 		crc >>= 1; | 
 | 		data_byte >>= 1; | 
 |  | 
 | 		if (temp) | 
 | 			crc ^= CRC32_POLY_LE; | 
 | 	} | 
 |  | 
 | 	return crc; | 
 | } | 
 |  | 
 | static int xlgmac_update_vlan_hash_table(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u16 vlan_hash_table = 0; | 
 | 	__le16 vid_le; | 
 | 	u32 regval; | 
 | 	u32 crc; | 
 | 	u16 vid; | 
 |  | 
 | 	/* Generate the VLAN Hash Table value */ | 
 | 	for_each_set_bit(vid, pdata->active_vlans, VLAN_N_VID) { | 
 | 		/* Get the CRC32 value of the VLAN ID */ | 
 | 		vid_le = cpu_to_le16(vid); | 
 | 		crc = bitrev32(~xlgmac_vid_crc32_le(vid_le)) >> 28; | 
 |  | 
 | 		vlan_hash_table |= (1 << crc); | 
 | 	} | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_VLANHTR); | 
 | 	/* Set the VLAN Hash Table filtering register */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANHTR_VLHT_POS, | 
 | 				     MAC_VLANHTR_VLHT_LEN, vlan_hash_table); | 
 | 	writel(regval, pdata->mac_regs + MAC_VLANHTR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_set_promiscuous_mode(struct xlgmac_pdata *pdata, | 
 | 				       unsigned int enable) | 
 | { | 
 | 	unsigned int val = enable ? 1 : 0; | 
 | 	u32 regval; | 
 |  | 
 | 	regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_PFR), | 
 | 				     MAC_PFR_PR_POS, MAC_PFR_PR_LEN); | 
 | 	if (regval == val) | 
 | 		return 0; | 
 |  | 
 | 	netif_dbg(pdata, drv, pdata->netdev, "%s promiscuous mode\n", | 
 | 		  enable ? "entering" : "leaving"); | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_PFR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_PR_POS, | 
 | 				     MAC_PFR_PR_LEN, val); | 
 | 	writel(regval, pdata->mac_regs + MAC_PFR); | 
 |  | 
 | 	/* Hardware will still perform VLAN filtering in promiscuous mode */ | 
 | 	if (enable) { | 
 | 		xlgmac_disable_rx_vlan_filtering(pdata); | 
 | 	} else { | 
 | 		if (pdata->netdev->features & NETIF_F_HW_VLAN_CTAG_FILTER) | 
 | 			xlgmac_enable_rx_vlan_filtering(pdata); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_set_all_multicast_mode(struct xlgmac_pdata *pdata, | 
 | 					 unsigned int enable) | 
 | { | 
 | 	unsigned int val = enable ? 1 : 0; | 
 | 	u32 regval; | 
 |  | 
 | 	regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_PFR), | 
 | 				     MAC_PFR_PM_POS, MAC_PFR_PM_LEN); | 
 | 	if (regval == val) | 
 | 		return 0; | 
 |  | 
 | 	netif_dbg(pdata, drv, pdata->netdev, "%s allmulti mode\n", | 
 | 		  enable ? "entering" : "leaving"); | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_PFR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_PM_POS, | 
 | 				     MAC_PFR_PM_LEN, val); | 
 | 	writel(regval, pdata->mac_regs + MAC_PFR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_set_mac_addn_addrs(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct net_device *netdev = pdata->netdev; | 
 | 	struct netdev_hw_addr *ha; | 
 | 	unsigned int addn_macs; | 
 | 	unsigned int mac_reg; | 
 |  | 
 | 	mac_reg = MAC_MACA1HR; | 
 | 	addn_macs = pdata->hw_feat.addn_mac; | 
 |  | 
 | 	if (netdev_uc_count(netdev) > addn_macs) { | 
 | 		xlgmac_set_promiscuous_mode(pdata, 1); | 
 | 	} else { | 
 | 		netdev_for_each_uc_addr(ha, netdev) { | 
 | 			xlgmac_set_mac_reg(pdata, ha, &mac_reg); | 
 | 			addn_macs--; | 
 | 		} | 
 |  | 
 | 		if (netdev_mc_count(netdev) > addn_macs) { | 
 | 			xlgmac_set_all_multicast_mode(pdata, 1); | 
 | 		} else { | 
 | 			netdev_for_each_mc_addr(ha, netdev) { | 
 | 				xlgmac_set_mac_reg(pdata, ha, &mac_reg); | 
 | 				addn_macs--; | 
 | 			} | 
 | 		} | 
 | 	} | 
 |  | 
 | 	/* Clear remaining additional MAC address entries */ | 
 | 	while (addn_macs--) | 
 | 		xlgmac_set_mac_reg(pdata, NULL, &mac_reg); | 
 | } | 
 |  | 
 | static void xlgmac_set_mac_hash_table(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int hash_table_shift, hash_table_count; | 
 | 	u32 hash_table[XLGMAC_MAC_HASH_TABLE_SIZE]; | 
 | 	struct net_device *netdev = pdata->netdev; | 
 | 	struct netdev_hw_addr *ha; | 
 | 	unsigned int hash_reg; | 
 | 	unsigned int i; | 
 | 	u32 crc; | 
 |  | 
 | 	hash_table_shift = 26 - (pdata->hw_feat.hash_table_size >> 7); | 
 | 	hash_table_count = pdata->hw_feat.hash_table_size / 32; | 
 | 	memset(hash_table, 0, sizeof(hash_table)); | 
 |  | 
 | 	/* Build the MAC Hash Table register values */ | 
 | 	netdev_for_each_uc_addr(ha, netdev) { | 
 | 		crc = bitrev32(~crc32_le(~0, ha->addr, ETH_ALEN)); | 
 | 		crc >>= hash_table_shift; | 
 | 		hash_table[crc >> 5] |= (1 << (crc & 0x1f)); | 
 | 	} | 
 |  | 
 | 	netdev_for_each_mc_addr(ha, netdev) { | 
 | 		crc = bitrev32(~crc32_le(~0, ha->addr, ETH_ALEN)); | 
 | 		crc >>= hash_table_shift; | 
 | 		hash_table[crc >> 5] |= (1 << (crc & 0x1f)); | 
 | 	} | 
 |  | 
 | 	/* Set the MAC Hash Table registers */ | 
 | 	hash_reg = MAC_HTR0; | 
 | 	for (i = 0; i < hash_table_count; i++) { | 
 | 		writel(hash_table[i], pdata->mac_regs + hash_reg); | 
 | 		hash_reg += MAC_HTR_INC; | 
 | 	} | 
 | } | 
 |  | 
 | static int xlgmac_add_mac_addresses(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	if (pdata->hw_feat.hash_table_size) | 
 | 		xlgmac_set_mac_hash_table(pdata); | 
 | 	else | 
 | 		xlgmac_set_mac_addn_addrs(pdata); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_config_mac_address(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	xlgmac_set_mac_address(pdata, pdata->netdev->dev_addr); | 
 |  | 
 | 	/* Filtering is done using perfect filtering and hash filtering */ | 
 | 	if (pdata->hw_feat.hash_table_size) { | 
 | 		regval = readl(pdata->mac_regs + MAC_PFR); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_HPF_POS, | 
 | 					     MAC_PFR_HPF_LEN, 1); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_HUC_POS, | 
 | 					     MAC_PFR_HUC_LEN, 1); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MAC_PFR_HMC_POS, | 
 | 					     MAC_PFR_HMC_LEN, 1); | 
 | 		writel(regval, pdata->mac_regs + MAC_PFR); | 
 | 	} | 
 | } | 
 |  | 
 | static void xlgmac_config_jumbo_enable(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int val; | 
 | 	u32 regval; | 
 |  | 
 | 	val = (pdata->netdev->mtu > XLGMAC_STD_PACKET_MTU) ? 1 : 0; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_RCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_JE_POS, | 
 | 				     MAC_RCR_JE_LEN, val); | 
 | 	writel(regval, pdata->mac_regs + MAC_RCR); | 
 | } | 
 |  | 
 | static void xlgmac_config_checksum_offload(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	if (pdata->netdev->features & NETIF_F_RXCSUM) | 
 | 		xlgmac_enable_rx_csum(pdata); | 
 | 	else | 
 | 		xlgmac_disable_rx_csum(pdata); | 
 | } | 
 |  | 
 | static void xlgmac_config_vlan_support(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_VLANIR); | 
 | 	/* Indicate that VLAN Tx CTAGs come from context descriptors */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANIR_CSVL_POS, | 
 | 				     MAC_VLANIR_CSVL_LEN, 0); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_VLANIR_VLTI_POS, | 
 | 				     MAC_VLANIR_VLTI_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MAC_VLANIR); | 
 |  | 
 | 	/* Set the current VLAN Hash Table register value */ | 
 | 	xlgmac_update_vlan_hash_table(pdata); | 
 |  | 
 | 	if (pdata->netdev->features & NETIF_F_HW_VLAN_CTAG_FILTER) | 
 | 		xlgmac_enable_rx_vlan_filtering(pdata); | 
 | 	else | 
 | 		xlgmac_disable_rx_vlan_filtering(pdata); | 
 |  | 
 | 	if (pdata->netdev->features & NETIF_F_HW_VLAN_CTAG_RX) | 
 | 		xlgmac_enable_rx_vlan_stripping(pdata); | 
 | 	else | 
 | 		xlgmac_disable_rx_vlan_stripping(pdata); | 
 | } | 
 |  | 
 | static int xlgmac_config_rx_mode(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct net_device *netdev = pdata->netdev; | 
 | 	unsigned int pr_mode, am_mode; | 
 |  | 
 | 	pr_mode = ((netdev->flags & IFF_PROMISC) != 0); | 
 | 	am_mode = ((netdev->flags & IFF_ALLMULTI) != 0); | 
 |  | 
 | 	xlgmac_set_promiscuous_mode(pdata, pr_mode); | 
 | 	xlgmac_set_all_multicast_mode(pdata, am_mode); | 
 |  | 
 | 	xlgmac_add_mac_addresses(pdata); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_prepare_tx_stop(struct xlgmac_pdata *pdata, | 
 | 				   struct xlgmac_channel *channel) | 
 | { | 
 | 	unsigned int tx_dsr, tx_pos, tx_qidx; | 
 | 	unsigned long tx_timeout; | 
 | 	unsigned int tx_status; | 
 |  | 
 | 	/* Calculate the status register to read and the position within */ | 
 | 	if (channel->queue_index < DMA_DSRX_FIRST_QUEUE) { | 
 | 		tx_dsr = DMA_DSR0; | 
 | 		tx_pos = (channel->queue_index * DMA_DSR_Q_LEN) + | 
 | 			 DMA_DSR0_TPS_START; | 
 | 	} else { | 
 | 		tx_qidx = channel->queue_index - DMA_DSRX_FIRST_QUEUE; | 
 |  | 
 | 		tx_dsr = DMA_DSR1 + ((tx_qidx / DMA_DSRX_QPR) * DMA_DSRX_INC); | 
 | 		tx_pos = ((tx_qidx % DMA_DSRX_QPR) * DMA_DSR_Q_LEN) + | 
 | 			 DMA_DSRX_TPS_START; | 
 | 	} | 
 |  | 
 | 	/* The Tx engine cannot be stopped if it is actively processing | 
 | 	 * descriptors. Wait for the Tx engine to enter the stopped or | 
 | 	 * suspended state.  Don't wait forever though... | 
 | 	 */ | 
 | 	tx_timeout = jiffies + (XLGMAC_DMA_STOP_TIMEOUT * HZ); | 
 | 	while (time_before(jiffies, tx_timeout)) { | 
 | 		tx_status = readl(pdata->mac_regs + tx_dsr); | 
 | 		tx_status = XLGMAC_GET_REG_BITS(tx_status, tx_pos, | 
 | 						DMA_DSR_TPS_LEN); | 
 | 		if ((tx_status == DMA_TPS_STOPPED) || | 
 | 		    (tx_status == DMA_TPS_SUSPENDED)) | 
 | 			break; | 
 |  | 
 | 		usleep_range(500, 1000); | 
 | 	} | 
 |  | 
 | 	if (!time_before(jiffies, tx_timeout)) | 
 | 		netdev_info(pdata->netdev, | 
 | 			    "timed out waiting for Tx DMA channel %u to stop\n", | 
 | 			    channel->queue_index); | 
 | } | 
 |  | 
 | static void xlgmac_enable_tx(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	/* Enable each Tx DMA channel */ | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->tx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_ST_POS, | 
 | 					     DMA_CH_TCR_ST_LEN, 1); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 	} | 
 |  | 
 | 	/* Enable each Tx queue */ | 
 | 	for (i = 0; i < pdata->tx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TXQEN_POS, | 
 | 					     MTL_Q_TQOMR_TXQEN_LEN, | 
 | 					MTL_Q_ENABLED); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 	} | 
 |  | 
 | 	/* Enable MAC Tx */ | 
 | 	regval = readl(pdata->mac_regs + MAC_TCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_TE_POS, | 
 | 				     MAC_TCR_TE_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MAC_TCR); | 
 | } | 
 |  | 
 | static void xlgmac_disable_tx(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	/* Prepare for Tx DMA channel stop */ | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->tx_ring) | 
 | 			break; | 
 |  | 
 | 		xlgmac_prepare_tx_stop(pdata, channel); | 
 | 	} | 
 |  | 
 | 	/* Disable MAC Tx */ | 
 | 	regval = readl(pdata->mac_regs + MAC_TCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_TE_POS, | 
 | 				     MAC_TCR_TE_LEN, 0); | 
 | 	writel(regval, pdata->mac_regs + MAC_TCR); | 
 |  | 
 | 	/* Disable each Tx queue */ | 
 | 	for (i = 0; i < pdata->tx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TXQEN_POS, | 
 | 					     MTL_Q_TQOMR_TXQEN_LEN, 0); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 	} | 
 |  | 
 | 	/* Disable each Tx DMA channel */ | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->tx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_ST_POS, | 
 | 					     DMA_CH_TCR_ST_LEN, 0); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 	} | 
 | } | 
 |  | 
 | static void xlgmac_prepare_rx_stop(struct xlgmac_pdata *pdata, | 
 | 				   unsigned int queue) | 
 | { | 
 | 	unsigned int rx_status, prxq, rxqsts; | 
 | 	unsigned long rx_timeout; | 
 |  | 
 | 	/* The Rx engine cannot be stopped if it is actively processing | 
 | 	 * packets. Wait for the Rx queue to empty the Rx fifo.  Don't | 
 | 	 * wait forever though... | 
 | 	 */ | 
 | 	rx_timeout = jiffies + (XLGMAC_DMA_STOP_TIMEOUT * HZ); | 
 | 	while (time_before(jiffies, rx_timeout)) { | 
 | 		rx_status = readl(XLGMAC_MTL_REG(pdata, queue, MTL_Q_RQDR)); | 
 | 		prxq = XLGMAC_GET_REG_BITS(rx_status, MTL_Q_RQDR_PRXQ_POS, | 
 | 					   MTL_Q_RQDR_PRXQ_LEN); | 
 | 		rxqsts = XLGMAC_GET_REG_BITS(rx_status, MTL_Q_RQDR_RXQSTS_POS, | 
 | 					     MTL_Q_RQDR_RXQSTS_LEN); | 
 | 		if ((prxq == 0) && (rxqsts == 0)) | 
 | 			break; | 
 |  | 
 | 		usleep_range(500, 1000); | 
 | 	} | 
 |  | 
 | 	if (!time_before(jiffies, rx_timeout)) | 
 | 		netdev_info(pdata->netdev, | 
 | 			    "timed out waiting for Rx queue %u to empty\n", | 
 | 			    queue); | 
 | } | 
 |  | 
 | static void xlgmac_enable_rx(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int regval, i; | 
 |  | 
 | 	/* Enable each Rx DMA channel */ | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->rx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RCR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RCR_SR_POS, | 
 | 					     DMA_CH_RCR_SR_LEN, 1); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RCR)); | 
 | 	} | 
 |  | 
 | 	/* Enable each Rx queue */ | 
 | 	regval = 0; | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) | 
 | 		regval |= (0x02 << (i << 1)); | 
 | 	writel(regval, pdata->mac_regs + MAC_RQC0R); | 
 |  | 
 | 	/* Enable MAC Rx */ | 
 | 	regval = readl(pdata->mac_regs + MAC_RCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_DCRCC_POS, | 
 | 				     MAC_RCR_DCRCC_LEN, 1); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_CST_POS, | 
 | 				     MAC_RCR_CST_LEN, 1); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_ACS_POS, | 
 | 				     MAC_RCR_ACS_LEN, 1); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_RE_POS, | 
 | 				     MAC_RCR_RE_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MAC_RCR); | 
 | } | 
 |  | 
 | static void xlgmac_disable_rx(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	/* Disable MAC Rx */ | 
 | 	regval = readl(pdata->mac_regs + MAC_RCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_DCRCC_POS, | 
 | 				     MAC_RCR_DCRCC_LEN, 0); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_CST_POS, | 
 | 				     MAC_RCR_CST_LEN, 0); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_ACS_POS, | 
 | 				     MAC_RCR_ACS_LEN, 0); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_RE_POS, | 
 | 				     MAC_RCR_RE_LEN, 0); | 
 | 	writel(regval, pdata->mac_regs + MAC_RCR); | 
 |  | 
 | 	/* Prepare for Rx DMA channel stop */ | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) | 
 | 		xlgmac_prepare_rx_stop(pdata, i); | 
 |  | 
 | 	/* Disable each Rx queue */ | 
 | 	writel(0, pdata->mac_regs + MAC_RQC0R); | 
 |  | 
 | 	/* Disable each Rx DMA channel */ | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->rx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RCR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RCR_SR_POS, | 
 | 					     DMA_CH_RCR_SR_LEN, 0); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RCR)); | 
 | 	} | 
 | } | 
 |  | 
 | static void xlgmac_tx_start_xmit(struct xlgmac_channel *channel, | 
 | 				 struct xlgmac_ring *ring) | 
 | { | 
 | 	struct xlgmac_pdata *pdata = channel->pdata; | 
 | 	struct xlgmac_desc_data *desc_data; | 
 |  | 
 | 	/* Make sure everything is written before the register write */ | 
 | 	wmb(); | 
 |  | 
 | 	/* Issue a poll command to Tx DMA by writing address | 
 | 	 * of next immediate free descriptor | 
 | 	 */ | 
 | 	desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); | 
 | 	writel(lower_32_bits(desc_data->dma_desc_addr), | 
 | 	       XLGMAC_DMA_REG(channel, DMA_CH_TDTR_LO)); | 
 |  | 
 | 	/* Start the Tx timer */ | 
 | 	if (pdata->tx_usecs && !channel->tx_timer_active) { | 
 | 		channel->tx_timer_active = 1; | 
 | 		mod_timer(&channel->tx_timer, | 
 | 			  jiffies + usecs_to_jiffies(pdata->tx_usecs)); | 
 | 	} | 
 |  | 
 | 	ring->tx.xmit_more = 0; | 
 | } | 
 |  | 
 | static void xlgmac_dev_xmit(struct xlgmac_channel *channel) | 
 | { | 
 | 	struct xlgmac_pdata *pdata = channel->pdata; | 
 | 	struct xlgmac_ring *ring = channel->tx_ring; | 
 | 	unsigned int tso_context, vlan_context; | 
 | 	struct xlgmac_desc_data *desc_data; | 
 | 	struct xlgmac_dma_desc *dma_desc; | 
 | 	struct xlgmac_pkt_info *pkt_info; | 
 | 	unsigned int csum, tso, vlan; | 
 | 	int start_index = ring->cur; | 
 | 	int cur_index = ring->cur; | 
 | 	unsigned int tx_set_ic; | 
 | 	int i; | 
 |  | 
 | 	pkt_info = &ring->pkt_info; | 
 | 	csum = XLGMAC_GET_REG_BITS(pkt_info->attributes, | 
 | 				   TX_PACKET_ATTRIBUTES_CSUM_ENABLE_POS, | 
 | 				TX_PACKET_ATTRIBUTES_CSUM_ENABLE_LEN); | 
 | 	tso = XLGMAC_GET_REG_BITS(pkt_info->attributes, | 
 | 				  TX_PACKET_ATTRIBUTES_TSO_ENABLE_POS, | 
 | 				TX_PACKET_ATTRIBUTES_TSO_ENABLE_LEN); | 
 | 	vlan = XLGMAC_GET_REG_BITS(pkt_info->attributes, | 
 | 				   TX_PACKET_ATTRIBUTES_VLAN_CTAG_POS, | 
 | 				TX_PACKET_ATTRIBUTES_VLAN_CTAG_LEN); | 
 |  | 
 | 	if (tso && (pkt_info->mss != ring->tx.cur_mss)) | 
 | 		tso_context = 1; | 
 | 	else | 
 | 		tso_context = 0; | 
 |  | 
 | 	if (vlan && (pkt_info->vlan_ctag != ring->tx.cur_vlan_ctag)) | 
 | 		vlan_context = 1; | 
 | 	else | 
 | 		vlan_context = 0; | 
 |  | 
 | 	/* Determine if an interrupt should be generated for this Tx: | 
 | 	 *   Interrupt: | 
 | 	 *     - Tx frame count exceeds the frame count setting | 
 | 	 *     - Addition of Tx frame count to the frame count since the | 
 | 	 *       last interrupt was set exceeds the frame count setting | 
 | 	 *   No interrupt: | 
 | 	 *     - No frame count setting specified (ethtool -C ethX tx-frames 0) | 
 | 	 *     - Addition of Tx frame count to the frame count since the | 
 | 	 *       last interrupt was set does not exceed the frame count setting | 
 | 	 */ | 
 | 	ring->coalesce_count += pkt_info->tx_packets; | 
 | 	if (!pdata->tx_frames) | 
 | 		tx_set_ic = 0; | 
 | 	else if (pkt_info->tx_packets > pdata->tx_frames) | 
 | 		tx_set_ic = 1; | 
 | 	else if ((ring->coalesce_count % pdata->tx_frames) < | 
 | 		 pkt_info->tx_packets) | 
 | 		tx_set_ic = 1; | 
 | 	else | 
 | 		tx_set_ic = 0; | 
 |  | 
 | 	desc_data = XLGMAC_GET_DESC_DATA(ring, cur_index); | 
 | 	dma_desc = desc_data->dma_desc; | 
 |  | 
 | 	/* Create a context descriptor if this is a TSO pkt_info */ | 
 | 	if (tso_context || vlan_context) { | 
 | 		if (tso_context) { | 
 | 			netif_dbg(pdata, tx_queued, pdata->netdev, | 
 | 				  "TSO context descriptor, mss=%u\n", | 
 | 				  pkt_info->mss); | 
 |  | 
 | 			/* Set the MSS size */ | 
 | 			dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( | 
 | 						dma_desc->desc2, | 
 | 						TX_CONTEXT_DESC2_MSS_POS, | 
 | 						TX_CONTEXT_DESC2_MSS_LEN, | 
 | 						pkt_info->mss); | 
 |  | 
 | 			/* Mark it as a CONTEXT descriptor */ | 
 | 			dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 						dma_desc->desc3, | 
 | 						TX_CONTEXT_DESC3_CTXT_POS, | 
 | 						TX_CONTEXT_DESC3_CTXT_LEN, | 
 | 						1); | 
 |  | 
 | 			/* Indicate this descriptor contains the MSS */ | 
 | 			dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 						dma_desc->desc3, | 
 | 						TX_CONTEXT_DESC3_TCMSSV_POS, | 
 | 						TX_CONTEXT_DESC3_TCMSSV_LEN, | 
 | 						1); | 
 |  | 
 | 			ring->tx.cur_mss = pkt_info->mss; | 
 | 		} | 
 |  | 
 | 		if (vlan_context) { | 
 | 			netif_dbg(pdata, tx_queued, pdata->netdev, | 
 | 				  "VLAN context descriptor, ctag=%u\n", | 
 | 				  pkt_info->vlan_ctag); | 
 |  | 
 | 			/* Mark it as a CONTEXT descriptor */ | 
 | 			dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 						dma_desc->desc3, | 
 | 						TX_CONTEXT_DESC3_CTXT_POS, | 
 | 						TX_CONTEXT_DESC3_CTXT_LEN, | 
 | 						1); | 
 |  | 
 | 			/* Set the VLAN tag */ | 
 | 			dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 						dma_desc->desc3, | 
 | 						TX_CONTEXT_DESC3_VT_POS, | 
 | 						TX_CONTEXT_DESC3_VT_LEN, | 
 | 						pkt_info->vlan_ctag); | 
 |  | 
 | 			/* Indicate this descriptor contains the VLAN tag */ | 
 | 			dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 						dma_desc->desc3, | 
 | 						TX_CONTEXT_DESC3_VLTV_POS, | 
 | 						TX_CONTEXT_DESC3_VLTV_LEN, | 
 | 						1); | 
 |  | 
 | 			ring->tx.cur_vlan_ctag = pkt_info->vlan_ctag; | 
 | 		} | 
 |  | 
 | 		cur_index++; | 
 | 		desc_data = XLGMAC_GET_DESC_DATA(ring, cur_index); | 
 | 		dma_desc = desc_data->dma_desc; | 
 | 	} | 
 |  | 
 | 	/* Update buffer address (for TSO this is the header) */ | 
 | 	dma_desc->desc0 =  cpu_to_le32(lower_32_bits(desc_data->skb_dma)); | 
 | 	dma_desc->desc1 =  cpu_to_le32(upper_32_bits(desc_data->skb_dma)); | 
 |  | 
 | 	/* Update the buffer length */ | 
 | 	dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( | 
 | 				dma_desc->desc2, | 
 | 				TX_NORMAL_DESC2_HL_B1L_POS, | 
 | 				TX_NORMAL_DESC2_HL_B1L_LEN, | 
 | 				desc_data->skb_dma_len); | 
 |  | 
 | 	/* VLAN tag insertion check */ | 
 | 	if (vlan) { | 
 | 		dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc2, | 
 | 					TX_NORMAL_DESC2_VTIR_POS, | 
 | 					TX_NORMAL_DESC2_VTIR_LEN, | 
 | 					TX_NORMAL_DESC2_VLAN_INSERT); | 
 | 		pdata->stats.tx_vlan_packets++; | 
 | 	} | 
 |  | 
 | 	/* Timestamp enablement check */ | 
 | 	if (XLGMAC_GET_REG_BITS(pkt_info->attributes, | 
 | 				TX_PACKET_ATTRIBUTES_PTP_POS, | 
 | 				TX_PACKET_ATTRIBUTES_PTP_LEN)) | 
 | 		dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc2, | 
 | 					TX_NORMAL_DESC2_TTSE_POS, | 
 | 					TX_NORMAL_DESC2_TTSE_LEN, | 
 | 					1); | 
 |  | 
 | 	/* Mark it as First Descriptor */ | 
 | 	dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 				dma_desc->desc3, | 
 | 				TX_NORMAL_DESC3_FD_POS, | 
 | 				TX_NORMAL_DESC3_FD_LEN, | 
 | 				1); | 
 |  | 
 | 	/* Mark it as a NORMAL descriptor */ | 
 | 	dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 				dma_desc->desc3, | 
 | 				TX_NORMAL_DESC3_CTXT_POS, | 
 | 				TX_NORMAL_DESC3_CTXT_LEN, | 
 | 				0); | 
 |  | 
 | 	/* Set OWN bit if not the first descriptor */ | 
 | 	if (cur_index != start_index) | 
 | 		dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc3, | 
 | 					TX_NORMAL_DESC3_OWN_POS, | 
 | 					TX_NORMAL_DESC3_OWN_LEN, | 
 | 					1); | 
 |  | 
 | 	if (tso) { | 
 | 		/* Enable TSO */ | 
 | 		dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc3, | 
 | 					TX_NORMAL_DESC3_TSE_POS, | 
 | 					TX_NORMAL_DESC3_TSE_LEN, 1); | 
 | 		dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc3, | 
 | 					TX_NORMAL_DESC3_TCPPL_POS, | 
 | 					TX_NORMAL_DESC3_TCPPL_LEN, | 
 | 					pkt_info->tcp_payload_len); | 
 | 		dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc3, | 
 | 					TX_NORMAL_DESC3_TCPHDRLEN_POS, | 
 | 					TX_NORMAL_DESC3_TCPHDRLEN_LEN, | 
 | 					pkt_info->tcp_header_len / 4); | 
 |  | 
 | 		pdata->stats.tx_tso_packets++; | 
 | 	} else { | 
 | 		/* Enable CRC and Pad Insertion */ | 
 | 		dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc3, | 
 | 					TX_NORMAL_DESC3_CPC_POS, | 
 | 					TX_NORMAL_DESC3_CPC_LEN, 0); | 
 |  | 
 | 		/* Enable HW CSUM */ | 
 | 		if (csum) | 
 | 			dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 						dma_desc->desc3, | 
 | 						TX_NORMAL_DESC3_CIC_POS, | 
 | 						TX_NORMAL_DESC3_CIC_LEN, | 
 | 						0x3); | 
 |  | 
 | 		/* Set the total length to be transmitted */ | 
 | 		dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc3, | 
 | 					TX_NORMAL_DESC3_FL_POS, | 
 | 					TX_NORMAL_DESC3_FL_LEN, | 
 | 					pkt_info->length); | 
 | 	} | 
 |  | 
 | 	for (i = cur_index - start_index + 1; i < pkt_info->desc_count; i++) { | 
 | 		cur_index++; | 
 | 		desc_data = XLGMAC_GET_DESC_DATA(ring, cur_index); | 
 | 		dma_desc = desc_data->dma_desc; | 
 |  | 
 | 		/* Update buffer address */ | 
 | 		dma_desc->desc0 = | 
 | 			cpu_to_le32(lower_32_bits(desc_data->skb_dma)); | 
 | 		dma_desc->desc1 = | 
 | 			cpu_to_le32(upper_32_bits(desc_data->skb_dma)); | 
 |  | 
 | 		/* Update the buffer length */ | 
 | 		dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc2, | 
 | 					TX_NORMAL_DESC2_HL_B1L_POS, | 
 | 					TX_NORMAL_DESC2_HL_B1L_LEN, | 
 | 					desc_data->skb_dma_len); | 
 |  | 
 | 		/* Set OWN bit */ | 
 | 		dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc3, | 
 | 					TX_NORMAL_DESC3_OWN_POS, | 
 | 					TX_NORMAL_DESC3_OWN_LEN, 1); | 
 |  | 
 | 		/* Mark it as NORMAL descriptor */ | 
 | 		dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc3, | 
 | 					TX_NORMAL_DESC3_CTXT_POS, | 
 | 					TX_NORMAL_DESC3_CTXT_LEN, 0); | 
 |  | 
 | 		/* Enable HW CSUM */ | 
 | 		if (csum) | 
 | 			dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 						dma_desc->desc3, | 
 | 						TX_NORMAL_DESC3_CIC_POS, | 
 | 						TX_NORMAL_DESC3_CIC_LEN, | 
 | 						0x3); | 
 | 	} | 
 |  | 
 | 	/* Set LAST bit for the last descriptor */ | 
 | 	dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 				dma_desc->desc3, | 
 | 				TX_NORMAL_DESC3_LD_POS, | 
 | 				TX_NORMAL_DESC3_LD_LEN, 1); | 
 |  | 
 | 	/* Set IC bit based on Tx coalescing settings */ | 
 | 	if (tx_set_ic) | 
 | 		dma_desc->desc2 = XLGMAC_SET_REG_BITS_LE( | 
 | 					dma_desc->desc2, | 
 | 					TX_NORMAL_DESC2_IC_POS, | 
 | 					TX_NORMAL_DESC2_IC_LEN, 1); | 
 |  | 
 | 	/* Save the Tx info to report back during cleanup */ | 
 | 	desc_data->tx.packets = pkt_info->tx_packets; | 
 | 	desc_data->tx.bytes = pkt_info->tx_bytes; | 
 |  | 
 | 	/* In case the Tx DMA engine is running, make sure everything | 
 | 	 * is written to the descriptor(s) before setting the OWN bit | 
 | 	 * for the first descriptor | 
 | 	 */ | 
 | 	dma_wmb(); | 
 |  | 
 | 	/* Set OWN bit for the first descriptor */ | 
 | 	desc_data = XLGMAC_GET_DESC_DATA(ring, start_index); | 
 | 	dma_desc = desc_data->dma_desc; | 
 | 	dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 				dma_desc->desc3, | 
 | 				TX_NORMAL_DESC3_OWN_POS, | 
 | 				TX_NORMAL_DESC3_OWN_LEN, 1); | 
 |  | 
 | 	if (netif_msg_tx_queued(pdata)) | 
 | 		xlgmac_dump_tx_desc(pdata, ring, start_index, | 
 | 				    pkt_info->desc_count, 1); | 
 |  | 
 | 	/* Make sure ownership is written to the descriptor */ | 
 | 	smp_wmb(); | 
 |  | 
 | 	ring->cur = cur_index + 1; | 
 | 	if (!netdev_xmit_more() || | 
 | 	    netif_xmit_stopped(netdev_get_tx_queue(pdata->netdev, | 
 | 						   channel->queue_index))) | 
 | 		xlgmac_tx_start_xmit(channel, ring); | 
 | 	else | 
 | 		ring->tx.xmit_more = 1; | 
 |  | 
 | 	XLGMAC_PR("%s: descriptors %u to %u written\n", | 
 | 		  channel->name, start_index & (ring->dma_desc_count - 1), | 
 | 		  (ring->cur - 1) & (ring->dma_desc_count - 1)); | 
 | } | 
 |  | 
 | static void xlgmac_get_rx_tstamp(struct xlgmac_pkt_info *pkt_info, | 
 | 				 struct xlgmac_dma_desc *dma_desc) | 
 | { | 
 | 	u32 tsa, tsd; | 
 | 	u64 nsec; | 
 |  | 
 | 	tsa = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				     RX_CONTEXT_DESC3_TSA_POS, | 
 | 				RX_CONTEXT_DESC3_TSA_LEN); | 
 | 	tsd = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				     RX_CONTEXT_DESC3_TSD_POS, | 
 | 				RX_CONTEXT_DESC3_TSD_LEN); | 
 | 	if (tsa && !tsd) { | 
 | 		nsec = le32_to_cpu(dma_desc->desc1); | 
 | 		nsec <<= 32; | 
 | 		nsec |= le32_to_cpu(dma_desc->desc0); | 
 | 		if (nsec != 0xffffffffffffffffULL) { | 
 | 			pkt_info->rx_tstamp = nsec; | 
 | 			pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 					pkt_info->attributes, | 
 | 					RX_PACKET_ATTRIBUTES_RX_TSTAMP_POS, | 
 | 					RX_PACKET_ATTRIBUTES_RX_TSTAMP_LEN, | 
 | 					1); | 
 | 		} | 
 | 	} | 
 | } | 
 |  | 
 | static void xlgmac_tx_desc_reset(struct xlgmac_desc_data *desc_data) | 
 | { | 
 | 	struct xlgmac_dma_desc *dma_desc = desc_data->dma_desc; | 
 |  | 
 | 	/* Reset the Tx descriptor | 
 | 	 *   Set buffer 1 (lo) address to zero | 
 | 	 *   Set buffer 1 (hi) address to zero | 
 | 	 *   Reset all other control bits (IC, TTSE, B2L & B1L) | 
 | 	 *   Reset all other control bits (OWN, CTXT, FD, LD, CPC, CIC, etc) | 
 | 	 */ | 
 | 	dma_desc->desc0 = 0; | 
 | 	dma_desc->desc1 = 0; | 
 | 	dma_desc->desc2 = 0; | 
 | 	dma_desc->desc3 = 0; | 
 |  | 
 | 	/* Make sure ownership is written to the descriptor */ | 
 | 	dma_wmb(); | 
 | } | 
 |  | 
 | static void xlgmac_tx_desc_init(struct xlgmac_channel *channel) | 
 | { | 
 | 	struct xlgmac_ring *ring = channel->tx_ring; | 
 | 	struct xlgmac_desc_data *desc_data; | 
 | 	int start_index = ring->cur; | 
 | 	int i; | 
 |  | 
 | 	/* Initialze all descriptors */ | 
 | 	for (i = 0; i < ring->dma_desc_count; i++) { | 
 | 		desc_data = XLGMAC_GET_DESC_DATA(ring, i); | 
 |  | 
 | 		/* Initialize Tx descriptor */ | 
 | 		xlgmac_tx_desc_reset(desc_data); | 
 | 	} | 
 |  | 
 | 	/* Update the total number of Tx descriptors */ | 
 | 	writel(ring->dma_desc_count - 1, XLGMAC_DMA_REG(channel, DMA_CH_TDRLR)); | 
 |  | 
 | 	/* Update the starting address of descriptor ring */ | 
 | 	desc_data = XLGMAC_GET_DESC_DATA(ring, start_index); | 
 | 	writel(upper_32_bits(desc_data->dma_desc_addr), | 
 | 	       XLGMAC_DMA_REG(channel, DMA_CH_TDLR_HI)); | 
 | 	writel(lower_32_bits(desc_data->dma_desc_addr), | 
 | 	       XLGMAC_DMA_REG(channel, DMA_CH_TDLR_LO)); | 
 | } | 
 |  | 
 | static void xlgmac_rx_desc_reset(struct xlgmac_pdata *pdata, | 
 | 				 struct xlgmac_desc_data *desc_data, | 
 | 				 unsigned int index) | 
 | { | 
 | 	struct xlgmac_dma_desc *dma_desc = desc_data->dma_desc; | 
 | 	unsigned int rx_frames = pdata->rx_frames; | 
 | 	unsigned int rx_usecs = pdata->rx_usecs; | 
 | 	dma_addr_t hdr_dma, buf_dma; | 
 | 	unsigned int inte; | 
 |  | 
 | 	if (!rx_usecs && !rx_frames) { | 
 | 		/* No coalescing, interrupt for every descriptor */ | 
 | 		inte = 1; | 
 | 	} else { | 
 | 		/* Set interrupt based on Rx frame coalescing setting */ | 
 | 		if (rx_frames && !((index + 1) % rx_frames)) | 
 | 			inte = 1; | 
 | 		else | 
 | 			inte = 0; | 
 | 	} | 
 |  | 
 | 	/* Reset the Rx descriptor | 
 | 	 *   Set buffer 1 (lo) address to header dma address (lo) | 
 | 	 *   Set buffer 1 (hi) address to header dma address (hi) | 
 | 	 *   Set buffer 2 (lo) address to buffer dma address (lo) | 
 | 	 *   Set buffer 2 (hi) address to buffer dma address (hi) and | 
 | 	 *     set control bits OWN and INTE | 
 | 	 */ | 
 | 	hdr_dma = desc_data->rx.hdr.dma_base + desc_data->rx.hdr.dma_off; | 
 | 	buf_dma = desc_data->rx.buf.dma_base + desc_data->rx.buf.dma_off; | 
 | 	dma_desc->desc0 = cpu_to_le32(lower_32_bits(hdr_dma)); | 
 | 	dma_desc->desc1 = cpu_to_le32(upper_32_bits(hdr_dma)); | 
 | 	dma_desc->desc2 = cpu_to_le32(lower_32_bits(buf_dma)); | 
 | 	dma_desc->desc3 = cpu_to_le32(upper_32_bits(buf_dma)); | 
 |  | 
 | 	dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 				dma_desc->desc3, | 
 | 				RX_NORMAL_DESC3_INTE_POS, | 
 | 				RX_NORMAL_DESC3_INTE_LEN, | 
 | 				inte); | 
 |  | 
 | 	/* Since the Rx DMA engine is likely running, make sure everything | 
 | 	 * is written to the descriptor(s) before setting the OWN bit | 
 | 	 * for the descriptor | 
 | 	 */ | 
 | 	dma_wmb(); | 
 |  | 
 | 	dma_desc->desc3 = XLGMAC_SET_REG_BITS_LE( | 
 | 				dma_desc->desc3, | 
 | 				RX_NORMAL_DESC3_OWN_POS, | 
 | 				RX_NORMAL_DESC3_OWN_LEN, | 
 | 				1); | 
 |  | 
 | 	/* Make sure ownership is written to the descriptor */ | 
 | 	dma_wmb(); | 
 | } | 
 |  | 
 | static void xlgmac_rx_desc_init(struct xlgmac_channel *channel) | 
 | { | 
 | 	struct xlgmac_pdata *pdata = channel->pdata; | 
 | 	struct xlgmac_ring *ring = channel->rx_ring; | 
 | 	unsigned int start_index = ring->cur; | 
 | 	struct xlgmac_desc_data *desc_data; | 
 | 	unsigned int i; | 
 |  | 
 | 	/* Initialize all descriptors */ | 
 | 	for (i = 0; i < ring->dma_desc_count; i++) { | 
 | 		desc_data = XLGMAC_GET_DESC_DATA(ring, i); | 
 |  | 
 | 		/* Initialize Rx descriptor */ | 
 | 		xlgmac_rx_desc_reset(pdata, desc_data, i); | 
 | 	} | 
 |  | 
 | 	/* Update the total number of Rx descriptors */ | 
 | 	writel(ring->dma_desc_count - 1, XLGMAC_DMA_REG(channel, DMA_CH_RDRLR)); | 
 |  | 
 | 	/* Update the starting address of descriptor ring */ | 
 | 	desc_data = XLGMAC_GET_DESC_DATA(ring, start_index); | 
 | 	writel(upper_32_bits(desc_data->dma_desc_addr), | 
 | 	       XLGMAC_DMA_REG(channel, DMA_CH_RDLR_HI)); | 
 | 	writel(lower_32_bits(desc_data->dma_desc_addr), | 
 | 	       XLGMAC_DMA_REG(channel, DMA_CH_RDLR_LO)); | 
 |  | 
 | 	/* Update the Rx Descriptor Tail Pointer */ | 
 | 	desc_data = XLGMAC_GET_DESC_DATA(ring, start_index + | 
 | 					  ring->dma_desc_count - 1); | 
 | 	writel(lower_32_bits(desc_data->dma_desc_addr), | 
 | 	       XLGMAC_DMA_REG(channel, DMA_CH_RDTR_LO)); | 
 | } | 
 |  | 
 | static int xlgmac_is_context_desc(struct xlgmac_dma_desc *dma_desc) | 
 | { | 
 | 	/* Rx and Tx share CTXT bit, so check TDES3.CTXT bit */ | 
 | 	return XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				TX_NORMAL_DESC3_CTXT_POS, | 
 | 				TX_NORMAL_DESC3_CTXT_LEN); | 
 | } | 
 |  | 
 | static int xlgmac_is_last_desc(struct xlgmac_dma_desc *dma_desc) | 
 | { | 
 | 	/* Rx and Tx share LD bit, so check TDES3.LD bit */ | 
 | 	return XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				TX_NORMAL_DESC3_LD_POS, | 
 | 				TX_NORMAL_DESC3_LD_LEN); | 
 | } | 
 |  | 
 | static int xlgmac_disable_tx_flow_control(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int max_q_count, q_count; | 
 | 	unsigned int reg, regval; | 
 | 	unsigned int i; | 
 |  | 
 | 	/* Clear MTL flow control */ | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_EHFC_POS, | 
 | 					     MTL_Q_RQOMR_EHFC_LEN, 0); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 	} | 
 |  | 
 | 	/* Clear MAC flow control */ | 
 | 	max_q_count = XLGMAC_MAX_FLOW_CONTROL_QUEUES; | 
 | 	q_count = min_t(unsigned int, pdata->tx_q_count, max_q_count); | 
 | 	reg = MAC_Q0TFCR; | 
 | 	for (i = 0; i < q_count; i++) { | 
 | 		regval = readl(pdata->mac_regs + reg); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, | 
 | 					     MAC_Q0TFCR_TFE_POS, | 
 | 					MAC_Q0TFCR_TFE_LEN, | 
 | 					0); | 
 | 		writel(regval, pdata->mac_regs + reg); | 
 |  | 
 | 		reg += MAC_QTFCR_INC; | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_enable_tx_flow_control(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int max_q_count, q_count; | 
 | 	unsigned int reg, regval; | 
 | 	unsigned int i; | 
 |  | 
 | 	/* Set MTL flow control */ | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_EHFC_POS, | 
 | 					     MTL_Q_RQOMR_EHFC_LEN, 1); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 	} | 
 |  | 
 | 	/* Set MAC flow control */ | 
 | 	max_q_count = XLGMAC_MAX_FLOW_CONTROL_QUEUES; | 
 | 	q_count = min_t(unsigned int, pdata->tx_q_count, max_q_count); | 
 | 	reg = MAC_Q0TFCR; | 
 | 	for (i = 0; i < q_count; i++) { | 
 | 		regval = readl(pdata->mac_regs + reg); | 
 |  | 
 | 		/* Enable transmit flow control */ | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MAC_Q0TFCR_TFE_POS, | 
 | 					     MAC_Q0TFCR_TFE_LEN, 1); | 
 | 		/* Set pause time */ | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MAC_Q0TFCR_PT_POS, | 
 | 					     MAC_Q0TFCR_PT_LEN, 0xffff); | 
 |  | 
 | 		writel(regval, pdata->mac_regs + reg); | 
 |  | 
 | 		reg += MAC_QTFCR_INC; | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_disable_rx_flow_control(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_RFCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RFCR_RFE_POS, | 
 | 				     MAC_RFCR_RFE_LEN, 0); | 
 | 	writel(regval, pdata->mac_regs + MAC_RFCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_enable_rx_flow_control(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_RFCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RFCR_RFE_POS, | 
 | 				     MAC_RFCR_RFE_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MAC_RFCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_config_tx_flow_control(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	if (pdata->tx_pause) | 
 | 		xlgmac_enable_tx_flow_control(pdata); | 
 | 	else | 
 | 		xlgmac_disable_tx_flow_control(pdata); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_config_rx_flow_control(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	if (pdata->rx_pause) | 
 | 		xlgmac_enable_rx_flow_control(pdata); | 
 | 	else | 
 | 		xlgmac_disable_rx_flow_control(pdata); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_config_rx_coalesce(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->rx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RIWT)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RIWT_RWT_POS, | 
 | 					     DMA_CH_RIWT_RWT_LEN, | 
 | 					     pdata->rx_riwt); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RIWT)); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_config_flow_control(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	xlgmac_config_tx_flow_control(pdata); | 
 | 	xlgmac_config_rx_flow_control(pdata); | 
 | } | 
 |  | 
 | static void xlgmac_config_rx_fep_enable(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_FEP_POS, | 
 | 					     MTL_Q_RQOMR_FEP_LEN, 1); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 	} | 
 | } | 
 |  | 
 | static void xlgmac_config_rx_fup_enable(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_FUP_POS, | 
 | 					     MTL_Q_RQOMR_FUP_LEN, 1); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 	} | 
 | } | 
 |  | 
 | static int xlgmac_config_tx_coalesce(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_config_rx_buffer_size(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->rx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RCR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RCR_RBSZ_POS, | 
 | 					     DMA_CH_RCR_RBSZ_LEN, | 
 | 					pdata->rx_buf_size); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RCR)); | 
 | 	} | 
 | } | 
 |  | 
 | static void xlgmac_config_tso_mode(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->tx_ring) | 
 | 			break; | 
 |  | 
 | 		if (pdata->hw_feat.tso) { | 
 | 			regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 			regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_TSE_POS, | 
 | 						     DMA_CH_TCR_TSE_LEN, 1); | 
 | 			writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 		} | 
 | 	} | 
 | } | 
 |  | 
 | static void xlgmac_config_sph_mode(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->rx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_CR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_CR_SPH_POS, | 
 | 					     DMA_CH_CR_SPH_LEN, 1); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_CR)); | 
 | 	} | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_RCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RCR_HDSMS_POS, | 
 | 				     MAC_RCR_HDSMS_LEN, | 
 | 				XLGMAC_SPH_HDSMS_SIZE); | 
 | 	writel(regval, pdata->mac_regs + MAC_RCR); | 
 | } | 
 |  | 
 | static unsigned int xlgmac_usec_to_riwt(struct xlgmac_pdata *pdata, | 
 | 					unsigned int usec) | 
 | { | 
 | 	unsigned long rate; | 
 | 	unsigned int ret; | 
 |  | 
 | 	rate = pdata->sysclk_rate; | 
 |  | 
 | 	/* Convert the input usec value to the watchdog timer value. Each | 
 | 	 * watchdog timer value is equivalent to 256 clock cycles. | 
 | 	 * Calculate the required value as: | 
 | 	 *   ( usec * ( system_clock_mhz / 10^6 ) / 256 | 
 | 	 */ | 
 | 	ret = (usec * (rate / 1000000)) / 256; | 
 |  | 
 | 	return ret; | 
 | } | 
 |  | 
 | static unsigned int xlgmac_riwt_to_usec(struct xlgmac_pdata *pdata, | 
 | 					unsigned int riwt) | 
 | { | 
 | 	unsigned long rate; | 
 | 	unsigned int ret; | 
 |  | 
 | 	rate = pdata->sysclk_rate; | 
 |  | 
 | 	/* Convert the input watchdog timer value to the usec value. Each | 
 | 	 * watchdog timer value is equivalent to 256 clock cycles. | 
 | 	 * Calculate the required value as: | 
 | 	 *   ( riwt * 256 ) / ( system_clock_mhz / 10^6 ) | 
 | 	 */ | 
 | 	ret = (riwt * 256) / (rate / 1000000); | 
 |  | 
 | 	return ret; | 
 | } | 
 |  | 
 | static int xlgmac_config_rx_threshold(struct xlgmac_pdata *pdata, | 
 | 				      unsigned int val) | 
 | { | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_RTC_POS, | 
 | 					     MTL_Q_RQOMR_RTC_LEN, val); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_config_mtl_mode(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	/* Set Tx to weighted round robin scheduling algorithm */ | 
 | 	regval = readl(pdata->mac_regs + MTL_OMR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MTL_OMR_ETSALG_POS, | 
 | 				     MTL_OMR_ETSALG_LEN, MTL_ETSALG_WRR); | 
 | 	writel(regval, pdata->mac_regs + MTL_OMR); | 
 |  | 
 | 	/* Set Tx traffic classes to use WRR algorithm with equal weights */ | 
 | 	for (i = 0; i < pdata->hw_feat.tc_cnt; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_TC_ETSCR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_TC_ETSCR_TSA_POS, | 
 | 					     MTL_TC_ETSCR_TSA_LEN, MTL_TSA_ETS); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_TC_ETSCR)); | 
 |  | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_TC_QWR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_TC_QWR_QW_POS, | 
 | 					     MTL_TC_QWR_QW_LEN, 1); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_TC_QWR)); | 
 | 	} | 
 |  | 
 | 	/* Set Rx to strict priority algorithm */ | 
 | 	regval = readl(pdata->mac_regs + MTL_OMR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MTL_OMR_RAA_POS, | 
 | 				     MTL_OMR_RAA_LEN, MTL_RAA_SP); | 
 | 	writel(regval, pdata->mac_regs + MTL_OMR); | 
 | } | 
 |  | 
 | static void xlgmac_config_queue_mapping(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int ppq, ppq_extra, prio, prio_queues; | 
 | 	unsigned int qptc, qptc_extra, queue; | 
 | 	unsigned int reg, regval; | 
 | 	unsigned int mask; | 
 | 	unsigned int i, j; | 
 |  | 
 | 	/* Map the MTL Tx Queues to Traffic Classes | 
 | 	 *   Note: Tx Queues >= Traffic Classes | 
 | 	 */ | 
 | 	qptc = pdata->tx_q_count / pdata->hw_feat.tc_cnt; | 
 | 	qptc_extra = pdata->tx_q_count % pdata->hw_feat.tc_cnt; | 
 |  | 
 | 	for (i = 0, queue = 0; i < pdata->hw_feat.tc_cnt; i++) { | 
 | 		for (j = 0; j < qptc; j++) { | 
 | 			netif_dbg(pdata, drv, pdata->netdev, | 
 | 				  "TXq%u mapped to TC%u\n", queue, i); | 
 | 			regval = readl(XLGMAC_MTL_REG(pdata, queue, | 
 | 						      MTL_Q_TQOMR)); | 
 | 			regval = XLGMAC_SET_REG_BITS(regval, | 
 | 						     MTL_Q_TQOMR_Q2TCMAP_POS, | 
 | 						     MTL_Q_TQOMR_Q2TCMAP_LEN, | 
 | 						     i); | 
 | 			writel(regval, XLGMAC_MTL_REG(pdata, queue, | 
 | 						      MTL_Q_TQOMR)); | 
 | 			queue++; | 
 | 		} | 
 |  | 
 | 		if (i < qptc_extra) { | 
 | 			netif_dbg(pdata, drv, pdata->netdev, | 
 | 				  "TXq%u mapped to TC%u\n", queue, i); | 
 | 			regval = readl(XLGMAC_MTL_REG(pdata, queue, | 
 | 						      MTL_Q_TQOMR)); | 
 | 			regval = XLGMAC_SET_REG_BITS(regval, | 
 | 						     MTL_Q_TQOMR_Q2TCMAP_POS, | 
 | 						     MTL_Q_TQOMR_Q2TCMAP_LEN, | 
 | 						     i); | 
 | 			writel(regval, XLGMAC_MTL_REG(pdata, queue, | 
 | 						      MTL_Q_TQOMR)); | 
 | 			queue++; | 
 | 		} | 
 | 	} | 
 |  | 
 | 	/* Map the 8 VLAN priority values to available MTL Rx queues */ | 
 | 	prio_queues = min_t(unsigned int, IEEE_8021QAZ_MAX_TCS, | 
 | 			    pdata->rx_q_count); | 
 | 	ppq = IEEE_8021QAZ_MAX_TCS / prio_queues; | 
 | 	ppq_extra = IEEE_8021QAZ_MAX_TCS % prio_queues; | 
 |  | 
 | 	reg = MAC_RQC2R; | 
 | 	regval = 0; | 
 | 	for (i = 0, prio = 0; i < prio_queues;) { | 
 | 		mask = 0; | 
 | 		for (j = 0; j < ppq; j++) { | 
 | 			netif_dbg(pdata, drv, pdata->netdev, | 
 | 				  "PRIO%u mapped to RXq%u\n", prio, i); | 
 | 			mask |= (1 << prio); | 
 | 			prio++; | 
 | 		} | 
 |  | 
 | 		if (i < ppq_extra) { | 
 | 			netif_dbg(pdata, drv, pdata->netdev, | 
 | 				  "PRIO%u mapped to RXq%u\n", prio, i); | 
 | 			mask |= (1 << prio); | 
 | 			prio++; | 
 | 		} | 
 |  | 
 | 		regval |= (mask << ((i++ % MAC_RQC2_Q_PER_REG) << 3)); | 
 |  | 
 | 		if ((i % MAC_RQC2_Q_PER_REG) && (i != prio_queues)) | 
 | 			continue; | 
 |  | 
 | 		writel(regval, pdata->mac_regs + reg); | 
 | 		reg += MAC_RQC2_INC; | 
 | 		regval = 0; | 
 | 	} | 
 |  | 
 | 	/* Configure one to one, MTL Rx queue to DMA Rx channel mapping | 
 | 	 *  ie Q0 <--> CH0, Q1 <--> CH1 ... Q11 <--> CH11 | 
 | 	 */ | 
 | 	reg = MTL_RQDCM0R; | 
 | 	regval = readl(pdata->mac_regs + reg); | 
 | 	regval |= (MTL_RQDCM0R_Q0MDMACH | MTL_RQDCM0R_Q1MDMACH | | 
 | 		    MTL_RQDCM0R_Q2MDMACH | MTL_RQDCM0R_Q3MDMACH); | 
 | 	writel(regval, pdata->mac_regs + reg); | 
 |  | 
 | 	reg += MTL_RQDCM_INC; | 
 | 	regval = readl(pdata->mac_regs + reg); | 
 | 	regval |= (MTL_RQDCM1R_Q4MDMACH | MTL_RQDCM1R_Q5MDMACH | | 
 | 		    MTL_RQDCM1R_Q6MDMACH | MTL_RQDCM1R_Q7MDMACH); | 
 | 	writel(regval, pdata->mac_regs + reg); | 
 |  | 
 | 	reg += MTL_RQDCM_INC; | 
 | 	regval = readl(pdata->mac_regs + reg); | 
 | 	regval |= (MTL_RQDCM2R_Q8MDMACH | MTL_RQDCM2R_Q9MDMACH | | 
 | 		    MTL_RQDCM2R_Q10MDMACH | MTL_RQDCM2R_Q11MDMACH); | 
 | 	writel(regval, pdata->mac_regs + reg); | 
 | } | 
 |  | 
 | static unsigned int xlgmac_calculate_per_queue_fifo( | 
 | 					unsigned int fifo_size, | 
 | 					unsigned int queue_count) | 
 | { | 
 | 	unsigned int q_fifo_size; | 
 | 	unsigned int p_fifo; | 
 |  | 
 | 	/* Calculate the configured fifo size */ | 
 | 	q_fifo_size = 1 << (fifo_size + 7); | 
 |  | 
 | 	/* The configured value may not be the actual amount of fifo RAM */ | 
 | 	q_fifo_size = min_t(unsigned int, XLGMAC_MAX_FIFO, q_fifo_size); | 
 |  | 
 | 	q_fifo_size = q_fifo_size / queue_count; | 
 |  | 
 | 	/* Each increment in the queue fifo size represents 256 bytes of | 
 | 	 * fifo, with 0 representing 256 bytes. Distribute the fifo equally | 
 | 	 * between the queues. | 
 | 	 */ | 
 | 	p_fifo = q_fifo_size / 256; | 
 | 	if (p_fifo) | 
 | 		p_fifo--; | 
 |  | 
 | 	return p_fifo; | 
 | } | 
 |  | 
 | static void xlgmac_config_tx_fifo_size(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int fifo_size; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	fifo_size = xlgmac_calculate_per_queue_fifo( | 
 | 				pdata->hw_feat.tx_fifo_size, | 
 | 				pdata->tx_q_count); | 
 |  | 
 | 	for (i = 0; i < pdata->tx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TQS_POS, | 
 | 					     MTL_Q_TQOMR_TQS_LEN, fifo_size); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 	} | 
 |  | 
 | 	netif_info(pdata, drv, pdata->netdev, | 
 | 		   "%d Tx hardware queues, %d byte fifo per queue\n", | 
 | 		   pdata->tx_q_count, ((fifo_size + 1) * 256)); | 
 | } | 
 |  | 
 | static void xlgmac_config_rx_fifo_size(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int fifo_size; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	fifo_size = xlgmac_calculate_per_queue_fifo( | 
 | 					pdata->hw_feat.rx_fifo_size, | 
 | 					pdata->rx_q_count); | 
 |  | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_RQS_POS, | 
 | 					     MTL_Q_RQOMR_RQS_LEN, fifo_size); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 	} | 
 |  | 
 | 	netif_info(pdata, drv, pdata->netdev, | 
 | 		   "%d Rx hardware queues, %d byte fifo per queue\n", | 
 | 		   pdata->rx_q_count, ((fifo_size + 1) * 256)); | 
 | } | 
 |  | 
 | static void xlgmac_config_flow_control_threshold(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQFCR)); | 
 | 		/* Activate flow control when less than 4k left in fifo */ | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQFCR_RFA_POS, | 
 | 					     MTL_Q_RQFCR_RFA_LEN, 2); | 
 | 		/* De-activate flow control when more than 6k left in fifo */ | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQFCR_RFD_POS, | 
 | 					     MTL_Q_RQFCR_RFD_LEN, 4); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQFCR)); | 
 | 	} | 
 | } | 
 |  | 
 | static int xlgmac_config_tx_threshold(struct xlgmac_pdata *pdata, | 
 | 				      unsigned int val) | 
 | { | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	for (i = 0; i < pdata->tx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TTC_POS, | 
 | 					     MTL_Q_TQOMR_TTC_LEN, val); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_config_rsf_mode(struct xlgmac_pdata *pdata, | 
 | 				  unsigned int val) | 
 | { | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	for (i = 0; i < pdata->rx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_RQOMR_RSF_POS, | 
 | 					     MTL_Q_RQOMR_RSF_LEN, val); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_RQOMR)); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_config_tsf_mode(struct xlgmac_pdata *pdata, | 
 | 				  unsigned int val) | 
 | { | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	for (i = 0; i < pdata->tx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_TSF_POS, | 
 | 					     MTL_Q_TQOMR_TSF_LEN, val); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_config_osp_mode(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->tx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_OSP_POS, | 
 | 					     DMA_CH_TCR_OSP_LEN, | 
 | 					pdata->tx_osp_mode); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_config_pblx8(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_CR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_CR_PBLX8_POS, | 
 | 					     DMA_CH_CR_PBLX8_LEN, | 
 | 					pdata->pblx8); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_CR)); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_get_tx_pbl_val(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(XLGMAC_DMA_REG(pdata->channel_head, DMA_CH_TCR)); | 
 | 	regval = XLGMAC_GET_REG_BITS(regval, DMA_CH_TCR_PBL_POS, | 
 | 				     DMA_CH_TCR_PBL_LEN); | 
 | 	return regval; | 
 | } | 
 |  | 
 | static int xlgmac_config_tx_pbl_val(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->tx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_TCR_PBL_POS, | 
 | 					     DMA_CH_TCR_PBL_LEN, | 
 | 					pdata->tx_pbl); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_TCR)); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_get_rx_pbl_val(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(XLGMAC_DMA_REG(pdata->channel_head, DMA_CH_RCR)); | 
 | 	regval = XLGMAC_GET_REG_BITS(regval, DMA_CH_RCR_PBL_POS, | 
 | 				     DMA_CH_RCR_PBL_LEN); | 
 | 	return regval; | 
 | } | 
 |  | 
 | static int xlgmac_config_rx_pbl_val(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 | 	u32 regval; | 
 |  | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		if (!channel->rx_ring) | 
 | 			break; | 
 |  | 
 | 		regval = readl(XLGMAC_DMA_REG(channel, DMA_CH_RCR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, DMA_CH_RCR_PBL_POS, | 
 | 					     DMA_CH_RCR_PBL_LEN, | 
 | 					pdata->rx_pbl); | 
 | 		writel(regval, XLGMAC_DMA_REG(channel, DMA_CH_RCR)); | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static u64 xlgmac_mmc_read(struct xlgmac_pdata *pdata, unsigned int reg_lo) | 
 | { | 
 | 	bool read_hi; | 
 | 	u64 val; | 
 |  | 
 | 	switch (reg_lo) { | 
 | 	/* These registers are always 64 bit */ | 
 | 	case MMC_TXOCTETCOUNT_GB_LO: | 
 | 	case MMC_TXOCTETCOUNT_G_LO: | 
 | 	case MMC_RXOCTETCOUNT_GB_LO: | 
 | 	case MMC_RXOCTETCOUNT_G_LO: | 
 | 		read_hi = true; | 
 | 		break; | 
 |  | 
 | 	default: | 
 | 		read_hi = false; | 
 | 	} | 
 |  | 
 | 	val = (u64)readl(pdata->mac_regs + reg_lo); | 
 |  | 
 | 	if (read_hi) | 
 | 		val |= ((u64)readl(pdata->mac_regs + reg_lo + 4) << 32); | 
 |  | 
 | 	return val; | 
 | } | 
 |  | 
 | static void xlgmac_tx_mmc_int(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int mmc_isr = readl(pdata->mac_regs + MMC_TISR); | 
 | 	struct xlgmac_stats *stats = &pdata->stats; | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXOCTETCOUNT_GB_POS, | 
 | 				MMC_TISR_TXOCTETCOUNT_GB_LEN)) | 
 | 		stats->txoctetcount_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXOCTETCOUNT_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXFRAMECOUNT_GB_POS, | 
 | 				MMC_TISR_TXFRAMECOUNT_GB_LEN)) | 
 | 		stats->txframecount_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXFRAMECOUNT_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXBROADCASTFRAMES_G_POS, | 
 | 				MMC_TISR_TXBROADCASTFRAMES_G_LEN)) | 
 | 		stats->txbroadcastframes_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXBROADCASTFRAMES_G_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXMULTICASTFRAMES_G_POS, | 
 | 				MMC_TISR_TXMULTICASTFRAMES_G_LEN)) | 
 | 		stats->txmulticastframes_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXMULTICASTFRAMES_G_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TX64OCTETS_GB_POS, | 
 | 				MMC_TISR_TX64OCTETS_GB_LEN)) | 
 | 		stats->tx64octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TX64OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TX65TO127OCTETS_GB_POS, | 
 | 				MMC_TISR_TX65TO127OCTETS_GB_LEN)) | 
 | 		stats->tx65to127octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TX65TO127OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TX128TO255OCTETS_GB_POS, | 
 | 				MMC_TISR_TX128TO255OCTETS_GB_LEN)) | 
 | 		stats->tx128to255octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TX128TO255OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TX256TO511OCTETS_GB_POS, | 
 | 				MMC_TISR_TX256TO511OCTETS_GB_LEN)) | 
 | 		stats->tx256to511octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TX256TO511OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TX512TO1023OCTETS_GB_POS, | 
 | 				MMC_TISR_TX512TO1023OCTETS_GB_LEN)) | 
 | 		stats->tx512to1023octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TX512TO1023OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TX1024TOMAXOCTETS_GB_POS, | 
 | 				MMC_TISR_TX1024TOMAXOCTETS_GB_LEN)) | 
 | 		stats->tx1024tomaxoctets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TX1024TOMAXOCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXUNICASTFRAMES_GB_POS, | 
 | 				MMC_TISR_TXUNICASTFRAMES_GB_LEN)) | 
 | 		stats->txunicastframes_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXUNICASTFRAMES_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXMULTICASTFRAMES_GB_POS, | 
 | 				MMC_TISR_TXMULTICASTFRAMES_GB_LEN)) | 
 | 		stats->txmulticastframes_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXMULTICASTFRAMES_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXBROADCASTFRAMES_GB_POS, | 
 | 				MMC_TISR_TXBROADCASTFRAMES_GB_LEN)) | 
 | 		stats->txbroadcastframes_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXBROADCASTFRAMES_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXUNDERFLOWERROR_POS, | 
 | 				MMC_TISR_TXUNDERFLOWERROR_LEN)) | 
 | 		stats->txunderflowerror += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXUNDERFLOWERROR_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXOCTETCOUNT_G_POS, | 
 | 				MMC_TISR_TXOCTETCOUNT_G_LEN)) | 
 | 		stats->txoctetcount_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXOCTETCOUNT_G_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXFRAMECOUNT_G_POS, | 
 | 				MMC_TISR_TXFRAMECOUNT_G_LEN)) | 
 | 		stats->txframecount_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXFRAMECOUNT_G_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXPAUSEFRAMES_POS, | 
 | 				MMC_TISR_TXPAUSEFRAMES_LEN)) | 
 | 		stats->txpauseframes += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXPAUSEFRAMES_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_TISR_TXVLANFRAMES_G_POS, | 
 | 				MMC_TISR_TXVLANFRAMES_G_LEN)) | 
 | 		stats->txvlanframes_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_TXVLANFRAMES_G_LO); | 
 | } | 
 |  | 
 | static void xlgmac_rx_mmc_int(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int mmc_isr = readl(pdata->mac_regs + MMC_RISR); | 
 | 	struct xlgmac_stats *stats = &pdata->stats; | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXFRAMECOUNT_GB_POS, | 
 | 				MMC_RISR_RXFRAMECOUNT_GB_LEN)) | 
 | 		stats->rxframecount_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXFRAMECOUNT_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXOCTETCOUNT_GB_POS, | 
 | 				MMC_RISR_RXOCTETCOUNT_GB_LEN)) | 
 | 		stats->rxoctetcount_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXOCTETCOUNT_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXOCTETCOUNT_G_POS, | 
 | 				MMC_RISR_RXOCTETCOUNT_G_LEN)) | 
 | 		stats->rxoctetcount_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXOCTETCOUNT_G_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXBROADCASTFRAMES_G_POS, | 
 | 				MMC_RISR_RXBROADCASTFRAMES_G_LEN)) | 
 | 		stats->rxbroadcastframes_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXBROADCASTFRAMES_G_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXMULTICASTFRAMES_G_POS, | 
 | 				MMC_RISR_RXMULTICASTFRAMES_G_LEN)) | 
 | 		stats->rxmulticastframes_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXMULTICASTFRAMES_G_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXCRCERROR_POS, | 
 | 				MMC_RISR_RXCRCERROR_LEN)) | 
 | 		stats->rxcrcerror += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXCRCERROR_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXRUNTERROR_POS, | 
 | 				MMC_RISR_RXRUNTERROR_LEN)) | 
 | 		stats->rxrunterror += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXRUNTERROR); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXJABBERERROR_POS, | 
 | 				MMC_RISR_RXJABBERERROR_LEN)) | 
 | 		stats->rxjabbererror += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXJABBERERROR); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXUNDERSIZE_G_POS, | 
 | 				MMC_RISR_RXUNDERSIZE_G_LEN)) | 
 | 		stats->rxundersize_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXUNDERSIZE_G); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXOVERSIZE_G_POS, | 
 | 				MMC_RISR_RXOVERSIZE_G_LEN)) | 
 | 		stats->rxoversize_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXOVERSIZE_G); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RX64OCTETS_GB_POS, | 
 | 				MMC_RISR_RX64OCTETS_GB_LEN)) | 
 | 		stats->rx64octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_RX64OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RX65TO127OCTETS_GB_POS, | 
 | 				MMC_RISR_RX65TO127OCTETS_GB_LEN)) | 
 | 		stats->rx65to127octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_RX65TO127OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RX128TO255OCTETS_GB_POS, | 
 | 				MMC_RISR_RX128TO255OCTETS_GB_LEN)) | 
 | 		stats->rx128to255octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_RX128TO255OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RX256TO511OCTETS_GB_POS, | 
 | 				MMC_RISR_RX256TO511OCTETS_GB_LEN)) | 
 | 		stats->rx256to511octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_RX256TO511OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RX512TO1023OCTETS_GB_POS, | 
 | 				MMC_RISR_RX512TO1023OCTETS_GB_LEN)) | 
 | 		stats->rx512to1023octets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_RX512TO1023OCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RX1024TOMAXOCTETS_GB_POS, | 
 | 				MMC_RISR_RX1024TOMAXOCTETS_GB_LEN)) | 
 | 		stats->rx1024tomaxoctets_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_RX1024TOMAXOCTETS_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXUNICASTFRAMES_G_POS, | 
 | 				MMC_RISR_RXUNICASTFRAMES_G_LEN)) | 
 | 		stats->rxunicastframes_g += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXUNICASTFRAMES_G_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXLENGTHERROR_POS, | 
 | 				MMC_RISR_RXLENGTHERROR_LEN)) | 
 | 		stats->rxlengtherror += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXLENGTHERROR_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXOUTOFRANGETYPE_POS, | 
 | 				MMC_RISR_RXOUTOFRANGETYPE_LEN)) | 
 | 		stats->rxoutofrangetype += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXOUTOFRANGETYPE_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXPAUSEFRAMES_POS, | 
 | 				MMC_RISR_RXPAUSEFRAMES_LEN)) | 
 | 		stats->rxpauseframes += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXPAUSEFRAMES_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXFIFOOVERFLOW_POS, | 
 | 				MMC_RISR_RXFIFOOVERFLOW_LEN)) | 
 | 		stats->rxfifooverflow += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXFIFOOVERFLOW_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXVLANFRAMES_GB_POS, | 
 | 				MMC_RISR_RXVLANFRAMES_GB_LEN)) | 
 | 		stats->rxvlanframes_gb += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXVLANFRAMES_GB_LO); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS(mmc_isr, | 
 | 				MMC_RISR_RXWATCHDOGERROR_POS, | 
 | 				MMC_RISR_RXWATCHDOGERROR_LEN)) | 
 | 		stats->rxwatchdogerror += | 
 | 			xlgmac_mmc_read(pdata, MMC_RXWATCHDOGERROR); | 
 | } | 
 |  | 
 | static void xlgmac_read_mmc_stats(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_stats *stats = &pdata->stats; | 
 | 	u32 regval; | 
 |  | 
 | 	/* Freeze counters */ | 
 | 	regval = readl(pdata->mac_regs + MMC_CR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MMC_CR_MCF_POS, | 
 | 				     MMC_CR_MCF_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MMC_CR); | 
 |  | 
 | 	stats->txoctetcount_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXOCTETCOUNT_GB_LO); | 
 |  | 
 | 	stats->txframecount_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXFRAMECOUNT_GB_LO); | 
 |  | 
 | 	stats->txbroadcastframes_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXBROADCASTFRAMES_G_LO); | 
 |  | 
 | 	stats->txmulticastframes_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXMULTICASTFRAMES_G_LO); | 
 |  | 
 | 	stats->tx64octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TX64OCTETS_GB_LO); | 
 |  | 
 | 	stats->tx65to127octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TX65TO127OCTETS_GB_LO); | 
 |  | 
 | 	stats->tx128to255octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TX128TO255OCTETS_GB_LO); | 
 |  | 
 | 	stats->tx256to511octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TX256TO511OCTETS_GB_LO); | 
 |  | 
 | 	stats->tx512to1023octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TX512TO1023OCTETS_GB_LO); | 
 |  | 
 | 	stats->tx1024tomaxoctets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TX1024TOMAXOCTETS_GB_LO); | 
 |  | 
 | 	stats->txunicastframes_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXUNICASTFRAMES_GB_LO); | 
 |  | 
 | 	stats->txmulticastframes_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXMULTICASTFRAMES_GB_LO); | 
 |  | 
 | 	stats->txbroadcastframes_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXBROADCASTFRAMES_GB_LO); | 
 |  | 
 | 	stats->txunderflowerror += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXUNDERFLOWERROR_LO); | 
 |  | 
 | 	stats->txoctetcount_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXOCTETCOUNT_G_LO); | 
 |  | 
 | 	stats->txframecount_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXFRAMECOUNT_G_LO); | 
 |  | 
 | 	stats->txpauseframes += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXPAUSEFRAMES_LO); | 
 |  | 
 | 	stats->txvlanframes_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_TXVLANFRAMES_G_LO); | 
 |  | 
 | 	stats->rxframecount_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXFRAMECOUNT_GB_LO); | 
 |  | 
 | 	stats->rxoctetcount_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXOCTETCOUNT_GB_LO); | 
 |  | 
 | 	stats->rxoctetcount_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXOCTETCOUNT_G_LO); | 
 |  | 
 | 	stats->rxbroadcastframes_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXBROADCASTFRAMES_G_LO); | 
 |  | 
 | 	stats->rxmulticastframes_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXMULTICASTFRAMES_G_LO); | 
 |  | 
 | 	stats->rxcrcerror += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXCRCERROR_LO); | 
 |  | 
 | 	stats->rxrunterror += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXRUNTERROR); | 
 |  | 
 | 	stats->rxjabbererror += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXJABBERERROR); | 
 |  | 
 | 	stats->rxundersize_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXUNDERSIZE_G); | 
 |  | 
 | 	stats->rxoversize_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXOVERSIZE_G); | 
 |  | 
 | 	stats->rx64octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_RX64OCTETS_GB_LO); | 
 |  | 
 | 	stats->rx65to127octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_RX65TO127OCTETS_GB_LO); | 
 |  | 
 | 	stats->rx128to255octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_RX128TO255OCTETS_GB_LO); | 
 |  | 
 | 	stats->rx256to511octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_RX256TO511OCTETS_GB_LO); | 
 |  | 
 | 	stats->rx512to1023octets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_RX512TO1023OCTETS_GB_LO); | 
 |  | 
 | 	stats->rx1024tomaxoctets_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_RX1024TOMAXOCTETS_GB_LO); | 
 |  | 
 | 	stats->rxunicastframes_g += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXUNICASTFRAMES_G_LO); | 
 |  | 
 | 	stats->rxlengtherror += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXLENGTHERROR_LO); | 
 |  | 
 | 	stats->rxoutofrangetype += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXOUTOFRANGETYPE_LO); | 
 |  | 
 | 	stats->rxpauseframes += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXPAUSEFRAMES_LO); | 
 |  | 
 | 	stats->rxfifooverflow += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXFIFOOVERFLOW_LO); | 
 |  | 
 | 	stats->rxvlanframes_gb += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXVLANFRAMES_GB_LO); | 
 |  | 
 | 	stats->rxwatchdogerror += | 
 | 		xlgmac_mmc_read(pdata, MMC_RXWATCHDOGERROR); | 
 |  | 
 | 	/* Un-freeze counters */ | 
 | 	regval = readl(pdata->mac_regs + MMC_CR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MMC_CR_MCF_POS, | 
 | 				     MMC_CR_MCF_LEN, 0); | 
 | 	writel(regval, pdata->mac_regs + MMC_CR); | 
 | } | 
 |  | 
 | static void xlgmac_config_mmc(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MMC_CR); | 
 | 	/* Set counters to reset on read */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MMC_CR_ROR_POS, | 
 | 				     MMC_CR_ROR_LEN, 1); | 
 | 	/* Reset the counters */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MMC_CR_CR_POS, | 
 | 				     MMC_CR_CR_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MMC_CR); | 
 | } | 
 |  | 
 | static int xlgmac_write_rss_reg(struct xlgmac_pdata *pdata, unsigned int type, | 
 | 				unsigned int index, unsigned int val) | 
 | { | 
 | 	unsigned int wait; | 
 | 	int ret = 0; | 
 | 	u32 regval; | 
 |  | 
 | 	mutex_lock(&pdata->rss_mutex); | 
 |  | 
 | 	regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_RSSAR), | 
 | 				     MAC_RSSAR_OB_POS, MAC_RSSAR_OB_LEN); | 
 | 	if (regval) { | 
 | 		ret = -EBUSY; | 
 | 		goto unlock; | 
 | 	} | 
 |  | 
 | 	writel(val, pdata->mac_regs + MAC_RSSDR); | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_RSSAR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSAR_RSSIA_POS, | 
 | 				     MAC_RSSAR_RSSIA_LEN, index); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSAR_ADDRT_POS, | 
 | 				     MAC_RSSAR_ADDRT_LEN, type); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSAR_CT_POS, | 
 | 				     MAC_RSSAR_CT_LEN, 0); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSAR_OB_POS, | 
 | 				     MAC_RSSAR_OB_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MAC_RSSAR); | 
 |  | 
 | 	wait = 1000; | 
 | 	while (wait--) { | 
 | 		regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_RSSAR), | 
 | 					     MAC_RSSAR_OB_POS, | 
 | 					     MAC_RSSAR_OB_LEN); | 
 | 		if (!regval) | 
 | 			goto unlock; | 
 |  | 
 | 		usleep_range(1000, 1500); | 
 | 	} | 
 |  | 
 | 	ret = -EBUSY; | 
 |  | 
 | unlock: | 
 | 	mutex_unlock(&pdata->rss_mutex); | 
 |  | 
 | 	return ret; | 
 | } | 
 |  | 
 | static int xlgmac_write_rss_hash_key(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int key_regs = sizeof(pdata->rss_key) / sizeof(u32); | 
 | 	unsigned int *key = (unsigned int *)&pdata->rss_key; | 
 | 	int ret; | 
 |  | 
 | 	while (key_regs--) { | 
 | 		ret = xlgmac_write_rss_reg(pdata, XLGMAC_RSS_HASH_KEY_TYPE, | 
 | 					   key_regs, *key++); | 
 | 		if (ret) | 
 | 			return ret; | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_write_rss_lookup_table(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int i; | 
 | 	int ret; | 
 |  | 
 | 	for (i = 0; i < ARRAY_SIZE(pdata->rss_table); i++) { | 
 | 		ret = xlgmac_write_rss_reg(pdata, | 
 | 					   XLGMAC_RSS_LOOKUP_TABLE_TYPE, i, | 
 | 					   pdata->rss_table[i]); | 
 | 		if (ret) | 
 | 			return ret; | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_set_rss_hash_key(struct xlgmac_pdata *pdata, const u8 *key) | 
 | { | 
 | 	memcpy(pdata->rss_key, key, sizeof(pdata->rss_key)); | 
 |  | 
 | 	return xlgmac_write_rss_hash_key(pdata); | 
 | } | 
 |  | 
 | static int xlgmac_set_rss_lookup_table(struct xlgmac_pdata *pdata, | 
 | 				       const u32 *table) | 
 | { | 
 | 	unsigned int i; | 
 | 	u32 tval; | 
 |  | 
 | 	for (i = 0; i < ARRAY_SIZE(pdata->rss_table); i++) { | 
 | 		tval = table[i]; | 
 | 		pdata->rss_table[i] = XLGMAC_SET_REG_BITS( | 
 | 						pdata->rss_table[i], | 
 | 						MAC_RSSDR_DMCH_POS, | 
 | 						MAC_RSSDR_DMCH_LEN, | 
 | 						tval); | 
 | 	} | 
 |  | 
 | 	return xlgmac_write_rss_lookup_table(pdata); | 
 | } | 
 |  | 
 | static int xlgmac_enable_rss(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 | 	int ret; | 
 |  | 
 | 	if (!pdata->hw_feat.rss) | 
 | 		return -EOPNOTSUPP; | 
 |  | 
 | 	/* Program the hash key */ | 
 | 	ret = xlgmac_write_rss_hash_key(pdata); | 
 | 	if (ret) | 
 | 		return ret; | 
 |  | 
 | 	/* Program the lookup table */ | 
 | 	ret = xlgmac_write_rss_lookup_table(pdata); | 
 | 	if (ret) | 
 | 		return ret; | 
 |  | 
 | 	/* Set the RSS options */ | 
 | 	writel(pdata->rss_options, pdata->mac_regs + MAC_RSSCR); | 
 |  | 
 | 	/* Enable RSS */ | 
 | 	regval = readl(pdata->mac_regs + MAC_RSSCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSCR_RSSE_POS, | 
 | 				     MAC_RSSCR_RSSE_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + MAC_RSSCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_disable_rss(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	if (!pdata->hw_feat.rss) | 
 | 		return -EOPNOTSUPP; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_RSSCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_RSSCR_RSSE_POS, | 
 | 				     MAC_RSSCR_RSSE_LEN, 0); | 
 | 	writel(regval, pdata->mac_regs + MAC_RSSCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_config_rss(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	int ret; | 
 |  | 
 | 	if (!pdata->hw_feat.rss) | 
 | 		return; | 
 |  | 
 | 	if (pdata->netdev->features & NETIF_F_RXHASH) | 
 | 		ret = xlgmac_enable_rss(pdata); | 
 | 	else | 
 | 		ret = xlgmac_disable_rss(pdata); | 
 |  | 
 | 	if (ret) | 
 | 		netdev_err(pdata->netdev, | 
 | 			   "error configuring RSS, RSS disabled\n"); | 
 | } | 
 |  | 
 | static void xlgmac_enable_dma_interrupts(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int dma_ch_isr, dma_ch_ier; | 
 | 	struct xlgmac_channel *channel; | 
 | 	unsigned int i; | 
 |  | 
 | 	channel = pdata->channel_head; | 
 | 	for (i = 0; i < pdata->channel_count; i++, channel++) { | 
 | 		/* Clear all the interrupts which are set */ | 
 | 		dma_ch_isr = readl(XLGMAC_DMA_REG(channel, DMA_CH_SR)); | 
 | 		writel(dma_ch_isr, XLGMAC_DMA_REG(channel, DMA_CH_SR)); | 
 |  | 
 | 		/* Clear all interrupt enable bits */ | 
 | 		dma_ch_ier = 0; | 
 |  | 
 | 		/* Enable following interrupts | 
 | 		 *   NIE  - Normal Interrupt Summary Enable | 
 | 		 *   AIE  - Abnormal Interrupt Summary Enable | 
 | 		 *   FBEE - Fatal Bus Error Enable | 
 | 		 */ | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS(dma_ch_ier, | 
 | 						 DMA_CH_IER_NIE_POS, | 
 | 					DMA_CH_IER_NIE_LEN, 1); | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS(dma_ch_ier, | 
 | 						 DMA_CH_IER_AIE_POS, | 
 | 					DMA_CH_IER_AIE_LEN, 1); | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS(dma_ch_ier, | 
 | 						 DMA_CH_IER_FBEE_POS, | 
 | 					DMA_CH_IER_FBEE_LEN, 1); | 
 |  | 
 | 		if (channel->tx_ring) { | 
 | 			/* Enable the following Tx interrupts | 
 | 			 *   TIE  - Transmit Interrupt Enable (unless using | 
 | 			 *          per channel interrupts) | 
 | 			 */ | 
 | 			if (!pdata->per_channel_irq) | 
 | 				dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 						dma_ch_ier, | 
 | 						DMA_CH_IER_TIE_POS, | 
 | 						DMA_CH_IER_TIE_LEN, | 
 | 						1); | 
 | 		} | 
 | 		if (channel->rx_ring) { | 
 | 			/* Enable following Rx interrupts | 
 | 			 *   RBUE - Receive Buffer Unavailable Enable | 
 | 			 *   RIE  - Receive Interrupt Enable (unless using | 
 | 			 *          per channel interrupts) | 
 | 			 */ | 
 | 			dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 					dma_ch_ier, | 
 | 					DMA_CH_IER_RBUE_POS, | 
 | 					DMA_CH_IER_RBUE_LEN, | 
 | 					1); | 
 | 			if (!pdata->per_channel_irq) | 
 | 				dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 						dma_ch_ier, | 
 | 						DMA_CH_IER_RIE_POS, | 
 | 						DMA_CH_IER_RIE_LEN, | 
 | 						1); | 
 | 		} | 
 |  | 
 | 		writel(dma_ch_isr, XLGMAC_DMA_REG(channel, DMA_CH_IER)); | 
 | 	} | 
 | } | 
 |  | 
 | static void xlgmac_enable_mtl_interrupts(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int q_count, i; | 
 | 	unsigned int mtl_q_isr; | 
 |  | 
 | 	q_count = max(pdata->hw_feat.tx_q_cnt, pdata->hw_feat.rx_q_cnt); | 
 | 	for (i = 0; i < q_count; i++) { | 
 | 		/* Clear all the interrupts which are set */ | 
 | 		mtl_q_isr = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_ISR)); | 
 | 		writel(mtl_q_isr, XLGMAC_MTL_REG(pdata, i, MTL_Q_ISR)); | 
 |  | 
 | 		/* No MTL interrupts to be enabled */ | 
 | 		writel(0, XLGMAC_MTL_REG(pdata, i, MTL_Q_IER)); | 
 | 	} | 
 | } | 
 |  | 
 | static void xlgmac_enable_mac_interrupts(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int mac_ier = 0; | 
 | 	u32 regval; | 
 |  | 
 | 	/* Enable Timestamp interrupt */ | 
 | 	mac_ier = XLGMAC_SET_REG_BITS(mac_ier, MAC_IER_TSIE_POS, | 
 | 				      MAC_IER_TSIE_LEN, 1); | 
 |  | 
 | 	writel(mac_ier, pdata->mac_regs + MAC_IER); | 
 |  | 
 | 	/* Enable all counter interrupts */ | 
 | 	regval = readl(pdata->mac_regs + MMC_RIER); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MMC_RIER_ALL_INTERRUPTS_POS, | 
 | 				     MMC_RIER_ALL_INTERRUPTS_LEN, 0xffffffff); | 
 | 	writel(regval, pdata->mac_regs + MMC_RIER); | 
 | 	regval = readl(pdata->mac_regs + MMC_TIER); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MMC_TIER_ALL_INTERRUPTS_POS, | 
 | 				     MMC_TIER_ALL_INTERRUPTS_LEN, 0xffffffff); | 
 | 	writel(regval, pdata->mac_regs + MMC_TIER); | 
 | } | 
 |  | 
 | static int xlgmac_set_xlgmii_25000_speed(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_TCR), | 
 | 				     MAC_TCR_SS_POS, MAC_TCR_SS_LEN); | 
 | 	if (regval == 0x1) | 
 | 		return 0; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_TCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_SS_POS, | 
 | 				     MAC_TCR_SS_LEN, 0x1); | 
 | 	writel(regval, pdata->mac_regs + MAC_TCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_set_xlgmii_40000_speed(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_TCR), | 
 | 				     MAC_TCR_SS_POS, MAC_TCR_SS_LEN); | 
 | 	if (regval == 0) | 
 | 		return 0; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_TCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_SS_POS, | 
 | 				     MAC_TCR_SS_LEN, 0); | 
 | 	writel(regval, pdata->mac_regs + MAC_TCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_set_xlgmii_50000_speed(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_TCR), | 
 | 				     MAC_TCR_SS_POS, MAC_TCR_SS_LEN); | 
 | 	if (regval == 0x2) | 
 | 		return 0; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_TCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_SS_POS, | 
 | 				     MAC_TCR_SS_LEN, 0x2); | 
 | 	writel(regval, pdata->mac_regs + MAC_TCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_set_xlgmii_100000_speed(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + MAC_TCR), | 
 | 				     MAC_TCR_SS_POS, MAC_TCR_SS_LEN); | 
 | 	if (regval == 0x3) | 
 | 		return 0; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + MAC_TCR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, MAC_TCR_SS_POS, | 
 | 				     MAC_TCR_SS_LEN, 0x3); | 
 | 	writel(regval, pdata->mac_regs + MAC_TCR); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_config_mac_speed(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	switch (pdata->phy_speed) { | 
 | 	case SPEED_100000: | 
 | 		xlgmac_set_xlgmii_100000_speed(pdata); | 
 | 		break; | 
 |  | 
 | 	case SPEED_50000: | 
 | 		xlgmac_set_xlgmii_50000_speed(pdata); | 
 | 		break; | 
 |  | 
 | 	case SPEED_40000: | 
 | 		xlgmac_set_xlgmii_40000_speed(pdata); | 
 | 		break; | 
 |  | 
 | 	case SPEED_25000: | 
 | 		xlgmac_set_xlgmii_25000_speed(pdata); | 
 | 		break; | 
 | 	} | 
 | } | 
 |  | 
 | static int xlgmac_dev_read(struct xlgmac_channel *channel) | 
 | { | 
 | 	struct xlgmac_pdata *pdata = channel->pdata; | 
 | 	struct xlgmac_ring *ring = channel->rx_ring; | 
 | 	struct net_device *netdev = pdata->netdev; | 
 | 	struct xlgmac_desc_data *desc_data; | 
 | 	struct xlgmac_dma_desc *dma_desc; | 
 | 	struct xlgmac_pkt_info *pkt_info; | 
 | 	unsigned int err, etlt, l34t; | 
 |  | 
 | 	desc_data = XLGMAC_GET_DESC_DATA(ring, ring->cur); | 
 | 	dma_desc = desc_data->dma_desc; | 
 | 	pkt_info = &ring->pkt_info; | 
 |  | 
 | 	/* Check for data availability */ | 
 | 	if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				   RX_NORMAL_DESC3_OWN_POS, | 
 | 				   RX_NORMAL_DESC3_OWN_LEN)) | 
 | 		return 1; | 
 |  | 
 | 	/* Make sure descriptor fields are read after reading the OWN bit */ | 
 | 	dma_rmb(); | 
 |  | 
 | 	if (netif_msg_rx_status(pdata)) | 
 | 		xlgmac_dump_rx_desc(pdata, ring, ring->cur); | 
 |  | 
 | 	if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				   RX_NORMAL_DESC3_CTXT_POS, | 
 | 				   RX_NORMAL_DESC3_CTXT_LEN)) { | 
 | 		/* Timestamp Context Descriptor */ | 
 | 		xlgmac_get_rx_tstamp(pkt_info, dma_desc); | 
 |  | 
 | 		pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 					pkt_info->attributes, | 
 | 					RX_PACKET_ATTRIBUTES_CONTEXT_POS, | 
 | 					RX_PACKET_ATTRIBUTES_CONTEXT_LEN, | 
 | 					1); | 
 | 		pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 				pkt_info->attributes, | 
 | 				RX_PACKET_ATTRIBUTES_CONTEXT_NEXT_POS, | 
 | 				RX_PACKET_ATTRIBUTES_CONTEXT_NEXT_LEN, | 
 | 				0); | 
 | 		return 0; | 
 | 	} | 
 |  | 
 | 	/* Normal Descriptor, be sure Context Descriptor bit is off */ | 
 | 	pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 				pkt_info->attributes, | 
 | 				RX_PACKET_ATTRIBUTES_CONTEXT_POS, | 
 | 				RX_PACKET_ATTRIBUTES_CONTEXT_LEN, | 
 | 				0); | 
 |  | 
 | 	/* Indicate if a Context Descriptor is next */ | 
 | 	if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				   RX_NORMAL_DESC3_CDA_POS, | 
 | 				   RX_NORMAL_DESC3_CDA_LEN)) | 
 | 		pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 				pkt_info->attributes, | 
 | 				RX_PACKET_ATTRIBUTES_CONTEXT_NEXT_POS, | 
 | 				RX_PACKET_ATTRIBUTES_CONTEXT_NEXT_LEN, | 
 | 				1); | 
 |  | 
 | 	/* Get the header length */ | 
 | 	if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				   RX_NORMAL_DESC3_FD_POS, | 
 | 				   RX_NORMAL_DESC3_FD_LEN)) { | 
 | 		desc_data->rx.hdr_len = XLGMAC_GET_REG_BITS_LE(dma_desc->desc2, | 
 | 							RX_NORMAL_DESC2_HL_POS, | 
 | 							RX_NORMAL_DESC2_HL_LEN); | 
 | 		if (desc_data->rx.hdr_len) | 
 | 			pdata->stats.rx_split_header_packets++; | 
 | 	} | 
 |  | 
 | 	/* Get the RSS hash */ | 
 | 	if (XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				   RX_NORMAL_DESC3_RSV_POS, | 
 | 				   RX_NORMAL_DESC3_RSV_LEN)) { | 
 | 		pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 				pkt_info->attributes, | 
 | 				RX_PACKET_ATTRIBUTES_RSS_HASH_POS, | 
 | 				RX_PACKET_ATTRIBUTES_RSS_HASH_LEN, | 
 | 				1); | 
 |  | 
 | 		pkt_info->rss_hash = le32_to_cpu(dma_desc->desc1); | 
 |  | 
 | 		l34t = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 					      RX_NORMAL_DESC3_L34T_POS, | 
 | 					  RX_NORMAL_DESC3_L34T_LEN); | 
 | 		switch (l34t) { | 
 | 		case RX_DESC3_L34T_IPV4_TCP: | 
 | 		case RX_DESC3_L34T_IPV4_UDP: | 
 | 		case RX_DESC3_L34T_IPV6_TCP: | 
 | 		case RX_DESC3_L34T_IPV6_UDP: | 
 | 			pkt_info->rss_hash_type = PKT_HASH_TYPE_L4; | 
 | 			break; | 
 | 		default: | 
 | 			pkt_info->rss_hash_type = PKT_HASH_TYPE_L3; | 
 | 		} | 
 | 	} | 
 |  | 
 | 	/* Get the pkt_info length */ | 
 | 	desc_data->rx.len = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 					RX_NORMAL_DESC3_PL_POS, | 
 | 					RX_NORMAL_DESC3_PL_LEN); | 
 |  | 
 | 	if (!XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				    RX_NORMAL_DESC3_LD_POS, | 
 | 				    RX_NORMAL_DESC3_LD_LEN)) { | 
 | 		/* Not all the data has been transferred for this pkt_info */ | 
 | 		pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 				pkt_info->attributes, | 
 | 				RX_PACKET_ATTRIBUTES_INCOMPLETE_POS, | 
 | 				RX_PACKET_ATTRIBUTES_INCOMPLETE_LEN, | 
 | 				1); | 
 | 		return 0; | 
 | 	} | 
 |  | 
 | 	/* This is the last of the data for this pkt_info */ | 
 | 	pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 			pkt_info->attributes, | 
 | 			RX_PACKET_ATTRIBUTES_INCOMPLETE_POS, | 
 | 			RX_PACKET_ATTRIBUTES_INCOMPLETE_LEN, | 
 | 			0); | 
 |  | 
 | 	/* Set checksum done indicator as appropriate */ | 
 | 	if (netdev->features & NETIF_F_RXCSUM) | 
 | 		pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 				pkt_info->attributes, | 
 | 				RX_PACKET_ATTRIBUTES_CSUM_DONE_POS, | 
 | 				RX_PACKET_ATTRIBUTES_CSUM_DONE_LEN, | 
 | 				1); | 
 |  | 
 | 	/* Check for errors (only valid in last descriptor) */ | 
 | 	err = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				     RX_NORMAL_DESC3_ES_POS, | 
 | 				     RX_NORMAL_DESC3_ES_LEN); | 
 | 	etlt = XLGMAC_GET_REG_BITS_LE(dma_desc->desc3, | 
 | 				      RX_NORMAL_DESC3_ETLT_POS, | 
 | 				      RX_NORMAL_DESC3_ETLT_LEN); | 
 | 	netif_dbg(pdata, rx_status, netdev, "err=%u, etlt=%#x\n", err, etlt); | 
 |  | 
 | 	if (!err || !etlt) { | 
 | 		/* No error if err is 0 or etlt is 0 */ | 
 | 		if ((etlt == 0x09) && | 
 | 		    (netdev->features & NETIF_F_HW_VLAN_CTAG_RX)) { | 
 | 			pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 					pkt_info->attributes, | 
 | 					RX_PACKET_ATTRIBUTES_VLAN_CTAG_POS, | 
 | 					RX_PACKET_ATTRIBUTES_VLAN_CTAG_LEN, | 
 | 					1); | 
 | 			pkt_info->vlan_ctag = | 
 | 				XLGMAC_GET_REG_BITS_LE(dma_desc->desc0, | 
 | 						       RX_NORMAL_DESC0_OVT_POS, | 
 | 						   RX_NORMAL_DESC0_OVT_LEN); | 
 | 			netif_dbg(pdata, rx_status, netdev, "vlan-ctag=%#06x\n", | 
 | 				  pkt_info->vlan_ctag); | 
 | 		} | 
 | 	} else { | 
 | 		if ((etlt == 0x05) || (etlt == 0x06)) | 
 | 			pkt_info->attributes = XLGMAC_SET_REG_BITS( | 
 | 					pkt_info->attributes, | 
 | 					RX_PACKET_ATTRIBUTES_CSUM_DONE_POS, | 
 | 					RX_PACKET_ATTRIBUTES_CSUM_DONE_LEN, | 
 | 					0); | 
 | 		else | 
 | 			pkt_info->errors = XLGMAC_SET_REG_BITS( | 
 | 					pkt_info->errors, | 
 | 					RX_PACKET_ERRORS_FRAME_POS, | 
 | 					RX_PACKET_ERRORS_FRAME_LEN, | 
 | 					1); | 
 | 	} | 
 |  | 
 | 	XLGMAC_PR("%s - descriptor=%u (cur=%d)\n", channel->name, | 
 | 		  ring->cur & (ring->dma_desc_count - 1), ring->cur); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_enable_int(struct xlgmac_channel *channel, | 
 | 			     enum xlgmac_int int_id) | 
 | { | 
 | 	unsigned int dma_ch_ier; | 
 |  | 
 | 	dma_ch_ier = readl(XLGMAC_DMA_REG(channel, DMA_CH_IER)); | 
 |  | 
 | 	switch (int_id) { | 
 | 	case XLGMAC_INT_DMA_CH_SR_TI: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_TIE_POS, | 
 | 				DMA_CH_IER_TIE_LEN, 1); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_TPS: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_TXSE_POS, | 
 | 				DMA_CH_IER_TXSE_LEN, 1); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_TBU: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_TBUE_POS, | 
 | 				DMA_CH_IER_TBUE_LEN, 1); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_RI: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_RIE_POS, | 
 | 				DMA_CH_IER_RIE_LEN, 1); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_RBU: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_RBUE_POS, | 
 | 				DMA_CH_IER_RBUE_LEN, 1); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_RPS: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_RSE_POS, | 
 | 				DMA_CH_IER_RSE_LEN, 1); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_TI_RI: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_TIE_POS, | 
 | 				DMA_CH_IER_TIE_LEN, 1); | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_RIE_POS, | 
 | 				DMA_CH_IER_RIE_LEN, 1); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_FBE: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_FBEE_POS, | 
 | 				DMA_CH_IER_FBEE_LEN, 1); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_ALL: | 
 | 		dma_ch_ier |= channel->saved_ier; | 
 | 		break; | 
 | 	default: | 
 | 		return -1; | 
 | 	} | 
 |  | 
 | 	writel(dma_ch_ier, XLGMAC_DMA_REG(channel, DMA_CH_IER)); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_disable_int(struct xlgmac_channel *channel, | 
 | 			      enum xlgmac_int int_id) | 
 | { | 
 | 	unsigned int dma_ch_ier; | 
 |  | 
 | 	dma_ch_ier = readl(XLGMAC_DMA_REG(channel, DMA_CH_IER)); | 
 |  | 
 | 	switch (int_id) { | 
 | 	case XLGMAC_INT_DMA_CH_SR_TI: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_TIE_POS, | 
 | 				DMA_CH_IER_TIE_LEN, 0); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_TPS: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_TXSE_POS, | 
 | 				DMA_CH_IER_TXSE_LEN, 0); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_TBU: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_TBUE_POS, | 
 | 				DMA_CH_IER_TBUE_LEN, 0); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_RI: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_RIE_POS, | 
 | 				DMA_CH_IER_RIE_LEN, 0); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_RBU: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_RBUE_POS, | 
 | 				DMA_CH_IER_RBUE_LEN, 0); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_RPS: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_RSE_POS, | 
 | 				DMA_CH_IER_RSE_LEN, 0); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_TI_RI: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_TIE_POS, | 
 | 				DMA_CH_IER_TIE_LEN, 0); | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_RIE_POS, | 
 | 				DMA_CH_IER_RIE_LEN, 0); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_CH_SR_FBE: | 
 | 		dma_ch_ier = XLGMAC_SET_REG_BITS( | 
 | 				dma_ch_ier, DMA_CH_IER_FBEE_POS, | 
 | 				DMA_CH_IER_FBEE_LEN, 0); | 
 | 		break; | 
 | 	case XLGMAC_INT_DMA_ALL: | 
 | 		channel->saved_ier = dma_ch_ier & XLGMAC_DMA_INTERRUPT_MASK; | 
 | 		dma_ch_ier &= ~XLGMAC_DMA_INTERRUPT_MASK; | 
 | 		break; | 
 | 	default: | 
 | 		return -1; | 
 | 	} | 
 |  | 
 | 	writel(dma_ch_ier, XLGMAC_DMA_REG(channel, DMA_CH_IER)); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_flush_tx_queues(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int i, count; | 
 | 	u32 regval; | 
 |  | 
 | 	for (i = 0; i < pdata->tx_q_count; i++) { | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 		regval = XLGMAC_SET_REG_BITS(regval, MTL_Q_TQOMR_FTQ_POS, | 
 | 					     MTL_Q_TQOMR_FTQ_LEN, 1); | 
 | 		writel(regval, XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 	} | 
 |  | 
 | 	/* Poll Until Poll Condition */ | 
 | 	for (i = 0; i < pdata->tx_q_count; i++) { | 
 | 		count = 2000; | 
 | 		regval = readl(XLGMAC_MTL_REG(pdata, i, MTL_Q_TQOMR)); | 
 | 		regval = XLGMAC_GET_REG_BITS(regval, MTL_Q_TQOMR_FTQ_POS, | 
 | 					     MTL_Q_TQOMR_FTQ_LEN); | 
 | 		while (--count && regval) | 
 | 			usleep_range(500, 600); | 
 |  | 
 | 		if (!count) | 
 | 			return -EBUSY; | 
 | 	} | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static void xlgmac_config_dma_bus(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	u32 regval; | 
 |  | 
 | 	regval = readl(pdata->mac_regs + DMA_SBMR); | 
 | 	/* Set enhanced addressing mode */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, DMA_SBMR_EAME_POS, | 
 | 				     DMA_SBMR_EAME_LEN, 1); | 
 | 	/* Set the System Bus mode */ | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, DMA_SBMR_UNDEF_POS, | 
 | 				     DMA_SBMR_UNDEF_LEN, 1); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, DMA_SBMR_BLEN_256_POS, | 
 | 				     DMA_SBMR_BLEN_256_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + DMA_SBMR); | 
 | } | 
 |  | 
 | static int xlgmac_hw_init(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	struct xlgmac_desc_ops *desc_ops = &pdata->desc_ops; | 
 | 	int ret; | 
 |  | 
 | 	/* Flush Tx queues */ | 
 | 	ret = xlgmac_flush_tx_queues(pdata); | 
 | 	if (ret) | 
 | 		return ret; | 
 |  | 
 | 	/* Initialize DMA related features */ | 
 | 	xlgmac_config_dma_bus(pdata); | 
 | 	xlgmac_config_osp_mode(pdata); | 
 | 	xlgmac_config_pblx8(pdata); | 
 | 	xlgmac_config_tx_pbl_val(pdata); | 
 | 	xlgmac_config_rx_pbl_val(pdata); | 
 | 	xlgmac_config_rx_coalesce(pdata); | 
 | 	xlgmac_config_tx_coalesce(pdata); | 
 | 	xlgmac_config_rx_buffer_size(pdata); | 
 | 	xlgmac_config_tso_mode(pdata); | 
 | 	xlgmac_config_sph_mode(pdata); | 
 | 	xlgmac_config_rss(pdata); | 
 | 	desc_ops->tx_desc_init(pdata); | 
 | 	desc_ops->rx_desc_init(pdata); | 
 | 	xlgmac_enable_dma_interrupts(pdata); | 
 |  | 
 | 	/* Initialize MTL related features */ | 
 | 	xlgmac_config_mtl_mode(pdata); | 
 | 	xlgmac_config_queue_mapping(pdata); | 
 | 	xlgmac_config_tsf_mode(pdata, pdata->tx_sf_mode); | 
 | 	xlgmac_config_rsf_mode(pdata, pdata->rx_sf_mode); | 
 | 	xlgmac_config_tx_threshold(pdata, pdata->tx_threshold); | 
 | 	xlgmac_config_rx_threshold(pdata, pdata->rx_threshold); | 
 | 	xlgmac_config_tx_fifo_size(pdata); | 
 | 	xlgmac_config_rx_fifo_size(pdata); | 
 | 	xlgmac_config_flow_control_threshold(pdata); | 
 | 	xlgmac_config_rx_fep_enable(pdata); | 
 | 	xlgmac_config_rx_fup_enable(pdata); | 
 | 	xlgmac_enable_mtl_interrupts(pdata); | 
 |  | 
 | 	/* Initialize MAC related features */ | 
 | 	xlgmac_config_mac_address(pdata); | 
 | 	xlgmac_config_rx_mode(pdata); | 
 | 	xlgmac_config_jumbo_enable(pdata); | 
 | 	xlgmac_config_flow_control(pdata); | 
 | 	xlgmac_config_mac_speed(pdata); | 
 | 	xlgmac_config_checksum_offload(pdata); | 
 | 	xlgmac_config_vlan_support(pdata); | 
 | 	xlgmac_config_mmc(pdata); | 
 | 	xlgmac_enable_mac_interrupts(pdata); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int xlgmac_hw_exit(struct xlgmac_pdata *pdata) | 
 | { | 
 | 	unsigned int count = 2000; | 
 | 	u32 regval; | 
 |  | 
 | 	/* Issue a software reset */ | 
 | 	regval = readl(pdata->mac_regs + DMA_MR); | 
 | 	regval = XLGMAC_SET_REG_BITS(regval, DMA_MR_SWR_POS, | 
 | 				     DMA_MR_SWR_LEN, 1); | 
 | 	writel(regval, pdata->mac_regs + DMA_MR); | 
 | 	usleep_range(10, 15); | 
 |  | 
 | 	/* Poll Until Poll Condition */ | 
 | 	while (--count && | 
 | 	       XLGMAC_GET_REG_BITS(readl(pdata->mac_regs + DMA_MR), | 
 | 				   DMA_MR_SWR_POS, DMA_MR_SWR_LEN)) | 
 | 		usleep_range(500, 600); | 
 |  | 
 | 	if (!count) | 
 | 		return -EBUSY; | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | void xlgmac_init_hw_ops(struct xlgmac_hw_ops *hw_ops) | 
 | { | 
 | 	hw_ops->init = xlgmac_hw_init; | 
 | 	hw_ops->exit = xlgmac_hw_exit; | 
 |  | 
 | 	hw_ops->tx_complete = xlgmac_tx_complete; | 
 |  | 
 | 	hw_ops->enable_tx = xlgmac_enable_tx; | 
 | 	hw_ops->disable_tx = xlgmac_disable_tx; | 
 | 	hw_ops->enable_rx = xlgmac_enable_rx; | 
 | 	hw_ops->disable_rx = xlgmac_disable_rx; | 
 |  | 
 | 	hw_ops->dev_xmit = xlgmac_dev_xmit; | 
 | 	hw_ops->dev_read = xlgmac_dev_read; | 
 | 	hw_ops->enable_int = xlgmac_enable_int; | 
 | 	hw_ops->disable_int = xlgmac_disable_int; | 
 |  | 
 | 	hw_ops->set_mac_address = xlgmac_set_mac_address; | 
 | 	hw_ops->config_rx_mode = xlgmac_config_rx_mode; | 
 | 	hw_ops->enable_rx_csum = xlgmac_enable_rx_csum; | 
 | 	hw_ops->disable_rx_csum = xlgmac_disable_rx_csum; | 
 |  | 
 | 	/* For MII speed configuration */ | 
 | 	hw_ops->set_xlgmii_25000_speed = xlgmac_set_xlgmii_25000_speed; | 
 | 	hw_ops->set_xlgmii_40000_speed = xlgmac_set_xlgmii_40000_speed; | 
 | 	hw_ops->set_xlgmii_50000_speed = xlgmac_set_xlgmii_50000_speed; | 
 | 	hw_ops->set_xlgmii_100000_speed = xlgmac_set_xlgmii_100000_speed; | 
 |  | 
 | 	/* For descriptor related operation */ | 
 | 	hw_ops->tx_desc_init = xlgmac_tx_desc_init; | 
 | 	hw_ops->rx_desc_init = xlgmac_rx_desc_init; | 
 | 	hw_ops->tx_desc_reset = xlgmac_tx_desc_reset; | 
 | 	hw_ops->rx_desc_reset = xlgmac_rx_desc_reset; | 
 | 	hw_ops->is_last_desc = xlgmac_is_last_desc; | 
 | 	hw_ops->is_context_desc = xlgmac_is_context_desc; | 
 | 	hw_ops->tx_start_xmit = xlgmac_tx_start_xmit; | 
 |  | 
 | 	/* For Flow Control */ | 
 | 	hw_ops->config_tx_flow_control = xlgmac_config_tx_flow_control; | 
 | 	hw_ops->config_rx_flow_control = xlgmac_config_rx_flow_control; | 
 |  | 
 | 	/* For Vlan related config */ | 
 | 	hw_ops->enable_rx_vlan_stripping = xlgmac_enable_rx_vlan_stripping; | 
 | 	hw_ops->disable_rx_vlan_stripping = xlgmac_disable_rx_vlan_stripping; | 
 | 	hw_ops->enable_rx_vlan_filtering = xlgmac_enable_rx_vlan_filtering; | 
 | 	hw_ops->disable_rx_vlan_filtering = xlgmac_disable_rx_vlan_filtering; | 
 | 	hw_ops->update_vlan_hash_table = xlgmac_update_vlan_hash_table; | 
 |  | 
 | 	/* For RX coalescing */ | 
 | 	hw_ops->config_rx_coalesce = xlgmac_config_rx_coalesce; | 
 | 	hw_ops->config_tx_coalesce = xlgmac_config_tx_coalesce; | 
 | 	hw_ops->usec_to_riwt = xlgmac_usec_to_riwt; | 
 | 	hw_ops->riwt_to_usec = xlgmac_riwt_to_usec; | 
 |  | 
 | 	/* For RX and TX threshold config */ | 
 | 	hw_ops->config_rx_threshold = xlgmac_config_rx_threshold; | 
 | 	hw_ops->config_tx_threshold = xlgmac_config_tx_threshold; | 
 |  | 
 | 	/* For RX and TX Store and Forward Mode config */ | 
 | 	hw_ops->config_rsf_mode = xlgmac_config_rsf_mode; | 
 | 	hw_ops->config_tsf_mode = xlgmac_config_tsf_mode; | 
 |  | 
 | 	/* For TX DMA Operating on Second Frame config */ | 
 | 	hw_ops->config_osp_mode = xlgmac_config_osp_mode; | 
 |  | 
 | 	/* For RX and TX PBL config */ | 
 | 	hw_ops->config_rx_pbl_val = xlgmac_config_rx_pbl_val; | 
 | 	hw_ops->get_rx_pbl_val = xlgmac_get_rx_pbl_val; | 
 | 	hw_ops->config_tx_pbl_val = xlgmac_config_tx_pbl_val; | 
 | 	hw_ops->get_tx_pbl_val = xlgmac_get_tx_pbl_val; | 
 | 	hw_ops->config_pblx8 = xlgmac_config_pblx8; | 
 |  | 
 | 	/* For MMC statistics support */ | 
 | 	hw_ops->tx_mmc_int = xlgmac_tx_mmc_int; | 
 | 	hw_ops->rx_mmc_int = xlgmac_rx_mmc_int; | 
 | 	hw_ops->read_mmc_stats = xlgmac_read_mmc_stats; | 
 |  | 
 | 	/* For Receive Side Scaling */ | 
 | 	hw_ops->enable_rss = xlgmac_enable_rss; | 
 | 	hw_ops->disable_rss = xlgmac_disable_rss; | 
 | 	hw_ops->set_rss_hash_key = xlgmac_set_rss_hash_key; | 
 | 	hw_ops->set_rss_lookup_table = xlgmac_set_rss_lookup_table; | 
 | } |