ixgbe: fix setting of promisc mode when using mac-vlans
[safe/jmp/linux-2.6] / drivers / net / ixgbe / ixgbe_main.c
index f098816..15032c7 100644 (file)
@@ -36,6 +36,7 @@
 #include <linux/tcp.h>
 #include <linux/pkt_sched.h>
 #include <linux/ipv6.h>
+#include <linux/slab.h>
 #include <net/checksum.h>
 #include <net/ip6_checksum.h>
 #include <linux/ethtool.h>
@@ -51,7 +52,7 @@ char ixgbe_driver_name[] = "ixgbe";
 static const char ixgbe_driver_string[] =
                               "Intel(R) 10 Gigabit PCI Express Network Driver";
 
-#define DRV_VERSION "2.0.44-k2"
+#define DRV_VERSION "2.0.62-k2"
 const char ixgbe_driver_version[] = DRV_VERSION;
 static char ixgbe_copyright[] = "Copyright (c) 1999-2010 Intel Corporation.";
 
@@ -174,6 +175,345 @@ static inline void ixgbe_disable_sriov(struct ixgbe_adapter *adapter)
        adapter->flags &= ~IXGBE_FLAG_SRIOV_ENABLED;
 }
 
+struct ixgbe_reg_info {
+       u32 ofs;
+       char *name;
+};
+
+static const struct ixgbe_reg_info ixgbe_reg_info_tbl[] = {
+
+       /* General Registers */
+       {IXGBE_CTRL, "CTRL"},
+       {IXGBE_STATUS, "STATUS"},
+       {IXGBE_CTRL_EXT, "CTRL_EXT"},
+
+       /* Interrupt Registers */
+       {IXGBE_EICR, "EICR"},
+
+       /* RX Registers */
+       {IXGBE_SRRCTL(0), "SRRCTL"},
+       {IXGBE_DCA_RXCTRL(0), "DRXCTL"},
+       {IXGBE_RDLEN(0), "RDLEN"},
+       {IXGBE_RDH(0), "RDH"},
+       {IXGBE_RDT(0), "RDT"},
+       {IXGBE_RXDCTL(0), "RXDCTL"},
+       {IXGBE_RDBAL(0), "RDBAL"},
+       {IXGBE_RDBAH(0), "RDBAH"},
+
+       /* TX Registers */
+       {IXGBE_TDBAL(0), "TDBAL"},
+       {IXGBE_TDBAH(0), "TDBAH"},
+       {IXGBE_TDLEN(0), "TDLEN"},
+       {IXGBE_TDH(0), "TDH"},
+       {IXGBE_TDT(0), "TDT"},
+       {IXGBE_TXDCTL(0), "TXDCTL"},
+
+       /* List Terminator */
+       {}
+};
+
+
+/*
+ * ixgbe_regdump - register printout routine
+ */
+static void ixgbe_regdump(struct ixgbe_hw *hw, struct ixgbe_reg_info *reginfo)
+{
+       int i = 0, j = 0;
+       char rname[16];
+       u32 regs[64];
+
+       switch (reginfo->ofs) {
+       case IXGBE_SRRCTL(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_SRRCTL(i));
+               break;
+       case IXGBE_DCA_RXCTRL(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_DCA_RXCTRL(i));
+               break;
+       case IXGBE_RDLEN(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_RDLEN(i));
+               break;
+       case IXGBE_RDH(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_RDH(i));
+               break;
+       case IXGBE_RDT(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_RDT(i));
+               break;
+       case IXGBE_RXDCTL(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_RXDCTL(i));
+               break;
+       case IXGBE_RDBAL(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_RDBAL(i));
+               break;
+       case IXGBE_RDBAH(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_RDBAH(i));
+               break;
+       case IXGBE_TDBAL(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_TDBAL(i));
+               break;
+       case IXGBE_TDBAH(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_TDBAH(i));
+               break;
+       case IXGBE_TDLEN(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_TDLEN(i));
+               break;
+       case IXGBE_TDH(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_TDH(i));
+               break;
+       case IXGBE_TDT(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_TDT(i));
+               break;
+       case IXGBE_TXDCTL(0):
+               for (i = 0; i < 64; i++)
+                       regs[i] = IXGBE_READ_REG(hw, IXGBE_TXDCTL(i));
+               break;
+       default:
+               printk(KERN_INFO "%-15s %08x\n", reginfo->name,
+                       IXGBE_READ_REG(hw, reginfo->ofs));
+               return;
+       }
+
+       for (i = 0; i < 8; i++) {
+               snprintf(rname, 16, "%s[%d-%d]", reginfo->name, i*8, i*8+7);
+               printk(KERN_ERR "%-15s ", rname);
+               for (j = 0; j < 8; j++)
+                       printk(KERN_CONT "%08x ", regs[i*8+j]);
+               printk(KERN_CONT "\n");
+       }
+
+}
+
+/*
+ * ixgbe_dump - Print registers, tx-rings and rx-rings
+ */
+static void ixgbe_dump(struct ixgbe_adapter *adapter)
+{
+       struct net_device *netdev = adapter->netdev;
+       struct ixgbe_hw *hw = &adapter->hw;
+       struct ixgbe_reg_info *reginfo;
+       int n = 0;
+       struct ixgbe_ring *tx_ring;
+       struct ixgbe_tx_buffer *tx_buffer_info;
+       union ixgbe_adv_tx_desc *tx_desc;
+       struct my_u0 { u64 a; u64 b; } *u0;
+       struct ixgbe_ring *rx_ring;
+       union ixgbe_adv_rx_desc *rx_desc;
+       struct ixgbe_rx_buffer *rx_buffer_info;
+       u32 staterr;
+       int i = 0;
+
+       if (!netif_msg_hw(adapter))
+               return;
+
+       /* Print netdevice Info */
+       if (netdev) {
+               dev_info(&adapter->pdev->dev, "Net device Info\n");
+               printk(KERN_INFO "Device Name     state            "
+                       "trans_start      last_rx\n");
+               printk(KERN_INFO "%-15s %016lX %016lX %016lX\n",
+               netdev->name,
+               netdev->state,
+               netdev->trans_start,
+               netdev->last_rx);
+       }
+
+       /* Print Registers */
+       dev_info(&adapter->pdev->dev, "Register Dump\n");
+       printk(KERN_INFO " Register Name   Value\n");
+       for (reginfo = (struct ixgbe_reg_info *)ixgbe_reg_info_tbl;
+            reginfo->name; reginfo++) {
+               ixgbe_regdump(hw, reginfo);
+       }
+
+       /* Print TX Ring Summary */
+       if (!netdev || !netif_running(netdev))
+               goto exit;
+
+       dev_info(&adapter->pdev->dev, "TX Rings Summary\n");
+       printk(KERN_INFO "Queue [NTU] [NTC] [bi(ntc)->dma  ] "
+               "leng ntw timestamp\n");
+       for (n = 0; n < adapter->num_tx_queues; n++) {
+               tx_ring = adapter->tx_ring[n];
+               tx_buffer_info =
+                       &tx_ring->tx_buffer_info[tx_ring->next_to_clean];
+               printk(KERN_INFO " %5d %5X %5X %016llX %04X %3X %016llX\n",
+                          n, tx_ring->next_to_use, tx_ring->next_to_clean,
+                          (u64)tx_buffer_info->dma,
+                          tx_buffer_info->length,
+                          tx_buffer_info->next_to_watch,
+                          (u64)tx_buffer_info->time_stamp);
+       }
+
+       /* Print TX Rings */
+       if (!netif_msg_tx_done(adapter))
+               goto rx_ring_summary;
+
+       dev_info(&adapter->pdev->dev, "TX Rings Dump\n");
+
+       /* Transmit Descriptor Formats
+        *
+        * Advanced Transmit Descriptor
+        *   +--------------------------------------------------------------+
+        * 0 |         Buffer Address [63:0]                                |
+        *   +--------------------------------------------------------------+
+        * 8 |  PAYLEN  | PORTS  | IDX | STA | DCMD  |DTYP |  RSV |  DTALEN |
+        *   +--------------------------------------------------------------+
+        *   63       46 45    40 39 36 35 32 31   24 23 20 19              0
+        */
+
+       for (n = 0; n < adapter->num_tx_queues; n++) {
+               tx_ring = adapter->tx_ring[n];
+               printk(KERN_INFO "------------------------------------\n");
+               printk(KERN_INFO "TX QUEUE INDEX = %d\n", tx_ring->queue_index);
+               printk(KERN_INFO "------------------------------------\n");
+               printk(KERN_INFO "T [desc]     [address 63:0  ] "
+                       "[PlPOIdStDDt Ln] [bi->dma       ] "
+                       "leng  ntw timestamp        bi->skb\n");
+
+               for (i = 0; tx_ring->desc && (i < tx_ring->count); i++) {
+                       tx_desc = IXGBE_TX_DESC_ADV(*tx_ring, i);
+                       tx_buffer_info = &tx_ring->tx_buffer_info[i];
+                       u0 = (struct my_u0 *)tx_desc;
+                       printk(KERN_INFO "T [0x%03X]    %016llX %016llX %016llX"
+                               " %04X  %3X %016llX %p", i,
+                               le64_to_cpu(u0->a),
+                               le64_to_cpu(u0->b),
+                               (u64)tx_buffer_info->dma,
+                               tx_buffer_info->length,
+                               tx_buffer_info->next_to_watch,
+                               (u64)tx_buffer_info->time_stamp,
+                               tx_buffer_info->skb);
+                       if (i == tx_ring->next_to_use &&
+                               i == tx_ring->next_to_clean)
+                               printk(KERN_CONT " NTC/U\n");
+                       else if (i == tx_ring->next_to_use)
+                               printk(KERN_CONT " NTU\n");
+                       else if (i == tx_ring->next_to_clean)
+                               printk(KERN_CONT " NTC\n");
+                       else
+                               printk(KERN_CONT "\n");
+
+                       if (netif_msg_pktdata(adapter) &&
+                               tx_buffer_info->dma != 0)
+                               print_hex_dump(KERN_INFO, "",
+                                       DUMP_PREFIX_ADDRESS, 16, 1,
+                                       phys_to_virt(tx_buffer_info->dma),
+                                       tx_buffer_info->length, true);
+               }
+       }
+
+       /* Print RX Rings Summary */
+rx_ring_summary:
+       dev_info(&adapter->pdev->dev, "RX Rings Summary\n");
+       printk(KERN_INFO "Queue [NTU] [NTC]\n");
+       for (n = 0; n < adapter->num_rx_queues; n++) {
+               rx_ring = adapter->rx_ring[n];
+               printk(KERN_INFO "%5d %5X %5X\n", n,
+                          rx_ring->next_to_use, rx_ring->next_to_clean);
+       }
+
+       /* Print RX Rings */
+       if (!netif_msg_rx_status(adapter))
+               goto exit;
+
+       dev_info(&adapter->pdev->dev, "RX Rings Dump\n");
+
+       /* Advanced Receive Descriptor (Read) Format
+        *    63                                           1        0
+        *    +-----------------------------------------------------+
+        *  0 |       Packet Buffer Address [63:1]           |A0/NSE|
+        *    +----------------------------------------------+------+
+        *  8 |       Header Buffer Address [63:1]           |  DD  |
+        *    +-----------------------------------------------------+
+        *
+        *
+        * Advanced Receive Descriptor (Write-Back) Format
+        *
+        *   63       48 47    32 31  30      21 20 16 15   4 3     0
+        *   +------------------------------------------------------+
+        * 0 | Packet     IP     |SPH| HDR_LEN   | RSV|Packet|  RSS |
+        *   | Checksum   Ident  |   |           |    | Type | Type |
+        *   +------------------------------------------------------+
+        * 8 | VLAN Tag | Length | Extended Error | Extended Status |
+        *   +------------------------------------------------------+
+        *   63       48 47    32 31            20 19               0
+        */
+       for (n = 0; n < adapter->num_rx_queues; n++) {
+               rx_ring = adapter->rx_ring[n];
+               printk(KERN_INFO "------------------------------------\n");
+               printk(KERN_INFO "RX QUEUE INDEX = %d\n", rx_ring->queue_index);
+               printk(KERN_INFO "------------------------------------\n");
+               printk(KERN_INFO "R  [desc]      [ PktBuf     A0] "
+                       "[  HeadBuf   DD] [bi->dma       ] [bi->skb] "
+                       "<-- Adv Rx Read format\n");
+               printk(KERN_INFO "RWB[desc]      [PcsmIpSHl PtRs] "
+                       "[vl er S cks ln] ---------------- [bi->skb] "
+                       "<-- Adv Rx Write-Back format\n");
+
+               for (i = 0; i < rx_ring->count; i++) {
+                       rx_buffer_info = &rx_ring->rx_buffer_info[i];
+                       rx_desc = IXGBE_RX_DESC_ADV(*rx_ring, i);
+                       u0 = (struct my_u0 *)rx_desc;
+                       staterr = le32_to_cpu(rx_desc->wb.upper.status_error);
+                       if (staterr & IXGBE_RXD_STAT_DD) {
+                               /* Descriptor Done */
+                               printk(KERN_INFO "RWB[0x%03X]     %016llX "
+                                       "%016llX ---------------- %p", i,
+                                       le64_to_cpu(u0->a),
+                                       le64_to_cpu(u0->b),
+                                       rx_buffer_info->skb);
+                       } else {
+                               printk(KERN_INFO "R  [0x%03X]     %016llX "
+                                       "%016llX %016llX %p", i,
+                                       le64_to_cpu(u0->a),
+                                       le64_to_cpu(u0->b),
+                                       (u64)rx_buffer_info->dma,
+                                       rx_buffer_info->skb);
+
+                               if (netif_msg_pktdata(adapter)) {
+                                       print_hex_dump(KERN_INFO, "",
+                                          DUMP_PREFIX_ADDRESS, 16, 1,
+                                          phys_to_virt(rx_buffer_info->dma),
+                                          rx_ring->rx_buf_len, true);
+
+                                       if (rx_ring->rx_buf_len
+                                               < IXGBE_RXBUFFER_2048)
+                                               print_hex_dump(KERN_INFO, "",
+                                                 DUMP_PREFIX_ADDRESS, 16, 1,
+                                                 phys_to_virt(
+                                                   rx_buffer_info->page_dma +
+                                                   rx_buffer_info->page_offset
+                                                 ),
+                                                 PAGE_SIZE/2, true);
+                               }
+                       }
+
+                       if (i == rx_ring->next_to_use)
+                               printk(KERN_CONT " NTU\n");
+                       else if (i == rx_ring->next_to_clean)
+                               printk(KERN_CONT " NTC\n");
+                       else
+                               printk(KERN_CONT "\n");
+
+               }
+       }
+
+exit:
+       return;
+}
+
 static void ixgbe_release_hw_control(struct ixgbe_adapter *adapter)
 {
        u32 ctrl_ext;
@@ -265,15 +605,15 @@ static void ixgbe_unmap_and_free_tx_resource(struct ixgbe_adapter *adapter,
 {
        if (tx_buffer_info->dma) {
                if (tx_buffer_info->mapped_as_page)
-                       pci_unmap_page(adapter->pdev,
+                       dma_unmap_page(&adapter->pdev->dev,
                                       tx_buffer_info->dma,
                                       tx_buffer_info->length,
-                                      PCI_DMA_TODEVICE);
+                                      DMA_TO_DEVICE);
                else
-                       pci_unmap_single(adapter->pdev,
+                       dma_unmap_single(&adapter->pdev->dev,
                                         tx_buffer_info->dma,
                                         tx_buffer_info->length,
-                                        PCI_DMA_TODEVICE);
+                                        DMA_TO_DEVICE);
                tx_buffer_info->dma = 0;
        }
        if (tx_buffer_info->skb) {
@@ -494,7 +834,7 @@ static void ixgbe_update_rx_dca(struct ixgbe_adapter *adapter,
 {
        u32 rxctrl;
        int cpu = get_cpu();
-       int q = rx_ring - adapter->rx_ring;
+       int q = rx_ring->reg_idx;
 
        if (rx_ring->cpu != cpu) {
                rxctrl = IXGBE_READ_REG(&adapter->hw, IXGBE_DCA_RXCTRL(q));
@@ -522,7 +862,7 @@ static void ixgbe_update_tx_dca(struct ixgbe_adapter *adapter,
 {
        u32 txctrl;
        int cpu = get_cpu();
-       int q = tx_ring - adapter->tx_ring;
+       int q = tx_ring->reg_idx;
        struct ixgbe_hw *hw = &adapter->hw;
 
        if (tx_ring->cpu != cpu) {
@@ -556,12 +896,12 @@ static void ixgbe_setup_dca(struct ixgbe_adapter *adapter)
        IXGBE_WRITE_REG(&adapter->hw, IXGBE_DCA_CTRL, 2);
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               adapter->tx_ring[i].cpu = -1;
-               ixgbe_update_tx_dca(adapter, &adapter->tx_ring[i]);
+               adapter->tx_ring[i]->cpu = -1;
+               ixgbe_update_tx_dca(adapter, adapter->tx_ring[i]);
        }
        for (i = 0; i < adapter->num_rx_queues; i++) {
-               adapter->rx_ring[i].cpu = -1;
-               ixgbe_update_rx_dca(adapter, &adapter->rx_ring[i]);
+               adapter->rx_ring[i]->cpu = -1;
+               ixgbe_update_rx_dca(adapter, adapter->rx_ring[i]);
        }
 }
 
@@ -720,10 +1060,10 @@ static void ixgbe_alloc_rx_buffers(struct ixgbe_adapter *adapter,
                                bi->page_offset ^= (PAGE_SIZE / 2);
                        }
 
-                       bi->page_dma = pci_map_page(pdev, bi->page,
+                       bi->page_dma = dma_map_page(&pdev->dev, bi->page,
                                                    bi->page_offset,
                                                    (PAGE_SIZE / 2),
-                                                   PCI_DMA_FROMDEVICE);
+                                                   DMA_FROM_DEVICE);
                }
 
                if (!bi->skb) {
@@ -742,9 +1082,9 @@ static void ixgbe_alloc_rx_buffers(struct ixgbe_adapter *adapter,
                                          - skb->data));
 
                        bi->skb = skb;
-                       bi->dma = pci_map_single(pdev, skb->data,
+                       bi->dma = dma_map_single(&pdev->dev, skb->data,
                                                 rx_ring->rx_buf_len,
-                                                PCI_DMA_FROMDEVICE);
+                                                DMA_FROM_DEVICE);
                }
                /* Refresh the desc even if buffer_addrs didn't change because
                 * each write-back erases this info. */
@@ -818,6 +1158,12 @@ static inline struct sk_buff *ixgbe_transform_rsc_queue(struct sk_buff *skb,
        return skb;
 }
 
+struct ixgbe_rsc_cb {
+       dma_addr_t dma;
+};
+
+#define IXGBE_RSC_CB(skb) ((struct ixgbe_rsc_cb *)(skb)->cb)
+
 static bool ixgbe_clean_rx_irq(struct ixgbe_q_vector *q_vector,
                                struct ixgbe_ring *rx_ring,
                                int *work_done, int work_to_do)
@@ -849,6 +1195,7 @@ static bool ixgbe_clean_rx_irq(struct ixgbe_q_vector *q_vector,
                        break;
                (*work_done)++;
 
+               rmb(); /* read descriptor and rx_buffer_info after status DD */
                if (rx_ring->flags & IXGBE_RING_RX_PS_ENABLED) {
                        hdr_info = le16_to_cpu(ixgbe_get_hdr_info(rx_desc));
                        len = (hdr_info & IXGBE_RXDADV_HDRBUFLEN_MASK) >>
@@ -866,16 +1213,29 @@ static bool ixgbe_clean_rx_irq(struct ixgbe_q_vector *q_vector,
                rx_buffer_info->skb = NULL;
 
                if (rx_buffer_info->dma) {
-                       pci_unmap_single(pdev, rx_buffer_info->dma,
-                                        rx_ring->rx_buf_len,
-                                        PCI_DMA_FROMDEVICE);
+                       if ((adapter->flags2 & IXGBE_FLAG2_RSC_ENABLED) &&
+                           (!(staterr & IXGBE_RXD_STAT_EOP)) &&
+                                (!(skb->prev)))
+                               /*
+                                * When HWRSC is enabled, delay unmapping
+                                * of the first packet. It carries the
+                                * header information, HW may still
+                                * access the header after the writeback.
+                                * Only unmap it when EOP is reached
+                                */
+                               IXGBE_RSC_CB(skb)->dma = rx_buffer_info->dma;
+                       else
+                               dma_unmap_single(&pdev->dev,
+                                                rx_buffer_info->dma,
+                                                rx_ring->rx_buf_len,
+                                                DMA_FROM_DEVICE);
                        rx_buffer_info->dma = 0;
                        skb_put(skb, len);
                }
 
                if (upper_len) {
-                       pci_unmap_page(pdev, rx_buffer_info->page_dma,
-                                      PAGE_SIZE / 2, PCI_DMA_FROMDEVICE);
+                       dma_unmap_page(&pdev->dev, rx_buffer_info->page_dma,
+                                      PAGE_SIZE / 2, DMA_FROM_DEVICE);
                        rx_buffer_info->page_dma = 0;
                        skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags,
                                           rx_buffer_info->page,
@@ -916,6 +1276,13 @@ static bool ixgbe_clean_rx_irq(struct ixgbe_q_vector *q_vector,
                        if (skb->prev)
                                skb = ixgbe_transform_rsc_queue(skb, &(rx_ring->rsc_count));
                        if (adapter->flags2 & IXGBE_FLAG2_RSC_ENABLED) {
+                               if (IXGBE_RSC_CB(skb)->dma) {
+                                       dma_unmap_single(&pdev->dev,
+                                                        IXGBE_RSC_CB(skb)->dma,
+                                                        rx_ring->rx_buf_len,
+                                                        DMA_FROM_DEVICE);
+                                       IXGBE_RSC_CB(skb)->dma = 0;
+                               }
                                if (rx_ring->flags & IXGBE_RING_RX_PS_ENABLED)
                                        rx_ring->rsc_count += skb_shinfo(skb)->nr_frags;
                                else
@@ -1027,12 +1394,12 @@ static void ixgbe_configure_msix(struct ixgbe_adapter *adapter)
         */
        for (v_idx = 0; v_idx < q_vectors; v_idx++) {
                q_vector = adapter->q_vector[v_idx];
-               /* XXX for_each_bit(...) */
+               /* XXX for_each_set_bit(...) */
                r_idx = find_first_bit(q_vector->rxr_idx,
                                       adapter->num_rx_queues);
 
                for (i = 0; i < q_vector->rxr_count; i++) {
-                       j = adapter->rx_ring[r_idx].reg_idx;
+                       j = adapter->rx_ring[r_idx]->reg_idx;
                        ixgbe_set_ivar(adapter, 0, j, v_idx);
                        r_idx = find_next_bit(q_vector->rxr_idx,
                                              adapter->num_rx_queues,
@@ -1042,7 +1409,7 @@ static void ixgbe_configure_msix(struct ixgbe_adapter *adapter)
                                       adapter->num_tx_queues);
 
                for (i = 0; i < q_vector->txr_count; i++) {
-                       j = adapter->tx_ring[r_idx].reg_idx;
+                       j = adapter->tx_ring[r_idx]->reg_idx;
                        ixgbe_set_ivar(adapter, 1, j, v_idx);
                        r_idx = find_next_bit(q_vector->txr_idx,
                                              adapter->num_tx_queues,
@@ -1164,6 +1531,15 @@ void ixgbe_write_eitr(struct ixgbe_q_vector *q_vector)
                itr_reg |= (itr_reg << 16);
        } else if (adapter->hw.mac.type == ixgbe_mac_82599EB) {
                /*
+                * 82599 can support a value of zero, so allow it for
+                * max interrupt rate, but there is an errata where it can
+                * not be zero with RSC
+                */
+               if (itr_reg == 8 &&
+                   !(adapter->flags2 & IXGBE_FLAG2_RSC_ENABLED))
+                       itr_reg = 0;
+
+               /*
                 * set the WDIS bit to not clear the timer bits and cause an
                 * immediate assertion of the interrupt
                 */
@@ -1182,7 +1558,7 @@ static void ixgbe_set_itr_msix(struct ixgbe_q_vector *q_vector)
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
        for (i = 0; i < q_vector->txr_count; i++) {
-               tx_ring = &(adapter->tx_ring[r_idx]);
+               tx_ring = adapter->tx_ring[r_idx];
                ret_itr = ixgbe_update_itr(adapter, q_vector->eitr,
                                           q_vector->tx_itr,
                                           tx_ring->total_packets,
@@ -1197,7 +1573,7 @@ static void ixgbe_set_itr_msix(struct ixgbe_q_vector *q_vector)
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
        for (i = 0; i < q_vector->rxr_count; i++) {
-               rx_ring = &(adapter->rx_ring[r_idx]);
+               rx_ring = adapter->rx_ring[r_idx];
                ret_itr = ixgbe_update_itr(adapter, q_vector->eitr,
                                           q_vector->rx_itr,
                                           rx_ring->total_packets,
@@ -1319,7 +1695,7 @@ static irqreturn_t ixgbe_msix_lsc(int irq, void *data)
                        netif_tx_stop_all_queues(netdev);
                        for (i = 0; i < adapter->num_tx_queues; i++) {
                                struct ixgbe_ring *tx_ring =
-                                                          &adapter->tx_ring[i];
+                                                           adapter->tx_ring[i];
                                if (test_and_clear_bit(__IXGBE_FDIR_INIT_DONE,
                                                       &tx_ring->reinit_state))
                                        schedule_work(&adapter->fdir_reinit_task);
@@ -1378,7 +1754,7 @@ static irqreturn_t ixgbe_msix_clean_tx(int irq, void *data)
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
        for (i = 0; i < q_vector->txr_count; i++) {
-               tx_ring = &(adapter->tx_ring[r_idx]);
+               tx_ring = adapter->tx_ring[r_idx];
                tx_ring->total_bytes = 0;
                tx_ring->total_packets = 0;
                r_idx = find_next_bit(q_vector->txr_idx, adapter->num_tx_queues,
@@ -1406,7 +1782,7 @@ static irqreturn_t ixgbe_msix_clean_rx(int irq, void *data)
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
        for (i = 0;  i < q_vector->rxr_count; i++) {
-               rx_ring = &(adapter->rx_ring[r_idx]);
+               rx_ring = adapter->rx_ring[r_idx];
                rx_ring->total_bytes = 0;
                rx_ring->total_packets = 0;
                r_idx = find_next_bit(q_vector->rxr_idx, adapter->num_rx_queues,
@@ -1436,7 +1812,7 @@ static irqreturn_t ixgbe_msix_clean_many(int irq, void *data)
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
        for (i = 0; i < q_vector->txr_count; i++) {
-               ring = &(adapter->tx_ring[r_idx]);
+               ring = adapter->tx_ring[r_idx];
                ring->total_bytes = 0;
                ring->total_packets = 0;
                r_idx = find_next_bit(q_vector->txr_idx, adapter->num_tx_queues,
@@ -1445,7 +1821,7 @@ static irqreturn_t ixgbe_msix_clean_many(int irq, void *data)
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
        for (i = 0; i < q_vector->rxr_count; i++) {
-               ring = &(adapter->rx_ring[r_idx]);
+               ring = adapter->rx_ring[r_idx];
                ring->total_bytes = 0;
                ring->total_packets = 0;
                r_idx = find_next_bit(q_vector->rxr_idx, adapter->num_rx_queues,
@@ -1476,7 +1852,7 @@ static int ixgbe_clean_rxonly(struct napi_struct *napi, int budget)
        long r_idx;
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
-       rx_ring = &(adapter->rx_ring[r_idx]);
+       rx_ring = adapter->rx_ring[r_idx];
 #ifdef CONFIG_IXGBE_DCA
        if (adapter->flags & IXGBE_FLAG_DCA_ENABLED)
                ixgbe_update_rx_dca(adapter, rx_ring);
@@ -1517,7 +1893,7 @@ static int ixgbe_clean_rxtx_many(struct napi_struct *napi, int budget)
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
        for (i = 0; i < q_vector->txr_count; i++) {
-               ring = &(adapter->tx_ring[r_idx]);
+               ring = adapter->tx_ring[r_idx];
 #ifdef CONFIG_IXGBE_DCA
                if (adapter->flags & IXGBE_FLAG_DCA_ENABLED)
                        ixgbe_update_tx_dca(adapter, ring);
@@ -1533,7 +1909,7 @@ static int ixgbe_clean_rxtx_many(struct napi_struct *napi, int budget)
        budget = max(budget, 1);
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
        for (i = 0; i < q_vector->rxr_count; i++) {
-               ring = &(adapter->rx_ring[r_idx]);
+               ring = adapter->rx_ring[r_idx];
 #ifdef CONFIG_IXGBE_DCA
                if (adapter->flags & IXGBE_FLAG_DCA_ENABLED)
                        ixgbe_update_rx_dca(adapter, ring);
@@ -1544,7 +1920,7 @@ static int ixgbe_clean_rxtx_many(struct napi_struct *napi, int budget)
        }
 
        r_idx = find_first_bit(q_vector->rxr_idx, adapter->num_rx_queues);
-       ring = &(adapter->rx_ring[r_idx]);
+       ring = adapter->rx_ring[r_idx];
        /* If all Rx work done, exit the polling mode */
        if (work_done < budget) {
                napi_complete(napi);
@@ -1577,7 +1953,7 @@ static int ixgbe_clean_txonly(struct napi_struct *napi, int budget)
        long r_idx;
 
        r_idx = find_first_bit(q_vector->txr_idx, adapter->num_tx_queues);
-       tx_ring = &(adapter->tx_ring[r_idx]);
+       tx_ring = adapter->tx_ring[r_idx];
 #ifdef CONFIG_IXGBE_DCA
        if (adapter->flags & IXGBE_FLAG_DCA_ENABLED)
                ixgbe_update_tx_dca(adapter, tx_ring);
@@ -1762,8 +2138,8 @@ static void ixgbe_set_itr(struct ixgbe_adapter *adapter)
        struct ixgbe_q_vector *q_vector = adapter->q_vector[0];
        u8 current_itr;
        u32 new_itr = q_vector->eitr;
-       struct ixgbe_ring *rx_ring = &adapter->rx_ring[0];
-       struct ixgbe_ring *tx_ring = &adapter->tx_ring[0];
+       struct ixgbe_ring *rx_ring = adapter->rx_ring[0];
+       struct ixgbe_ring *tx_ring = adapter->tx_ring[0];
 
        q_vector->tx_itr = ixgbe_update_itr(adapter, new_itr,
                                            q_vector->tx_itr,
@@ -1875,10 +2251,10 @@ static irqreturn_t ixgbe_intr(int irq, void *data)
        ixgbe_check_fan_failure(adapter, eicr);
 
        if (napi_schedule_prep(&(q_vector->napi))) {
-               adapter->tx_ring[0].total_packets = 0;
-               adapter->tx_ring[0].total_bytes = 0;
-               adapter->rx_ring[0].total_packets = 0;
-               adapter->rx_ring[0].total_bytes = 0;
+               adapter->tx_ring[0]->total_packets = 0;
+               adapter->tx_ring[0]->total_bytes = 0;
+               adapter->rx_ring[0]->total_packets = 0;
+               adapter->rx_ring[0]->total_bytes = 0;
                /* would disable interrupts here but EIAM disabled it */
                __napi_schedule(&(q_vector->napi));
        }
@@ -2010,7 +2386,7 @@ static void ixgbe_configure_tx(struct ixgbe_adapter *adapter)
 
        /* Setup the HW Tx Head and Tail descriptor pointers */
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               struct ixgbe_ring *ring = &adapter->tx_ring[i];
+               struct ixgbe_ring *ring = adapter->tx_ring[i];
                j = ring->reg_idx;
                tdba = ring->dma;
                tdlen = ring->count * sizeof(union ixgbe_adv_tx_desc);
@@ -2020,8 +2396,8 @@ static void ixgbe_configure_tx(struct ixgbe_adapter *adapter)
                IXGBE_WRITE_REG(hw, IXGBE_TDLEN(j), tdlen);
                IXGBE_WRITE_REG(hw, IXGBE_TDH(j), 0);
                IXGBE_WRITE_REG(hw, IXGBE_TDT(j), 0);
-               adapter->tx_ring[i].head = IXGBE_TDH(j);
-               adapter->tx_ring[i].tail = IXGBE_TDT(j);
+               adapter->tx_ring[i]->head = IXGBE_TDH(j);
+               adapter->tx_ring[i]->tail = IXGBE_TDT(j);
                /*
                 * Disable Tx Head Writeback RO bit, since this hoses
                 * bookkeeping if things aren't delivered in order.
@@ -2168,7 +2544,7 @@ static void ixgbe_configure_rscctl(struct ixgbe_adapter *adapter, int index)
        u32 rscctrl;
        int rx_buf_len;
 
-       rx_ring = &adapter->rx_ring[index];
+       rx_ring = adapter->rx_ring[index];
        j = rx_ring->reg_idx;
        rx_buf_len = rx_ring->rx_buf_len;
        rscctrl = IXGBE_READ_REG(hw, IXGBE_RSCCTL(j));
@@ -2266,7 +2642,7 @@ static void ixgbe_configure_rx(struct ixgbe_adapter *adapter)
 #endif
        IXGBE_WRITE_REG(hw, IXGBE_HLREG0, hlreg0);
 
-       rdlen = adapter->rx_ring[0].count * sizeof(union ixgbe_adv_rx_desc);
+       rdlen = adapter->rx_ring[0]->count * sizeof(union ixgbe_adv_rx_desc);
        /* disable receives while setting up the descriptors */
        rxctrl = IXGBE_READ_REG(hw, IXGBE_RXCTRL);
        IXGBE_WRITE_REG(hw, IXGBE_RXCTRL, rxctrl & ~IXGBE_RXCTRL_RXEN);
@@ -2276,7 +2652,7 @@ static void ixgbe_configure_rx(struct ixgbe_adapter *adapter)
         * the Base and Length of the Rx Descriptor Ring
         */
        for (i = 0; i < adapter->num_rx_queues; i++) {
-               rx_ring = &adapter->rx_ring[i];
+               rx_ring = adapter->rx_ring[i];
                rdba = rx_ring->dma;
                j = rx_ring->reg_idx;
                IXGBE_WRITE_REG(hw, IXGBE_RDBAL(j), (rdba & DMA_BIT_MASK(32)));
@@ -2346,7 +2722,7 @@ static void ixgbe_configure_rx(struct ixgbe_adapter *adapter)
                IXGBE_WRITE_REG(hw, IXGBE_VFRE(reg_offset), (1 << vf_shift));
                IXGBE_WRITE_REG(hw, IXGBE_VFTE(reg_offset), (1 << vf_shift));
                IXGBE_WRITE_REG(hw, IXGBE_PFDTXGSWC, IXGBE_PFDTXGSWC_VT_LBEN);
-               ixgbe_set_vmolr(hw, adapter->num_vfs);
+               ixgbe_set_vmolr(hw, adapter->num_vfs, true);
        }
 
        /* Program MRQC for the distribution of queues */
@@ -2456,12 +2832,74 @@ static void ixgbe_vlan_rx_kill_vid(struct net_device *netdev, u16 vid)
        hw->mac.ops.set_vfta(&adapter->hw, vid, pool_ndx, false);
 }
 
+/**
+ * ixgbe_vlan_filter_disable - helper to disable hw vlan filtering
+ * @adapter: driver data
+ */
+static void ixgbe_vlan_filter_disable(struct ixgbe_adapter *adapter)
+{
+       struct ixgbe_hw *hw = &adapter->hw;
+       u32 vlnctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
+       int i, j;
+
+       switch (hw->mac.type) {
+       case ixgbe_mac_82598EB:
+               vlnctrl &= ~(IXGBE_VLNCTRL_VME | IXGBE_VLNCTRL_VFE);
+               vlnctrl &= ~IXGBE_VLNCTRL_CFIEN;
+               IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
+               break;
+       case ixgbe_mac_82599EB:
+               vlnctrl &= ~IXGBE_VLNCTRL_VFE;
+               vlnctrl &= ~IXGBE_VLNCTRL_CFIEN;
+               IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
+               for (i = 0; i < adapter->num_rx_queues; i++) {
+                       j = adapter->rx_ring[i]->reg_idx;
+                       vlnctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j));
+                       vlnctrl &= ~IXGBE_RXDCTL_VME;
+                       IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(j), vlnctrl);
+               }
+               break;
+       default:
+               break;
+       }
+}
+
+/**
+ * ixgbe_vlan_filter_enable - helper to enable hw vlan filtering
+ * @adapter: driver data
+ */
+static void ixgbe_vlan_filter_enable(struct ixgbe_adapter *adapter)
+{
+       struct ixgbe_hw *hw = &adapter->hw;
+       u32 vlnctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
+       int i, j;
+
+       switch (hw->mac.type) {
+       case ixgbe_mac_82598EB:
+               vlnctrl |= IXGBE_VLNCTRL_VME | IXGBE_VLNCTRL_VFE;
+               vlnctrl &= ~IXGBE_VLNCTRL_CFIEN;
+               IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
+               break;
+       case ixgbe_mac_82599EB:
+               vlnctrl |= IXGBE_VLNCTRL_VFE;
+               vlnctrl &= ~IXGBE_VLNCTRL_CFIEN;
+               IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
+               for (i = 0; i < adapter->num_rx_queues; i++) {
+                       j = adapter->rx_ring[i]->reg_idx;
+                       vlnctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j));
+                       vlnctrl |= IXGBE_RXDCTL_VME;
+                       IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(j), vlnctrl);
+               }
+               break;
+       default:
+               break;
+       }
+}
+
 static void ixgbe_vlan_rx_register(struct net_device *netdev,
                                    struct vlan_group *grp)
 {
        struct ixgbe_adapter *adapter = netdev_priv(netdev);
-       u32 ctrl;
-       int i, j;
 
        if (!test_bit(__IXGBE_DOWN, &adapter->state))
                ixgbe_irq_disable(adapter);
@@ -2472,25 +2910,7 @@ static void ixgbe_vlan_rx_register(struct net_device *netdev,
         * still receive traffic from a DCB-enabled host even if we're
         * not in DCB mode.
         */
-       ctrl = IXGBE_READ_REG(&adapter->hw, IXGBE_VLNCTRL);
-
-       /* Disable CFI check */
-       ctrl &= ~IXGBE_VLNCTRL_CFIEN;
-
-       /* enable VLAN tag stripping */
-       if (adapter->hw.mac.type == ixgbe_mac_82598EB) {
-               ctrl |= IXGBE_VLNCTRL_VME;
-       } else if (adapter->hw.mac.type == ixgbe_mac_82599EB) {
-               for (i = 0; i < adapter->num_rx_queues; i++) {
-                       u32 ctrl;
-                       j = adapter->rx_ring[i].reg_idx;
-                       ctrl = IXGBE_READ_REG(&adapter->hw, IXGBE_RXDCTL(j));
-                       ctrl |= IXGBE_RXDCTL_VME;
-                       IXGBE_WRITE_REG(&adapter->hw, IXGBE_RXDCTL(j), ctrl);
-               }
-       }
-
-       IXGBE_WRITE_REG(&adapter->hw, IXGBE_VLNCTRL, ctrl);
+       ixgbe_vlan_filter_enable(adapter);
 
        ixgbe_vlan_rx_add_vid(netdev, 0);
 
@@ -2512,21 +2932,6 @@ static void ixgbe_restore_vlan(struct ixgbe_adapter *adapter)
        }
 }
 
-static u8 *ixgbe_addr_list_itr(struct ixgbe_hw *hw, u8 **mc_addr_ptr, u32 *vmdq)
-{
-       struct dev_mc_list *mc_ptr;
-       u8 *addr = *mc_addr_ptr;
-       *vmdq = 0;
-
-       mc_ptr = container_of(addr, struct dev_mc_list, dmi_addr[0]);
-       if (mc_ptr->next)
-               *mc_addr_ptr = mc_ptr->next->dmi_addr;
-       else
-               *mc_addr_ptr = NULL;
-
-       return addr;
-}
-
 /**
  * ixgbe_set_rx_mode - Unicast, Multicast and Promiscuous mode set
  * @netdev: network interface device structure
@@ -2540,42 +2945,36 @@ void ixgbe_set_rx_mode(struct net_device *netdev)
 {
        struct ixgbe_adapter *adapter = netdev_priv(netdev);
        struct ixgbe_hw *hw = &adapter->hw;
-       u32 fctrl, vlnctrl;
-       u8 *addr_list = NULL;
-       int addr_count = 0;
+       u32 fctrl;
 
        /* Check for Promiscuous and All Multicast modes */
 
        fctrl = IXGBE_READ_REG(hw, IXGBE_FCTRL);
-       vlnctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
 
        if (netdev->flags & IFF_PROMISC) {
-               hw->addr_ctrl.user_set_promisc = 1;
+               hw->addr_ctrl.user_set_promisc = true;
                fctrl |= (IXGBE_FCTRL_UPE | IXGBE_FCTRL_MPE);
-               vlnctrl &= ~IXGBE_VLNCTRL_VFE;
+               /* don't hardware filter vlans in promisc mode */
+               ixgbe_vlan_filter_disable(adapter);
        } else {
                if (netdev->flags & IFF_ALLMULTI) {
                        fctrl |= IXGBE_FCTRL_MPE;
                        fctrl &= ~IXGBE_FCTRL_UPE;
-               } else {
+               } else if (!hw->addr_ctrl.uc_set_promisc) {
                        fctrl &= ~(IXGBE_FCTRL_UPE | IXGBE_FCTRL_MPE);
                }
-               vlnctrl |= IXGBE_VLNCTRL_VFE;
-               hw->addr_ctrl.user_set_promisc = 0;
+               ixgbe_vlan_filter_enable(adapter);
+               hw->addr_ctrl.user_set_promisc = false;
        }
 
        IXGBE_WRITE_REG(hw, IXGBE_FCTRL, fctrl);
-       IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
 
        /* reprogram secondary unicast list */
        hw->mac.ops.update_uc_addr_list(hw, netdev);
 
        /* reprogram multicast list */
-       addr_count = netdev->mc_count;
-       if (addr_count)
-               addr_list = netdev->mc_list->dmi_addr;
-       hw->mac.ops.update_mc_addr_list(hw, addr_list, addr_count,
-                                       ixgbe_addr_list_itr);
+       hw->mac.ops.update_mc_addr_list(hw, netdev);
+
        if (adapter->num_vfs)
                ixgbe_restore_vf_multicasts(adapter);
 }
@@ -2635,7 +3034,7 @@ static void ixgbe_napi_disable_all(struct ixgbe_adapter *adapter)
 static void ixgbe_configure_dcb(struct ixgbe_adapter *adapter)
 {
        struct ixgbe_hw *hw = &adapter->hw;
-       u32 txdctl, vlnctrl;
+       u32 txdctl;
        int i, j;
 
        ixgbe_dcb_check_config(&adapter->dcb_cfg);
@@ -2646,29 +3045,15 @@ static void ixgbe_configure_dcb(struct ixgbe_adapter *adapter)
        ixgbe_dcb_hw_config(&adapter->hw, &adapter->dcb_cfg);
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               j = adapter->tx_ring[i].reg_idx;
+               j = adapter->tx_ring[i]->reg_idx;
                txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(j));
                /* PThresh workaround for Tx hang with DFP enabled. */
                txdctl |= 32;
                IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(j), txdctl);
        }
        /* Enable VLAN tag insert/strip */
-       vlnctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
-       if (hw->mac.type == ixgbe_mac_82598EB) {
-               vlnctrl |= IXGBE_VLNCTRL_VME | IXGBE_VLNCTRL_VFE;
-               vlnctrl &= ~IXGBE_VLNCTRL_CFIEN;
-               IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
-       } else if (hw->mac.type == ixgbe_mac_82599EB) {
-               vlnctrl |= IXGBE_VLNCTRL_VFE;
-               vlnctrl &= ~IXGBE_VLNCTRL_CFIEN;
-               IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, vlnctrl);
-               for (i = 0; i < adapter->num_rx_queues; i++) {
-                       j = adapter->rx_ring[i].reg_idx;
-                       vlnctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j));
-                       vlnctrl |= IXGBE_RXDCTL_VME;
-                       IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(j), vlnctrl);
-               }
-       }
+       ixgbe_vlan_filter_enable(adapter);
+
        hw->mac.ops.set_vfta(&adapter->hw, 0, 0, true);
 }
 
@@ -2703,7 +3088,7 @@ static void ixgbe_configure(struct ixgbe_adapter *adapter)
 #endif /* IXGBE_FCOE */
        if (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE) {
                for (i = 0; i < adapter->num_tx_queues; i++)
-                       adapter->tx_ring[i].atr_sample_rate =
+                       adapter->tx_ring[i]->atr_sample_rate =
                                                       adapter->atr_sample_rate;
                ixgbe_init_fdir_signature_82599(hw, adapter->fdir_pballoc);
        } else if (adapter->flags & IXGBE_FLAG_FDIR_PERFECT_CAPABLE) {
@@ -2713,8 +3098,8 @@ static void ixgbe_configure(struct ixgbe_adapter *adapter)
        ixgbe_configure_tx(adapter);
        ixgbe_configure_rx(adapter);
        for (i = 0; i < adapter->num_rx_queues; i++)
-               ixgbe_alloc_rx_buffers(adapter, &adapter->rx_ring[i],
-                                      (adapter->rx_ring[i].count - 1));
+               ixgbe_alloc_rx_buffers(adapter, adapter->rx_ring[i],
+                                      (adapter->rx_ring[i]->count - 1));
 }
 
 static inline bool ixgbe_is_sfp(struct ixgbe_hw *hw)
@@ -2797,7 +3182,7 @@ link_cfg_out:
 static inline void ixgbe_rx_desc_queue_enable(struct ixgbe_adapter *adapter,
                                              int rxr)
 {
-       int j = adapter->rx_ring[rxr].reg_idx;
+       int j = adapter->rx_ring[rxr]->reg_idx;
        int k;
 
        for (k = 0; k < IXGBE_MAX_RX_DESC_POLL; k++) {
@@ -2811,8 +3196,8 @@ static inline void ixgbe_rx_desc_queue_enable(struct ixgbe_adapter *adapter,
                DPRINTK(DRV, ERR, "RXDCTL.ENABLE on Rx queue %d "
                        "not set within the polling period\n", rxr);
        }
-       ixgbe_release_rx_desc(&adapter->hw, &adapter->rx_ring[rxr],
-                             (adapter->rx_ring[rxr].count - 1));
+       ixgbe_release_rx_desc(&adapter->hw, adapter->rx_ring[rxr],
+                             (adapter->rx_ring[rxr]->count - 1));
 }
 
 static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
@@ -2899,10 +3284,15 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
        }
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               j = adapter->tx_ring[i].reg_idx;
+               j = adapter->tx_ring[i]->reg_idx;
                txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(j));
-               /* enable WTHRESH=8 descriptors, to encourage burst writeback */
-               txdctl |= (8 << 16);
+               if (adapter->rx_itr_setting == 0) {
+                       /* cannot set wthresh when itr==0 */
+                       txdctl &= ~0x007F0000;
+               } else {
+                       /* enable WTHRESH=8 descriptors, to encourage burst writeback */
+                       txdctl |= (8 << 16);
+               }
                IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(j), txdctl);
        }
 
@@ -2913,7 +3303,7 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
                IXGBE_WRITE_REG(hw, IXGBE_DMATXCTL, dmatxctl);
        }
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               j = adapter->tx_ring[i].reg_idx;
+               j = adapter->tx_ring[i]->reg_idx;
                txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(j));
                txdctl |= IXGBE_TXDCTL_ENABLE;
                IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(j), txdctl);
@@ -2932,7 +3322,7 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
        }
 
        for (i = 0; i < num_rx_rings; i++) {
-               j = adapter->rx_ring[i].reg_idx;
+               j = adapter->rx_ring[i]->reg_idx;
                rxdctl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(j));
                /* enable PTHRESH=32 descriptors (half the internal cache)
                 * and HTHRESH=0 descriptors (to minimize latency on fetch),
@@ -2956,6 +3346,10 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
        else
                ixgbe_configure_msi_and_legacy(adapter);
 
+       /* enable the optics */
+       if (hw->phy.multispeed_fiber)
+               hw->mac.ops.enable_tx_laser(hw);
+
        clear_bit(__IXGBE_DOWN, &adapter->state);
        ixgbe_napi_enable_all(adapter);
 
@@ -3006,7 +3400,7 @@ static int ixgbe_up_complete(struct ixgbe_adapter *adapter)
 
        for (i = 0; i < adapter->num_tx_queues; i++)
                set_bit(__IXGBE_FDIR_INIT_DONE,
-                       &(adapter->tx_ring[i].reinit_state));
+                       &(adapter->tx_ring[i]->reinit_state));
 
        /* enable transmits */
        netif_tx_start_all_queues(netdev);
@@ -3031,6 +3425,14 @@ void ixgbe_reinit_locked(struct ixgbe_adapter *adapter)
        while (test_and_set_bit(__IXGBE_RESETTING, &adapter->state))
                msleep(1);
        ixgbe_down(adapter);
+       /*
+        * If SR-IOV enabled then wait a bit before bringing the adapter
+        * back up to give the VFs time to respond to the reset.  The
+        * two second wait is based upon the watchdog timer cycle in
+        * the VF driver.
+        */
+       if (adapter->flags & IXGBE_FLAG_SRIOV_ENABLED)
+               msleep(2000);
        ixgbe_up(adapter);
        clear_bit(__IXGBE_RESETTING, &adapter->state);
 }
@@ -3093,9 +3495,9 @@ static void ixgbe_clean_rx_ring(struct ixgbe_adapter *adapter,
 
                rx_buffer_info = &rx_ring->rx_buffer_info[i];
                if (rx_buffer_info->dma) {
-                       pci_unmap_single(pdev, rx_buffer_info->dma,
+                       dma_unmap_single(&pdev->dev, rx_buffer_info->dma,
                                         rx_ring->rx_buf_len,
-                                        PCI_DMA_FROMDEVICE);
+                                        DMA_FROM_DEVICE);
                        rx_buffer_info->dma = 0;
                }
                if (rx_buffer_info->skb) {
@@ -3103,6 +3505,13 @@ static void ixgbe_clean_rx_ring(struct ixgbe_adapter *adapter,
                        rx_buffer_info->skb = NULL;
                        do {
                                struct sk_buff *this = skb;
+                               if (IXGBE_RSC_CB(this)->dma) {
+                                       dma_unmap_single(&pdev->dev,
+                                                        IXGBE_RSC_CB(this)->dma,
+                                                        rx_ring->rx_buf_len,
+                                                        DMA_FROM_DEVICE);
+                                       IXGBE_RSC_CB(this)->dma = 0;
+                               }
                                skb = skb->prev;
                                dev_kfree_skb(this);
                        } while (skb);
@@ -3110,8 +3519,8 @@ static void ixgbe_clean_rx_ring(struct ixgbe_adapter *adapter,
                if (!rx_buffer_info->page)
                        continue;
                if (rx_buffer_info->page_dma) {
-                       pci_unmap_page(pdev, rx_buffer_info->page_dma,
-                                      PAGE_SIZE / 2, PCI_DMA_FROMDEVICE);
+                       dma_unmap_page(&pdev->dev, rx_buffer_info->page_dma,
+                                      PAGE_SIZE / 2, DMA_FROM_DEVICE);
                        rx_buffer_info->page_dma = 0;
                }
                put_page(rx_buffer_info->page);
@@ -3177,7 +3586,7 @@ static void ixgbe_clean_all_rx_rings(struct ixgbe_adapter *adapter)
        int i;
 
        for (i = 0; i < adapter->num_rx_queues; i++)
-               ixgbe_clean_rx_ring(adapter, &adapter->rx_ring[i]);
+               ixgbe_clean_rx_ring(adapter, adapter->rx_ring[i]);
 }
 
 /**
@@ -3189,7 +3598,7 @@ static void ixgbe_clean_all_tx_rings(struct ixgbe_adapter *adapter)
        int i;
 
        for (i = 0; i < adapter->num_tx_queues; i++)
-               ixgbe_clean_tx_ring(adapter, &adapter->tx_ring[i]);
+               ixgbe_clean_tx_ring(adapter, adapter->tx_ring[i]);
 }
 
 void ixgbe_down(struct ixgbe_adapter *adapter)
@@ -3203,44 +3612,51 @@ void ixgbe_down(struct ixgbe_adapter *adapter)
        /* signal that we are down to the interrupt handler */
        set_bit(__IXGBE_DOWN, &adapter->state);
 
+       /* power down the optics */
+       if (hw->phy.multispeed_fiber)
+               hw->mac.ops.disable_tx_laser(hw);
+
        /* disable receive for all VFs and wait one second */
        if (adapter->num_vfs) {
-               for (i = 0 ; i < adapter->num_vfs; i++)
-                       adapter->vfinfo[i].clear_to_send = 0;
-
                /* ping all the active vfs to let them know we are going down */
                ixgbe_ping_all_vfs(adapter);
+
                /* Disable all VFTE/VFRE TX/RX */
                ixgbe_disable_tx_rx(adapter);
+
+               /* Mark all the VFs as inactive */
+               for (i = 0 ; i < adapter->num_vfs; i++)
+                       adapter->vfinfo[i].clear_to_send = 0;
        }
 
        /* disable receives */
        rxctrl = IXGBE_READ_REG(hw, IXGBE_RXCTRL);
        IXGBE_WRITE_REG(hw, IXGBE_RXCTRL, rxctrl & ~IXGBE_RXCTRL_RXEN);
 
-       netif_tx_disable(netdev);
-
        IXGBE_WRITE_FLUSH(hw);
        msleep(10);
 
        netif_tx_stop_all_queues(netdev);
 
-       ixgbe_irq_disable(adapter);
-
-       ixgbe_napi_disable_all(adapter);
-
        clear_bit(__IXGBE_SFP_MODULE_NOT_FOUND, &adapter->state);
        del_timer_sync(&adapter->sfp_timer);
        del_timer_sync(&adapter->watchdog_timer);
        cancel_work_sync(&adapter->watchdog_task);
 
+       netif_carrier_off(netdev);
+       netif_tx_disable(netdev);
+
+       ixgbe_irq_disable(adapter);
+
+       ixgbe_napi_disable_all(adapter);
+
        if (adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE ||
            adapter->flags & IXGBE_FLAG_FDIR_PERFECT_CAPABLE)
                cancel_work_sync(&adapter->fdir_reinit_task);
 
        /* disable transmits in the hardware now that interrupts are off */
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               j = adapter->tx_ring[i].reg_idx;
+               j = adapter->tx_ring[i]->reg_idx;
                txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(j));
                IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(j),
                                (txdctl & ~IXGBE_TXDCTL_ENABLE));
@@ -3251,7 +3667,8 @@ void ixgbe_down(struct ixgbe_adapter *adapter)
                                (IXGBE_READ_REG(hw, IXGBE_DMATXCTL) &
                                 ~IXGBE_DMATXCTL_TE));
 
-       netif_carrier_off(netdev);
+       /* clear n-tuple filters that are cached */
+       ethtool_ntuple_flush(netdev);
 
        if (!pci_channel_offline(adapter->pdev))
                ixgbe_reset(adapter);
@@ -3280,13 +3697,13 @@ static int ixgbe_poll(struct napi_struct *napi, int budget)
 
 #ifdef CONFIG_IXGBE_DCA
        if (adapter->flags & IXGBE_FLAG_DCA_ENABLED) {
-               ixgbe_update_tx_dca(adapter, adapter->tx_ring);
-               ixgbe_update_rx_dca(adapter, adapter->rx_ring);
+               ixgbe_update_tx_dca(adapter, adapter->tx_ring[0]);
+               ixgbe_update_rx_dca(adapter, adapter->rx_ring[0]);
        }
 #endif
 
-       tx_clean_complete = ixgbe_clean_tx_irq(q_vector, adapter->tx_ring);
-       ixgbe_clean_rx_irq(q_vector, adapter->rx_ring, &work_done, budget);
+       tx_clean_complete = ixgbe_clean_tx_irq(q_vector, adapter->tx_ring[0]);
+       ixgbe_clean_rx_irq(q_vector, adapter->rx_ring[0], &work_done, budget);
 
        if (!tx_clean_complete)
                work_done = budget;
@@ -3326,6 +3743,8 @@ static void ixgbe_reset_task(struct work_struct *work)
 
        adapter->tx_timeout_count++;
 
+       ixgbe_dump(adapter);
+       netdev_err(adapter->netdev, "Reset adapter\n");
        ixgbe_reinit_locked(adapter);
 }
 
@@ -3426,12 +3845,12 @@ static inline bool ixgbe_set_fcoe_queues(struct ixgbe_adapter *adapter)
                adapter->num_tx_queues = 1;
 #ifdef CONFIG_IXGBE_DCB
                if (adapter->flags & IXGBE_FLAG_DCB_ENABLED) {
-                       DPRINTK(PROBE, INFO, "FCoE enabled with DCB \n");
+                       DPRINTK(PROBE, INFO, "FCoE enabled with DCB\n");
                        ixgbe_set_dcb_queues(adapter);
                }
 #endif
                if (adapter->flags & IXGBE_FLAG_RSS_ENABLED) {
-                       DPRINTK(PROBE, INFO, "FCoE enabled with RSS \n");
+                       DPRINTK(PROBE, INFO, "FCoE enabled with RSS\n");
                        if ((adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE) ||
                            (adapter->flags & IXGBE_FLAG_FDIR_PERFECT_CAPABLE))
                                ixgbe_set_fdir_queues(adapter);
@@ -3574,9 +3993,9 @@ static inline bool ixgbe_cache_ring_rss(struct ixgbe_adapter *adapter)
 
        if (adapter->flags & IXGBE_FLAG_RSS_ENABLED) {
                for (i = 0; i < adapter->num_rx_queues; i++)
-                       adapter->rx_ring[i].reg_idx = i;
+                       adapter->rx_ring[i]->reg_idx = i;
                for (i = 0; i < adapter->num_tx_queues; i++)
-                       adapter->tx_ring[i].reg_idx = i;
+                       adapter->tx_ring[i]->reg_idx = i;
                ret = true;
        } else {
                ret = false;
@@ -3603,8 +4022,8 @@ static inline bool ixgbe_cache_ring_dcb(struct ixgbe_adapter *adapter)
                if (adapter->hw.mac.type == ixgbe_mac_82598EB) {
                        /* the number of queues is assumed to be symmetric */
                        for (i = 0; i < dcb_i; i++) {
-                               adapter->rx_ring[i].reg_idx = i << 3;
-                               adapter->tx_ring[i].reg_idx = i << 2;
+                               adapter->rx_ring[i]->reg_idx = i << 3;
+                               adapter->tx_ring[i]->reg_idx = i << 2;
                        }
                        ret = true;
                } else if (adapter->hw.mac.type == ixgbe_mac_82599EB) {
@@ -3622,18 +4041,18 @@ static inline bool ixgbe_cache_ring_dcb(struct ixgbe_adapter *adapter)
                                 * Rx TC0-TC7 are offset by 16 queues each
                                 */
                                for (i = 0; i < 3; i++) {
-                                       adapter->tx_ring[i].reg_idx = i << 5;
-                                       adapter->rx_ring[i].reg_idx = i << 4;
+                                       adapter->tx_ring[i]->reg_idx = i << 5;
+                                       adapter->rx_ring[i]->reg_idx = i << 4;
                                }
                                for ( ; i < 5; i++) {
-                                       adapter->tx_ring[i].reg_idx =
+                                       adapter->tx_ring[i]->reg_idx =
                                                                 ((i + 2) << 4);
-                                       adapter->rx_ring[i].reg_idx = i << 4;
+                                       adapter->rx_ring[i]->reg_idx = i << 4;
                                }
                                for ( ; i < dcb_i; i++) {
-                                       adapter->tx_ring[i].reg_idx =
+                                       adapter->tx_ring[i]->reg_idx =
                                                                 ((i + 8) << 3);
-                                       adapter->rx_ring[i].reg_idx = i << 4;
+                                       adapter->rx_ring[i]->reg_idx = i << 4;
                                }
 
                                ret = true;
@@ -3646,12 +4065,12 @@ static inline bool ixgbe_cache_ring_dcb(struct ixgbe_adapter *adapter)
                                 *
                                 * Rx TC0-TC3 are offset by 32 queues each
                                 */
-                               adapter->tx_ring[0].reg_idx = 0;
-                               adapter->tx_ring[1].reg_idx = 64;
-                               adapter->tx_ring[2].reg_idx = 96;
-                               adapter->tx_ring[3].reg_idx = 112;
+                               adapter->tx_ring[0]->reg_idx = 0;
+                               adapter->tx_ring[1]->reg_idx = 64;
+                               adapter->tx_ring[2]->reg_idx = 96;
+                               adapter->tx_ring[3]->reg_idx = 112;
                                for (i = 0 ; i < dcb_i; i++)
-                                       adapter->rx_ring[i].reg_idx = i << 5;
+                                       adapter->rx_ring[i]->reg_idx = i << 5;
 
                                ret = true;
                        } else {
@@ -3684,9 +4103,9 @@ static bool inline ixgbe_cache_ring_fdir(struct ixgbe_adapter *adapter)
            ((adapter->flags & IXGBE_FLAG_FDIR_HASH_CAPABLE) ||
             (adapter->flags & IXGBE_FLAG_FDIR_PERFECT_CAPABLE))) {
                for (i = 0; i < adapter->num_rx_queues; i++)
-                       adapter->rx_ring[i].reg_idx = i;
+                       adapter->rx_ring[i]->reg_idx = i;
                for (i = 0; i < adapter->num_tx_queues; i++)
-                       adapter->tx_ring[i].reg_idx = i;
+                       adapter->tx_ring[i]->reg_idx = i;
                ret = true;
        }
 
@@ -3714,8 +4133,8 @@ static inline bool ixgbe_cache_ring_fcoe(struct ixgbe_adapter *adapter)
 
                        ixgbe_cache_ring_dcb(adapter);
                        /* find out queues in TC for FCoE */
-                       fcoe_rx_i = adapter->rx_ring[fcoe->tc].reg_idx + 1;
-                       fcoe_tx_i = adapter->tx_ring[fcoe->tc].reg_idx + 1;
+                       fcoe_rx_i = adapter->rx_ring[fcoe->tc]->reg_idx + 1;
+                       fcoe_tx_i = adapter->tx_ring[fcoe->tc]->reg_idx + 1;
                        /*
                         * In 82599, the number of Tx queues for each traffic
                         * class for both 8-TC and 4-TC modes are:
@@ -3746,8 +4165,8 @@ static inline bool ixgbe_cache_ring_fcoe(struct ixgbe_adapter *adapter)
                        fcoe_tx_i = f->mask;
                }
                for (i = 0; i < f->indices; i++, fcoe_rx_i++, fcoe_tx_i++) {
-                       adapter->rx_ring[f->mask + i].reg_idx = fcoe_rx_i;
-                       adapter->tx_ring[f->mask + i].reg_idx = fcoe_tx_i;
+                       adapter->rx_ring[f->mask + i]->reg_idx = fcoe_rx_i;
+                       adapter->tx_ring[f->mask + i]->reg_idx = fcoe_tx_i;
                }
                ret = true;
        }
@@ -3765,8 +4184,8 @@ static inline bool ixgbe_cache_ring_fcoe(struct ixgbe_adapter *adapter)
  */
 static inline bool ixgbe_cache_ring_sriov(struct ixgbe_adapter *adapter)
 {
-       adapter->rx_ring[0].reg_idx = adapter->num_vfs * 2;
-       adapter->tx_ring[0].reg_idx = adapter->num_vfs * 2;
+       adapter->rx_ring[0]->reg_idx = adapter->num_vfs * 2;
+       adapter->tx_ring[0]->reg_idx = adapter->num_vfs * 2;
        if (adapter->num_vfs)
                return true;
        else
@@ -3787,8 +4206,8 @@ static inline bool ixgbe_cache_ring_sriov(struct ixgbe_adapter *adapter)
 static void ixgbe_cache_ring_register(struct ixgbe_adapter *adapter)
 {
        /* start with default case */
-       adapter->rx_ring[0].reg_idx = 0;
-       adapter->tx_ring[0].reg_idx = 0;
+       adapter->rx_ring[0]->reg_idx = 0;
+       adapter->tx_ring[0]->reg_idx = 0;
 
        if (ixgbe_cache_ring_sriov(adapter))
                return;
@@ -3821,33 +4240,63 @@ static void ixgbe_cache_ring_register(struct ixgbe_adapter *adapter)
 static int ixgbe_alloc_queues(struct ixgbe_adapter *adapter)
 {
        int i;
-
-       adapter->tx_ring = kcalloc(adapter->num_tx_queues,
-                                  sizeof(struct ixgbe_ring), GFP_KERNEL);
-       if (!adapter->tx_ring)
-               goto err_tx_ring_allocation;
-
-       adapter->rx_ring = kcalloc(adapter->num_rx_queues,
-                                  sizeof(struct ixgbe_ring), GFP_KERNEL);
-       if (!adapter->rx_ring)
-               goto err_rx_ring_allocation;
+       int orig_node = adapter->node;
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               adapter->tx_ring[i].count = adapter->tx_ring_count;
-               adapter->tx_ring[i].queue_index = i;
+               struct ixgbe_ring *ring = adapter->tx_ring[i];
+               if (orig_node == -1) {
+                       int cur_node = next_online_node(adapter->node);
+                       if (cur_node == MAX_NUMNODES)
+                               cur_node = first_online_node;
+                       adapter->node = cur_node;
+               }
+               ring = kzalloc_node(sizeof(struct ixgbe_ring), GFP_KERNEL,
+                                   adapter->node);
+               if (!ring)
+                       ring = kzalloc(sizeof(struct ixgbe_ring), GFP_KERNEL);
+               if (!ring)
+                       goto err_tx_ring_allocation;
+               ring->count = adapter->tx_ring_count;
+               ring->queue_index = i;
+               ring->numa_node = adapter->node;
+
+               adapter->tx_ring[i] = ring;
        }
 
+       /* Restore the adapter's original node */
+       adapter->node = orig_node;
+
        for (i = 0; i < adapter->num_rx_queues; i++) {
-               adapter->rx_ring[i].count = adapter->rx_ring_count;
-               adapter->rx_ring[i].queue_index = i;
+               struct ixgbe_ring *ring = adapter->rx_ring[i];
+               if (orig_node == -1) {
+                       int cur_node = next_online_node(adapter->node);
+                       if (cur_node == MAX_NUMNODES)
+                               cur_node = first_online_node;
+                       adapter->node = cur_node;
+               }
+               ring = kzalloc_node(sizeof(struct ixgbe_ring), GFP_KERNEL,
+                                   adapter->node);
+               if (!ring)
+                       ring = kzalloc(sizeof(struct ixgbe_ring), GFP_KERNEL);
+               if (!ring)
+                       goto err_rx_ring_allocation;
+               ring->count = adapter->rx_ring_count;
+               ring->queue_index = i;
+               ring->numa_node = adapter->node;
+
+               adapter->rx_ring[i] = ring;
        }
 
+       /* Restore the adapter's original node */
+       adapter->node = orig_node;
+
        ixgbe_cache_ring_register(adapter);
 
        return 0;
 
 err_rx_ring_allocation:
-       kfree(adapter->tx_ring);
+       for (i = 0; i < adapter->num_tx_queues; i++)
+               kfree(adapter->tx_ring[i]);
 err_tx_ring_allocation:
        return -ENOMEM;
 }
@@ -3946,7 +4395,11 @@ static int ixgbe_alloc_q_vectors(struct ixgbe_adapter *adapter)
        }
 
        for (q_idx = 0; q_idx < num_q_vectors; q_idx++) {
-               q_vector = kzalloc(sizeof(struct ixgbe_q_vector), GFP_KERNEL);
+               q_vector = kzalloc_node(sizeof(struct ixgbe_q_vector),
+                                       GFP_KERNEL, adapter->node);
+               if (!q_vector)
+                       q_vector = kzalloc(sizeof(struct ixgbe_q_vector),
+                                          GFP_KERNEL);
                if (!q_vector)
                        goto err_out;
                q_vector->adapter = adapter;
@@ -4073,10 +4526,16 @@ err_set_interrupt:
  **/
 void ixgbe_clear_interrupt_scheme(struct ixgbe_adapter *adapter)
 {
-       kfree(adapter->tx_ring);
-       kfree(adapter->rx_ring);
-       adapter->tx_ring = NULL;
-       adapter->rx_ring = NULL;
+       int i;
+
+       for (i = 0; i < adapter->num_tx_queues; i++) {
+               kfree(adapter->tx_ring[i]);
+               adapter->tx_ring[i] = NULL;
+       }
+       for (i = 0; i < adapter->num_rx_queues; i++) {
+               kfree(adapter->rx_ring[i]);
+               adapter->rx_ring[i] = NULL;
+       }
 
        ixgbe_free_q_vectors(adapter);
        ixgbe_reset_interrupt_capability(adapter);
@@ -4147,6 +4606,7 @@ static int __devinit ixgbe_sw_init(struct ixgbe_adapter *adapter)
 {
        struct ixgbe_hw *hw = &adapter->hw;
        struct pci_dev *pdev = adapter->pdev;
+       struct net_device *dev = adapter->netdev;
        unsigned int rss;
 #ifdef CONFIG_IXGBE_DCB
        int j;
@@ -4174,10 +4634,18 @@ static int __devinit ixgbe_sw_init(struct ixgbe_adapter *adapter)
                adapter->max_msix_q_vectors = MAX_MSIX_Q_VECTORS_82599;
                adapter->flags2 |= IXGBE_FLAG2_RSC_CAPABLE;
                adapter->flags2 |= IXGBE_FLAG2_RSC_ENABLED;
-               adapter->flags |= IXGBE_FLAG_FDIR_HASH_CAPABLE;
+               if (dev->features & NETIF_F_NTUPLE) {
+                       /* Flow Director perfect filter enabled */
+                       adapter->flags |= IXGBE_FLAG_FDIR_PERFECT_CAPABLE;
+                       adapter->atr_sample_rate = 0;
+                       spin_lock_init(&adapter->fdir_perfect_lock);
+               } else {
+                       /* Flow Director hash filters enabled */
+                       adapter->flags |= IXGBE_FLAG_FDIR_HASH_CAPABLE;
+                       adapter->atr_sample_rate = 20;
+               }
                adapter->ring_feature[RING_F_FDIR].indices =
                                                         IXGBE_MAX_FDIR_INDICES;
-               adapter->atr_sample_rate = 20;
                adapter->fdir_pballoc = 0;
 #ifdef IXGBE_FCOE
                adapter->flags |= IXGBE_FLAG_FCOE_CAPABLE;
@@ -4246,6 +4714,9 @@ static int __devinit ixgbe_sw_init(struct ixgbe_adapter *adapter)
        /* enable rx csum by default */
        adapter->flags |= IXGBE_FLAG_RX_CSUM_ENABLED;
 
+       /* get assigned NUMA node */
+       adapter->node = dev_to_node(&pdev->dev);
+
        set_bit(__IXGBE_DOWN, &adapter->state);
 
        return 0;
@@ -4265,7 +4736,9 @@ int ixgbe_setup_tx_resources(struct ixgbe_adapter *adapter,
        int size;
 
        size = sizeof(struct ixgbe_tx_buffer) * tx_ring->count;
-       tx_ring->tx_buffer_info = vmalloc(size);
+       tx_ring->tx_buffer_info = vmalloc_node(size, tx_ring->numa_node);
+       if (!tx_ring->tx_buffer_info)
+               tx_ring->tx_buffer_info = vmalloc(size);
        if (!tx_ring->tx_buffer_info)
                goto err;
        memset(tx_ring->tx_buffer_info, 0, size);
@@ -4274,8 +4747,8 @@ int ixgbe_setup_tx_resources(struct ixgbe_adapter *adapter,
        tx_ring->size = tx_ring->count * sizeof(union ixgbe_adv_tx_desc);
        tx_ring->size = ALIGN(tx_ring->size, 4096);
 
-       tx_ring->desc = pci_alloc_consistent(pdev, tx_ring->size,
-                                            &tx_ring->dma);
+       tx_ring->desc = dma_alloc_coherent(&pdev->dev, tx_ring->size,
+                                          &tx_ring->dma, GFP_KERNEL);
        if (!tx_ring->desc)
                goto err;
 
@@ -4307,7 +4780,7 @@ static int ixgbe_setup_all_tx_resources(struct ixgbe_adapter *adapter)
        int i, err = 0;
 
        for (i = 0; i < adapter->num_tx_queues; i++) {
-               err = ixgbe_setup_tx_resources(adapter, &adapter->tx_ring[i]);
+               err = ixgbe_setup_tx_resources(adapter, adapter->tx_ring[i]);
                if (!err)
                        continue;
                DPRINTK(PROBE, ERR, "Allocation for Tx Queue %u failed\n", i);
@@ -4331,7 +4804,9 @@ int ixgbe_setup_rx_resources(struct ixgbe_adapter *adapter,
        int size;
 
        size = sizeof(struct ixgbe_rx_buffer) * rx_ring->count;
-       rx_ring->rx_buffer_info = vmalloc(size);
+       rx_ring->rx_buffer_info = vmalloc_node(size, adapter->node);
+       if (!rx_ring->rx_buffer_info)
+               rx_ring->rx_buffer_info = vmalloc(size);
        if (!rx_ring->rx_buffer_info) {
                DPRINTK(PROBE, ERR,
                        "vmalloc allocation failed for the rx desc ring\n");
@@ -4343,7 +4818,8 @@ int ixgbe_setup_rx_resources(struct ixgbe_adapter *adapter,
        rx_ring->size = rx_ring->count * sizeof(union ixgbe_adv_rx_desc);
        rx_ring->size = ALIGN(rx_ring->size, 4096);
 
-       rx_ring->desc = pci_alloc_consistent(pdev, rx_ring->size, &rx_ring->dma);
+       rx_ring->desc = dma_alloc_coherent(&pdev->dev, rx_ring->size,
+                                          &rx_ring->dma, GFP_KERNEL);
 
        if (!rx_ring->desc) {
                DPRINTK(PROBE, ERR,
@@ -4377,7 +4853,7 @@ static int ixgbe_setup_all_rx_resources(struct ixgbe_adapter *adapter)
        int i, err = 0;
 
        for (i = 0; i < adapter->num_rx_queues; i++) {
-               err = ixgbe_setup_rx_resources(adapter, &adapter->rx_ring[i]);
+               err = ixgbe_setup_rx_resources(adapter, adapter->rx_ring[i]);
                if (!err)
                        continue;
                DPRINTK(PROBE, ERR, "Allocation for Rx Queue %u failed\n", i);
@@ -4404,7 +4880,8 @@ void ixgbe_free_tx_resources(struct ixgbe_adapter *adapter,
        vfree(tx_ring->tx_buffer_info);
        tx_ring->tx_buffer_info = NULL;
 
-       pci_free_consistent(pdev, tx_ring->size, tx_ring->desc, tx_ring->dma);
+       dma_free_coherent(&pdev->dev, tx_ring->size, tx_ring->desc,
+                         tx_ring->dma);
 
        tx_ring->desc = NULL;
 }
@@ -4420,8 +4897,8 @@ static void ixgbe_free_all_tx_resources(struct ixgbe_adapter *adapter)
        int i;
 
        for (i = 0; i < adapter->num_tx_queues; i++)
-               if (adapter->tx_ring[i].desc)
-                       ixgbe_free_tx_resources(adapter, &adapter->tx_ring[i]);
+               if (adapter->tx_ring[i]->desc)
+                       ixgbe_free_tx_resources(adapter, adapter->tx_ring[i]);
 }
 
 /**
@@ -4441,7 +4918,8 @@ void ixgbe_free_rx_resources(struct ixgbe_adapter *adapter,
        vfree(rx_ring->rx_buffer_info);
        rx_ring->rx_buffer_info = NULL;
 
-       pci_free_consistent(pdev, rx_ring->size, rx_ring->desc, rx_ring->dma);
+       dma_free_coherent(&pdev->dev, rx_ring->size, rx_ring->desc,
+                         rx_ring->dma);
 
        rx_ring->desc = NULL;
 }
@@ -4457,8 +4935,8 @@ static void ixgbe_free_all_rx_resources(struct ixgbe_adapter *adapter)
        int i;
 
        for (i = 0; i < adapter->num_rx_queues; i++)
-               if (adapter->rx_ring[i].desc)
-                       ixgbe_free_rx_resources(adapter, &adapter->rx_ring[i]);
+               if (adapter->rx_ring[i]->desc)
+                       ixgbe_free_rx_resources(adapter, adapter->rx_ring[i]);
 }
 
 /**
@@ -4735,8 +5213,8 @@ void ixgbe_update_stats(struct ixgbe_adapter *adapter)
                        adapter->hw_rx_no_dma_resources +=
                                             IXGBE_READ_REG(hw, IXGBE_QPRDC(i));
                for (i = 0; i < adapter->num_rx_queues; i++) {
-                       rsc_count += adapter->rx_ring[i].rsc_count;
-                       rsc_flush += adapter->rx_ring[i].rsc_flush;
+                       rsc_count += adapter->rx_ring[i]->rsc_count;
+                       rsc_flush += adapter->rx_ring[i]->rsc_flush;
                }
                adapter->rsc_total_count = rsc_count;
                adapter->rsc_total_flush = rsc_flush;
@@ -4744,11 +5222,11 @@ void ixgbe_update_stats(struct ixgbe_adapter *adapter)
 
        /* gather some stats to the adapter struct that are per queue */
        for (i = 0; i < adapter->num_tx_queues; i++)
-               restart_queue += adapter->tx_ring[i].restart_queue;
+               restart_queue += adapter->tx_ring[i]->restart_queue;
        adapter->restart_queue = restart_queue;
 
        for (i = 0; i < adapter->num_rx_queues; i++)
-               non_eop_descs += adapter->rx_ring[i].non_eop_descs;
+               non_eop_descs += adapter->rx_ring[i]->non_eop_descs;
        adapter->non_eop_descs = non_eop_descs;
 
        adapter->stats.crcerrs += IXGBE_READ_REG(hw, IXGBE_CRCERRS);
@@ -4932,6 +5410,7 @@ static void ixgbe_multispeed_fiber_task(struct work_struct *work)
        autoneg = hw->phy.autoneg_advertised;
        if ((!autoneg) && (hw->mac.ops.get_link_capabilities))
                hw->mac.ops.get_link_capabilities(hw, &autoneg, &negotiation);
+       hw->mac.autotry_restart = false;
        if (hw->mac.ops.setup_link)
                hw->mac.ops.setup_link(hw, autoneg, negotiation, true);
        adapter->flags |= IXGBE_FLAG_NEED_LINK_UPDATE;
@@ -4987,15 +5466,17 @@ static void ixgbe_fdir_reinit_task(struct work_struct *work)
        if (ixgbe_reinit_fdir_tables_82599(hw) == 0) {
                for (i = 0; i < adapter->num_tx_queues; i++)
                        set_bit(__IXGBE_FDIR_INIT_DONE,
-                               &(adapter->tx_ring[i].reinit_state));
+                               &(adapter->tx_ring[i]->reinit_state));
        } else {
                DPRINTK(PROBE, ERR, "failed to finish FDIR re-initialization, "
-                       "ignored adding FDIR ATR filters \n");
+                       "ignored adding FDIR ATR filters\n");
        }
        /* Done FDIR Re-initialization, enable transmits */
        netif_tx_start_all_queues(adapter->netdev);
 }
 
+static DEFINE_MUTEX(ixgbe_watchdog_lock);
+
 /**
  * ixgbe_watchdog_task - worker thread to bring link up
  * @work: pointer to work_struct containing our data
@@ -5007,13 +5488,16 @@ static void ixgbe_watchdog_task(struct work_struct *work)
                                                     watchdog_task);
        struct net_device *netdev = adapter->netdev;
        struct ixgbe_hw *hw = &adapter->hw;
-       u32 link_speed = adapter->link_speed;
-       bool link_up = adapter->link_up;
+       u32 link_speed;
+       bool link_up;
        int i;
        struct ixgbe_ring *tx_ring;
        int some_tx_pending = 0;
 
-       adapter->flags |= IXGBE_FLAG_IN_WATCHDOG_TASK;
+       mutex_lock(&ixgbe_watchdog_lock);
+
+       link_up = adapter->link_up;
+       link_speed = adapter->link_speed;
 
        if (adapter->flags & IXGBE_FLAG_NEED_LINK_UPDATE) {
                hw->mac.ops.check_link(hw, &link_speed, &link_up, false);
@@ -5084,7 +5568,7 @@ static void ixgbe_watchdog_task(struct work_struct *work)
 
        if (!netif_carrier_ok(netdev)) {
                for (i = 0; i < adapter->num_tx_queues; i++) {
-                       tx_ring = &adapter->tx_ring[i];
+                       tx_ring = adapter->tx_ring[i];
                        if (tx_ring->next_to_use != tx_ring->next_to_clean) {
                                some_tx_pending = 1;
                                break;
@@ -5102,7 +5586,7 @@ static void ixgbe_watchdog_task(struct work_struct *work)
        }
 
        ixgbe_update_stats(adapter);
-       adapter->flags &= ~IXGBE_FLAG_IN_WATCHDOG_TASK;
+       mutex_unlock(&ixgbe_watchdog_lock);
 }
 
 static int ixgbe_tso(struct ixgbe_adapter *adapter,
@@ -5305,10 +5789,10 @@ static int ixgbe_tx_map(struct ixgbe_adapter *adapter,
 
                tx_buffer_info->length = size;
                tx_buffer_info->mapped_as_page = false;
-               tx_buffer_info->dma = pci_map_single(pdev,
+               tx_buffer_info->dma = dma_map_single(&pdev->dev,
                                                     skb->data + offset,
-                                                    size, PCI_DMA_TODEVICE);
-               if (pci_dma_mapping_error(pdev, tx_buffer_info->dma))
+                                                    size, DMA_TO_DEVICE);
+               if (dma_mapping_error(&pdev->dev, tx_buffer_info->dma))
                        goto dma_error;
                tx_buffer_info->time_stamp = jiffies;
                tx_buffer_info->next_to_watch = i;
@@ -5341,12 +5825,12 @@ static int ixgbe_tx_map(struct ixgbe_adapter *adapter,
                        size = min(len, (uint)IXGBE_MAX_DATA_PER_TXD);
 
                        tx_buffer_info->length = size;
-                       tx_buffer_info->dma = pci_map_page(adapter->pdev,
+                       tx_buffer_info->dma = dma_map_page(&adapter->pdev->dev,
                                                           frag->page,
                                                           offset, size,
-                                                          PCI_DMA_TODEVICE);
+                                                          DMA_TO_DEVICE);
                        tx_buffer_info->mapped_as_page = true;
-                       if (pci_dma_mapping_error(pdev, tx_buffer_info->dma))
+                       if (dma_mapping_error(&pdev->dev, tx_buffer_info->dma))
                                goto dma_error;
                        tx_buffer_info->time_stamp = jiffies;
                        tx_buffer_info->next_to_watch = i;
@@ -5542,14 +6026,21 @@ static u16 ixgbe_select_queue(struct net_device *dev, struct sk_buff *skb)
 
 #ifdef IXGBE_FCOE
        if ((adapter->flags & IXGBE_FLAG_FCOE_ENABLED) &&
-           (skb->protocol == htons(ETH_P_FCOE))) {
+           ((skb->protocol == htons(ETH_P_FCOE)) ||
+            (skb->protocol == htons(ETH_P_FIP)))) {
                txq &= (adapter->ring_feature[RING_F_FCOE].indices - 1);
                txq += adapter->ring_feature[RING_F_FCOE].mask;
                return txq;
        }
 #endif
-       if (adapter->flags & IXGBE_FLAG_DCB_ENABLED)
-               return (skb->vlan_tci & IXGBE_TX_FLAGS_VLAN_PRIO_MASK) >> 13;
+       if (adapter->flags & IXGBE_FLAG_DCB_ENABLED) {
+               if (skb->priority == TC_PRIO_CONTROL)
+                       txq = adapter->ring_feature[RING_F_DCB].indices-1;
+               else
+                       txq = (skb->vlan_tci & IXGBE_TX_FLAGS_VLAN_PRIO_MASK)
+                              >> 13;
+               return txq;
+       }
 
        return skb_tx_hash(dev, skb);
 }
@@ -5576,30 +6067,32 @@ static netdev_tx_t ixgbe_xmit_frame(struct sk_buff *skb,
                tx_flags <<= IXGBE_TX_FLAGS_VLAN_SHIFT;
                tx_flags |= IXGBE_TX_FLAGS_VLAN;
        } else if (adapter->flags & IXGBE_FLAG_DCB_ENABLED) {
-               if (skb->priority != TC_PRIO_CONTROL) {
-                       tx_flags |= ((skb->queue_mapping & 0x7) << 13);
-                       tx_flags <<= IXGBE_TX_FLAGS_VLAN_SHIFT;
-                       tx_flags |= IXGBE_TX_FLAGS_VLAN;
-               } else {
-                       skb->queue_mapping =
-                               adapter->ring_feature[RING_F_DCB].indices-1;
-               }
+               tx_flags |= ((skb->queue_mapping & 0x7) << 13);
+               tx_flags <<= IXGBE_TX_FLAGS_VLAN_SHIFT;
+               tx_flags |= IXGBE_TX_FLAGS_VLAN;
        }
 
-       tx_ring = &adapter->tx_ring[skb->queue_mapping];
+       tx_ring = adapter->tx_ring[skb->queue_mapping];
 
-       if ((adapter->flags & IXGBE_FLAG_FCOE_ENABLED) &&
-           (skb->protocol == htons(ETH_P_FCOE))) {
-               tx_flags |= IXGBE_TX_FLAGS_FCOE;
 #ifdef IXGBE_FCOE
+       if (adapter->flags & IXGBE_FLAG_FCOE_ENABLED) {
 #ifdef CONFIG_IXGBE_DCB
-               tx_flags &= ~(IXGBE_TX_FLAGS_VLAN_PRIO_MASK
-                             << IXGBE_TX_FLAGS_VLAN_SHIFT);
-               tx_flags |= ((adapter->fcoe.up << 13)
-                             << IXGBE_TX_FLAGS_VLAN_SHIFT);
-#endif
+               /* for FCoE with DCB, we force the priority to what
+                * was specified by the switch */
+               if ((skb->protocol == htons(ETH_P_FCOE)) ||
+                   (skb->protocol == htons(ETH_P_FIP))) {
+                       tx_flags &= ~(IXGBE_TX_FLAGS_VLAN_PRIO_MASK
+                                     << IXGBE_TX_FLAGS_VLAN_SHIFT);
+                       tx_flags |= ((adapter->fcoe.up << 13)
+                                    << IXGBE_TX_FLAGS_VLAN_SHIFT);
+               }
 #endif
+               /* flag for FCoE offloads */
+               if (skb->protocol == htons(ETH_P_FCOE))
+                       tx_flags |= IXGBE_TX_FLAGS_FCOE;
        }
+#endif
+
        /* four things can cause us to need a context descriptor */
        if (skb_is_gso(skb) ||
            (skb->ip_summed == CHECKSUM_PARTIAL) ||
@@ -5818,6 +6311,10 @@ static const struct net_device_ops ixgbe_netdev_ops = {
        .ndo_vlan_rx_add_vid    = ixgbe_vlan_rx_add_vid,
        .ndo_vlan_rx_kill_vid   = ixgbe_vlan_rx_kill_vid,
        .ndo_do_ioctl           = ixgbe_ioctl,
+       .ndo_set_vf_mac         = ixgbe_ndo_set_vf_mac,
+       .ndo_set_vf_vlan        = ixgbe_ndo_set_vf_vlan,
+       .ndo_set_vf_tx_rate     = ixgbe_ndo_set_vf_bw,
+       .ndo_get_vf_config      = ixgbe_ndo_get_vf_config,
 #ifdef CONFIG_NET_POLL_CONTROLLER
        .ndo_poll_controller    = ixgbe_netpoll,
 #endif
@@ -5905,6 +6402,7 @@ static int __devinit ixgbe_probe(struct pci_dev *pdev,
        const struct ixgbe_info *ii = ixgbe_info_tbl[ent->driver_data];
        static int cards_found;
        int i, err, pci_using_dac;
+       unsigned int indices = num_possible_cpus();
 #ifdef IXGBE_FCOE
        u16 device_caps;
 #endif
@@ -5914,13 +6412,14 @@ static int __devinit ixgbe_probe(struct pci_dev *pdev,
        if (err)
                return err;
 
-       if (!pci_set_dma_mask(pdev, DMA_BIT_MASK(64)) &&
-           !pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(64))) {
+       if (!dma_set_mask(&pdev->dev, DMA_BIT_MASK(64)) &&
+           !dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(64))) {
                pci_using_dac = 1;
        } else {
-               err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32));
+               err = dma_set_mask(&pdev->dev, DMA_BIT_MASK(32));
                if (err) {
-                       err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32));
+                       err = dma_set_coherent_mask(&pdev->dev,
+                                                   DMA_BIT_MASK(32));
                        if (err) {
                                dev_err(&pdev->dev, "No usable DMA "
                                        "configuration, aborting\n");
@@ -5943,7 +6442,17 @@ static int __devinit ixgbe_probe(struct pci_dev *pdev,
        pci_set_master(pdev);
        pci_save_state(pdev);
 
-       netdev = alloc_etherdev_mq(sizeof(struct ixgbe_adapter), MAX_TX_QUEUES);
+       if (ii->mac == ixgbe_mac_82598EB)
+               indices = min_t(unsigned int, indices, IXGBE_MAX_RSS_INDICES);
+       else
+               indices = min_t(unsigned int, indices, IXGBE_MAX_FDIR_INDICES);
+
+       indices = max_t(unsigned int, indices, IXGBE_MAX_DCB_INDICES);
+#ifdef IXGBE_FCOE
+       indices += min_t(unsigned int, num_possible_cpus(),
+                        IXGBE_MAX_FCOE_INDICES);
+#endif
+       netdev = alloc_etherdev_mq(sizeof(struct ixgbe_adapter), indices);
        if (!netdev) {
                err = -ENOMEM;
                goto err_alloc_etherdev;
@@ -6024,6 +6533,10 @@ static int __devinit ixgbe_probe(struct pci_dev *pdev,
        if (err)
                goto err_sw_init;
 
+       /* Make it possible the adapter to be woken up via WOL */
+       if (adapter->hw.mac.type == ixgbe_mac_82599EB)
+               IXGBE_WRITE_REG(&adapter->hw, IXGBE_WUS, ~0);
+
        /*
         * If there is a fan on this device and it has failed log the
         * failure.
@@ -6122,6 +6635,10 @@ static int __devinit ixgbe_probe(struct pci_dev *pdev,
                goto err_eeprom;
        }
 
+       /* power down the optics */
+       if (hw->phy.multispeed_fiber)
+               hw->mac.ops.disable_tx_laser(hw);
+
        init_timer(&adapter->watchdog_timer);
        adapter->watchdog_timer.function = &ixgbe_watchdog;
        adapter->watchdog_timer.data = (unsigned long)adapter;
@@ -6137,9 +6654,6 @@ static int __devinit ixgbe_probe(struct pci_dev *pdev,
        case IXGBE_DEV_ID_82599_KX4:
                adapter->wol = (IXGBE_WUFC_MAG | IXGBE_WUFC_EX |
                                IXGBE_WUFC_MC | IXGBE_WUFC_BC);
-               /* Enable ACPI wakeup in GRC */
-               IXGBE_WRITE_REG(hw, IXGBE_GRC,
-                            (IXGBE_READ_REG(hw, IXGBE_GRC) & ~IXGBE_GRC_APME));
                break;
        default:
                adapter->wol = 0;