#include <linux/slab.h>
#include <linux/edac.h>
#include <linux/mmzone.h>
+#include <linux/edac_mce.h>
+#include <linux/spinlock.h>
#include "edac_core.h"
+/* To use the new pci_[read/write]_config_qword instead of two dword */
+#define USE_QWORD 0
/*
* Alter this version for the module when modifications are made
* i7core Memory Controller Registers
*/
+ /* OFFSETS for Device 0 Function 0 */
+
+#define MC_CFG_CONTROL 0x90
+
/* OFFSETS for Device 3 Function 0 */
#define MC_CONTROL 0x48
#define MC_STATUS 0x4c
#define MC_MAX_DOD 0x64
+/*
+ * OFFSETS for Device 3 Function 4, as inicated on Xeon 5500 datasheet:
+ * http://www.arrownac.com/manufacturers/intel/s/nehalem/5500-datasheet-v2.pdf
+ */
+
+#define MC_TEST_ERR_RCV1 0x60
+ #define DIMM2_COR_ERR(r) ((r) & 0x7fff)
+
+#define MC_TEST_ERR_RCV0 0x64
+ #define DIMM1_COR_ERR(r) (((r) >> 16) & 0x7fff)
+ #define DIMM0_COR_ERR(r) ((r) & 0x7fff)
+
/* OFFSETS for Devices 4,5 and 6 Function 0 */
+#define MC_CHANNEL_DIMM_INIT_PARAMS 0x58
+ #define THREE_DIMMS_PRESENT (1 << 24)
+ #define SINGLE_QUAD_RANK_PRESENT (1 << 23)
+ #define QUAD_RANK_PRESENT (1 << 22)
+ #define REGISTERED_DIMM (1 << 15)
+
+#define MC_CHANNEL_MAPPER 0x60
+ #define RDLCH(r, ch) ((((r) >> (3 + (ch * 6))) & 0x07) - 1)
+ #define WRLCH(r, ch) ((((r) >> (ch * 6)) & 0x07) - 1)
+
+#define MC_CHANNEL_RANK_PRESENT 0x7c
+ #define RANK_PRESENT_MASK 0xffff
+
#define MC_CHANNEL_ADDR_MATCH 0xf0
#define MC_CHANNEL_ERROR_MASK 0xf8
#define MC_CHANNEL_ERROR_INJECT 0xfc
#define NO_MASK_CACHELINE 0x00
#define REPEAT_EN 0x01
+ /* OFFSETS for Devices 4,5 and 6 Function 1 */
+#define MC_DOD_CH_DIMM0 0x48
+#define MC_DOD_CH_DIMM1 0x4c
+#define MC_DOD_CH_DIMM2 0x50
+ #define RANKOFFSET_MASK ((1 << 12) | (1 << 11) | (1 << 10))
+ #define RANKOFFSET(x) ((x & RANKOFFSET_MASK) >> 10)
+ #define DIMM_PRESENT_MASK (1 << 9)
+ #define DIMM_PRESENT(x) (((x) & DIMM_PRESENT_MASK) >> 9)
+ #define MC_DOD_NUMBANK_MASK ((1 << 8) | (1 << 7))
+ #define MC_DOD_NUMBANK(x) (((x) & MC_DOD_NUMBANK_MASK) >> 7)
+ #define MC_DOD_NUMRANK_MASK ((1 << 6) | (1 << 5))
+ #define MC_DOD_NUMRANK(x) (((x) & MC_DOD_NUMRANK_MASK) >> 5)
+ #define MC_DOD_NUMROW_MASK ((1 << 4) | (1 << 3) | (1 << 2))
+ #define MC_DOD_NUMROW(x) (((x) & MC_DOD_NUMROW_MASK) >> 2)
+ #define MC_DOD_NUMCOL_MASK 3
+ #define MC_DOD_NUMCOL(x) ((x) & MC_DOD_NUMCOL_MASK)
+
+#define MC_RANK_PRESENT 0x7c
+
+#define MC_SAG_CH_0 0x80
+#define MC_SAG_CH_1 0x84
+#define MC_SAG_CH_2 0x88
+#define MC_SAG_CH_3 0x8c
+#define MC_SAG_CH_4 0x90
+#define MC_SAG_CH_5 0x94
+#define MC_SAG_CH_6 0x98
+#define MC_SAG_CH_7 0x9c
+
+#define MC_RIR_LIMIT_CH_0 0x40
+#define MC_RIR_LIMIT_CH_1 0x44
+#define MC_RIR_LIMIT_CH_2 0x48
+#define MC_RIR_LIMIT_CH_3 0x4C
+#define MC_RIR_LIMIT_CH_4 0x50
+#define MC_RIR_LIMIT_CH_5 0x54
+#define MC_RIR_LIMIT_CH_6 0x58
+#define MC_RIR_LIMIT_CH_7 0x5C
+#define MC_RIR_LIMIT_MASK ((1 << 10) - 1)
+
+#define MC_RIR_WAY_CH 0x80
+ #define MC_RIR_WAY_OFFSET_MASK (((1 << 14) - 1) & ~0x7)
+ #define MC_RIR_WAY_RANK_MASK 0x7
+
/*
* i7core structs
*/
#define NUM_CHANS 3
-#define NUM_FUNCS 1
+#define MAX_DIMMS 3 /* Max DIMMS per channel */
+#define NUM_SOCKETS 2 /* Max number of MC sockets */
+#define MAX_MCR_FUNC 4
+#define MAX_CHAN_FUNC 3
struct i7core_info {
u32 mc_control;
u32 mc_status;
u32 max_dod;
+ u32 ch_map;
};
struct i7core_inject {
int enable;
+ u8 socket;
u32 section;
u32 type;
u32 eccmask;
int channel, dimm, rank, bank, page, col;
};
+struct i7core_channel {
+ u32 ranks;
+ u32 dimms;
+};
+
+struct pci_id_descr {
+ int dev;
+ int func;
+ int dev_id;
+ struct pci_dev *pdev[NUM_SOCKETS];
+};
+
struct i7core_pvt {
- struct pci_dev *pci_mcr; /* Dev 3:0 */
- struct pci_dev *pci_ch[NUM_CHANS][NUM_FUNCS];
+ struct pci_dev *pci_noncore[NUM_SOCKETS];
+ struct pci_dev *pci_mcr[NUM_SOCKETS][MAX_MCR_FUNC + 1];
+ struct pci_dev *pci_ch[NUM_SOCKETS][NUM_CHANS][MAX_CHAN_FUNC + 1];
+
struct i7core_info info;
struct i7core_inject inject;
+ struct i7core_channel channel[NUM_SOCKETS][NUM_CHANS];
+
+ int sockets; /* Number of sockets */
+ int channels; /* Number of active channels */
+
+ int ce_count_available[NUM_SOCKETS];
+ /* ECC corrected errors counts per dimm */
+ unsigned long ce_count[NUM_SOCKETS][MAX_DIMMS];
+ int last_ce_count[NUM_SOCKETS][MAX_DIMMS];
+
+ /* mcelog glue */
+ struct edac_mce edac_mce;
+ struct mce mce_entry[MCE_LOG_LEN];
+ unsigned mce_count;
+ spinlock_t mce_lock;
};
/* Device name and register DID (Device ID) */
u16 fsb_mapping_errors; /* DID for the branchmap,control */
};
-static int chan_pci_ids[NUM_CHANS] = {
- PCI_DEVICE_ID_INTEL_I7_MC_CH0_CTRL, /* Dev 4 */
- PCI_DEVICE_ID_INTEL_I7_MC_CH1_CTRL, /* Dev 5 */
- PCI_DEVICE_ID_INTEL_I7_MC_CH2_CTRL, /* Dev 6 */
+#define PCI_DESCR(device, function, device_id) \
+ .dev = (device), \
+ .func = (function), \
+ .dev_id = (device_id)
+
+struct pci_id_descr pci_devs[] = {
+ /* Memory controller */
+ { PCI_DESCR(3, 0, PCI_DEVICE_ID_INTEL_I7_MCR) },
+ { PCI_DESCR(3, 1, PCI_DEVICE_ID_INTEL_I7_MC_TAD) },
+ { PCI_DESCR(3, 2, PCI_DEVICE_ID_INTEL_I7_MC_RAS) }, /* if RDIMM is supported */
+ { PCI_DESCR(3, 4, PCI_DEVICE_ID_INTEL_I7_MC_TEST) },
+
+ /* Channel 0 */
+ { PCI_DESCR(4, 0, PCI_DEVICE_ID_INTEL_I7_MC_CH0_CTRL) },
+ { PCI_DESCR(4, 1, PCI_DEVICE_ID_INTEL_I7_MC_CH0_ADDR) },
+ { PCI_DESCR(4, 2, PCI_DEVICE_ID_INTEL_I7_MC_CH0_RANK) },
+ { PCI_DESCR(4, 3, PCI_DEVICE_ID_INTEL_I7_MC_CH0_TC) },
+
+ /* Channel 1 */
+ { PCI_DESCR(5, 0, PCI_DEVICE_ID_INTEL_I7_MC_CH1_CTRL) },
+ { PCI_DESCR(5, 1, PCI_DEVICE_ID_INTEL_I7_MC_CH1_ADDR) },
+ { PCI_DESCR(5, 2, PCI_DEVICE_ID_INTEL_I7_MC_CH1_RANK) },
+ { PCI_DESCR(5, 3, PCI_DEVICE_ID_INTEL_I7_MC_CH1_TC) },
+
+ /* Channel 2 */
+ { PCI_DESCR(6, 0, PCI_DEVICE_ID_INTEL_I7_MC_CH2_CTRL) },
+ { PCI_DESCR(6, 1, PCI_DEVICE_ID_INTEL_I7_MC_CH2_ADDR) },
+ { PCI_DESCR(6, 2, PCI_DEVICE_ID_INTEL_I7_MC_CH2_RANK) },
+ { PCI_DESCR(6, 3, PCI_DEVICE_ID_INTEL_I7_MC_CH2_TC) },
+
+ /* Generic Non-core registers */
+ /*
+ * This is the PCI device on i7core and on Xeon 35xx (8086:2c41)
+ * On Xeon 55xx, however, it has a different id (8086:2c40). So,
+ * the probing code needs to test for the other address in case of
+ * failure of this one
+ */
+ { PCI_DESCR(0, 0, PCI_DEVICE_ID_INTEL_I7_NOCORE) },
+
+};
+#define N_DEVS ARRAY_SIZE(pci_devs)
+
+/*
+ * pci_device_id table for which devices we are looking for
+ * This should match the first device at pci_devs table
+ */
+static const struct pci_device_id i7core_pci_tbl[] __devinitdata = {
+ {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_X58_HUB_MGMT)},
+ {0,} /* 0 terminated list. */
};
+
/* Table of devices attributes supported by this driver */
static const struct i7core_dev_info i7core_devs[] = {
{
****************************************************************************/
/* MC_CONTROL bits */
-#define CH2_ACTIVE(pvt) ((pvt)->info.mc_control & 1 << 10)
-#define CH1_ACTIVE(pvt) ((pvt)->info.mc_control & 1 << 9)
-#define CH0_ACTIVE(pvt) ((pvt)->info.mc_control & 1 << 8)
-#define ECCx8(pvt) ((pvt)->info.mc_control & 1 << 1)
+#define CH_ACTIVE(pvt, ch) ((pvt)->info.mc_control & (1 << (8 + ch)))
+#define ECCx8(pvt) ((pvt)->info.mc_control & (1 << 1))
/* MC_STATUS bits */
-#define ECC_ENABLED(pvt) ((pvt)->info.mc_status & 1 << 3)
-#define CH2_DISABLED(pvt) ((pvt)->info.mc_status & 1 << 2)
-#define CH1_DISABLED(pvt) ((pvt)->info.mc_status & 1 << 1)
-#define CH0_DISABLED(pvt) ((pvt)->info.mc_status & 1 << 0)
+#define ECC_ENABLED(pvt) ((pvt)->info.mc_status & (1 << 3))
+#define CH_DISABLED(pvt, ch) ((pvt)->info.mc_status & (1 << ch))
/* MC_MAX_DOD read functions */
-static inline int maxnumdimms(struct i7core_pvt *pvt)
+static inline int numdimms(u32 dimms)
{
- return (pvt->info.max_dod & 0x3) + 1;
+ return (dimms & 0x3) + 1;
}
-static inline int maxnumrank(struct i7core_pvt *pvt)
+static inline int numrank(u32 rank)
{
static int ranks[4] = { 1, 2, 4, -EINVAL };
- return ranks[(pvt->info.max_dod >> 2) & 0x3];
+ return ranks[rank & 0x3];
}
-static inline int maxnumbank(struct i7core_pvt *pvt)
+static inline int numbank(u32 bank)
{
static int banks[4] = { 4, 8, 16, -EINVAL };
- return banks[(pvt->info.max_dod >> 4) & 0x3];
+ return banks[bank & 0x3];
}
-static inline int maxnumrow(struct i7core_pvt *pvt)
+static inline int numrow(u32 row)
{
static int rows[8] = {
1 << 12, 1 << 13, 1 << 14, 1 << 15,
1 << 16, -EINVAL, -EINVAL, -EINVAL,
};
- return rows[((pvt->info.max_dod >> 6) & 0x7)];
+ return rows[row & 0x7];
}
-static inline int maxnumcol(struct i7core_pvt *pvt)
+static inline int numcol(u32 col)
{
static int cols[8] = {
1 << 10, 1 << 11, 1 << 12, -EINVAL,
};
- return cols[((pvt->info.max_dod >> 9) & 0x3) << 12];
+ return cols[col & 0x3];
}
/****************************************************************************
Memory check routines
****************************************************************************/
-static int get_dimm_config(struct mem_ctl_info *mci)
+static struct pci_dev *get_pdev_slot_func(u8 socket, unsigned slot,
+ unsigned func)
+{
+ int i;
+
+ for (i = 0; i < N_DEVS; i++) {
+ if (!pci_devs[i].pdev[socket])
+ continue;
+
+ if (PCI_SLOT(pci_devs[i].pdev[socket]->devfn) == slot &&
+ PCI_FUNC(pci_devs[i].pdev[socket]->devfn) == func) {
+ return pci_devs[i].pdev[socket];
+ }
+ }
+
+ return NULL;
+}
+
+static int i7core_get_active_channels(u8 socket, unsigned *channels,
+ unsigned *csrows)
+{
+ struct pci_dev *pdev = NULL;
+ int i, j;
+ u32 status, control;
+
+ *channels = 0;
+ *csrows = 0;
+
+ pdev = get_pdev_slot_func(socket, 3, 0);
+ if (!pdev) {
+ i7core_printk(KERN_ERR, "Couldn't find socket %d fn 3.0!!!\n",
+ socket);
+ return -ENODEV;
+ }
+
+ /* Device 3 function 0 reads */
+ pci_read_config_dword(pdev, MC_STATUS, &status);
+ pci_read_config_dword(pdev, MC_CONTROL, &control);
+
+ for (i = 0; i < NUM_CHANS; i++) {
+ u32 dimm_dod[3];
+ /* Check if the channel is active */
+ if (!(control & (1 << (8 + i))))
+ continue;
+
+ /* Check if the channel is disabled */
+ if (status & (1 << i))
+ continue;
+
+ pdev = get_pdev_slot_func(socket, i + 4, 1);
+ if (!pdev) {
+ i7core_printk(KERN_ERR, "Couldn't find socket %d "
+ "fn %d.%d!!!\n",
+ socket, i + 4, 1);
+ return -ENODEV;
+ }
+ /* Devices 4-6 function 1 */
+ pci_read_config_dword(pdev,
+ MC_DOD_CH_DIMM0, &dimm_dod[0]);
+ pci_read_config_dword(pdev,
+ MC_DOD_CH_DIMM1, &dimm_dod[1]);
+ pci_read_config_dword(pdev,
+ MC_DOD_CH_DIMM2, &dimm_dod[2]);
+
+ (*channels)++;
+
+ for (j = 0; j < 3; j++) {
+ if (!DIMM_PRESENT(dimm_dod[j]))
+ continue;
+ (*csrows)++;
+ }
+ }
+
+ debugf0("Number of active channels on socked %d: %d\n",
+ socket, *channels);
+
+ return 0;
+}
+
+static int get_dimm_config(struct mem_ctl_info *mci, int *csrow, u8 socket)
{
struct i7core_pvt *pvt = mci->pvt_info;
+ struct csrow_info *csr;
+ struct pci_dev *pdev;
+ int i, j;
+ unsigned long last_page = 0;
+ enum edac_type mode;
+ enum mem_type mtype;
+
+ /* Get data from the MC register, function 0 */
+ pdev = pvt->pci_mcr[socket][0];
+ if (!pdev)
+ return -ENODEV;
- pci_read_config_dword(pvt->pci_mcr, MC_CONTROL, &pvt->info.mc_control);
- pci_read_config_dword(pvt->pci_mcr, MC_STATUS, &pvt->info.mc_status);
- pci_read_config_dword(pvt->pci_mcr, MC_MAX_DOD, &pvt->info.max_dod);
+ /* Device 3 function 0 reads */
+ pci_read_config_dword(pdev, MC_CONTROL, &pvt->info.mc_control);
+ pci_read_config_dword(pdev, MC_STATUS, &pvt->info.mc_status);
+ pci_read_config_dword(pdev, MC_MAX_DOD, &pvt->info.max_dod);
+ pci_read_config_dword(pdev, MC_CHANNEL_MAPPER, &pvt->info.ch_map);
- debugf0("Channels active [%c][%c][%c] - enabled [%c][%c][%c]\n",
- CH0_ACTIVE(pvt)?'0':'-',
- CH1_ACTIVE(pvt)?'1':'-',
- CH2_ACTIVE(pvt)?'2':'-',
- CH0_DISABLED(pvt)?'-':'0',
- CH1_DISABLED(pvt)?'-':'1',
- CH2_DISABLED(pvt)?'-':'2');
+ debugf0("MC control=0x%08x status=0x%08x dod=0x%08x map=0x%08x\n",
+ pvt->info.mc_control, pvt->info.mc_status,
+ pvt->info.max_dod, pvt->info.ch_map);
- if (ECC_ENABLED(pvt))
- debugf0("ECC enabled with x%d SDCC\n", ECCx8(pvt)?8:4);
- else
+ if (ECC_ENABLED(pvt)) {
+ debugf0("ECC enabled with x%d SDCC\n", ECCx8(pvt) ? 8 : 4);
+ if (ECCx8(pvt))
+ mode = EDAC_S8ECD8ED;
+ else
+ mode = EDAC_S4ECD4ED;
+ } else {
debugf0("ECC disabled\n");
+ mode = EDAC_NONE;
+ }
/* FIXME: need to handle the error codes */
- debugf0("DOD Maximum limits: DIMMS: %d, %d-ranked, %d-banked\n",
- maxnumdimms(pvt), maxnumrank(pvt), maxnumbank(pvt));
- debugf0("DOD Maximum rows x colums = 0x%x x 0x%x\n",
- maxnumrow(pvt), maxnumcol(pvt));
+ debugf0("DOD Max limits: DIMMS: %d, %d-ranked, %d-banked\n",
+ numdimms(pvt->info.max_dod),
+ numrank(pvt->info.max_dod >> 2),
+ numbank(pvt->info.max_dod >> 4));
+ debugf0("DOD Max rows x colums = 0x%x x 0x%x\n",
+ numrow(pvt->info.max_dod >> 6),
+ numcol(pvt->info.max_dod >> 9));
+
+ debugf0("Memory channel configuration:\n");
+
+ for (i = 0; i < NUM_CHANS; i++) {
+ u32 data, dimm_dod[3], value[8];
+
+ if (!CH_ACTIVE(pvt, i)) {
+ debugf0("Channel %i is not active\n", i);
+ continue;
+ }
+ if (CH_DISABLED(pvt, i)) {
+ debugf0("Channel %i is disabled\n", i);
+ continue;
+ }
+
+ /* Devices 4-6 function 0 */
+ pci_read_config_dword(pvt->pci_ch[socket][i][0],
+ MC_CHANNEL_DIMM_INIT_PARAMS, &data);
+
+ pvt->channel[socket][i].ranks = (data & QUAD_RANK_PRESENT) ?
+ 4 : 2;
+
+ if (data & REGISTERED_DIMM)
+ mtype = MEM_RDDR3;
+ else
+ mtype = MEM_DDR3;
+#if 0
+ if (data & THREE_DIMMS_PRESENT)
+ pvt->channel[i].dimms = 3;
+ else if (data & SINGLE_QUAD_RANK_PRESENT)
+ pvt->channel[i].dimms = 1;
+ else
+ pvt->channel[i].dimms = 2;
+#endif
+
+ /* Devices 4-6 function 1 */
+ pci_read_config_dword(pvt->pci_ch[socket][i][1],
+ MC_DOD_CH_DIMM0, &dimm_dod[0]);
+ pci_read_config_dword(pvt->pci_ch[socket][i][1],
+ MC_DOD_CH_DIMM1, &dimm_dod[1]);
+ pci_read_config_dword(pvt->pci_ch[socket][i][1],
+ MC_DOD_CH_DIMM2, &dimm_dod[2]);
+
+ debugf0("Ch%d phy rd%d, wr%d (0x%08x): "
+ "%d ranks, %cDIMMs\n",
+ i,
+ RDLCH(pvt->info.ch_map, i), WRLCH(pvt->info.ch_map, i),
+ data,
+ pvt->channel[socket][i].ranks,
+ (data & REGISTERED_DIMM) ? 'R' : 'U');
+
+ for (j = 0; j < 3; j++) {
+ u32 banks, ranks, rows, cols;
+ u32 size, npages;
+
+ if (!DIMM_PRESENT(dimm_dod[j]))
+ continue;
+
+ banks = numbank(MC_DOD_NUMBANK(dimm_dod[j]));
+ ranks = numrank(MC_DOD_NUMRANK(dimm_dod[j]));
+ rows = numrow(MC_DOD_NUMROW(dimm_dod[j]));
+ cols = numcol(MC_DOD_NUMCOL(dimm_dod[j]));
+
+ /* DDR3 has 8 I/O banks */
+ size = (rows * cols * banks * ranks) >> (20 - 3);
+
+ pvt->channel[socket][i].dimms++;
+
+ debugf0("\tdimm %d (0x%08x) %d Mb offset: %x, "
+ "numbank: %d,\n\t\t"
+ "numrank: %d, numrow: %#x, numcol: %#x\n",
+ j, dimm_dod[j], size,
+ RANKOFFSET(dimm_dod[j]),
+ banks, ranks, rows, cols);
+
+#if PAGE_SHIFT > 20
+ npages = size >> (PAGE_SHIFT - 20);
+#else
+ npages = size << (20 - PAGE_SHIFT);
+#endif
+
+ csr = &mci->csrows[*csrow];
+ csr->first_page = last_page + 1;
+ last_page += npages;
+ csr->last_page = last_page;
+ csr->nr_pages = npages;
+
+ csr->page_mask = 0;
+ csr->grain = 8;
+ csr->csrow_idx = *csrow;
+ csr->nr_channels = 1;
+
+ csr->channels[0].chan_idx = i;
+ csr->channels[0].ce_count = 0;
+
+ switch (banks) {
+ case 4:
+ csr->dtype = DEV_X4;
+ break;
+ case 8:
+ csr->dtype = DEV_X8;
+ break;
+ case 16:
+ csr->dtype = DEV_X16;
+ break;
+ default:
+ csr->dtype = DEV_UNKNOWN;
+ }
+
+ csr->edac_mode = mode;
+ csr->mtype = mtype;
+
+ (*csrow)++;
+ }
+
+ pci_read_config_dword(pdev, MC_SAG_CH_0, &value[0]);
+ pci_read_config_dword(pdev, MC_SAG_CH_1, &value[1]);
+ pci_read_config_dword(pdev, MC_SAG_CH_2, &value[2]);
+ pci_read_config_dword(pdev, MC_SAG_CH_3, &value[3]);
+ pci_read_config_dword(pdev, MC_SAG_CH_4, &value[4]);
+ pci_read_config_dword(pdev, MC_SAG_CH_5, &value[5]);
+ pci_read_config_dword(pdev, MC_SAG_CH_6, &value[6]);
+ pci_read_config_dword(pdev, MC_SAG_CH_7, &value[7]);
+ debugf0("\t[%i] DIVBY3\tREMOVED\tOFFSET\n", i);
+ for (j = 0; j < 8; j++)
+ debugf0("\t\t%#x\t%#x\t%#x\n",
+ (value[j] >> 27) & 0x1,
+ (value[j] >> 24) & 0x7,
+ (value[j] && ((1 << 24) - 1)));
+ }
return 0;
}
we're disabling error injection on all write calls to the sysfs nodes that
controls the error code injection.
*/
-static void disable_inject(struct mem_ctl_info *mci)
+static int disable_inject(struct mem_ctl_info *mci)
{
struct i7core_pvt *pvt = mci->pvt_info;
pvt->inject.enable = 0;
- pci_write_config_dword(pvt->pci_ch[pvt->inject.channel][0],
+ if (!pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0])
+ return -ENODEV;
+
+ pci_write_config_dword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
MC_CHANNEL_ERROR_MASK, 0);
+
+ return 0;
+}
+
+/*
+ * i7core inject inject.socket
+ *
+ * accept and store error injection inject.socket value
+ */
+static ssize_t i7core_inject_socket_store(struct mem_ctl_info *mci,
+ const char *data, size_t count)
+{
+ struct i7core_pvt *pvt = mci->pvt_info;
+ unsigned long value;
+ int rc;
+
+ rc = strict_strtoul(data, 10, &value);
+ if ((rc < 0) || (value > pvt->sockets))
+ return 0;
+
+ pvt->inject.section = (u32) value;
+ return count;
+}
+
+static ssize_t i7core_inject_socket_show(struct mem_ctl_info *mci,
+ char *data)
+{
+ struct i7core_pvt *pvt = mci->pvt_info;
+ return sprintf(data, "%d\n", pvt->inject.socket);
}
/*
int rc;
if (pvt->inject.enable)
- disable_inject(mci);
+ disable_inject(mci);
rc = strict_strtoul(data, 10, &value);
if ((rc < 0) || (value > 3))
int rc;
if (pvt->inject.enable)
- disable_inject(mci);
+ disable_inject(mci);
rc = strict_strtoul(data, 10, &value);
if ((rc < 0) || (value > 7))
int rc;
if (pvt->inject.enable)
- disable_inject(mci);
+ disable_inject(mci);
rc = strict_strtoul(data, 10, &value);
if (rc < 0)
int rc;
if (pvt->inject.enable)
- disable_inject(mci);
+ disable_inject(mci);
do {
cmd = strsep((char **) &data, ":");
if (!val)
return cmd - data;
- if (!strcasecmp(val,"any"))
+ if (!strcasecmp(val, "any"))
value = -1;
else {
rc = strict_strtol(val, 10, &value);
return cmd - data;
}
- if (!strcasecmp(cmd,"channel")) {
+ if (!strcasecmp(cmd, "channel")) {
if (value < 3)
pvt->inject.channel = value;
else
return cmd - data;
- } else if (!strcasecmp(cmd,"dimm")) {
+ } else if (!strcasecmp(cmd, "dimm")) {
if (value < 4)
pvt->inject.dimm = value;
else
return cmd - data;
- } else if (!strcasecmp(cmd,"rank")) {
+ } else if (!strcasecmp(cmd, "rank")) {
if (value < 4)
pvt->inject.rank = value;
else
return cmd - data;
- } else if (!strcasecmp(cmd,"bank")) {
+ } else if (!strcasecmp(cmd, "bank")) {
if (value < 4)
pvt->inject.bank = value;
else
return cmd - data;
- } else if (!strcasecmp(cmd,"page")) {
+ } else if (!strcasecmp(cmd, "page")) {
if (value <= 0xffff)
pvt->inject.page = value;
else
return cmd - data;
- } else if (!strcasecmp(cmd,"col") ||
- !strcasecmp(cmd,"column")) {
+ } else if (!strcasecmp(cmd, "col") ||
+ !strcasecmp(cmd, "column")) {
if (value <= 0x3fff)
pvt->inject.col = value;
else
int rc;
long enable;
+ if (!pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0])
+ return 0;
+
rc = strict_strtoul(data, 10, &enable);
if ((rc < 0))
return 0;
/* Sets pvt->inject.dimm mask */
if (pvt->inject.dimm < 0)
- mask |= 1l << 41;
+ mask |= 1L << 41;
else {
- if (maxnumdimms(pvt) > 2)
- mask |= (pvt->inject.dimm & 0x3l) << 35;
+ if (pvt->channel[pvt->inject.socket][pvt->inject.channel].dimms > 2)
+ mask |= (pvt->inject.dimm & 0x3L) << 35;
else
- mask |= (pvt->inject.dimm & 0x1l) << 36;
+ mask |= (pvt->inject.dimm & 0x1L) << 36;
}
/* Sets pvt->inject.rank mask */
if (pvt->inject.rank < 0)
- mask |= 1l << 40;
+ mask |= 1L << 40;
else {
- if (maxnumdimms(pvt) > 2)
- mask |= (pvt->inject.rank & 0x1l) << 34;
+ if (pvt->channel[pvt->inject.socket][pvt->inject.channel].dimms > 2)
+ mask |= (pvt->inject.rank & 0x1L) << 34;
else
- mask |= (pvt->inject.rank & 0x3l) << 34;
+ mask |= (pvt->inject.rank & 0x3L) << 34;
}
/* Sets pvt->inject.bank mask */
if (pvt->inject.bank < 0)
- mask |= 1l << 39;
+ mask |= 1L << 39;
else
- mask |= (pvt->inject.bank & 0x15l) << 30;
+ mask |= (pvt->inject.bank & 0x15L) << 30;
/* Sets pvt->inject.page mask */
if (pvt->inject.page < 0)
- mask |= 1l << 38;
+ mask |= 1L << 38;
else
- mask |= (pvt->inject.page & 0xffffl) << 14;
+ mask |= (pvt->inject.page & 0xffffL) << 14;
/* Sets pvt->inject.column mask */
if (pvt->inject.col < 0)
- mask |= 1l << 37;
+ mask |= 1L << 37;
else
- mask |= (pvt->inject.col & 0x3fffl);
+ mask |= (pvt->inject.col & 0x3fffL);
- pci_write_config_qword(pvt->pci_ch[pvt->inject.channel][0],
- MC_CHANNEL_ADDR_MATCH, mask);
+ /* Unlock writes to registers */
+ pci_write_config_dword(pvt->pci_noncore[pvt->inject.socket],
+ MC_CFG_CONTROL, 0x2);
+ msleep(100);
- pci_write_config_dword(pvt->pci_ch[pvt->inject.channel][0],
+ /* Zeroes error count registers */
+ pci_write_config_dword(pvt->pci_mcr[pvt->inject.socket][4],
+ MC_TEST_ERR_RCV1, 0);
+ pci_write_config_dword(pvt->pci_mcr[pvt->inject.socket][4],
+ MC_TEST_ERR_RCV0, 0);
+ pvt->ce_count_available[pvt->inject.socket] = 0;
+
+
+#if USE_QWORD
+ pci_write_config_qword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
+ MC_CHANNEL_ADDR_MATCH, mask);
+#else
+ pci_write_config_dword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
+ MC_CHANNEL_ADDR_MATCH, mask);
+ pci_write_config_dword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
+ MC_CHANNEL_ADDR_MATCH + 4, mask >> 32L);
+#endif
+
+#if 1
+#if USE_QWORD
+ u64 rdmask;
+ pci_read_config_qword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
+ MC_CHANNEL_ADDR_MATCH, &rdmask);
+ debugf0("Inject addr match write 0x%016llx, read: 0x%016llx\n",
+ mask, rdmask);
+#else
+ u32 rdmask1, rdmask2;
+
+ pci_read_config_dword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
+ MC_CHANNEL_ADDR_MATCH, &rdmask1);
+ pci_read_config_dword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
+ MC_CHANNEL_ADDR_MATCH + 4, &rdmask2);
+
+ debugf0("Inject addr match write 0x%016llx, read: 0x%08x 0x%08x\n",
+ mask, rdmask1, rdmask2);
+#endif
+#endif
+
+ pci_write_config_dword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
MC_CHANNEL_ERROR_MASK, pvt->inject.eccmask);
/*
* bit 4: INJECT_ADDR_PARITY
*/
- injectmask = (pvt->inject.type & 1) &&
- (pvt->inject.section & 0x3) << 1 &&
+ injectmask = (pvt->inject.type & 1) |
+ (pvt->inject.section & 0x3) << 1 |
(pvt->inject.type & 0x6) << (3 - 1);
- pci_write_config_dword(pvt->pci_ch[pvt->inject.channel][0],
+ pci_write_config_dword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
MC_CHANNEL_ERROR_MASK, injectmask);
-
- debugf0("Error inject addr match 0x%016llx, ecc 0x%08x, inject 0x%08x\n",
+#if 0
+ /* lock writes to registers */
+ pci_write_config_dword(pvt->pci_noncore, MC_CFG_CONTROL, 0);
+#endif
+ debugf0("Error inject addr match 0x%016llx, ecc 0x%08x,"
+ " inject 0x%08x\n",
mask, pvt->inject.eccmask, injectmask);
+
return count;
}
char *data)
{
struct i7core_pvt *pvt = mci->pvt_info;
+ u32 injectmask;
+
+ pci_read_config_dword(pvt->pci_ch[pvt->inject.socket][pvt->inject.channel][0],
+ MC_CHANNEL_ERROR_MASK, &injectmask);
+
+ debugf0("Inject error read: 0x%018x\n", injectmask);
+
+ if (injectmask & 0x0c)
+ pvt->inject.enable = 1;
+
return sprintf(data, "%d\n", pvt->inject.enable);
}
+static ssize_t i7core_ce_regs_show(struct mem_ctl_info *mci, char *data)
+{
+ unsigned i, count, total = 0;
+ struct i7core_pvt *pvt = mci->pvt_info;
+
+ for (i = 0; i < pvt->sockets; i++) {
+ if (!pvt->ce_count_available[i])
+ count = sprintf(data, "socket 0 data unavailable\n");
+ else
+ count = sprintf(data, "socket %d, dimm0: %lu\n"
+ "dimm1: %lu\ndimm2: %lu\n",
+ i,
+ pvt->ce_count[i][0],
+ pvt->ce_count[i][1],
+ pvt->ce_count[i][2]);
+ data += count;
+ total += count;
+ }
+
+ return total;
+}
+
/*
* Sysfs struct
*/
static struct mcidev_sysfs_attribute i7core_inj_attrs[] = {
-
{
.attr = {
+ .name = "inject_socket",
+ .mode = (S_IRUGO | S_IWUSR)
+ },
+ .show = i7core_inject_socket_show,
+ .store = i7core_inject_socket_store,
+ }, {
+ .attr = {
.name = "inject_section",
.mode = (S_IRUGO | S_IWUSR)
},
},
.show = i7core_inject_enable_show,
.store = i7core_inject_enable_store,
+ }, {
+ .attr = {
+ .name = "corrected_error_counts",
+ .mode = (S_IRUGO | S_IWUSR)
+ },
+ .show = i7core_ce_regs_show,
+ .store = NULL,
},
};
* i7core_put_devices 'put' all the devices that we have
* reserved via 'get'
*/
-static void i7core_put_devices(struct mem_ctl_info *mci)
+static void i7core_put_devices(void)
{
- struct i7core_pvt *pvt = mci->pvt_info;
- int i, n;
+ int i, j;
- pci_dev_put(pvt->pci_mcr);
-
- /* Release all PCI device functions at MTR channel controllers */
- for (i = 0; i < NUM_CHANS; i++)
- for (n = 0; n < NUM_FUNCS; n++)
- pci_dev_put(pvt->pci_ch[i][n]);
+ for (i = 0; i < NUM_SOCKETS; i++)
+ for (j = 0; j < N_DEVS; j++)
+ pci_dev_put(pci_devs[j].pdev[i]);
}
/*
*
* Need to 'get' device 16 func 1 and func 2
*/
-static int i7core_get_devices(struct mem_ctl_info *mci, int dev_idx)
+static int i7core_get_devices(void)
{
- struct i7core_pvt *pvt;
- struct pci_dev *pdev;
- int i, n, func;
+ int rc, i;
+ struct pci_dev *pdev = NULL;
+ u8 bus = 0;
+ u8 socket = 0;
- pvt = mci->pvt_info;
- memset(pvt, 0, sizeof(*pvt));
+ for (i = 0; i < N_DEVS; i++) {
+ pdev = pci_get_device(PCI_VENDOR_ID_INTEL,
+ pci_devs[i].dev_id, NULL);
- pdev = pci_get_device(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_I7_MCR,
- NULL);
- if (!pdev) {
- i7core_printk(KERN_ERR,
- "Couldn't get PCI ID %04x:%04x function 0\n",
- PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_I7_MCR);
- return -ENODEV;
- }
- pvt->pci_mcr=pdev;
+ if (!pdev && !i) {
+ pcibios_scan_specific_bus(254);
+ pcibios_scan_specific_bus(255);
- /* Get dimm basic config */
- get_dimm_config(mci);
+ pdev = pci_get_device(PCI_VENDOR_ID_INTEL,
+ pci_devs[i].dev_id, NULL);
+ }
- /* Retrieve all needed functions at MTR channel controllers */
- for (i = 0; i < NUM_CHANS; i++) {
- pdev = NULL;
- for (n = 0; n < NUM_FUNCS; n++) {
+ /*
+ * On Xeon 55xx, the Intel Quckpath Arch Generic Non-core regs
+ * is at addr 8086:2c40, instead of 8086:2c41. So, we need
+ * to probe for the alternate address in case of failure
+ */
+ if (pci_devs[i].dev_id == PCI_DEVICE_ID_INTEL_I7_NOCORE
+ && !pdev)
pdev = pci_get_device(PCI_VENDOR_ID_INTEL,
- chan_pci_ids[i], pdev);
- if (!pdev) {
- /* End of list, leave */
+ PCI_DEVICE_ID_INTEL_I7_NOCORE_ALT, NULL);
+
+ if (likely(pdev)) {
+ bus = pdev->bus->number;
+
+ if (bus == 0x3f)
+ socket = 0;
+ else
+ socket = 255 - bus;
+
+ if (socket >= NUM_SOCKETS) {
i7core_printk(KERN_ERR,
- "Device not found: PCI ID %04x:%04x "
- "found only %d functions "
- "(broken BIOS?)\n",
- PCI_VENDOR_ID_INTEL,
- chan_pci_ids[i], n);
- i7core_put_devices(mci);
- return -ENODEV;
+ "Found unexpected socket for "
+ "dev %02x:%02x.%d PCI ID %04x:%04x\n",
+ bus, pci_devs[i].dev, pci_devs[i].func,
+ PCI_VENDOR_ID_INTEL, pci_devs[i].dev_id);
+
+ rc = -ENODEV;
+ goto error;
}
+
+ pci_devs[i].pdev[socket] = pdev;
+ } else {
+ i7core_printk(KERN_ERR,
+ "Device not found: "
+ "dev %02x:%02x.%d PCI ID %04x:%04x\n",
+ bus, pci_devs[i].dev, pci_devs[i].func,
+ PCI_VENDOR_ID_INTEL, pci_devs[i].dev_id);
+
+ /* Dev 3 function 2 only exists on chips with RDIMMs */
+ if ((pci_devs[i].dev == 3) && (pci_devs[i].func == 2))
+ continue;
+
+ /* End of list, leave */
+ rc = -ENODEV;
+ goto error;
+ }
+
+ /* Sanity check */
+ if (unlikely(PCI_SLOT(pdev->devfn) != pci_devs[i].dev ||
+ PCI_FUNC(pdev->devfn) != pci_devs[i].func)) {
+ i7core_printk(KERN_ERR,
+ "Device PCI ID %04x:%04x "
+ "has fn %d.%d instead of fn %d.%d\n",
+ PCI_VENDOR_ID_INTEL, pci_devs[i].dev_id,
+ PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn),
+ pci_devs[i].dev, pci_devs[i].func);
+ rc = -EINVAL;
+ goto error;
+ }
+
+ /* Be sure that the device is enabled */
+ rc = pci_enable_device(pdev);
+ if (unlikely(rc < 0)) {
+ i7core_printk(KERN_ERR,
+ "Couldn't enable PCI ID %04x:%04x "
+ "fn %d.%d\n",
+ PCI_VENDOR_ID_INTEL, pci_devs[i].dev_id,
+ PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn));
+ goto error;
+ }
+
+ i7core_printk(KERN_INFO,
+ "Registered socket %d "
+ "dev %02x:%02x.%d PCI ID %04x:%04x\n",
+ socket, bus, pci_devs[i].dev, pci_devs[i].func,
+ PCI_VENDOR_ID_INTEL, pci_devs[i].dev_id);
+ }
+
+ return 0;
+
+error:
+ i7core_put_devices();
+ return -EINVAL;
+}
+
+static int mci_bind_devs(struct mem_ctl_info *mci)
+{
+ struct i7core_pvt *pvt = mci->pvt_info;
+ struct pci_dev *pdev;
+ int i, j, func, slot;
+
+ for (i = 0; i < pvt->sockets; i++) {
+ for (j = 0; j < N_DEVS; j++) {
+ pdev = pci_devs[j].pdev[i];
+ if (!pdev)
+ continue;
+
func = PCI_FUNC(pdev->devfn);
- pvt->pci_ch[i][func] = pdev;
+ slot = PCI_SLOT(pdev->devfn);
+ if (slot == 3) {
+ if (unlikely(func > MAX_MCR_FUNC))
+ goto error;
+ pvt->pci_mcr[i][func] = pdev;
+ } else if (likely(slot >= 4 && slot < 4 + NUM_CHANS)) {
+ if (unlikely(func > MAX_CHAN_FUNC))
+ goto error;
+ pvt->pci_ch[i][slot - 4][func] = pdev;
+ } else if (!slot && !func)
+ pvt->pci_noncore[i] = pdev;
+ else
+ goto error;
+
+ debugf0("Associated fn %d.%d, dev = %p, socket %d\n",
+ PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn),
+ pdev, i);
}
}
- i7core_printk(KERN_INFO, "Driver loaded.\n");
return 0;
+
+error:
+ i7core_printk(KERN_ERR, "Device %d, function %d "
+ "is out of the expected range\n",
+ slot, func);
+ return -EINVAL;
+}
+
+/****************************************************************************
+ Error check routines
+ ****************************************************************************/
+
+/* This function is based on the device 3 function 4 registers as described on:
+ * Intel Xeon Processor 5500 Series Datasheet Volume 2
+ * http://www.intel.com/Assets/PDF/datasheet/321322.pdf
+ * also available at:
+ * http://www.arrownac.com/manufacturers/intel/s/nehalem/5500-datasheet-v2.pdf
+ */
+static void check_mc_test_err(struct mem_ctl_info *mci, u8 socket)
+{
+ struct i7core_pvt *pvt = mci->pvt_info;
+ u32 rcv1, rcv0;
+ int new0, new1, new2;
+
+ if (!pvt->pci_mcr[socket][4]) {
+ debugf0("%s MCR registers not found\n",__func__);
+ return;
+ }
+
+ /* Corrected error reads */
+ pci_read_config_dword(pvt->pci_mcr[socket][4], MC_TEST_ERR_RCV1, &rcv1);
+ pci_read_config_dword(pvt->pci_mcr[socket][4], MC_TEST_ERR_RCV0, &rcv0);
+
+ /* Store the new values */
+ new2 = DIMM2_COR_ERR(rcv1);
+ new1 = DIMM1_COR_ERR(rcv0);
+ new0 = DIMM0_COR_ERR(rcv0);
+
+#if 0
+ debugf2("%s CE rcv1=0x%08x rcv0=0x%08x, %d %d %d\n",
+ (pvt->ce_count_available ? "UPDATE" : "READ"),
+ rcv1, rcv0, new0, new1, new2);
+#endif
+
+ /* Updates CE counters if it is not the first time here */
+ if (pvt->ce_count_available[socket]) {
+ /* Updates CE counters */
+ int add0, add1, add2;
+
+ add2 = new2 - pvt->last_ce_count[socket][2];
+ add1 = new1 - pvt->last_ce_count[socket][1];
+ add0 = new0 - pvt->last_ce_count[socket][0];
+
+ if (add2 < 0)
+ add2 += 0x7fff;
+ pvt->ce_count[socket][2] += add2;
+
+ if (add1 < 0)
+ add1 += 0x7fff;
+ pvt->ce_count[socket][1] += add1;
+
+ if (add0 < 0)
+ add0 += 0x7fff;
+ pvt->ce_count[socket][0] += add0;
+ } else
+ pvt->ce_count_available[socket] = 1;
+
+ /* Store the new values */
+ pvt->last_ce_count[socket][2] = new2;
+ pvt->last_ce_count[socket][1] = new1;
+ pvt->last_ce_count[socket][0] = new0;
+}
+
+/*
+ * According with tables E-11 and E-12 of chapter E.3.3 of Intel 64 and IA-32
+ * Architectures Software Developer’s Manual Volume 3B.
+ * Nehalem are defined as family 0x06, model 0x1a
+ *
+ * The MCA registers used here are the following ones:
+ * struct mce field MCA Register
+ * m->status MSR_IA32_MC8_STATUS
+ * m->addr MSR_IA32_MC8_ADDR
+ * m->misc MSR_IA32_MC8_MISC
+ * In the case of Nehalem, the error information is masked at .status and .misc
+ * fields
+ */
+static void i7core_mce_output_error(struct mem_ctl_info *mci,
+ struct mce *m)
+{
+ char *type, *optype, *err, *msg;
+ unsigned long error = m->status & 0x1ff0000l;
+ u32 optypenum = (m->status >> 4) & 0x07;
+ u32 core_err_cnt = (m->status >> 38) && 0x7fff;
+ u32 dimm = (m->misc >> 16) & 0x3;
+ u32 channel = (m->misc >> 18) & 0x3;
+ u32 syndrome = m->misc >> 32;
+ u32 errnum = find_first_bit(&error, 32);
+
+ if (m->mcgstatus & 1)
+ type = "FATAL";
+ else
+ type = "NON_FATAL";
+
+ switch (optypenum) {
+ case 0:
+ optype = "generic undef request";
+ break;
+ case 1:
+ optype = "read error";
+ break;
+ case 2:
+ optype = "write error";
+ break;
+ case 3:
+ optype = "addr/cmd error";
+ break;
+ case 4:
+ optype = "scrubbing error";
+ break;
+ default:
+ optype = "reserved";
+ break;
+ }
+
+ switch (errnum) {
+ case 16:
+ err = "read ECC error";
+ break;
+ case 17:
+ err = "RAS ECC error";
+ break;
+ case 18:
+ err = "write parity error";
+ break;
+ case 19:
+ err = "redundacy loss";
+ break;
+ case 20:
+ err = "reserved";
+ break;
+ case 21:
+ err = "memory range error";
+ break;
+ case 22:
+ err = "RTID out of range";
+ break;
+ case 23:
+ err = "address parity error";
+ break;
+ case 24:
+ err = "byte enable parity error";
+ break;
+ default:
+ err = "unknown";
+ }
+
+ /* FIXME: should convert addr into bank and rank information */
+ msg = kasprintf(GFP_ATOMIC,
+ "%s (addr = 0x%08llx, Dimm=%d, Channel=%d, "
+ "syndrome=0x%08x, count=%d, Err=%08llx:%08llx (%s: %s))\n",
+ type, (long long) m->addr, dimm, channel,
+ syndrome, core_err_cnt, (long long)m->status,
+ (long long)m->misc, optype, err);
+
+ debugf0("%s", msg);
+
+ /* Call the helper to output message */
+ edac_mc_handle_fbd_ue(mci, 0 /* FIXME: should be rank here */,
+ 0, 0 /* FIXME: should be channel here */, msg);
+
+ kfree(msg);
+}
+
+/*
+ * i7core_check_error Retrieve and process errors reported by the
+ * hardware. Called by the Core module.
+ */
+static void i7core_check_error(struct mem_ctl_info *mci)
+{
+ struct i7core_pvt *pvt = mci->pvt_info;
+ int i;
+ unsigned count = 0;
+ struct mce *m = NULL;
+ unsigned long flags;
+
+ debugf0(__FILE__ ": %s()\n", __func__);
+
+ /* Copy all mce errors into a temporary buffer */
+ spin_lock_irqsave(&pvt->mce_lock, flags);
+ if (pvt->mce_count) {
+ m = kmalloc(sizeof(*m) * pvt->mce_count, GFP_ATOMIC);
+ if (m) {
+ count = pvt->mce_count;
+ memcpy(m, &pvt->mce_entry, sizeof(*m) * count);
+ }
+ pvt->mce_count = 0;
+ }
+ spin_unlock_irqrestore(&pvt->mce_lock, flags);
+
+ /* proccess mcelog errors */
+ for (i = 0; i < count; i++)
+ i7core_mce_output_error(mci, &m[i]);
+
+ kfree(m);
+
+ /* check memory count errors */
+ for (i = 0; i < pvt->sockets; i++)
+ check_mc_test_err(mci, i);
+}
+
+/*
+ * i7core_mce_check_error Replicates mcelog routine to get errors
+ * This routine simply queues mcelog errors, and
+ * return. The error itself should be handled later
+ * by i7core_check_error.
+ */
+static int i7core_mce_check_error(void *priv, struct mce *mce)
+{
+ struct mem_ctl_info *mci = priv;
+ struct i7core_pvt *pvt = mci->pvt_info;
+ unsigned long flags;
+
+ debugf0(__FILE__ ": %s()\n", __func__);
+
+ /*
+ * Just let mcelog handle it if the error is
+ * outside the memory controller
+ */
+ if (((mce->status & 0xffff) >> 7) != 1)
+ return 0;
+
+ /* Bank 8 registers are the only ones that we know how to handle */
+ if (mce->bank != 8)
+ return 0;
+
+ spin_lock_irqsave(&pvt->mce_lock, flags);
+ if (pvt->mce_count < MCE_LOG_LEN) {
+ memcpy(&pvt->mce_entry[pvt->mce_count], mce, sizeof(*mce));
+ pvt->mce_count++;
+ }
+ spin_unlock_irqrestore(&pvt->mce_lock, flags);
+
+ /* Handle fatal errors immediately */
+ if (mce->mcgstatus & 1)
+ i7core_check_error(mci);
+
+ /* Advice mcelog that the error were handled */
+ return 1;
}
/*
{
struct mem_ctl_info *mci;
struct i7core_pvt *pvt;
- int rc;
- int num_channels;
- int num_csrows;
- int num_dimms_per_channel;
+ int num_channels = 0;
+ int num_csrows = 0;
+ int csrow = 0;
int dev_idx = id->driver_data;
+ int rc, i;
+ u8 sockets;
- if (dev_idx >= ARRAY_SIZE(i7core_devs))
+ if (unlikely(dev_idx >= ARRAY_SIZE(i7core_devs)))
return -EINVAL;
- /* wake up device */
- rc = pci_enable_device(pdev);
- if (rc == -EIO)
+ /* get the pci devices we want to reserve for our use */
+ rc = i7core_get_devices();
+ if (unlikely(rc < 0))
return rc;
- debugf0("MC: " __FILE__ ": %s(), pdev bus %u dev=0x%x fn=0x%x\n",
- __func__,
- pdev->bus->number,
- PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn));
+ sockets = 1;
+ for (i = NUM_SOCKETS - 1; i > 0; i--)
+ if (pci_devs[0].pdev[i]) {
+ sockets = i + 1;
+ break;
+ }
- /* We only are looking for func 0 of the set */
- if (PCI_FUNC(pdev->devfn) != 0)
- return -ENODEV;
+ for (i = 0; i < sockets; i++) {
+ int channels;
+ int csrows;
- num_channels = NUM_CHANS;
+ /* Check the number of active and not disabled channels */
+ rc = i7core_get_active_channels(i, &channels, &csrows);
+ if (unlikely(rc < 0))
+ goto fail0;
- /* FIXME: FAKE data, since we currently don't now how to get this */
- num_dimms_per_channel = 4;
- num_csrows = num_dimms_per_channel;
+ num_channels += channels;
+ num_csrows += csrows;
+ }
/* allocate a new MC control structure */
mci = edac_mc_alloc(sizeof(*pvt), num_csrows, num_channels, 0);
- if (mci == NULL)
- return -ENOMEM;
+ if (unlikely(!mci)) {
+ rc = -ENOMEM;
+ goto fail0;
+ }
debugf0("MC: " __FILE__ ": %s(): mci = %p\n", __func__, mci);
mci->dev = &pdev->dev; /* record ptr to the generic device */
- dev_set_drvdata(mci->dev, mci);
-
pvt = mci->pvt_info;
-
-// pvt->system_address = pdev; /* Record this device in our private */
-// pvt->maxch = num_channels;
-// pvt->maxdimmperch = num_dimms_per_channel;
-
- /* 'get' the pci devices we want to reserve for our use */
- if (i7core_get_devices(mci, dev_idx))
- goto fail0;
-
+ memset(pvt, 0, sizeof(*pvt));
+ pvt->sockets = sockets;
mci->mc_idx = 0;
- mci->mtype_cap = MEM_FLAG_FB_DDR2; /* FIXME: it uses DDR3 */
+
+ /*
+ * FIXME: how to handle RDDR3 at MCI level? It is possible to have
+ * Mixed RDDR3/UDDR3 with Nehalem, provided that they are on different
+ * memory channels
+ */
+ mci->mtype_cap = MEM_FLAG_DDR3;
mci->edac_ctl_cap = EDAC_FLAG_NONE;
mci->edac_cap = EDAC_FLAG_NONE;
mci->mod_name = "i7core_edac.c";
mci->dev_name = pci_name(pdev);
mci->ctl_page_to_phys = NULL;
mci->mc_driver_sysfs_attributes = i7core_inj_attrs;
+ /* Set the function pointer to an actual operation function */
+ mci->edac_check = i7core_check_error;
+
+ /* Store pci devices at mci for faster access */
+ rc = mci_bind_devs(mci);
+ if (unlikely(rc < 0))
+ goto fail1;
+
+ /* Get dimm basic config */
+ for (i = 0; i < sockets; i++)
+ get_dimm_config(mci, &csrow, i);
/* add this new MC control structure to EDAC's list of MCs */
- if (edac_mc_add_mc(mci)) {
+ if (unlikely(edac_mc_add_mc(mci))) {
debugf0("MC: " __FILE__
": %s(): failed edac_mc_add_mc()\n", __func__);
/* FIXME: perhaps some code should go here that disables error
* reporting if we just enabled it
*/
+
+ rc = -EINVAL;
goto fail1;
}
/* allocating generic PCI control info */
i7core_pci = edac_pci_create_generic_ctl(&pdev->dev, EDAC_MOD_STR);
- if (!i7core_pci) {
+ if (unlikely(!i7core_pci)) {
printk(KERN_WARNING
"%s(): Unable to create PCI control\n",
__func__);
}
/* Default error mask is any memory */
- pvt->inject.channel = -1;
+ pvt->inject.channel = 0;
pvt->inject.dimm = -1;
pvt->inject.rank = -1;
pvt->inject.bank = -1;
pvt->inject.page = -1;
pvt->inject.col = -1;
+ /* Registers on edac_mce in order to receive memory errors */
+ pvt->edac_mce.priv = mci;
+ pvt->edac_mce.check_error = i7core_mce_check_error;
+ spin_lock_init(&pvt->mce_lock);
+
+ rc = edac_mce_register(&pvt->edac_mce);
+ if (unlikely (rc < 0)) {
+ debugf0("MC: " __FILE__
+ ": %s(): failed edac_mce_register()\n", __func__);
+ goto fail1;
+ }
+
+ i7core_printk(KERN_INFO, "Driver loaded.\n");
+
return 0;
fail1:
- i7core_put_devices(mci);
+ edac_mc_free(mci);
fail0:
- edac_mc_free(mci);
- return -ENODEV;
+ i7core_put_devices();
+ return rc;
}
/*
static void __devexit i7core_remove(struct pci_dev *pdev)
{
struct mem_ctl_info *mci;
+ struct i7core_pvt *pvt;
debugf0(__FILE__ ": %s()\n", __func__);
if (i7core_pci)
edac_pci_release_generic_ctl(i7core_pci);
+
mci = edac_mc_del_mc(&pdev->dev);
if (!mci)
return;
+ /* Unregisters on edac_mce in order to receive memory errors */
+ pvt = mci->pvt_info;
+ edac_mce_unregister(&pvt->edac_mce);
+
/* retrieve references to resources, and free those resources */
- i7core_put_devices(mci);
+ i7core_put_devices();
edac_mc_free(mci);
}
-/*
- * pci_device_id table for which devices we are looking for
- *
- * The "E500P" device is the first device supported.
- */
-static const struct pci_device_id i7core_pci_tbl[] __devinitdata = {
- {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_I7_MCR)},
- {0,} /* 0 terminated list. */
-};
-
MODULE_DEVICE_TABLE(pci, i7core_pci_tbl);
/*