123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596159715981599160016011602160316041605160616071608160916101611161216131614161516161617161816191620162116221623162416251626162716281629163016311632163316341635163616371638163916401641164216431644164516461647164816491650165116521653165416551656165716581659166016611662166316641665166616671668166916701671167216731674167516761677167816791680168116821683168416851686168716881689169016911692169316941695169616971698169917001701170217031704170517061707170817091710171117121713171417151716171717181719172017211722172317241725172617271728172917301731173217331734173517361737173817391740174117421743174417451746174717481749175017511752175317541755175617571758175917601761176217631764176517661767176817691770177117721773177417751776177717781779178017811782178317841785178617871788178917901791179217931794179517961797179817991800180118021803180418051806180718081809181018111812181318141815181618171818181918201821182218231824182518261827182818291830183118321833183418351836183718381839184018411842184318441845184618471848184918501851185218531854185518561857185818591860186118621863186418651866186718681869187018711872187318741875187618771878187918801881188218831884188518861887188818891890189118921893189418951896189718981899190019011902190319041905190619071908190919101911191219131914191519161917191819191920192119221923192419251926192719281929193019311932193319341935193619371938193919401941194219431944194519461947194819491950195119521953195419551956195719581959196019611962196319641965196619671968196919701971197219731974197519761977197819791980198119821983198419851986198719881989199019911992199319941995199619971998199920002001200220032004200520062007200820092010201120122013201420152016201720182019202020212022202320242025202620272028202920302031203220332034203520362037203820392040204120422043204420452046204720482049205020512052205320542055205620572058205920602061206220632064206520662067206820692070207120722073207420752076207720782079208020812082208320842085208620872088208920902091209220932094209520962097209820992100210121022103210421052106210721082109211021112112211321142115211621172118211921202121212221232124212521262127212821292130213121322133213421352136213721382139214021412142214321442145214621472148214921502151215221532154215521562157215821592160216121622163216421652166216721682169217021712172217321742175217621772178217921802181218221832184218521862187218821892190219121922193219421952196219721982199220022012202220322042205220622072208220922102211221222132214221522162217221822192220222122222223222422252226222722282229223022312232223322342235223622372238223922402241224222432244224522462247224822492250225122522253225422552256225722582259226022612262226322642265226622672268226922702271227222732274227522762277227822792280228122822283228422852286228722882289229022912292229322942295229622972298229923002301230223032304230523062307230823092310231123122313231423152316231723182319232023212322232323242325232623272328232923302331233223332334233523362337233823392340234123422343234423452346234723482349235023512352235323542355235623572358235923602361236223632364236523662367236823692370237123722373237423752376237723782379238023812382238323842385238623872388238923902391239223932394239523962397239823992400240124022403240424052406240724082409241024112412241324142415241624172418241924202421242224232424242524262427242824292430243124322433243424352436243724382439244024412442244324442445244624472448244924502451245224532454245524562457245824592460246124622463246424652466 |
- /* Intel i7 core/Nehalem Memory Controller kernel module
- *
- * This driver supports the memory controllers found on the Intel
- * processor families i7core, i7core 7xx/8xx, i5core, Xeon 35xx,
- * Xeon 55xx and Xeon 56xx also known as Nehalem, Nehalem-EP, Lynnfield
- * and Westmere-EP.
- *
- * This file may be distributed under the terms of the
- * GNU General Public License version 2 only.
- *
- * Copyright (c) 2009-2010 by:
- * Mauro Carvalho Chehab
- *
- * Red Hat Inc. http://www.redhat.com
- *
- * Forked and adapted from the i5400_edac driver
- *
- * Based on the following public Intel datasheets:
- * Intel Core i7 Processor Extreme Edition and Intel Core i7 Processor
- * Datasheet, Volume 2:
- * http://download.intel.com/design/processor/datashts/320835.pdf
- * Intel Xeon Processor 5500 Series Datasheet Volume 2
- * http://www.intel.com/Assets/PDF/datasheet/321322.pdf
- * also available at:
- * http://www.arrownac.com/manufacturers/intel/s/nehalem/5500-datasheet-v2.pdf
- */
- #include <linux/module.h>
- #include <linux/init.h>
- #include <linux/pci.h>
- #include <linux/pci_ids.h>
- #include <linux/slab.h>
- #include <linux/delay.h>
- #include <linux/dmi.h>
- #include <linux/edac.h>
- #include <linux/mmzone.h>
- #include <linux/smp.h>
- #include <asm/mce.h>
- #include <asm/processor.h>
- #include <asm/div64.h>
- #include "edac_core.h"
- /* Static vars */
- static LIST_HEAD(i7core_edac_list);
- static DEFINE_MUTEX(i7core_edac_lock);
- static int probed;
- static int use_pci_fixup;
- module_param(use_pci_fixup, int, 0444);
- MODULE_PARM_DESC(use_pci_fixup, "Enable PCI fixup to seek for hidden devices");
- /*
- * This is used for Nehalem-EP and Nehalem-EX devices, where the non-core
- * registers start at bus 255, and are not reported by BIOS.
- * We currently find devices with only 2 sockets. In order to support more QPI
- * Quick Path Interconnect, just increment this number.
- */
- #define MAX_SOCKET_BUSES 2
- /*
- * Alter this version for the module when modifications are made
- */
- #define I7CORE_REVISION " Ver: 1.0.0"
- #define EDAC_MOD_STR "i7core_edac"
- /*
- * Debug macros
- */
- #define i7core_printk(level, fmt, arg...) \
- edac_printk(level, "i7core", fmt, ##arg)
- #define i7core_mc_printk(mci, level, fmt, arg...) \
- edac_mc_chipset_printk(mci, level, "i7core", fmt, ##arg)
- /*
- * i7core Memory Controller Registers
- */
- /* OFFSETS for Device 0 Function 0 */
- #define MC_CFG_CONTROL 0x90
- #define MC_CFG_UNLOCK 0x02
- #define MC_CFG_LOCK 0x00
- /* OFFSETS for Device 3 Function 0 */
- #define MC_CONTROL 0x48
- #define MC_STATUS 0x4c
- #define MC_MAX_DOD 0x64
- /*
- * OFFSETS for Device 3 Function 4, as indicated on Xeon 5500 datasheet:
- * http://www.arrownac.com/manufacturers/intel/s/nehalem/5500-datasheet-v2.pdf
- */
- #define MC_TEST_ERR_RCV1 0x60
- #define DIMM2_COR_ERR(r) ((r) & 0x7fff)
- #define MC_TEST_ERR_RCV0 0x64
- #define DIMM1_COR_ERR(r) (((r) >> 16) & 0x7fff)
- #define DIMM0_COR_ERR(r) ((r) & 0x7fff)
- /* OFFSETS for Device 3 Function 2, as indicated on Xeon 5500 datasheet */
- #define MC_SSRCONTROL 0x48
- #define SSR_MODE_DISABLE 0x00
- #define SSR_MODE_ENABLE 0x01
- #define SSR_MODE_MASK 0x03
- #define MC_SCRUB_CONTROL 0x4c
- #define STARTSCRUB (1 << 24)
- #define SCRUBINTERVAL_MASK 0xffffff
- #define MC_COR_ECC_CNT_0 0x80
- #define MC_COR_ECC_CNT_1 0x84
- #define MC_COR_ECC_CNT_2 0x88
- #define MC_COR_ECC_CNT_3 0x8c
- #define MC_COR_ECC_CNT_4 0x90
- #define MC_COR_ECC_CNT_5 0x94
- #define DIMM_TOP_COR_ERR(r) (((r) >> 16) & 0x7fff)
- #define DIMM_BOT_COR_ERR(r) ((r) & 0x7fff)
- /* OFFSETS for Devices 4,5 and 6 Function 0 */
- #define MC_CHANNEL_DIMM_INIT_PARAMS 0x58
- #define THREE_DIMMS_PRESENT (1 << 24)
- #define SINGLE_QUAD_RANK_PRESENT (1 << 23)
- #define QUAD_RANK_PRESENT (1 << 22)
- #define REGISTERED_DIMM (1 << 15)
- #define MC_CHANNEL_MAPPER 0x60
- #define RDLCH(r, ch) ((((r) >> (3 + (ch * 6))) & 0x07) - 1)
- #define WRLCH(r, ch) ((((r) >> (ch * 6)) & 0x07) - 1)
- #define MC_CHANNEL_RANK_PRESENT 0x7c
- #define RANK_PRESENT_MASK 0xffff
- #define MC_CHANNEL_ADDR_MATCH 0xf0
- #define MC_CHANNEL_ERROR_MASK 0xf8
- #define MC_CHANNEL_ERROR_INJECT 0xfc
- #define INJECT_ADDR_PARITY 0x10
- #define INJECT_ECC 0x08
- #define MASK_CACHELINE 0x06
- #define MASK_FULL_CACHELINE 0x06
- #define MASK_MSB32_CACHELINE 0x04
- #define MASK_LSB32_CACHELINE 0x02
- #define NO_MASK_CACHELINE 0x00
- #define REPEAT_EN 0x01
- /* OFFSETS for Devices 4,5 and 6 Function 1 */
- #define MC_DOD_CH_DIMM0 0x48
- #define MC_DOD_CH_DIMM1 0x4c
- #define MC_DOD_CH_DIMM2 0x50
- #define RANKOFFSET_MASK ((1 << 12) | (1 << 11) | (1 << 10))
- #define RANKOFFSET(x) ((x & RANKOFFSET_MASK) >> 10)
- #define DIMM_PRESENT_MASK (1 << 9)
- #define DIMM_PRESENT(x) (((x) & DIMM_PRESENT_MASK) >> 9)
- #define MC_DOD_NUMBANK_MASK ((1 << 8) | (1 << 7))
- #define MC_DOD_NUMBANK(x) (((x) & MC_DOD_NUMBANK_MASK) >> 7)
- #define MC_DOD_NUMRANK_MASK ((1 << 6) | (1 << 5))
- #define MC_DOD_NUMRANK(x) (((x) & MC_DOD_NUMRANK_MASK) >> 5)
- #define MC_DOD_NUMROW_MASK ((1 << 4) | (1 << 3) | (1 << 2))
- #define MC_DOD_NUMROW(x) (((x) & MC_DOD_NUMROW_MASK) >> 2)
- #define MC_DOD_NUMCOL_MASK 3
- #define MC_DOD_NUMCOL(x) ((x) & MC_DOD_NUMCOL_MASK)
- #define MC_RANK_PRESENT 0x7c
- #define MC_SAG_CH_0 0x80
- #define MC_SAG_CH_1 0x84
- #define MC_SAG_CH_2 0x88
- #define MC_SAG_CH_3 0x8c
- #define MC_SAG_CH_4 0x90
- #define MC_SAG_CH_5 0x94
- #define MC_SAG_CH_6 0x98
- #define MC_SAG_CH_7 0x9c
- #define MC_RIR_LIMIT_CH_0 0x40
- #define MC_RIR_LIMIT_CH_1 0x44
- #define MC_RIR_LIMIT_CH_2 0x48
- #define MC_RIR_LIMIT_CH_3 0x4C
- #define MC_RIR_LIMIT_CH_4 0x50
- #define MC_RIR_LIMIT_CH_5 0x54
- #define MC_RIR_LIMIT_CH_6 0x58
- #define MC_RIR_LIMIT_CH_7 0x5C
- #define MC_RIR_LIMIT_MASK ((1 << 10) - 1)
- #define MC_RIR_WAY_CH 0x80
- #define MC_RIR_WAY_OFFSET_MASK (((1 << 14) - 1) & ~0x7)
- #define MC_RIR_WAY_RANK_MASK 0x7
- /*
- * i7core structs
- */
- #define NUM_CHANS 3
- #define MAX_DIMMS 3 /* Max DIMMS per channel */
- #define MAX_MCR_FUNC 4
- #define MAX_CHAN_FUNC 3
- struct i7core_info {
- u32 mc_control;
- u32 mc_status;
- u32 max_dod;
- u32 ch_map;
- };
- struct i7core_inject {
- int enable;
- u32 section;
- u32 type;
- u32 eccmask;
- /* Error address mask */
- int channel, dimm, rank, bank, page, col;
- };
- struct i7core_channel {
- bool is_3dimms_present;
- bool is_single_4rank;
- bool has_4rank;
- u32 dimms;
- };
- struct pci_id_descr {
- int dev;
- int func;
- int dev_id;
- int optional;
- };
- struct pci_id_table {
- const struct pci_id_descr *descr;
- int n_devs;
- };
- struct i7core_dev {
- struct list_head list;
- u8 socket;
- struct pci_dev **pdev;
- int n_devs;
- struct mem_ctl_info *mci;
- };
- struct i7core_pvt {
- struct device *addrmatch_dev, *chancounts_dev;
- struct pci_dev *pci_noncore;
- struct pci_dev *pci_mcr[MAX_MCR_FUNC + 1];
- struct pci_dev *pci_ch[NUM_CHANS][MAX_CHAN_FUNC + 1];
- struct i7core_dev *i7core_dev;
- struct i7core_info info;
- struct i7core_inject inject;
- struct i7core_channel channel[NUM_CHANS];
- int ce_count_available;
- /* ECC corrected errors counts per udimm */
- unsigned long udimm_ce_count[MAX_DIMMS];
- int udimm_last_ce_count[MAX_DIMMS];
- /* ECC corrected errors counts per rdimm */
- unsigned long rdimm_ce_count[NUM_CHANS][MAX_DIMMS];
- int rdimm_last_ce_count[NUM_CHANS][MAX_DIMMS];
- bool is_registered, enable_scrub;
- /* Fifo double buffers */
- struct mce mce_entry[MCE_LOG_LEN];
- struct mce mce_outentry[MCE_LOG_LEN];
- /* Fifo in/out counters */
- unsigned mce_in, mce_out;
- /* Count indicator to show errors not got */
- unsigned mce_overrun;
- /* DCLK Frequency used for computing scrub rate */
- int dclk_freq;
- /* Struct to control EDAC polling */
- struct edac_pci_ctl_info *i7core_pci;
- };
- #define PCI_DESCR(device, function, device_id) \
- .dev = (device), \
- .func = (function), \
- .dev_id = (device_id)
- static const struct pci_id_descr pci_dev_descr_i7core_nehalem[] = {
- /* Memory controller */
- { PCI_DESCR(3, 0, PCI_DEVICE_ID_INTEL_I7_MCR) },
- { PCI_DESCR(3, 1, PCI_DEVICE_ID_INTEL_I7_MC_TAD) },
- /* Exists only for RDIMM */
- { PCI_DESCR(3, 2, PCI_DEVICE_ID_INTEL_I7_MC_RAS), .optional = 1 },
- { PCI_DESCR(3, 4, PCI_DEVICE_ID_INTEL_I7_MC_TEST) },
- /* Channel 0 */
- { PCI_DESCR(4, 0, PCI_DEVICE_ID_INTEL_I7_MC_CH0_CTRL) },
- { PCI_DESCR(4, 1, PCI_DEVICE_ID_INTEL_I7_MC_CH0_ADDR) },
- { PCI_DESCR(4, 2, PCI_DEVICE_ID_INTEL_I7_MC_CH0_RANK) },
- { PCI_DESCR(4, 3, PCI_DEVICE_ID_INTEL_I7_MC_CH0_TC) },
- /* Channel 1 */
- { PCI_DESCR(5, 0, PCI_DEVICE_ID_INTEL_I7_MC_CH1_CTRL) },
- { PCI_DESCR(5, 1, PCI_DEVICE_ID_INTEL_I7_MC_CH1_ADDR) },
- { PCI_DESCR(5, 2, PCI_DEVICE_ID_INTEL_I7_MC_CH1_RANK) },
- { PCI_DESCR(5, 3, PCI_DEVICE_ID_INTEL_I7_MC_CH1_TC) },
- /* Channel 2 */
- { PCI_DESCR(6, 0, PCI_DEVICE_ID_INTEL_I7_MC_CH2_CTRL) },
- { PCI_DESCR(6, 1, PCI_DEVICE_ID_INTEL_I7_MC_CH2_ADDR) },
- { PCI_DESCR(6, 2, PCI_DEVICE_ID_INTEL_I7_MC_CH2_RANK) },
- { PCI_DESCR(6, 3, PCI_DEVICE_ID_INTEL_I7_MC_CH2_TC) },
- /* Generic Non-core registers */
- /*
- * This is the PCI device on i7core and on Xeon 35xx (8086:2c41)
- * On Xeon 55xx, however, it has a different id (8086:2c40). So,
- * the probing code needs to test for the other address in case of
- * failure of this one
- */
- { PCI_DESCR(0, 0, PCI_DEVICE_ID_INTEL_I7_NONCORE) },
- };
- static const struct pci_id_descr pci_dev_descr_lynnfield[] = {
- { PCI_DESCR( 3, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MCR) },
- { PCI_DESCR( 3, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_TAD) },
- { PCI_DESCR( 3, 4, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_TEST) },
- { PCI_DESCR( 4, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_CTRL) },
- { PCI_DESCR( 4, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_ADDR) },
- { PCI_DESCR( 4, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_RANK) },
- { PCI_DESCR( 4, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_TC) },
- { PCI_DESCR( 5, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_CTRL) },
- { PCI_DESCR( 5, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_ADDR) },
- { PCI_DESCR( 5, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_RANK) },
- { PCI_DESCR( 5, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_TC) },
- /*
- * This is the PCI device has an alternate address on some
- * processors like Core i7 860
- */
- { PCI_DESCR( 0, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE) },
- };
- static const struct pci_id_descr pci_dev_descr_i7core_westmere[] = {
- /* Memory controller */
- { PCI_DESCR(3, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MCR_REV2) },
- { PCI_DESCR(3, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_TAD_REV2) },
- /* Exists only for RDIMM */
- { PCI_DESCR(3, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_RAS_REV2), .optional = 1 },
- { PCI_DESCR(3, 4, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_TEST_REV2) },
- /* Channel 0 */
- { PCI_DESCR(4, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_CTRL_REV2) },
- { PCI_DESCR(4, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_ADDR_REV2) },
- { PCI_DESCR(4, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_RANK_REV2) },
- { PCI_DESCR(4, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH0_TC_REV2) },
- /* Channel 1 */
- { PCI_DESCR(5, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_CTRL_REV2) },
- { PCI_DESCR(5, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_ADDR_REV2) },
- { PCI_DESCR(5, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_RANK_REV2) },
- { PCI_DESCR(5, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH1_TC_REV2) },
- /* Channel 2 */
- { PCI_DESCR(6, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH2_CTRL_REV2) },
- { PCI_DESCR(6, 1, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH2_ADDR_REV2) },
- { PCI_DESCR(6, 2, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH2_RANK_REV2) },
- { PCI_DESCR(6, 3, PCI_DEVICE_ID_INTEL_LYNNFIELD_MC_CH2_TC_REV2) },
- /* Generic Non-core registers */
- { PCI_DESCR(0, 0, PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE_REV2) },
- };
- #define PCI_ID_TABLE_ENTRY(A) { .descr=A, .n_devs = ARRAY_SIZE(A) }
- static const struct pci_id_table pci_dev_table[] = {
- PCI_ID_TABLE_ENTRY(pci_dev_descr_i7core_nehalem),
- PCI_ID_TABLE_ENTRY(pci_dev_descr_lynnfield),
- PCI_ID_TABLE_ENTRY(pci_dev_descr_i7core_westmere),
- {0,} /* 0 terminated list. */
- };
- /*
- * pci_device_id table for which devices we are looking for
- */
- static const struct pci_device_id i7core_pci_tbl[] = {
- {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_X58_HUB_MGMT)},
- {PCI_DEVICE(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_LYNNFIELD_QPI_LINK0)},
- {0,} /* 0 terminated list. */
- };
- /****************************************************************************
- Ancillary status routines
- ****************************************************************************/
- /* MC_CONTROL bits */
- #define CH_ACTIVE(pvt, ch) ((pvt)->info.mc_control & (1 << (8 + ch)))
- #define ECCx8(pvt) ((pvt)->info.mc_control & (1 << 1))
- /* MC_STATUS bits */
- #define ECC_ENABLED(pvt) ((pvt)->info.mc_status & (1 << 4))
- #define CH_DISABLED(pvt, ch) ((pvt)->info.mc_status & (1 << ch))
- /* MC_MAX_DOD read functions */
- static inline int numdimms(u32 dimms)
- {
- return (dimms & 0x3) + 1;
- }
- static inline int numrank(u32 rank)
- {
- static const int ranks[] = { 1, 2, 4, -EINVAL };
- return ranks[rank & 0x3];
- }
- static inline int numbank(u32 bank)
- {
- static const int banks[] = { 4, 8, 16, -EINVAL };
- return banks[bank & 0x3];
- }
- static inline int numrow(u32 row)
- {
- static const int rows[] = {
- 1 << 12, 1 << 13, 1 << 14, 1 << 15,
- 1 << 16, -EINVAL, -EINVAL, -EINVAL,
- };
- return rows[row & 0x7];
- }
- static inline int numcol(u32 col)
- {
- static const int cols[] = {
- 1 << 10, 1 << 11, 1 << 12, -EINVAL,
- };
- return cols[col & 0x3];
- }
- static struct i7core_dev *get_i7core_dev(u8 socket)
- {
- struct i7core_dev *i7core_dev;
- list_for_each_entry(i7core_dev, &i7core_edac_list, list) {
- if (i7core_dev->socket == socket)
- return i7core_dev;
- }
- return NULL;
- }
- static struct i7core_dev *alloc_i7core_dev(u8 socket,
- const struct pci_id_table *table)
- {
- struct i7core_dev *i7core_dev;
- i7core_dev = kzalloc(sizeof(*i7core_dev), GFP_KERNEL);
- if (!i7core_dev)
- return NULL;
- i7core_dev->pdev = kzalloc(sizeof(*i7core_dev->pdev) * table->n_devs,
- GFP_KERNEL);
- if (!i7core_dev->pdev) {
- kfree(i7core_dev);
- return NULL;
- }
- i7core_dev->socket = socket;
- i7core_dev->n_devs = table->n_devs;
- list_add_tail(&i7core_dev->list, &i7core_edac_list);
- return i7core_dev;
- }
- static void free_i7core_dev(struct i7core_dev *i7core_dev)
- {
- list_del(&i7core_dev->list);
- kfree(i7core_dev->pdev);
- kfree(i7core_dev);
- }
- /****************************************************************************
- Memory check routines
- ****************************************************************************/
- static int get_dimm_config(struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- struct pci_dev *pdev;
- int i, j;
- enum edac_type mode;
- enum mem_type mtype;
- struct dimm_info *dimm;
- /* Get data from the MC register, function 0 */
- pdev = pvt->pci_mcr[0];
- if (!pdev)
- return -ENODEV;
- /* Device 3 function 0 reads */
- pci_read_config_dword(pdev, MC_CONTROL, &pvt->info.mc_control);
- pci_read_config_dword(pdev, MC_STATUS, &pvt->info.mc_status);
- pci_read_config_dword(pdev, MC_MAX_DOD, &pvt->info.max_dod);
- pci_read_config_dword(pdev, MC_CHANNEL_MAPPER, &pvt->info.ch_map);
- edac_dbg(0, "QPI %d control=0x%08x status=0x%08x dod=0x%08x map=0x%08x\n",
- pvt->i7core_dev->socket, pvt->info.mc_control,
- pvt->info.mc_status, pvt->info.max_dod, pvt->info.ch_map);
- if (ECC_ENABLED(pvt)) {
- edac_dbg(0, "ECC enabled with x%d SDCC\n", ECCx8(pvt) ? 8 : 4);
- if (ECCx8(pvt))
- mode = EDAC_S8ECD8ED;
- else
- mode = EDAC_S4ECD4ED;
- } else {
- edac_dbg(0, "ECC disabled\n");
- mode = EDAC_NONE;
- }
- /* FIXME: need to handle the error codes */
- edac_dbg(0, "DOD Max limits: DIMMS: %d, %d-ranked, %d-banked x%x x 0x%x\n",
- numdimms(pvt->info.max_dod),
- numrank(pvt->info.max_dod >> 2),
- numbank(pvt->info.max_dod >> 4),
- numrow(pvt->info.max_dod >> 6),
- numcol(pvt->info.max_dod >> 9));
- for (i = 0; i < NUM_CHANS; i++) {
- u32 data, dimm_dod[3], value[8];
- if (!pvt->pci_ch[i][0])
- continue;
- if (!CH_ACTIVE(pvt, i)) {
- edac_dbg(0, "Channel %i is not active\n", i);
- continue;
- }
- if (CH_DISABLED(pvt, i)) {
- edac_dbg(0, "Channel %i is disabled\n", i);
- continue;
- }
- /* Devices 4-6 function 0 */
- pci_read_config_dword(pvt->pci_ch[i][0],
- MC_CHANNEL_DIMM_INIT_PARAMS, &data);
- if (data & THREE_DIMMS_PRESENT)
- pvt->channel[i].is_3dimms_present = true;
- if (data & SINGLE_QUAD_RANK_PRESENT)
- pvt->channel[i].is_single_4rank = true;
- if (data & QUAD_RANK_PRESENT)
- pvt->channel[i].has_4rank = true;
- if (data & REGISTERED_DIMM)
- mtype = MEM_RDDR3;
- else
- mtype = MEM_DDR3;
- /* Devices 4-6 function 1 */
- pci_read_config_dword(pvt->pci_ch[i][1],
- MC_DOD_CH_DIMM0, &dimm_dod[0]);
- pci_read_config_dword(pvt->pci_ch[i][1],
- MC_DOD_CH_DIMM1, &dimm_dod[1]);
- pci_read_config_dword(pvt->pci_ch[i][1],
- MC_DOD_CH_DIMM2, &dimm_dod[2]);
- edac_dbg(0, "Ch%d phy rd%d, wr%d (0x%08x): %s%s%s%cDIMMs\n",
- i,
- RDLCH(pvt->info.ch_map, i), WRLCH(pvt->info.ch_map, i),
- data,
- pvt->channel[i].is_3dimms_present ? "3DIMMS " : "",
- pvt->channel[i].is_3dimms_present ? "SINGLE_4R " : "",
- pvt->channel[i].has_4rank ? "HAS_4R " : "",
- (data & REGISTERED_DIMM) ? 'R' : 'U');
- for (j = 0; j < 3; j++) {
- u32 banks, ranks, rows, cols;
- u32 size, npages;
- if (!DIMM_PRESENT(dimm_dod[j]))
- continue;
- dimm = EDAC_DIMM_PTR(mci->layers, mci->dimms, mci->n_layers,
- i, j, 0);
- banks = numbank(MC_DOD_NUMBANK(dimm_dod[j]));
- ranks = numrank(MC_DOD_NUMRANK(dimm_dod[j]));
- rows = numrow(MC_DOD_NUMROW(dimm_dod[j]));
- cols = numcol(MC_DOD_NUMCOL(dimm_dod[j]));
- /* DDR3 has 8 I/O banks */
- size = (rows * cols * banks * ranks) >> (20 - 3);
- edac_dbg(0, "\tdimm %d %d Mb offset: %x, bank: %d, rank: %d, row: %#x, col: %#x\n",
- j, size,
- RANKOFFSET(dimm_dod[j]),
- banks, ranks, rows, cols);
- npages = MiB_TO_PAGES(size);
- dimm->nr_pages = npages;
- switch (banks) {
- case 4:
- dimm->dtype = DEV_X4;
- break;
- case 8:
- dimm->dtype = DEV_X8;
- break;
- case 16:
- dimm->dtype = DEV_X16;
- break;
- default:
- dimm->dtype = DEV_UNKNOWN;
- }
- snprintf(dimm->label, sizeof(dimm->label),
- "CPU#%uChannel#%u_DIMM#%u",
- pvt->i7core_dev->socket, i, j);
- dimm->grain = 8;
- dimm->edac_mode = mode;
- dimm->mtype = mtype;
- }
- pci_read_config_dword(pdev, MC_SAG_CH_0, &value[0]);
- pci_read_config_dword(pdev, MC_SAG_CH_1, &value[1]);
- pci_read_config_dword(pdev, MC_SAG_CH_2, &value[2]);
- pci_read_config_dword(pdev, MC_SAG_CH_3, &value[3]);
- pci_read_config_dword(pdev, MC_SAG_CH_4, &value[4]);
- pci_read_config_dword(pdev, MC_SAG_CH_5, &value[5]);
- pci_read_config_dword(pdev, MC_SAG_CH_6, &value[6]);
- pci_read_config_dword(pdev, MC_SAG_CH_7, &value[7]);
- edac_dbg(1, "\t[%i] DIVBY3\tREMOVED\tOFFSET\n", i);
- for (j = 0; j < 8; j++)
- edac_dbg(1, "\t\t%#x\t%#x\t%#x\n",
- (value[j] >> 27) & 0x1,
- (value[j] >> 24) & 0x7,
- (value[j] & ((1 << 24) - 1)));
- }
- return 0;
- }
- /****************************************************************************
- Error insertion routines
- ****************************************************************************/
- #define to_mci(k) container_of(k, struct mem_ctl_info, dev)
- /* The i7core has independent error injection features per channel.
- However, to have a simpler code, we don't allow enabling error injection
- on more than one channel.
- Also, since a change at an inject parameter will be applied only at enable,
- we're disabling error injection on all write calls to the sysfs nodes that
- controls the error code injection.
- */
- static int disable_inject(const struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- pvt->inject.enable = 0;
- if (!pvt->pci_ch[pvt->inject.channel][0])
- return -ENODEV;
- pci_write_config_dword(pvt->pci_ch[pvt->inject.channel][0],
- MC_CHANNEL_ERROR_INJECT, 0);
- return 0;
- }
- /*
- * i7core inject inject.section
- *
- * accept and store error injection inject.section value
- * bit 0 - refers to the lower 32-byte half cacheline
- * bit 1 - refers to the upper 32-byte half cacheline
- */
- static ssize_t i7core_inject_section_store(struct device *dev,
- struct device_attribute *mattr,
- const char *data, size_t count)
- {
- struct mem_ctl_info *mci = to_mci(dev);
- struct i7core_pvt *pvt = mci->pvt_info;
- unsigned long value;
- int rc;
- if (pvt->inject.enable)
- disable_inject(mci);
- rc = kstrtoul(data, 10, &value);
- if ((rc < 0) || (value > 3))
- return -EIO;
- pvt->inject.section = (u32) value;
- return count;
- }
- static ssize_t i7core_inject_section_show(struct device *dev,
- struct device_attribute *mattr,
- char *data)
- {
- struct mem_ctl_info *mci = to_mci(dev);
- struct i7core_pvt *pvt = mci->pvt_info;
- return sprintf(data, "0x%08x\n", pvt->inject.section);
- }
- /*
- * i7core inject.type
- *
- * accept and store error injection inject.section value
- * bit 0 - repeat enable - Enable error repetition
- * bit 1 - inject ECC error
- * bit 2 - inject parity error
- */
- static ssize_t i7core_inject_type_store(struct device *dev,
- struct device_attribute *mattr,
- const char *data, size_t count)
- {
- struct mem_ctl_info *mci = to_mci(dev);
- struct i7core_pvt *pvt = mci->pvt_info;
- unsigned long value;
- int rc;
- if (pvt->inject.enable)
- disable_inject(mci);
- rc = kstrtoul(data, 10, &value);
- if ((rc < 0) || (value > 7))
- return -EIO;
- pvt->inject.type = (u32) value;
- return count;
- }
- static ssize_t i7core_inject_type_show(struct device *dev,
- struct device_attribute *mattr,
- char *data)
- {
- struct mem_ctl_info *mci = to_mci(dev);
- struct i7core_pvt *pvt = mci->pvt_info;
- return sprintf(data, "0x%08x\n", pvt->inject.type);
- }
- /*
- * i7core_inject_inject.eccmask_store
- *
- * The type of error (UE/CE) will depend on the inject.eccmask value:
- * Any bits set to a 1 will flip the corresponding ECC bit
- * Correctable errors can be injected by flipping 1 bit or the bits within
- * a symbol pair (2 consecutive aligned 8-bit pairs - i.e. 7:0 and 15:8 or
- * 23:16 and 31:24). Flipping bits in two symbol pairs will cause an
- * uncorrectable error to be injected.
- */
- static ssize_t i7core_inject_eccmask_store(struct device *dev,
- struct device_attribute *mattr,
- const char *data, size_t count)
- {
- struct mem_ctl_info *mci = to_mci(dev);
- struct i7core_pvt *pvt = mci->pvt_info;
- unsigned long value;
- int rc;
- if (pvt->inject.enable)
- disable_inject(mci);
- rc = kstrtoul(data, 10, &value);
- if (rc < 0)
- return -EIO;
- pvt->inject.eccmask = (u32) value;
- return count;
- }
- static ssize_t i7core_inject_eccmask_show(struct device *dev,
- struct device_attribute *mattr,
- char *data)
- {
- struct mem_ctl_info *mci = to_mci(dev);
- struct i7core_pvt *pvt = mci->pvt_info;
- return sprintf(data, "0x%08x\n", pvt->inject.eccmask);
- }
- /*
- * i7core_addrmatch
- *
- * The type of error (UE/CE) will depend on the inject.eccmask value:
- * Any bits set to a 1 will flip the corresponding ECC bit
- * Correctable errors can be injected by flipping 1 bit or the bits within
- * a symbol pair (2 consecutive aligned 8-bit pairs - i.e. 7:0 and 15:8 or
- * 23:16 and 31:24). Flipping bits in two symbol pairs will cause an
- * uncorrectable error to be injected.
- */
- #define DECLARE_ADDR_MATCH(param, limit) \
- static ssize_t i7core_inject_store_##param( \
- struct device *dev, \
- struct device_attribute *mattr, \
- const char *data, size_t count) \
- { \
- struct mem_ctl_info *mci = dev_get_drvdata(dev); \
- struct i7core_pvt *pvt; \
- long value; \
- int rc; \
- \
- edac_dbg(1, "\n"); \
- pvt = mci->pvt_info; \
- \
- if (pvt->inject.enable) \
- disable_inject(mci); \
- \
- if (!strcasecmp(data, "any") || !strcasecmp(data, "any\n"))\
- value = -1; \
- else { \
- rc = kstrtoul(data, 10, &value); \
- if ((rc < 0) || (value >= limit)) \
- return -EIO; \
- } \
- \
- pvt->inject.param = value; \
- \
- return count; \
- } \
- \
- static ssize_t i7core_inject_show_##param( \
- struct device *dev, \
- struct device_attribute *mattr, \
- char *data) \
- { \
- struct mem_ctl_info *mci = dev_get_drvdata(dev); \
- struct i7core_pvt *pvt; \
- \
- pvt = mci->pvt_info; \
- edac_dbg(1, "pvt=%p\n", pvt); \
- if (pvt->inject.param < 0) \
- return sprintf(data, "any\n"); \
- else \
- return sprintf(data, "%d\n", pvt->inject.param);\
- }
- #define ATTR_ADDR_MATCH(param) \
- static DEVICE_ATTR(param, S_IRUGO | S_IWUSR, \
- i7core_inject_show_##param, \
- i7core_inject_store_##param)
- DECLARE_ADDR_MATCH(channel, 3);
- DECLARE_ADDR_MATCH(dimm, 3);
- DECLARE_ADDR_MATCH(rank, 4);
- DECLARE_ADDR_MATCH(bank, 32);
- DECLARE_ADDR_MATCH(page, 0x10000);
- DECLARE_ADDR_MATCH(col, 0x4000);
- ATTR_ADDR_MATCH(channel);
- ATTR_ADDR_MATCH(dimm);
- ATTR_ADDR_MATCH(rank);
- ATTR_ADDR_MATCH(bank);
- ATTR_ADDR_MATCH(page);
- ATTR_ADDR_MATCH(col);
- static int write_and_test(struct pci_dev *dev, const int where, const u32 val)
- {
- u32 read;
- int count;
- edac_dbg(0, "setting pci %02x:%02x.%x reg=%02x value=%08x\n",
- dev->bus->number, PCI_SLOT(dev->devfn), PCI_FUNC(dev->devfn),
- where, val);
- for (count = 0; count < 10; count++) {
- if (count)
- msleep(100);
- pci_write_config_dword(dev, where, val);
- pci_read_config_dword(dev, where, &read);
- if (read == val)
- return 0;
- }
- i7core_printk(KERN_ERR, "Error during set pci %02x:%02x.%x reg=%02x "
- "write=%08x. Read=%08x\n",
- dev->bus->number, PCI_SLOT(dev->devfn), PCI_FUNC(dev->devfn),
- where, val, read);
- return -EINVAL;
- }
- /*
- * This routine prepares the Memory Controller for error injection.
- * The error will be injected when some process tries to write to the
- * memory that matches the given criteria.
- * The criteria can be set in terms of a mask where dimm, rank, bank, page
- * and col can be specified.
- * A -1 value for any of the mask items will make the MCU to ignore
- * that matching criteria for error injection.
- *
- * It should be noticed that the error will only happen after a write operation
- * on a memory that matches the condition. if REPEAT_EN is not enabled at
- * inject mask, then it will produce just one error. Otherwise, it will repeat
- * until the injectmask would be cleaned.
- *
- * FIXME: This routine assumes that MAXNUMDIMMS value of MC_MAX_DOD
- * is reliable enough to check if the MC is using the
- * three channels. However, this is not clear at the datasheet.
- */
- static ssize_t i7core_inject_enable_store(struct device *dev,
- struct device_attribute *mattr,
- const char *data, size_t count)
- {
- struct mem_ctl_info *mci = to_mci(dev);
- struct i7core_pvt *pvt = mci->pvt_info;
- u32 injectmask;
- u64 mask = 0;
- int rc;
- long enable;
- if (!pvt->pci_ch[pvt->inject.channel][0])
- return 0;
- rc = kstrtoul(data, 10, &enable);
- if ((rc < 0))
- return 0;
- if (enable) {
- pvt->inject.enable = 1;
- } else {
- disable_inject(mci);
- return count;
- }
- /* Sets pvt->inject.dimm mask */
- if (pvt->inject.dimm < 0)
- mask |= 1LL << 41;
- else {
- if (pvt->channel[pvt->inject.channel].dimms > 2)
- mask |= (pvt->inject.dimm & 0x3LL) << 35;
- else
- mask |= (pvt->inject.dimm & 0x1LL) << 36;
- }
- /* Sets pvt->inject.rank mask */
- if (pvt->inject.rank < 0)
- mask |= 1LL << 40;
- else {
- if (pvt->channel[pvt->inject.channel].dimms > 2)
- mask |= (pvt->inject.rank & 0x1LL) << 34;
- else
- mask |= (pvt->inject.rank & 0x3LL) << 34;
- }
- /* Sets pvt->inject.bank mask */
- if (pvt->inject.bank < 0)
- mask |= 1LL << 39;
- else
- mask |= (pvt->inject.bank & 0x15LL) << 30;
- /* Sets pvt->inject.page mask */
- if (pvt->inject.page < 0)
- mask |= 1LL << 38;
- else
- mask |= (pvt->inject.page & 0xffff) << 14;
- /* Sets pvt->inject.column mask */
- if (pvt->inject.col < 0)
- mask |= 1LL << 37;
- else
- mask |= (pvt->inject.col & 0x3fff);
- /*
- * bit 0: REPEAT_EN
- * bits 1-2: MASK_HALF_CACHELINE
- * bit 3: INJECT_ECC
- * bit 4: INJECT_ADDR_PARITY
- */
- injectmask = (pvt->inject.type & 1) |
- (pvt->inject.section & 0x3) << 1 |
- (pvt->inject.type & 0x6) << (3 - 1);
- /* Unlock writes to registers - this register is write only */
- pci_write_config_dword(pvt->pci_noncore,
- MC_CFG_CONTROL, 0x2);
- write_and_test(pvt->pci_ch[pvt->inject.channel][0],
- MC_CHANNEL_ADDR_MATCH, mask);
- write_and_test(pvt->pci_ch[pvt->inject.channel][0],
- MC_CHANNEL_ADDR_MATCH + 4, mask >> 32L);
- write_and_test(pvt->pci_ch[pvt->inject.channel][0],
- MC_CHANNEL_ERROR_MASK, pvt->inject.eccmask);
- write_and_test(pvt->pci_ch[pvt->inject.channel][0],
- MC_CHANNEL_ERROR_INJECT, injectmask);
- /*
- * This is something undocumented, based on my tests
- * Without writing 8 to this register, errors aren't injected. Not sure
- * why.
- */
- pci_write_config_dword(pvt->pci_noncore,
- MC_CFG_CONTROL, 8);
- edac_dbg(0, "Error inject addr match 0x%016llx, ecc 0x%08x, inject 0x%08x\n",
- mask, pvt->inject.eccmask, injectmask);
- return count;
- }
- static ssize_t i7core_inject_enable_show(struct device *dev,
- struct device_attribute *mattr,
- char *data)
- {
- struct mem_ctl_info *mci = to_mci(dev);
- struct i7core_pvt *pvt = mci->pvt_info;
- u32 injectmask;
- if (!pvt->pci_ch[pvt->inject.channel][0])
- return 0;
- pci_read_config_dword(pvt->pci_ch[pvt->inject.channel][0],
- MC_CHANNEL_ERROR_INJECT, &injectmask);
- edac_dbg(0, "Inject error read: 0x%018x\n", injectmask);
- if (injectmask & 0x0c)
- pvt->inject.enable = 1;
- return sprintf(data, "%d\n", pvt->inject.enable);
- }
- #define DECLARE_COUNTER(param) \
- static ssize_t i7core_show_counter_##param( \
- struct device *dev, \
- struct device_attribute *mattr, \
- char *data) \
- { \
- struct mem_ctl_info *mci = dev_get_drvdata(dev); \
- struct i7core_pvt *pvt = mci->pvt_info; \
- \
- edac_dbg(1, "\n"); \
- if (!pvt->ce_count_available || (pvt->is_registered)) \
- return sprintf(data, "data unavailable\n"); \
- return sprintf(data, "%lu\n", \
- pvt->udimm_ce_count[param]); \
- }
- #define ATTR_COUNTER(param) \
- static DEVICE_ATTR(udimm##param, S_IRUGO | S_IWUSR, \
- i7core_show_counter_##param, \
- NULL)
- DECLARE_COUNTER(0);
- DECLARE_COUNTER(1);
- DECLARE_COUNTER(2);
- ATTR_COUNTER(0);
- ATTR_COUNTER(1);
- ATTR_COUNTER(2);
- /*
- * inject_addrmatch device sysfs struct
- */
- static struct attribute *i7core_addrmatch_attrs[] = {
- &dev_attr_channel.attr,
- &dev_attr_dimm.attr,
- &dev_attr_rank.attr,
- &dev_attr_bank.attr,
- &dev_attr_page.attr,
- &dev_attr_col.attr,
- NULL
- };
- static struct attribute_group addrmatch_grp = {
- .attrs = i7core_addrmatch_attrs,
- };
- static const struct attribute_group *addrmatch_groups[] = {
- &addrmatch_grp,
- NULL
- };
- static void addrmatch_release(struct device *device)
- {
- edac_dbg(1, "Releasing device %s\n", dev_name(device));
- kfree(device);
- }
- static struct device_type addrmatch_type = {
- .groups = addrmatch_groups,
- .release = addrmatch_release,
- };
- /*
- * all_channel_counts sysfs struct
- */
- static struct attribute *i7core_udimm_counters_attrs[] = {
- &dev_attr_udimm0.attr,
- &dev_attr_udimm1.attr,
- &dev_attr_udimm2.attr,
- NULL
- };
- static struct attribute_group all_channel_counts_grp = {
- .attrs = i7core_udimm_counters_attrs,
- };
- static const struct attribute_group *all_channel_counts_groups[] = {
- &all_channel_counts_grp,
- NULL
- };
- static void all_channel_counts_release(struct device *device)
- {
- edac_dbg(1, "Releasing device %s\n", dev_name(device));
- kfree(device);
- }
- static struct device_type all_channel_counts_type = {
- .groups = all_channel_counts_groups,
- .release = all_channel_counts_release,
- };
- /*
- * inject sysfs attributes
- */
- static DEVICE_ATTR(inject_section, S_IRUGO | S_IWUSR,
- i7core_inject_section_show, i7core_inject_section_store);
- static DEVICE_ATTR(inject_type, S_IRUGO | S_IWUSR,
- i7core_inject_type_show, i7core_inject_type_store);
- static DEVICE_ATTR(inject_eccmask, S_IRUGO | S_IWUSR,
- i7core_inject_eccmask_show, i7core_inject_eccmask_store);
- static DEVICE_ATTR(inject_enable, S_IRUGO | S_IWUSR,
- i7core_inject_enable_show, i7core_inject_enable_store);
- static struct attribute *i7core_dev_attrs[] = {
- &dev_attr_inject_section.attr,
- &dev_attr_inject_type.attr,
- &dev_attr_inject_eccmask.attr,
- &dev_attr_inject_enable.attr,
- NULL
- };
- ATTRIBUTE_GROUPS(i7core_dev);
- static int i7core_create_sysfs_devices(struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- int rc;
- pvt->addrmatch_dev = kzalloc(sizeof(*pvt->addrmatch_dev), GFP_KERNEL);
- if (!pvt->addrmatch_dev)
- return -ENOMEM;
- pvt->addrmatch_dev->type = &addrmatch_type;
- pvt->addrmatch_dev->bus = mci->dev.bus;
- device_initialize(pvt->addrmatch_dev);
- pvt->addrmatch_dev->parent = &mci->dev;
- dev_set_name(pvt->addrmatch_dev, "inject_addrmatch");
- dev_set_drvdata(pvt->addrmatch_dev, mci);
- edac_dbg(1, "creating %s\n", dev_name(pvt->addrmatch_dev));
- rc = device_add(pvt->addrmatch_dev);
- if (rc < 0)
- goto err_put_addrmatch;
- if (!pvt->is_registered) {
- pvt->chancounts_dev = kzalloc(sizeof(*pvt->chancounts_dev),
- GFP_KERNEL);
- if (!pvt->chancounts_dev) {
- rc = -ENOMEM;
- goto err_del_addrmatch;
- }
- pvt->chancounts_dev->type = &all_channel_counts_type;
- pvt->chancounts_dev->bus = mci->dev.bus;
- device_initialize(pvt->chancounts_dev);
- pvt->chancounts_dev->parent = &mci->dev;
- dev_set_name(pvt->chancounts_dev, "all_channel_counts");
- dev_set_drvdata(pvt->chancounts_dev, mci);
- edac_dbg(1, "creating %s\n", dev_name(pvt->chancounts_dev));
- rc = device_add(pvt->chancounts_dev);
- if (rc < 0)
- goto err_put_chancounts;
- }
- return 0;
- err_put_chancounts:
- put_device(pvt->chancounts_dev);
- err_del_addrmatch:
- device_del(pvt->addrmatch_dev);
- err_put_addrmatch:
- put_device(pvt->addrmatch_dev);
- return rc;
- }
- static void i7core_delete_sysfs_devices(struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- edac_dbg(1, "\n");
- if (!pvt->is_registered) {
- device_del(pvt->chancounts_dev);
- put_device(pvt->chancounts_dev);
- }
- device_del(pvt->addrmatch_dev);
- put_device(pvt->addrmatch_dev);
- }
- /****************************************************************************
- Device initialization routines: put/get, init/exit
- ****************************************************************************/
- /*
- * i7core_put_all_devices 'put' all the devices that we have
- * reserved via 'get'
- */
- static void i7core_put_devices(struct i7core_dev *i7core_dev)
- {
- int i;
- edac_dbg(0, "\n");
- for (i = 0; i < i7core_dev->n_devs; i++) {
- struct pci_dev *pdev = i7core_dev->pdev[i];
- if (!pdev)
- continue;
- edac_dbg(0, "Removing dev %02x:%02x.%d\n",
- pdev->bus->number,
- PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn));
- pci_dev_put(pdev);
- }
- }
- static void i7core_put_all_devices(void)
- {
- struct i7core_dev *i7core_dev, *tmp;
- list_for_each_entry_safe(i7core_dev, tmp, &i7core_edac_list, list) {
- i7core_put_devices(i7core_dev);
- free_i7core_dev(i7core_dev);
- }
- }
- static void __init i7core_xeon_pci_fixup(const struct pci_id_table *table)
- {
- struct pci_dev *pdev = NULL;
- int i;
- /*
- * On Xeon 55xx, the Intel Quick Path Arch Generic Non-core pci buses
- * aren't announced by acpi. So, we need to use a legacy scan probing
- * to detect them
- */
- while (table && table->descr) {
- pdev = pci_get_device(PCI_VENDOR_ID_INTEL, table->descr[0].dev_id, NULL);
- if (unlikely(!pdev)) {
- for (i = 0; i < MAX_SOCKET_BUSES; i++)
- pcibios_scan_specific_bus(255-i);
- }
- pci_dev_put(pdev);
- table++;
- }
- }
- static unsigned i7core_pci_lastbus(void)
- {
- int last_bus = 0, bus;
- struct pci_bus *b = NULL;
- while ((b = pci_find_next_bus(b)) != NULL) {
- bus = b->number;
- edac_dbg(0, "Found bus %d\n", bus);
- if (bus > last_bus)
- last_bus = bus;
- }
- edac_dbg(0, "Last bus %d\n", last_bus);
- return last_bus;
- }
- /*
- * i7core_get_all_devices Find and perform 'get' operation on the MCH's
- * device/functions we want to reference for this driver
- *
- * Need to 'get' device 16 func 1 and func 2
- */
- static int i7core_get_onedevice(struct pci_dev **prev,
- const struct pci_id_table *table,
- const unsigned devno,
- const unsigned last_bus)
- {
- struct i7core_dev *i7core_dev;
- const struct pci_id_descr *dev_descr = &table->descr[devno];
- struct pci_dev *pdev = NULL;
- u8 bus = 0;
- u8 socket = 0;
- pdev = pci_get_device(PCI_VENDOR_ID_INTEL,
- dev_descr->dev_id, *prev);
- /*
- * On Xeon 55xx, the Intel QuickPath Arch Generic Non-core regs
- * is at addr 8086:2c40, instead of 8086:2c41. So, we need
- * to probe for the alternate address in case of failure
- */
- if (dev_descr->dev_id == PCI_DEVICE_ID_INTEL_I7_NONCORE && !pdev) {
- pci_dev_get(*prev); /* pci_get_device will put it */
- pdev = pci_get_device(PCI_VENDOR_ID_INTEL,
- PCI_DEVICE_ID_INTEL_I7_NONCORE_ALT, *prev);
- }
- if (dev_descr->dev_id == PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE &&
- !pdev) {
- pci_dev_get(*prev); /* pci_get_device will put it */
- pdev = pci_get_device(PCI_VENDOR_ID_INTEL,
- PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE_ALT,
- *prev);
- }
- if (!pdev) {
- if (*prev) {
- *prev = pdev;
- return 0;
- }
- if (dev_descr->optional)
- return 0;
- if (devno == 0)
- return -ENODEV;
- i7core_printk(KERN_INFO,
- "Device not found: dev %02x.%d PCI ID %04x:%04x\n",
- dev_descr->dev, dev_descr->func,
- PCI_VENDOR_ID_INTEL, dev_descr->dev_id);
- /* End of list, leave */
- return -ENODEV;
- }
- bus = pdev->bus->number;
- socket = last_bus - bus;
- i7core_dev = get_i7core_dev(socket);
- if (!i7core_dev) {
- i7core_dev = alloc_i7core_dev(socket, table);
- if (!i7core_dev) {
- pci_dev_put(pdev);
- return -ENOMEM;
- }
- }
- if (i7core_dev->pdev[devno]) {
- i7core_printk(KERN_ERR,
- "Duplicated device for "
- "dev %02x:%02x.%d PCI ID %04x:%04x\n",
- bus, dev_descr->dev, dev_descr->func,
- PCI_VENDOR_ID_INTEL, dev_descr->dev_id);
- pci_dev_put(pdev);
- return -ENODEV;
- }
- i7core_dev->pdev[devno] = pdev;
- /* Sanity check */
- if (unlikely(PCI_SLOT(pdev->devfn) != dev_descr->dev ||
- PCI_FUNC(pdev->devfn) != dev_descr->func)) {
- i7core_printk(KERN_ERR,
- "Device PCI ID %04x:%04x "
- "has dev %02x:%02x.%d instead of dev %02x:%02x.%d\n",
- PCI_VENDOR_ID_INTEL, dev_descr->dev_id,
- bus, PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn),
- bus, dev_descr->dev, dev_descr->func);
- return -ENODEV;
- }
- /* Be sure that the device is enabled */
- if (unlikely(pci_enable_device(pdev) < 0)) {
- i7core_printk(KERN_ERR,
- "Couldn't enable "
- "dev %02x:%02x.%d PCI ID %04x:%04x\n",
- bus, dev_descr->dev, dev_descr->func,
- PCI_VENDOR_ID_INTEL, dev_descr->dev_id);
- return -ENODEV;
- }
- edac_dbg(0, "Detected socket %d dev %02x:%02x.%d PCI ID %04x:%04x\n",
- socket, bus, dev_descr->dev,
- dev_descr->func,
- PCI_VENDOR_ID_INTEL, dev_descr->dev_id);
- /*
- * As stated on drivers/pci/search.c, the reference count for
- * @from is always decremented if it is not %NULL. So, as we need
- * to get all devices up to null, we need to do a get for the device
- */
- pci_dev_get(pdev);
- *prev = pdev;
- return 0;
- }
- static int i7core_get_all_devices(void)
- {
- int i, rc, last_bus;
- struct pci_dev *pdev = NULL;
- const struct pci_id_table *table = pci_dev_table;
- last_bus = i7core_pci_lastbus();
- while (table && table->descr) {
- for (i = 0; i < table->n_devs; i++) {
- pdev = NULL;
- do {
- rc = i7core_get_onedevice(&pdev, table, i,
- last_bus);
- if (rc < 0) {
- if (i == 0) {
- i = table->n_devs;
- break;
- }
- i7core_put_all_devices();
- return -ENODEV;
- }
- } while (pdev);
- }
- table++;
- }
- return 0;
- }
- static int mci_bind_devs(struct mem_ctl_info *mci,
- struct i7core_dev *i7core_dev)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- struct pci_dev *pdev;
- int i, func, slot;
- char *family;
- pvt->is_registered = false;
- pvt->enable_scrub = false;
- for (i = 0; i < i7core_dev->n_devs; i++) {
- pdev = i7core_dev->pdev[i];
- if (!pdev)
- continue;
- func = PCI_FUNC(pdev->devfn);
- slot = PCI_SLOT(pdev->devfn);
- if (slot == 3) {
- if (unlikely(func > MAX_MCR_FUNC))
- goto error;
- pvt->pci_mcr[func] = pdev;
- } else if (likely(slot >= 4 && slot < 4 + NUM_CHANS)) {
- if (unlikely(func > MAX_CHAN_FUNC))
- goto error;
- pvt->pci_ch[slot - 4][func] = pdev;
- } else if (!slot && !func) {
- pvt->pci_noncore = pdev;
- /* Detect the processor family */
- switch (pdev->device) {
- case PCI_DEVICE_ID_INTEL_I7_NONCORE:
- family = "Xeon 35xx/ i7core";
- pvt->enable_scrub = false;
- break;
- case PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE_ALT:
- family = "i7-800/i5-700";
- pvt->enable_scrub = false;
- break;
- case PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE:
- family = "Xeon 34xx";
- pvt->enable_scrub = false;
- break;
- case PCI_DEVICE_ID_INTEL_I7_NONCORE_ALT:
- family = "Xeon 55xx";
- pvt->enable_scrub = true;
- break;
- case PCI_DEVICE_ID_INTEL_LYNNFIELD_NONCORE_REV2:
- family = "Xeon 56xx / i7-900";
- pvt->enable_scrub = true;
- break;
- default:
- family = "unknown";
- pvt->enable_scrub = false;
- }
- edac_dbg(0, "Detected a processor type %s\n", family);
- } else
- goto error;
- edac_dbg(0, "Associated fn %d.%d, dev = %p, socket %d\n",
- PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn),
- pdev, i7core_dev->socket);
- if (PCI_SLOT(pdev->devfn) == 3 &&
- PCI_FUNC(pdev->devfn) == 2)
- pvt->is_registered = true;
- }
- return 0;
- error:
- i7core_printk(KERN_ERR, "Device %d, function %d "
- "is out of the expected range\n",
- slot, func);
- return -EINVAL;
- }
- /****************************************************************************
- Error check routines
- ****************************************************************************/
- static void i7core_rdimm_update_ce_count(struct mem_ctl_info *mci,
- const int chan,
- const int new0,
- const int new1,
- const int new2)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- int add0 = 0, add1 = 0, add2 = 0;
- /* Updates CE counters if it is not the first time here */
- if (pvt->ce_count_available) {
- /* Updates CE counters */
- add2 = new2 - pvt->rdimm_last_ce_count[chan][2];
- add1 = new1 - pvt->rdimm_last_ce_count[chan][1];
- add0 = new0 - pvt->rdimm_last_ce_count[chan][0];
- if (add2 < 0)
- add2 += 0x7fff;
- pvt->rdimm_ce_count[chan][2] += add2;
- if (add1 < 0)
- add1 += 0x7fff;
- pvt->rdimm_ce_count[chan][1] += add1;
- if (add0 < 0)
- add0 += 0x7fff;
- pvt->rdimm_ce_count[chan][0] += add0;
- } else
- pvt->ce_count_available = 1;
- /* Store the new values */
- pvt->rdimm_last_ce_count[chan][2] = new2;
- pvt->rdimm_last_ce_count[chan][1] = new1;
- pvt->rdimm_last_ce_count[chan][0] = new0;
- /*updated the edac core */
- if (add0 != 0)
- edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, add0,
- 0, 0, 0,
- chan, 0, -1, "error", "");
- if (add1 != 0)
- edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, add1,
- 0, 0, 0,
- chan, 1, -1, "error", "");
- if (add2 != 0)
- edac_mc_handle_error(HW_EVENT_ERR_CORRECTED, mci, add2,
- 0, 0, 0,
- chan, 2, -1, "error", "");
- }
- static void i7core_rdimm_check_mc_ecc_err(struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- u32 rcv[3][2];
- int i, new0, new1, new2;
- /*Read DEV 3: FUN 2: MC_COR_ECC_CNT regs directly*/
- pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_0,
- &rcv[0][0]);
- pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_1,
- &rcv[0][1]);
- pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_2,
- &rcv[1][0]);
- pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_3,
- &rcv[1][1]);
- pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_4,
- &rcv[2][0]);
- pci_read_config_dword(pvt->pci_mcr[2], MC_COR_ECC_CNT_5,
- &rcv[2][1]);
- for (i = 0 ; i < 3; i++) {
- edac_dbg(3, "MC_COR_ECC_CNT%d = 0x%x; MC_COR_ECC_CNT%d = 0x%x\n",
- (i * 2), rcv[i][0], (i * 2) + 1, rcv[i][1]);
- /*if the channel has 3 dimms*/
- if (pvt->channel[i].dimms > 2) {
- new0 = DIMM_BOT_COR_ERR(rcv[i][0]);
- new1 = DIMM_TOP_COR_ERR(rcv[i][0]);
- new2 = DIMM_BOT_COR_ERR(rcv[i][1]);
- } else {
- new0 = DIMM_TOP_COR_ERR(rcv[i][0]) +
- DIMM_BOT_COR_ERR(rcv[i][0]);
- new1 = DIMM_TOP_COR_ERR(rcv[i][1]) +
- DIMM_BOT_COR_ERR(rcv[i][1]);
- new2 = 0;
- }
- i7core_rdimm_update_ce_count(mci, i, new0, new1, new2);
- }
- }
- /* This function is based on the device 3 function 4 registers as described on:
- * Intel Xeon Processor 5500 Series Datasheet Volume 2
- * http://www.intel.com/Assets/PDF/datasheet/321322.pdf
- * also available at:
- * http://www.arrownac.com/manufacturers/intel/s/nehalem/5500-datasheet-v2.pdf
- */
- static void i7core_udimm_check_mc_ecc_err(struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- u32 rcv1, rcv0;
- int new0, new1, new2;
- if (!pvt->pci_mcr[4]) {
- edac_dbg(0, "MCR registers not found\n");
- return;
- }
- /* Corrected test errors */
- pci_read_config_dword(pvt->pci_mcr[4], MC_TEST_ERR_RCV1, &rcv1);
- pci_read_config_dword(pvt->pci_mcr[4], MC_TEST_ERR_RCV0, &rcv0);
- /* Store the new values */
- new2 = DIMM2_COR_ERR(rcv1);
- new1 = DIMM1_COR_ERR(rcv0);
- new0 = DIMM0_COR_ERR(rcv0);
- /* Updates CE counters if it is not the first time here */
- if (pvt->ce_count_available) {
- /* Updates CE counters */
- int add0, add1, add2;
- add2 = new2 - pvt->udimm_last_ce_count[2];
- add1 = new1 - pvt->udimm_last_ce_count[1];
- add0 = new0 - pvt->udimm_last_ce_count[0];
- if (add2 < 0)
- add2 += 0x7fff;
- pvt->udimm_ce_count[2] += add2;
- if (add1 < 0)
- add1 += 0x7fff;
- pvt->udimm_ce_count[1] += add1;
- if (add0 < 0)
- add0 += 0x7fff;
- pvt->udimm_ce_count[0] += add0;
- if (add0 | add1 | add2)
- i7core_printk(KERN_ERR, "New Corrected error(s): "
- "dimm0: +%d, dimm1: +%d, dimm2 +%d\n",
- add0, add1, add2);
- } else
- pvt->ce_count_available = 1;
- /* Store the new values */
- pvt->udimm_last_ce_count[2] = new2;
- pvt->udimm_last_ce_count[1] = new1;
- pvt->udimm_last_ce_count[0] = new0;
- }
- /*
- * According with tables E-11 and E-12 of chapter E.3.3 of Intel 64 and IA-32
- * Architectures Software Developer’s Manual Volume 3B.
- * Nehalem are defined as family 0x06, model 0x1a
- *
- * The MCA registers used here are the following ones:
- * struct mce field MCA Register
- * m->status MSR_IA32_MC8_STATUS
- * m->addr MSR_IA32_MC8_ADDR
- * m->misc MSR_IA32_MC8_MISC
- * In the case of Nehalem, the error information is masked at .status and .misc
- * fields
- */
- static void i7core_mce_output_error(struct mem_ctl_info *mci,
- const struct mce *m)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- char *optype, *err;
- enum hw_event_mc_err_type tp_event;
- unsigned long error = m->status & 0x1ff0000l;
- bool uncorrected_error = m->mcgstatus & 1ll << 61;
- bool ripv = m->mcgstatus & 1;
- u32 optypenum = (m->status >> 4) & 0x07;
- u32 core_err_cnt = (m->status >> 38) & 0x7fff;
- u32 dimm = (m->misc >> 16) & 0x3;
- u32 channel = (m->misc >> 18) & 0x3;
- u32 syndrome = m->misc >> 32;
- u32 errnum = find_first_bit(&error, 32);
- if (uncorrected_error) {
- if (ripv)
- tp_event = HW_EVENT_ERR_FATAL;
- else
- tp_event = HW_EVENT_ERR_UNCORRECTED;
- } else {
- tp_event = HW_EVENT_ERR_CORRECTED;
- }
- switch (optypenum) {
- case 0:
- optype = "generic undef request";
- break;
- case 1:
- optype = "read error";
- break;
- case 2:
- optype = "write error";
- break;
- case 3:
- optype = "addr/cmd error";
- break;
- case 4:
- optype = "scrubbing error";
- break;
- default:
- optype = "reserved";
- break;
- }
- switch (errnum) {
- case 16:
- err = "read ECC error";
- break;
- case 17:
- err = "RAS ECC error";
- break;
- case 18:
- err = "write parity error";
- break;
- case 19:
- err = "redundacy loss";
- break;
- case 20:
- err = "reserved";
- break;
- case 21:
- err = "memory range error";
- break;
- case 22:
- err = "RTID out of range";
- break;
- case 23:
- err = "address parity error";
- break;
- case 24:
- err = "byte enable parity error";
- break;
- default:
- err = "unknown";
- }
- /*
- * Call the helper to output message
- * FIXME: what to do if core_err_cnt > 1? Currently, it generates
- * only one event
- */
- if (uncorrected_error || !pvt->is_registered)
- edac_mc_handle_error(tp_event, mci, core_err_cnt,
- m->addr >> PAGE_SHIFT,
- m->addr & ~PAGE_MASK,
- syndrome,
- channel, dimm, -1,
- err, optype);
- }
- /*
- * i7core_check_error Retrieve and process errors reported by the
- * hardware. Called by the Core module.
- */
- static void i7core_check_error(struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- int i;
- unsigned count = 0;
- struct mce *m;
- /*
- * MCE first step: Copy all mce errors into a temporary buffer
- * We use a double buffering here, to reduce the risk of
- * losing an error.
- */
- smp_rmb();
- count = (pvt->mce_out + MCE_LOG_LEN - pvt->mce_in)
- % MCE_LOG_LEN;
- if (!count)
- goto check_ce_error;
- m = pvt->mce_outentry;
- if (pvt->mce_in + count > MCE_LOG_LEN) {
- unsigned l = MCE_LOG_LEN - pvt->mce_in;
- memcpy(m, &pvt->mce_entry[pvt->mce_in], sizeof(*m) * l);
- smp_wmb();
- pvt->mce_in = 0;
- count -= l;
- m += l;
- }
- memcpy(m, &pvt->mce_entry[pvt->mce_in], sizeof(*m) * count);
- smp_wmb();
- pvt->mce_in += count;
- smp_rmb();
- if (pvt->mce_overrun) {
- i7core_printk(KERN_ERR, "Lost %d memory errors\n",
- pvt->mce_overrun);
- smp_wmb();
- pvt->mce_overrun = 0;
- }
- /*
- * MCE second step: parse errors and display
- */
- for (i = 0; i < count; i++)
- i7core_mce_output_error(mci, &pvt->mce_outentry[i]);
- /*
- * Now, let's increment CE error counts
- */
- check_ce_error:
- if (!pvt->is_registered)
- i7core_udimm_check_mc_ecc_err(mci);
- else
- i7core_rdimm_check_mc_ecc_err(mci);
- }
- /*
- * i7core_mce_check_error Replicates mcelog routine to get errors
- * This routine simply queues mcelog errors, and
- * return. The error itself should be handled later
- * by i7core_check_error.
- * WARNING: As this routine should be called at NMI time, extra care should
- * be taken to avoid deadlocks, and to be as fast as possible.
- */
- static int i7core_mce_check_error(struct notifier_block *nb, unsigned long val,
- void *data)
- {
- struct mce *mce = (struct mce *)data;
- struct i7core_dev *i7_dev;
- struct mem_ctl_info *mci;
- struct i7core_pvt *pvt;
- i7_dev = get_i7core_dev(mce->socketid);
- if (!i7_dev)
- return NOTIFY_DONE;
- mci = i7_dev->mci;
- pvt = mci->pvt_info;
- /*
- * Just let mcelog handle it if the error is
- * outside the memory controller
- */
- if (((mce->status & 0xffff) >> 7) != 1)
- return NOTIFY_DONE;
- /* Bank 8 registers are the only ones that we know how to handle */
- if (mce->bank != 8)
- return NOTIFY_DONE;
- smp_rmb();
- if ((pvt->mce_out + 1) % MCE_LOG_LEN == pvt->mce_in) {
- smp_wmb();
- pvt->mce_overrun++;
- return NOTIFY_DONE;
- }
- /* Copy memory error at the ringbuffer */
- memcpy(&pvt->mce_entry[pvt->mce_out], mce, sizeof(*mce));
- smp_wmb();
- pvt->mce_out = (pvt->mce_out + 1) % MCE_LOG_LEN;
- /* Handle fatal errors immediately */
- if (mce->mcgstatus & 1)
- i7core_check_error(mci);
- /* Advise mcelog that the errors were handled */
- return NOTIFY_STOP;
- }
- static struct notifier_block i7_mce_dec = {
- .notifier_call = i7core_mce_check_error,
- };
- struct memdev_dmi_entry {
- u8 type;
- u8 length;
- u16 handle;
- u16 phys_mem_array_handle;
- u16 mem_err_info_handle;
- u16 total_width;
- u16 data_width;
- u16 size;
- u8 form;
- u8 device_set;
- u8 device_locator;
- u8 bank_locator;
- u8 memory_type;
- u16 type_detail;
- u16 speed;
- u8 manufacturer;
- u8 serial_number;
- u8 asset_tag;
- u8 part_number;
- u8 attributes;
- u32 extended_size;
- u16 conf_mem_clk_speed;
- } __attribute__((__packed__));
- /*
- * Decode the DRAM Clock Frequency, be paranoid, make sure that all
- * memory devices show the same speed, and if they don't then consider
- * all speeds to be invalid.
- */
- static void decode_dclk(const struct dmi_header *dh, void *_dclk_freq)
- {
- int *dclk_freq = _dclk_freq;
- u16 dmi_mem_clk_speed;
- if (*dclk_freq == -1)
- return;
- if (dh->type == DMI_ENTRY_MEM_DEVICE) {
- struct memdev_dmi_entry *memdev_dmi_entry =
- (struct memdev_dmi_entry *)dh;
- unsigned long conf_mem_clk_speed_offset =
- (unsigned long)&memdev_dmi_entry->conf_mem_clk_speed -
- (unsigned long)&memdev_dmi_entry->type;
- unsigned long speed_offset =
- (unsigned long)&memdev_dmi_entry->speed -
- (unsigned long)&memdev_dmi_entry->type;
- /* Check that a DIMM is present */
- if (memdev_dmi_entry->size == 0)
- return;
- /*
- * Pick the configured speed if it's available, otherwise
- * pick the DIMM speed, or we don't have a speed.
- */
- if (memdev_dmi_entry->length > conf_mem_clk_speed_offset) {
- dmi_mem_clk_speed =
- memdev_dmi_entry->conf_mem_clk_speed;
- } else if (memdev_dmi_entry->length > speed_offset) {
- dmi_mem_clk_speed = memdev_dmi_entry->speed;
- } else {
- *dclk_freq = -1;
- return;
- }
- if (*dclk_freq == 0) {
- /* First pass, speed was 0 */
- if (dmi_mem_clk_speed > 0) {
- /* Set speed if a valid speed is read */
- *dclk_freq = dmi_mem_clk_speed;
- } else {
- /* Otherwise we don't have a valid speed */
- *dclk_freq = -1;
- }
- } else if (*dclk_freq > 0 &&
- *dclk_freq != dmi_mem_clk_speed) {
- /*
- * If we have a speed, check that all DIMMS are the same
- * speed, otherwise set the speed as invalid.
- */
- *dclk_freq = -1;
- }
- }
- }
- /*
- * The default DCLK frequency is used as a fallback if we
- * fail to find anything reliable in the DMI. The value
- * is taken straight from the datasheet.
- */
- #define DEFAULT_DCLK_FREQ 800
- static int get_dclk_freq(void)
- {
- int dclk_freq = 0;
- dmi_walk(decode_dclk, (void *)&dclk_freq);
- if (dclk_freq < 1)
- return DEFAULT_DCLK_FREQ;
- return dclk_freq;
- }
- /*
- * set_sdram_scrub_rate This routine sets byte/sec bandwidth scrub rate
- * to hardware according to SCRUBINTERVAL formula
- * found in datasheet.
- */
- static int set_sdram_scrub_rate(struct mem_ctl_info *mci, u32 new_bw)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- struct pci_dev *pdev;
- u32 dw_scrub;
- u32 dw_ssr;
- /* Get data from the MC register, function 2 */
- pdev = pvt->pci_mcr[2];
- if (!pdev)
- return -ENODEV;
- pci_read_config_dword(pdev, MC_SCRUB_CONTROL, &dw_scrub);
- if (new_bw == 0) {
- /* Prepare to disable petrol scrub */
- dw_scrub &= ~STARTSCRUB;
- /* Stop the patrol scrub engine */
- write_and_test(pdev, MC_SCRUB_CONTROL,
- dw_scrub & ~SCRUBINTERVAL_MASK);
- /* Get current status of scrub rate and set bit to disable */
- pci_read_config_dword(pdev, MC_SSRCONTROL, &dw_ssr);
- dw_ssr &= ~SSR_MODE_MASK;
- dw_ssr |= SSR_MODE_DISABLE;
- } else {
- const int cache_line_size = 64;
- const u32 freq_dclk_mhz = pvt->dclk_freq;
- unsigned long long scrub_interval;
- /*
- * Translate the desired scrub rate to a register value and
- * program the corresponding register value.
- */
- scrub_interval = (unsigned long long)freq_dclk_mhz *
- cache_line_size * 1000000;
- do_div(scrub_interval, new_bw);
- if (!scrub_interval || scrub_interval > SCRUBINTERVAL_MASK)
- return -EINVAL;
- dw_scrub = SCRUBINTERVAL_MASK & scrub_interval;
- /* Start the patrol scrub engine */
- pci_write_config_dword(pdev, MC_SCRUB_CONTROL,
- STARTSCRUB | dw_scrub);
- /* Get current status of scrub rate and set bit to enable */
- pci_read_config_dword(pdev, MC_SSRCONTROL, &dw_ssr);
- dw_ssr &= ~SSR_MODE_MASK;
- dw_ssr |= SSR_MODE_ENABLE;
- }
- /* Disable or enable scrubbing */
- pci_write_config_dword(pdev, MC_SSRCONTROL, dw_ssr);
- return new_bw;
- }
- /*
- * get_sdram_scrub_rate This routine convert current scrub rate value
- * into byte/sec bandwidth according to
- * SCRUBINTERVAL formula found in datasheet.
- */
- static int get_sdram_scrub_rate(struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- struct pci_dev *pdev;
- const u32 cache_line_size = 64;
- const u32 freq_dclk_mhz = pvt->dclk_freq;
- unsigned long long scrub_rate;
- u32 scrubval;
- /* Get data from the MC register, function 2 */
- pdev = pvt->pci_mcr[2];
- if (!pdev)
- return -ENODEV;
- /* Get current scrub control data */
- pci_read_config_dword(pdev, MC_SCRUB_CONTROL, &scrubval);
- /* Mask highest 8-bits to 0 */
- scrubval &= SCRUBINTERVAL_MASK;
- if (!scrubval)
- return 0;
- /* Calculate scrub rate value into byte/sec bandwidth */
- scrub_rate = (unsigned long long)freq_dclk_mhz *
- 1000000 * cache_line_size;
- do_div(scrub_rate, scrubval);
- return (int)scrub_rate;
- }
- static void enable_sdram_scrub_setting(struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- u32 pci_lock;
- /* Unlock writes to pci registers */
- pci_read_config_dword(pvt->pci_noncore, MC_CFG_CONTROL, &pci_lock);
- pci_lock &= ~0x3;
- pci_write_config_dword(pvt->pci_noncore, MC_CFG_CONTROL,
- pci_lock | MC_CFG_UNLOCK);
- mci->set_sdram_scrub_rate = set_sdram_scrub_rate;
- mci->get_sdram_scrub_rate = get_sdram_scrub_rate;
- }
- static void disable_sdram_scrub_setting(struct mem_ctl_info *mci)
- {
- struct i7core_pvt *pvt = mci->pvt_info;
- u32 pci_lock;
- /* Lock writes to pci registers */
- pci_read_config_dword(pvt->pci_noncore, MC_CFG_CONTROL, &pci_lock);
- pci_lock &= ~0x3;
- pci_write_config_dword(pvt->pci_noncore, MC_CFG_CONTROL,
- pci_lock | MC_CFG_LOCK);
- }
- static void i7core_pci_ctl_create(struct i7core_pvt *pvt)
- {
- pvt->i7core_pci = edac_pci_create_generic_ctl(
- &pvt->i7core_dev->pdev[0]->dev,
- EDAC_MOD_STR);
- if (unlikely(!pvt->i7core_pci))
- i7core_printk(KERN_WARNING,
- "Unable to setup PCI error report via EDAC\n");
- }
- static void i7core_pci_ctl_release(struct i7core_pvt *pvt)
- {
- if (likely(pvt->i7core_pci))
- edac_pci_release_generic_ctl(pvt->i7core_pci);
- else
- i7core_printk(KERN_ERR,
- "Couldn't find mem_ctl_info for socket %d\n",
- pvt->i7core_dev->socket);
- pvt->i7core_pci = NULL;
- }
- static void i7core_unregister_mci(struct i7core_dev *i7core_dev)
- {
- struct mem_ctl_info *mci = i7core_dev->mci;
- struct i7core_pvt *pvt;
- if (unlikely(!mci || !mci->pvt_info)) {
- edac_dbg(0, "MC: dev = %p\n", &i7core_dev->pdev[0]->dev);
- i7core_printk(KERN_ERR, "Couldn't find mci handler\n");
- return;
- }
- pvt = mci->pvt_info;
- edac_dbg(0, "MC: mci = %p, dev = %p\n", mci, &i7core_dev->pdev[0]->dev);
- /* Disable scrubrate setting */
- if (pvt->enable_scrub)
- disable_sdram_scrub_setting(mci);
- /* Disable EDAC polling */
- i7core_pci_ctl_release(pvt);
- /* Remove MC sysfs nodes */
- i7core_delete_sysfs_devices(mci);
- edac_mc_del_mc(mci->pdev);
- edac_dbg(1, "%s: free mci struct\n", mci->ctl_name);
- kfree(mci->ctl_name);
- edac_mc_free(mci);
- i7core_dev->mci = NULL;
- }
- static int i7core_register_mci(struct i7core_dev *i7core_dev)
- {
- struct mem_ctl_info *mci;
- struct i7core_pvt *pvt;
- int rc;
- struct edac_mc_layer layers[2];
- /* allocate a new MC control structure */
- layers[0].type = EDAC_MC_LAYER_CHANNEL;
- layers[0].size = NUM_CHANS;
- layers[0].is_virt_csrow = false;
- layers[1].type = EDAC_MC_LAYER_SLOT;
- layers[1].size = MAX_DIMMS;
- layers[1].is_virt_csrow = true;
- mci = edac_mc_alloc(i7core_dev->socket, ARRAY_SIZE(layers), layers,
- sizeof(*pvt));
- if (unlikely(!mci))
- return -ENOMEM;
- edac_dbg(0, "MC: mci = %p, dev = %p\n", mci, &i7core_dev->pdev[0]->dev);
- pvt = mci->pvt_info;
- memset(pvt, 0, sizeof(*pvt));
- /* Associates i7core_dev and mci for future usage */
- pvt->i7core_dev = i7core_dev;
- i7core_dev->mci = mci;
- /*
- * FIXME: how to handle RDDR3 at MCI level? It is possible to have
- * Mixed RDDR3/UDDR3 with Nehalem, provided that they are on different
- * memory channels
- */
- mci->mtype_cap = MEM_FLAG_DDR3;
- mci->edac_ctl_cap = EDAC_FLAG_NONE;
- mci->edac_cap = EDAC_FLAG_NONE;
- mci->mod_name = "i7core_edac.c";
- mci->mod_ver = I7CORE_REVISION;
- mci->ctl_name = kasprintf(GFP_KERNEL, "i7 core #%d",
- i7core_dev->socket);
- mci->dev_name = pci_name(i7core_dev->pdev[0]);
- mci->ctl_page_to_phys = NULL;
- /* Store pci devices at mci for faster access */
- rc = mci_bind_devs(mci, i7core_dev);
- if (unlikely(rc < 0))
- goto fail0;
- /* Get dimm basic config */
- get_dimm_config(mci);
- /* record ptr to the generic device */
- mci->pdev = &i7core_dev->pdev[0]->dev;
- /* Set the function pointer to an actual operation function */
- mci->edac_check = i7core_check_error;
- /* Enable scrubrate setting */
- if (pvt->enable_scrub)
- enable_sdram_scrub_setting(mci);
- /* add this new MC control structure to EDAC's list of MCs */
- if (unlikely(edac_mc_add_mc_with_groups(mci, i7core_dev_groups))) {
- edac_dbg(0, "MC: failed edac_mc_add_mc()\n");
- /* FIXME: perhaps some code should go here that disables error
- * reporting if we just enabled it
- */
- rc = -EINVAL;
- goto fail0;
- }
- if (i7core_create_sysfs_devices(mci)) {
- edac_dbg(0, "MC: failed to create sysfs nodes\n");
- edac_mc_del_mc(mci->pdev);
- rc = -EINVAL;
- goto fail0;
- }
- /* Default error mask is any memory */
- pvt->inject.channel = 0;
- pvt->inject.dimm = -1;
- pvt->inject.rank = -1;
- pvt->inject.bank = -1;
- pvt->inject.page = -1;
- pvt->inject.col = -1;
- /* allocating generic PCI control info */
- i7core_pci_ctl_create(pvt);
- /* DCLK for scrub rate setting */
- pvt->dclk_freq = get_dclk_freq();
- return 0;
- fail0:
- kfree(mci->ctl_name);
- edac_mc_free(mci);
- i7core_dev->mci = NULL;
- return rc;
- }
- /*
- * i7core_probe Probe for ONE instance of device to see if it is
- * present.
- * return:
- * 0 for FOUND a device
- * < 0 for error code
- */
- static int i7core_probe(struct pci_dev *pdev, const struct pci_device_id *id)
- {
- int rc, count = 0;
- struct i7core_dev *i7core_dev;
- /* get the pci devices we want to reserve for our use */
- mutex_lock(&i7core_edac_lock);
- /*
- * All memory controllers are allocated at the first pass.
- */
- if (unlikely(probed >= 1)) {
- mutex_unlock(&i7core_edac_lock);
- return -ENODEV;
- }
- probed++;
- rc = i7core_get_all_devices();
- if (unlikely(rc < 0))
- goto fail0;
- list_for_each_entry(i7core_dev, &i7core_edac_list, list) {
- count++;
- rc = i7core_register_mci(i7core_dev);
- if (unlikely(rc < 0))
- goto fail1;
- }
- /*
- * Nehalem-EX uses a different memory controller. However, as the
- * memory controller is not visible on some Nehalem/Nehalem-EP, we
- * need to indirectly probe via a X58 PCI device. The same devices
- * are found on (some) Nehalem-EX. So, on those machines, the
- * probe routine needs to return -ENODEV, as the actual Memory
- * Controller registers won't be detected.
- */
- if (!count) {
- rc = -ENODEV;
- goto fail1;
- }
- i7core_printk(KERN_INFO,
- "Driver loaded, %d memory controller(s) found.\n",
- count);
- mutex_unlock(&i7core_edac_lock);
- return 0;
- fail1:
- list_for_each_entry(i7core_dev, &i7core_edac_list, list)
- i7core_unregister_mci(i7core_dev);
- i7core_put_all_devices();
- fail0:
- mutex_unlock(&i7core_edac_lock);
- return rc;
- }
- /*
- * i7core_remove destructor for one instance of device
- *
- */
- static void i7core_remove(struct pci_dev *pdev)
- {
- struct i7core_dev *i7core_dev;
- edac_dbg(0, "\n");
- /*
- * we have a trouble here: pdev value for removal will be wrong, since
- * it will point to the X58 register used to detect that the machine
- * is a Nehalem or upper design. However, due to the way several PCI
- * devices are grouped together to provide MC functionality, we need
- * to use a different method for releasing the devices
- */
- mutex_lock(&i7core_edac_lock);
- if (unlikely(!probed)) {
- mutex_unlock(&i7core_edac_lock);
- return;
- }
- list_for_each_entry(i7core_dev, &i7core_edac_list, list)
- i7core_unregister_mci(i7core_dev);
- /* Release PCI resources */
- i7core_put_all_devices();
- probed--;
- mutex_unlock(&i7core_edac_lock);
- }
- MODULE_DEVICE_TABLE(pci, i7core_pci_tbl);
- /*
- * i7core_driver pci_driver structure for this module
- *
- */
- static struct pci_driver i7core_driver = {
- .name = "i7core_edac",
- .probe = i7core_probe,
- .remove = i7core_remove,
- .id_table = i7core_pci_tbl,
- };
- /*
- * i7core_init Module entry function
- * Try to initialize this module for its devices
- */
- static int __init i7core_init(void)
- {
- int pci_rc;
- edac_dbg(2, "\n");
- /* Ensure that the OPSTATE is set correctly for POLL or NMI */
- opstate_init();
- if (use_pci_fixup)
- i7core_xeon_pci_fixup(pci_dev_table);
- pci_rc = pci_register_driver(&i7core_driver);
- if (pci_rc >= 0) {
- mce_register_decode_chain(&i7_mce_dec);
- return 0;
- }
- i7core_printk(KERN_ERR, "Failed to register device with error %d.\n",
- pci_rc);
- return pci_rc;
- }
- /*
- * i7core_exit() Module exit function
- * Unregister the driver
- */
- static void __exit i7core_exit(void)
- {
- edac_dbg(2, "\n");
- pci_unregister_driver(&i7core_driver);
- mce_unregister_decode_chain(&i7_mce_dec);
- }
- module_init(i7core_init);
- module_exit(i7core_exit);
- MODULE_LICENSE("GPL");
- MODULE_AUTHOR("Mauro Carvalho Chehab");
- MODULE_AUTHOR("Red Hat Inc. (http://www.redhat.com)");
- MODULE_DESCRIPTION("MC Driver for Intel i7 Core memory controllers - "
- I7CORE_REVISION);
- module_param(edac_op_state, int, 0444);
- MODULE_PARM_DESC(edac_op_state, "EDAC Error Reporting state: 0=Poll,1=NMI");
|