2 * edac_mc kernel module
3 * (C) 2005, 2006 Linux Networx (http://lnxi.com)
4 * This file may be distributed under the terms of the
5 * GNU General Public License.
7 * Written by Thayne Harbaugh
8 * Based on work by Dan Hollis <goemon at anime dot net> and others.
9 * http://www.anime.net/~goemon/linux-ecc/
11 * Modified by Dave Peterson and Doug Thompson
15 #include <linux/module.h>
16 #include <linux/proc_fs.h>
17 #include <linux/kernel.h>
18 #include <linux/types.h>
19 #include <linux/smp.h>
20 #include <linux/init.h>
21 #include <linux/sysctl.h>
22 #include <linux/highmem.h>
23 #include <linux/timer.h>
24 #include <linux/slab.h>
25 #include <linux/jiffies.h>
26 #include <linux/spinlock.h>
27 #include <linux/list.h>
28 #include <linux/sysdev.h>
29 #include <linux/ctype.h>
30 #include <linux/edac.h>
31 #include <asm/uaccess.h>
34 #include "edac_core.h"
35 #include "edac_module.h"
37 /* lock to memory controller's control array */
38 static DEFINE_MUTEX(mem_ctls_mutex);
39 static LIST_HEAD(mc_devices);
41 #ifdef CONFIG_EDAC_DEBUG
43 static void edac_mc_dump_channel(struct channel_info *chan)
45 debugf4("\tchannel = %p\n", chan);
46 debugf4("\tchannel->chan_idx = %d\n", chan->chan_idx);
47 debugf4("\tchannel->ce_count = %d\n", chan->ce_count);
48 debugf4("\tchannel->label = '%s'\n", chan->label);
49 debugf4("\tchannel->csrow = %p\n\n", chan->csrow);
52 static void edac_mc_dump_csrow(struct csrow_info *csrow)
54 debugf4("\tcsrow = %p\n", csrow);
55 debugf4("\tcsrow->csrow_idx = %d\n", csrow->csrow_idx);
56 debugf4("\tcsrow->first_page = 0x%lx\n", csrow->first_page);
57 debugf4("\tcsrow->last_page = 0x%lx\n", csrow->last_page);
58 debugf4("\tcsrow->page_mask = 0x%lx\n", csrow->page_mask);
59 debugf4("\tcsrow->nr_pages = 0x%x\n", csrow->nr_pages);
60 debugf4("\tcsrow->nr_channels = %d\n", csrow->nr_channels);
61 debugf4("\tcsrow->channels = %p\n", csrow->channels);
62 debugf4("\tcsrow->mci = %p\n\n", csrow->mci);
65 static void edac_mc_dump_mci(struct mem_ctl_info *mci)
67 debugf3("\tmci = %p\n", mci);
68 debugf3("\tmci->mtype_cap = %lx\n", mci->mtype_cap);
69 debugf3("\tmci->edac_ctl_cap = %lx\n", mci->edac_ctl_cap);
70 debugf3("\tmci->edac_cap = %lx\n", mci->edac_cap);
71 debugf4("\tmci->edac_check = %p\n", mci->edac_check);
72 debugf3("\tmci->nr_csrows = %d, csrows = %p\n",
73 mci->nr_csrows, mci->csrows);
74 debugf3("\tdev = %p\n", mci->dev);
75 debugf3("\tmod_name:ctl_name = %s:%s\n", mci->mod_name, mci->ctl_name);
76 debugf3("\tpvt_info = %p\n\n", mci->pvt_info);
80 * keep those in sync with the enum mem_type
82 const char *edac_mem_types[] = {
84 "Reserved csrow type",
87 "Extended data out RAM",
88 "Burst Extended data out RAM",
89 "Single data rate SDRAM",
90 "Registered single data rate SDRAM",
91 "Double data rate SDRAM",
92 "Registered Double data rate SDRAM",
94 "Unbuffered DDR2 RAM",
95 "Fully buffered DDR2",
96 "Registered DDR2 RAM",
98 "Unbuffered DDR3 RAM",
99 "Registered DDR3 RAM",
101 EXPORT_SYMBOL_GPL(edac_mem_types);
103 #endif /* CONFIG_EDAC_DEBUG */
105 /* 'ptr' points to a possibly unaligned item X such that sizeof(X) is 'size'.
106 * Adjust 'ptr' so that its alignment is at least as stringent as what the
107 * compiler would provide for X and return the aligned result.
109 * If 'size' is a constant, the compiler will optimize this whole function
110 * down to either a no-op or the addition of a constant to the value of 'ptr'.
112 void *edac_align_ptr(void *ptr, unsigned size)
116 /* Here we assume that the alignment of a "long long" is the most
117 * stringent alignment that the compiler will ever provide by default.
118 * As far as I know, this is a reasonable assumption.
120 if (size > sizeof(long))
121 align = sizeof(long long);
122 else if (size > sizeof(int))
123 align = sizeof(long);
124 else if (size > sizeof(short))
126 else if (size > sizeof(char))
127 align = sizeof(short);
136 return (void *)(((unsigned long)ptr) + align - r);
140 * edac_mc_alloc: Allocate a struct mem_ctl_info structure
141 * @size_pvt: size of private storage needed
142 * @nr_csrows: Number of CWROWS needed for this MC
143 * @nr_chans: Number of channels for the MC
145 * Everything is kmalloc'ed as one big chunk - more efficient.
146 * Only can be used if all structures have the same lifetime - otherwise
147 * you have to allocate and initialize your own structures.
149 * Use edac_mc_free() to free mc structures allocated by this function.
152 * NULL allocation failed
153 * struct mem_ctl_info pointer
155 struct mem_ctl_info *edac_mc_alloc(unsigned sz_pvt, unsigned nr_csrows,
156 unsigned nr_chans, int edac_index)
158 struct mem_ctl_info *mci;
159 struct csrow_info *csi, *csrow;
160 struct channel_info *chi, *chp, *chan;
166 /* Figure out the offsets of the various items from the start of an mc
167 * structure. We want the alignment of each item to be at least as
168 * stringent as what the compiler would provide if we could simply
169 * hardcode everything into a single struct.
171 mci = (struct mem_ctl_info *)0;
172 csi = edac_align_ptr(&mci[1], sizeof(*csi));
173 chi = edac_align_ptr(&csi[nr_csrows], sizeof(*chi));
174 pvt = edac_align_ptr(&chi[nr_chans * nr_csrows], sz_pvt);
175 size = ((unsigned long)pvt) + sz_pvt;
177 mci = kzalloc(size, GFP_KERNEL);
181 /* Adjust pointers so they point within the memory we just allocated
182 * rather than an imaginary chunk of memory located at address 0.
184 csi = (struct csrow_info *)(((char *)mci) + ((unsigned long)csi));
185 chi = (struct channel_info *)(((char *)mci) + ((unsigned long)chi));
186 pvt = sz_pvt ? (((char *)mci) + ((unsigned long)pvt)) : NULL;
188 /* setup index and various internal pointers */
189 mci->mc_idx = edac_index;
192 mci->nr_csrows = nr_csrows;
194 for (row = 0; row < nr_csrows; row++) {
196 csrow->csrow_idx = row;
198 csrow->nr_channels = nr_chans;
199 chp = &chi[row * nr_chans];
200 csrow->channels = chp;
202 for (chn = 0; chn < nr_chans; chn++) {
204 chan->chan_idx = chn;
209 mci->op_state = OP_ALLOC;
212 * Initialize the 'root' kobj for the edac_mc controller
214 err = edac_mc_register_sysfs_main_kobj(mci);
220 /* at this point, the root kobj is valid, and in order to
221 * 'free' the object, then the function:
222 * edac_mc_unregister_sysfs_main_kobj() must be called
223 * which will perform kobj unregistration and the actual free
224 * will occur during the kobject callback operation
228 EXPORT_SYMBOL_GPL(edac_mc_alloc);
232 * 'Free' a previously allocated 'mci' structure
233 * @mci: pointer to a struct mem_ctl_info structure
235 void edac_mc_free(struct mem_ctl_info *mci)
237 edac_mc_unregister_sysfs_main_kobj(mci);
239 EXPORT_SYMBOL_GPL(edac_mc_free);
245 * scan list of controllers looking for the one that manages
247 * @dev: pointer to a struct device related with the MCI
249 struct mem_ctl_info *find_mci_by_dev(struct device *dev)
251 struct mem_ctl_info *mci;
252 struct list_head *item;
254 debugf3("%s()\n", __func__);
256 list_for_each(item, &mc_devices) {
257 mci = list_entry(item, struct mem_ctl_info, link);
265 EXPORT_SYMBOL_GPL(find_mci_by_dev);
268 * handler for EDAC to check if NMI type handler has asserted interrupt
270 static int edac_mc_assert_error_check_and_clear(void)
274 if (edac_op_state == EDAC_OPSTATE_POLL)
277 old_state = edac_err_assert;
284 * edac_mc_workq_function
285 * performs the operation scheduled by a workq request
287 static void edac_mc_workq_function(struct work_struct *work_req)
289 struct delayed_work *d_work = to_delayed_work(work_req);
290 struct mem_ctl_info *mci = to_edac_mem_ctl_work(d_work);
292 mutex_lock(&mem_ctls_mutex);
294 /* if this control struct has movd to offline state, we are done */
295 if (mci->op_state == OP_OFFLINE) {
296 mutex_unlock(&mem_ctls_mutex);
300 /* Only poll controllers that are running polled and have a check */
301 if (edac_mc_assert_error_check_and_clear() && (mci->edac_check != NULL))
302 mci->edac_check(mci);
304 mutex_unlock(&mem_ctls_mutex);
307 queue_delayed_work(edac_workqueue, &mci->work,
308 msecs_to_jiffies(edac_mc_get_poll_msec()));
312 * edac_mc_workq_setup
313 * initialize a workq item for this mci
314 * passing in the new delay period in msec
318 * called with the mem_ctls_mutex held
320 static void edac_mc_workq_setup(struct mem_ctl_info *mci, unsigned msec)
322 debugf0("%s()\n", __func__);
324 /* if this instance is not in the POLL state, then simply return */
325 if (mci->op_state != OP_RUNNING_POLL)
328 INIT_DELAYED_WORK(&mci->work, edac_mc_workq_function);
329 queue_delayed_work(edac_workqueue, &mci->work, msecs_to_jiffies(msec));
333 * edac_mc_workq_teardown
334 * stop the workq processing on this mci
338 * called WITHOUT lock held
340 static void edac_mc_workq_teardown(struct mem_ctl_info *mci)
344 if (mci->op_state != OP_RUNNING_POLL)
347 status = cancel_delayed_work(&mci->work);
349 debugf0("%s() not canceled, flush the queue\n",
352 /* workq instance might be running, wait for it */
353 flush_workqueue(edac_workqueue);
358 * edac_mc_reset_delay_period(unsigned long value)
360 * user space has updated our poll period value, need to
361 * reset our workq delays
363 void edac_mc_reset_delay_period(int value)
365 struct mem_ctl_info *mci;
366 struct list_head *item;
368 mutex_lock(&mem_ctls_mutex);
370 /* scan the list and turn off all workq timers, doing so under lock
372 list_for_each(item, &mc_devices) {
373 mci = list_entry(item, struct mem_ctl_info, link);
375 if (mci->op_state == OP_RUNNING_POLL)
376 cancel_delayed_work(&mci->work);
379 mutex_unlock(&mem_ctls_mutex);
382 /* re-walk the list, and reset the poll delay */
383 mutex_lock(&mem_ctls_mutex);
385 list_for_each(item, &mc_devices) {
386 mci = list_entry(item, struct mem_ctl_info, link);
388 edac_mc_workq_setup(mci, (unsigned long) value);
391 mutex_unlock(&mem_ctls_mutex);
396 /* Return 0 on success, 1 on failure.
397 * Before calling this function, caller must
398 * assign a unique value to mci->mc_idx.
402 * called with the mem_ctls_mutex lock held
404 static int add_mc_to_global_list(struct mem_ctl_info *mci)
406 struct list_head *item, *insert_before;
407 struct mem_ctl_info *p;
409 insert_before = &mc_devices;
411 p = find_mci_by_dev(mci->dev);
412 if (unlikely(p != NULL))
415 list_for_each(item, &mc_devices) {
416 p = list_entry(item, struct mem_ctl_info, link);
418 if (p->mc_idx >= mci->mc_idx) {
419 if (unlikely(p->mc_idx == mci->mc_idx))
422 insert_before = item;
427 list_add_tail_rcu(&mci->link, insert_before);
428 atomic_inc(&edac_handlers);
432 edac_printk(KERN_WARNING, EDAC_MC,
433 "%s (%s) %s %s already assigned %d\n", dev_name(p->dev),
434 edac_dev_name(mci), p->mod_name, p->ctl_name, p->mc_idx);
438 edac_printk(KERN_WARNING, EDAC_MC,
439 "bug in low-level driver: attempt to assign\n"
440 " duplicate mc_idx %d in %s()\n", p->mc_idx, __func__);
444 static void complete_mc_list_del(struct rcu_head *head)
446 struct mem_ctl_info *mci;
448 mci = container_of(head, struct mem_ctl_info, rcu);
449 INIT_LIST_HEAD(&mci->link);
452 static void del_mc_from_global_list(struct mem_ctl_info *mci)
454 atomic_dec(&edac_handlers);
455 list_del_rcu(&mci->link);
456 call_rcu(&mci->rcu, complete_mc_list_del);
461 * edac_mc_find: Search for a mem_ctl_info structure whose index is 'idx'.
463 * If found, return a pointer to the structure.
466 * Caller must hold mem_ctls_mutex.
468 struct mem_ctl_info *edac_mc_find(int idx)
470 struct list_head *item;
471 struct mem_ctl_info *mci;
473 list_for_each(item, &mc_devices) {
474 mci = list_entry(item, struct mem_ctl_info, link);
476 if (mci->mc_idx >= idx) {
477 if (mci->mc_idx == idx)
486 EXPORT_SYMBOL(edac_mc_find);
489 * edac_mc_add_mc: Insert the 'mci' structure into the mci global list and
490 * create sysfs entries associated with mci structure
491 * @mci: pointer to the mci structure to be added to the list
492 * @mc_idx: A unique numeric identifier to be assigned to the 'mci' structure.
499 /* FIXME - should a warning be printed if no error detection? correction? */
500 int edac_mc_add_mc(struct mem_ctl_info *mci)
502 debugf0("%s()\n", __func__);
504 #ifdef CONFIG_EDAC_DEBUG
505 if (edac_debug_level >= 3)
506 edac_mc_dump_mci(mci);
508 if (edac_debug_level >= 4) {
511 for (i = 0; i < mci->nr_csrows; i++) {
514 edac_mc_dump_csrow(&mci->csrows[i]);
515 for (j = 0; j < mci->csrows[i].nr_channels; j++)
516 edac_mc_dump_channel(&mci->csrows[i].
521 mutex_lock(&mem_ctls_mutex);
523 if (add_mc_to_global_list(mci))
526 /* set load time so that error rate can be tracked */
527 mci->start_time = jiffies;
529 if (edac_create_sysfs_mci_device(mci)) {
530 edac_mc_printk(mci, KERN_WARNING,
531 "failed to create sysfs device\n");
535 /* If there IS a check routine, then we are running POLLED */
536 if (mci->edac_check != NULL) {
537 /* This instance is NOW RUNNING */
538 mci->op_state = OP_RUNNING_POLL;
540 edac_mc_workq_setup(mci, edac_mc_get_poll_msec());
542 mci->op_state = OP_RUNNING_INTERRUPT;
545 /* Report action taken */
546 edac_mc_printk(mci, KERN_INFO, "Giving out device to '%s' '%s':"
547 " DEV %s\n", mci->mod_name, mci->ctl_name, edac_dev_name(mci));
549 mutex_unlock(&mem_ctls_mutex);
553 del_mc_from_global_list(mci);
556 mutex_unlock(&mem_ctls_mutex);
559 EXPORT_SYMBOL_GPL(edac_mc_add_mc);
562 * edac_mc_del_mc: Remove sysfs entries for specified mci structure and
563 * remove mci structure from global list
564 * @pdev: Pointer to 'struct device' representing mci structure to remove.
566 * Return pointer to removed mci structure, or NULL if device not found.
568 struct mem_ctl_info *edac_mc_del_mc(struct device *dev)
570 struct mem_ctl_info *mci;
572 debugf0("%s()\n", __func__);
574 mutex_lock(&mem_ctls_mutex);
576 /* find the requested mci struct in the global list */
577 mci = find_mci_by_dev(dev);
579 mutex_unlock(&mem_ctls_mutex);
583 /* marking MCI offline */
584 mci->op_state = OP_OFFLINE;
586 del_mc_from_global_list(mci);
587 mutex_unlock(&mem_ctls_mutex);
589 /* flush workq processes and remove sysfs */
590 edac_mc_workq_teardown(mci);
591 edac_remove_sysfs_mci_device(mci);
593 edac_printk(KERN_INFO, EDAC_MC,
594 "Removed device %d for %s %s: DEV %s\n", mci->mc_idx,
595 mci->mod_name, mci->ctl_name, edac_dev_name(mci));
599 EXPORT_SYMBOL_GPL(edac_mc_del_mc);
601 static void edac_mc_scrub_block(unsigned long page, unsigned long offset,
606 unsigned long flags = 0;
608 debugf3("%s()\n", __func__);
610 /* ECC error page was not in our memory. Ignore it. */
611 if (!pfn_valid(page))
614 /* Find the actual page structure then map it and fix */
615 pg = pfn_to_page(page);
618 local_irq_save(flags);
620 virt_addr = kmap_atomic(pg, KM_BOUNCE_READ);
622 /* Perform architecture specific atomic scrub operation */
623 atomic_scrub(virt_addr + offset, size);
625 /* Unmap and complete */
626 kunmap_atomic(virt_addr, KM_BOUNCE_READ);
629 local_irq_restore(flags);
632 /* FIXME - should return -1 */
633 int edac_mc_find_csrow_by_page(struct mem_ctl_info *mci, unsigned long page)
635 struct csrow_info *csrows = mci->csrows;
638 debugf1("MC%d: %s(): 0x%lx\n", mci->mc_idx, __func__, page);
641 for (i = 0; i < mci->nr_csrows; i++) {
642 struct csrow_info *csrow = &csrows[i];
644 if (csrow->nr_pages == 0)
647 debugf3("MC%d: %s(): first(0x%lx) page(0x%lx) last(0x%lx) "
648 "mask(0x%lx)\n", mci->mc_idx, __func__,
649 csrow->first_page, page, csrow->last_page,
652 if ((page >= csrow->first_page) &&
653 (page <= csrow->last_page) &&
654 ((page & csrow->page_mask) ==
655 (csrow->first_page & csrow->page_mask))) {
662 edac_mc_printk(mci, KERN_ERR,
663 "could not look up page error address %lx\n",
664 (unsigned long)page);
668 EXPORT_SYMBOL_GPL(edac_mc_find_csrow_by_page);
670 /* FIXME - setable log (warning/emerg) levels */
671 /* FIXME - integrate with evlog: http://evlog.sourceforge.net/ */
672 void edac_mc_handle_ce(struct mem_ctl_info *mci,
673 unsigned long page_frame_number,
674 unsigned long offset_in_page, unsigned long syndrome,
675 int row, int channel, const char *msg)
677 unsigned long remapped_page;
679 debugf3("MC%d: %s()\n", mci->mc_idx, __func__);
681 /* FIXME - maybe make panic on INTERNAL ERROR an option */
682 if (row >= mci->nr_csrows || row < 0) {
683 /* something is wrong */
684 edac_mc_printk(mci, KERN_ERR,
685 "INTERNAL ERROR: row out of range "
686 "(%d >= %d)\n", row, mci->nr_csrows);
687 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
691 if (channel >= mci->csrows[row].nr_channels || channel < 0) {
692 /* something is wrong */
693 edac_mc_printk(mci, KERN_ERR,
694 "INTERNAL ERROR: channel out of range "
695 "(%d >= %d)\n", channel,
696 mci->csrows[row].nr_channels);
697 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
701 if (edac_mc_get_log_ce())
702 /* FIXME - put in DIMM location */
703 edac_mc_printk(mci, KERN_WARNING,
704 "CE page 0x%lx, offset 0x%lx, grain %d, syndrome "
705 "0x%lx, row %d, channel %d, label \"%s\": %s\n",
706 page_frame_number, offset_in_page,
707 mci->csrows[row].grain, syndrome, row, channel,
708 mci->csrows[row].channels[channel].label, msg);
711 mci->csrows[row].ce_count++;
712 mci->csrows[row].channels[channel].ce_count++;
714 if (mci->scrub_mode & SCRUB_SW_SRC) {
716 * Some MC's can remap memory so that it is still available
717 * at a different address when PCI devices map into memory.
718 * MC's that can't do this lose the memory where PCI devices
719 * are mapped. This mapping is MC dependant and so we call
720 * back into the MC driver for it to map the MC page to
721 * a physical (CPU) page which can then be mapped to a virtual
722 * page - which can then be scrubbed.
724 remapped_page = mci->ctl_page_to_phys ?
725 mci->ctl_page_to_phys(mci, page_frame_number) :
728 edac_mc_scrub_block(remapped_page, offset_in_page,
729 mci->csrows[row].grain);
732 EXPORT_SYMBOL_GPL(edac_mc_handle_ce);
734 void edac_mc_handle_ce_no_info(struct mem_ctl_info *mci, const char *msg)
736 if (edac_mc_get_log_ce())
737 edac_mc_printk(mci, KERN_WARNING,
738 "CE - no information available: %s\n", msg);
740 mci->ce_noinfo_count++;
743 EXPORT_SYMBOL_GPL(edac_mc_handle_ce_no_info);
745 void edac_mc_handle_ue(struct mem_ctl_info *mci,
746 unsigned long page_frame_number,
747 unsigned long offset_in_page, int row, const char *msg)
749 int len = EDAC_MC_LABEL_LEN * 4;
750 char labels[len + 1];
755 debugf3("MC%d: %s()\n", mci->mc_idx, __func__);
757 /* FIXME - maybe make panic on INTERNAL ERROR an option */
758 if (row >= mci->nr_csrows || row < 0) {
759 /* something is wrong */
760 edac_mc_printk(mci, KERN_ERR,
761 "INTERNAL ERROR: row out of range "
762 "(%d >= %d)\n", row, mci->nr_csrows);
763 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
767 chars = snprintf(pos, len + 1, "%s",
768 mci->csrows[row].channels[0].label);
772 for (chan = 1; (chan < mci->csrows[row].nr_channels) && (len > 0);
774 chars = snprintf(pos, len + 1, ":%s",
775 mci->csrows[row].channels[chan].label);
780 if (edac_mc_get_log_ue())
781 edac_mc_printk(mci, KERN_EMERG,
782 "UE page 0x%lx, offset 0x%lx, grain %d, row %d, "
783 "labels \"%s\": %s\n", page_frame_number,
784 offset_in_page, mci->csrows[row].grain, row,
787 if (edac_mc_get_panic_on_ue())
788 panic("EDAC MC%d: UE page 0x%lx, offset 0x%lx, grain %d, "
789 "row %d, labels \"%s\": %s\n", mci->mc_idx,
790 page_frame_number, offset_in_page,
791 mci->csrows[row].grain, row, labels, msg);
794 mci->csrows[row].ue_count++;
796 EXPORT_SYMBOL_GPL(edac_mc_handle_ue);
798 void edac_mc_handle_ue_no_info(struct mem_ctl_info *mci, const char *msg)
800 if (edac_mc_get_panic_on_ue())
801 panic("EDAC MC%d: Uncorrected Error", mci->mc_idx);
803 if (edac_mc_get_log_ue())
804 edac_mc_printk(mci, KERN_WARNING,
805 "UE - no information available: %s\n", msg);
806 mci->ue_noinfo_count++;
809 EXPORT_SYMBOL_GPL(edac_mc_handle_ue_no_info);
811 /*************************************************************
812 * On Fully Buffered DIMM modules, this help function is
813 * called to process UE events
815 void edac_mc_handle_fbd_ue(struct mem_ctl_info *mci,
817 unsigned int channela,
818 unsigned int channelb, char *msg)
820 int len = EDAC_MC_LABEL_LEN * 4;
821 char labels[len + 1];
825 if (csrow >= mci->nr_csrows) {
826 /* something is wrong */
827 edac_mc_printk(mci, KERN_ERR,
828 "INTERNAL ERROR: row out of range (%d >= %d)\n",
829 csrow, mci->nr_csrows);
830 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
834 if (channela >= mci->csrows[csrow].nr_channels) {
835 /* something is wrong */
836 edac_mc_printk(mci, KERN_ERR,
837 "INTERNAL ERROR: channel-a out of range "
839 channela, mci->csrows[csrow].nr_channels);
840 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
844 if (channelb >= mci->csrows[csrow].nr_channels) {
845 /* something is wrong */
846 edac_mc_printk(mci, KERN_ERR,
847 "INTERNAL ERROR: channel-b out of range "
849 channelb, mci->csrows[csrow].nr_channels);
850 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
855 mci->csrows[csrow].ue_count++;
857 /* Generate the DIMM labels from the specified channels */
858 chars = snprintf(pos, len + 1, "%s",
859 mci->csrows[csrow].channels[channela].label);
862 chars = snprintf(pos, len + 1, "-%s",
863 mci->csrows[csrow].channels[channelb].label);
865 if (edac_mc_get_log_ue())
866 edac_mc_printk(mci, KERN_EMERG,
867 "UE row %d, channel-a= %d channel-b= %d "
868 "labels \"%s\": %s\n", csrow, channela, channelb,
871 if (edac_mc_get_panic_on_ue())
872 panic("UE row %d, channel-a= %d channel-b= %d "
873 "labels \"%s\": %s\n", csrow, channela,
874 channelb, labels, msg);
876 EXPORT_SYMBOL(edac_mc_handle_fbd_ue);
878 /*************************************************************
879 * On Fully Buffered DIMM modules, this help function is
880 * called to process CE events
882 void edac_mc_handle_fbd_ce(struct mem_ctl_info *mci,
883 unsigned int csrow, unsigned int channel, char *msg)
886 /* Ensure boundary values */
887 if (csrow >= mci->nr_csrows) {
888 /* something is wrong */
889 edac_mc_printk(mci, KERN_ERR,
890 "INTERNAL ERROR: row out of range (%d >= %d)\n",
891 csrow, mci->nr_csrows);
892 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
895 if (channel >= mci->csrows[csrow].nr_channels) {
896 /* something is wrong */
897 edac_mc_printk(mci, KERN_ERR,
898 "INTERNAL ERROR: channel out of range (%d >= %d)\n",
899 channel, mci->csrows[csrow].nr_channels);
900 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
904 if (edac_mc_get_log_ce())
905 /* FIXME - put in DIMM location */
906 edac_mc_printk(mci, KERN_WARNING,
907 "CE row %d, channel %d, label \"%s\": %s\n",
909 mci->csrows[csrow].channels[channel].label, msg);
912 mci->csrows[csrow].ce_count++;
913 mci->csrows[csrow].channels[channel].ce_count++;
915 EXPORT_SYMBOL(edac_mc_handle_fbd_ce);