2 * edac_mc kernel module
3 * (C) 2005, 2006 Linux Networx (http://lnxi.com)
4 * This file may be distributed under the terms of the
5 * GNU General Public License.
7 * Written by Thayne Harbaugh
8 * Based on work by Dan Hollis <goemon at anime dot net> and others.
9 * http://www.anime.net/~goemon/linux-ecc/
11 * Modified by Dave Peterson and Doug Thompson
15 #include <linux/module.h>
16 #include <linux/proc_fs.h>
17 #include <linux/kernel.h>
18 #include <linux/types.h>
19 #include <linux/smp.h>
20 #include <linux/init.h>
21 #include <linux/sysctl.h>
22 #include <linux/highmem.h>
23 #include <linux/timer.h>
24 #include <linux/slab.h>
25 #include <linux/jiffies.h>
26 #include <linux/spinlock.h>
27 #include <linux/list.h>
28 #include <linux/sysdev.h>
29 #include <linux/ctype.h>
30 #include <linux/edac.h>
31 #include <asm/uaccess.h>
34 #include "edac_core.h"
35 #include "edac_module.h"
37 /* lock to memory controller's control array */
38 static DEFINE_MUTEX(mem_ctls_mutex);
39 static LIST_HEAD(mc_devices);
41 #ifdef CONFIG_EDAC_DEBUG
43 static void edac_mc_dump_channel(struct channel_info *chan)
45 debugf4("\tchannel = %p\n", chan);
46 debugf4("\tchannel->chan_idx = %d\n", chan->chan_idx);
47 debugf4("\tchannel->ce_count = %d\n", chan->ce_count);
48 debugf4("\tchannel->label = '%s'\n", chan->label);
49 debugf4("\tchannel->csrow = %p\n\n", chan->csrow);
52 static void edac_mc_dump_csrow(struct csrow_info *csrow)
54 debugf4("\tcsrow = %p\n", csrow);
55 debugf4("\tcsrow->csrow_idx = %d\n", csrow->csrow_idx);
56 debugf4("\tcsrow->first_page = 0x%lx\n", csrow->first_page);
57 debugf4("\tcsrow->last_page = 0x%lx\n", csrow->last_page);
58 debugf4("\tcsrow->page_mask = 0x%lx\n", csrow->page_mask);
59 debugf4("\tcsrow->nr_pages = 0x%x\n", csrow->nr_pages);
60 debugf4("\tcsrow->nr_channels = %d\n", csrow->nr_channels);
61 debugf4("\tcsrow->channels = %p\n", csrow->channels);
62 debugf4("\tcsrow->mci = %p\n\n", csrow->mci);
65 static void edac_mc_dump_mci(struct mem_ctl_info *mci)
67 debugf3("\tmci = %p\n", mci);
68 debugf3("\tmci->mtype_cap = %lx\n", mci->mtype_cap);
69 debugf3("\tmci->edac_ctl_cap = %lx\n", mci->edac_ctl_cap);
70 debugf3("\tmci->edac_cap = %lx\n", mci->edac_cap);
71 debugf4("\tmci->edac_check = %p\n", mci->edac_check);
72 debugf3("\tmci->nr_csrows = %d, csrows = %p\n",
73 mci->nr_csrows, mci->csrows);
74 debugf3("\tdev = %p\n", mci->dev);
75 debugf3("\tmod_name:ctl_name = %s:%s\n", mci->mod_name, mci->ctl_name);
76 debugf3("\tpvt_info = %p\n\n", mci->pvt_info);
80 * keep those in sync with the enum mem_type
82 const char *edac_mem_types[] = {
84 "Reserved csrow type",
87 "Extended data out RAM",
88 "Burst Extended data out RAM",
89 "Single data rate SDRAM",
90 "Registered single data rate SDRAM",
91 "Double data rate SDRAM",
92 "Registered Double data rate SDRAM",
94 "Unbuffered DDR2 RAM",
95 "Fully buffered DDR2",
96 "Registered DDR2 RAM",
98 "Unbuffered DDR3 RAM",
99 "Registered DDR3 RAM",
101 EXPORT_SYMBOL_GPL(edac_mem_types);
103 #endif /* CONFIG_EDAC_DEBUG */
105 /* 'ptr' points to a possibly unaligned item X such that sizeof(X) is 'size'.
106 * Adjust 'ptr' so that its alignment is at least as stringent as what the
107 * compiler would provide for X and return the aligned result.
109 * If 'size' is a constant, the compiler will optimize this whole function
110 * down to either a no-op or the addition of a constant to the value of 'ptr'.
112 void *edac_align_ptr(void *ptr, unsigned size)
116 /* Here we assume that the alignment of a "long long" is the most
117 * stringent alignment that the compiler will ever provide by default.
118 * As far as I know, this is a reasonable assumption.
120 if (size > sizeof(long))
121 align = sizeof(long long);
122 else if (size > sizeof(int))
123 align = sizeof(long);
124 else if (size > sizeof(short))
126 else if (size > sizeof(char))
127 align = sizeof(short);
136 return (void *)(((unsigned long)ptr) + align - r);
140 * edac_mc_alloc: Allocate a struct mem_ctl_info structure
141 * @size_pvt: size of private storage needed
142 * @nr_csrows: Number of CWROWS needed for this MC
143 * @nr_chans: Number of channels for the MC
145 * Everything is kmalloc'ed as one big chunk - more efficient.
146 * Only can be used if all structures have the same lifetime - otherwise
147 * you have to allocate and initialize your own structures.
149 * Use edac_mc_free() to free mc structures allocated by this function.
152 * NULL allocation failed
153 * struct mem_ctl_info pointer
155 struct mem_ctl_info *edac_mc_alloc(unsigned sz_pvt, unsigned nr_csrows,
156 unsigned nr_chans, int edac_index)
158 struct mem_ctl_info *mci;
159 struct csrow_info *csi, *csrow;
160 struct channel_info *chi, *chp, *chan;
166 /* Figure out the offsets of the various items from the start of an mc
167 * structure. We want the alignment of each item to be at least as
168 * stringent as what the compiler would provide if we could simply
169 * hardcode everything into a single struct.
171 mci = (struct mem_ctl_info *)0;
172 csi = edac_align_ptr(&mci[1], sizeof(*csi));
173 chi = edac_align_ptr(&csi[nr_csrows], sizeof(*chi));
174 pvt = edac_align_ptr(&chi[nr_chans * nr_csrows], sz_pvt);
175 size = ((unsigned long)pvt) + sz_pvt;
177 mci = kzalloc(size, GFP_KERNEL);
181 /* Adjust pointers so they point within the memory we just allocated
182 * rather than an imaginary chunk of memory located at address 0.
184 csi = (struct csrow_info *)(((char *)mci) + ((unsigned long)csi));
185 chi = (struct channel_info *)(((char *)mci) + ((unsigned long)chi));
186 pvt = sz_pvt ? (((char *)mci) + ((unsigned long)pvt)) : NULL;
188 /* setup index and various internal pointers */
189 mci->mc_idx = edac_index;
192 mci->nr_csrows = nr_csrows;
194 for (row = 0; row < nr_csrows; row++) {
196 csrow->csrow_idx = row;
198 csrow->nr_channels = nr_chans;
199 chp = &chi[row * nr_chans];
200 csrow->channels = chp;
202 for (chn = 0; chn < nr_chans; chn++) {
204 chan->chan_idx = chn;
209 mci->op_state = OP_ALLOC;
210 INIT_LIST_HEAD(&mci->grp_kobj_list);
213 * Initialize the 'root' kobj for the edac_mc controller
215 err = edac_mc_register_sysfs_main_kobj(mci);
221 /* at this point, the root kobj is valid, and in order to
222 * 'free' the object, then the function:
223 * edac_mc_unregister_sysfs_main_kobj() must be called
224 * which will perform kobj unregistration and the actual free
225 * will occur during the kobject callback operation
229 EXPORT_SYMBOL_GPL(edac_mc_alloc);
233 * 'Free' a previously allocated 'mci' structure
234 * @mci: pointer to a struct mem_ctl_info structure
236 void edac_mc_free(struct mem_ctl_info *mci)
238 debugf1("%s()\n", __func__);
240 edac_mc_unregister_sysfs_main_kobj(mci);
242 EXPORT_SYMBOL_GPL(edac_mc_free);
248 * scan list of controllers looking for the one that manages
250 * @dev: pointer to a struct device related with the MCI
252 struct mem_ctl_info *find_mci_by_dev(struct device *dev)
254 struct mem_ctl_info *mci;
255 struct list_head *item;
257 debugf3("%s()\n", __func__);
259 list_for_each(item, &mc_devices) {
260 mci = list_entry(item, struct mem_ctl_info, link);
268 EXPORT_SYMBOL_GPL(find_mci_by_dev);
271 * handler for EDAC to check if NMI type handler has asserted interrupt
273 static int edac_mc_assert_error_check_and_clear(void)
277 if (edac_op_state == EDAC_OPSTATE_POLL)
280 old_state = edac_err_assert;
287 * edac_mc_workq_function
288 * performs the operation scheduled by a workq request
290 static void edac_mc_workq_function(struct work_struct *work_req)
292 struct delayed_work *d_work = to_delayed_work(work_req);
293 struct mem_ctl_info *mci = to_edac_mem_ctl_work(d_work);
295 mutex_lock(&mem_ctls_mutex);
297 /* if this control struct has movd to offline state, we are done */
298 if (mci->op_state == OP_OFFLINE) {
299 mutex_unlock(&mem_ctls_mutex);
303 /* Only poll controllers that are running polled and have a check */
304 if (edac_mc_assert_error_check_and_clear() && (mci->edac_check != NULL))
305 mci->edac_check(mci);
307 mutex_unlock(&mem_ctls_mutex);
310 queue_delayed_work(edac_workqueue, &mci->work,
311 msecs_to_jiffies(edac_mc_get_poll_msec()));
315 * edac_mc_workq_setup
316 * initialize a workq item for this mci
317 * passing in the new delay period in msec
321 * called with the mem_ctls_mutex held
323 static void edac_mc_workq_setup(struct mem_ctl_info *mci, unsigned msec)
325 debugf0("%s()\n", __func__);
327 /* if this instance is not in the POLL state, then simply return */
328 if (mci->op_state != OP_RUNNING_POLL)
331 INIT_DELAYED_WORK(&mci->work, edac_mc_workq_function);
332 queue_delayed_work(edac_workqueue, &mci->work, msecs_to_jiffies(msec));
336 * edac_mc_workq_teardown
337 * stop the workq processing on this mci
341 * called WITHOUT lock held
343 static void edac_mc_workq_teardown(struct mem_ctl_info *mci)
347 if (mci->op_state != OP_RUNNING_POLL)
350 status = cancel_delayed_work(&mci->work);
352 debugf0("%s() not canceled, flush the queue\n",
355 /* workq instance might be running, wait for it */
356 flush_workqueue(edac_workqueue);
361 * edac_mc_reset_delay_period(unsigned long value)
363 * user space has updated our poll period value, need to
364 * reset our workq delays
366 void edac_mc_reset_delay_period(int value)
368 struct mem_ctl_info *mci;
369 struct list_head *item;
371 mutex_lock(&mem_ctls_mutex);
373 /* scan the list and turn off all workq timers, doing so under lock
375 list_for_each(item, &mc_devices) {
376 mci = list_entry(item, struct mem_ctl_info, link);
378 if (mci->op_state == OP_RUNNING_POLL)
379 cancel_delayed_work(&mci->work);
382 mutex_unlock(&mem_ctls_mutex);
385 /* re-walk the list, and reset the poll delay */
386 mutex_lock(&mem_ctls_mutex);
388 list_for_each(item, &mc_devices) {
389 mci = list_entry(item, struct mem_ctl_info, link);
391 edac_mc_workq_setup(mci, (unsigned long) value);
394 mutex_unlock(&mem_ctls_mutex);
399 /* Return 0 on success, 1 on failure.
400 * Before calling this function, caller must
401 * assign a unique value to mci->mc_idx.
405 * called with the mem_ctls_mutex lock held
407 static int add_mc_to_global_list(struct mem_ctl_info *mci)
409 struct list_head *item, *insert_before;
410 struct mem_ctl_info *p;
412 insert_before = &mc_devices;
414 p = find_mci_by_dev(mci->dev);
415 if (unlikely(p != NULL))
418 list_for_each(item, &mc_devices) {
419 p = list_entry(item, struct mem_ctl_info, link);
421 if (p->mc_idx >= mci->mc_idx) {
422 if (unlikely(p->mc_idx == mci->mc_idx))
425 insert_before = item;
430 list_add_tail_rcu(&mci->link, insert_before);
431 atomic_inc(&edac_handlers);
435 edac_printk(KERN_WARNING, EDAC_MC,
436 "%s (%s) %s %s already assigned %d\n", dev_name(p->dev),
437 edac_dev_name(mci), p->mod_name, p->ctl_name, p->mc_idx);
441 edac_printk(KERN_WARNING, EDAC_MC,
442 "bug in low-level driver: attempt to assign\n"
443 " duplicate mc_idx %d in %s()\n", p->mc_idx, __func__);
447 static void complete_mc_list_del(struct rcu_head *head)
449 struct mem_ctl_info *mci;
451 mci = container_of(head, struct mem_ctl_info, rcu);
452 INIT_LIST_HEAD(&mci->link);
455 static void del_mc_from_global_list(struct mem_ctl_info *mci)
457 atomic_dec(&edac_handlers);
458 list_del_rcu(&mci->link);
459 call_rcu(&mci->rcu, complete_mc_list_del);
464 * edac_mc_find: Search for a mem_ctl_info structure whose index is 'idx'.
466 * If found, return a pointer to the structure.
469 * Caller must hold mem_ctls_mutex.
471 struct mem_ctl_info *edac_mc_find(int idx)
473 struct list_head *item;
474 struct mem_ctl_info *mci;
476 list_for_each(item, &mc_devices) {
477 mci = list_entry(item, struct mem_ctl_info, link);
479 if (mci->mc_idx >= idx) {
480 if (mci->mc_idx == idx)
489 EXPORT_SYMBOL(edac_mc_find);
492 * edac_mc_add_mc: Insert the 'mci' structure into the mci global list and
493 * create sysfs entries associated with mci structure
494 * @mci: pointer to the mci structure to be added to the list
495 * @mc_idx: A unique numeric identifier to be assigned to the 'mci' structure.
502 /* FIXME - should a warning be printed if no error detection? correction? */
503 int edac_mc_add_mc(struct mem_ctl_info *mci)
505 debugf0("%s()\n", __func__);
507 #ifdef CONFIG_EDAC_DEBUG
508 if (edac_debug_level >= 3)
509 edac_mc_dump_mci(mci);
511 if (edac_debug_level >= 4) {
514 for (i = 0; i < mci->nr_csrows; i++) {
517 edac_mc_dump_csrow(&mci->csrows[i]);
518 for (j = 0; j < mci->csrows[i].nr_channels; j++)
519 edac_mc_dump_channel(&mci->csrows[i].
524 mutex_lock(&mem_ctls_mutex);
526 if (add_mc_to_global_list(mci))
529 /* set load time so that error rate can be tracked */
530 mci->start_time = jiffies;
532 if (edac_create_sysfs_mci_device(mci)) {
533 edac_mc_printk(mci, KERN_WARNING,
534 "failed to create sysfs device\n");
538 /* If there IS a check routine, then we are running POLLED */
539 if (mci->edac_check != NULL) {
540 /* This instance is NOW RUNNING */
541 mci->op_state = OP_RUNNING_POLL;
543 edac_mc_workq_setup(mci, edac_mc_get_poll_msec());
545 mci->op_state = OP_RUNNING_INTERRUPT;
548 /* Report action taken */
549 edac_mc_printk(mci, KERN_INFO, "Giving out device to '%s' '%s':"
550 " DEV %s\n", mci->mod_name, mci->ctl_name, edac_dev_name(mci));
552 mutex_unlock(&mem_ctls_mutex);
556 del_mc_from_global_list(mci);
559 mutex_unlock(&mem_ctls_mutex);
562 EXPORT_SYMBOL_GPL(edac_mc_add_mc);
565 * edac_mc_del_mc: Remove sysfs entries for specified mci structure and
566 * remove mci structure from global list
567 * @pdev: Pointer to 'struct device' representing mci structure to remove.
569 * Return pointer to removed mci structure, or NULL if device not found.
571 struct mem_ctl_info *edac_mc_del_mc(struct device *dev)
573 struct mem_ctl_info *mci;
575 debugf0("%s()\n", __func__);
577 mutex_lock(&mem_ctls_mutex);
579 /* find the requested mci struct in the global list */
580 mci = find_mci_by_dev(dev);
582 mutex_unlock(&mem_ctls_mutex);
586 /* marking MCI offline */
587 mci->op_state = OP_OFFLINE;
589 del_mc_from_global_list(mci);
590 mutex_unlock(&mem_ctls_mutex);
592 /* flush workq processes and remove sysfs */
593 edac_mc_workq_teardown(mci);
594 edac_remove_sysfs_mci_device(mci);
596 edac_printk(KERN_INFO, EDAC_MC,
597 "Removed device %d for %s %s: DEV %s\n", mci->mc_idx,
598 mci->mod_name, mci->ctl_name, edac_dev_name(mci));
602 EXPORT_SYMBOL_GPL(edac_mc_del_mc);
604 static void edac_mc_scrub_block(unsigned long page, unsigned long offset,
609 unsigned long flags = 0;
611 debugf3("%s()\n", __func__);
613 /* ECC error page was not in our memory. Ignore it. */
614 if (!pfn_valid(page))
617 /* Find the actual page structure then map it and fix */
618 pg = pfn_to_page(page);
621 local_irq_save(flags);
623 virt_addr = kmap_atomic(pg, KM_BOUNCE_READ);
625 /* Perform architecture specific atomic scrub operation */
626 atomic_scrub(virt_addr + offset, size);
628 /* Unmap and complete */
629 kunmap_atomic(virt_addr, KM_BOUNCE_READ);
632 local_irq_restore(flags);
635 /* FIXME - should return -1 */
636 int edac_mc_find_csrow_by_page(struct mem_ctl_info *mci, unsigned long page)
638 struct csrow_info *csrows = mci->csrows;
641 debugf1("MC%d: %s(): 0x%lx\n", mci->mc_idx, __func__, page);
644 for (i = 0; i < mci->nr_csrows; i++) {
645 struct csrow_info *csrow = &csrows[i];
647 if (csrow->nr_pages == 0)
650 debugf3("MC%d: %s(): first(0x%lx) page(0x%lx) last(0x%lx) "
651 "mask(0x%lx)\n", mci->mc_idx, __func__,
652 csrow->first_page, page, csrow->last_page,
655 if ((page >= csrow->first_page) &&
656 (page <= csrow->last_page) &&
657 ((page & csrow->page_mask) ==
658 (csrow->first_page & csrow->page_mask))) {
665 edac_mc_printk(mci, KERN_ERR,
666 "could not look up page error address %lx\n",
667 (unsigned long)page);
671 EXPORT_SYMBOL_GPL(edac_mc_find_csrow_by_page);
673 /* FIXME - setable log (warning/emerg) levels */
674 /* FIXME - integrate with evlog: http://evlog.sourceforge.net/ */
675 void edac_mc_handle_ce(struct mem_ctl_info *mci,
676 unsigned long page_frame_number,
677 unsigned long offset_in_page, unsigned long syndrome,
678 int row, int channel, const char *msg)
680 unsigned long remapped_page;
682 debugf3("MC%d: %s()\n", mci->mc_idx, __func__);
684 /* FIXME - maybe make panic on INTERNAL ERROR an option */
685 if (row >= mci->nr_csrows || row < 0) {
686 /* something is wrong */
687 edac_mc_printk(mci, KERN_ERR,
688 "INTERNAL ERROR: row out of range "
689 "(%d >= %d)\n", row, mci->nr_csrows);
690 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
694 if (channel >= mci->csrows[row].nr_channels || channel < 0) {
695 /* something is wrong */
696 edac_mc_printk(mci, KERN_ERR,
697 "INTERNAL ERROR: channel out of range "
698 "(%d >= %d)\n", channel,
699 mci->csrows[row].nr_channels);
700 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
704 if (edac_mc_get_log_ce())
705 /* FIXME - put in DIMM location */
706 edac_mc_printk(mci, KERN_WARNING,
707 "CE page 0x%lx, offset 0x%lx, grain %d, syndrome "
708 "0x%lx, row %d, channel %d, label \"%s\": %s\n",
709 page_frame_number, offset_in_page,
710 mci->csrows[row].grain, syndrome, row, channel,
711 mci->csrows[row].channels[channel].label, msg);
714 mci->csrows[row].ce_count++;
715 mci->csrows[row].channels[channel].ce_count++;
717 if (mci->scrub_mode & SCRUB_SW_SRC) {
719 * Some MC's can remap memory so that it is still available
720 * at a different address when PCI devices map into memory.
721 * MC's that can't do this lose the memory where PCI devices
722 * are mapped. This mapping is MC dependant and so we call
723 * back into the MC driver for it to map the MC page to
724 * a physical (CPU) page which can then be mapped to a virtual
725 * page - which can then be scrubbed.
727 remapped_page = mci->ctl_page_to_phys ?
728 mci->ctl_page_to_phys(mci, page_frame_number) :
731 edac_mc_scrub_block(remapped_page, offset_in_page,
732 mci->csrows[row].grain);
735 EXPORT_SYMBOL_GPL(edac_mc_handle_ce);
737 void edac_mc_handle_ce_no_info(struct mem_ctl_info *mci, const char *msg)
739 if (edac_mc_get_log_ce())
740 edac_mc_printk(mci, KERN_WARNING,
741 "CE - no information available: %s\n", msg);
743 mci->ce_noinfo_count++;
746 EXPORT_SYMBOL_GPL(edac_mc_handle_ce_no_info);
748 void edac_mc_handle_ue(struct mem_ctl_info *mci,
749 unsigned long page_frame_number,
750 unsigned long offset_in_page, int row, const char *msg)
752 int len = EDAC_MC_LABEL_LEN * 4;
753 char labels[len + 1];
758 debugf3("MC%d: %s()\n", mci->mc_idx, __func__);
760 /* FIXME - maybe make panic on INTERNAL ERROR an option */
761 if (row >= mci->nr_csrows || row < 0) {
762 /* something is wrong */
763 edac_mc_printk(mci, KERN_ERR,
764 "INTERNAL ERROR: row out of range "
765 "(%d >= %d)\n", row, mci->nr_csrows);
766 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
770 chars = snprintf(pos, len + 1, "%s",
771 mci->csrows[row].channels[0].label);
775 for (chan = 1; (chan < mci->csrows[row].nr_channels) && (len > 0);
777 chars = snprintf(pos, len + 1, ":%s",
778 mci->csrows[row].channels[chan].label);
783 if (edac_mc_get_log_ue())
784 edac_mc_printk(mci, KERN_EMERG,
785 "UE page 0x%lx, offset 0x%lx, grain %d, row %d, "
786 "labels \"%s\": %s\n", page_frame_number,
787 offset_in_page, mci->csrows[row].grain, row,
790 if (edac_mc_get_panic_on_ue())
791 panic("EDAC MC%d: UE page 0x%lx, offset 0x%lx, grain %d, "
792 "row %d, labels \"%s\": %s\n", mci->mc_idx,
793 page_frame_number, offset_in_page,
794 mci->csrows[row].grain, row, labels, msg);
797 mci->csrows[row].ue_count++;
799 EXPORT_SYMBOL_GPL(edac_mc_handle_ue);
801 void edac_mc_handle_ue_no_info(struct mem_ctl_info *mci, const char *msg)
803 if (edac_mc_get_panic_on_ue())
804 panic("EDAC MC%d: Uncorrected Error", mci->mc_idx);
806 if (edac_mc_get_log_ue())
807 edac_mc_printk(mci, KERN_WARNING,
808 "UE - no information available: %s\n", msg);
809 mci->ue_noinfo_count++;
812 EXPORT_SYMBOL_GPL(edac_mc_handle_ue_no_info);
814 /*************************************************************
815 * On Fully Buffered DIMM modules, this help function is
816 * called to process UE events
818 void edac_mc_handle_fbd_ue(struct mem_ctl_info *mci,
820 unsigned int channela,
821 unsigned int channelb, char *msg)
823 int len = EDAC_MC_LABEL_LEN * 4;
824 char labels[len + 1];
828 if (csrow >= mci->nr_csrows) {
829 /* something is wrong */
830 edac_mc_printk(mci, KERN_ERR,
831 "INTERNAL ERROR: row out of range (%d >= %d)\n",
832 csrow, mci->nr_csrows);
833 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
837 if (channela >= mci->csrows[csrow].nr_channels) {
838 /* something is wrong */
839 edac_mc_printk(mci, KERN_ERR,
840 "INTERNAL ERROR: channel-a out of range "
842 channela, mci->csrows[csrow].nr_channels);
843 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
847 if (channelb >= mci->csrows[csrow].nr_channels) {
848 /* something is wrong */
849 edac_mc_printk(mci, KERN_ERR,
850 "INTERNAL ERROR: channel-b out of range "
852 channelb, mci->csrows[csrow].nr_channels);
853 edac_mc_handle_ue_no_info(mci, "INTERNAL ERROR");
858 mci->csrows[csrow].ue_count++;
860 /* Generate the DIMM labels from the specified channels */
861 chars = snprintf(pos, len + 1, "%s",
862 mci->csrows[csrow].channels[channela].label);
865 chars = snprintf(pos, len + 1, "-%s",
866 mci->csrows[csrow].channels[channelb].label);
868 if (edac_mc_get_log_ue())
869 edac_mc_printk(mci, KERN_EMERG,
870 "UE row %d, channel-a= %d channel-b= %d "
871 "labels \"%s\": %s\n", csrow, channela, channelb,
874 if (edac_mc_get_panic_on_ue())
875 panic("UE row %d, channel-a= %d channel-b= %d "
876 "labels \"%s\": %s\n", csrow, channela,
877 channelb, labels, msg);
879 EXPORT_SYMBOL(edac_mc_handle_fbd_ue);
881 /*************************************************************
882 * On Fully Buffered DIMM modules, this help function is
883 * called to process CE events
885 void edac_mc_handle_fbd_ce(struct mem_ctl_info *mci,
886 unsigned int csrow, unsigned int channel, char *msg)
889 /* Ensure boundary values */
890 if (csrow >= mci->nr_csrows) {
891 /* something is wrong */
892 edac_mc_printk(mci, KERN_ERR,
893 "INTERNAL ERROR: row out of range (%d >= %d)\n",
894 csrow, mci->nr_csrows);
895 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
898 if (channel >= mci->csrows[csrow].nr_channels) {
899 /* something is wrong */
900 edac_mc_printk(mci, KERN_ERR,
901 "INTERNAL ERROR: channel out of range (%d >= %d)\n",
902 channel, mci->csrows[csrow].nr_channels);
903 edac_mc_handle_ce_no_info(mci, "INTERNAL ERROR");
907 if (edac_mc_get_log_ce())
908 /* FIXME - put in DIMM location */
909 edac_mc_printk(mci, KERN_WARNING,
910 "CE row %d, channel %d, label \"%s\": %s\n",
912 mci->csrows[csrow].channels[channel].label, msg);
915 mci->csrows[csrow].ce_count++;
916 mci->csrows[csrow].channels[channel].ce_count++;
918 EXPORT_SYMBOL(edac_mc_handle_fbd_ce);