cxgb4: Add debugfs support to dump meminfo
authorHariprasad Shenai <hariprasad@chelsio.com>
Tue, 4 Aug 2015 09:06:17 +0000 (14:36 +0530)
committerDavid S. Miller <davem@davemloft.net>
Tue, 4 Aug 2015 08:24:34 +0000 (01:24 -0700)
Add debug support to dump memory address ranges of various hardware
modules of the adapter.

Signed-off-by: Hariprasad Shenai <hariprasad@chelsio.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c
drivers/net/ethernet/chelsio/cxgb4/t4_regs.h

index f701a6f..b657734 100644 (file)
@@ -2275,6 +2275,290 @@ static const struct file_operations blocked_fl_fops = {
        .llseek  = generic_file_llseek,
 };
 
+struct mem_desc {
+       unsigned int base;
+       unsigned int limit;
+       unsigned int idx;
+};
+
+static int mem_desc_cmp(const void *a, const void *b)
+{
+       return ((const struct mem_desc *)a)->base -
+              ((const struct mem_desc *)b)->base;
+}
+
+static void mem_region_show(struct seq_file *seq, const char *name,
+                           unsigned int from, unsigned int to)
+{
+       char buf[40];
+
+       string_get_size((u64)to - from + 1, 1, STRING_UNITS_2, buf,
+                       sizeof(buf));
+       seq_printf(seq, "%-15s %#x-%#x [%s]\n", name, from, to, buf);
+}
+
+static int meminfo_show(struct seq_file *seq, void *v)
+{
+       static const char * const memory[] = { "EDC0:", "EDC1:", "MC:",
+                                       "MC0:", "MC1:"};
+       static const char * const region[] = {
+               "DBQ contexts:", "IMSG contexts:", "FLM cache:", "TCBs:",
+               "Pstructs:", "Timers:", "Rx FL:", "Tx FL:", "Pstruct FL:",
+               "Tx payload:", "Rx payload:", "LE hash:", "iSCSI region:",
+               "TDDP region:", "TPT region:", "STAG region:", "RQ region:",
+               "RQUDP region:", "PBL region:", "TXPBL region:",
+               "DBVFIFO region:", "ULPRX state:", "ULPTX state:",
+               "On-chip queues:"
+       };
+
+       int i, n;
+       u32 lo, hi, used, alloc;
+       struct mem_desc avail[4];
+       struct mem_desc mem[ARRAY_SIZE(region) + 3];      /* up to 3 holes */
+       struct mem_desc *md = mem;
+       struct adapter *adap = seq->private;
+
+       for (i = 0; i < ARRAY_SIZE(mem); i++) {
+               mem[i].limit = 0;
+               mem[i].idx = i;
+       }
+
+       /* Find and sort the populated memory ranges */
+       i = 0;
+       lo = t4_read_reg(adap, MA_TARGET_MEM_ENABLE_A);
+       if (lo & EDRAM0_ENABLE_F) {
+               hi = t4_read_reg(adap, MA_EDRAM0_BAR_A);
+               avail[i].base = EDRAM0_BASE_G(hi) << 20;
+               avail[i].limit = avail[i].base + (EDRAM0_SIZE_G(hi) << 20);
+               avail[i].idx = 0;
+               i++;
+       }
+       if (lo & EDRAM1_ENABLE_F) {
+               hi = t4_read_reg(adap, MA_EDRAM1_BAR_A);
+               avail[i].base = EDRAM1_BASE_G(hi) << 20;
+               avail[i].limit = avail[i].base + (EDRAM1_SIZE_G(hi) << 20);
+               avail[i].idx = 1;
+               i++;
+       }
+
+       if (is_t5(adap->params.chip)) {
+               if (lo & EXT_MEM0_ENABLE_F) {
+                       hi = t4_read_reg(adap, MA_EXT_MEMORY0_BAR_A);
+                       avail[i].base = EXT_MEM0_BASE_G(hi) << 20;
+                       avail[i].limit =
+                               avail[i].base + (EXT_MEM0_SIZE_G(hi) << 20);
+                       avail[i].idx = 3;
+                       i++;
+               }
+               if (lo & EXT_MEM1_ENABLE_F) {
+                       hi = t4_read_reg(adap, MA_EXT_MEMORY1_BAR_A);
+                       avail[i].base = EXT_MEM1_BASE_G(hi) << 20;
+                       avail[i].limit =
+                               avail[i].base + (EXT_MEM1_SIZE_G(hi) << 20);
+                       avail[i].idx = 4;
+                       i++;
+               }
+       } else {
+               if (lo & EXT_MEM_ENABLE_F) {
+                       hi = t4_read_reg(adap, MA_EXT_MEMORY_BAR_A);
+                       avail[i].base = EXT_MEM_BASE_G(hi) << 20;
+                       avail[i].limit =
+                               avail[i].base + (EXT_MEM_SIZE_G(hi) << 20);
+                       avail[i].idx = 2;
+                       i++;
+               }
+       }
+       if (!i)                                    /* no memory available */
+               return 0;
+       sort(avail, i, sizeof(struct mem_desc), mem_desc_cmp, NULL);
+
+       (md++)->base = t4_read_reg(adap, SGE_DBQ_CTXT_BADDR_A);
+       (md++)->base = t4_read_reg(adap, SGE_IMSG_CTXT_BADDR_A);
+       (md++)->base = t4_read_reg(adap, SGE_FLM_CACHE_BADDR_A);
+       (md++)->base = t4_read_reg(adap, TP_CMM_TCB_BASE_A);
+       (md++)->base = t4_read_reg(adap, TP_CMM_MM_BASE_A);
+       (md++)->base = t4_read_reg(adap, TP_CMM_TIMER_BASE_A);
+       (md++)->base = t4_read_reg(adap, TP_CMM_MM_RX_FLST_BASE_A);
+       (md++)->base = t4_read_reg(adap, TP_CMM_MM_TX_FLST_BASE_A);
+       (md++)->base = t4_read_reg(adap, TP_CMM_MM_PS_FLST_BASE_A);
+
+       /* the next few have explicit upper bounds */
+       md->base = t4_read_reg(adap, TP_PMM_TX_BASE_A);
+       md->limit = md->base - 1 +
+                   t4_read_reg(adap, TP_PMM_TX_PAGE_SIZE_A) *
+                   PMTXMAXPAGE_G(t4_read_reg(adap, TP_PMM_TX_MAX_PAGE_A));
+       md++;
+
+       md->base = t4_read_reg(adap, TP_PMM_RX_BASE_A);
+       md->limit = md->base - 1 +
+                   t4_read_reg(adap, TP_PMM_RX_PAGE_SIZE_A) *
+                   PMRXMAXPAGE_G(t4_read_reg(adap, TP_PMM_RX_MAX_PAGE_A));
+       md++;
+
+       if (t4_read_reg(adap, LE_DB_CONFIG_A) & HASHEN_F) {
+               if (CHELSIO_CHIP_VERSION(adap->params.chip) <= CHELSIO_T5) {
+                       hi = t4_read_reg(adap, LE_DB_TID_HASHBASE_A) / 4;
+                       md->base = t4_read_reg(adap, LE_DB_HASH_TID_BASE_A);
+                } else {
+                       hi = t4_read_reg(adap, LE_DB_HASH_TID_BASE_A);
+                       md->base = t4_read_reg(adap,
+                                              LE_DB_HASH_TBL_BASE_ADDR_A);
+               }
+               md->limit = 0;
+       } else {
+               md->base = 0;
+               md->idx = ARRAY_SIZE(region);  /* hide it */
+       }
+       md++;
+
+#define ulp_region(reg) do { \
+       md->base = t4_read_reg(adap, ULP_ ## reg ## _LLIMIT_A);\
+       (md++)->limit = t4_read_reg(adap, ULP_ ## reg ## _ULIMIT_A); \
+} while (0)
+
+       ulp_region(RX_ISCSI);
+       ulp_region(RX_TDDP);
+       ulp_region(TX_TPT);
+       ulp_region(RX_STAG);
+       ulp_region(RX_RQ);
+       ulp_region(RX_RQUDP);
+       ulp_region(RX_PBL);
+       ulp_region(TX_PBL);
+#undef ulp_region
+       md->base = 0;
+       md->idx = ARRAY_SIZE(region);
+       if (!is_t4(adap->params.chip)) {
+               u32 size = 0;
+               u32 sge_ctrl = t4_read_reg(adap, SGE_CONTROL2_A);
+               u32 fifo_size = t4_read_reg(adap, SGE_DBVFIFO_SIZE_A);
+
+               if (is_t5(adap->params.chip)) {
+                       if (sge_ctrl & VFIFO_ENABLE_F)
+                               size = DBVFIFO_SIZE_G(fifo_size);
+               } else {
+                       size = T6_DBVFIFO_SIZE_G(fifo_size);
+               }
+
+               if (size) {
+                       md->base = BASEADDR_G(t4_read_reg(adap,
+                                       SGE_DBVFIFO_BADDR_A));
+                       md->limit = md->base + (size << 2) - 1;
+               }
+       }
+
+       md++;
+
+       md->base = t4_read_reg(adap, ULP_RX_CTX_BASE_A);
+       md->limit = 0;
+       md++;
+       md->base = t4_read_reg(adap, ULP_TX_ERR_TABLE_BASE_A);
+       md->limit = 0;
+       md++;
+
+       md->base = adap->vres.ocq.start;
+       if (adap->vres.ocq.size)
+               md->limit = md->base + adap->vres.ocq.size - 1;
+       else
+               md->idx = ARRAY_SIZE(region);  /* hide it */
+       md++;
+
+       /* add any address-space holes, there can be up to 3 */
+       for (n = 0; n < i - 1; n++)
+               if (avail[n].limit < avail[n + 1].base)
+                       (md++)->base = avail[n].limit;
+       if (avail[n].limit)
+               (md++)->base = avail[n].limit;
+
+       n = md - mem;
+       sort(mem, n, sizeof(struct mem_desc), mem_desc_cmp, NULL);
+
+       for (lo = 0; lo < i; lo++)
+               mem_region_show(seq, memory[avail[lo].idx], avail[lo].base,
+                               avail[lo].limit - 1);
+
+       seq_putc(seq, '\n');
+       for (i = 0; i < n; i++) {
+               if (mem[i].idx >= ARRAY_SIZE(region))
+                       continue;                        /* skip holes */
+               if (!mem[i].limit)
+                       mem[i].limit = i < n - 1 ? mem[i + 1].base - 1 : ~0;
+               mem_region_show(seq, region[mem[i].idx], mem[i].base,
+                               mem[i].limit);
+       }
+
+       seq_putc(seq, '\n');
+       lo = t4_read_reg(adap, CIM_SDRAM_BASE_ADDR_A);
+       hi = t4_read_reg(adap, CIM_SDRAM_ADDR_SIZE_A) + lo - 1;
+       mem_region_show(seq, "uP RAM:", lo, hi);
+
+       lo = t4_read_reg(adap, CIM_EXTMEM2_BASE_ADDR_A);
+       hi = t4_read_reg(adap, CIM_EXTMEM2_ADDR_SIZE_A) + lo - 1;
+       mem_region_show(seq, "uP Extmem2:", lo, hi);
+
+       lo = t4_read_reg(adap, TP_PMM_RX_MAX_PAGE_A);
+       seq_printf(seq, "\n%u Rx pages of size %uKiB for %u channels\n",
+                  PMRXMAXPAGE_G(lo),
+                  t4_read_reg(adap, TP_PMM_RX_PAGE_SIZE_A) >> 10,
+                  (lo & PMRXNUMCHN_F) ? 2 : 1);
+
+       lo = t4_read_reg(adap, TP_PMM_TX_MAX_PAGE_A);
+       hi = t4_read_reg(adap, TP_PMM_TX_PAGE_SIZE_A);
+       seq_printf(seq, "%u Tx pages of size %u%ciB for %u channels\n",
+                  PMTXMAXPAGE_G(lo),
+                  hi >= (1 << 20) ? (hi >> 20) : (hi >> 10),
+                  hi >= (1 << 20) ? 'M' : 'K', 1 << PMTXNUMCHN_G(lo));
+       seq_printf(seq, "%u p-structs\n\n",
+                  t4_read_reg(adap, TP_CMM_MM_MAX_PSTRUCT_A));
+
+       for (i = 0; i < 4; i++) {
+               if (CHELSIO_CHIP_VERSION(adap->params.chip) > CHELSIO_T5)
+                       lo = t4_read_reg(adap, MPS_RX_MAC_BG_PG_CNT0_A + i * 4);
+               else
+                       lo = t4_read_reg(adap, MPS_RX_PG_RSV0_A + i * 4);
+               if (is_t5(adap->params.chip)) {
+                       used = T5_USED_G(lo);
+                       alloc = T5_ALLOC_G(lo);
+               } else {
+                       used = USED_G(lo);
+                       alloc = ALLOC_G(lo);
+               }
+               /* For T6 these are MAC buffer groups */
+               seq_printf(seq, "Port %d using %u pages out of %u allocated\n",
+                          i, used, alloc);
+       }
+       for (i = 0; i < adap->params.arch.nchan; i++) {
+               if (CHELSIO_CHIP_VERSION(adap->params.chip) > CHELSIO_T5)
+                       lo = t4_read_reg(adap,
+                                        MPS_RX_LPBK_BG_PG_CNT0_A + i * 4);
+               else
+                       lo = t4_read_reg(adap, MPS_RX_PG_RSV4_A + i * 4);
+               if (is_t5(adap->params.chip)) {
+                       used = T5_USED_G(lo);
+                       alloc = T5_ALLOC_G(lo);
+               } else {
+                       used = USED_G(lo);
+                       alloc = ALLOC_G(lo);
+               }
+               /* For T6 these are MAC buffer groups */
+               seq_printf(seq,
+                          "Loopback %d using %u pages out of %u allocated\n",
+                          i, used, alloc);
+       }
+       return 0;
+}
+
+static int meminfo_open(struct inode *inode, struct file *file)
+{
+       return single_open(file, meminfo_show, inode->i_private);
+}
+
+static const struct file_operations meminfo_fops = {
+       .owner   = THIS_MODULE,
+       .open    = meminfo_open,
+       .read    = seq_read,
+       .llseek  = seq_lseek,
+       .release = single_release,
+};
 /* Add an array of Debug FS files.
  */
 void add_debugfs_files(struct adapter *adap,
@@ -2342,6 +2626,7 @@ int t4_setup_debugfs(struct adapter *adap)
                { "clip_tbl", &clip_tbl_debugfs_fops, S_IRUSR, 0 },
 #endif
                { "blocked_fl", &blocked_fl_fops, S_IRUSR | S_IWUSR, 0 },
+               { "meminfo", &meminfo_fops, S_IRUSR, 0 },
        };
 
        /* Debug FS nodes common to all T5 and later adapters.