281 lines
9.0 KiB
C
281 lines
9.0 KiB
C
/*
|
|
* Copyright (c) 2006 QLogic, Inc. All rights reserved.
|
|
* Copyright (c) 2003, 2004, 2005, 2006 PathScale, Inc. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* OpenIB.org BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*/
|
|
|
|
#include "ipath_kernel.h"
|
|
|
|
struct infinipath_stats ipath_stats;
|
|
|
|
/**
|
|
* ipath_snap_cntr - snapshot a chip counter
|
|
* @dd: the infinipath device
|
|
* @creg: the counter to snapshot
|
|
*
|
|
* called from add_timer and user counter read calls, to deal with
|
|
* counters that wrap in "human time". The words sent and received, and
|
|
* the packets sent and received are all that we worry about. For now,
|
|
* at least, we don't worry about error counters, because if they wrap
|
|
* that quickly, we probably don't care. We may eventually just make this
|
|
* handle all the counters. word counters can wrap in about 20 seconds
|
|
* of full bandwidth traffic, packet counters in a few hours.
|
|
*/
|
|
|
|
u64 ipath_snap_cntr(struct ipath_devdata *dd, ipath_creg creg)
|
|
{
|
|
u32 val, reg64 = 0;
|
|
u64 val64;
|
|
unsigned long t0, t1;
|
|
u64 ret;
|
|
|
|
t0 = jiffies;
|
|
/* If fast increment counters are only 32 bits, snapshot them,
|
|
* and maintain them as 64bit values in the driver */
|
|
if (!(dd->ipath_flags & IPATH_32BITCOUNTERS) &&
|
|
(creg == dd->ipath_cregs->cr_wordsendcnt ||
|
|
creg == dd->ipath_cregs->cr_wordrcvcnt ||
|
|
creg == dd->ipath_cregs->cr_pktsendcnt ||
|
|
creg == dd->ipath_cregs->cr_pktrcvcnt)) {
|
|
val64 = ipath_read_creg(dd, creg);
|
|
val = val64 == ~0ULL ? ~0U : 0;
|
|
reg64 = 1;
|
|
} else /* val64 just to keep gcc quiet... */
|
|
val64 = val = ipath_read_creg32(dd, creg);
|
|
/*
|
|
* See if a second has passed. This is just a way to detect things
|
|
* that are quite broken. Normally this should take just a few
|
|
* cycles (the check is for long enough that we don't care if we get
|
|
* pre-empted.) An Opteron HT O read timeout is 4 seconds with
|
|
* normal NB values
|
|
*/
|
|
t1 = jiffies;
|
|
if (time_before(t0 + HZ, t1) && val == -1) {
|
|
ipath_dev_err(dd, "Error! Read counter 0x%x timed out\n",
|
|
creg);
|
|
ret = 0ULL;
|
|
goto bail;
|
|
}
|
|
if (reg64) {
|
|
ret = val64;
|
|
goto bail;
|
|
}
|
|
|
|
if (creg == dd->ipath_cregs->cr_wordsendcnt) {
|
|
if (val != dd->ipath_lastsword) {
|
|
dd->ipath_sword += val - dd->ipath_lastsword;
|
|
dd->ipath_lastsword = val;
|
|
}
|
|
val64 = dd->ipath_sword;
|
|
} else if (creg == dd->ipath_cregs->cr_wordrcvcnt) {
|
|
if (val != dd->ipath_lastrword) {
|
|
dd->ipath_rword += val - dd->ipath_lastrword;
|
|
dd->ipath_lastrword = val;
|
|
}
|
|
val64 = dd->ipath_rword;
|
|
} else if (creg == dd->ipath_cregs->cr_pktsendcnt) {
|
|
if (val != dd->ipath_lastspkts) {
|
|
dd->ipath_spkts += val - dd->ipath_lastspkts;
|
|
dd->ipath_lastspkts = val;
|
|
}
|
|
val64 = dd->ipath_spkts;
|
|
} else if (creg == dd->ipath_cregs->cr_pktrcvcnt) {
|
|
if (val != dd->ipath_lastrpkts) {
|
|
dd->ipath_rpkts += val - dd->ipath_lastrpkts;
|
|
dd->ipath_lastrpkts = val;
|
|
}
|
|
val64 = dd->ipath_rpkts;
|
|
} else
|
|
val64 = (u64) val;
|
|
|
|
ret = val64;
|
|
|
|
bail:
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* ipath_qcheck - print delta of egrfull/hdrqfull errors for kernel ports
|
|
* @dd: the infinipath device
|
|
*
|
|
* print the delta of egrfull/hdrqfull errors for kernel ports no more than
|
|
* every 5 seconds. User processes are printed at close, but kernel doesn't
|
|
* close, so... Separate routine so may call from other places someday, and
|
|
* so function name when printed by _IPATH_INFO is meaningfull
|
|
*/
|
|
static void ipath_qcheck(struct ipath_devdata *dd)
|
|
{
|
|
static u64 last_tot_hdrqfull;
|
|
size_t blen = 0;
|
|
char buf[128];
|
|
|
|
*buf = 0;
|
|
if (dd->ipath_pd[0]->port_hdrqfull != dd->ipath_p0_hdrqfull) {
|
|
blen = snprintf(buf, sizeof buf, "port 0 hdrqfull %u",
|
|
dd->ipath_pd[0]->port_hdrqfull -
|
|
dd->ipath_p0_hdrqfull);
|
|
dd->ipath_p0_hdrqfull = dd->ipath_pd[0]->port_hdrqfull;
|
|
}
|
|
if (ipath_stats.sps_etidfull != dd->ipath_last_tidfull) {
|
|
blen += snprintf(buf + blen, sizeof buf - blen,
|
|
"%srcvegrfull %llu",
|
|
blen ? ", " : "",
|
|
(unsigned long long)
|
|
(ipath_stats.sps_etidfull -
|
|
dd->ipath_last_tidfull));
|
|
dd->ipath_last_tidfull = ipath_stats.sps_etidfull;
|
|
}
|
|
|
|
/*
|
|
* this is actually the number of hdrq full interrupts, not actual
|
|
* events, but at the moment that's mostly what I'm interested in.
|
|
* Actual count, etc. is in the counters, if needed. For production
|
|
* users this won't ordinarily be printed.
|
|
*/
|
|
|
|
if ((ipath_debug & (__IPATH_PKTDBG | __IPATH_DBG)) &&
|
|
ipath_stats.sps_hdrqfull != last_tot_hdrqfull) {
|
|
blen += snprintf(buf + blen, sizeof buf - blen,
|
|
"%shdrqfull %llu (all ports)",
|
|
blen ? ", " : "",
|
|
(unsigned long long)
|
|
(ipath_stats.sps_hdrqfull -
|
|
last_tot_hdrqfull));
|
|
last_tot_hdrqfull = ipath_stats.sps_hdrqfull;
|
|
}
|
|
if (blen)
|
|
ipath_dbg("%s\n", buf);
|
|
|
|
if (dd->ipath_port0head != (u32)
|
|
le64_to_cpu(*dd->ipath_hdrqtailptr)) {
|
|
if (dd->ipath_lastport0rcv_cnt ==
|
|
ipath_stats.sps_port0pkts) {
|
|
ipath_cdbg(PKT, "missing rcv interrupts? "
|
|
"port0 hd=%llx tl=%x; port0pkts %llx\n",
|
|
(unsigned long long)
|
|
le64_to_cpu(*dd->ipath_hdrqtailptr),
|
|
dd->ipath_port0head,
|
|
(unsigned long long)
|
|
ipath_stats.sps_port0pkts);
|
|
}
|
|
dd->ipath_lastport0rcv_cnt = ipath_stats.sps_port0pkts;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ipath_get_faststats - get word counters from chip before they overflow
|
|
* @opaque - contains a pointer to the infinipath device ipath_devdata
|
|
*
|
|
* called from add_timer
|
|
*/
|
|
void ipath_get_faststats(unsigned long opaque)
|
|
{
|
|
struct ipath_devdata *dd = (struct ipath_devdata *) opaque;
|
|
u32 val;
|
|
static unsigned cnt;
|
|
|
|
/*
|
|
* don't access the chip while running diags, or memory diags can
|
|
* fail
|
|
*/
|
|
if (!dd->ipath_kregbase || !(dd->ipath_flags & IPATH_INITTED) ||
|
|
ipath_diag_inuse)
|
|
/* but re-arm the timer, for diags case; won't hurt other */
|
|
goto done;
|
|
|
|
if (dd->ipath_flags & IPATH_32BITCOUNTERS) {
|
|
ipath_snap_cntr(dd, dd->ipath_cregs->cr_wordsendcnt);
|
|
ipath_snap_cntr(dd, dd->ipath_cregs->cr_wordrcvcnt);
|
|
ipath_snap_cntr(dd, dd->ipath_cregs->cr_pktsendcnt);
|
|
ipath_snap_cntr(dd, dd->ipath_cregs->cr_pktrcvcnt);
|
|
}
|
|
|
|
ipath_qcheck(dd);
|
|
|
|
/*
|
|
* deal with repeat error suppression. Doesn't really matter if
|
|
* last error was almost a full interval ago, or just a few usecs
|
|
* ago; still won't get more than 2 per interval. We may want
|
|
* longer intervals for this eventually, could do with mod, counter
|
|
* or separate timer. Also see code in ipath_handle_errors() and
|
|
* ipath_handle_hwerrors().
|
|
*/
|
|
|
|
if (dd->ipath_lasterror)
|
|
dd->ipath_lasterror = 0;
|
|
if (dd->ipath_lasthwerror)
|
|
dd->ipath_lasthwerror = 0;
|
|
if ((dd->ipath_maskederrs & ~dd->ipath_ignorederrs)
|
|
&& time_after(jiffies, dd->ipath_unmasktime)) {
|
|
char ebuf[256];
|
|
int iserr;
|
|
iserr = ipath_decode_err(ebuf, sizeof ebuf,
|
|
(dd->ipath_maskederrs & ~dd->
|
|
ipath_ignorederrs));
|
|
if ((dd->ipath_maskederrs & ~dd->ipath_ignorederrs) &
|
|
~(INFINIPATH_E_RRCVEGRFULL | INFINIPATH_E_RRCVHDRFULL |
|
|
INFINIPATH_E_PKTERRS ))
|
|
ipath_dev_err(dd, "Re-enabling masked errors "
|
|
"(%s)\n", ebuf);
|
|
else {
|
|
/*
|
|
* rcvegrfull and rcvhdrqfull are "normal", for some
|
|
* types of processes (mostly benchmarks) that send
|
|
* huge numbers of messages, while not processing
|
|
* them. So only complain about these at debug
|
|
* level.
|
|
*/
|
|
if (iserr)
|
|
ipath_dbg("Re-enabling queue full errors (%s)\n",
|
|
ebuf);
|
|
else
|
|
ipath_cdbg(ERRPKT, "Re-enabling packet"
|
|
" problem interrupt (%s)\n", ebuf);
|
|
}
|
|
dd->ipath_maskederrs = dd->ipath_ignorederrs;
|
|
ipath_write_kreg(dd, dd->ipath_kregs->kr_errormask,
|
|
~dd->ipath_maskederrs);
|
|
}
|
|
|
|
/* limit qfull messages to ~one per minute per port */
|
|
if ((++cnt & 0x10)) {
|
|
for (val = dd->ipath_cfgports - 1; ((int)val) >= 0;
|
|
val--) {
|
|
if (dd->ipath_lastegrheads[val] != -1)
|
|
dd->ipath_lastegrheads[val] = -1;
|
|
if (dd->ipath_lastrcvhdrqtails[val] != -1)
|
|
dd->ipath_lastrcvhdrqtails[val] = -1;
|
|
}
|
|
}
|
|
|
|
done:
|
|
mod_timer(&dd->ipath_stats_timer, jiffies + HZ * 5);
|
|
}
|