linux-stable/drivers/infiniband/hw/hfi1/chip.h

1431 lines
42 KiB
C
Raw Normal View History

/* SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause */
/*
* Copyright(c) 2015 - 2020 Intel Corporation.
*/
#ifndef _CHIP_H
#define _CHIP_H
/*
* This file contains all of the defines that is specific to the HFI chip
*/
/* sizes */
#define BITS_PER_REGISTER (BITS_PER_BYTE * sizeof(u64))
#define NUM_INTERRUPT_SOURCES 768
#define RXE_NUM_CONTEXTS 160
#define RXE_PER_CONTEXT_SIZE 0x1000 /* 4k */
#define RXE_NUM_TID_FLOWS 32
#define RXE_NUM_DATA_VL 8
#define TXE_NUM_CONTEXTS 160
#define TXE_NUM_SDMA_ENGINES 16
#define NUM_CONTEXTS_PER_SET 8
#define VL_ARB_HIGH_PRIO_TABLE_SIZE 16
#define VL_ARB_LOW_PRIO_TABLE_SIZE 16
#define VL_ARB_TABLE_SIZE 16
#define TXE_NUM_32_BIT_COUNTER 7
#define TXE_NUM_64_BIT_COUNTER 30
#define TXE_NUM_DATA_VL 8
#define TXE_PIO_SIZE (32 * 0x100000) /* 32 MB */
#define PIO_BLOCK_SIZE 64 /* bytes */
#define SDMA_BLOCK_SIZE 64 /* bytes */
#define RCV_BUF_BLOCK_SIZE 64 /* bytes */
#define PIO_CMASK 0x7ff /* counter mask for free and fill counters */
#define MAX_EAGER_ENTRIES 2048 /* max receive eager entries */
#define MAX_TID_PAIR_ENTRIES 1024 /* max receive expected pairs */
/*
* Virtual? Allocation Unit, defined as AU = 8*2^vAU, 64 bytes, AU is fixed
* at 64 bytes for all generation one devices
*/
#define CM_VAU 3
/* HFI link credit count, AKA receive buffer depth (RBUF_DEPTH) */
#define CM_GLOBAL_CREDITS 0x880
/* Number of PKey entries in the HW */
#define MAX_PKEY_VALUES 16
#include "chip_registers.h"
#define RXE_PER_CONTEXT_USER (RXE + RXE_PER_CONTEXT_OFFSET)
#define TXE_PIO_SEND (TXE + TXE_PIO_SEND_OFFSET)
/* PBC flags */
#define PBC_INTR BIT_ULL(31)
#define PBC_DC_INFO_SHIFT (30)
#define PBC_DC_INFO BIT_ULL(PBC_DC_INFO_SHIFT)
#define PBC_TEST_EBP BIT_ULL(29)
#define PBC_PACKET_BYPASS BIT_ULL(28)
#define PBC_CREDIT_RETURN BIT_ULL(25)
#define PBC_INSERT_BYPASS_ICRC BIT_ULL(24)
#define PBC_TEST_BAD_ICRC BIT_ULL(23)
#define PBC_FECN BIT_ULL(22)
/* PbcInsertHcrc field settings */
#define PBC_IHCRC_LKDETH 0x0 /* insert @ local KDETH offset */
#define PBC_IHCRC_GKDETH 0x1 /* insert @ global KDETH offset */
#define PBC_IHCRC_NONE 0x2 /* no HCRC inserted */
/* PBC fields */
#define PBC_STATIC_RATE_CONTROL_COUNT_SHIFT 32
#define PBC_STATIC_RATE_CONTROL_COUNT_MASK 0xffffull
#define PBC_STATIC_RATE_CONTROL_COUNT_SMASK \
(PBC_STATIC_RATE_CONTROL_COUNT_MASK << \
PBC_STATIC_RATE_CONTROL_COUNT_SHIFT)
#define PBC_INSERT_HCRC_SHIFT 26
#define PBC_INSERT_HCRC_MASK 0x3ull
#define PBC_INSERT_HCRC_SMASK \
(PBC_INSERT_HCRC_MASK << PBC_INSERT_HCRC_SHIFT)
#define PBC_VL_SHIFT 12
#define PBC_VL_MASK 0xfull
#define PBC_VL_SMASK (PBC_VL_MASK << PBC_VL_SHIFT)
#define PBC_LENGTH_DWS_SHIFT 0
#define PBC_LENGTH_DWS_MASK 0xfffull
#define PBC_LENGTH_DWS_SMASK \
(PBC_LENGTH_DWS_MASK << PBC_LENGTH_DWS_SHIFT)
/* Credit Return Fields */
#define CR_COUNTER_SHIFT 0
#define CR_COUNTER_MASK 0x7ffull
#define CR_COUNTER_SMASK (CR_COUNTER_MASK << CR_COUNTER_SHIFT)
#define CR_STATUS_SHIFT 11
#define CR_STATUS_MASK 0x1ull
#define CR_STATUS_SMASK (CR_STATUS_MASK << CR_STATUS_SHIFT)
#define CR_CREDIT_RETURN_DUE_TO_PBC_SHIFT 12
#define CR_CREDIT_RETURN_DUE_TO_PBC_MASK 0x1ull
#define CR_CREDIT_RETURN_DUE_TO_PBC_SMASK \
(CR_CREDIT_RETURN_DUE_TO_PBC_MASK << \
CR_CREDIT_RETURN_DUE_TO_PBC_SHIFT)
#define CR_CREDIT_RETURN_DUE_TO_THRESHOLD_SHIFT 13
#define CR_CREDIT_RETURN_DUE_TO_THRESHOLD_MASK 0x1ull
#define CR_CREDIT_RETURN_DUE_TO_THRESHOLD_SMASK \
(CR_CREDIT_RETURN_DUE_TO_THRESHOLD_MASK << \
CR_CREDIT_RETURN_DUE_TO_THRESHOLD_SHIFT)
#define CR_CREDIT_RETURN_DUE_TO_ERR_SHIFT 14
#define CR_CREDIT_RETURN_DUE_TO_ERR_MASK 0x1ull
#define CR_CREDIT_RETURN_DUE_TO_ERR_SMASK \
(CR_CREDIT_RETURN_DUE_TO_ERR_MASK << \
CR_CREDIT_RETURN_DUE_TO_ERR_SHIFT)
#define CR_CREDIT_RETURN_DUE_TO_FORCE_SHIFT 15
#define CR_CREDIT_RETURN_DUE_TO_FORCE_MASK 0x1ull
#define CR_CREDIT_RETURN_DUE_TO_FORCE_SMASK \
(CR_CREDIT_RETURN_DUE_TO_FORCE_MASK << \
CR_CREDIT_RETURN_DUE_TO_FORCE_SHIFT)
/* Specific IRQ sources */
#define CCE_ERR_INT 0
#define RXE_ERR_INT 1
#define MISC_ERR_INT 2
#define PIO_ERR_INT 4
#define SDMA_ERR_INT 5
#define EGRESS_ERR_INT 6
#define TXE_ERR_INT 7
#define PBC_INT 240
#define GPIO_ASSERT_INT 241
#define QSFP1_INT 242
#define QSFP2_INT 243
#define TCRIT_INT 244
/* interrupt source ranges */
#define IS_FIRST_SOURCE CCE_ERR_INT
#define IS_GENERAL_ERR_START 0
#define IS_SDMAENG_ERR_START 16
#define IS_SENDCTXT_ERR_START 32
#define IS_SDMA_START 192
#define IS_SDMA_PROGRESS_START 208
#define IS_SDMA_IDLE_START 224
#define IS_VARIOUS_START 240
#define IS_DC_START 248
#define IS_RCVAVAIL_START 256
#define IS_RCVURGENT_START 416
#define IS_SENDCREDIT_START 576
#define IS_RESERVED_START 736
#define IS_LAST_SOURCE 767
/* derived interrupt source values */
#define IS_GENERAL_ERR_END 7
#define IS_SDMAENG_ERR_END 31
#define IS_SENDCTXT_ERR_END 191
#define IS_SDMA_END 207
#define IS_SDMA_PROGRESS_END 223
#define IS_SDMA_IDLE_END 239
#define IS_VARIOUS_END 244
#define IS_DC_END 255
#define IS_RCVAVAIL_END 415
#define IS_RCVURGENT_END 575
#define IS_SENDCREDIT_END 735
#define IS_RESERVED_END IS_LAST_SOURCE
/* DCC_CFG_PORT_CONFIG logical link states */
#define LSTATE_DOWN 0x1
#define LSTATE_INIT 0x2
#define LSTATE_ARMED 0x3
#define LSTATE_ACTIVE 0x4
/* DCC_CFG_RESET reset states */
#define LCB_RX_FPE_TX_FPE_INTO_RESET (DCC_CFG_RESET_RESET_LCB | \
DCC_CFG_RESET_RESET_TX_FPE | \
DCC_CFG_RESET_RESET_RX_FPE | \
DCC_CFG_RESET_ENABLE_CCLK_BCC)
/* 0x17 */
#define LCB_RX_FPE_TX_FPE_OUT_OF_RESET DCC_CFG_RESET_ENABLE_CCLK_BCC /* 0x10 */
/* DC8051_STS_CUR_STATE port values (physical link states) */
#define PLS_DISABLED 0x30
#define PLS_OFFLINE 0x90
#define PLS_OFFLINE_QUIET 0x90
#define PLS_OFFLINE_PLANNED_DOWN_INFORM 0x91
#define PLS_OFFLINE_READY_TO_QUIET_LT 0x92
#define PLS_OFFLINE_REPORT_FAILURE 0x93
#define PLS_OFFLINE_READY_TO_QUIET_BCC 0x94
#define PLS_OFFLINE_QUIET_DURATION 0x95
#define PLS_POLLING 0x20
#define PLS_POLLING_QUIET 0x20
#define PLS_POLLING_ACTIVE 0x21
#define PLS_CONFIGPHY 0x40
#define PLS_CONFIGPHY_DEBOUCE 0x40
#define PLS_CONFIGPHY_ESTCOMM 0x41
#define PLS_CONFIGPHY_ESTCOMM_TXRX_HUNT 0x42
#define PLS_CONFIGPHY_ESTCOMM_LOCAL_COMPLETE 0x43
#define PLS_CONFIGPHY_OPTEQ 0x44
#define PLS_CONFIGPHY_OPTEQ_OPTIMIZING 0x44
#define PLS_CONFIGPHY_OPTEQ_LOCAL_COMPLETE 0x45
#define PLS_CONFIGPHY_VERIFYCAP 0x46
#define PLS_CONFIGPHY_VERIFYCAP_EXCHANGE 0x46
#define PLS_CONFIGPHY_VERIFYCAP_LOCAL_COMPLETE 0x47
#define PLS_CONFIGLT 0x48
#define PLS_CONFIGLT_CONFIGURE 0x48
#define PLS_CONFIGLT_LINK_TRANSFER_ACTIVE 0x49
#define PLS_LINKUP 0x50
#define PLS_PHYTEST 0xB0
#define PLS_INTERNAL_SERDES_LOOPBACK 0xe1
#define PLS_QUICK_LINKUP 0xe2
/* DC_DC8051_CFG_HOST_CMD_0.REQ_TYPE - 8051 host commands */
#define HCMD_LOAD_CONFIG_DATA 0x01
#define HCMD_READ_CONFIG_DATA 0x02
#define HCMD_CHANGE_PHY_STATE 0x03
#define HCMD_SEND_LCB_IDLE_MSG 0x04
#define HCMD_MISC 0x05
#define HCMD_READ_LCB_IDLE_MSG 0x06
#define HCMD_READ_LCB_CSR 0x07
#define HCMD_WRITE_LCB_CSR 0x08
#define HCMD_INTERFACE_TEST 0xff
/* DC_DC8051_CFG_HOST_CMD_1.RETURN_CODE - 8051 host command return */
#define HCMD_SUCCESS 2
/* DC_DC8051_DBG_ERR_INFO_SET_BY_8051.ERROR - error flags */
#define SPICO_ROM_FAILED BIT(0)
#define UNKNOWN_FRAME BIT(1)
#define TARGET_BER_NOT_MET BIT(2)
#define FAILED_SERDES_INTERNAL_LOOPBACK BIT(3)
#define FAILED_SERDES_INIT BIT(4)
#define FAILED_LNI_POLLING BIT(5)
#define FAILED_LNI_DEBOUNCE BIT(6)
#define FAILED_LNI_ESTBCOMM BIT(7)
#define FAILED_LNI_OPTEQ BIT(8)
#define FAILED_LNI_VERIFY_CAP1 BIT(9)
#define FAILED_LNI_VERIFY_CAP2 BIT(10)
#define FAILED_LNI_CONFIGLT BIT(11)
#define HOST_HANDSHAKE_TIMEOUT BIT(12)
#define EXTERNAL_DEVICE_REQ_TIMEOUT BIT(13)
#define FAILED_LNI (FAILED_LNI_POLLING | FAILED_LNI_DEBOUNCE \
| FAILED_LNI_ESTBCOMM | FAILED_LNI_OPTEQ \
| FAILED_LNI_VERIFY_CAP1 \
| FAILED_LNI_VERIFY_CAP2 \
| FAILED_LNI_CONFIGLT | HOST_HANDSHAKE_TIMEOUT \
| EXTERNAL_DEVICE_REQ_TIMEOUT)
/* DC_DC8051_DBG_ERR_INFO_SET_BY_8051.HOST_MSG - host message flags */
#define HOST_REQ_DONE BIT(0)
#define BC_PWR_MGM_MSG BIT(1)
#define BC_SMA_MSG BIT(2)
#define BC_BCC_UNKNOWN_MSG BIT(3)
#define BC_IDLE_UNKNOWN_MSG BIT(4)
#define EXT_DEVICE_CFG_REQ BIT(5)
#define VERIFY_CAP_FRAME BIT(6)
#define LINKUP_ACHIEVED BIT(7)
#define LINK_GOING_DOWN BIT(8)
#define LINK_WIDTH_DOWNGRADED BIT(9)
/* DC_DC8051_CFG_EXT_DEV_1.REQ_TYPE - 8051 host requests */
#define HREQ_LOAD_CONFIG 0x01
#define HREQ_SAVE_CONFIG 0x02
#define HREQ_READ_CONFIG 0x03
#define HREQ_SET_TX_EQ_ABS 0x04
#define HREQ_SET_TX_EQ_REL 0x05
#define HREQ_ENABLE 0x06
#define HREQ_LCB_RESET 0x07
#define HREQ_CONFIG_DONE 0xfe
#define HREQ_INTERFACE_TEST 0xff
/* DC_DC8051_CFG_EXT_DEV_0.RETURN_CODE - 8051 host request return codes */
#define HREQ_INVALID 0x01
#define HREQ_SUCCESS 0x02
#define HREQ_NOT_SUPPORTED 0x03
#define HREQ_FEATURE_NOT_SUPPORTED 0x04 /* request specific feature */
#define HREQ_REQUEST_REJECTED 0xfe
#define HREQ_EXECUTION_ONGOING 0xff
/* MISC host command functions */
#define HCMD_MISC_REQUEST_LCB_ACCESS 0x1
#define HCMD_MISC_GRANT_LCB_ACCESS 0x2
/* idle flit message types */
#define IDLE_PHYSICAL_LINK_MGMT 0x1
#define IDLE_CRU 0x2
#define IDLE_SMA 0x3
#define IDLE_POWER_MGMT 0x4
/* idle flit message send fields (both send and read) */
#define IDLE_PAYLOAD_MASK 0xffffffffffull /* 40 bits */
#define IDLE_PAYLOAD_SHIFT 8
#define IDLE_MSG_TYPE_MASK 0xf
#define IDLE_MSG_TYPE_SHIFT 0
/* idle flit message read fields */
#define READ_IDLE_MSG_TYPE_MASK 0xf
#define READ_IDLE_MSG_TYPE_SHIFT 0
/* SMA idle flit payload commands */
#define SMA_IDLE_ARM 1
#define SMA_IDLE_ACTIVE 2
/* DC_DC8051_CFG_MODE.GENERAL bits */
#define DISABLE_SELF_GUID_CHECK 0x2
/* Bad L2 frame error code */
#define BAD_L2_ERR 0x6
/*
* Eager buffer minimum and maximum sizes supported by the hardware.
* All power-of-two sizes in between are supported as well.
* MAX_EAGER_BUFFER_TOTAL is the maximum size of memory
* allocatable for Eager buffer to a single context. All others
* are limits for the RcvArray entries.
*/
#define MIN_EAGER_BUFFER (4 * 1024)
#define MAX_EAGER_BUFFER (256 * 1024)
#define MAX_EAGER_BUFFER_TOTAL (64 * (1 << 20)) /* max per ctxt 64MB */
#define MAX_EXPECTED_BUFFER (2048 * 1024)
#define HFI1_MIN_HDRQ_EGRBUF_CNT 32
#define HFI1_MAX_HDRQ_EGRBUF_CNT 16352
/*
* Receive expected base and count and eager base and count increment -
* the CSR fields hold multiples of this value.
*/
#define RCV_SHIFT 3
#define RCV_INCREMENT BIT(RCV_SHIFT)
/*
* Receive header queue entry increment - the CSR holds multiples of
* this value.
*/
#define HDRQ_SIZE_SHIFT 5
#define HDRQ_INCREMENT BIT(HDRQ_SIZE_SHIFT)
/*
* Freeze handling flags
*/
#define FREEZE_ABORT 0x01 /* do not do recovery */
#define FREEZE_SELF 0x02 /* initiate the freeze */
#define FREEZE_LINK_DOWN 0x04 /* link is down */
/*
* Chip implementation codes.
*/
#define ICODE_RTL_SILICON 0x00
#define ICODE_RTL_VCS_SIMULATION 0x01
#define ICODE_FPGA_EMULATION 0x02
#define ICODE_FUNCTIONAL_SIMULATOR 0x03
/*
* 8051 data memory size.
*/
#define DC8051_DATA_MEM_SIZE 0x1000
/*
* 8051 firmware registers
*/
#define NUM_GENERAL_FIELDS 0x17
#define NUM_LANE_FIELDS 0x8
/* 8051 general register Field IDs */
#define LINK_OPTIMIZATION_SETTINGS 0x00
#define LINK_TUNING_PARAMETERS 0x02
#define DC_HOST_COMM_SETTINGS 0x03
#define TX_SETTINGS 0x06
#define VERIFY_CAP_LOCAL_PHY 0x07
#define VERIFY_CAP_LOCAL_FABRIC 0x08
#define VERIFY_CAP_LOCAL_LINK_MODE 0x09
#define LOCAL_DEVICE_ID 0x0a
#define RESERVED_REGISTERS 0x0b
#define LOCAL_LNI_INFO 0x0c
#define REMOTE_LNI_INFO 0x0d
#define MISC_STATUS 0x0e
#define VERIFY_CAP_REMOTE_PHY 0x0f
#define VERIFY_CAP_REMOTE_FABRIC 0x10
#define VERIFY_CAP_REMOTE_LINK_WIDTH 0x11
#define LAST_LOCAL_STATE_COMPLETE 0x12
#define LAST_REMOTE_STATE_COMPLETE 0x13
#define LINK_QUALITY_INFO 0x14
#define REMOTE_DEVICE_ID 0x15
#define LINK_DOWN_REASON 0x16 /* first byte of offset 0x16 */
#define VERSION_PATCH 0x16 /* last byte of offset 0x16 */
/* 8051 lane specific register field IDs */
#define TX_EQ_SETTINGS 0x00
#define CHANNEL_LOSS_SETTINGS 0x05
/* Lane ID for general configuration registers */
#define GENERAL_CONFIG 4
/* LINK_TUNING_PARAMETERS fields */
#define TUNING_METHOD_SHIFT 24
/* LINK_OPTIMIZATION_SETTINGS fields */
#define ENABLE_EXT_DEV_CONFIG_SHIFT 24
/* LOAD_DATA 8051 command shifts and fields */
#define LOAD_DATA_FIELD_ID_SHIFT 40
#define LOAD_DATA_FIELD_ID_MASK 0xfull
#define LOAD_DATA_LANE_ID_SHIFT 32
#define LOAD_DATA_LANE_ID_MASK 0xfull
#define LOAD_DATA_DATA_SHIFT 0x0
#define LOAD_DATA_DATA_MASK 0xffffffffull
/* READ_DATA 8051 command shifts and fields */
#define READ_DATA_FIELD_ID_SHIFT 40
#define READ_DATA_FIELD_ID_MASK 0xffull
#define READ_DATA_LANE_ID_SHIFT 32
#define READ_DATA_LANE_ID_MASK 0xffull
#define READ_DATA_DATA_SHIFT 0x0
#define READ_DATA_DATA_MASK 0xffffffffull
/* TX settings fields */
#define ENABLE_LANE_TX_SHIFT 0
#define ENABLE_LANE_TX_MASK 0xff
#define TX_POLARITY_INVERSION_SHIFT 8
#define TX_POLARITY_INVERSION_MASK 0xff
#define RX_POLARITY_INVERSION_SHIFT 16
#define RX_POLARITY_INVERSION_MASK 0xff
#define MAX_RATE_SHIFT 24
#define MAX_RATE_MASK 0xff
/* verify capability PHY fields */
#define CONTINIOUS_REMOTE_UPDATE_SUPPORT_SHIFT 0x4
#define CONTINIOUS_REMOTE_UPDATE_SUPPORT_MASK 0x1
#define POWER_MANAGEMENT_SHIFT 0x0
#define POWER_MANAGEMENT_MASK 0xf
/* 8051 lane register Field IDs */
#define SPICO_FW_VERSION 0x7 /* SPICO firmware version */
/* SPICO firmware version fields */
#define SPICO_ROM_VERSION_SHIFT 0
#define SPICO_ROM_VERSION_MASK 0xffff
#define SPICO_ROM_PROD_ID_SHIFT 16
#define SPICO_ROM_PROD_ID_MASK 0xffff
/* verify capability fabric fields */
#define VAU_SHIFT 0
#define VAU_MASK 0x0007
#define Z_SHIFT 3
#define Z_MASK 0x0001
#define VCU_SHIFT 4
#define VCU_MASK 0x0007
#define VL15BUF_SHIFT 8
#define VL15BUF_MASK 0x0fff
#define CRC_SIZES_SHIFT 20
#define CRC_SIZES_MASK 0x7
/* verify capability local link width fields */
#define LINK_WIDTH_SHIFT 0 /* also for remote link width */
#define LINK_WIDTH_MASK 0xffff /* also for remote link width */
#define LOCAL_FLAG_BITS_SHIFT 16
#define LOCAL_FLAG_BITS_MASK 0xff
#define MISC_CONFIG_BITS_SHIFT 24
#define MISC_CONFIG_BITS_MASK 0xff
/* verify capability remote link width fields */
#define REMOTE_TX_RATE_SHIFT 16
#define REMOTE_TX_RATE_MASK 0xff
/* LOCAL_DEVICE_ID fields */
#define LOCAL_DEVICE_REV_SHIFT 0
#define LOCAL_DEVICE_REV_MASK 0xff
#define LOCAL_DEVICE_ID_SHIFT 8
#define LOCAL_DEVICE_ID_MASK 0xffff
/* REMOTE_DEVICE_ID fields */
#define REMOTE_DEVICE_REV_SHIFT 0
#define REMOTE_DEVICE_REV_MASK 0xff
#define REMOTE_DEVICE_ID_SHIFT 8
#define REMOTE_DEVICE_ID_MASK 0xffff
/* local LNI link width fields */
#define ENABLE_LANE_RX_SHIFT 16
#define ENABLE_LANE_RX_MASK 0xff
/* mask, shift for reading 'mgmt_enabled' value from REMOTE_LNI_INFO field */
#define MGMT_ALLOWED_SHIFT 23
#define MGMT_ALLOWED_MASK 0x1
/* mask, shift for 'link_quality' within LINK_QUALITY_INFO field */
#define LINK_QUALITY_SHIFT 24
#define LINK_QUALITY_MASK 0x7
/*
* mask, shift for reading 'planned_down_remote_reason_code'
* from LINK_QUALITY_INFO field
*/
#define DOWN_REMOTE_REASON_SHIFT 16
#define DOWN_REMOTE_REASON_MASK 0xff
#define HOST_INTERFACE_VERSION 1
#define HOST_INTERFACE_VERSION_SHIFT 16
#define HOST_INTERFACE_VERSION_MASK 0xff
/* verify capability PHY power management bits */
#define PWRM_BER_CONTROL 0x1
#define PWRM_BANDWIDTH_CONTROL 0x2
/* 8051 link down reasons */
#define LDR_LINK_TRANSFER_ACTIVE_LOW 0xa
#define LDR_RECEIVED_LINKDOWN_IDLE_MSG 0xb
#define LDR_RECEIVED_HOST_OFFLINE_REQ 0xc
/* verify capability fabric CRC size bits */
enum {
CAP_CRC_14B = (1 << 0), /* 14b CRC */
CAP_CRC_48B = (1 << 1), /* 48b CRC */
CAP_CRC_12B_16B_PER_LANE = (1 << 2) /* 12b-16b per lane CRC */
};
#define SUPPORTED_CRCS (CAP_CRC_14B | CAP_CRC_48B)
/* misc status version fields */
#define STS_FM_VERSION_MINOR_SHIFT 16
#define STS_FM_VERSION_MINOR_MASK 0xff
#define STS_FM_VERSION_MAJOR_SHIFT 24
#define STS_FM_VERSION_MAJOR_MASK 0xff
#define STS_FM_VERSION_PATCH_SHIFT 24
#define STS_FM_VERSION_PATCH_MASK 0xff
/* LCB_CFG_CRC_MODE TX_VAL and RX_VAL CRC mode values */
#define LCB_CRC_16B 0x0 /* 16b CRC */
#define LCB_CRC_14B 0x1 /* 14b CRC */
#define LCB_CRC_48B 0x2 /* 48b CRC */
#define LCB_CRC_12B_16B_PER_LANE 0x3 /* 12b-16b per lane CRC */
/*
* the following enum is (almost) a copy/paste of the definition
* in the OPA spec, section 20.2.2.6.8 (PortInfo)
*/
enum {
PORT_LTP_CRC_MODE_NONE = 0,
PORT_LTP_CRC_MODE_14 = 1, /* 14-bit LTP CRC mode (optional) */
PORT_LTP_CRC_MODE_16 = 2, /* 16-bit LTP CRC mode */
PORT_LTP_CRC_MODE_48 = 4,
/* 48-bit overlapping LTP CRC mode (optional) */
PORT_LTP_CRC_MODE_PER_LANE = 8
/* 12 to 16 bit per lane LTP CRC mode (optional) */
};
/* timeouts */
#define LINK_RESTART_DELAY 1000 /* link restart delay, in ms */
#define TIMEOUT_8051_START 5000 /* 8051 start timeout, in ms */
#define DC8051_COMMAND_TIMEOUT 1000 /* DC8051 command timeout, in ms */
#define FREEZE_STATUS_TIMEOUT 20 /* wait for freeze indicators, in ms */
#define VL_STATUS_CLEAR_TIMEOUT 5000 /* per-VL status clear, in ms */
#define CCE_STATUS_TIMEOUT 10 /* time to clear CCE Status, in ms */
/* cclock tick time, in picoseconds per tick: 1/speed * 10^12 */
#define ASIC_CCLOCK_PS 1242 /* 805 MHz */
#define FPGA_CCLOCK_PS 30300 /* 33 MHz */
/*
* Mask of enabled MISC errors. Do not enable the two RSA engine errors -
* see firmware.c:run_rsa() for details.
*/
#define DRIVER_MISC_MASK \
(~(MISC_ERR_STATUS_MISC_FW_AUTH_FAILED_ERR_SMASK \
| MISC_ERR_STATUS_MISC_KEY_MISMATCH_ERR_SMASK))
/* valid values for the loopback module parameter */
#define LOOPBACK_NONE 0 /* no loopback - default */
#define LOOPBACK_SERDES 1
#define LOOPBACK_LCB 2
#define LOOPBACK_CABLE 3 /* external cable */
/* set up bits in MISC_CONFIG_BITS */
#define LOOPBACK_SERDES_CONFIG_BIT_MASK_SHIFT 0
#define EXT_CFG_LCB_RESET_SUPPORTED_SHIFT 3
/* read and write hardware registers */
u64 read_csr(const struct hfi1_devdata *dd, u32 offset);
void write_csr(const struct hfi1_devdata *dd, u32 offset, u64 value);
/*
* The *_kctxt_* flavor of the CSR read/write functions are for
* per-context or per-SDMA CSRs that are not mappable to user-space.
* Their spacing is not a PAGE_SIZE multiple.
*/
static inline u64 read_kctxt_csr(const struct hfi1_devdata *dd, int ctxt,
u32 offset0)
{
/* kernel per-context CSRs are separated by 0x100 */
return read_csr(dd, offset0 + (0x100 * ctxt));
}
static inline void write_kctxt_csr(struct hfi1_devdata *dd, int ctxt,
u32 offset0, u64 value)
{
/* kernel per-context CSRs are separated by 0x100 */
write_csr(dd, offset0 + (0x100 * ctxt), value);
}
int read_lcb_csr(struct hfi1_devdata *dd, u32 offset, u64 *data);
int write_lcb_csr(struct hfi1_devdata *dd, u32 offset, u64 data);
void __iomem *get_csr_addr(
IB/hfi1: Fix bar0 mapping to use write combining When the debugpat kernel boot flag is turned on the following traces are printed: [ 1884.793168] x86/PAT: Overlap at 0x90000000-0x92000000 [ 1884.803510] x86/PAT: reserve_memtype added [mem 0x91200000-0x9127ffff], track uncached-minus, req write-combining, ret uncached-minus [ 1884.818167] hfi1 0000:05:00.0: hfi1_0: WC Remapped RcvArray: ffffc9000a980000 The ioremap_wc() clearly is not returning a write combining mapping due to an overlap where the RcvArray is mapped in a uncached mapping prior to creating the proposed write combining mapping. The patch replaces the single base register for uncached CSRs that used to overlap the RcvArray with two mappings. One, kregbase1, from the bar0 up to the RcvArray and another, kregbase2, from the end of the RcvArray to the pio send buffer space. A new dd field, base2_start, is used to convert the zero-based offset in the CSR routines to the correct kregbase1/kregbase2 mapping. A single direct write of the RcvArray CSRs is replaced with hfi1_put_tid() to insure correct access using the new disjoint mapping. Additionally, the kregend field is deleted since it is only ever written. patdebug now shows the RcvArray as write combining: [ 35.688990] x86/PAT: reserve_memtype added [mem 0x91200000-0x9127ffff], track write-combining, req write-combining, ret write-combining To insulate from any potential issues with write combining, all writeq are now flushed in hfi1_put_tid() and rcv_array_wc_fill(). Reviewed-by: Mitko Haralanov <mitko.haralanov@intel.com> Reviewed-by: Ashutosh Dixit <ashutosh.dixit@intel.com> Signed-off-by: Mike Marciniszyn <mike.marciniszyn@intel.com> Signed-off-by: Dennis Dalessandro <dennis.dalessandro@intel.com> Signed-off-by: Doug Ledford <dledford@redhat.com>
2017-07-24 14:45:31 +00:00
const struct hfi1_devdata *dd,
u32 offset);
static inline void __iomem *get_kctxt_csr_addr(
IB/hfi1: Fix bar0 mapping to use write combining When the debugpat kernel boot flag is turned on the following traces are printed: [ 1884.793168] x86/PAT: Overlap at 0x90000000-0x92000000 [ 1884.803510] x86/PAT: reserve_memtype added [mem 0x91200000-0x9127ffff], track uncached-minus, req write-combining, ret uncached-minus [ 1884.818167] hfi1 0000:05:00.0: hfi1_0: WC Remapped RcvArray: ffffc9000a980000 The ioremap_wc() clearly is not returning a write combining mapping due to an overlap where the RcvArray is mapped in a uncached mapping prior to creating the proposed write combining mapping. The patch replaces the single base register for uncached CSRs that used to overlap the RcvArray with two mappings. One, kregbase1, from the bar0 up to the RcvArray and another, kregbase2, from the end of the RcvArray to the pio send buffer space. A new dd field, base2_start, is used to convert the zero-based offset in the CSR routines to the correct kregbase1/kregbase2 mapping. A single direct write of the RcvArray CSRs is replaced with hfi1_put_tid() to insure correct access using the new disjoint mapping. Additionally, the kregend field is deleted since it is only ever written. patdebug now shows the RcvArray as write combining: [ 35.688990] x86/PAT: reserve_memtype added [mem 0x91200000-0x9127ffff], track write-combining, req write-combining, ret write-combining To insulate from any potential issues with write combining, all writeq are now flushed in hfi1_put_tid() and rcv_array_wc_fill(). Reviewed-by: Mitko Haralanov <mitko.haralanov@intel.com> Reviewed-by: Ashutosh Dixit <ashutosh.dixit@intel.com> Signed-off-by: Mike Marciniszyn <mike.marciniszyn@intel.com> Signed-off-by: Dennis Dalessandro <dennis.dalessandro@intel.com> Signed-off-by: Doug Ledford <dledford@redhat.com>
2017-07-24 14:45:31 +00:00
const struct hfi1_devdata *dd,
int ctxt,
u32 offset0)
{
return get_csr_addr(dd, offset0 + (0x100 * ctxt));
}
/*
* The *_uctxt_* flavor of the CSR read/write functions are for
* per-context CSRs that are mappable to user space. All these CSRs
* are spaced by a PAGE_SIZE multiple in order to be mappable to
* different processes without exposing other contexts' CSRs
*/
static inline u64 read_uctxt_csr(const struct hfi1_devdata *dd, int ctxt,
u32 offset0)
{
/* user per-context CSRs are separated by 0x1000 */
return read_csr(dd, offset0 + (0x1000 * ctxt));
}
static inline void write_uctxt_csr(struct hfi1_devdata *dd, int ctxt,
u32 offset0, u64 value)
{
/* user per-context CSRs are separated by 0x1000 */
write_csr(dd, offset0 + (0x1000 * ctxt), value);
}
static inline u32 chip_rcv_contexts(struct hfi1_devdata *dd)
{
return read_csr(dd, RCV_CONTEXTS);
}
static inline u32 chip_send_contexts(struct hfi1_devdata *dd)
{
return read_csr(dd, SEND_CONTEXTS);
}
static inline u32 chip_sdma_engines(struct hfi1_devdata *dd)
{
return read_csr(dd, SEND_DMA_ENGINES);
}
static inline u32 chip_pio_mem_size(struct hfi1_devdata *dd)
{
return read_csr(dd, SEND_PIO_MEM_SIZE);
}
static inline u32 chip_sdma_mem_size(struct hfi1_devdata *dd)
{
return read_csr(dd, SEND_DMA_MEM_SIZE);
}
static inline u32 chip_rcv_array_count(struct hfi1_devdata *dd)
{
return read_csr(dd, RCV_ARRAY_CNT);
}
u8 encode_rcv_header_entry_size(u8 size);
int hfi1_validate_rcvhdrcnt(struct hfi1_devdata *dd, uint thecnt);
void set_hdrq_regs(struct hfi1_devdata *dd, u8 ctxt, u8 entsize, u16 hdrcnt);
u64 create_pbc(struct hfi1_pportdata *ppd, u64 flags, int srate_mbs, u32 vl,
u32 dw_len);
/* firmware.c */
#define SBUS_MASTER_BROADCAST 0xfd
#define NUM_PCIE_SERDES 16 /* number of PCIe serdes on the SBus */
extern const u8 pcie_serdes_broadcast[];
extern const u8 pcie_pcs_addrs[2][NUM_PCIE_SERDES];
/* SBus commands */
#define RESET_SBUS_RECEIVER 0x20
#define WRITE_SBUS_RECEIVER 0x21
#define READ_SBUS_RECEIVER 0x22
void sbus_request(struct hfi1_devdata *dd,
u8 receiver_addr, u8 data_addr, u8 command, u32 data_in);
int sbus_request_slow(struct hfi1_devdata *dd,
u8 receiver_addr, u8 data_addr, u8 command, u32 data_in);
void set_sbus_fast_mode(struct hfi1_devdata *dd);
void clear_sbus_fast_mode(struct hfi1_devdata *dd);
int hfi1_firmware_init(struct hfi1_devdata *dd);
int load_pcie_firmware(struct hfi1_devdata *dd);
int load_firmware(struct hfi1_devdata *dd);
void dispose_firmware(void);
int acquire_hw_mutex(struct hfi1_devdata *dd);
void release_hw_mutex(struct hfi1_devdata *dd);
/*
* Bitmask of dynamic access for ASIC block chip resources. Each HFI has its
* own range of bits for the resource so it can clear its own bits on
* starting and exiting. If either HFI has the resource bit set, the
* resource is in use. The separate bit ranges are:
* HFI0 bits 7:0
* HFI1 bits 15:8
*/
#define CR_SBUS 0x01 /* SBUS, THERM, and PCIE registers */
#define CR_EPROM 0x02 /* EEP, GPIO registers */
#define CR_I2C1 0x04 /* QSFP1_OE register */
#define CR_I2C2 0x08 /* QSFP2_OE register */
#define CR_DYN_SHIFT 8 /* dynamic flag shift */
#define CR_DYN_MASK ((1ull << CR_DYN_SHIFT) - 1)
/*
* Bitmask of static ASIC states these are outside of the dynamic ASIC
* block chip resources above. These are to be set once and never cleared.
* Must be holding the SBus dynamic flag when setting.
*/
#define CR_THERM_INIT 0x010000
int acquire_chip_resource(struct hfi1_devdata *dd, u32 resource, u32 mswait);
void release_chip_resource(struct hfi1_devdata *dd, u32 resource);
bool check_chip_resource(struct hfi1_devdata *dd, u32 resource,
const char *func);
void init_chip_resources(struct hfi1_devdata *dd);
void finish_chip_resources(struct hfi1_devdata *dd);
/* ms wait time for access to an SBus resoure */
#define SBUS_TIMEOUT 4000 /* long enough for a FW download and SBR */
/* ms wait time for a qsfp (i2c) chain to become available */
#define QSFP_WAIT 20000 /* long enough for FW update to the F4 uc */
void fabric_serdes_reset(struct hfi1_devdata *dd);
int read_8051_data(struct hfi1_devdata *dd, u32 addr, u32 len, u64 *result);
/* chip.c */
void read_misc_status(struct hfi1_devdata *dd, u8 *ver_major, u8 *ver_minor,
u8 *ver_patch);
int write_host_interface_version(struct hfi1_devdata *dd, u8 version);
void read_guid(struct hfi1_devdata *dd);
int wait_fm_ready(struct hfi1_devdata *dd, u32 mstimeout);
void set_link_down_reason(struct hfi1_pportdata *ppd, u8 lcl_reason,
u8 neigh_reason, u8 rem_reason);
int set_link_state(struct hfi1_pportdata *, u32 state);
int port_ltp_to_cap(int port_ltp);
void handle_verify_cap(struct work_struct *work);
void handle_freeze(struct work_struct *work);
void handle_link_up(struct work_struct *work);
void handle_link_down(struct work_struct *work);
void handle_link_downgrade(struct work_struct *work);
void handle_link_bounce(struct work_struct *work);
void handle_start_link(struct work_struct *work);
void handle_sma_message(struct work_struct *work);
int reset_qsfp(struct hfi1_pportdata *ppd);
void qsfp_event(struct work_struct *work);
void start_freeze_handling(struct hfi1_pportdata *ppd, int flags);
int send_idle_sma(struct hfi1_devdata *dd, u64 message);
int load_8051_config(struct hfi1_devdata *, u8, u8, u32);
int read_8051_config(struct hfi1_devdata *, u8, u8, u32 *);
int start_link(struct hfi1_pportdata *ppd);
int bringup_serdes(struct hfi1_pportdata *ppd);
void set_intr_state(struct hfi1_devdata *dd, u32 enable);
bool apply_link_downgrade_policy(struct hfi1_pportdata *ppd,
bool refresh_widths);
void update_usrhead(struct hfi1_ctxtdata *rcd, u32 hd, u32 updegr, u32 egrhd,
u32 intr_adjust, u32 npkts);
int stop_drain_data_vls(struct hfi1_devdata *dd);
int open_fill_data_vls(struct hfi1_devdata *dd);
u32 ns_to_cclock(struct hfi1_devdata *dd, u32 ns);
u32 cclock_to_ns(struct hfi1_devdata *dd, u32 cclock);
void get_linkup_link_widths(struct hfi1_pportdata *ppd);
void read_ltp_rtt(struct hfi1_devdata *dd);
void clear_linkup_counters(struct hfi1_devdata *dd);
u32 hdrqempty(struct hfi1_ctxtdata *rcd);
int is_ax(struct hfi1_devdata *dd);
int is_bx(struct hfi1_devdata *dd);
bool is_urg_masked(struct hfi1_ctxtdata *rcd);
u32 read_physical_state(struct hfi1_devdata *dd);
u32 chip_to_opa_pstate(struct hfi1_devdata *dd, u32 chip_pstate);
const char *opa_lstate_name(u32 lstate);
const char *opa_pstate_name(u32 pstate);
u32 driver_pstate(struct hfi1_pportdata *ppd);
u32 driver_lstate(struct hfi1_pportdata *ppd);
int acquire_lcb_access(struct hfi1_devdata *dd, int sleep_ok);
int release_lcb_access(struct hfi1_devdata *dd, int sleep_ok);
#define LCB_START DC_LCB_CSRS
#define LCB_END DC_8051_CSRS /* next block is 8051 */
extern uint num_vls;
extern uint disable_integrity;
u64 read_dev_cntr(struct hfi1_devdata *dd, int index, int vl);
u64 write_dev_cntr(struct hfi1_devdata *dd, int index, int vl, u64 data);
u64 read_port_cntr(struct hfi1_pportdata *ppd, int index, int vl);
u64 write_port_cntr(struct hfi1_pportdata *ppd, int index, int vl, u64 data);
u32 read_logical_state(struct hfi1_devdata *dd);
void force_recv_intr(struct hfi1_ctxtdata *rcd);
/* Per VL indexes */
enum {
C_VL_0 = 0,
C_VL_1,
C_VL_2,
C_VL_3,
C_VL_4,
C_VL_5,
C_VL_6,
C_VL_7,
C_VL_15,
C_VL_COUNT
};
static inline int vl_from_idx(int idx)
{
return (idx == C_VL_15 ? 15 : idx);
}
static inline int idx_from_vl(int vl)
{
return (vl == 15 ? C_VL_15 : vl);
}
/* Per device counter indexes */
enum {
C_RCV_OVF = 0,
C_RX_LEN_ERR,
C_RX_SHORT_ERR,
C_RX_ICRC_ERR,
C_RX_EBP,
C_RX_TID_FULL,
C_RX_TID_INVALID,
C_RX_TID_FLGMS,
C_RX_CTX_EGRS,
C_RCV_TID_FLSMS,
C_CCE_PCI_CR_ST,
C_CCE_PCI_TR_ST,
C_CCE_PIO_WR_ST,
C_CCE_ERR_INT,
C_CCE_SDMA_INT,
C_CCE_MISC_INT,
C_CCE_RCV_AV_INT,
C_CCE_RCV_URG_INT,
C_CCE_SEND_CR_INT,
C_DC_UNC_ERR,
C_DC_RCV_ERR,
C_DC_FM_CFG_ERR,
C_DC_RMT_PHY_ERR,
C_DC_DROPPED_PKT,
C_DC_MC_XMIT_PKTS,
C_DC_MC_RCV_PKTS,
C_DC_XMIT_CERR,
C_DC_RCV_CERR,
C_DC_RCV_FCC,
C_DC_XMIT_FCC,
C_DC_XMIT_FLITS,
C_DC_RCV_FLITS,
C_DC_XMIT_PKTS,
C_DC_RCV_PKTS,
C_DC_RX_FLIT_VL,
C_DC_RX_PKT_VL,
C_DC_RCV_FCN,
C_DC_RCV_FCN_VL,
C_DC_RCV_BCN,
C_DC_RCV_BCN_VL,
C_DC_RCV_BBL,
C_DC_RCV_BBL_VL,
C_DC_MARK_FECN,
C_DC_MARK_FECN_VL,
C_DC_TOTAL_CRC,
C_DC_CRC_LN0,
C_DC_CRC_LN1,
C_DC_CRC_LN2,
C_DC_CRC_LN3,
C_DC_CRC_MULT_LN,
C_DC_TX_REPLAY,
C_DC_RX_REPLAY,
C_DC_SEQ_CRC_CNT,
C_DC_ESC0_ONLY_CNT,
C_DC_ESC0_PLUS1_CNT,
C_DC_ESC0_PLUS2_CNT,
C_DC_REINIT_FROM_PEER_CNT,
C_DC_SBE_CNT,
C_DC_MISC_FLG_CNT,
C_DC_PRF_GOOD_LTP_CNT,
C_DC_PRF_ACCEPTED_LTP_CNT,
C_DC_PRF_RX_FLIT_CNT,
C_DC_PRF_TX_FLIT_CNT,
C_DC_PRF_CLK_CNTR,
C_DC_PG_DBG_FLIT_CRDTS_CNT,
C_DC_PG_STS_PAUSE_COMPLETE_CNT,
C_DC_PG_STS_TX_SBE_CNT,
C_DC_PG_STS_TX_MBE_CNT,
C_SW_CPU_INTR,
C_SW_CPU_RCV_LIM,
C_SW_CTX0_SEQ_DROP,
C_SW_VTX_WAIT,
C_SW_PIO_WAIT,
staging/rdma/hfi1: Adaptive PIO for short messages The change requires a new pio_busy field in the iowait structure to track the number of outstanding pios. The new counter together with the sdma counter serve as the basis for a packet by packet decision as to which egress mechanism to use. Since packets given to different egress mechanisms are not ordered, this scheme will preserve the order. The iowait drain/wait mechanisms are extended for a pio case. An additional qp wait flag is added for the PIO drain wait case. Currently the only pio wait is for buffers, so the no_bufs_available() routine name is changed to pio_wait() and a third argument is passed with one of the two pio wait flags to generalize the routine. A module parameter is added to hold a configurable threshold. For now, the module parameter is zero. A heuristic routine is added to return the func pointer of the proper egress routine to use. The heuristic is as follows: - SMI always uses pio - GSI,UD qps <= threshold use pio - UD qps > threadhold use sdma o No coordination with sdma is required because order is not required and this qp pio count is not maintained for UD - RC/UC ONLY packets <= threshold chose as follows: o If sdmas pending, use SDMA o Otherwise use pio and enable the pio tracking count at the time the pio buffer is allocated - RC/UC ONLY packets > threshold use SDMA o If pio's are pending the pio_wait with the new wait flag is called to delay for pios to drain The threshold is potentially reduced by the QP's mtu. The sc_buffer_alloc() has two additional args (a callback, a void *) which are exploited by the RC/UC cases to pass a new complete routine and a qp *. When the shadow ring completes the credit associated with a packet, the new complete routine is called. The verbs_pio_complete() will then decrement the busy count and trigger any drain waiters in qp destroy or reset. Reviewed-by: Jubin John <jubin.john@intel.com> Reviewed-by: Dennis Dalessandro <dennis.dalessandro@intel.com> Signed-off-by: Mike Marciniszyn <mike.marciniszyn@intel.com> Signed-off-by: Doug Ledford <dledford@redhat.com>
2016-02-14 20:45:36 +00:00
C_SW_PIO_DRAIN,
C_SW_KMEM_WAIT,
C_SW_TID_WAIT,
C_SW_SEND_SCHED,
C_SDMA_DESC_FETCHED_CNT,
C_SDMA_INT_CNT,
C_SDMA_ERR_CNT,
C_SDMA_IDLE_INT_CNT,
C_SDMA_PROGRESS_INT_CNT,
/* MISC_ERR_STATUS */
C_MISC_PLL_LOCK_FAIL_ERR,
C_MISC_MBIST_FAIL_ERR,
C_MISC_INVALID_EEP_CMD_ERR,
C_MISC_EFUSE_DONE_PARITY_ERR,
C_MISC_EFUSE_WRITE_ERR,
C_MISC_EFUSE_READ_BAD_ADDR_ERR,
C_MISC_EFUSE_CSR_PARITY_ERR,
C_MISC_FW_AUTH_FAILED_ERR,
C_MISC_KEY_MISMATCH_ERR,
C_MISC_SBUS_WRITE_FAILED_ERR,
C_MISC_CSR_WRITE_BAD_ADDR_ERR,
C_MISC_CSR_READ_BAD_ADDR_ERR,
C_MISC_CSR_PARITY_ERR,
/* CceErrStatus */
/*
* A special counter that is the aggregate count
* of all the cce_err_status errors. The remainder
* are actual bits in the CceErrStatus register.
*/
C_CCE_ERR_STATUS_AGGREGATED_CNT,
C_CCE_MSIX_CSR_PARITY_ERR,
C_CCE_INT_MAP_UNC_ERR,
C_CCE_INT_MAP_COR_ERR,
C_CCE_MSIX_TABLE_UNC_ERR,
C_CCE_MSIX_TABLE_COR_ERR,
C_CCE_RXDMA_CONV_FIFO_PARITY_ERR,
C_CCE_RCPL_ASYNC_FIFO_PARITY_ERR,
C_CCE_SEG_WRITE_BAD_ADDR_ERR,
C_CCE_SEG_READ_BAD_ADDR_ERR,
C_LA_TRIGGERED,
C_CCE_TRGT_CPL_TIMEOUT_ERR,
C_PCIC_RECEIVE_PARITY_ERR,
C_PCIC_TRANSMIT_BACK_PARITY_ERR,
C_PCIC_TRANSMIT_FRONT_PARITY_ERR,
C_PCIC_CPL_DAT_Q_UNC_ERR,
C_PCIC_CPL_HD_Q_UNC_ERR,
C_PCIC_POST_DAT_Q_UNC_ERR,
C_PCIC_POST_HD_Q_UNC_ERR,
C_PCIC_RETRY_SOT_MEM_UNC_ERR,
C_PCIC_RETRY_MEM_UNC_ERR,
C_PCIC_N_POST_DAT_Q_PARITY_ERR,
C_PCIC_N_POST_H_Q_PARITY_ERR,
C_PCIC_CPL_DAT_Q_COR_ERR,
C_PCIC_CPL_HD_Q_COR_ERR,
C_PCIC_POST_DAT_Q_COR_ERR,
C_PCIC_POST_HD_Q_COR_ERR,
C_PCIC_RETRY_SOT_MEM_COR_ERR,
C_PCIC_RETRY_MEM_COR_ERR,
C_CCE_CLI1_ASYNC_FIFO_DBG_PARITY_ERR,
C_CCE_CLI1_ASYNC_FIFO_RXDMA_PARITY_ERR,
C_CCE_CLI1_ASYNC_FIFO_SDMA_HD_PARITY_ERR,
C_CCE_CLI1_ASYNC_FIFO_PIO_CRDT_PARITY_ERR,
C_CCE_CLI2_ASYNC_FIFO_PARITY_ERR,
C_CCE_CSR_CFG_BUS_PARITY_ERR,
C_CCE_CLI0_ASYNC_FIFO_PARTIY_ERR,
C_CCE_RSPD_DATA_PARITY_ERR,
C_CCE_TRGT_ACCESS_ERR,
C_CCE_TRGT_ASYNC_FIFO_PARITY_ERR,
C_CCE_CSR_WRITE_BAD_ADDR_ERR,
C_CCE_CSR_READ_BAD_ADDR_ERR,
C_CCE_CSR_PARITY_ERR,
/* RcvErrStatus */
C_RX_CSR_PARITY_ERR,
C_RX_CSR_WRITE_BAD_ADDR_ERR,
C_RX_CSR_READ_BAD_ADDR_ERR,
C_RX_DMA_CSR_UNC_ERR,
C_RX_DMA_DQ_FSM_ENCODING_ERR,
C_RX_DMA_EQ_FSM_ENCODING_ERR,
C_RX_DMA_CSR_PARITY_ERR,
C_RX_RBUF_DATA_COR_ERR,
C_RX_RBUF_DATA_UNC_ERR,
C_RX_DMA_DATA_FIFO_RD_COR_ERR,
C_RX_DMA_DATA_FIFO_RD_UNC_ERR,
C_RX_DMA_HDR_FIFO_RD_COR_ERR,
C_RX_DMA_HDR_FIFO_RD_UNC_ERR,
C_RX_RBUF_DESC_PART2_COR_ERR,
C_RX_RBUF_DESC_PART2_UNC_ERR,
C_RX_RBUF_DESC_PART1_COR_ERR,
C_RX_RBUF_DESC_PART1_UNC_ERR,
C_RX_HQ_INTR_FSM_ERR,
C_RX_HQ_INTR_CSR_PARITY_ERR,
C_RX_LOOKUP_CSR_PARITY_ERR,
C_RX_LOOKUP_RCV_ARRAY_COR_ERR,
C_RX_LOOKUP_RCV_ARRAY_UNC_ERR,
C_RX_LOOKUP_DES_PART2_PARITY_ERR,
C_RX_LOOKUP_DES_PART1_UNC_COR_ERR,
C_RX_LOOKUP_DES_PART1_UNC_ERR,
C_RX_RBUF_NEXT_FREE_BUF_COR_ERR,
C_RX_RBUF_NEXT_FREE_BUF_UNC_ERR,
C_RX_RBUF_FL_INIT_WR_ADDR_PARITY_ERR,
C_RX_RBUF_FL_INITDONE_PARITY_ERR,
C_RX_RBUF_FL_WRITE_ADDR_PARITY_ERR,
C_RX_RBUF_FL_RD_ADDR_PARITY_ERR,
C_RX_RBUF_EMPTY_ERR,
C_RX_RBUF_FULL_ERR,
C_RX_RBUF_BAD_LOOKUP_ERR,
C_RX_RBUF_CTX_ID_PARITY_ERR,
C_RX_RBUF_CSR_QEOPDW_PARITY_ERR,
C_RX_RBUF_CSR_Q_NUM_OF_PKT_PARITY_ERR,
C_RX_RBUF_CSR_Q_T1_PTR_PARITY_ERR,
C_RX_RBUF_CSR_Q_HD_PTR_PARITY_ERR,
C_RX_RBUF_CSR_Q_VLD_BIT_PARITY_ERR,
C_RX_RBUF_CSR_Q_NEXT_BUF_PARITY_ERR,
C_RX_RBUF_CSR_Q_ENT_CNT_PARITY_ERR,
C_RX_RBUF_CSR_Q_HEAD_BUF_NUM_PARITY_ERR,
C_RX_RBUF_BLOCK_LIST_READ_COR_ERR,
C_RX_RBUF_BLOCK_LIST_READ_UNC_ERR,
C_RX_RBUF_LOOKUP_DES_COR_ERR,
C_RX_RBUF_LOOKUP_DES_UNC_ERR,
C_RX_RBUF_LOOKUP_DES_REG_UNC_COR_ERR,
C_RX_RBUF_LOOKUP_DES_REG_UNC_ERR,
C_RX_RBUF_FREE_LIST_COR_ERR,
C_RX_RBUF_FREE_LIST_UNC_ERR,
C_RX_RCV_FSM_ENCODING_ERR,
C_RX_DMA_FLAG_COR_ERR,
C_RX_DMA_FLAG_UNC_ERR,
C_RX_DC_SOP_EOP_PARITY_ERR,
C_RX_RCV_CSR_PARITY_ERR,
C_RX_RCV_QP_MAP_TABLE_COR_ERR,
C_RX_RCV_QP_MAP_TABLE_UNC_ERR,
C_RX_RCV_DATA_COR_ERR,
C_RX_RCV_DATA_UNC_ERR,
C_RX_RCV_HDR_COR_ERR,
C_RX_RCV_HDR_UNC_ERR,
C_RX_DC_INTF_PARITY_ERR,
C_RX_DMA_CSR_COR_ERR,
/* SendPioErrStatus */
C_PIO_PEC_SOP_HEAD_PARITY_ERR,
C_PIO_PCC_SOP_HEAD_PARITY_ERR,
C_PIO_LAST_RETURNED_CNT_PARITY_ERR,
C_PIO_CURRENT_FREE_CNT_PARITY_ERR,
C_PIO_RSVD_31_ERR,
C_PIO_RSVD_30_ERR,
C_PIO_PPMC_SOP_LEN_ERR,
C_PIO_PPMC_BQC_MEM_PARITY_ERR,
C_PIO_VL_FIFO_PARITY_ERR,
C_PIO_VLF_SOP_PARITY_ERR,
C_PIO_VLF_V1_LEN_PARITY_ERR,
C_PIO_BLOCK_QW_COUNT_PARITY_ERR,
C_PIO_WRITE_QW_VALID_PARITY_ERR,
C_PIO_STATE_MACHINE_ERR,
C_PIO_WRITE_DATA_PARITY_ERR,
C_PIO_HOST_ADDR_MEM_COR_ERR,
C_PIO_HOST_ADDR_MEM_UNC_ERR,
C_PIO_PKT_EVICT_SM_OR_ARM_SM_ERR,
C_PIO_INIT_SM_IN_ERR,
C_PIO_PPMC_PBL_FIFO_ERR,
C_PIO_CREDIT_RET_FIFO_PARITY_ERR,
C_PIO_V1_LEN_MEM_BANK1_COR_ERR,
C_PIO_V1_LEN_MEM_BANK0_COR_ERR,
C_PIO_V1_LEN_MEM_BANK1_UNC_ERR,
C_PIO_V1_LEN_MEM_BANK0_UNC_ERR,
C_PIO_SM_PKT_RESET_PARITY_ERR,
C_PIO_PKT_EVICT_FIFO_PARITY_ERR,
C_PIO_SBRDCTRL_CRREL_FIFO_PARITY_ERR,
C_PIO_SBRDCTL_CRREL_PARITY_ERR,
C_PIO_PEC_FIFO_PARITY_ERR,
C_PIO_PCC_FIFO_PARITY_ERR,
C_PIO_SB_MEM_FIFO1_ERR,
C_PIO_SB_MEM_FIFO0_ERR,
C_PIO_CSR_PARITY_ERR,
C_PIO_WRITE_ADDR_PARITY_ERR,
C_PIO_WRITE_BAD_CTXT_ERR,
/* SendDmaErrStatus */
C_SDMA_PCIE_REQ_TRACKING_COR_ERR,
C_SDMA_PCIE_REQ_TRACKING_UNC_ERR,
C_SDMA_CSR_PARITY_ERR,
C_SDMA_RPY_TAG_ERR,
/* SendEgressErrStatus */
C_TX_READ_PIO_MEMORY_CSR_UNC_ERR,
C_TX_READ_SDMA_MEMORY_CSR_UNC_ERR,
C_TX_EGRESS_FIFO_COR_ERR,
C_TX_READ_PIO_MEMORY_COR_ERR,
C_TX_READ_SDMA_MEMORY_COR_ERR,
C_TX_SB_HDR_COR_ERR,
C_TX_CREDIT_OVERRUN_ERR,
C_TX_LAUNCH_FIFO8_COR_ERR,
C_TX_LAUNCH_FIFO7_COR_ERR,
C_TX_LAUNCH_FIFO6_COR_ERR,
C_TX_LAUNCH_FIFO5_COR_ERR,
C_TX_LAUNCH_FIFO4_COR_ERR,
C_TX_LAUNCH_FIFO3_COR_ERR,
C_TX_LAUNCH_FIFO2_COR_ERR,
C_TX_LAUNCH_FIFO1_COR_ERR,
C_TX_LAUNCH_FIFO0_COR_ERR,
C_TX_CREDIT_RETURN_VL_ERR,
C_TX_HCRC_INSERTION_ERR,
C_TX_EGRESS_FIFI_UNC_ERR,
C_TX_READ_PIO_MEMORY_UNC_ERR,
C_TX_READ_SDMA_MEMORY_UNC_ERR,
C_TX_SB_HDR_UNC_ERR,
C_TX_CREDIT_RETURN_PARITY_ERR,
C_TX_LAUNCH_FIFO8_UNC_OR_PARITY_ERR,
C_TX_LAUNCH_FIFO7_UNC_OR_PARITY_ERR,
C_TX_LAUNCH_FIFO6_UNC_OR_PARITY_ERR,
C_TX_LAUNCH_FIFO5_UNC_OR_PARITY_ERR,
C_TX_LAUNCH_FIFO4_UNC_OR_PARITY_ERR,
C_TX_LAUNCH_FIFO3_UNC_OR_PARITY_ERR,
C_TX_LAUNCH_FIFO2_UNC_OR_PARITY_ERR,
C_TX_LAUNCH_FIFO1_UNC_OR_PARITY_ERR,
C_TX_LAUNCH_FIFO0_UNC_OR_PARITY_ERR,
C_TX_SDMA15_DISALLOWED_PACKET_ERR,
C_TX_SDMA14_DISALLOWED_PACKET_ERR,
C_TX_SDMA13_DISALLOWED_PACKET_ERR,
C_TX_SDMA12_DISALLOWED_PACKET_ERR,
C_TX_SDMA11_DISALLOWED_PACKET_ERR,
C_TX_SDMA10_DISALLOWED_PACKET_ERR,
C_TX_SDMA9_DISALLOWED_PACKET_ERR,
C_TX_SDMA8_DISALLOWED_PACKET_ERR,
C_TX_SDMA7_DISALLOWED_PACKET_ERR,
C_TX_SDMA6_DISALLOWED_PACKET_ERR,
C_TX_SDMA5_DISALLOWED_PACKET_ERR,
C_TX_SDMA4_DISALLOWED_PACKET_ERR,
C_TX_SDMA3_DISALLOWED_PACKET_ERR,
C_TX_SDMA2_DISALLOWED_PACKET_ERR,
C_TX_SDMA1_DISALLOWED_PACKET_ERR,
C_TX_SDMA0_DISALLOWED_PACKET_ERR,
C_TX_CONFIG_PARITY_ERR,
C_TX_SBRD_CTL_CSR_PARITY_ERR,
C_TX_LAUNCH_CSR_PARITY_ERR,
C_TX_ILLEGAL_CL_ERR,
C_TX_SBRD_CTL_STATE_MACHINE_PARITY_ERR,
C_TX_RESERVED_10,
C_TX_RESERVED_9,
C_TX_SDMA_LAUNCH_INTF_PARITY_ERR,
C_TX_PIO_LAUNCH_INTF_PARITY_ERR,
C_TX_RESERVED_6,
C_TX_INCORRECT_LINK_STATE_ERR,
C_TX_LINK_DOWN_ERR,
C_TX_EGRESS_FIFO_UNDERRUN_OR_PARITY_ERR,
C_TX_RESERVED_2,
C_TX_PKT_INTEGRITY_MEM_UNC_ERR,
C_TX_PKT_INTEGRITY_MEM_COR_ERR,
/* SendErrStatus */
C_SEND_CSR_WRITE_BAD_ADDR_ERR,
C_SEND_CSR_READ_BAD_ADD_ERR,
C_SEND_CSR_PARITY_ERR,
/* SendCtxtErrStatus */
C_PIO_WRITE_OUT_OF_BOUNDS_ERR,
C_PIO_WRITE_OVERFLOW_ERR,
C_PIO_WRITE_CROSSES_BOUNDARY_ERR,
C_PIO_DISALLOWED_PACKET_ERR,
C_PIO_INCONSISTENT_SOP_ERR,
/*SendDmaEngErrStatus */
C_SDMA_HEADER_REQUEST_FIFO_COR_ERR,
C_SDMA_HEADER_STORAGE_COR_ERR,
C_SDMA_PACKET_TRACKING_COR_ERR,
C_SDMA_ASSEMBLY_COR_ERR,
C_SDMA_DESC_TABLE_COR_ERR,
C_SDMA_HEADER_REQUEST_FIFO_UNC_ERR,
C_SDMA_HEADER_STORAGE_UNC_ERR,
C_SDMA_PACKET_TRACKING_UNC_ERR,
C_SDMA_ASSEMBLY_UNC_ERR,
C_SDMA_DESC_TABLE_UNC_ERR,
C_SDMA_TIMEOUT_ERR,
C_SDMA_HEADER_LENGTH_ERR,
C_SDMA_HEADER_ADDRESS_ERR,
C_SDMA_HEADER_SELECT_ERR,
C_SMDA_RESERVED_9,
C_SDMA_PACKET_DESC_OVERFLOW_ERR,
C_SDMA_LENGTH_MISMATCH_ERR,
C_SDMA_HALT_ERR,
C_SDMA_MEM_READ_ERR,
C_SDMA_FIRST_DESC_ERR,
C_SDMA_TAIL_OUT_OF_BOUNDS_ERR,
C_SDMA_TOO_LONG_ERR,
C_SDMA_GEN_MISMATCH_ERR,
C_SDMA_WRONG_DW_ERR,
DEV_CNTR_LAST /* Must be kept last */
};
/* Per port counter indexes */
enum {
C_TX_UNSUP_VL = 0,
C_TX_INVAL_LEN,
C_TX_MM_LEN_ERR,
C_TX_UNDERRUN,
C_TX_FLOW_STALL,
C_TX_DROPPED,
C_TX_HDR_ERR,
C_TX_PKT,
C_TX_WORDS,
C_TX_WAIT,
C_TX_FLIT_VL,
C_TX_PKT_VL,
C_TX_WAIT_VL,
C_RX_PKT,
C_RX_WORDS,
C_SW_LINK_DOWN,
C_SW_LINK_UP,
C_SW_UNKNOWN_FRAME,
C_SW_XMIT_DSCD,
C_SW_XMIT_DSCD_VL,
C_SW_XMIT_CSTR_ERR,
C_SW_RCV_CSTR_ERR,
C_SW_IBP_LOOP_PKTS,
C_SW_IBP_RC_RESENDS,
C_SW_IBP_RNR_NAKS,
C_SW_IBP_OTHER_NAKS,
C_SW_IBP_RC_TIMEOUTS,
C_SW_IBP_PKT_DROPS,
C_SW_IBP_DMA_WAIT,
C_SW_IBP_RC_SEQNAK,
C_SW_IBP_RC_DUPREQ,
C_SW_IBP_RDMA_SEQ,
C_SW_IBP_UNALIGNED,
C_SW_IBP_SEQ_NAK,
C_SW_IBP_RC_CRWAITS,
C_SW_CPU_RC_ACKS,
C_SW_CPU_RC_QACKS,
C_SW_CPU_RC_DELAYED_COMP,
C_RCV_HDR_OVF_0,
C_RCV_HDR_OVF_1,
C_RCV_HDR_OVF_2,
C_RCV_HDR_OVF_3,
C_RCV_HDR_OVF_4,
C_RCV_HDR_OVF_5,
C_RCV_HDR_OVF_6,
C_RCV_HDR_OVF_7,
C_RCV_HDR_OVF_8,
C_RCV_HDR_OVF_9,
C_RCV_HDR_OVF_10,
C_RCV_HDR_OVF_11,
C_RCV_HDR_OVF_12,
C_RCV_HDR_OVF_13,
C_RCV_HDR_OVF_14,
C_RCV_HDR_OVF_15,
C_RCV_HDR_OVF_16,
C_RCV_HDR_OVF_17,
C_RCV_HDR_OVF_18,
C_RCV_HDR_OVF_19,
C_RCV_HDR_OVF_20,
C_RCV_HDR_OVF_21,
C_RCV_HDR_OVF_22,
C_RCV_HDR_OVF_23,
C_RCV_HDR_OVF_24,
C_RCV_HDR_OVF_25,
C_RCV_HDR_OVF_26,
C_RCV_HDR_OVF_27,
C_RCV_HDR_OVF_28,
C_RCV_HDR_OVF_29,
C_RCV_HDR_OVF_30,
C_RCV_HDR_OVF_31,
C_RCV_HDR_OVF_32,
C_RCV_HDR_OVF_33,
C_RCV_HDR_OVF_34,
C_RCV_HDR_OVF_35,
C_RCV_HDR_OVF_36,
C_RCV_HDR_OVF_37,
C_RCV_HDR_OVF_38,
C_RCV_HDR_OVF_39,
C_RCV_HDR_OVF_40,
C_RCV_HDR_OVF_41,
C_RCV_HDR_OVF_42,
C_RCV_HDR_OVF_43,
C_RCV_HDR_OVF_44,
C_RCV_HDR_OVF_45,
C_RCV_HDR_OVF_46,
C_RCV_HDR_OVF_47,
C_RCV_HDR_OVF_48,
C_RCV_HDR_OVF_49,
C_RCV_HDR_OVF_50,
C_RCV_HDR_OVF_51,
C_RCV_HDR_OVF_52,
C_RCV_HDR_OVF_53,
C_RCV_HDR_OVF_54,
C_RCV_HDR_OVF_55,
C_RCV_HDR_OVF_56,
C_RCV_HDR_OVF_57,
C_RCV_HDR_OVF_58,
C_RCV_HDR_OVF_59,
C_RCV_HDR_OVF_60,
C_RCV_HDR_OVF_61,
C_RCV_HDR_OVF_62,
C_RCV_HDR_OVF_63,
C_RCV_HDR_OVF_64,
C_RCV_HDR_OVF_65,
C_RCV_HDR_OVF_66,
C_RCV_HDR_OVF_67,
C_RCV_HDR_OVF_68,
C_RCV_HDR_OVF_69,
C_RCV_HDR_OVF_70,
C_RCV_HDR_OVF_71,
C_RCV_HDR_OVF_72,
C_RCV_HDR_OVF_73,
C_RCV_HDR_OVF_74,
C_RCV_HDR_OVF_75,
C_RCV_HDR_OVF_76,
C_RCV_HDR_OVF_77,
C_RCV_HDR_OVF_78,
C_RCV_HDR_OVF_79,
C_RCV_HDR_OVF_80,
C_RCV_HDR_OVF_81,
C_RCV_HDR_OVF_82,
C_RCV_HDR_OVF_83,
C_RCV_HDR_OVF_84,
C_RCV_HDR_OVF_85,
C_RCV_HDR_OVF_86,
C_RCV_HDR_OVF_87,
C_RCV_HDR_OVF_88,
C_RCV_HDR_OVF_89,
C_RCV_HDR_OVF_90,
C_RCV_HDR_OVF_91,
C_RCV_HDR_OVF_92,
C_RCV_HDR_OVF_93,
C_RCV_HDR_OVF_94,
C_RCV_HDR_OVF_95,
C_RCV_HDR_OVF_96,
C_RCV_HDR_OVF_97,
C_RCV_HDR_OVF_98,
C_RCV_HDR_OVF_99,
C_RCV_HDR_OVF_100,
C_RCV_HDR_OVF_101,
C_RCV_HDR_OVF_102,
C_RCV_HDR_OVF_103,
C_RCV_HDR_OVF_104,
C_RCV_HDR_OVF_105,
C_RCV_HDR_OVF_106,
C_RCV_HDR_OVF_107,
C_RCV_HDR_OVF_108,
C_RCV_HDR_OVF_109,
C_RCV_HDR_OVF_110,
C_RCV_HDR_OVF_111,
C_RCV_HDR_OVF_112,
C_RCV_HDR_OVF_113,
C_RCV_HDR_OVF_114,
C_RCV_HDR_OVF_115,
C_RCV_HDR_OVF_116,
C_RCV_HDR_OVF_117,
C_RCV_HDR_OVF_118,
C_RCV_HDR_OVF_119,
C_RCV_HDR_OVF_120,
C_RCV_HDR_OVF_121,
C_RCV_HDR_OVF_122,
C_RCV_HDR_OVF_123,
C_RCV_HDR_OVF_124,
C_RCV_HDR_OVF_125,
C_RCV_HDR_OVF_126,
C_RCV_HDR_OVF_127,
C_RCV_HDR_OVF_128,
C_RCV_HDR_OVF_129,
C_RCV_HDR_OVF_130,
C_RCV_HDR_OVF_131,
C_RCV_HDR_OVF_132,
C_RCV_HDR_OVF_133,
C_RCV_HDR_OVF_134,
C_RCV_HDR_OVF_135,
C_RCV_HDR_OVF_136,
C_RCV_HDR_OVF_137,
C_RCV_HDR_OVF_138,
C_RCV_HDR_OVF_139,
C_RCV_HDR_OVF_140,
C_RCV_HDR_OVF_141,
C_RCV_HDR_OVF_142,
C_RCV_HDR_OVF_143,
C_RCV_HDR_OVF_144,
C_RCV_HDR_OVF_145,
C_RCV_HDR_OVF_146,
C_RCV_HDR_OVF_147,
C_RCV_HDR_OVF_148,
C_RCV_HDR_OVF_149,
C_RCV_HDR_OVF_150,
C_RCV_HDR_OVF_151,
C_RCV_HDR_OVF_152,
C_RCV_HDR_OVF_153,
C_RCV_HDR_OVF_154,
C_RCV_HDR_OVF_155,
C_RCV_HDR_OVF_156,
C_RCV_HDR_OVF_157,
C_RCV_HDR_OVF_158,
C_RCV_HDR_OVF_159,
PORT_CNTR_LAST /* Must be kept last */
};
u64 get_all_cpu_total(u64 __percpu *cntr);
void hfi1_start_cleanup(struct hfi1_devdata *dd);
void hfi1_clear_tids(struct hfi1_ctxtdata *rcd);
void hfi1_init_ctxt(struct send_context *sc);
void hfi1_put_tid(struct hfi1_devdata *dd, u32 index,
u32 type, unsigned long pa, u16 order);
void hfi1_quiet_serdes(struct hfi1_pportdata *ppd);
void hfi1_rcvctrl(struct hfi1_devdata *dd, unsigned int op,
struct hfi1_ctxtdata *rcd);
u32 hfi1_read_cntrs(struct hfi1_devdata *dd, char **namep, u64 **cntrp);
u32 hfi1_read_portcntrs(struct hfi1_pportdata *ppd, char **namep, u64 **cntrp);
int hfi1_get_ib_cfg(struct hfi1_pportdata *ppd, int which);
int hfi1_set_ib_cfg(struct hfi1_pportdata *ppd, int which, u32 val);
int hfi1_set_ctxt_jkey(struct hfi1_devdata *dd, struct hfi1_ctxtdata *rcd,
u16 jkey);
int hfi1_clear_ctxt_jkey(struct hfi1_devdata *dd, struct hfi1_ctxtdata *ctxt);
int hfi1_set_ctxt_pkey(struct hfi1_devdata *dd, struct hfi1_ctxtdata *ctxt,
u16 pkey);
int hfi1_clear_ctxt_pkey(struct hfi1_devdata *dd, struct hfi1_ctxtdata *ctxt);
void hfi1_read_link_quality(struct hfi1_devdata *dd, u8 *link_quality);
void hfi1_init_vnic_rsm(struct hfi1_devdata *dd);
void hfi1_deinit_vnic_rsm(struct hfi1_devdata *dd);
irqreturn_t general_interrupt(int irq, void *data);
irqreturn_t sdma_interrupt(int irq, void *data);
irqreturn_t receive_context_interrupt(int irq, void *data);
irqreturn_t receive_context_thread(int irq, void *data);
irqreturn_t receive_context_interrupt_napi(int irq, void *data);
int set_intr_bits(struct hfi1_devdata *dd, u16 first, u16 last, bool set);
void init_qsfp_int(struct hfi1_devdata *dd);
void clear_all_interrupts(struct hfi1_devdata *dd);
void remap_intr(struct hfi1_devdata *dd, int isrc, int msix_intr);
void remap_sdma_interrupts(struct hfi1_devdata *dd, int engine, int msix_intr);
void reset_interrupts(struct hfi1_devdata *dd);
IB/hfi1: Correct tid qp rcd to match verbs context The qp priv rcd pointer doesn't match the context being used for verbs causing issues when 9B and kdeth packets are processed by different receive contexts and hence different CPUs. When running on different CPUs the following panic can occur: WARNING: CPU: 3 PID: 2584 at lib/list_debug.c:59 __list_del_entry+0xa1/0xd0 list_del corruption. prev->next should be ffff9a7ac31f7a30, but was ffff9a7c3bc89230 CPU: 3 PID: 2584 Comm: z_wr_iss Kdump: loaded Tainted: P OE ------------ 3.10.0-862.2.3.el7_lustre.x86_64 #1 Call Trace: <IRQ> [<ffffffffb7b0d78e>] dump_stack+0x19/0x1b [<ffffffffb74916d8>] __warn+0xd8/0x100 [<ffffffffb749175f>] warn_slowpath_fmt+0x5f/0x80 [<ffffffffb7768671>] __list_del_entry+0xa1/0xd0 [<ffffffffc0c7a945>] process_rcv_qp_work+0xb5/0x160 [hfi1] [<ffffffffc0c7bc2b>] handle_receive_interrupt_nodma_rtail+0x20b/0x2b0 [hfi1] [<ffffffffc0c70683>] receive_context_interrupt+0x23/0x40 [hfi1] [<ffffffffb7540a94>] __handle_irq_event_percpu+0x44/0x1c0 [<ffffffffb7540c42>] handle_irq_event_percpu+0x32/0x80 [<ffffffffb7540ccc>] handle_irq_event+0x3c/0x60 [<ffffffffb7543a1f>] handle_edge_irq+0x7f/0x150 [<ffffffffb742d504>] handle_irq+0xe4/0x1a0 [<ffffffffb7b23f7d>] do_IRQ+0x4d/0xf0 [<ffffffffb7b16362>] common_interrupt+0x162/0x162 <EOI> [<ffffffffb775a326>] ? memcpy+0x6/0x110 [<ffffffffc109210d>] ? abd_copy_from_buf_off_cb+0x1d/0x30 [zfs] [<ffffffffc10920f0>] ? abd_copy_to_buf_off_cb+0x30/0x30 [zfs] [<ffffffffc1093257>] abd_iterate_func+0x97/0x120 [zfs] [<ffffffffc10934d9>] abd_copy_from_buf_off+0x39/0x60 [zfs] [<ffffffffc109b828>] arc_write_ready+0x178/0x300 [zfs] [<ffffffffb7b11032>] ? mutex_lock+0x12/0x2f [<ffffffffb7b11032>] ? mutex_lock+0x12/0x2f [<ffffffffc1164d05>] zio_ready+0x65/0x3d0 [zfs] [<ffffffffc04d725e>] ? tsd_get_by_thread+0x2e/0x50 [spl] [<ffffffffc04d1318>] ? taskq_member+0x18/0x30 [spl] [<ffffffffc115ef22>] zio_execute+0xa2/0x100 [zfs] [<ffffffffc04d1d2c>] taskq_thread+0x2ac/0x4f0 [spl] [<ffffffffb74cee80>] ? wake_up_state+0x20/0x20 [<ffffffffc115ee80>] ? zio_taskq_member.isra.7.constprop.10+0x80/0x80 [zfs] [<ffffffffc04d1a80>] ? taskq_thread_spawn+0x60/0x60 [spl] [<ffffffffb74bae31>] kthread+0xd1/0xe0 [<ffffffffb74bad60>] ? insert_kthread_work+0x40/0x40 [<ffffffffb7b1f5f7>] ret_from_fork_nospec_begin+0x21/0x21 [<ffffffffb74bad60>] ? insert_kthread_work+0x40/0x40 Fix by reading the map entry in the same manner as the hardware so that the kdeth and verbs contexts match. Cc: <stable@vger.kernel.org> Fixes: 5190f052a365 ("IB/hfi1: Allow the driver to initialize QP priv struct") Reviewed-by: Kaike Wan <kaike.wan@intel.com> Signed-off-by: Mike Marciniszyn <mike.marciniszyn@intel.com> Signed-off-by: Dennis Dalessandro <dennis.dalessandro@intel.com> Signed-off-by: Jason Gunthorpe <jgg@mellanox.com>
2019-06-10 16:28:18 +00:00
u8 hfi1_get_qp_map(struct hfi1_devdata *dd, u8 idx);
void hfi1_init_aip_rsm(struct hfi1_devdata *dd);
void hfi1_deinit_aip_rsm(struct hfi1_devdata *dd);
/*
* Interrupt source table.
*
* Each entry is an interrupt source "type". It is ordered by increasing
* number.
*/
struct is_table {
int start; /* interrupt source type start */
int end; /* interrupt source type end */
/* routine that returns the name of the interrupt source */
char *(*is_name)(char *name, size_t size, unsigned int source);
/* routine to call when receiving an interrupt */
void (*is_int)(struct hfi1_devdata *dd, unsigned int source);
};
#endif /* _CHIP_H */