Current Path : /sys/contrib/octeon-sdk/ |
FreeBSD hs32.drive.ne.jp 9.1-RELEASE FreeBSD 9.1-RELEASE #1: Wed Jan 14 12:18:08 JST 2015 root@hs32.drive.ne.jp:/sys/amd64/compile/hs32 amd64 |
Current File : //sys/contrib/octeon-sdk/cvmx-error-custom.c |
/***********************license start*************** * Copyright (c) 2003-2010 Cavium Networks (support@cavium.com). All rights * reserved. * * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are * met: * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * * Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following * disclaimer in the documentation and/or other materials provided * with the distribution. * * Neither the name of Cavium Networks nor the names of * its contributors may be used to endorse or promote products * derived from this software without specific prior written * permission. * This Software, including technical data, may be subject to U.S. export control * laws, including the U.S. Export Administration Act and its associated * regulations, and may be subject to export or import regulations in other * countries. * TO THE MAXIMUM EXTENT PERMITTED BY LAW, THE SOFTWARE IS PROVIDED "AS IS" * AND WITH ALL FAULTS AND CAVIUM NETWORKS MAKES NO PROMISES, REPRESENTATIONS OR * WARRANTIES, EITHER EXPRESS, IMPLIED, STATUTORY, OR OTHERWISE, WITH RESPECT TO * THE SOFTWARE, INCLUDING ITS CONDITION, ITS CONFORMITY TO ANY REPRESENTATION OR * DESCRIPTION, OR THE EXISTENCE OF ANY LATENT OR PATENT DEFECTS, AND CAVIUM * SPECIFICALLY DISCLAIMS ALL IMPLIED (IF ANY) WARRANTIES OF TITLE, * MERCHANTABILITY, NONINFRINGEMENT, FITNESS FOR A PARTICULAR PURPOSE, LACK OF * VIRUSES, ACCURACY OR COMPLETENESS, QUIET ENJOYMENT, QUIET POSSESSION OR * CORRESPONDENCE TO DESCRIPTION. THE ENTIRE RISK ARISING OUT OF USE OR * PERFORMANCE OF THE SOFTWARE LIES WITH YOU. ***********************license end**************************************/ /** * @file * * Prototypes for custom error handler function not handled by the default * message display error function. * * <hr>$Revision: 44252 $<hr> */ #ifdef CVMX_BUILD_FOR_LINUX_KERNEL #include <asm/octeon/cvmx.h> #include <asm/octeon/cvmx-error.h> #include <asm/octeon/cvmx-error-custom.h> #include <asm/octeon/cvmx-helper.h> #include <asm/octeon/cvmx-l2c.h> #include <asm/octeon/cvmx-pcie.h> #include <asm/octeon/cvmx-pexp-defs.h> #include <asm/octeon/cvmx-dfa-defs.h> #include <asm/octeon/cvmx-gmxx-defs.h> #include <asm/octeon/cvmx-lmcx-defs.h> #include <asm/octeon/cvmx-pemx-defs.h> #define PRINT_ERROR(format, ...) cvmx_safe_printf("ERROR " format, ##__VA_ARGS__) #else #include "cvmx.h" #include "cvmx-error.h" #include "cvmx-error-custom.h" #include "cvmx-helper.h" #include "cvmx-l2c.h" #include "cvmx-pcie.h" #include "cvmx-interrupt.h" #endif /** * @INTERNAL * XAUI interfaces need to be reset whenever a local or remote fault * is detected. Calling autoconf takes the link through a reset. * * @param info * * @return */ static int __cvmx_error_handle_gmxx_rxx_int_reg(const struct cvmx_error_info *info) { #ifdef CVMX_ENABLE_PKO_FUNCTIONS int ipd_port = info->group_index; cvmx_helper_link_autoconf(ipd_port); #endif cvmx_write_csr(info->status_addr, info->status_mask); return 1; } /** * @INTERNAL * When NPEI_INT_SUM[C0_LDWN] is set, the PCIe block requires a shutdown and * initialization to bring the link back up. This handler does this for port 0. * Note that config space is not enumerated again, so the devices will still be * unusable. * * @param info * * @return */ static int __cvmx_error_handle_npei_int_sum_c0_ldwn(const struct cvmx_error_info *info) { cvmx_ciu_soft_prst_t ciu_soft_prst; PRINT_ERROR("NPEI_INT_SUM[C0_LDWN]: Reset request due to link0 down status.\n"); ciu_soft_prst.u64 = cvmx_read_csr(CVMX_CIU_SOFT_PRST); if (!ciu_soft_prst.s.soft_prst) { /* Attempt to automatically bring the link back up */ cvmx_pcie_rc_shutdown(0); cvmx_pcie_rc_initialize(0); } cvmx_write_csr(CVMX_PEXP_NPEI_INT_SUM, cvmx_read_csr(CVMX_PEXP_NPEI_INT_SUM)); return 1; } /** * @INTERNAL * When NPEI_INT_SUM[C1_LDWN] is set, the PCIe block requires a shutdown and * initialization to bring the link back up. This handler does this for port 1. * Note that config space is not enumerated again, so the devices will still be * unusable. * * @param info * * @return */ static int __cvmx_error_handle_npei_int_sum_c1_ldwn(const struct cvmx_error_info *info) { cvmx_ciu_soft_prst_t ciu_soft_prst; PRINT_ERROR("NPEI_INT_SUM[C1_LDWN]: Reset request due to link1 down status.\n"); ciu_soft_prst.u64 = cvmx_read_csr(CVMX_CIU_SOFT_PRST1); if (!ciu_soft_prst.s.soft_prst) { /* Attempt to automatically bring the link back up */ cvmx_pcie_rc_shutdown(1); cvmx_pcie_rc_initialize(1); } cvmx_write_csr(CVMX_PEXP_NPEI_INT_SUM, cvmx_read_csr(CVMX_PEXP_NPEI_INT_SUM)); return 1; } /** * @INTERNAL * Some errors require more complicated error handing functions than the * automatically generated functions in cvmx-error-init-*.c. This function * replaces these handers with hand coded functions for these special cases. * * @return Zero on success, negative on failure. */ int __cvmx_error_custom_initialize(void) { if (OCTEON_IS_MODEL(OCTEON_CN52XX) || OCTEON_IS_MODEL(OCTEON_CN56XX) || OCTEON_IS_MODEL(OCTEON_CN63XX)) { cvmx_error_change_handler(CVMX_ERROR_REGISTER_IO64, CVMX_GMXX_RXX_INT_REG(0,0), 1ull<<21 /* rem_fault */, __cvmx_error_handle_gmxx_rxx_int_reg, 0, NULL, NULL); cvmx_error_change_handler(CVMX_ERROR_REGISTER_IO64, CVMX_GMXX_RXX_INT_REG(0,0), 1ull<<20 /* loc_fault */, __cvmx_error_handle_gmxx_rxx_int_reg, 0, NULL, NULL); } if (OCTEON_IS_MODEL(OCTEON_CN56XX)) { cvmx_error_change_handler(CVMX_ERROR_REGISTER_IO64, CVMX_GMXX_RXX_INT_REG(0,1), 1ull<<21 /* rem_fault */, __cvmx_error_handle_gmxx_rxx_int_reg, 0, NULL, NULL); cvmx_error_change_handler(CVMX_ERROR_REGISTER_IO64, CVMX_GMXX_RXX_INT_REG(0,1), 1ull<<20 /* loc_fault */, __cvmx_error_handle_gmxx_rxx_int_reg, 0, NULL, NULL); } if (octeon_has_feature(OCTEON_FEATURE_NPEI)) { cvmx_error_change_handler(CVMX_ERROR_REGISTER_IO64, CVMX_PEXP_NPEI_INT_SUM, 1ull<<59 /* c0_ldwn */, __cvmx_error_handle_npei_int_sum_c0_ldwn, 0, NULL, NULL); cvmx_error_change_handler(CVMX_ERROR_REGISTER_IO64, CVMX_PEXP_NPEI_INT_SUM, 1ull<<60 /* c1_ldwn */, __cvmx_error_handle_npei_int_sum_c1_ldwn, 0, NULL, NULL); } /* CN63XX pass 1.x has a bug where the PCIe config CRS counter does not stop. Disable reporting errors from CRS */ if (OCTEON_IS_MODEL(OCTEON_CN63XX_PASS1_X)) { cvmx_error_disable(CVMX_ERROR_REGISTER_IO64, CVMX_PEMX_INT_SUM(0), 1ull<<12); cvmx_error_disable(CVMX_ERROR_REGISTER_IO64, CVMX_PEMX_INT_SUM(0), 1ull<<13); cvmx_error_disable(CVMX_ERROR_REGISTER_IO64, CVMX_PEMX_INT_SUM(1), 1ull<<12); cvmx_error_disable(CVMX_ERROR_REGISTER_IO64, CVMX_PEMX_INT_SUM(1), 1ull<<13); } return 0; } /** * @INTERNAL * DFA_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_dfa_err_cp2dbe(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_DFA_ERR, cvmx_read_csr(CVMX_DFA_ERR)); PRINT_ERROR("DFA_ERR[CP2DBE]: DFA PP-CP2 Double Bit Error Detected\n"); return 1; } /** * @INTERNAL * DFA_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_dfa_err_cp2perr(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_DFA_ERR, cvmx_read_csr(CVMX_DFA_ERR)); PRINT_ERROR("DFA_ERR[CP2PERR]: PP-CP2 Parity Error Detected\n"); return 1; } /** * @INTERNAL * DFA_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_dfa_err_cp2sbe(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_DFA_ERR, cvmx_read_csr(CVMX_DFA_ERR)); PRINT_ERROR("DFA_ERR[CP2SBE]: DFA PP-CP2 Single Bit Error Corrected\n"); return 1; } /** * @INTERNAL * DFA_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_dfa_err_dblovf(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_DFA_ERR, cvmx_read_csr(CVMX_DFA_ERR)); PRINT_ERROR("DFA_ERR[DBLOVF]: Doorbell Overflow detected\n"); return 1; } /** * @INTERNAL * DFA_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_dfa_err_dtedbe(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_DFA_ERR, cvmx_read_csr(CVMX_DFA_ERR)); PRINT_ERROR("DFA_ERR[DTEDBE]: DFA DTE 29b Double Bit Error Detected\n"); return 1; } /** * @INTERNAL * DFA_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_dfa_err_dteperr(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_DFA_ERR, cvmx_read_csr(CVMX_DFA_ERR)); PRINT_ERROR("DFA_ERR[DTEPERR]: DTE Parity Error Detected\n"); return 1; } /** * @INTERNAL * DFA_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_dfa_err_dtesbe(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_DFA_ERR, cvmx_read_csr(CVMX_DFA_ERR)); PRINT_ERROR("DFA_ERR[DTESBE]: DFA DTE 29b Single Bit Error Corrected\n"); return 1; } /** * @INTERNAL * L2D_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_l2d_err_ded_err(const struct cvmx_error_info *info) { cvmx_l2d_err_t derr; cvmx_l2d_fadr_t fadr; uint64_t syn0 = cvmx_read_csr(CVMX_L2D_FSYN0); uint64_t syn1 = cvmx_read_csr(CVMX_L2D_FSYN1); derr.u64 = cvmx_read_csr(CVMX_L2D_ERR); fadr.u64 = cvmx_read_csr(CVMX_L2D_FADR); PRINT_ERROR("L2D_ERR[DED_ERR] ECC double: fadr: 0x%llx, syn0:0x%llx, syn1: 0x%llx\n", (unsigned long long)fadr.u64, (unsigned long long)syn0, (unsigned long long)syn1); /* Flush the line that had the error */ cvmx_l2c_flush_line(fadr.s.fset, fadr.s.fadr >> 1); cvmx_write_csr(CVMX_L2D_ERR, derr.u64); return 1; } /** * @INTERNAL * L2D_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_l2d_err_sec_err(const struct cvmx_error_info *info) { cvmx_l2d_err_t derr; cvmx_l2d_fadr_t fadr; uint64_t syn0 = cvmx_read_csr(CVMX_L2D_FSYN0); uint64_t syn1 = cvmx_read_csr(CVMX_L2D_FSYN1); derr.u64 = cvmx_read_csr(CVMX_L2D_ERR); fadr.u64 = cvmx_read_csr(CVMX_L2D_FADR); PRINT_ERROR("L2D_ERR[SEC_ERR] ECC single: fadr: 0x%llx, syn0:0x%llx, syn1: 0x%llx\n", (unsigned long long)fadr.u64, (unsigned long long)syn0, (unsigned long long)syn1); /* Flush the line that had the error */ cvmx_l2c_flush_line(fadr.s.fset, fadr.s.fadr >> 1); cvmx_write_csr(CVMX_L2D_ERR, derr.u64); return 1; } /** * @INTERNAL * L2T_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_l2t_err_ded_err(const struct cvmx_error_info *info) { cvmx_l2t_err_t terr; terr.u64 = cvmx_read_csr(CVMX_L2T_ERR); cvmx_write_csr(CVMX_L2T_ERR, terr.u64); PRINT_ERROR("L2T_ERR[DED_ERR]: double bit:\tfadr: 0x%x, fset: 0x%x, fsyn: 0x%x\n", terr.s.fadr, terr.s.fset, terr.s.fsyn); if (!terr.s.fsyn) { /* Syndrome is zero, which means error was in non-hit line, so flush all associations */ int i; int l2_assoc = cvmx_l2c_get_num_assoc(); for (i = 0; i < l2_assoc; i++) cvmx_l2c_flush_line(i, terr.s.fadr); } else cvmx_l2c_flush_line(terr.s.fset, terr.s.fadr); return 1; } /** * @INTERNAL * L2T_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_l2t_err_lckerr2(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_L2T_ERR, cvmx_read_csr(CVMX_L2T_ERR)); PRINT_ERROR("L2T_ERR[LCKERR2]: HW detected a case where a Rd/Wr Miss from PP#n could not find an available/unlocked set (for replacement).\n"); return 1; } /** * @INTERNAL * L2T_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_l2t_err_lckerr(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_L2T_ERR, cvmx_read_csr(CVMX_L2T_ERR)); PRINT_ERROR("L2T_ERR[LCKERR]: SW attempted to LOCK DOWN the last available set of the INDEX (which is ignored by HW - but reported to SW).\n"); return 1; } /** * @INTERNAL * L2T_ERR contains R/W1C bits along with R/W bits. This means that it requires * special handling instead of the normal __cvmx_error_display() function. * * @param info * * @return */ int __cvmx_error_handle_l2t_err_sec_err(const struct cvmx_error_info *info) { cvmx_l2t_err_t terr; terr.u64 = cvmx_read_csr(CVMX_L2T_ERR); cvmx_write_csr(CVMX_L2T_ERR, terr.u64); PRINT_ERROR("L2T_ERR[SEC_ERR]: single bit:\tfadr: 0x%x, fset: 0x%x, fsyn: 0x%x\n", terr.s.fadr, terr.s.fset, terr.s.fsyn); if (!terr.s.fsyn) { /* Syndrome is zero, which means error was in non-hit line, so flush all associations */ int i; int l2_assoc = cvmx_l2c_get_num_assoc(); for (i = 0; i < l2_assoc; i++) cvmx_l2c_flush_line(i, terr.s.fadr); } else cvmx_l2c_flush_line(terr.s.fset, terr.s.fadr); return 1; } /** * @INTERNAL * LMCX_MEM_CFG0 contains R/W1C bits along with R/W bits. This means that it * requires special handling instead of the normal __cvmx_error_display() * function. * * @param info * * @return */ static int __cvmx_error_handle_lmcx_mem_cfg0(const struct cvmx_error_info *info) { int ddr_controller = info->group_index; cvmx_lmcx_mem_cfg0_t mem_cfg0; cvmx_lmcx_fadr_t fadr; int sec_err; int ded_err; mem_cfg0.u64 = cvmx_read_csr(CVMX_LMCX_MEM_CFG0(ddr_controller)); fadr.u64 = cvmx_read_csr(CVMX_LMCX_FADR(ddr_controller)); cvmx_write_csr(CVMX_LMCX_MEM_CFG0(ddr_controller),mem_cfg0.u64); sec_err = cvmx_dpop(mem_cfg0.s.sec_err); ded_err = cvmx_dpop(mem_cfg0.s.ded_err); if (ded_err || sec_err) { PRINT_ERROR("DDR%d ECC: %d Single bit corrections, %d Double bit errors\n" "DDR%d ECC:\tFailing dimm: %u\n" "DDR%d ECC:\tFailing rank: %u\n" "DDR%d ECC:\tFailing bank: %u\n" "DDR%d ECC:\tFailing row: 0x%x\n" "DDR%d ECC:\tFailing column: 0x%x\n", ddr_controller, sec_err, ded_err, ddr_controller, fadr.cn38xx.fdimm, ddr_controller, fadr.cn38xx.fbunk, ddr_controller, fadr.cn38xx.fbank, ddr_controller, fadr.cn38xx.frow, ddr_controller, fadr.cn38xx.fcol); } return 1; } /** * @INTERNAL * LMCX_MEM_CFG0 contains R/W1C bits along with R/W bits. This means that it * requires special handling instead of the normal __cvmx_error_display() * function. * * @param info * * @return */ int __cvmx_error_handle_lmcx_mem_cfg0_ded_err(const struct cvmx_error_info *info) { return __cvmx_error_handle_lmcx_mem_cfg0(info); } /** * @INTERNAL * LMCX_MEM_CFG0 contains R/W1C bits along with R/W bits. This means that it * requires special handling instead of the normal __cvmx_error_display() * function. * * @param info * * @return */ int __cvmx_error_handle_lmcx_mem_cfg0_sec_err(const struct cvmx_error_info *info) { return __cvmx_error_handle_lmcx_mem_cfg0(info); } /** * @INTERNAL * POW_ECC_ERR contains R/W1C bits along with R/W bits. This means that it * requires special handling instead of the normal __cvmx_error_display() * function. * * @param info * * @return */ int __cvmx_error_handle_pow_ecc_err_dbe(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_POW_ECC_ERR, cvmx_read_csr(CVMX_POW_ECC_ERR)); PRINT_ERROR("POW_ECC_ERR[DBE]: POW double bit error\n"); return 1; } /** * @INTERNAL * POW_ECC_ERR contains R/W1C bits along with R/W bits. This means that it * requires special handling instead of the normal __cvmx_error_display() * function. * * @param info * * @return */ int __cvmx_error_handle_pow_ecc_err_iop(const struct cvmx_error_info *info) { cvmx_pow_ecc_err_t err; err.u64 = cvmx_read_csr(CVMX_POW_ECC_ERR); cvmx_write_csr(CVMX_POW_ECC_ERR, err.u64); if (err.s.iop & (1 << 0)) PRINT_ERROR("POW_ECC_ERR[IOP0]: Received SWTAG/SWTAG_FULL/SWTAG_DESCH/DESCH/UPD_WQP from PP in NULL_NULL state\n"); if (err.s.iop & (1 << 1)) PRINT_ERROR("POW_ECC_ERR[IOP1]: Received SWTAG/SWTAG_DESCH/DESCH/UPD_WQP from PP in NULL state\n"); if (err.s.iop & (1 << 2)) PRINT_ERROR("POW_ECC_ERR[IOP2]: Received SWTAG/SWTAG_FULL/SWTAG_DESCH/GET_WORK from PP with pending tag switch to ORDERED or ATOMIC\n"); if (err.s.iop & (1 << 3)) PRINT_ERROR("POW_ECC_ERR[IOP3]: Received SWTAG/SWTAG_FULL/SWTAG_DESCH from PP with tag specified as NULL_NULL\n"); if (err.s.iop & (1 << 4)) PRINT_ERROR("POW_ECC_ERR[IOP4]: Received SWTAG_FULL/SWTAG_DESCH from PP with tag specified as NULL\n"); if (err.s.iop & (1 << 5)) PRINT_ERROR("POW_ECC_ERR[IOP5]: Received SWTAG/SWTAG_FULL/SWTAG_DESCH/DESCH/UPD_WQP/GET_WORK/NULL_RD from PP with GET_WORK pending\n"); if (err.s.iop & (1 << 6)) PRINT_ERROR("POW_ECC_ERR[IOP6]: Received SWTAG/SWTAG_FULL/SWTAG_DESCH/DESCH/UPD_WQP/GET_WORK/NULL_RD from PP with NULL_RD pending\n"); if (err.s.iop & (1 << 7)) PRINT_ERROR("POW_ECC_ERR[IOP7]: Received CLR_NSCHED from PP with SWTAG_DESCH/DESCH/CLR_NSCHED pending\n"); if (err.s.iop & (1 << 8)) PRINT_ERROR("POW_ECC_ERR[IOP8]: Received SWTAG/SWTAG_FULL/SWTAG_DESCH/DESCH/UPD_WQP/GET_WORK/NULL_RD from PP with CLR_NSCHED pending\n"); if (err.s.iop & (1 << 9)) PRINT_ERROR("POW_ECC_ERR[IOP9]: Received illegal opcode\n"); if (err.s.iop & (1 << 10)) PRINT_ERROR("POW_ECC_ERR[IOP10]: Received ADD_WORK with tag specified as NULL_NULL\n"); if (err.s.iop & (1 << 11)) PRINT_ERROR("POW_ECC_ERR[IOP11]: Received DBG load from PP with DBG load pending\n"); if (err.s.iop & (1 << 12)) PRINT_ERROR("POW_ECC_ERR[IOP12]: Received CSR load from PP with CSR load pending\n"); return 1; } /** * @INTERNAL * POW_ECC_ERR contains R/W1C bits along with R/W bits. This means that it * requires special handling instead of the normal __cvmx_error_display() * function. * * @param info * * @return */ int __cvmx_error_handle_pow_ecc_err_rpe(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_POW_ECC_ERR, cvmx_read_csr(CVMX_POW_ECC_ERR)); PRINT_ERROR("POW_ECC_ERR[RPE]: Remote pointer error\n"); return 1; } /** * @INTERNAL * POW_ECC_ERR contains R/W1C bits along with R/W bits. This means that it * requires special handling instead of the normal __cvmx_error_display() * function. * * @param info * * @return */ int __cvmx_error_handle_pow_ecc_err_sbe(const struct cvmx_error_info *info) { cvmx_write_csr(CVMX_POW_ECC_ERR, cvmx_read_csr(CVMX_POW_ECC_ERR)); PRINT_ERROR("POW_ECC_ERR[SBE]: POW single bit error\n"); return 1; }