Searched refs:err_cqe (Results 1 – 11 of 11) sorted by relevance
/linux/drivers/infiniband/hw/cxgb4/ |
A D | ev.c | 69 CQE_QPID(err_cqe), CQE_OPCODE(err_cqe), in dump_err_cqe() 70 CQE_STATUS(err_cqe), CQE_TYPE(err_cqe), ntohl(err_cqe->len), in dump_err_cqe() 71 CQE_WRID_HI(err_cqe), CQE_WRID_LOW(err_cqe)); in dump_err_cqe() 82 if (RQ_TYPE(err_cqe) && (CQE_OPCODE(err_cqe) == FW_RI_RDMA_WRITE || in dump_err_cqe() 130 CQE_QPID(err_cqe), in c4iw_ev_dispatch() 131 CQE_OPCODE(err_cqe), CQE_STATUS(err_cqe), in c4iw_ev_dispatch() 132 CQE_TYPE(err_cqe), CQE_WRID_HI(err_cqe), in c4iw_ev_dispatch() 138 if (SQ_TYPE(err_cqe)) in c4iw_ev_dispatch() 146 CQE_OPCODE(err_cqe), CQE_STATUS(err_cqe), in c4iw_ev_dispatch() 147 CQE_TYPE(err_cqe), CQE_WRID_HI(err_cqe), in c4iw_ev_dispatch() [all …]
|
A D | qp.c | 1424 static inline void build_term_codes(struct t4_cqe *err_cqe, u8 *layer_type, in build_term_codes() argument 1433 if (!err_cqe) { in build_term_codes() 1439 status = CQE_STATUS(err_cqe); in build_term_codes() 1440 opcode = CQE_OPCODE(err_cqe); in build_term_codes() 1441 rqtype = RQ_TYPE(err_cqe); in build_term_codes() 1560 static void post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe, in post_terminate() argument 1589 build_term_codes(err_cqe, &term->layer_etype, &term->ecode); in post_terminate()
|
A D | iw_cxgb4.h | 1013 int c4iw_post_terminate(struct c4iw_qp *qhp, struct t4_cqe *err_cqe); 1020 void c4iw_ev_dispatch(struct c4iw_dev *dev, struct t4_cqe *err_cqe);
|
/linux/include/linux/mlx5/ |
A D | cq.h | 197 struct mlx5_err_cqe *err_cqe) in mlx5_dump_err_cqe() argument 199 print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, err_cqe, in mlx5_dump_err_cqe() 200 sizeof(*err_cqe), false); in mlx5_dump_err_cqe()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
A D | txrx.h | 322 struct mlx5_err_cqe *err_cqe) in mlx5e_dump_error_cqe() argument 332 get_cqe_opcode((struct mlx5_cqe64 *)err_cqe), in mlx5e_dump_error_cqe() 333 err_cqe->syndrome, err_cqe->vendor_err_synd); in mlx5e_dump_error_cqe() 334 mlx5_dump_err_cqe(cq->mdev, err_cqe); in mlx5e_dump_error_cqe()
|
A D | ptp.c | 91 ptpsq->cq_stats->err_cqe++; in mlx5e_ptp_handle_ts_cqe()
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
A D | en_tx.c | 396 static void mlx4_en_handle_err_cqe(struct mlx4_en_priv *priv, struct mlx4_err_cqe *err_cqe, in mlx4_en_handle_err_cqe() argument 406 ring->sp_cqn, cqe_index, err_cqe->vendor_err_syndrome, err_cqe->syndrome); in mlx4_en_handle_err_cqe() 407 print_hex_dump(KERN_WARNING, "", DUMP_PREFIX_OFFSET, 16, 1, err_cqe, sizeof(*err_cqe), in mlx4_en_handle_err_cqe() 410 wqe_index = be16_to_cpu(err_cqe->wqe_index) & ring->size_mask; in mlx4_en_handle_err_cqe()
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | cq.c | 446 struct mlx5_err_cqe *err_cqe; in mlx5_poll_one() local 516 err_cqe = (struct mlx5_err_cqe *)cqe64; in mlx5_poll_one() 517 mlx5_handle_error_cqe(dev, err_cqe, wc); in mlx5_poll_one() 522 err_cqe->syndrome, err_cqe->vendor_err_synd); in mlx5_poll_one()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
A D | en_stats.h | 428 u64 err_cqe; member
|
A D | en_rx.c | 1595 struct mlx5_err_cqe *err_cqe = (struct mlx5_err_cqe *)cqe; in trigger_report() local 1598 if (cqe_syndrome_needs_recover(err_cqe->syndrome) && in trigger_report() 1600 mlx5e_dump_error_cqe(&rq->cq, rq->rqn, err_cqe); in trigger_report()
|
A D | en_stats.c | 1929 { MLX5E_DECLARE_PTP_CQ_STAT(struct mlx5e_ptp_cq_stats, err_cqe) },
|
Completed in 39 milliseconds