rq_num 6242 drivers/infiniband/hw/mlx5/qp.c MLX5_SET(rqtc, rqtc, rq_num[i], init_attr->ind_tbl[i]->wq_num); rq_num 447 drivers/net/ethernet/cavium/thunder/nic.h u8 rq_num; rq_num 994 drivers/net/ethernet/cavium/thunder/nic_main.c (mbx.rq.rq_num << NIC_Q_NUM_SHIFT); rq_num 1009 drivers/net/ethernet/cavium/thunder/nic_main.c (mbx.rq.rq_num << NIC_Q_NUM_SHIFT); rq_num 1018 drivers/net/ethernet/cavium/thunder/nic_main.c (mbx.rq.rq_num << NIC_Q_NUM_SHIFT); rq_num 778 drivers/net/ethernet/cavium/thunder/nicvf_queues.c mbx.rq.rq_num = qidx; rq_num 469 drivers/net/ethernet/huawei/hinic/hinic_port.c struct hinic_rq_num rq_num = { 0 }; rq_num 470 drivers/net/ethernet/huawei/hinic/hinic_port.c u16 out_size = sizeof(rq_num); rq_num 473 drivers/net/ethernet/huawei/hinic/hinic_port.c rq_num.func_id = HINIC_HWIF_FUNC_IDX(hwif); rq_num 474 drivers/net/ethernet/huawei/hinic/hinic_port.c rq_num.num_rqs = num_rqs; rq_num 475 drivers/net/ethernet/huawei/hinic/hinic_port.c rq_num.rq_depth = ilog2(HINIC_SQ_DEPTH); rq_num 478 drivers/net/ethernet/huawei/hinic/hinic_port.c &rq_num, sizeof(rq_num), rq_num 479 drivers/net/ethernet/huawei/hinic/hinic_port.c &rq_num, &out_size); rq_num 480 drivers/net/ethernet/huawei/hinic/hinic_port.c if (err || !out_size || rq_num.status) { rq_num 483 drivers/net/ethernet/huawei/hinic/hinic_port.c rq_num.status); rq_num 528 drivers/net/ethernet/ibm/ehea/ehea_main.c static inline int ehea_check_cqe(struct ehea_cqe *cqe, int *rq_num) rq_num 530 drivers/net/ethernet/ibm/ehea/ehea_main.c *rq_num = (cqe->type & EHEA_CQE_TYPE_RQ) >> 5; rq_num 2455 drivers/net/ethernet/mellanox/mlx5/core/en_main.c MLX5_SET(rqtc, rqtc, rq_num[i], priv->drop_rq.rqn); rq_num 2548 drivers/net/ethernet/mellanox/mlx5/core/en_main.c MLX5_SET(rqtc, rqtc, rq_num[i], rqn); rq_num 503 drivers/net/ethernet/mellanox/mlx5/core/en_tc.c MLX5_SET(rqtc, rqtc, rq_num[i], rqn); rq_num 1647 include/linux/mlx5/mlx5_ifc.h u8 rq_num[0x18]; rq_num 3112 include/linux/mlx5/mlx5_ifc.h struct mlx5_ifc_rq_num_bits rq_num[0];