qid 329 arch/mips/cavium-octeon/executive/cvmx-helper-rgmii.c pko_mem_queue_qos.s.qid = queue; qid 83 arch/mips/cavium-octeon/executive/cvmx-pko.c config.s.qid = base_queue + queue; qid 186 arch/mips/include/asm/netlogic/xlp-hal/pic.h #define PIC_IRT_MSG_Q_INDEX(qid) ((qid) + PIC_IRT_MSG_Q0_INDEX) qid 1023 arch/mips/include/asm/octeon/cvmx-pko-defs.h uint64_t qid:8; qid 1025 arch/mips/include/asm/octeon/cvmx-pko-defs.h uint64_t qid:8; qid 1047 arch/mips/include/asm/octeon/cvmx-pko-defs.h uint64_t qid:8; qid 1049 arch/mips/include/asm/octeon/cvmx-pko-defs.h uint64_t qid:8; qid 1199 arch/mips/include/asm/octeon/cvmx-pko-defs.h uint64_t qid:7; qid 1201 arch/mips/include/asm/octeon/cvmx-pko-defs.h uint64_t qid:7; qid 78 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_tapq(ap_qid_t qid, unsigned long *info) qid 80 arch/s390/include/asm/ap.h register unsigned long reg0 asm ("0") = qid; qid 101 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_test_queue(ap_qid_t qid, qid 106 arch/s390/include/asm/ap.h qid |= 1UL << 23; /* set T bit*/ qid 107 arch/s390/include/asm/ap.h return ap_tapq(qid, info); qid 116 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_rapq(ap_qid_t qid) qid 118 arch/s390/include/asm/ap.h register unsigned long reg0 asm ("0") = qid | (1UL << 24); qid 135 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_zapq(ap_qid_t qid) qid 137 arch/s390/include/asm/ap.h register unsigned long reg0 asm ("0") = qid | (2UL << 24); qid 219 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_aqic(ap_qid_t qid, qid 223 arch/s390/include/asm/ap.h register unsigned long reg0 asm ("0") = qid | (3UL << 24); qid 268 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_qact(ap_qid_t qid, int ifbit, qid 271 arch/s390/include/asm/ap.h register unsigned long reg0 asm ("0") = qid | (5UL << 24) qid 302 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_nqap(ap_qid_t qid, qid 306 arch/s390/include/asm/ap.h register unsigned long reg0 asm ("0") = qid | 0x40000000UL; qid 340 arch/s390/include/asm/ap.h static inline struct ap_queue_status ap_dqap(ap_qid_t qid, qid 344 arch/s390/include/asm/ap.h register unsigned long reg0 asm("0") = qid | 0x80000000UL; qid 223 arch/s390/include/uapi/asm/zcrypt.h unsigned int qid:16; qid 341 arch/s390/include/uapi/asm/zcrypt.h unsigned int qid:14; qid 249 drivers/block/virtio_blk.c int qid = vq->index; qid 254 drivers/block/virtio_blk.c spin_lock_irqsave(&vblk->vqs[qid].lock, flags); qid 257 drivers/block/virtio_blk.c while ((vbr = virtqueue_get_buf(vblk->vqs[qid].vq, &len)) != NULL) { qid 270 drivers/block/virtio_blk.c spin_unlock_irqrestore(&vblk->vqs[qid].lock, flags); qid 295 drivers/block/virtio_blk.c int qid = hctx->queue_num; qid 351 drivers/block/virtio_blk.c spin_lock_irqsave(&vblk->vqs[qid].lock, flags); qid 353 drivers/block/virtio_blk.c err = virtblk_add_req_scsi(vblk->vqs[qid].vq, vbr, vbr->sg, num); qid 355 drivers/block/virtio_blk.c err = virtblk_add_req(vblk->vqs[qid].vq, vbr, vbr->sg, num); qid 357 drivers/block/virtio_blk.c virtqueue_kick(vblk->vqs[qid].vq); qid 363 drivers/block/virtio_blk.c spin_unlock_irqrestore(&vblk->vqs[qid].lock, flags); qid 374 drivers/block/virtio_blk.c if (bd->last && virtqueue_kick_prepare(vblk->vqs[qid].vq)) qid 376 drivers/block/virtio_blk.c spin_unlock_irqrestore(&vblk->vqs[qid].lock, flags); qid 379 drivers/block/virtio_blk.c virtqueue_notify(vblk->vqs[qid].vq); qid 886 drivers/block/xen-blkfront.c int qid = hctx->queue_num; qid 890 drivers/block/xen-blkfront.c BUG_ON(info->nr_rings <= qid); qid 891 drivers/block/xen-blkfront.c rinfo = &info->rinfo[qid]; qid 207 drivers/clk/zynqmp/clkc.c qdata.qid = PM_QID_CLOCK_GET_NUM_CLOCKS; qid 231 drivers/clk/zynqmp/clkc.c qdata.qid = PM_QID_CLOCK_GET_NAME; qid 264 drivers/clk/zynqmp/clkc.c qdata.qid = PM_QID_CLOCK_GET_TOPOLOGY; qid 296 drivers/clk/zynqmp/clkc.c qdata.qid = PM_QID_CLOCK_GET_FIXEDFACTOR_PARAMS; qid 338 drivers/clk/zynqmp/clkc.c qdata.qid = PM_QID_CLOCK_GET_PARENTS; qid 364 drivers/clk/zynqmp/clkc.c qdata.qid = PM_QID_CLOCK_GET_ATTRIBUTES; qid 527 drivers/crypto/cavium/zip/zip_regs.h u64 qid : 3; qid 531 drivers/crypto/cavium/zip/zip_regs.h u64 qid : 3; qid 388 drivers/crypto/chelsio/chcr_algo.c static inline void dsgl_walk_end(struct dsgl_walk *walk, unsigned short qid, qid 405 drivers/crypto/chelsio/chcr_algo.c phys_cpl->rss_hdr_int.qid = htons(qid); qid 728 drivers/crypto/chelsio/chcr_algo.c int qid = u_ctx->lldi.rxq_ids[ctx->rx_qidx]; qid 738 drivers/crypto/chelsio/chcr_algo.c FILL_WR_RX_Q_ID(ctx->tx_chan_id, qid, qid 742 drivers/crypto/chelsio/chcr_algo.c qid); qid 833 drivers/crypto/chelsio/chcr_algo.c chcr_add_cipher_dst_ent(wrparam->req, phys_cpl, wrparam, wrparam->qid); qid 1165 drivers/crypto/chelsio/chcr_algo.c wrparam.qid = u_ctx->lldi.rxq_ids[c_ctx(tfm)->rx_qidx]; qid 1189 drivers/crypto/chelsio/chcr_algo.c unsigned short qid, qid 1281 drivers/crypto/chelsio/chcr_algo.c wrparam.qid = qid; qid 2323 drivers/crypto/chelsio/chcr_algo.c unsigned short qid, qid 2444 drivers/crypto/chelsio/chcr_algo.c chcr_add_aead_dst_ent(req, phys_cpl, qid); qid 2562 drivers/crypto/chelsio/chcr_algo.c unsigned short qid) qid 2576 drivers/crypto/chelsio/chcr_algo.c dsgl_walk_end(&dsgl_walk, qid, ctx->pci_chan_id); qid 2605 drivers/crypto/chelsio/chcr_algo.c unsigned short qid) qid 2618 drivers/crypto/chelsio/chcr_algo.c dsgl_walk_end(&dsgl_walk, qid, ctx->pci_chan_id); qid 2884 drivers/crypto/chelsio/chcr_algo.c unsigned short qid, qid 2961 drivers/crypto/chelsio/chcr_algo.c chcr_add_aead_dst_ent(req, phys_cpl, qid); qid 2981 drivers/crypto/chelsio/chcr_algo.c unsigned short qid, qid 3078 drivers/crypto/chelsio/chcr_algo.c chcr_add_aead_dst_ent(req, phys_cpl, qid); qid 230 drivers/crypto/chelsio/chcr_algo.h #define FILL_WR_RX_Q_ID(cid, qid, lcb, fid) \ qid 233 drivers/crypto/chelsio/chcr_algo.h FW_CRYPTO_LOOKASIDE_WR_RX_Q_ID_V((qid)) | \ qid 238 drivers/crypto/chelsio/chcr_algo.h #define FILL_ULPTX_CMD_DEST(cid, qid) \ qid 244 drivers/crypto/chelsio/chcr_algo.h ULP_TXPKT_FID_V(qid)) qid 293 drivers/crypto/chelsio/chcr_algo.h unsigned short qid; qid 322 drivers/crypto/chelsio/chcr_crypto.h unsigned short qid); qid 332 drivers/crypto/chelsio/chcr_crypto.h unsigned short qid); qid 545 drivers/crypto/chelsio/chcr_ipsec.c int qid = q->q.cntxt_id; qid 578 drivers/crypto/chelsio/chcr_ipsec.c wr->req.ulptx.cmd_dest = FILL_ULPTX_CMD_DEST(pi->port_id, qid); qid 31 drivers/crypto/virtio/virtio_crypto_core.c unsigned int qid = vq->index; qid 33 drivers/crypto/virtio/virtio_crypto_core.c spin_lock_irqsave(&vcrypto->data_vq[qid].lock, flags); qid 38 drivers/crypto/virtio/virtio_crypto_core.c &vcrypto->data_vq[qid].lock, flags); qid 42 drivers/crypto/virtio/virtio_crypto_core.c &vcrypto->data_vq[qid].lock, flags); qid 45 drivers/crypto/virtio/virtio_crypto_core.c spin_unlock_irqrestore(&vcrypto->data_vq[qid].lock, flags); qid 100 drivers/firmware/xilinx/zynqmp-debug.c qdata.qid = pm_api_arg[0]; qid 109 drivers/firmware/xilinx/zynqmp-debug.c switch (qdata.qid) { qid 288 drivers/firmware/xilinx/zynqmp.c ret = zynqmp_pm_invoke_fn(PM_QUERY_DATA, qdata.qid, qdata.arg1, qid 296 drivers/firmware/xilinx/zynqmp.c return qdata.qid == PM_QID_CLOCK_GET_NAME ? 0 : ret; qid 181 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c unsigned int qid; qid 188 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c &properties, &qid); qid 195 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c pr_debug("DIQ Created with queue id: %d\n", qid); qid 197 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c kq = pqm_get_kernel_queue(dbgdev->pqm, qid); qid 201 drivers/gpu/drm/amd/amdkfd/kfd_dbgdev.c pqm_destroy_queue(dbgdev->pqm, qid); qid 906 drivers/gpu/drm/amd/amdkfd/kfd_priv.h unsigned int *qid); qid 907 drivers/gpu/drm/amd/amdkfd/kfd_priv.h int pqm_destroy_queue(struct process_queue_manager *pqm, unsigned int qid); qid 908 drivers/gpu/drm/amd/amdkfd/kfd_priv.h int pqm_update_queue(struct process_queue_manager *pqm, unsigned int qid, qid 910 drivers/gpu/drm/amd/amdkfd/kfd_priv.h int pqm_set_cu_mask(struct process_queue_manager *pqm, unsigned int qid, qid 912 drivers/gpu/drm/amd/amdkfd/kfd_priv.h int pqm_set_gws(struct process_queue_manager *pqm, unsigned int qid, qid 915 drivers/gpu/drm/amd/amdkfd/kfd_priv.h unsigned int qid); qid 917 drivers/gpu/drm/amd/amdkfd/kfd_priv.h unsigned int qid, qid 32 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c struct process_queue_manager *pqm, unsigned int qid) qid 37 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c if ((pqn->q && pqn->q->properties.queue_id == qid) || qid 38 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c (pqn->kq && pqn->kq->queue->properties.queue_id == qid)) qid 46 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c unsigned int *qid) qid 62 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c *qid = found; qid 78 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c int pqm_set_gws(struct process_queue_manager *pqm, unsigned int qid, qid 87 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c pqn = get_queue_by_qid(pqm, qid); qid 168 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c struct file *f, unsigned int qid) qid 177 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c q_properties->queue_id = qid; qid 195 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c unsigned int *qid) qid 226 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c retval = find_available_queue_slot(pqm, qid); qid 253 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c retval = create_cp_queue(pqm, dev, &q, properties, f, *qid); qid 274 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c retval = create_cp_queue(pqm, dev, &q, properties, f, *qid); qid 289 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c kq->queue->properties.queue_id = *qid; qid 330 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c clear_bit(*qid, pqm->queue_slot_bitmap); qid 337 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c int pqm_destroy_queue(struct process_queue_manager *pqm, unsigned int qid) qid 349 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c pqn = get_queue_by_qid(pqm, qid); qid 400 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c clear_bit(qid, pqm->queue_slot_bitmap); qid 410 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c int pqm_update_queue(struct process_queue_manager *pqm, unsigned int qid, qid 416 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c pqn = get_queue_by_qid(pqm, qid); qid 418 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c pr_debug("No queue %d exists for update operation\n", qid); qid 435 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c int pqm_set_cu_mask(struct process_queue_manager *pqm, unsigned int qid, qid 441 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c pqn = get_queue_by_qid(pqm, qid); qid 443 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c pr_debug("No queue %d exists for update operation\n", qid); qid 465 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c unsigned int qid) qid 469 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c pqn = get_queue_by_qid(pqm, qid); qid 477 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c unsigned int qid, qid 484 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c pqn = get_queue_by_qid(pqm, qid); qid 487 drivers/gpu/drm/amd/amdkfd/kfd_process_queue_manager.c qid); qid 1879 drivers/infiniband/hw/cxgb4/cm.c __func__, ep->com.qp->wq.sq.qid, ep, qid 3032 drivers/infiniband/hw/cxgb4/cm.c ep->com.qp->wq.sq.qid); qid 195 drivers/infiniband/hw/cxgb4/cq.c CQE_QPID_V(wq->sq.qid)); qid 229 drivers/infiniband/hw/cxgb4/cq.c CQE_QPID_V(wq->sq.qid)); qid 425 drivers/infiniband/hw/cxgb4/cq.c WARN_ONCE(1, "Unexpected DRAIN CQE qp id %u!\n", wq->sq.qid); qid 454 drivers/infiniband/hw/cxgb4/cq.c (CQE_QPID(cqe) == wq->sq.qid) && cqe_completes_wr(cqe, wq)) qid 107 drivers/infiniband/hw/cxgb4/device.c le.qid = wq->sq.qid; qid 113 drivers/infiniband/hw/cxgb4/device.c le.qid = wq->rq.qid; qid 152 drivers/infiniband/hw/cxgb4/device.c lep->qid, lep->opcode, qid 250 drivers/infiniband/hw/cxgb4/device.c if (id != qp->wq.sq.qid) qid 271 drivers/infiniband/hw/cxgb4/device.c qp->wq.sq.qid, qp->srq ? "srq" : "rq", qid 272 drivers/infiniband/hw/cxgb4/device.c qp->srq ? qp->srq->idx : qp->wq.rq.qid, qid 292 drivers/infiniband/hw/cxgb4/device.c qp->wq.sq.qid, qp->wq.rq.qid, qid 306 drivers/infiniband/hw/cxgb4/device.c qp->wq.sq.qid, qp->wq.rq.qid, qid 479 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.qid.total, dev->rdev.stats.qid.cur, qid 480 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.qid.max, dev->rdev.stats.qid.fail); qid 526 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.qid.max = 0; qid 527 drivers/infiniband/hw/cxgb4/device.c dev->rdev.stats.qid.fail = 0; qid 757 drivers/infiniband/hw/cxgb4/device.c if (!(entry->qid & rdev->qpmask)) { qid 759 drivers/infiniband/hw/cxgb4/device.c entry->qid); qid 761 drivers/infiniband/hw/cxgb4/device.c rdev->stats.qid.cur -= rdev->qpmask + 1; qid 848 drivers/infiniband/hw/cxgb4/device.c rdev->stats.qid.total = rdev->lldi.vr->qp.size; qid 1185 drivers/infiniband/hw/cxgb4/device.c u32 qid = be32_to_cpu(rc->pldbuflen_qid); qid 1186 drivers/infiniband/hw/cxgb4/device.c c4iw_ev_handler(dev, qid); qid 1378 drivers/infiniband/hw/cxgb4/device.c qp->wq.sq.qid, qid 1383 drivers/infiniband/hw/cxgb4/device.c pci_name(ctx->lldi.pdev), qp->wq.sq.qid); qid 1391 drivers/infiniband/hw/cxgb4/device.c qp->wq.rq.qid, qid 1397 drivers/infiniband/hw/cxgb4/device.c pci_name(ctx->lldi.pdev), qp->wq.rq.qid); qid 211 drivers/infiniband/hw/cxgb4/ev.c CQE_STATUS(err_cqe), qhp->wq.sq.qid); qid 223 drivers/infiniband/hw/cxgb4/ev.c int c4iw_ev_handler(struct c4iw_dev *dev, u32 qid) qid 229 drivers/infiniband/hw/cxgb4/ev.c chp = xa_load(&dev->cqs, qid); qid 240 drivers/infiniband/hw/cxgb4/ev.c pr_debug("unknown cqid 0x%x\n", qid); qid 105 drivers/infiniband/hw/cxgb4/iw_cxgb4.h u32 qid; qid 129 drivers/infiniband/hw/cxgb4/iw_cxgb4.h struct c4iw_stat qid; qid 165 drivers/infiniband/hw/cxgb4/iw_cxgb4.h u16 qid; qid 1025 drivers/infiniband/hw/cxgb4/iw_cxgb4.h int c4iw_ev_handler(struct c4iw_dev *rnicp, u32 qid); qid 1029 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void c4iw_put_cqid(struct c4iw_rdev *rdev, u32 qid, qid 1032 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void c4iw_put_qpid(struct c4iw_rdev *rdev, u32 qid, qid 1038 drivers/infiniband/hw/cxgb4/iw_cxgb4.h void __iomem *c4iw_bar2_addrs(struct c4iw_rdev *rdev, unsigned int qid, qid 159 drivers/infiniband/hw/cxgb4/qp.c c4iw_put_qpid(rdev, wq->sq.qid, uctx); qid 167 drivers/infiniband/hw/cxgb4/qp.c c4iw_put_qpid(rdev, wq->rq.qid, uctx); qid 177 drivers/infiniband/hw/cxgb4/qp.c void __iomem *c4iw_bar2_addrs(struct c4iw_rdev *rdev, unsigned int qid, qid 184 drivers/infiniband/hw/cxgb4/qp.c ret = cxgb4_bar2_sge_qregs(rdev->lldi.ports[0], qid, qtype, qid 213 drivers/infiniband/hw/cxgb4/qp.c wq->sq.qid = c4iw_get_qpid(rdev, uctx); qid 214 drivers/infiniband/hw/cxgb4/qp.c if (!wq->sq.qid) qid 218 drivers/infiniband/hw/cxgb4/qp.c wq->rq.qid = c4iw_get_qpid(rdev, uctx); qid 219 drivers/infiniband/hw/cxgb4/qp.c if (!wq->rq.qid) { qid 282 drivers/infiniband/hw/cxgb4/qp.c wq->sq.bar2_va = c4iw_bar2_addrs(rdev, wq->sq.qid, qid 287 drivers/infiniband/hw/cxgb4/qp.c wq->rq.bar2_va = c4iw_bar2_addrs(rdev, wq->rq.qid, qid 297 drivers/infiniband/hw/cxgb4/qp.c pci_name(rdev->lldi.pdev), wq->sq.qid, wq->rq.qid); qid 347 drivers/infiniband/hw/cxgb4/qp.c res->u.sqrq.eqid = cpu_to_be32(wq->sq.qid); qid 376 drivers/infiniband/hw/cxgb4/qp.c res->u.sqrq.eqid = cpu_to_be32(wq->rq.qid); qid 381 drivers/infiniband/hw/cxgb4/qp.c ret = c4iw_ref_send_wait(rdev, skb, wr_waitp, 0, wq->sq.qid, __func__); qid 386 drivers/infiniband/hw/cxgb4/qp.c wq->sq.qid, wq->rq.qid, wq->db, qid 407 drivers/infiniband/hw/cxgb4/qp.c c4iw_put_qpid(rdev, wq->rq.qid, uctx); qid 409 drivers/infiniband/hw/cxgb4/qp.c c4iw_put_qpid(rdev, wq->sq.qid, uctx); qid 1000 drivers/infiniband/hw/cxgb4/qp.c CQE_QPID_V(qhp->wq.sq.qid)); qid 1051 drivers/infiniband/hw/cxgb4/qp.c CQE_QPID_V(qhp->wq.sq.qid)); qid 1566 drivers/infiniband/hw/cxgb4/qp.c pr_debug("qhp %p qid 0x%x tid %u\n", qhp, qhp->wq.sq.qid, qid 1698 drivers/infiniband/hw/cxgb4/qp.c pr_debug("qhp %p qid 0x%x tid %u\n", qhp, qhp->wq.sq.qid, ep->hwtid); qid 1718 drivers/infiniband/hw/cxgb4/qp.c qhp->ep->hwtid, qhp->wq.sq.qid, __func__); qid 1755 drivers/infiniband/hw/cxgb4/qp.c qhp->wq.sq.qid, qhp->ep->hwtid, qhp->ep->ird, qhp->ep->ord); qid 1800 drivers/infiniband/hw/cxgb4/qp.c wqe->u.init.qpid = cpu_to_be32(qhp->wq.sq.qid); qid 1801 drivers/infiniband/hw/cxgb4/qp.c wqe->u.init.sq_eqid = cpu_to_be32(qhp->wq.sq.qid); qid 1806 drivers/infiniband/hw/cxgb4/qp.c wqe->u.init.rq_eqid = cpu_to_be32(qhp->wq.rq.qid); qid 1821 drivers/infiniband/hw/cxgb4/qp.c qhp->ep->hwtid, qhp->wq.sq.qid, __func__); qid 1845 drivers/infiniband/hw/cxgb4/qp.c qhp, qhp->wq.sq.qid, qhp->wq.rq.qid, qhp->ep, qhp->attr.state, qid 2031 drivers/infiniband/hw/cxgb4/qp.c qhp->wq.sq.qid); qid 2089 drivers/infiniband/hw/cxgb4/qp.c __xa_erase(&rhp->qps, qhp->wq.sq.qid); qid 2099 drivers/infiniband/hw/cxgb4/qp.c pr_debug("ib_qp %p qpid 0x%0x\n", ib_qp, qhp->wq.sq.qid); qid 2220 drivers/infiniband/hw/cxgb4/qp.c ret = xa_insert_irq(&rhp->qps, qhp->wq.sq.qid, qhp, GFP_KERNEL); qid 2263 drivers/infiniband/hw/cxgb4/qp.c uresp.sqid = qhp->wq.sq.qid; qid 2267 drivers/infiniband/hw/cxgb4/qp.c uresp.rqid = qhp->wq.rq.qid; qid 2334 drivers/infiniband/hw/cxgb4/qp.c qhp->ibqp.qp_num = qhp->wq.sq.qid; qid 2339 drivers/infiniband/hw/cxgb4/qp.c qhp->wq.sq.qid, qhp->wq.sq.size, qhp->wq.sq.memsize, qid 2340 drivers/infiniband/hw/cxgb4/qp.c attrs->cap.max_send_wr, qhp->wq.rq.qid, qhp->wq.rq.size, qid 2356 drivers/infiniband/hw/cxgb4/qp.c xa_erase_irq(&rhp->qps, qhp->wq.sq.qid); qid 2504 drivers/infiniband/hw/cxgb4/qp.c res->u.srq.eqid = cpu_to_be32(wq->qid); qid 2514 drivers/infiniband/hw/cxgb4/qp.c c4iw_put_qpid(rdev, wq->qid, uctx); qid 2530 drivers/infiniband/hw/cxgb4/qp.c wq->qid = c4iw_get_qpid(rdev, uctx); qid 2531 drivers/infiniband/hw/cxgb4/qp.c if (!wq->qid) qid 2560 drivers/infiniband/hw/cxgb4/qp.c wq->bar2_va = c4iw_bar2_addrs(rdev, wq->qid, CXGB4_BAR2_QTYPE_EGRESS, qid 2570 drivers/infiniband/hw/cxgb4/qp.c pci_name(rdev->lldi.pdev), wq->qid); qid 2599 drivers/infiniband/hw/cxgb4/qp.c res->u.srq.eqid = cpu_to_be32(wq->qid); qid 2623 drivers/infiniband/hw/cxgb4/qp.c ret = c4iw_ref_send_wait(rdev, skb, wr_waitp, 0, wq->qid, __func__); qid 2629 drivers/infiniband/hw/cxgb4/qp.c __func__, srq->idx, wq->qid, srq->pdid, wq->queue, qid 2647 drivers/infiniband/hw/cxgb4/qp.c c4iw_put_qpid(rdev, wq->qid, uctx); qid 2754 drivers/infiniband/hw/cxgb4/qp.c uresp.srqid = srq->wq.qid; qid 2778 drivers/infiniband/hw/cxgb4/qp.c __func__, srq->wq.qid, srq->idx, srq->wq.size, qid 2809 drivers/infiniband/hw/cxgb4/qp.c pr_debug("%s id %d\n", __func__, srq->wq.qid); qid 111 drivers/infiniband/hw/cxgb4/resource.c u32 qid; qid 119 drivers/infiniband/hw/cxgb4/resource.c qid = entry->qid; qid 122 drivers/infiniband/hw/cxgb4/resource.c qid = c4iw_get_resource(&rdev->resource.qid_table); qid 123 drivers/infiniband/hw/cxgb4/resource.c if (!qid) qid 126 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.qid.cur += rdev->qpmask + 1; qid 128 drivers/infiniband/hw/cxgb4/resource.c for (i = qid+1; i & rdev->qpmask; i++) { qid 132 drivers/infiniband/hw/cxgb4/resource.c entry->qid = i; qid 143 drivers/infiniband/hw/cxgb4/resource.c entry->qid = qid; qid 145 drivers/infiniband/hw/cxgb4/resource.c for (i = qid+1; i & rdev->qpmask; i++) { qid 149 drivers/infiniband/hw/cxgb4/resource.c entry->qid = i; qid 155 drivers/infiniband/hw/cxgb4/resource.c pr_debug("qid 0x%x\n", qid); qid 157 drivers/infiniband/hw/cxgb4/resource.c if (rdev->stats.qid.cur > rdev->stats.qid.max) qid 158 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.qid.max = rdev->stats.qid.cur; qid 160 drivers/infiniband/hw/cxgb4/resource.c return qid; qid 163 drivers/infiniband/hw/cxgb4/resource.c void c4iw_put_cqid(struct c4iw_rdev *rdev, u32 qid, qid 171 drivers/infiniband/hw/cxgb4/resource.c pr_debug("qid 0x%x\n", qid); qid 172 drivers/infiniband/hw/cxgb4/resource.c entry->qid = qid; qid 181 drivers/infiniband/hw/cxgb4/resource.c u32 qid; qid 189 drivers/infiniband/hw/cxgb4/resource.c qid = entry->qid; qid 192 drivers/infiniband/hw/cxgb4/resource.c qid = c4iw_get_resource(&rdev->resource.qid_table); qid 193 drivers/infiniband/hw/cxgb4/resource.c if (!qid) { qid 195 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.qid.fail++; qid 200 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.qid.cur += rdev->qpmask + 1; qid 202 drivers/infiniband/hw/cxgb4/resource.c for (i = qid+1; i & rdev->qpmask; i++) { qid 206 drivers/infiniband/hw/cxgb4/resource.c entry->qid = i; qid 217 drivers/infiniband/hw/cxgb4/resource.c entry->qid = qid; qid 219 drivers/infiniband/hw/cxgb4/resource.c for (i = qid; i & rdev->qpmask; i++) { qid 223 drivers/infiniband/hw/cxgb4/resource.c entry->qid = i; qid 229 drivers/infiniband/hw/cxgb4/resource.c pr_debug("qid 0x%x\n", qid); qid 231 drivers/infiniband/hw/cxgb4/resource.c if (rdev->stats.qid.cur > rdev->stats.qid.max) qid 232 drivers/infiniband/hw/cxgb4/resource.c rdev->stats.qid.max = rdev->stats.qid.cur; qid 234 drivers/infiniband/hw/cxgb4/resource.c return qid; qid 237 drivers/infiniband/hw/cxgb4/resource.c void c4iw_put_qpid(struct c4iw_rdev *rdev, u32 qid, qid 245 drivers/infiniband/hw/cxgb4/resource.c pr_debug("qid 0x%x\n", qid); qid 246 drivers/infiniband/hw/cxgb4/resource.c entry->qid = qid; qid 42 drivers/infiniband/hw/cxgb4/restrack.c if (rdma_nl_put_driver_u32(msg, "sqid", wq->sq.qid)) qid 70 drivers/infiniband/hw/cxgb4/restrack.c if (rdma_nl_put_driver_u32(msg, "rqid", wq->rq.qid)) qid 51 drivers/infiniband/hw/cxgb4/t4.h __be16 qid; qid 343 drivers/infiniband/hw/cxgb4/t4.h u32 qid; qid 370 drivers/infiniband/hw/cxgb4/t4.h u32 qid; qid 407 drivers/infiniband/hw/cxgb4/t4.h u32 qid; qid 632 drivers/infiniband/hw/cxgb4/t4.h writel(QID_V(wq->sq.qid) | PIDX_V(inc), wq->db); qid 657 drivers/infiniband/hw/cxgb4/t4.h writel(QID_V(wq->rq.qid) | PIDX_V(inc), wq->db); qid 1316 drivers/net/bonding/bond_options.c u16 qid; qid 1327 drivers/net/bonding/bond_options.c if (sscanf(++delim, "%hd\n", &qid) != 1) qid 1332 drivers/net/bonding/bond_options.c qid > bond->dev->real_num_tx_queues) qid 1348 drivers/net/bonding/bond_options.c else if (qid && qid == slave->queue_id) { qid 1357 drivers/net/bonding/bond_options.c update_slave->queue_id = qid; qid 1275 drivers/net/ethernet/amazon/ena/ena_com.c u16 qid; qid 1279 drivers/net/ethernet/amazon/ena/ena_com.c qid = rss->host_rss_ind_tbl[i]; qid 1280 drivers/net/ethernet/amazon/ena/ena_com.c if (qid >= ENA_TOTAL_NUM_QUEUES) qid 1283 drivers/net/ethernet/amazon/ena/ena_com.c io_sq = &ena_dev->io_sq_queues[qid]; qid 1417 drivers/net/ethernet/amazon/ena/ena_com.c int ena_com_get_io_handlers(struct ena_com_dev *ena_dev, u16 qid, qid 1421 drivers/net/ethernet/amazon/ena/ena_com.c if (qid >= ENA_TOTAL_NUM_QUEUES) { qid 1422 drivers/net/ethernet/amazon/ena/ena_com.c pr_err("Invalid queue number %d but the max is %d\n", qid, qid 1427 drivers/net/ethernet/amazon/ena/ena_com.c *io_sq = &ena_dev->io_sq_queues[qid]; qid 1428 drivers/net/ethernet/amazon/ena/ena_com.c *io_cq = &ena_dev->io_cq_queues[qid]; qid 1825 drivers/net/ethernet/amazon/ena/ena_com.c if (ctx->qid >= ENA_TOTAL_NUM_QUEUES) { qid 1827 drivers/net/ethernet/amazon/ena/ena_com.c ctx->qid, ENA_TOTAL_NUM_QUEUES); qid 1831 drivers/net/ethernet/amazon/ena/ena_com.c io_sq = &ena_dev->io_sq_queues[ctx->qid]; qid 1832 drivers/net/ethernet/amazon/ena/ena_com.c io_cq = &ena_dev->io_cq_queues[ctx->qid]; qid 1840 drivers/net/ethernet/amazon/ena/ena_com.c io_cq->qid = ctx->qid; qid 1846 drivers/net/ethernet/amazon/ena/ena_com.c io_sq->qid = ctx->qid; qid 1879 drivers/net/ethernet/amazon/ena/ena_com.c void ena_com_destroy_io_queue(struct ena_com_dev *ena_dev, u16 qid) qid 1884 drivers/net/ethernet/amazon/ena/ena_com.c if (qid >= ENA_TOTAL_NUM_QUEUES) { qid 1885 drivers/net/ethernet/amazon/ena/ena_com.c pr_err("Qid (%d) is bigger than max num of queues (%d)\n", qid, qid 1890 drivers/net/ethernet/amazon/ena/ena_com.c io_sq = &ena_dev->io_sq_queues[qid]; qid 1891 drivers/net/ethernet/amazon/ena/ena_com.c io_cq = &ena_dev->io_cq_queues[qid]; qid 156 drivers/net/ethernet/amazon/ena/ena_com.h u16 qid; qid 197 drivers/net/ethernet/amazon/ena/ena_com.h u16 qid; qid 370 drivers/net/ethernet/amazon/ena/ena_com.h u16 qid; qid 459 drivers/net/ethernet/amazon/ena/ena_com.h void ena_com_destroy_io_queue(struct ena_com_dev *ena_dev, u16 qid); qid 469 drivers/net/ethernet/amazon/ena/ena_com.h int ena_com_get_io_handlers(struct ena_com_dev *ena_dev, u16 qid, qid 93 drivers/net/ethernet/amazon/ena/ena_eth_com.c io_sq->qid, io_sq->entries_in_tx_burst_left); qid 279 drivers/net/ethernet/amazon/ena/ena_eth_com.c io_cq->qid, *first_cdesc_idx, count); qid 532 drivers/net/ethernet/amazon/ena/ena_eth_com.c pr_debug("fetch rx packet: queue %d completed desc: %d\n", io_cq->qid, qid 552 drivers/net/ethernet/amazon/ena/ena_eth_com.c pr_debug("[%s][QID#%d] Updating SQ head to: %d\n", __func__, io_sq->qid, qid 168 drivers/net/ethernet/amazon/ena/ena_eth_com.h pr_debug("queue: %d num_descs: %d num_entries_needed: %d\n", io_sq->qid, qid 180 drivers/net/ethernet/amazon/ena/ena_eth_com.h io_sq->qid, tail); qid 186 drivers/net/ethernet/amazon/ena/ena_eth_com.h io_sq->qid, max_entries_in_tx_burst); qid 205 drivers/net/ethernet/amazon/ena/ena_eth_com.h io_cq->qid, head); qid 151 drivers/net/ethernet/amazon/ena/ena_netdev.c struct ena_ring *ring, u16 qid) qid 153 drivers/net/ethernet/amazon/ena/ena_netdev.c ring->qid = qid; qid 157 drivers/net/ethernet/amazon/ena/ena_netdev.c ring->napi = &adapter->ena_napi[qid].napi; qid 208 drivers/net/ethernet/amazon/ena/ena_netdev.c static int ena_setup_tx_resources(struct ena_adapter *adapter, int qid) qid 210 drivers/net/ethernet/amazon/ena/ena_netdev.c struct ena_ring *tx_ring = &adapter->tx_ring[qid]; qid 211 drivers/net/ethernet/amazon/ena/ena_netdev.c struct ena_irq *ena_irq = &adapter->irq_tbl[ENA_IO_IRQ_IDX(qid)]; qid 274 drivers/net/ethernet/amazon/ena/ena_netdev.c static void ena_free_tx_resources(struct ena_adapter *adapter, int qid) qid 276 drivers/net/ethernet/amazon/ena/ena_netdev.c struct ena_ring *tx_ring = &adapter->tx_ring[qid]; qid 354 drivers/net/ethernet/amazon/ena/ena_netdev.c u32 qid) qid 356 drivers/net/ethernet/amazon/ena/ena_netdev.c struct ena_ring *rx_ring = &adapter->rx_ring[qid]; qid 357 drivers/net/ethernet/amazon/ena/ena_netdev.c struct ena_irq *ena_irq = &adapter->irq_tbl[ENA_IO_IRQ_IDX(qid)]; qid 411 drivers/net/ethernet/amazon/ena/ena_netdev.c u32 qid) qid 413 drivers/net/ethernet/amazon/ena/ena_netdev.c struct ena_ring *rx_ring = &adapter->rx_ring[qid]; qid 547 drivers/net/ethernet/amazon/ena/ena_netdev.c rx_ring->qid); qid 556 drivers/net/ethernet/amazon/ena/ena_netdev.c rx_ring->qid); qid 569 drivers/net/ethernet/amazon/ena/ena_netdev.c rx_ring->qid, i, num); qid 582 drivers/net/ethernet/amazon/ena/ena_netdev.c u32 qid) qid 584 drivers/net/ethernet/amazon/ena/ena_netdev.c struct ena_ring *rx_ring = &adapter->rx_ring[qid]; qid 670 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring->qid, i); qid 675 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring->qid, i); qid 683 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring->qid)); qid 765 drivers/net/ethernet/amazon/ena/ena_netdev.c txq = netdev_get_tx_queue(tx_ring->netdev, tx_ring->qid); qid 792 drivers/net/ethernet/amazon/ena/ena_netdev.c "tx_poll: q %d skb %p completed\n", tx_ring->qid, qid 813 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring->qid, tx_pkts); qid 1064 drivers/net/ethernet/amazon/ena/ena_netdev.c "%s qid %d\n", __func__, rx_ring->qid); qid 1082 drivers/net/ethernet/amazon/ena/ena_netdev.c rx_ring->qid, ena_rx_ctx.descs, ena_rx_ctx.l3_proto, qid 1105 drivers/net/ethernet/amazon/ena/ena_netdev.c skb_record_rx_queue(skb, rx_ring->qid); qid 1554 drivers/net/ethernet/amazon/ena/ena_netdev.c napi->qid = i; qid 1628 drivers/net/ethernet/amazon/ena/ena_netdev.c static int ena_create_io_tx_queue(struct ena_adapter *adapter, int qid) qid 1639 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring = &adapter->tx_ring[qid]; qid 1640 drivers/net/ethernet/amazon/ena/ena_netdev.c msix_vector = ENA_IO_IRQ_IDX(qid); qid 1641 drivers/net/ethernet/amazon/ena/ena_netdev.c ena_qid = ENA_IO_TXQ_IDX(qid); qid 1646 drivers/net/ethernet/amazon/ena/ena_netdev.c ctx.qid = ena_qid; qid 1656 drivers/net/ethernet/amazon/ena/ena_netdev.c qid, rc); qid 1666 drivers/net/ethernet/amazon/ena/ena_netdev.c qid, rc); qid 1695 drivers/net/ethernet/amazon/ena/ena_netdev.c static int ena_create_io_rx_queue(struct ena_adapter *adapter, int qid) qid 1706 drivers/net/ethernet/amazon/ena/ena_netdev.c rx_ring = &adapter->rx_ring[qid]; qid 1707 drivers/net/ethernet/amazon/ena/ena_netdev.c msix_vector = ENA_IO_IRQ_IDX(qid); qid 1708 drivers/net/ethernet/amazon/ena/ena_netdev.c ena_qid = ENA_IO_RXQ_IDX(qid); qid 1712 drivers/net/ethernet/amazon/ena/ena_netdev.c ctx.qid = ena_qid; qid 1723 drivers/net/ethernet/amazon/ena/ena_netdev.c qid, rc); qid 1733 drivers/net/ethernet/amazon/ena/ena_netdev.c qid, rc); qid 2255 drivers/net/ethernet/amazon/ena/ena_netdev.c int qid, rc, nb_hw_desc; qid 2259 drivers/net/ethernet/amazon/ena/ena_netdev.c qid = skb_get_queue_mapping(skb); qid 2260 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring = &adapter->tx_ring[qid]; qid 2261 drivers/net/ethernet/amazon/ena/ena_netdev.c txq = netdev_get_tx_queue(dev, qid); qid 2293 drivers/net/ethernet/amazon/ena/ena_netdev.c qid); qid 2341 drivers/net/ethernet/amazon/ena/ena_netdev.c __func__, qid); qid 2391 drivers/net/ethernet/amazon/ena/ena_netdev.c u16 qid; qid 2397 drivers/net/ethernet/amazon/ena/ena_netdev.c qid = skb_get_rx_queue(skb); qid 2399 drivers/net/ethernet/amazon/ena/ena_netdev.c qid = netdev_pick_tx(dev, skb, NULL); qid 2401 drivers/net/ethernet/amazon/ena/ena_netdev.c return qid; qid 2865 drivers/net/ethernet/amazon/ena/ena_netdev.c rx_ring->qid); qid 2898 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring->qid); qid 2910 drivers/net/ethernet/amazon/ena/ena_netdev.c tx_ring->qid, i); qid 158 drivers/net/ethernet/amazon/ena/ena_netdev.h u32 qid; qid 265 drivers/net/ethernet/amazon/ena/ena_netdev.h u16 qid; qid 235 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c struct bnx2x_virtf *vf, int qid, qid 241 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid); qid 245 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c q_params->q_obj = &bnx2x_vfq(vf, qid, sp_obj); qid 268 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c bnx2x_vf_igu_ack_sb(bp, vf, vf_igu_sb(vf, bnx2x_vfq(vf, qid, sb_idx)), qid 275 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int qid) qid 287 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c q_params.q_obj = &bnx2x_vfq(vf, qid, sp_obj); qid 307 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c if (bnx2x_vfq(vf, qid, cxt)) { qid 308 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c bnx2x_vfq(vf, qid, cxt)->ustorm_ag_context.cdu_usage = 0; qid 309 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c bnx2x_vfq(vf, qid, cxt)->xstorm_ag_context.cdu_reserved = 0; qid 356 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int qid, bool drv_only, int type) qid 369 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, vlan_mac_obj); qid 372 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, mac_obj); qid 374 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, vlan_obj); qid 400 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c struct bnx2x_virtf *vf, int qid, qid 415 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, vlan_mac_obj); qid 420 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, vlan_obj); qid 424 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c ramrod.vlan_mac_obj = &bnx2x_vfq(vf, qid, mac_obj); qid 457 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int qid, bool drv_only) qid 468 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_mac_vlan_config(bp, vf, qid, qid 482 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c bnx2x_vf_mac_vlan_config(bp, vf, qid, qid 494 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int bnx2x_vf_queue_setup(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid, qid 499 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid); qid 501 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_queue_create(bp, vf, qid, qctor); qid 510 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c BNX2X_ERR("QSETUP[%d:%d] error: rc %d\n", vf->abs_vfid, qid, rc); qid 515 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int qid) qid 519 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid); qid 522 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c if ((qid == LEADING_IDX) && qid 524 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid, true, qid 528 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid, true, qid 532 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid, true, qid 539 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c if (bnx2x_vfq(vf, qid, sp_obj).state != BNX2X_Q_STATE_RESET) { qid 543 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c qstate.q_obj = &bnx2x_vfq(vf, qid, sp_obj); qid 554 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c BNX2X_ERR("vf[%d:%d] error: rc %d\n", vf->abs_vfid, qid, rc); qid 608 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c static void bnx2x_vf_prep_rx_mode(struct bnx2x *bp, u8 qid, qid 613 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c struct bnx2x_vf_queue *vfq = vfq_get(vf, qid); qid 634 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int qid, unsigned long accept_flags) qid 640 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c bnx2x_vf_prep_rx_mode(bp, qid, &ramrod, vf, accept_flags); qid 642 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c vfq_get(vf, qid)->accept_flags = ramrod.rx_accept_flags; qid 646 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int bnx2x_vf_queue_teardown(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid) qid 650 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c DP(BNX2X_MSG_IOV, "vf[%d:%d]\n", vf->abs_vfid, qid); qid 653 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c if (qid == LEADING_IDX) { qid 654 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_rxmode(bp, vf, qid, 0); qid 660 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid, qid 665 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid, qid 670 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_vlan_mac_clear(bp, vf, qid, qid 682 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c rc = bnx2x_vf_queue_destroy(bp, vf, qid); qid 688 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c vf->abs_vfid, qid, rc); qid 1968 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c static void bnx2x_vf_qtbl_set_q(struct bnx2x *bp, u8 abs_vfid, u8 qid, qid 1971 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c u32 reg = PXP_REG_HST_ZONE_PERMISSION_TABLE + qid * 4; qid 2262 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c int qid, rc = 0; qid 2273 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c for (qid = 0; qid < vf_rxq_count(vf); qid++) { qid 2274 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c qstate.q_obj = &bnx2x_vfq(vf, qid, sp_obj); qid 2275 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c qstate.params.update_tpa.sge_map = sge_addr[qid]; qid 2277 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c vf->abs_vfid, qid, U64_HI(sge_addr[qid]), qid 2278 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c U64_LO(sge_addr[qid])); qid 2282 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c U64_HI(sge_addr[qid]), U64_LO(sge_addr[qid]), qid 2283 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.c vf->abs_vfid, qid); qid 453 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h int qid, bool drv_only); qid 455 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h int bnx2x_vf_queue_setup(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid, qid 458 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h int bnx2x_vf_queue_teardown(struct bnx2x *bp, struct bnx2x_virtf *vf, int qid); qid 464 drivers/net/ethernet/broadcom/bnx2x/bnx2x_sriov.h int qid, unsigned long accept_flags); qid 1947 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c int qid = mbx->msg->req.q_op.vf_qid; qid 1951 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c vf->abs_vfid, qid); qid 1953 drivers/net/ethernet/broadcom/bnx2x/bnx2x_vfpf.c rc = bnx2x_vf_queue_teardown(bp, vf, qid); qid 27 drivers/net/ethernet/brocade/bna/bfi.h u8 qid; qid 36 drivers/net/ethernet/brocade/bna/bfi.h #define bfi_mhdr_2_qid(_mh) ((_mh)->mtag.h2i.qid) qid 1257 drivers/net/ethernet/cavium/thunder/nicvf_main.c int qid = skb_get_queue_mapping(skb); qid 1258 drivers/net/ethernet/cavium/thunder/nicvf_main.c struct netdev_queue *txq = netdev_get_tx_queue(netdev, qid); qid 1274 drivers/net/ethernet/cavium/thunder/nicvf_main.c qid += nic->xdp_tx_queues; qid 1278 drivers/net/ethernet/cavium/thunder/nicvf_main.c if (qid >= MAX_SND_QUEUES_PER_QS) { qid 1279 drivers/net/ethernet/cavium/thunder/nicvf_main.c tmp = qid / MAX_SND_QUEUES_PER_QS; qid 1288 drivers/net/ethernet/cavium/thunder/nicvf_main.c qid = qid % MAX_SND_QUEUES_PER_QS; qid 1291 drivers/net/ethernet/cavium/thunder/nicvf_main.c sq = &snic->qs->sq[qid]; qid 1293 drivers/net/ethernet/cavium/thunder/nicvf_main.c !nicvf_sq_append_skb(snic, sq, skb, qid)) { qid 1305 drivers/net/ethernet/cavium/thunder/nicvf_main.c "Transmit ring full, stopping SQ%d\n", qid); qid 1660 drivers/net/ethernet/chelsio/cxgb/sge.c unsigned int qid, struct net_device *dev) qid 1663 drivers/net/ethernet/chelsio/cxgb/sge.c struct cmdQ *q = &sge->cmdQ[qid]; qid 1697 drivers/net/ethernet/chelsio/cxgb/sge.c if (sge->tx_sched && !qid && skb->dev) { qid 1732 drivers/net/ethernet/chelsio/cxgb/sge.c if (qid) qid 1423 drivers/net/ethernet/chelsio/cxgb3/sge.c unsigned int qid) qid 1430 drivers/net/ethernet/chelsio/cxgb3/sge.c struct sge_qset *qs = txq_to_qset(q, qid); qid 1432 drivers/net/ethernet/chelsio/cxgb3/sge.c set_bit(qid, &qs->txq_stopped); qid 1436 drivers/net/ethernet/chelsio/cxgb3/sge.c test_and_clear_bit(qid, &qs->txq_stopped)) qid 1167 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h __u8 qid:4; qid 1171 drivers/net/ethernet/chelsio/cxgb3/t3_cpl.h __u8 qid:4; qid 336 drivers/net/ethernet/chelsio/cxgb4/cudbg_entity.h u32 qid; qid 566 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c struct cudbg_error *cudbg_err, int qid) qid 580 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c no_of_read_words = t4_read_cim_ibq(padap, qid, qid 637 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c u32 cudbg_cim_obq_size(struct adapter *padap, int qid) qid 642 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c QUENUMSELECT_V(qid)); qid 650 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c struct cudbg_error *cudbg_err, int qid) qid 658 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c qsize = cudbg_cim_obq_size(padap, qid); qid 664 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.c no_of_read_words = t4_read_cim_obq(padap, qid, qid 169 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h u32 cudbg_cim_obq_size(struct adapter *padap, int qid); qid 242 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qid = txq->cntxt_id; qid 254 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qid = rxq->cntxt_id; qid 266 drivers/net/ethernet/chelsio/cxgb4/cudbg_lib.h entry->qid = flq->cntxt_id; qid 1668 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h unsigned int qid, qid 1708 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_read_cim_ibq(struct adapter *adap, unsigned int qid, u32 *data, qid 1710 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h int t4_read_cim_obq(struct adapter *adap, unsigned int qid, u32 *data, qid 1751 drivers/net/ethernet/chelsio/cxgb4/cxgb4.h void t4_mk_filtdelwr(unsigned int ftid, struct fw_filter_wr *wr, int qid); qid 395 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c unsigned int qid = (uintptr_t)inode->i_private & 7; qid 396 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c struct adapter *adap = inode->i_private - qid; qid 402 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c ret = t4_read_cim_ibq(adap, qid, (u32 *)p->data, CIM_IBQ_SIZE * 4); qid 422 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c unsigned int qid = (uintptr_t)inode->i_private & 7; qid 423 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c struct adapter *adap = inode->i_private - qid; qid 429 drivers/net/ethernet/chelsio/cxgb4/cxgb4_debugfs.c ret = t4_read_cim_obq(adap, qid, (u32 *)p->data, 6 * CIM_OBQ_SIZE * 4); qid 575 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c unsigned int qid = EGR_QID_G(ntohl(p->opcode_qid)); qid 578 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c txq = q->adap->sge.egr_map[qid - q->adap->sge.egr_start]; qid 896 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static unsigned int rxq_to_chan(const struct sge *p, unsigned int qid) qid 898 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c qid -= p->ingr_start; qid 899 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c return netdev2pinfo(p->ingr_map[qid]->netdev)->tx_chan; qid 1814 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c static int read_eq_indices(struct adapter *adap, u16 qid, u16 *pidx, u16 *cidx) qid 1816 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c u32 addr = t4_read_reg(adap, SGE_DBQ_CTXT_BADDR_A) + 24 * qid + 8; qid 1832 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c int cxgb4_sync_txq_pidx(struct net_device *dev, u16 qid, u16 pidx, qid 1839 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c ret = read_eq_indices(adap, qid, &hw_pidx, &hw_cidx); qid 1858 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c QID_V(qid) | val); qid 1955 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c unsigned int qid, qid 1962 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c qid, qid 2193 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c u16 qid = (dropped_db >> 15) & 0x1ffff; qid 2199 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c ret = t4_bar2_sge_qregs(adap, qid, T4_BAR2_QTYPE_EGRESS, qid 2203 drivers/net/ethernet/chelsio/cxgb4/cxgb4_main.c "qid=%d, pidx_inc=%d\n", qid, pidx_inc); qid 184 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h void cxgb4_remove_tid(struct tid_info *t, unsigned int qid, unsigned int tid, qid 414 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h int cxgb4_sync_txq_pidx(struct net_device *dev, u16 qid, u16 pidx, u16 size); qid 421 drivers/net/ethernet/chelsio/cxgb4/cxgb4_uld.h unsigned int qid, qid 109 drivers/net/ethernet/chelsio/cxgb4/sched.c const unsigned int qid, qid 127 drivers/net/ethernet/chelsio/cxgb4/sched.c if (qe->cntxt_id == qid) { qid 149 drivers/net/ethernet/chelsio/cxgb4/sched.c unsigned int qid; qid 157 drivers/net/ethernet/chelsio/cxgb4/sched.c qid = txq->q.cntxt_id; qid 160 drivers/net/ethernet/chelsio/cxgb4/sched.c e = t4_sched_queue_lookup(pi, qid, &index); qid 191 drivers/net/ethernet/chelsio/cxgb4/sched.c unsigned int qid; qid 202 drivers/net/ethernet/chelsio/cxgb4/sched.c qid = txq->q.cntxt_id; qid 210 drivers/net/ethernet/chelsio/cxgb4/sched.c qe->cntxt_id = qid; qid 3312 drivers/net/ethernet/chelsio/cxgb4/sge.c unsigned int qid = ntohl(rc->pldbuflen_qid); qid 3314 drivers/net/ethernet/chelsio/cxgb4/sge.c qid -= adap->sge.ingr_start; qid 3315 drivers/net/ethernet/chelsio/cxgb4/sge.c napi_schedule(&adap->sge.ingr_map[qid]->napi); qid 3493 drivers/net/ethernet/chelsio/cxgb4/sge.c unsigned int qid, qid 3500 drivers/net/ethernet/chelsio/cxgb4/sge.c ret = t4_bar2_sge_qregs(adapter, qid, qtype, 0, qid 6502 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c void t4_mk_filtdelwr(unsigned int ftid, struct fw_filter_wr *wr, int qid) qid 6508 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c FW_FILTER_WR_NOREPLY_V(qid < 0)); qid 6510 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (qid >= 0) qid 6512 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c cpu_to_be16(FW_FILTER_WR_RX_RPL_IQ_V(qid)); qid 9247 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c unsigned int qid, qid 9278 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c bar2_page_offset = ((u64)(qid >> qpp_shift) << page_shift); qid 9279 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c bar2_qid = qid & qpp_mask; qid 9756 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int t4_read_cim_ibq(struct adapter *adap, unsigned int qid, u32 *data, size_t n) qid 9762 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if (qid > 5 || (n & 3)) qid 9765 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c addr = qid * nwords; qid 9798 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c int t4_read_cim_obq(struct adapter *adap, unsigned int qid, u32 *data, size_t n) qid 9805 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c if ((qid > (cim_num_obq - 1)) || (n & 3)) qid 9809 drivers/net/ethernet/chelsio/cxgb4/t4_hw.c QUENUMSELECT_V(qid)); qid 150 drivers/net/ethernet/chelsio/cxgb4/t4_hw.h __be32 qid; qid 235 drivers/net/ethernet/chelsio/cxgb4/t4_msg.h __be16 qid; qid 535 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c unsigned int qid = EGR_QID_G(be32_to_cpu(p->opcode_qid)); qid 548 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c eq_idx = EQ_IDX(s, qid); qid 551 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c "Egress Update QID %d out of range\n", qid); qid 557 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c "Egress Update QID %d TXQ=NULL\n", qid); qid 561 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c if (unlikely(tq->abs_id != qid)) { qid 564 drivers/net/ethernet/chelsio/cxgb4vf/cxgb4vf_main.c qid, tq->abs_id); qid 1947 drivers/net/ethernet/chelsio/cxgb4vf/sge.c unsigned int qid, iq_idx; qid 1979 drivers/net/ethernet/chelsio/cxgb4vf/sge.c qid = RSPD_QID_G(be32_to_cpu(rc->pldbuflen_qid)); qid 1980 drivers/net/ethernet/chelsio/cxgb4vf/sge.c iq_idx = IQ_IDX(s, qid); qid 1983 drivers/net/ethernet/chelsio/cxgb4vf/sge.c "Ingress QID %d out of range\n", qid); qid 1989 drivers/net/ethernet/chelsio/cxgb4vf/sge.c "Ingress QID %d RSPQ=NULL\n", qid); qid 1992 drivers/net/ethernet/chelsio/cxgb4vf/sge.c if (unlikely(rspq->abs_id != qid)) { qid 1995 drivers/net/ethernet/chelsio/cxgb4vf/sge.c qid, rspq->abs_id); qid 2179 drivers/net/ethernet/chelsio/cxgb4vf/sge.c unsigned int qid, qid 2186 drivers/net/ethernet/chelsio/cxgb4vf/sge.c ret = t4vf_bar2_sge_qregs(adapter, qid, qtype, qid 374 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_common.h unsigned int qid, qid 768 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c unsigned int qid, qid 799 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c bar2_page_offset = ((u64)(qid >> qpp_shift) << page_shift); qid 800 drivers/net/ethernet/chelsio/cxgb4vf/t4vf_hw.c bar2_qid = qid & qpp_mask; qid 971 drivers/net/ethernet/emulex/benet/be.h void be_cq_notify(struct be_adapter *adapter, u16 qid, bool arm, qid 208 drivers/net/ethernet/emulex/benet/be_main.c static void be_rxq_notify(struct be_adapter *adapter, u16 qid, u16 posted) qid 215 drivers/net/ethernet/emulex/benet/be_main.c val |= qid & DB_RQ_RING_ID_MASK; qid 237 drivers/net/ethernet/emulex/benet/be_main.c static void be_eq_notify(struct be_adapter *adapter, u16 qid, qid 243 drivers/net/ethernet/emulex/benet/be_main.c val |= qid & DB_EQ_RING_ID_MASK; qid 244 drivers/net/ethernet/emulex/benet/be_main.c val |= ((qid & DB_EQ_RING_ID_EXT_MASK) << DB_EQ_RING_ID_EXT_MASK_SHIFT); qid 259 drivers/net/ethernet/emulex/benet/be_main.c void be_cq_notify(struct be_adapter *adapter, u16 qid, bool arm, u16 num_popped) qid 263 drivers/net/ethernet/emulex/benet/be_main.c val |= qid & DB_CQ_RING_ID_MASK; qid 264 drivers/net/ethernet/emulex/benet/be_main.c val |= ((qid & DB_CQ_RING_ID_EXT_MASK) << qid 2554 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpni_queue_id qid = {0}; qid 2573 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c &queue, &qid); qid 2577 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c fq->tx_fqid[j] = qid.fqid; qid 2685 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpni_queue_id qid; qid 2689 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c DPNI_QUEUE_RX, 0, fq->flowid, &queue, &qid); qid 2695 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c fq->fqid = qid.fqid; qid 2733 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c struct dpni_queue_id qid; qid 2739 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c &queue, &qid); qid 2744 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c fq->tx_fqid[i] = qid.fqid; qid 2748 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c fq->tx_qdbin = qid.qdbin; qid 2752 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c &queue, &qid); qid 2758 drivers/net/ethernet/freescale/dpaa2/dpaa2-eth.c fq->fqid = qid.fqid; qid 1428 drivers/net/ethernet/freescale/dpaa2/dpni.c struct dpni_queue_id *qid) qid 1461 drivers/net/ethernet/freescale/dpaa2/dpni.c qid->fqid = le32_to_cpu(rsp_params->fqid); qid 1462 drivers/net/ethernet/freescale/dpaa2/dpni.c qid->qdbin = le16_to_cpu(rsp_params->qdbin); qid 834 drivers/net/ethernet/freescale/dpaa2/dpni.h struct dpni_queue_id *qid); qid 461 drivers/net/ethernet/freescale/fec.h int qid; qid 424 drivers/net/ethernet/freescale/fec_main.c estatus |= FEC_TX_BD_FTYPE(txq->bd.qid); qid 557 drivers/net/ethernet/freescale/fec_main.c estatus |= FEC_TX_BD_FTYPE(txq->bd.qid); qid 637 drivers/net/ethernet/freescale/fec_main.c estatus |= FEC_TX_BD_FTYPE(txq->bd.qid); qid 700 drivers/net/ethernet/freescale/fec_main.c estatus |= FEC_TX_BD_FTYPE(txq->bd.qid); qid 3277 drivers/net/ethernet/freescale/fec_main.c rxq->bd.qid = i; qid 3293 drivers/net/ethernet/freescale/fec_main.c txq->bd.qid = i; qid 192 drivers/net/ethernet/hisilicon/hns/hnae.h __u8 qid; qid 168 drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.c static void hns_ppe_set_qid(struct ppe_common_cb *ppe_common, u32 qid) qid 175 drivers/net/ethernet/hisilicon/hns/hns_dsaf_ppe.c PPE_CFG_QID_MODE_DEF_QID_S, qid); qid 605 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c u16 i = 0, j = 0, qid = 0; qid 608 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c for (qid = 0; qid < nic_dev->num_qps; qid++) { qid 612 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c hinic_txq_get_stats(&nic_dev->txqs[qid], &txq_stats); qid 621 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c for (qid = 0; qid < nic_dev->num_qps; qid++) { qid 625 drivers/net/ethernet/huawei/hinic/hinic_ethtool.c hinic_rxq_get_stats(&nic_dev->rxqs[qid], &rxq_stats); qid 3114 drivers/net/ethernet/intel/i40e/i40e_main.c int qid = ring->queue_index; qid 3117 drivers/net/ethernet/intel/i40e/i40e_main.c qid -= ring->vsi->alloc_queue_pairs; qid 3119 drivers/net/ethernet/intel/i40e/i40e_main.c if (!xdp_on || !test_bit(qid, ring->vsi->af_xdp_zc_qps)) qid 3122 drivers/net/ethernet/intel/i40e/i40e_main.c return xdp_get_umem_from_qid(ring->vsi->netdev, qid); qid 190 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c u16 qid) qid 195 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c return (vsi && (qid < vsi->alloc_queue_pairs)); qid 824 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c u16 qid = i40e_vc_get_pf_queue_id(vf, qid 827 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c reg = qid; qid 828 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c qid = i40e_vc_get_pf_queue_id(vf, vsi_id, qid 830 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c reg |= qid << 16; qid 852 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c u16 vsi_id, qid; qid 868 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c qid = i40e_vc_get_pf_queue_id(vf, vsi_id, j); qid 870 drivers/net/ethernet/intel/i40e/i40e_virtchnl_pf.c reg = (qid & I40E_VPLAN_QTABLE_QINDEX_MASK); qid 78 drivers/net/ethernet/intel/i40e/i40e_xsk.c u16 qid) qid 88 drivers/net/ethernet/intel/i40e/i40e_xsk.c if (qid >= vsi->num_queue_pairs) qid 91 drivers/net/ethernet/intel/i40e/i40e_xsk.c if (qid >= netdev->real_num_rx_queues || qid 92 drivers/net/ethernet/intel/i40e/i40e_xsk.c qid >= netdev->real_num_tx_queues) qid 105 drivers/net/ethernet/intel/i40e/i40e_xsk.c set_bit(qid, vsi->af_xdp_zc_qps); qid 110 drivers/net/ethernet/intel/i40e/i40e_xsk.c err = i40e_queue_pair_disable(vsi, qid); qid 114 drivers/net/ethernet/intel/i40e/i40e_xsk.c err = i40e_queue_pair_enable(vsi, qid); qid 119 drivers/net/ethernet/intel/i40e/i40e_xsk.c err = i40e_xsk_wakeup(vsi->netdev, qid, XDP_WAKEUP_RX); qid 134 drivers/net/ethernet/intel/i40e/i40e_xsk.c static int i40e_xsk_umem_disable(struct i40e_vsi *vsi, u16 qid) qid 141 drivers/net/ethernet/intel/i40e/i40e_xsk.c umem = xdp_get_umem_from_qid(netdev, qid); qid 148 drivers/net/ethernet/intel/i40e/i40e_xsk.c err = i40e_queue_pair_disable(vsi, qid); qid 153 drivers/net/ethernet/intel/i40e/i40e_xsk.c clear_bit(qid, vsi->af_xdp_zc_qps); qid 157 drivers/net/ethernet/intel/i40e/i40e_xsk.c err = i40e_queue_pair_enable(vsi, qid); qid 176 drivers/net/ethernet/intel/i40e/i40e_xsk.c u16 qid) qid 178 drivers/net/ethernet/intel/i40e/i40e_xsk.c return umem ? i40e_xsk_umem_enable(vsi, umem, qid) : qid 179 drivers/net/ethernet/intel/i40e/i40e_xsk.c i40e_xsk_umem_disable(vsi, qid); qid 14 drivers/net/ethernet/intel/i40e/i40e_xsk.h u16 qid); qid 1747 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c static bool ice_vc_isvalid_q_id(struct ice_vf *vf, u16 vsi_id, u8 qid) qid 1751 drivers/net/ethernet/intel/ice/ice_virtchnl_pf.c return (vsi && (qid < vsi->alloc_txq)); qid 34 drivers/net/ethernet/intel/ixgbe/ixgbe_txrx_common.h u16 qid); qid 15 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c int qid = ring->ring_idx; qid 17 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c if (!xdp_on || !test_bit(qid, adapter->af_xdp_zc_qps)) qid 20 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c return xdp_get_umem_from_qid(adapter->netdev, qid); qid 67 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c u16 qid) qid 74 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c if (qid >= adapter->num_rx_queues) qid 77 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c if (qid >= netdev->real_num_rx_queues || qid 78 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c qid >= netdev->real_num_tx_queues) qid 95 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c ixgbe_txrx_ring_disable(adapter, qid); qid 97 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c set_bit(qid, adapter->af_xdp_zc_qps); qid 100 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c ixgbe_txrx_ring_enable(adapter, qid); qid 103 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c err = ixgbe_xsk_wakeup(adapter->netdev, qid, XDP_WAKEUP_RX); qid 111 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c static int ixgbe_xsk_umem_disable(struct ixgbe_adapter *adapter, u16 qid) qid 116 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c umem = xdp_get_umem_from_qid(adapter->netdev, qid); qid 124 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c ixgbe_txrx_ring_disable(adapter, qid); qid 126 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c clear_bit(qid, adapter->af_xdp_zc_qps); qid 130 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c ixgbe_txrx_ring_enable(adapter, qid); qid 136 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c u16 qid) qid 138 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c return umem ? ixgbe_xsk_umem_enable(adapter, umem, qid) : qid 139 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c ixgbe_xsk_umem_disable(adapter, qid); qid 698 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c int ixgbe_xsk_wakeup(struct net_device *dev, u32 qid, u32 flags) qid 709 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c if (qid >= adapter->num_xdp_queues) qid 712 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c ring = adapter->xdp_ring[qid]; qid 44 drivers/net/ethernet/mellanox/mlx5/core/en/params.h u16 qid, qid 49 drivers/net/ethernet/mellanox/mlx5/core/en/params.h int ch = qid - nch * group; qid 59 drivers/net/ethernet/mellanox/mlx5/core/en/params.h u16 qid, qid 65 drivers/net/ethernet/mellanox/mlx5/core/en/params.h *ix = qid % nch; qid 66 drivers/net/ethernet/mellanox/mlx5/core/en/params.h *group = qid / nch; qid 70 drivers/net/ethernet/mellanox/mlx5/core/en/params.h struct mlx5e_params *params, u64 qid) qid 72 drivers/net/ethernet/mellanox/mlx5/core/en/params.h return qid < params->num_channels * profile->rq_groups; qid 10 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c int mlx5e_xsk_wakeup(struct net_device *dev, u32 qid, u32 flags) qid 20 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.c if (unlikely(!mlx5e_qid_get_ch_if_in_group(params, qid, MLX5E_RQ_GROUP_XSK, &ix))) qid 12 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/tx.h int mlx5e_xsk_wakeup(struct net_device *dev, u32 qid, u32 flags); qid 231 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/umem.c int mlx5e_xsk_setup_umem(struct net_device *dev, struct xdp_umem *umem, u16 qid) qid 237 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/umem.c if (unlikely(!mlx5e_qid_get_ch_if_in_group(params, qid, MLX5E_RQ_GROUP_XSK, &ix))) qid 25 drivers/net/ethernet/mellanox/mlx5/core/en/xsk/umem.h int mlx5e_xsk_setup_umem(struct net_device *dev, struct xdp_umem *umem, u16 qid); qid 54 drivers/net/ethernet/netronome/nfp/abm/ctrl.c unsigned int qid; qid 58 drivers/net/ethernet/netronome/nfp/abm/ctrl.c qid = band * NFP_NET_MAX_RX_RINGS + alink->queue_base + queue; qid 60 drivers/net/ethernet/netronome/nfp/abm/ctrl.c sym_offset = qid * stride + offset; qid 134 drivers/net/ethernet/netronome/nfp/abm/ctrl.c unsigned int qid; qid 136 drivers/net/ethernet/netronome/nfp/abm/ctrl.c qid = band * NFP_NET_MAX_RX_RINGS + alink->queue_base + queue; qid 138 drivers/net/ethernet/netronome/nfp/abm/ctrl.c return __nfp_abm_ctrl_set_q_act(alink->abm, qid, act); qid 1535 drivers/net/ethernet/pensando/ionic/ionic_if.h __le32 qid; qid 2068 drivers/net/ethernet/qlogic/qed/qed_l2.c if (p_params->qid != QED_RFS_NTUPLE_QID_RSS) { qid 2069 drivers/net/ethernet/qlogic/qed/qed_l2.c rc = qed_fw_l2_queue(p_hwfn, p_params->qid, qid 2849 drivers/net/ethernet/qlogic/qed/qed_l2.c params->qid = QED_RFS_NTUPLE_QID_RSS; qid 1489 drivers/net/ethernet/qlogic/qed/qed_ll2.c u8 qid; qid 1543 drivers/net/ethernet/qlogic/qed/qed_ll2.c qid = p_hwfn->hw_info.resc_start[QED_LL2_QUEUE] + connection_handle; qid 1544 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_conn->queue_id = qid; qid 1545 drivers/net/ethernet/qlogic/qed/qed_ll2.c p_ll2_conn->tx_stats_id = qid; qid 1548 drivers/net/ethernet/qlogic/qed/qed_ll2.c TSTORM_LL2_RX_PRODS_OFFSET(qid); qid 2102 drivers/net/ethernet/qlogic/qed/qed_ll2.c u8 qid = p_ll2_conn->queue_id; qid 2107 drivers/net/ethernet/qlogic/qed/qed_ll2.c CORE_LL2_TSTORM_PER_QUEUE_STAT_OFFSET(qid); qid 2121 drivers/net/ethernet/qlogic/qed/qed_ll2.c u8 qid = p_ll2_conn->queue_id; qid 2126 drivers/net/ethernet/qlogic/qed/qed_ll2.c CORE_LL2_USTORM_PER_QUEUE_STAT_OFFSET(qid); qid 223 drivers/net/ethernet/qlogic/qed/qed_sriov.c u16 qid, qid 235 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_qcid = &p_vf->vf_queues[qid].cids[i]; qid 865 drivers/net/ethernet/qlogic/qed/qed_sriov.c int qid; qid 867 drivers/net/ethernet/qlogic/qed/qed_sriov.c for (qid = 0; qid < vf->num_rxqs; qid++) { qid 868 drivers/net/ethernet/qlogic/qed/qed_sriov.c qed_fw_l2_queue(p_hwfn, vf->vf_queues[qid].fw_rx_qid, qid 896 drivers/net/ethernet/qlogic/qed/qed_sriov.c int qid = 0; qid 907 drivers/net/ethernet/qlogic/qed/qed_sriov.c for (qid = 0; qid < num_rx_queues; qid++) { qid 909 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->igu_sbs[qid] = p_block->igu_sb_id; qid 911 drivers/net/ethernet/qlogic/qed/qed_sriov.c SET_FIELD(val, IGU_MAPPING_LINE_VECTOR_NUMBER, qid); qid 1004 drivers/net/ethernet/qlogic/qed/qed_sriov.c u16 qid, num_irqs; qid 1027 drivers/net/ethernet/qlogic/qed/qed_sriov.c qid = p_params->req_rx_queue[i]; qid 1028 drivers/net/ethernet/qlogic/qed/qed_sriov.c if (qid < min_vf_qzone || qid > max_vf_qzone) { qid 1031 drivers/net/ethernet/qlogic/qed/qed_sriov.c qid, qid 1037 drivers/net/ethernet/qlogic/qed/qed_sriov.c qid = p_params->req_tx_queue[i]; qid 1038 drivers/net/ethernet/qlogic/qed/qed_sriov.c if (qid > max_vf_qzone) { qid 1041 drivers/net/ethernet/qlogic/qed/qed_sriov.c qid, p_params->rel_vf_id, max_vf_qzone); qid 1046 drivers/net/ethernet/qlogic/qed/qed_sriov.c if (qid < min_vf_qzone) qid 1050 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_params->rel_vf_id, qid, i); qid 2094 drivers/net/ethernet/qlogic/qed/qed_sriov.c if (p_qid_tlv->qid >= MAX_QUEUES_PER_QZONE) { qid 2097 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_vf->relative_vf_id, p_qid_tlv->qid); qid 2101 drivers/net/ethernet/qlogic/qed/qed_sriov.c return p_qid_tlv->qid; qid 2709 drivers/net/ethernet/qlogic/qed/qed_sriov.c u16 qid = req->rx_qid + i; qid 2711 drivers/net/ethernet/qlogic/qed/qed_sriov.c handlers[i] = vf->vf_queues[qid].cids[qid_usage_idx].p_cid; qid 3434 drivers/net/ethernet/qlogic/qed/qed_sriov.c u16 coal = 0, qid, i; qid 3441 drivers/net/ethernet/qlogic/qed/qed_sriov.c qid = req->qid; qid 3445 drivers/net/ethernet/qlogic/qed/qed_sriov.c if (!qed_iov_validate_rxq(p_hwfn, p_vf, qid, qid 3449 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_vf->abs_vf_id, qid); qid 3453 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_cid = qed_iov_get_vf_rx_queue_cid(&p_vf->vf_queues[qid]); qid 3458 drivers/net/ethernet/qlogic/qed/qed_sriov.c if (!qed_iov_validate_txq(p_hwfn, p_vf, qid, qid 3462 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_vf->abs_vf_id, qid); qid 3466 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_queue = &p_vf->vf_queues[qid]; qid 3503 drivers/net/ethernet/qlogic/qed/qed_sriov.c u16 qid; qid 3509 drivers/net/ethernet/qlogic/qed/qed_sriov.c qid = req->qid; qid 3511 drivers/net/ethernet/qlogic/qed/qed_sriov.c if (!qed_iov_validate_rxq(p_hwfn, vf, qid, qid 3515 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->abs_vf_id, qid); qid 3519 drivers/net/ethernet/qlogic/qed/qed_sriov.c if (!qed_iov_validate_txq(p_hwfn, vf, qid, qid 3523 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->abs_vf_id, qid); qid 3530 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->abs_vf_id, rx_coal, tx_coal, qid); qid 3533 drivers/net/ethernet/qlogic/qed/qed_sriov.c p_cid = qed_iov_get_vf_rx_queue_cid(&vf->vf_queues[qid]); qid 3540 drivers/net/ethernet/qlogic/qed/qed_sriov.c vf->abs_vf_id, vf->vf_queues[qid].fw_rx_qid); qid 3547 drivers/net/ethernet/qlogic/qed/qed_sriov.c struct qed_vf_queue *p_queue = &vf->vf_queues[qid]; qid 169 drivers/net/ethernet/qlogic/qed/qed_vf.c p_qid_tlv->qid = p_cid->qid_usage_idx; qid 854 drivers/net/ethernet/qlogic/qed/qed_vf.c u16 qid = p_cid->rel.queue_id; qid 860 drivers/net/ethernet/qlogic/qed/qed_vf.c req->tx_qid = qid; qid 890 drivers/net/ethernet/qlogic/qed/qed_vf.c u8 cid = p_iov->acquire_resp.resc.cid[qid]; qid 899 drivers/net/ethernet/qlogic/qed/qed_vf.c qid, p_cid->qid_usage_idx, *pp_doorbell, resp->offset); qid 1365 drivers/net/ethernet/qlogic/qed/qed_vf.c req->qid = p_cid->rel.queue_id; qid 1429 drivers/net/ethernet/qlogic/qed/qed_vf.c req->qid = p_cid->rel.queue_id; qid 1434 drivers/net/ethernet/qlogic/qed/qed_vf.c rx_coal, tx_coal, req->qid); qid 260 drivers/net/ethernet/qlogic/qed/qed_vf.h u8 qid; qid 509 drivers/net/ethernet/qlogic/qed/qed_vf.h u16 qid; qid 515 drivers/net/ethernet/qlogic/qed/qed_vf.h u16 qid; qid 127 drivers/net/ethernet/qlogic/qede/qede_filter.c params.qid = rxq_id; qid 1094 drivers/net/ethernet/sfc/falcon/farch.c int qid; qid 1096 drivers/net/ethernet/sfc/falcon/farch.c qid = EF4_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_SUBDATA); qid 1097 drivers/net/ethernet/sfc/falcon/farch.c if (qid < EF4_TXQ_TYPES * efx->n_tx_channels) { qid 1098 drivers/net/ethernet/sfc/falcon/farch.c tx_queue = ef4_get_tx_queue(efx, qid / EF4_TXQ_TYPES, qid 1099 drivers/net/ethernet/sfc/falcon/farch.c qid % EF4_TXQ_TYPES); qid 1116 drivers/net/ethernet/sfc/falcon/farch.c int qid; qid 1119 drivers/net/ethernet/sfc/falcon/farch.c qid = EF4_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_RX_DESCQ_ID); qid 1121 drivers/net/ethernet/sfc/falcon/farch.c if (qid >= efx->n_channels) qid 1123 drivers/net/ethernet/sfc/falcon/farch.c channel = ef4_get_channel(efx, qid); qid 1130 drivers/net/ethernet/sfc/falcon/farch.c "RXQ %d flush retry\n", qid); qid 1082 drivers/net/ethernet/sfc/farch.c int qid; qid 1084 drivers/net/ethernet/sfc/farch.c qid = EFX_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_SUBDATA); qid 1085 drivers/net/ethernet/sfc/farch.c if (qid < EFX_TXQ_TYPES * (efx->n_tx_channels + efx->n_extra_tx_channels)) { qid 1086 drivers/net/ethernet/sfc/farch.c tx_queue = efx_get_tx_queue(efx, qid / EFX_TXQ_TYPES, qid 1087 drivers/net/ethernet/sfc/farch.c qid % EFX_TXQ_TYPES); qid 1104 drivers/net/ethernet/sfc/farch.c int qid; qid 1107 drivers/net/ethernet/sfc/farch.c qid = EFX_QWORD_FIELD(*event, FSF_AZ_DRIVER_EV_RX_DESCQ_ID); qid 1109 drivers/net/ethernet/sfc/farch.c if (qid >= efx->n_channels) qid 1111 drivers/net/ethernet/sfc/farch.c channel = efx_get_channel(efx, qid); qid 1118 drivers/net/ethernet/sfc/farch.c "RXQ %d flush retry\n", qid); qid 1405 drivers/net/ethernet/sfc/siena_sriov.c unsigned qid, seq, type, data; qid 1407 drivers/net/ethernet/sfc/siena_sriov.c qid = EFX_QWORD_FIELD(*event, FSF_CZ_USER_QID); qid 1417 drivers/net/ethernet/sfc/siena_sriov.c qid, seq, type, data); qid 1419 drivers/net/ethernet/sfc/siena_sriov.c if (map_vi_index(efx, qid, &vf, NULL)) qid 1493 drivers/net/ethernet/sfc/siena_sriov.c unsigned queue, qid; qid 1496 drivers/net/ethernet/sfc/siena_sriov.c if (map_vi_index(efx, queue, &vf, &qid)) qid 1499 drivers/net/ethernet/sfc/siena_sriov.c if (!test_bit(qid, vf->txq_mask)) qid 1502 drivers/net/ethernet/sfc/siena_sriov.c __clear_bit(qid, vf->txq_mask); qid 1512 drivers/net/ethernet/sfc/siena_sriov.c unsigned ev_failed, queue, qid; qid 1517 drivers/net/ethernet/sfc/siena_sriov.c if (map_vi_index(efx, queue, &vf, &qid)) qid 1519 drivers/net/ethernet/sfc/siena_sriov.c if (!test_bit(qid, vf->rxq_mask)) qid 1523 drivers/net/ethernet/sfc/siena_sriov.c set_bit(qid, vf->rxq_retry_mask); qid 1526 drivers/net/ethernet/sfc/siena_sriov.c __clear_bit(qid, vf->rxq_mask); qid 139 drivers/net/hyperv/netvsc_trace.h __field( u16, qid ) qid 146 drivers/net/hyperv/netvsc_trace.h __entry->qid = chan->offermsg.offer.sub_channel_index; qid 152 drivers/net/hyperv/netvsc_trace.h __get_str(name), __entry->qid, qid 164 drivers/net/hyperv/netvsc_trace.h __field( u16, qid ) qid 169 drivers/net/hyperv/netvsc_trace.h __entry->qid = chan->offermsg.offer.sub_channel_index; qid 173 drivers/net/hyperv/netvsc_trace.h __get_str(name), __entry->qid, qid 1137 drivers/net/vmxnet3/vmxnet3_drv.c VMXNET3_REG_TXPROD + tq->qid * 8, qid 1316 drivers/net/vmxnet3/vmxnet3_drv.c BUG_ON(rcd->rqID != rq->qid && rcd->rqID != rq->qid2 && qid 1338 drivers/net/vmxnet3/vmxnet3_drv.c (rcd->rqID != rq->qid && qid 1566 drivers/net/vmxnet3/vmxnet3_drv.c rxprod_reg[ring_idx] + rq->qid * 8, qid 2158 drivers/net/vmxnet3/vmxnet3_drv.c rq->qid = i; qid 2835 drivers/net/vmxnet3/vmxnet3_drv.c tq->qid = i; qid 244 drivers/net/vmxnet3/vmxnet3_int.h int qid; qid 290 drivers/net/vmxnet3/vmxnet3_int.h u32 qid; /* rqID in RCD for buffer from 1st ring */ qid 444 drivers/net/wireless/ath/ar5523/ar5523.c __be32 qid = cpu_to_be32(0); qid 448 drivers/net/wireless/ath/ar5523/ar5523.c &qid, sizeof(qid), 0); qid 476 drivers/net/wireless/ath/ar5523/ar5523.c qinfo.qid = cpu_to_be32(0); qid 284 drivers/net/wireless/ath/ar5523/ar5523_hw.h __be32 qid; qid 380 drivers/net/wireless/ath/ath9k/ar9003_mac.c ts->qid = MS(ads->ds_info, AR_TxQcuNum); qid 113 drivers/net/wireless/ath/ath9k/mac.h u8 qid; qid 2667 drivers/net/wireless/ath/ath9k/xmit.c if (ts.qid == sc->beacon.beaconq) { qid 2680 drivers/net/wireless/ath/ath9k/xmit.c txq = &sc->tx.txq[ts.qid]; qid 870 drivers/net/wireless/ath/wil6210/wmi.c data->info.qid, data->info.mid, data->info.cid); qid 1438 drivers/net/wireless/ath/wil6210/wmi.c d_len, data->info.qid, data->info.mid, data->info.cid); qid 2442 drivers/net/wireless/ath/wil6210/wmi.h u8 qid; qid 1186 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c int ret, qid; qid 1196 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c qid = le16_to_cpu(rsp->queue_number); qid 1199 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (qid >= ARRAY_SIZE(trans_pcie->txq)) { qid 1200 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c WARN_ONCE(1, "queue index %d unsupported", qid); qid 1205 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c if (test_and_set_bit(qid, trans_pcie->queue_used)) { qid 1206 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c WARN_ONCE(1, "queue %d already used", qid); qid 1211 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c txq->id = qid; qid 1212 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c trans_pcie->txq[qid] = txq; qid 1219 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c IWL_DEBUG_TX_QUEUES(trans, "Activate queue %d\n", qid); qid 1222 drivers/net/wireless/intel/iwlwifi/pcie/tx-gen2.c return qid; qid 139 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_tx_cleanup(struct mt76_dev *dev, enum mt76_txq_id qid, bool flush) qid 141 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_sw_queue *sq = &dev->q_tx[qid]; qid 160 drivers/net/wireless/mediatek/mt76/dma.c n_swq_queued[entry.qid]++; qid 166 drivers/net/wireless/mediatek/mt76/dma.c dev->drv->tx_complete_skb(dev, qid, &entry); qid 192 drivers/net/wireless/mediatek/mt76/dma.c qid < IEEE80211_NUM_ACS && q->queued < q->ndesc - 8; qid 202 drivers/net/wireless/mediatek/mt76/dma.c ieee80211_wake_queue(dev->hw, qid); qid 257 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_tx_queue_skb_raw(struct mt76_dev *dev, enum mt76_txq_id qid, qid 260 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_queue *q = dev->q_tx[qid].q; qid 281 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid, qid 285 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_queue *q = dev->q_tx[qid].q; qid 333 drivers/net/wireless/mediatek/mt76/dma.c ret = dev->drv->tx_prepare_skb(dev, txwi, qid, wcid, sta, &tx_info); qid 355 drivers/net/wireless/mediatek/mt76/dma.c dev->drv->tx_complete_skb(dev, qid, &e); qid 425 drivers/net/wireless/mediatek/mt76/dma.c mt76_dma_rx_reset(struct mt76_dev *dev, enum mt76_rxq_id qid) qid 427 drivers/net/wireless/mediatek/mt76/dma.c struct mt76_queue *q = &dev->q_rx[qid]; qid 530 drivers/net/wireless/mediatek/mt76/dma.c int qid, done = 0, cur; qid 533 drivers/net/wireless/mediatek/mt76/dma.c qid = napi - dev->napi; qid 538 drivers/net/wireless/mediatek/mt76/dma.c cur = mt76_dma_rx_process(dev, &dev->q_rx[qid], budget - done); qid 539 drivers/net/wireless/mediatek/mt76/dma.c mt76_rx_poll_complete(dev, qid, napi); qid 547 drivers/net/wireless/mediatek/mt76/dma.c dev->drv->rx_poll_complete(dev, qid); qid 95 drivers/net/wireless/mediatek/mt76/mt76.h enum mt76_txq_id qid; qid 159 drivers/net/wireless/mediatek/mt76/mt76.h int (*tx_queue_skb)(struct mt76_dev *dev, enum mt76_txq_id qid, qid 163 drivers/net/wireless/mediatek/mt76/mt76.h int (*tx_queue_skb_raw)(struct mt76_dev *dev, enum mt76_txq_id qid, qid 169 drivers/net/wireless/mediatek/mt76/mt76.h void (*rx_reset)(struct mt76_dev *dev, enum mt76_rxq_id qid); qid 171 drivers/net/wireless/mediatek/mt76/mt76.h void (*tx_cleanup)(struct mt76_dev *dev, enum mt76_txq_id qid, qid 295 drivers/net/wireless/mediatek/mt76/mt76.h enum mt76_txq_id qid, struct mt76_wcid *wcid, qid 299 drivers/net/wireless/mediatek/mt76/mt76.h void (*tx_complete_skb)(struct mt76_dev *dev, enum mt76_txq_id qid, qid 711 drivers/net/wireless/mediatek/mt76/mt76.h void mt76_txq_schedule(struct mt76_dev *dev, enum mt76_txq_id qid); qid 792 drivers/net/wireless/mediatek/mt76/mt76.h static inline u8 q2ep(u8 qid) qid 795 drivers/net/wireless/mediatek/mt76/mt76.h return qid + 1; qid 779 drivers/net/wireless/mediatek/mt76/mt7603/mac.c struct sk_buff *skb, enum mt76_txq_id qid, qid 788 drivers/net/wireless/mediatek/mt76/mt7603/mac.c struct mt76_queue *q = dev->mt76.q_tx[qid].q; qid 803 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (vif_idx && qid >= MT_TXQ_BEACON) qid 877 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (qid >= MT_TXQ_BEACON) qid 908 drivers/net/wireless/mediatek/mt76/mt7603/mac.c enum mt76_txq_id qid, struct mt76_wcid *wcid, qid 940 drivers/net/wireless/mediatek/mt76/mt7603/mac.c mt7603_mac_write_txwi(dev, txwi_ptr, tx_info->skb, qid, wcid, qid 1153 drivers/net/wireless/mediatek/mt76/mt7603/mac.c void mt7603_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid, qid 1164 drivers/net/wireless/mediatek/mt76/mt7603/mac.c if (qid < 4) qid 228 drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h enum mt76_txq_id qid, struct mt76_wcid *wcid, qid 232 drivers/net/wireless/mediatek/mt76/mt7603/mt7603.h void mt7603_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid, qid 222 drivers/net/wireless/mediatek/mt76/mt7615/mac.c void mt7615_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid, qid 763 drivers/net/wireless/mediatek/mt76/mt7615/mac.c enum mt76_txq_id qid, struct mt76_wcid *wcid, qid 56 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c enum mt76_txq_id qid; qid 108 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c qid = MT_TXQ_MCU; qid 110 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c qid = MT_TXQ_FWDL; qid 112 drivers/net/wireless/mediatek/mt76/mt7615/mcu.c return mt76_tx_queue_skb_raw(dev, qid, skb, 0); qid 254 drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h enum mt76_txq_id qid, struct mt76_wcid *wcid, qid 258 drivers/net/wireless/mediatek/mt76/mt7615/mt7615.h void mt7615_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid, qid 177 drivers/net/wireless/mediatek/mt76/mt76x02.h enum mt76_txq_id qid, struct mt76_wcid *wcid, qid 826 drivers/net/wireless/mediatek/mt76/mt76x02_mac.c void mt76x02_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid, qid 190 drivers/net/wireless/mediatek/mt76/mt76x02_mac.h void mt76x02_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid, qid 139 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c enum mt76_txq_id qid, struct mt76_wcid *wcid, qid 149 drivers/net/wireless/mediatek/mt76/mt76x02_txrx.c if (qid == MT_TXQ_PSD && wcid && wcid->idx < 128) qid 18 drivers/net/wireless/mediatek/mt76/mt76x02_usb.h enum mt76_txq_id qid, struct mt76_wcid *wcid, qid 21 drivers/net/wireless/mediatek/mt76/mt76x02_usb.h void mt76x02u_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid, qid 18 drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c void mt76x02u_tx_complete_skb(struct mt76_dev *mdev, enum mt76_txq_id qid, qid 64 drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c enum mt76_txq_id qid, struct mt76_wcid *wcid, qid 69 drivers/net/wireless/mediatek/mt76/mt76x02_usb_core.c int pid, len = tx_info->skb->len, ep = q2ep(mdev->q_tx[qid].q->hw_idx); qid 489 drivers/net/wireless/mediatek/mt76/mt76x02_util.c u8 cw_min = 5, cw_max = 10, qid; qid 492 drivers/net/wireless/mediatek/mt76/mt76x02_util.c qid = dev->mt76.q_tx[queue].q->hw_idx; qid 503 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76_wr(dev, MT_EDCA_CFG_AC(qid), val); qid 505 drivers/net/wireless/mediatek/mt76/mt76x02_util.c val = mt76_rr(dev, MT_WMM_TXOP(qid)); qid 506 drivers/net/wireless/mediatek/mt76/mt76x02_util.c val &= ~(MT_WMM_TXOP_MASK << MT_WMM_TXOP_SHIFT(qid)); qid 507 drivers/net/wireless/mediatek/mt76/mt76x02_util.c val |= params->txop << MT_WMM_TXOP_SHIFT(qid); qid 508 drivers/net/wireless/mediatek/mt76/mt76x02_util.c mt76_wr(dev, MT_WMM_TXOP(qid), val); qid 511 drivers/net/wireless/mediatek/mt76/mt76x02_util.c val &= ~(MT_WMM_AIFSN_MASK << MT_WMM_AIFSN_SHIFT(qid)); qid 512 drivers/net/wireless/mediatek/mt76/mt76x02_util.c val |= params->aifs << MT_WMM_AIFSN_SHIFT(qid); qid 516 drivers/net/wireless/mediatek/mt76/mt76x02_util.c val &= ~(MT_WMM_CWMIN_MASK << MT_WMM_CWMIN_SHIFT(qid)); qid 517 drivers/net/wireless/mediatek/mt76/mt76x02_util.c val |= cw_min << MT_WMM_CWMIN_SHIFT(qid); qid 521 drivers/net/wireless/mediatek/mt76/mt76x02_util.c val &= ~(MT_WMM_CWMAX_MASK << MT_WMM_CWMAX_SHIFT(qid)); qid 522 drivers/net/wireless/mediatek/mt76/mt76x02_util.c val |= cw_max << MT_WMM_CWMAX_SHIFT(qid); qid 254 drivers/net/wireless/mediatek/mt76/tx.c int qid = skb_get_queue_mapping(skb); qid 256 drivers/net/wireless/mediatek/mt76/tx.c if (WARN_ON(qid >= MT_TXQ_PSD)) { qid 257 drivers/net/wireless/mediatek/mt76/tx.c qid = MT_TXQ_BE; qid 258 drivers/net/wireless/mediatek/mt76/tx.c skb_set_queue_mapping(skb, qid); qid 278 drivers/net/wireless/mediatek/mt76/tx.c q = dev->q_tx[qid].q; qid 281 drivers/net/wireless/mediatek/mt76/tx.c dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, sta); qid 384 drivers/net/wireless/mediatek/mt76/tx.c enum mt76_txq_id qid = mt76_txq_get_qid(txq); qid 419 drivers/net/wireless/mediatek/mt76/tx.c idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, txq->sta); qid 453 drivers/net/wireless/mediatek/mt76/tx.c idx = dev->queue_ops->tx_queue_skb(dev, qid, skb, wcid, qid 462 drivers/net/wireless/mediatek/mt76/tx.c hwq->entry[idx].qid = sq - dev->q_tx; qid 473 drivers/net/wireless/mediatek/mt76/tx.c mt76_txq_schedule_list(struct mt76_dev *dev, enum mt76_txq_id qid) qid 475 drivers/net/wireless/mediatek/mt76/tx.c struct mt76_sw_queue *sq = &dev->q_tx[qid]; qid 494 drivers/net/wireless/mediatek/mt76/tx.c txq = ieee80211_next_txq(dev->hw, qid); qid 526 drivers/net/wireless/mediatek/mt76/tx.c void mt76_txq_schedule(struct mt76_dev *dev, enum mt76_txq_id qid) qid 528 drivers/net/wireless/mediatek/mt76/tx.c struct mt76_sw_queue *sq = &dev->q_tx[qid]; qid 531 drivers/net/wireless/mediatek/mt76/tx.c if (qid >= 4) qid 540 drivers/net/wireless/mediatek/mt76/tx.c ieee80211_txq_schedule_start(dev->hw, qid); qid 541 drivers/net/wireless/mediatek/mt76/tx.c len = mt76_txq_schedule_list(dev, qid); qid 542 drivers/net/wireless/mediatek/mt76/tx.c ieee80211_txq_schedule_end(dev->hw, qid); qid 768 drivers/net/wireless/mediatek/mt76/usb.c mt76u_tx_queue_skb(struct mt76_dev *dev, enum mt76_txq_id qid, qid 772 drivers/net/wireless/mediatek/mt76/usb.c struct mt76_queue *q = dev->q_tx[qid].q; qid 783 drivers/net/wireless/mediatek/mt76/usb.c err = dev->drv->tx_prepare_skb(dev, NULL, qid, wcid, sta, &tx_info); qid 341 drivers/net/wireless/mediatek/mt7601u/dma.c static u8 q2ep(u8 qid) qid 344 drivers/net/wireless/mediatek/mt7601u/dma.c return qid + 1; qid 29 drivers/net/wireless/mediatek/mt7601u/tx.c int qid = skb_get_queue_mapping(skb); qid 31 drivers/net/wireless/mediatek/mt7601u/tx.c if (WARN_ON(qid >= MT_TXQ_PSD)) { qid 32 drivers/net/wireless/mediatek/mt7601u/tx.c qid = MT_TXQ_BE; qid 33 drivers/net/wireless/mediatek/mt7601u/tx.c skb_set_queue_mapping(skb, qid); qid 36 drivers/net/wireless/mediatek/mt7601u/tx.c return q2hwq(qid); qid 632 drivers/net/wireless/ralink/rt2x00/rt2400pci.c switch (queue->qid) { qid 655 drivers/net/wireless/ralink/rt2x00/rt2400pci.c switch (queue->qid) { qid 681 drivers/net/wireless/ralink/rt2x00/rt2400pci.c switch (queue->qid) { qid 719 drivers/net/wireless/ralink/rt2x00/rt2400pci.c if (entry->queue->qid == QID_RX) { qid 737 drivers/net/wireless/ralink/rt2x00/rt2400pci.c if (entry->queue->qid == QID_RX) { qid 1758 drivers/net/wireless/ralink/rt2x00/rt2400pci.c switch (queue->qid) { qid 721 drivers/net/wireless/ralink/rt2x00/rt2500pci.c switch (queue->qid) { qid 744 drivers/net/wireless/ralink/rt2x00/rt2500pci.c switch (queue->qid) { qid 770 drivers/net/wireless/ralink/rt2x00/rt2500pci.c switch (queue->qid) { qid 808 drivers/net/wireless/ralink/rt2x00/rt2500pci.c if (entry->queue->qid == QID_RX) { qid 826 drivers/net/wireless/ralink/rt2x00/rt2500pci.c if (entry->queue->qid == QID_RX) { qid 2057 drivers/net/wireless/ralink/rt2x00/rt2500pci.c switch (queue->qid) { qid 722 drivers/net/wireless/ralink/rt2x00/rt2500usb.c switch (queue->qid) { qid 745 drivers/net/wireless/ralink/rt2x00/rt2500usb.c switch (queue->qid) { qid 1848 drivers/net/wireless/ralink/rt2x00/rt2500usb.c switch (queue->qid) { qid 816 drivers/net/wireless/ralink/rt2x00/rt2800lib.c rt2x00_set_field32(&word, TXWI_W1_PACKETID_QUEUE, entry->queue->qid); qid 989 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry->queue->qid, entry->entry_idx); qid 1103 drivers/net/wireless/ralink/rt2x00/rt2800lib.c u8 qid; qid 1111 drivers/net/wireless/ralink/rt2x00/rt2800lib.c qid = rt2x00_get_field32(reg, TX_STA_FIFO_PID_QUEUE); qid 1112 drivers/net/wireless/ralink/rt2x00/rt2800lib.c queue = rt2x00queue_get_tx_queue(rt2x00dev, qid); qid 1116 drivers/net/wireless/ralink/rt2x00/rt2800lib.c qid); qid 1125 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry->entry_idx, qid); qid 1153 drivers/net/wireless/ralink/rt2x00/rt2800lib.c entry->entry_idx, entry->queue->qid); qid 1241 drivers/net/wireless/ralink/rt2x00/rt2800lib.c switch (queue->qid) { qid 31 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c int idx, qid; qid 33 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c switch (queue->qid) { qid 38 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c qid = queue->qid; qid 39 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c idx = rt2x00mmio_register_read(rt2x00dev, TX_DTX_IDX(qid)); qid 425 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c switch (queue->qid) { qid 456 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c switch (queue->qid) { qid 463 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c rt2x00mmio_register_write(rt2x00dev, TX_CTX_IDX(queue->qid), qid 485 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c switch (queue->qid) { qid 528 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c switch (queue->qid) { qid 567 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c switch (queue->qid) { qid 612 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c if (entry->queue->qid == QID_RX) { qid 631 drivers/net/wireless/ralink/rt2x00/rt2800mmio.c if (entry->queue->qid == QID_RX) { qid 51 drivers/net/wireless/ralink/rt2x00/rt2800usb.c switch (queue->qid) { qid 74 drivers/net/wireless/ralink/rt2x00/rt2800usb.c switch (queue->qid) { qid 395 drivers/net/wireless/ralink/rt2x00/rt2800usb.c if (entry->queue->qid == QID_BEACON) qid 719 drivers/net/wireless/ralink/rt2x00/rt2800usb.c switch (queue->qid) { qid 171 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c dump_hdr->queue_index = entry->queue->qid; qid 324 drivers/net/wireless/ralink/rt2x00/rt2x00debug.c queue->qid, (unsigned int)queue->flags, qid 95 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c enum data_queue_qid qid = skb_get_queue_mapping(skb); qid 112 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c qid = QID_ATIM; qid 114 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c queue = rt2x00queue_get_tx_queue(rt2x00dev, qid); qid 118 drivers/net/wireless/ralink/rt2x00/rt2x00mac.c "Please file bug report to %s\n", qid, DRV_PROJECT); qid 502 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c entry->queue->qid, DRV_PROJECT); qid 667 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c queue->qid); qid 679 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c queue->qid, DRV_PROJECT); qid 894 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c switch (queue->qid) { qid 903 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c ieee80211_stop_queue(queue->rt2x00dev->hw, queue->qid); qid 927 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c switch (queue->qid) { qid 936 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c ieee80211_wake_queue(queue->rt2x00dev->hw, queue->qid); qid 990 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c (queue->qid == QID_AC_VO) || qid 991 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c (queue->qid == QID_AC_VI) || qid 992 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c (queue->qid == QID_AC_BE) || qid 993 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c (queue->qid == QID_AC_BK); qid 1020 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c queue->qid); qid 1214 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c struct data_queue *queue, enum data_queue_qid qid) qid 1221 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c queue->qid = qid; qid 1235 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c enum data_queue_qid qid; qid 1271 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c qid = QID_AC_VO; qid 1273 drivers/net/wireless/ralink/rt2x00/rt2x00queue.c rt2x00queue_init(rt2x00dev, queue, qid++); qid 455 drivers/net/wireless/ralink/rt2x00/rt2x00queue.h enum data_queue_qid qid; qid 426 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c switch (queue->qid) { qid 466 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if ((entry->queue->qid == QID_BEACON) && qid 485 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c switch (queue->qid) { qid 526 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c queue->qid); qid 571 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (entry->queue->qid == QID_RX) qid 584 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (queue->qid == QID_RX) { qid 665 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (queue->qid != QID_BEACON || qid 700 drivers/net/wireless/ralink/rt2x00/rt2x00usb.c if (queue->qid != QID_BEACON || qid 1046 drivers/net/wireless/ralink/rt2x00/rt61pci.c switch (queue->qid) { qid 1069 drivers/net/wireless/ralink/rt2x00/rt61pci.c switch (queue->qid) { qid 1100 drivers/net/wireless/ralink/rt2x00/rt61pci.c switch (queue->qid) { qid 1292 drivers/net/wireless/ralink/rt2x00/rt61pci.c if (entry->queue->qid == QID_RX) { qid 1310 drivers/net/wireless/ralink/rt2x00/rt61pci.c if (entry->queue->qid == QID_RX) { qid 1787 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid); qid 1812 drivers/net/wireless/ralink/rt2x00/rt61pci.c rt2x00_set_field32(&word, TXD_W5_PID_TYPE, entry->queue->qid); qid 1819 drivers/net/wireless/ralink/rt2x00/rt61pci.c if (entry->queue->qid != QID_BEACON) { qid 1865 drivers/net/wireless/ralink/rt2x00/rt61pci.c skbdesc->desc_len = (entry->queue->qid == QID_BEACON) ? TXINFO_SIZE : qid 2930 drivers/net/wireless/ralink/rt2x00/rt61pci.c switch (queue->qid) { qid 1018 drivers/net/wireless/ralink/rt2x00/rt73usb.c switch (queue->qid) { qid 1041 drivers/net/wireless/ralink/rt2x00/rt73usb.c switch (queue->qid) { qid 1481 drivers/net/wireless/ralink/rt2x00/rt73usb.c rt2x00_set_field32(&word, TXD_W1_HOST_Q_ID, entry->queue->qid); qid 2350 drivers/net/wireless/ralink/rt2x00/rt73usb.c switch (queue->qid) { qid 304 drivers/net/wireless/ti/wl1251/acx.h u8 qid; qid 217 drivers/net/wireless/ti/wl1251/init.c static int wl1251_hw_init_txq_fill(u8 qid, qid 221 drivers/net/wireless/ti/wl1251/init.c config->qid = qid; qid 223 drivers/net/wireless/ti/wl1251/init.c switch (qid) { qid 249 drivers/net/wireless/ti/wl1251/init.c wl1251_error("Invalid TX queue id: %d", qid); qid 480 drivers/nvme/host/core.c struct nvme_command *cmd, blk_mq_req_flags_t flags, int qid) qid 485 drivers/nvme/host/core.c if (qid == NVME_QID_ANY) { qid 489 drivers/nvme/host/core.c qid ? qid - 1 : 0); qid 820 drivers/nvme/host/core.c unsigned timeout, int qid, int at_head, qid 826 drivers/nvme/host/core.c req = nvme_alloc_request(q, cmd, flags, qid); qid 309 drivers/nvme/host/fabrics.c case (offsetof(struct nvmf_connect_command, qid)): qid 312 drivers/nvme/host/fabrics.c inv_sqe, cmd->connect.qid); qid 377 drivers/nvme/host/fabrics.c cmd.connect.qid = 0; qid 437 drivers/nvme/host/fabrics.c int nvmf_connect_io_queue(struct nvme_ctrl *ctrl, u16 qid, bool poll) qid 447 drivers/nvme/host/fabrics.c cmd.connect.qid = cpu_to_le16(qid); qid 463 drivers/nvme/host/fabrics.c data, sizeof(*data), 0, qid, 1, qid 169 drivers/nvme/host/fabrics.h int nvmf_connect_io_queue(struct nvme_ctrl *ctrl, u16 qid, bool poll); qid 1327 drivers/nvme/host/fc.c conn_rqst->connect_cmd.qid = cpu_to_be16(queue->qnum); qid 484 drivers/nvme/host/nvme.h struct nvme_command *cmd, blk_mq_req_flags_t flags, int qid); qid 492 drivers/nvme/host/nvme.h unsigned timeout, int qid, int at_head, qid 144 drivers/nvme/host/pci.c static inline unsigned int sq_idx(unsigned int qid, u32 stride) qid 146 drivers/nvme/host/pci.c return qid * 2 * stride; qid 149 drivers/nvme/host/pci.c static inline unsigned int cq_idx(unsigned int qid, u32 stride) qid 151 drivers/nvme/host/pci.c return (qid * 2 + 1) * stride; qid 179 drivers/nvme/host/pci.c u16 qid; qid 271 drivers/nvme/host/pci.c struct nvme_queue *nvmeq, int qid) qid 273 drivers/nvme/host/pci.c if (!dev->dbbuf_dbs || !qid) qid 276 drivers/nvme/host/pci.c nvmeq->dbbuf_sq_db = &dev->dbbuf_dbs[sq_idx(qid, dev->db_stride)]; qid 277 drivers/nvme/host/pci.c nvmeq->dbbuf_cq_db = &dev->dbbuf_dbs[cq_idx(qid, dev->db_stride)]; qid 278 drivers/nvme/host/pci.c nvmeq->dbbuf_sq_ei = &dev->dbbuf_eis[sq_idx(qid, dev->db_stride)]; qid 279 drivers/nvme/host/pci.c nvmeq->dbbuf_cq_ei = &dev->dbbuf_eis[cq_idx(qid, dev->db_stride)]; qid 517 drivers/nvme/host/pci.c if (!iod->nvmeq->qid) qid 810 drivers/nvme/host/pci.c if (iod->nvmeq->qid && qid 942 drivers/nvme/host/pci.c if (!nvmeq->qid) qid 944 drivers/nvme/host/pci.c return nvmeq->dev->tagset.tags[nvmeq->qid - 1]; qid 965 drivers/nvme/host/pci.c if (unlikely(nvmeq->qid == 0 && qid 1111 drivers/nvme/host/pci.c c.delete_queue.qid = cpu_to_le16(id); qid 1116 drivers/nvme/host/pci.c static int adapter_alloc_cq(struct nvme_dev *dev, u16 qid, qid 1132 drivers/nvme/host/pci.c c.create_cq.cqid = cpu_to_le16(qid); qid 1140 drivers/nvme/host/pci.c static int adapter_alloc_sq(struct nvme_dev *dev, u16 qid, qid 1162 drivers/nvme/host/pci.c c.create_sq.sqid = cpu_to_le16(qid); qid 1165 drivers/nvme/host/pci.c c.create_sq.cqid = cpu_to_le16(qid); qid 1267 drivers/nvme/host/pci.c req->tag, nvmeq->qid); qid 1284 drivers/nvme/host/pci.c req->tag, nvmeq->qid); qid 1299 drivers/nvme/host/pci.c if (!nvmeq->qid || iod->aborted) { qid 1302 drivers/nvme/host/pci.c req->tag, nvmeq->qid); qid 1319 drivers/nvme/host/pci.c cmd.abort.sqid = cpu_to_le16(nvmeq->qid); qid 1323 drivers/nvme/host/pci.c req->tag, nvmeq->qid); qid 1383 drivers/nvme/host/pci.c if (!nvmeq->qid && nvmeq->dev->ctrl.admin_q) qid 1452 drivers/nvme/host/pci.c int qid) qid 1456 drivers/nvme/host/pci.c if (qid && dev->cmb_use_sqes && (dev->cmbsz & NVME_CMBSZ_SQS)) { qid 1477 drivers/nvme/host/pci.c static int nvme_alloc_queue(struct nvme_dev *dev, int qid, int depth) qid 1479 drivers/nvme/host/pci.c struct nvme_queue *nvmeq = &dev->queues[qid]; qid 1481 drivers/nvme/host/pci.c if (dev->ctrl.queue_count > qid) qid 1484 drivers/nvme/host/pci.c nvmeq->sqes = qid ? dev->io_sqes : NVME_ADM_SQES; qid 1491 drivers/nvme/host/pci.c if (nvme_alloc_sq_cmds(dev, nvmeq, qid)) qid 1499 drivers/nvme/host/pci.c nvmeq->q_db = &dev->dbs[qid * 2 * dev->db_stride]; qid 1500 drivers/nvme/host/pci.c nvmeq->qid = qid; qid 1519 drivers/nvme/host/pci.c nvme_irq, nvmeq, "nvme%dq%d", nr, nvmeq->qid); qid 1522 drivers/nvme/host/pci.c NULL, nvmeq, "nvme%dq%d", nr, nvmeq->qid); qid 1526 drivers/nvme/host/pci.c static void nvme_init_queue(struct nvme_queue *nvmeq, u16 qid) qid 1534 drivers/nvme/host/pci.c nvmeq->q_db = &dev->dbs[qid * 2 * dev->db_stride]; qid 1536 drivers/nvme/host/pci.c nvme_dbbuf_init(dev, nvmeq, qid); qid 1541 drivers/nvme/host/pci.c static int nvme_create_queue(struct nvme_queue *nvmeq, int qid, bool polled) qid 1554 drivers/nvme/host/pci.c vector = dev->num_vecs == 1 ? 0 : qid; qid 1558 drivers/nvme/host/pci.c result = adapter_alloc_cq(dev, qid, nvmeq, vector); qid 1562 drivers/nvme/host/pci.c result = adapter_alloc_sq(dev, qid, nvmeq); qid 1569 drivers/nvme/host/pci.c nvme_init_queue(nvmeq, qid); qid 1582 drivers/nvme/host/pci.c adapter_delete_sq(dev, qid); qid 1584 drivers/nvme/host/pci.c adapter_delete_cq(dev, qid); qid 2224 drivers/nvme/host/pci.c cmd.delete_queue.qid = cpu_to_le16(nvmeq->qid); qid 1603 drivers/nvme/host/rdma.c priv.qid = cpu_to_le16(nvme_rdma_queue_idx(queue)); qid 1608 drivers/nvme/host/rdma.c if (priv.qid == 0) { qid 1136 drivers/nvme/host/tcp.c static void nvme_tcp_free_queue(struct nvme_ctrl *nctrl, int qid) qid 1139 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[qid]; qid 1250 drivers/nvme/host/tcp.c int qid, size_t queue_size) qid 1253 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[qid]; qid 1263 drivers/nvme/host/tcp.c if (qid > 0) qid 1323 drivers/nvme/host/tcp.c if (!qid) qid 1326 drivers/nvme/host/tcp.c n = (qid - 1) % num_online_cpus(); qid 1341 drivers/nvme/host/tcp.c qid, ret); qid 1352 drivers/nvme/host/tcp.c "failed to allocate queue %d crypto\n", qid); qid 1431 drivers/nvme/host/tcp.c static void nvme_tcp_stop_queue(struct nvme_ctrl *nctrl, int qid) qid 1434 drivers/nvme/host/tcp.c struct nvme_tcp_queue *queue = &ctrl->queues[qid]; qid 199 drivers/nvme/host/trace.c u16 qid = get_unaligned_le16(spc + 2); qid 205 drivers/nvme/host/trace.c recfmt, qid, sqsize, cattr, kato); qid 26 drivers/nvme/host/trace.h #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ qid 29 drivers/nvme/host/trace.h ((qid) ? \ qid 53 drivers/nvme/host/trace.h __field(int, qid) qid 64 drivers/nvme/host/trace.h __entry->qid = nvme_req_qid(req); qid 77 drivers/nvme/host/trace.h __entry->qid, __entry->cid, __entry->nsid, qid 79 drivers/nvme/host/trace.h show_opcode_name(__entry->qid, __entry->opcode, qid 81 drivers/nvme/host/trace.h parse_nvme_cmd(__entry->qid, __entry->opcode, qid 91 drivers/nvme/host/trace.h __field(int, qid) qid 100 drivers/nvme/host/trace.h __entry->qid = nvme_req_qid(req); qid 110 drivers/nvme/host/trace.h __entry->qid, __entry->cid, __entry->result, qid 150 drivers/nvme/host/trace.h __field(int, qid) qid 157 drivers/nvme/host/trace.h __entry->qid = nvme_req_qid(req); qid 163 drivers/nvme/host/trace.h __entry->qid, __entry->sq_head, __entry->sq_tail qid 891 drivers/nvme/target/admin-cmd.c req->sq->qid); qid 692 drivers/nvme/target/core.c new_error_slot->sqid = cpu_to_le16(req->sq->qid); qid 708 drivers/nvme/target/core.c req->cqe->sq_id = cpu_to_le16(req->sq->qid); qid 729 drivers/nvme/target/core.c u16 qid, u16 size) qid 731 drivers/nvme/target/core.c cq->qid = qid; qid 734 drivers/nvme/target/core.c ctrl->cqs[qid] = cq; qid 738 drivers/nvme/target/core.c u16 qid, u16 size) qid 741 drivers/nvme/target/core.c sq->qid = qid; qid 744 drivers/nvme/target/core.c ctrl->sqs[qid] = sq; qid 898 drivers/nvme/target/core.c else if (likely(req->sq->qid != 0)) qid 954 drivers/nvme/target/core.c if (req->sq->qid && p2p_dev) { qid 1133 drivers/nvme/target/core.c cmd->common.opcode, req->sq->qid); qid 1139 drivers/nvme/target/core.c cmd->common.opcode, req->sq->qid); qid 105 drivers/nvme/target/fabrics-cmd.c u16 qid = le16_to_cpu(c->qid); qid 124 drivers/nvme/target/fabrics-cmd.c nvmet_cq_setup(ctrl, req->cq, qid, sqsize + 1); qid 125 drivers/nvme/target/fabrics-cmd.c nvmet_sq_setup(ctrl, req->sq, qid, sqsize + 1); qid 136 drivers/nvme/target/fabrics-cmd.c qid, ctrl->cntlid, ret); qid 216 drivers/nvme/target/fabrics-cmd.c u16 qid = le16_to_cpu(c->qid); qid 245 drivers/nvme/target/fabrics-cmd.c if (unlikely(qid > ctrl->subsys->max_qid)) { qid 246 drivers/nvme/target/fabrics-cmd.c pr_warn("invalid queue id (%d)\n", qid); qid 248 drivers/nvme/target/fabrics-cmd.c req->cqe->result.u32 = IPO_IATTR_CONNECT_SQE(qid); qid 259 drivers/nvme/target/fabrics-cmd.c pr_debug("adding queue %d to ctrl %d.\n", qid, ctrl->cntlid); qid 290 drivers/nvme/target/fabrics-cmd.c if (cmd->connect.qid == 0) qid 119 drivers/nvme/target/fc.c u16 qid; qid 178 drivers/nvme/target/fc.c nvmet_fc_makeconnid(struct nvmet_fc_tgt_assoc *assoc, u16 qid) qid 180 drivers/nvme/target/fc.c return (assoc->association_id | qid); qid 498 drivers/nvme/target/fc.c fcpreq->hwqid = queue->qid ? qid 499 drivers/nvme/target/fc.c ((queue->qid - 1) % tgtport->ops->max_hw_queues) : 0; qid 582 drivers/nvme/target/fc.c u16 qid, u16 sqsize) qid 588 drivers/nvme/target/fc.c if (qid > NVMET_NR_QUEUES) qid 600 drivers/nvme/target/fc.c assoc->a_id, qid); qid 604 drivers/nvme/target/fc.c queue->qid = qid; qid 623 drivers/nvme/target/fc.c WARN_ON(assoc->queues[qid]); qid 625 drivers/nvme/target/fc.c assoc->queues[qid] = queue; qid 649 drivers/nvme/target/fc.c queue->assoc->queues[queue->qid] = NULL; qid 758 drivers/nvme/target/fc.c u16 qid = nvmet_fc_getqueueid(connection_id); qid 761 drivers/nvme/target/fc.c if (qid > NVMET_NR_QUEUES) qid 767 drivers/nvme/target/fc.c queue = assoc->queues[qid]; qid 1451 drivers/nvme/target/fc.c be16_to_cpu(rqst->connect_cmd.qid), qid 1491 drivers/nvme/target/fc.c be16_to_cpu(rqst->connect_cmd.qid))); qid 2106 drivers/nvme/target/fc.c cqe->sq_id = cpu_to_le16(fod->queue->qid); qid 339 drivers/nvme/target/io-cmd-bdev.c req->sq->qid); qid 391 drivers/nvme/target/io-cmd-file.c cmd->common.opcode, req->sq->qid); qid 92 drivers/nvme/target/nvmet.h u16 qid; qid 99 drivers/nvme/target/nvmet.h u16 qid; qid 385 drivers/nvme/target/nvmet.h void nvmet_cq_setup(struct nvmet_ctrl *ctrl, struct nvmet_cq *cq, u16 qid, qid 387 drivers/nvme/target/nvmet.h void nvmet_sq_setup(struct nvmet_ctrl *ctrl, struct nvmet_sq *sq, u16 qid, qid 1105 drivers/nvme/target/rdma.c queue->host_qid = le16_to_cpu(req->qid); qid 1689 drivers/nvme/target/tcp.c if (sq->qid == 0) { qid 140 drivers/nvme/target/trace.c u16 qid = get_unaligned_le16(spc + 2); qid 146 drivers/nvme/target/trace.c recfmt, qid, sqsize, cattr, kato); qid 28 drivers/nvme/target/trace.h #define parse_nvme_cmd(qid, opcode, fctype, cdw10) \ qid 31 drivers/nvme/target/trace.h (qid ? \ qid 55 drivers/nvme/target/trace.h if ((init && req->sq->qid) || (!init && req->cq->qid)) { qid 72 drivers/nvme/target/trace.h __field(int, qid) qid 85 drivers/nvme/target/trace.h __entry->qid = req->sq->qid; qid 99 drivers/nvme/target/trace.h __entry->qid, __entry->cid, __entry->nsid, qid 101 drivers/nvme/target/trace.h show_opcode_name(__entry->qid, __entry->opcode, qid 103 drivers/nvme/target/trace.h parse_nvme_cmd(__entry->qid, __entry->opcode, qid 113 drivers/nvme/target/trace.h __field(int, qid) qid 120 drivers/nvme/target/trace.h __entry->qid = req->cq->qid; qid 129 drivers/nvme/target/trace.h __entry->qid, __entry->cid, __entry->result, __entry->status) qid 295 drivers/s390/crypto/ap_bus.c static int ap_query_queue(ap_qid_t qid, int *queue_depth, int *device_type, qid 302 drivers/s390/crypto/ap_bus.c if (!ap_test_config_card_id(AP_QID_CARD(qid))) qid 305 drivers/s390/crypto/ap_bus.c status = ap_test_queue(qid, ap_apft_available(), &info); qid 638 drivers/s390/crypto/ap_bus.c AP_QID_CARD(to_ap_queue(dev)->qid) == (int)(long) data) qid 711 drivers/s390/crypto/ap_bus.c card = AP_QID_CARD(to_ap_queue(dev)->qid); qid 712 drivers/s390/crypto/ap_bus.c queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); qid 787 drivers/s390/crypto/ap_bus.c card = AP_QID_CARD(to_ap_queue(dev)->qid); qid 788 drivers/s390/crypto/ap_bus.c queue = AP_QID_QUEUE(to_ap_queue(dev)->qid); qid 1316 drivers/s390/crypto/ap_bus.c static int ap_get_compatible_type(ap_qid_t qid, int rawtype, unsigned int func) qid 1337 drivers/s390/crypto/ap_bus.c status = ap_qact(qid, 0, &apinfo); qid 1345 drivers/s390/crypto/ap_bus.c AP_QID_CARD(qid), AP_QID_QUEUE(qid), rawtype); qid 1348 drivers/s390/crypto/ap_bus.c AP_QID_CARD(qid), AP_QID_QUEUE(qid), rawtype, comp_type); qid 1367 drivers/s390/crypto/ap_bus.c return is_queue_dev(dev) && to_ap_queue(dev)->qid == (int)(long) data; qid 1377 drivers/s390/crypto/ap_bus.c && AP_QID_QUEUE(to_ap_queue(dev)->qid) == (int)(long) data; qid 1386 drivers/s390/crypto/ap_bus.c ap_qid_t qid; qid 1418 drivers/s390/crypto/ap_bus.c qid = AP_MKQID(id, dom); qid 1419 drivers/s390/crypto/ap_bus.c if (ap_query_queue(qid, &depth, &type, &func) == 0) qid 1455 drivers/s390/crypto/ap_bus.c qid = AP_MKQID(id, dom); qid 1457 drivers/s390/crypto/ap_bus.c (void *)(long) qid, qid 1471 drivers/s390/crypto/ap_bus.c rc = ap_query_queue(qid, &depth, &type, &func); qid 1493 drivers/s390/crypto/ap_bus.c comp_type = ap_get_compatible_type(qid, type, func); qid 1515 drivers/s390/crypto/ap_bus.c aq = ap_queue_create(qid, comp_type); qid 176 drivers/s390/crypto/ap_bus.h ap_qid_t qid; /* AP queue id. */ qid 37 drivers/s390/crypto/ap_queue.c status = ap_aqic(aq->qid, qirqctrl, ind); qid 47 drivers/s390/crypto/ap_queue.c AP_QID_CARD(aq->qid), qid 48 drivers/s390/crypto/ap_queue.c AP_QID_QUEUE(aq->qid)); qid 71 drivers/s390/crypto/ap_queue.c __ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length, qid 75 drivers/s390/crypto/ap_queue.c qid |= 0x400000UL; qid 76 drivers/s390/crypto/ap_queue.c return ap_nqap(qid, psmid, msg, length); qid 79 drivers/s390/crypto/ap_queue.c int ap_send(ap_qid_t qid, unsigned long long psmid, void *msg, size_t length) qid 83 drivers/s390/crypto/ap_queue.c status = __ap_send(qid, psmid, msg, length, 0); qid 98 drivers/s390/crypto/ap_queue.c int ap_recv(ap_qid_t qid, unsigned long long *psmid, void *msg, size_t length) qid 104 drivers/s390/crypto/ap_queue.c status = ap_dqap(qid, psmid, msg, length); qid 139 drivers/s390/crypto/ap_queue.c status = ap_dqap(aq->qid, &aq->reply->psmid, qid 243 drivers/s390/crypto/ap_queue.c status = __ap_send(aq->qid, ap_msg->psmid, qid 298 drivers/s390/crypto/ap_queue.c status = ap_rapq(aq->qid); qid 332 drivers/s390/crypto/ap_queue.c status = ap_tapq(aq->qid, NULL); qid 370 drivers/s390/crypto/ap_queue.c status = ap_tapq(aq->qid, NULL); qid 569 drivers/s390/crypto/ap_queue.c AP_QID_CARD(aq->qid), AP_QID_QUEUE(aq->qid)); qid 630 drivers/s390/crypto/ap_queue.c struct ap_queue *ap_queue_create(ap_qid_t qid, int device_type) qid 640 drivers/s390/crypto/ap_queue.c aq->qid = qid; qid 769 drivers/s390/crypto/ap_queue.c ap_zapq(aq->qid); qid 60 drivers/s390/crypto/vfio_ap_drv.c q->apqn = to_ap_queue(&apdev->device)->qid; qid 440 drivers/s390/crypto/vfio_ap_ops.c ap_qid_t qid; qid 444 drivers/s390/crypto/vfio_ap_ops.c qid = AP_MKQID(*qres->apid, *qres->apqi); qid 445 drivers/s390/crypto/vfio_ap_ops.c if (qid == ap_queue->qid) qid 448 drivers/s390/crypto/vfio_ap_ops.c id = AP_QID_CARD(ap_queue->qid); qid 452 drivers/s390/crypto/vfio_ap_ops.c id = AP_QID_QUEUE(ap_queue->qid); qid 638 drivers/s390/crypto/zcrypt_api.c int qid = 0, rc = -ENODEV; qid 684 drivers/s390/crypto/zcrypt_api.c AP_QID_QUEUE(zq->queue->qid))) qid 702 drivers/s390/crypto/zcrypt_api.c qid = pref_zq->queue->qid; qid 711 drivers/s390/crypto/zcrypt_api.c AP_QID_CARD(qid), AP_QID_QUEUE(qid)); qid 722 drivers/s390/crypto/zcrypt_api.c int qid = 0, rc = -ENODEV; qid 768 drivers/s390/crypto/zcrypt_api.c AP_QID_QUEUE(zq->queue->qid))) qid 786 drivers/s390/crypto/zcrypt_api.c qid = pref_zq->queue->qid; qid 795 drivers/s390/crypto/zcrypt_api.c AP_QID_CARD(qid), AP_QID_QUEUE(qid)); qid 808 drivers/s390/crypto/zcrypt_api.c int qid = 0, rc = -ENODEV; qid 853 drivers/s390/crypto/zcrypt_api.c tdom != AP_QID_QUEUE(zq->queue->qid))) qid 857 drivers/s390/crypto/zcrypt_api.c AP_QID_QUEUE(zq->queue->qid))) qid 876 drivers/s390/crypto/zcrypt_api.c qid = pref_zq->queue->qid; qid 878 drivers/s390/crypto/zcrypt_api.c *domain = AP_QID_QUEUE(qid); qid 889 drivers/s390/crypto/zcrypt_api.c AP_QID_CARD(qid), AP_QID_QUEUE(qid)); qid 933 drivers/s390/crypto/zcrypt_api.c int qid = 0, rc = -ENODEV; qid 990 drivers/s390/crypto/zcrypt_api.c !is_desired_ep11_queue(zq->queue->qid, qid 995 drivers/s390/crypto/zcrypt_api.c AP_QID_QUEUE(zq->queue->qid))) qid 1013 drivers/s390/crypto/zcrypt_api.c qid = pref_zq->queue->qid; qid 1025 drivers/s390/crypto/zcrypt_api.c AP_QID_CARD(qid), AP_QID_QUEUE(qid)); qid 1037 drivers/s390/crypto/zcrypt_api.c int qid = 0, rc = -ENODEV; qid 1078 drivers/s390/crypto/zcrypt_api.c qid = pref_zq->queue->qid; qid 1088 drivers/s390/crypto/zcrypt_api.c AP_QID_CARD(qid), AP_QID_QUEUE(qid)); qid 1105 drivers/s390/crypto/zcrypt_api.c card = AP_QID_CARD(zq->queue->qid); qid 1108 drivers/s390/crypto/zcrypt_api.c queue = AP_QID_QUEUE(zq->queue->qid); qid 1112 drivers/s390/crypto/zcrypt_api.c stat->qid = zq->queue->qid; qid 1132 drivers/s390/crypto/zcrypt_api.c card = AP_QID_CARD(zq->queue->qid); qid 1133 drivers/s390/crypto/zcrypt_api.c queue = AP_QID_QUEUE(zq->queue->qid); qid 1137 drivers/s390/crypto/zcrypt_api.c stat->qid = zq->queue->qid; qid 1156 drivers/s390/crypto/zcrypt_api.c if (card == AP_QID_CARD(zq->queue->qid) && qid 1157 drivers/s390/crypto/zcrypt_api.c queue == AP_QID_QUEUE(zq->queue->qid)) { qid 1160 drivers/s390/crypto/zcrypt_api.c devstat->qid = zq->queue->qid; qid 1183 drivers/s390/crypto/zcrypt_api.c card = AP_QID_CARD(zq->queue->qid); qid 1184 drivers/s390/crypto/zcrypt_api.c if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index qid 1204 drivers/s390/crypto/zcrypt_api.c card = AP_QID_CARD(zq->queue->qid); qid 1205 drivers/s390/crypto/zcrypt_api.c if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index qid 1231 drivers/s390/crypto/zcrypt_api.c card = AP_QID_CARD(zq->queue->qid); qid 1232 drivers/s390/crypto/zcrypt_api.c if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index qid 1256 drivers/s390/crypto/zcrypt_api.c if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index) qid 1279 drivers/s390/crypto/zcrypt_api.c if (AP_QID_QUEUE(zq->queue->qid) != ap_domain_index) qid 1580 drivers/s390/crypto/zcrypt_ccamisc.c card = AP_QID_CARD(device_status[i].qid); qid 1581 drivers/s390/crypto/zcrypt_ccamisc.c dom = AP_QID_QUEUE(device_status[i].qid); qid 1612 drivers/s390/crypto/zcrypt_ccamisc.c card = AP_QID_CARD(device_status[i].qid); qid 1613 drivers/s390/crypto/zcrypt_ccamisc.c dom = AP_QID_QUEUE(device_status[i].qid); qid 1630 drivers/s390/crypto/zcrypt_ccamisc.c card = AP_QID_CARD(device_status[oi].qid); qid 1631 drivers/s390/crypto/zcrypt_ccamisc.c dom = AP_QID_QUEUE(device_status[oi].qid); qid 1699 drivers/s390/crypto/zcrypt_ccamisc.c card = AP_QID_CARD(device_status[i].qid); qid 1700 drivers/s390/crypto/zcrypt_ccamisc.c dom = AP_QID_QUEUE(device_status[i].qid); qid 97 drivers/s390/crypto/zcrypt_cex2c.c msg->cprbx.domain = AP_QID_QUEUE(aq->qid); qid 99 drivers/s390/crypto/zcrypt_cex2c.c rc = ap_send(aq->qid, 0x0102030405060708ULL, ap_msg.message, qid 107 drivers/s390/crypto/zcrypt_cex2c.c rc = ap_recv(aq->qid, &psmid, ap_msg.message, 4096); qid 223 drivers/s390/crypto/zcrypt_cex2c.c ap_rapq(aq->qid); qid 117 drivers/s390/crypto/zcrypt_cex4.c cca_get_info(AP_QID_CARD(zq->queue->qid), qid 118 drivers/s390/crypto/zcrypt_cex4.c AP_QID_QUEUE(zq->queue->qid), qid 84 drivers/s390/crypto/zcrypt_error.h int card = AP_QID_CARD(zq->queue->qid); qid 85 drivers/s390/crypto/zcrypt_error.h int queue = AP_QID_QUEUE(zq->queue->qid); qid 360 drivers/s390/crypto/zcrypt_msgtype50.c AP_QID_CARD(zq->queue->qid), qid 361 drivers/s390/crypto/zcrypt_msgtype50.c AP_QID_QUEUE(zq->queue->qid)); qid 364 drivers/s390/crypto/zcrypt_msgtype50.c AP_QID_CARD(zq->queue->qid), qid 365 drivers/s390/crypto/zcrypt_msgtype50.c AP_QID_QUEUE(zq->queue->qid), qid 397 drivers/s390/crypto/zcrypt_msgtype50.c AP_QID_CARD(zq->queue->qid), qid 398 drivers/s390/crypto/zcrypt_msgtype50.c AP_QID_QUEUE(zq->queue->qid)); qid 401 drivers/s390/crypto/zcrypt_msgtype50.c AP_QID_CARD(zq->queue->qid), qid 402 drivers/s390/crypto/zcrypt_msgtype50.c AP_QID_QUEUE(zq->queue->qid), qid 297 drivers/s390/crypto/zcrypt_msgtype6.c msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); qid 367 drivers/s390/crypto/zcrypt_msgtype6.c msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); qid 657 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 658 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid), qid 664 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 665 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid)); qid 668 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 669 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid), qid 808 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 809 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid)); qid 812 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 813 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid), qid 842 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 843 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid)); qid 846 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 847 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid), qid 871 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 872 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid)); qid 875 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 876 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid), qid 901 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 902 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid)); qid 905 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_CARD(zq->queue->qid), qid 906 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid), qid 1217 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid); qid 1235 drivers/s390/crypto/zcrypt_msgtype6.c AP_QID_QUEUE(zq->queue->qid); qid 1297 drivers/s390/crypto/zcrypt_msgtype6.c msg->cprbx.domain = AP_QID_QUEUE(zq->queue->qid); qid 64 drivers/s390/crypto/zcrypt_queue.c AP_QID_CARD(zq->queue->qid), qid 65 drivers/s390/crypto/zcrypt_queue.c AP_QID_QUEUE(zq->queue->qid), qid 168 drivers/s390/crypto/zcrypt_queue.c AP_QID_CARD(zq->queue->qid), AP_QID_QUEUE(zq->queue->qid)); qid 211 drivers/s390/crypto/zcrypt_queue.c AP_QID_CARD(zq->queue->qid), AP_QID_QUEUE(zq->queue->qid)); qid 2712 drivers/scsi/aacraid/aacraid.h int aac_queue_get(struct aac_dev * dev, u32 * index, u32 qid, struct hw_fib * hw_fib, int wait, struct fib * fibptr, unsigned long *nonotify); qid 357 drivers/scsi/aacraid/commsup.c static int aac_get_entry (struct aac_dev * dev, u32 qid, struct aac_entry **entry, u32 * index, unsigned long *nonotify) qid 369 drivers/scsi/aacraid/commsup.c q = &dev->queues->queue[qid]; qid 375 drivers/scsi/aacraid/commsup.c if (qid == AdapNormCmdQueue) qid 384 drivers/scsi/aacraid/commsup.c if (qid == AdapNormCmdQueue) { qid 395 drivers/scsi/aacraid/commsup.c qid, atomic_read(&q->numpending)); qid 419 drivers/scsi/aacraid/commsup.c int aac_queue_get(struct aac_dev * dev, u32 * index, u32 qid, struct hw_fib * hw_fib, int wait, struct fib * fibptr, unsigned long *nonotify) qid 424 drivers/scsi/aacraid/commsup.c if (qid == AdapNormCmdQueue) { qid 426 drivers/scsi/aacraid/commsup.c while (!aac_get_entry(dev, qid, &entry, index, nonotify)) { qid 435 drivers/scsi/aacraid/commsup.c while (!aac_get_entry(dev, qid, &entry, index, nonotify)) { qid 830 drivers/scsi/aacraid/commsup.c void aac_consumer_free(struct aac_dev * dev, struct aac_queue *q, u32 qid) qid 844 drivers/scsi/aacraid/commsup.c switch (qid) { qid 684 drivers/scsi/be2iscsi/be_main.h u8 qid[10]; qid 51 drivers/scsi/bfa/bfa.h (__mh).mtag.h2i.qid = (__bfa)->iocfc.hw_qid[__reqq];\ qid 692 drivers/scsi/bfa/bfa_core.c bfa_reqq_resume(struct bfa_s *bfa, int qid) qid 697 drivers/scsi/bfa/bfa_core.c waitq = bfa_reqq(bfa, qid); qid 702 drivers/scsi/bfa/bfa_core.c if (bfa_reqq_full(bfa, qid)) qid 712 drivers/scsi/bfa/bfa_core.c bfa_isr_rspq(struct bfa_s *bfa, int qid) qid 719 drivers/scsi/bfa/bfa_core.c ci = bfa_rspq_ci(bfa, qid); qid 720 drivers/scsi/bfa/bfa_core.c pi = bfa_rspq_pi(bfa, qid); qid 725 drivers/scsi/bfa/bfa_core.c m = bfa_rspq_elem(bfa, qid, ci); qid 735 drivers/scsi/bfa/bfa_core.c bfa_isr_rspq_ack(bfa, qid, ci); qid 740 drivers/scsi/bfa/bfa_core.c waitq = bfa_reqq(bfa, qid); qid 742 drivers/scsi/bfa/bfa_core.c bfa_reqq_resume(bfa, qid); qid 748 drivers/scsi/bfa/bfa_core.c bfa_isr_reqq(struct bfa_s *bfa, int qid) qid 752 drivers/scsi/bfa/bfa_core.c bfa_isr_reqq_ack(bfa, qid); qid 757 drivers/scsi/bfa/bfa_core.c waitq = bfa_reqq(bfa, qid); qid 759 drivers/scsi/bfa/bfa_core.c bfa_reqq_resume(bfa, qid); qid 52 drivers/scsi/bfa/bfi.h u8 qid; qid 1137 drivers/scsi/csiostor/csio_wr.c uint32_t wr_type, fw_qid, qid; qid 1170 drivers/scsi/csiostor/csio_wr.c qid = fw_qid - wrm->fw_iq_start; qid 1171 drivers/scsi/csiostor/csio_wr.c q_completed = hw->wrm.intr_map[qid]; qid 1173 drivers/scsi/csiostor/csio_wr.c if (unlikely(qid == qid 274 drivers/scsi/csiostor/csio_wr.h __be32 qid; qid 562 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_wq_by_id(struct lpfc_hba *phba, int qid) qid 567 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.hdwq[wq_idx].io_wq->queue_id == qid) qid 570 drivers/scsi/lpfc/lpfc_debugfs.h pr_err("IO WQ[Idx:%d|Qid:%d]\n", wq_idx, qid); qid 575 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.els_wq->queue_id == qid) { qid 576 drivers/scsi/lpfc/lpfc_debugfs.h pr_err("ELS WQ[Qid:%d]\n", qid); qid 581 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.nvmels_wq->queue_id == qid) { qid 582 drivers/scsi/lpfc/lpfc_debugfs.h pr_err("NVME LS WQ[Qid:%d]\n", qid); qid 596 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_mq_by_id(struct lpfc_hba *phba, int qid) qid 598 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.mbx_wq->queue_id == qid) { qid 599 drivers/scsi/lpfc/lpfc_debugfs.h printk(KERN_ERR "MBX WQ[Qid:%d]\n", qid); qid 613 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_rq_by_id(struct lpfc_hba *phba, int qid) qid 615 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.hdr_rq->queue_id == qid) { qid 616 drivers/scsi/lpfc/lpfc_debugfs.h printk(KERN_ERR "HDR RQ[Qid:%d]\n", qid); qid 620 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.dat_rq->queue_id == qid) { qid 621 drivers/scsi/lpfc/lpfc_debugfs.h printk(KERN_ERR "DAT RQ[Qid:%d]\n", qid); qid 635 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_cq_by_id(struct lpfc_hba *phba, int qid) qid 640 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.hdwq[cq_idx].io_cq->queue_id == qid) qid 644 drivers/scsi/lpfc/lpfc_debugfs.h pr_err("IO CQ[Idx:%d|Qid:%d]\n", cq_idx, qid); qid 649 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.els_cq->queue_id == qid) { qid 650 drivers/scsi/lpfc/lpfc_debugfs.h pr_err("ELS CQ[Qid:%d]\n", qid); qid 655 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.nvmels_cq->queue_id == qid) { qid 656 drivers/scsi/lpfc/lpfc_debugfs.h pr_err("NVME LS CQ[Qid:%d]\n", qid); qid 661 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.mbx_cq->queue_id == qid) { qid 662 drivers/scsi/lpfc/lpfc_debugfs.h pr_err("MBX CQ[Qid:%d]\n", qid); qid 676 drivers/scsi/lpfc/lpfc_debugfs.h lpfc_debug_dump_eq_by_id(struct lpfc_hba *phba, int qid) qid 681 drivers/scsi/lpfc/lpfc_debugfs.h if (phba->sli4_hba.hdwq[eq_idx].hba_eq->queue_id == qid) qid 685 drivers/scsi/lpfc/lpfc_debugfs.h printk(KERN_ERR "FCP EQ[Idx:%d|Qid:%d]\n", eq_idx, qid); qid 4013 drivers/scsi/qla2xxx/qla_nx.c uint32_t r_stride, r_value, r_cnt, qid = 0; qid 4025 drivers/scsi/qla2xxx/qla_nx.c qla82xx_md_rw_32(ha, s_addr, qid, 1); qid 4032 drivers/scsi/qla2xxx/qla_nx.c qid += q_hdr->q_strd.queue_id_stride; qid 2627 drivers/scsi/qla2xxx/qla_nx2.c uint32_t r_stride, r_value, r_cnt, qid = 0; qid 2640 drivers/scsi/qla2xxx/qla_nx2.c qla8044_wr_reg_indirect(vha, s_addr, qid); qid 2647 drivers/scsi/qla2xxx/qla_nx2.c qid += q_hdr->q_strd.queue_id_stride; qid 3853 drivers/scsi/qla2xxx/qla_target.c int qid = GET_QID(handle); qid 3859 drivers/scsi/qla2xxx/qla_target.c if (qid == rsp->req->id) { qid 3861 drivers/scsi/qla2xxx/qla_target.c } else if (vha->hw->req_q_map[qid]) { qid 3865 drivers/scsi/qla2xxx/qla_target.c req = vha->hw->req_q_map[qid]; qid 2415 drivers/scsi/qla4xxx/ql4_nx.c uint32_t r_stride, r_value, r_cnt, qid = 0; qid 2428 drivers/scsi/qla4xxx/ql4_nx.c ha->isp_ops->wr_reg_indirect(ha, s_addr, qid); qid 2435 drivers/scsi/qla4xxx/ql4_nx.c qid += q_hdr->q_strd.queue_id_stride; qid 101 drivers/staging/media/ipu3/ipu3-css.c enum imgu_abi_queue_id qid; qid 1943 drivers/staging/media/ipu3/ipu3-css.c if (b->queue >= IPU3_CSS_QUEUES || !imgu_css_queues[b->queue].qid) qid 1946 drivers/staging/media/ipu3/ipu3-css.c b->queue_pos = imgu_css_queue_pos(css, imgu_css_queues[b->queue].qid, qid 1976 drivers/staging/media/ipu3/ipu3-css.c r = imgu_css_queue_data(css, imgu_css_queues[b->queue].qid, qid 1982 drivers/staging/media/ipu3/ipu3-css.c imgu_css_queues[b->queue].qid); qid 2015 drivers/staging/media/ipu3/ipu3-css.c int evtype, pipe, pipeid, queue, qid, r; qid 2038 drivers/staging/media/ipu3/ipu3-css.c qid = imgu_css_queues[queue].qid; qid 2045 drivers/staging/media/ipu3/ipu3-css.c if (qid >= IMGU_ABI_QUEUE_NUM) { qid 2046 drivers/staging/media/ipu3/ipu3-css.c dev_err(css->dev, "Invalid qid: %i\n", qid); qid 2054 drivers/staging/media/ipu3/ipu3-css.c r = imgu_css_dequeue_data(css, qid, &daddr); qid 2062 drivers/staging/media/ipu3/ipu3-css.c IMGU_ABI_EVENT_BUFFER_DEQUEUED(qid)); qid 86 fs/9p/cache.c if (buflen != sizeof(v9inode->qid.version)) qid 89 fs/9p/cache.c if (memcmp(buffer, &v9inode->qid.version, qid 90 fs/9p/cache.c sizeof(v9inode->qid.version))) qid 117 fs/9p/cache.c &v9inode->qid.path, qid 118 fs/9p/cache.c sizeof(v9inode->qid.path), qid 119 fs/9p/cache.c &v9inode->qid.version, qid 120 fs/9p/cache.c sizeof(v9inode->qid.version), qid 138 fs/9p/cache.c fscache_relinquish_cookie(v9inode->fscache, &v9inode->qid.version, qid 190 fs/9p/cache.c &v9inode->qid.path, qid 191 fs/9p/cache.c sizeof(v9inode->qid.path), qid 192 fs/9p/cache.c &v9inode->qid.version, qid 193 fs/9p/cache.c sizeof(v9inode->qid.version), qid 627 fs/9p/v9fs.c memset(&v9inode->qid, 0, sizeof(v9inode->qid)); qid 115 fs/9p/v9fs.h struct p9_qid qid; qid 51 fs/9p/v9fs_vfs.h ino_t v9fs_qid2ino(struct p9_qid *qid); qid 131 fs/9p/vfs_dir.c v9fs_qid2ino(&st.qid), dt_type(&st)); qid 188 fs/9p/vfs_dir.c v9fs_qid2ino(&curdirent.qid), qid 202 fs/9p/vfs_inode.c wstat->qid.type = ~0; qid 203 fs/9p/vfs_inode.c wstat->qid.version = ~0; qid 204 fs/9p/vfs_inode.c *((long long *)&wstat->qid.path) = ~0; qid 459 fs/9p/vfs_inode.c if (memcmp(&v9inode->qid.version, qid 460 fs/9p/vfs_inode.c &st->qid.version, sizeof(v9inode->qid.version))) qid 463 fs/9p/vfs_inode.c if (v9inode->qid.type != st->qid.type) qid 466 fs/9p/vfs_inode.c if (v9inode->qid.path != st->qid.path) qid 481 fs/9p/vfs_inode.c memcpy(&v9inode->qid, &st->qid, sizeof(st->qid)); qid 486 fs/9p/vfs_inode.c struct p9_qid *qid, qid 503 fs/9p/vfs_inode.c i_ino = v9fs_qid2ino(qid); qid 541 fs/9p/vfs_inode.c inode = v9fs_qid_iget(sb, &st->qid, st, new); qid 1214 fs/9p/vfs_inode.c ino_t v9fs_qid2ino(struct p9_qid *qid) qid 1216 fs/9p/vfs_inode.c u64 path = qid->path + 2; qid 69 fs/9p/vfs_inode_dotl.c if (memcmp(&v9inode->qid.version, qid 70 fs/9p/vfs_inode_dotl.c &st->qid.version, sizeof(v9inode->qid.version))) qid 73 fs/9p/vfs_inode_dotl.c if (v9inode->qid.type != st->qid.type) qid 76 fs/9p/vfs_inode_dotl.c if (v9inode->qid.path != st->qid.path) qid 92 fs/9p/vfs_inode_dotl.c memcpy(&v9inode->qid, &st->qid, sizeof(st->qid)); qid 98 fs/9p/vfs_inode_dotl.c struct p9_qid *qid, qid 114 fs/9p/vfs_inode_dotl.c i_ino = v9fs_qid2ino(qid); qid 156 fs/9p/vfs_inode_dotl.c inode = v9fs_qid_iget_dotl(sb, &st->qid, fid, st, new); qid 235 fs/9p/vfs_inode_dotl.c struct p9_qid qid; qid 289 fs/9p/vfs_inode_dotl.c mode, gid, &qid); qid 378 fs/9p/vfs_inode_dotl.c struct p9_qid qid; qid 407 fs/9p/vfs_inode_dotl.c err = p9_client_mkdir_dotl(dfid, name, mode, gid, &qid); qid 674 fs/9p/vfs_inode_dotl.c struct p9_qid qid; qid 694 fs/9p/vfs_inode_dotl.c err = p9_client_symlink(dfid, name, symname, gid, &qid); qid 812 fs/9p/vfs_inode_dotl.c struct p9_qid qid; qid 839 fs/9p/vfs_inode_dotl.c err = p9_client_mknod_dotl(dfid, name, mode, rdev, gid, &qid); qid 163 fs/9p/vfs_super.c d_inode(root)->i_ino = v9fs_qid2ino(&st->qid); qid 174 fs/9p/vfs_super.c d_inode(root)->i_ino = v9fs_qid2ino(&st->qid); qid 1388 fs/ext4/super.c static int ext4_get_next_id(struct super_block *sb, struct kqid *qid); qid 5575 fs/ext4/super.c struct kqid qid; qid 5580 fs/ext4/super.c qid = make_kqid_projid(projid); qid 5581 fs/ext4/super.c dquot = dqget(sb, qid); qid 6071 fs/ext4/super.c static int ext4_get_next_id(struct super_block *sb, struct kqid *qid) qid 6075 fs/ext4/super.c if (!sb_has_quota_loaded(sb, qid->type)) qid 6077 fs/ext4/super.c ops = sb_dqopt(sb)->ops[qid->type]; qid 6080 fs/ext4/super.c return dquot_get_next_id(sb, qid); qid 1205 fs/f2fs/super.c struct kqid qid; qid 1210 fs/f2fs/super.c qid = make_kqid_projid(projid); qid 1211 fs/f2fs/super.c dquot = dqget(sb, qid); qid 86 fs/gfs2/quota.c const struct kqid qid) qid 91 fs/gfs2/quota.c h = jhash(&qid, sizeof(struct kqid), h); qid 196 fs/gfs2/quota.c struct kqid qid = qd->qd_id; qid 197 fs/gfs2/quota.c return (2 * (u64)from_kqid(&init_user_ns, qid)) + qid 198 fs/gfs2/quota.c ((qid.type == USRQUOTA) ? 0 : 1); qid 211 fs/gfs2/quota.c static struct gfs2_quota_data *qd_alloc(unsigned hash, struct gfs2_sbd *sdp, struct kqid qid) qid 223 fs/gfs2/quota.c qd->qd_id = qid; qid 242 fs/gfs2/quota.c struct kqid qid) qid 248 fs/gfs2/quota.c if (!qid_eq(qd->qd_id, qid)) qid 262 fs/gfs2/quota.c static int qd_get(struct gfs2_sbd *sdp, struct kqid qid, qid 266 fs/gfs2/quota.c unsigned int hash = gfs2_qd_hash(sdp, qid); qid 269 fs/gfs2/quota.c *qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid); qid 275 fs/gfs2/quota.c new_qd = qd_alloc(hash, sdp, qid); qid 281 fs/gfs2/quota.c *qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid); qid 494 fs/gfs2/quota.c static int qdsb_get(struct gfs2_sbd *sdp, struct kqid qid, qid 499 fs/gfs2/quota.c error = qd_get(sdp, qid, qdp); qid 1317 fs/gfs2/quota.c int gfs2_quota_refresh(struct gfs2_sbd *sdp, struct kqid qid) qid 1323 fs/gfs2/quota.c error = qd_get(sdp, qid, &qd); qid 1610 fs/gfs2/quota.c static int gfs2_get_dqblk(struct super_block *sb, struct kqid qid, qid 1624 fs/gfs2/quota.c if ((qid.type != USRQUOTA) && qid 1625 fs/gfs2/quota.c (qid.type != GRPQUOTA)) qid 1628 fs/gfs2/quota.c error = qd_get(sdp, qid, &qd); qid 1649 fs/gfs2/quota.c static int gfs2_set_dqblk(struct super_block *sb, struct kqid qid, qid 1665 fs/gfs2/quota.c if ((qid.type != USRQUOTA) && qid 1666 fs/gfs2/quota.c (qid.type != GRPQUOTA)) qid 1672 fs/gfs2/quota.c error = qd_get(sdp, qid, &qd); qid 32 fs/gfs2/quota.h extern int gfs2_quota_refresh(struct gfs2_sbd *sdp, struct kqid qid); qid 185 fs/gfs2/sys.c struct kqid qid; qid 196 fs/gfs2/sys.c qid = make_kqid(current_user_ns(), USRQUOTA, id); qid 197 fs/gfs2/sys.c if (!qid_valid(qid)) qid 200 fs/gfs2/sys.c error = gfs2_quota_refresh(sdp, qid); qid 207 fs/gfs2/sys.c struct kqid qid; qid 218 fs/gfs2/sys.c qid = make_kqid(current_user_ns(), GRPQUOTA, id); qid 219 fs/gfs2/sys.c if (!qid_valid(qid)) qid 222 fs/gfs2/sys.c error = gfs2_quota_refresh(sdp, qid); qid 27 fs/ntfs/quota.c const le32 qid = QUOTA_DEFAULTS_ID; qid 43 fs/ntfs/quota.c err = ntfs_index_lookup(&qid, sizeof(qid), ictx); qid 877 fs/ocfs2/quota_global.c static int ocfs2_get_next_id(struct super_block *sb, struct kqid *qid) qid 879 fs/ocfs2/quota_global.c int type = qid->type; qid 883 fs/ocfs2/quota_global.c trace_ocfs2_get_next_id(from_kqid(&init_user_ns, *qid), type); qid 894 fs/ocfs2/quota_global.c status = qtree_get_next_id(&info->dqi_gi, qid); qid 262 fs/quota/dquot.c hashfn(const struct super_block *sb, struct kqid qid) qid 264 fs/quota/dquot.c unsigned int id = from_kqid(&init_user_ns, qid); qid 265 fs/quota/dquot.c int type = qid.type; qid 288 fs/quota/dquot.c struct kqid qid) qid 295 fs/quota/dquot.c if (dquot->dq_sb == sb && qid_eq(dquot->dq_id, qid)) qid 848 fs/quota/dquot.c struct dquot *dqget(struct super_block *sb, struct kqid qid) qid 850 fs/quota/dquot.c unsigned int hashent = hashfn(sb, qid); qid 853 fs/quota/dquot.c if (!qid_has_mapping(sb->s_user_ns, qid)) qid 856 fs/quota/dquot.c if (!sb_has_quota_active(sb, qid.type)) qid 861 fs/quota/dquot.c if (!sb_has_quota_active(sb, qid.type)) { qid 869 fs/quota/dquot.c dquot = find_dquot(hashent, sb, qid); qid 873 fs/quota/dquot.c empty = get_empty_dquot(sb, qid.type); qid 880 fs/quota/dquot.c dquot->dq_id = qid; qid 1446 fs/quota/dquot.c struct kqid qid; qid 1468 fs/quota/dquot.c qid = make_kqid_uid(inode->i_uid); qid 1471 fs/quota/dquot.c qid = make_kqid_gid(inode->i_gid); qid 1477 fs/quota/dquot.c qid = make_kqid_projid(projid); qid 1480 fs/quota/dquot.c dquot = dqget(sb, qid); qid 2126 fs/quota/dquot.c int dquot_get_next_id(struct super_block *sb, struct kqid *qid) qid 2130 fs/quota/dquot.c if (!sb_has_quota_active(sb, qid->type)) qid 2132 fs/quota/dquot.c if (!dqopt->ops[qid->type]->get_next_id) qid 2134 fs/quota/dquot.c return dqopt->ops[qid->type]->get_next_id(sb, qid); qid 2630 fs/quota/dquot.c int dquot_get_dqblk(struct super_block *sb, struct kqid qid, qid 2635 fs/quota/dquot.c dquot = dqget(sb, qid); qid 2645 fs/quota/dquot.c int dquot_get_next_dqblk(struct super_block *sb, struct kqid *qid, qid 2653 fs/quota/dquot.c err = sb->dq_op->get_next_id(sb, qid); qid 2656 fs/quota/dquot.c dquot = dqget(sb, *qid); qid 2763 fs/quota/dquot.c int dquot_set_dqblk(struct super_block *sb, struct kqid qid, qid 2769 fs/quota/dquot.c dquot = dqget(sb, qid); qid 120 fs/quota/kqid.c bool qid_valid(struct kqid qid) qid 122 fs/quota/kqid.c switch (qid.type) { qid 124 fs/quota/kqid.c return uid_valid(qid.uid); qid 126 fs/quota/kqid.c return gid_valid(qid.gid); qid 128 fs/quota/kqid.c return projid_valid(qid.projid); qid 37 fs/quota/netlink.c void quota_send_warning(struct kqid qid, dev_t dev, qid 63 fs/quota/netlink.c ret = nla_put_u32(skb, QUOTA_NL_A_QTYPE, qid.type); qid 67 fs/quota/netlink.c from_kqid_munged(&init_user_ns, qid), qid 202 fs/quota/quota.c struct kqid qid; qid 209 fs/quota/quota.c qid = make_kqid(current_user_ns(), type, id); qid 210 fs/quota/quota.c if (!qid_has_mapping(sb->s_user_ns, qid)) qid 212 fs/quota/quota.c ret = sb->s_qcop->get_dqblk(sb, qid, &fdq); qid 228 fs/quota/quota.c struct kqid qid; qid 235 fs/quota/quota.c qid = make_kqid(current_user_ns(), type, id); qid 236 fs/quota/quota.c if (!qid_has_mapping(sb->s_user_ns, qid)) qid 238 fs/quota/quota.c ret = sb->s_qcop->get_nextdqblk(sb, &qid, &fdq); qid 243 fs/quota/quota.c idq.dqb_id = from_kqid(current_user_ns(), qid); qid 280 fs/quota/quota.c struct kqid qid; qid 286 fs/quota/quota.c qid = make_kqid(current_user_ns(), type, id); qid 287 fs/quota/quota.c if (!qid_has_mapping(sb->s_user_ns, qid)) qid 290 fs/quota/quota.c return sb->s_qcop->set_dqblk(sb, qid, &fdq); qid 565 fs/quota/quota.c struct kqid qid; qid 571 fs/quota/quota.c qid = make_kqid(current_user_ns(), type, id); qid 572 fs/quota/quota.c if (!qid_has_mapping(sb->s_user_ns, qid)) qid 575 fs/quota/quota.c if (from_kqid(sb->s_user_ns, qid) == 0 && qid 590 fs/quota/quota.c return sb->s_qcop->set_dqblk(sb, qid, &qdq); qid 627 fs/quota/quota.c struct kqid qid; qid 632 fs/quota/quota.c qid = make_kqid(current_user_ns(), type, id); qid 633 fs/quota/quota.c if (!qid_has_mapping(sb->s_user_ns, qid)) qid 635 fs/quota/quota.c ret = sb->s_qcop->get_dqblk(sb, qid, &qdq); qid 653 fs/quota/quota.c struct kqid qid; qid 659 fs/quota/quota.c qid = make_kqid(current_user_ns(), type, id); qid 660 fs/quota/quota.c if (!qid_has_mapping(sb->s_user_ns, qid)) qid 662 fs/quota/quota.c ret = sb->s_qcop->get_nextdqblk(sb, &qid, &qdq); qid 665 fs/quota/quota.c id_out = from_kqid(current_user_ns(), qid); qid 36 fs/quota/quota_tree.c static int get_index(struct qtree_mem_dqinfo *info, struct kqid qid, int depth) qid 38 fs/quota/quota_tree.c qid_t id = from_kqid(&init_user_ns, qid); qid 723 fs/quota/quota_tree.c int qtree_get_next_id(struct qtree_mem_dqinfo *info, struct kqid *qid) qid 725 fs/quota/quota_tree.c qid_t id = from_kqid(&init_user_ns, *qid); qid 731 fs/quota/quota_tree.c *qid = make_kqid(&init_user_ns, qid->type, id); qid 363 fs/quota/quota_v2.c static int v2_get_next_id(struct super_block *sb, struct kqid *qid) qid 369 fs/quota/quota_v2.c ret = qtree_get_next_id(sb_dqinfo(sb, qid->type)->dqi_priv, qid); qid 220 fs/xfs/xfs_quotaops.c struct kqid qid, qid 231 fs/xfs/xfs_quotaops.c id = from_kqid(&init_user_ns, qid); qid 232 fs/xfs/xfs_quotaops.c return xfs_qm_scall_getquota(mp, id, xfs_quota_type(qid.type), qdq); qid 239 fs/xfs/xfs_quotaops.c struct kqid *qid, qid 251 fs/xfs/xfs_quotaops.c id = from_kqid(&init_user_ns, *qid); qid 252 fs/xfs/xfs_quotaops.c ret = xfs_qm_scall_getquota_next(mp, &id, xfs_quota_type(qid->type), qid 258 fs/xfs/xfs_quotaops.c *qid = make_kqid(current_user_ns(), qid->type, id); qid 265 fs/xfs/xfs_quotaops.c struct kqid qid, qid 277 fs/xfs/xfs_quotaops.c return xfs_qm_scall_setqlim(mp, from_kqid(&init_user_ns, qid), qid 278 fs/xfs/xfs_quotaops.c xfs_quota_type(qid.type), qdq); qid 57 include/linux/dqblk_qtree.h int qtree_get_next_id(struct qtree_mem_dqinfo *info, struct kqid *qid); qid 261 include/linux/firmware/xlnx-zynqmp.h u32 qid; qid 207 include/linux/nvme-fc.h __be16 qid; qid 58 include/linux/nvme-rdma.h __le16 qid; qid 973 include/linux/nvme.h __le16 qid; qid 1072 include/linux/nvme.h #define show_opcode_name(qid, opcode, fctype) \ qid 1075 include/linux/nvme.h ((qid) ? \ qid 1148 include/linux/nvme.h __le16 qid; qid 89 include/linux/qed/qed_eth_if.h u16 qid; qid 79 include/linux/quota.h extern qid_t from_kqid(struct user_namespace *to, struct kqid qid); qid 80 include/linux/quota.h extern qid_t from_kqid_munged(struct user_namespace *to, struct kqid qid); qid 81 include/linux/quota.h extern bool qid_valid(struct kqid qid); qid 98 include/linux/quota.h enum quota_type type, qid_t qid) qid 105 include/linux/quota.h kqid.uid = make_kuid(from, qid); qid 108 include/linux/quota.h kqid.gid = make_kgid(from, qid); qid 111 include/linux/quota.h kqid.projid = make_kprojid(from, qid); qid 187 include/linux/quota.h static inline bool qid_has_mapping(struct user_namespace *ns, struct kqid qid) qid 189 include/linux/quota.h return from_kqid(ns, qid) != (qid_t) -1; qid 318 include/linux/quota.h int (*get_next_id)(struct super_block *sb, struct kqid *qid); /* Get next ID with existing structure in the quota file */ qid 337 include/linux/quota.h int (*get_next_id) (struct super_block *sb, struct kqid *qid); qid 508 include/linux/quota.h extern void quota_send_warning(struct kqid qid, dev_t dev, qid 511 include/linux/quota.h static inline void quota_send_warning(struct kqid qid, dev_t dev, qid 48 include/linux/quotaops.h struct dquot *dqget(struct super_block *sb, struct kqid qid); qid 95 include/linux/quotaops.h int dquot_get_next_id(struct super_block *sb, struct kqid *qid); qid 388 include/net/9p/9p.h struct p9_qid qid; qid 405 include/net/9p/9p.h struct p9_qid qid; qid 150 include/net/9p/client.h struct p9_qid qid; qid 168 include/net/9p/client.h struct p9_qid qid; qid 195 include/net/9p/client.h kgid_t gid, struct p9_qid *qid); qid 197 include/net/9p/client.h kgid_t gid, struct p9_qid *qid); qid 898 net/9p/client.c memset(&fid->qid, 0, sizeof(struct p9_qid)); qid 1122 net/9p/client.c struct p9_qid qid; qid 1141 net/9p/client.c err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", &qid); qid 1149 net/9p/client.c qid.type, (unsigned long long)qid.path, qid.version); qid 1151 net/9p/client.c memmove(&fid->qid, &qid, sizeof(struct p9_qid)); qid 1220 net/9p/client.c memmove(&fid->qid, &wqids[nwqids - 1], sizeof(struct p9_qid)); qid 1222 net/9p/client.c fid->qid = oldfid->qid; qid 1245 net/9p/client.c struct p9_qid qid; qid 1265 net/9p/client.c err = p9pdu_readf(&req->rc, clnt->proto_version, "Qd", &qid, &iounit); qid 1272 net/9p/client.c p9_is_proto_dotl(clnt) ? "RLOPEN" : "ROPEN", qid.type, qid 1273 net/9p/client.c (unsigned long long)qid.path, qid.version, iounit); qid 1286 net/9p/client.c kgid_t gid, struct p9_qid *qid) qid 1309 net/9p/client.c err = p9pdu_readf(&req->rc, clnt->proto_version, "Qd", qid, &iounit); qid 1316 net/9p/client.c qid->type, qid 1317 net/9p/client.c (unsigned long long)qid->path, qid 1318 net/9p/client.c qid->version, iounit); qid 1336 net/9p/client.c struct p9_qid qid; qid 1354 net/9p/client.c err = p9pdu_readf(&req->rc, clnt->proto_version, "Qd", &qid, &iounit); qid 1361 net/9p/client.c qid.type, qid 1362 net/9p/client.c (unsigned long long)qid.path, qid 1363 net/9p/client.c qid.version, iounit); qid 1376 net/9p/client.c const char *symtgt, kgid_t gid, struct p9_qid *qid) qid 1393 net/9p/client.c err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", qid); qid 1400 net/9p/client.c qid->type, (unsigned long long)qid->path, qid->version); qid 1720 net/9p/client.c ret->size, ret->type, ret->dev, ret->qid.type, qid 1721 net/9p/client.c (unsigned long long)ret->qid.path, ret->qid.version, ret->mode, qid 1779 net/9p/client.c ret->st_result_mask, ret->qid.type, ret->qid.path, qid 1780 net/9p/client.c ret->qid.version, ret->st_mode, ret->st_nlink, qid 1842 net/9p/client.c wst->size, wst->type, wst->dev, wst->qid.type, qid 1843 net/9p/client.c (unsigned long long)wst->qid.path, wst->qid.version, wst->mode, qid 2135 net/9p/client.c dev_t rdev, kgid_t gid, struct p9_qid *qid) qid 2150 net/9p/client.c err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", qid); qid 2155 net/9p/client.c p9_debug(P9_DEBUG_9P, "<<< RMKNOD qid %x.%llx.%x\n", qid->type, qid 2156 net/9p/client.c (unsigned long long)qid->path, qid->version); qid 2166 net/9p/client.c kgid_t gid, struct p9_qid *qid) qid 2181 net/9p/client.c err = p9pdu_readf(&req->rc, clnt->proto_version, "Q", qid); qid 2186 net/9p/client.c p9_debug(P9_DEBUG_9P, "<<< RMKDIR qid %x.%llx.%x\n", qid->type, qid 2187 net/9p/client.c (unsigned long long)qid->path, qid->version); qid 181 net/9p/protocol.c struct p9_qid *qid = qid 185 net/9p/protocol.c &qid->type, &qid->version, qid 186 net/9p/protocol.c &qid->path); qid 201 net/9p/protocol.c &stbuf->dev, &stbuf->qid, qid 315 net/9p/protocol.c &stbuf->qid, qid 411 net/9p/protocol.c const struct p9_qid *qid = qid 415 net/9p/protocol.c qid->type, qid->version, qid 416 net/9p/protocol.c qid->path); qid 425 net/9p/protocol.c stbuf->dev, &stbuf->qid, qid 606 net/9p/protocol.c ret = p9pdu_readf(&fake_pdu, clnt->proto_version, "Qqbs", &dirent->qid, qid 1988 net/sched/sch_api.c u32 qid; qid 2021 net/sched/sch_api.c qid = TC_H_MAJ(clid); qid 2026 net/sched/sch_api.c if (qid && qid1) { qid 2028 net/sched/sch_api.c if (qid != qid1) qid 2031 net/sched/sch_api.c qid = qid1; qid 2032 net/sched/sch_api.c } else if (qid == 0) qid 2033 net/sched/sch_api.c qid = dev->qdisc->handle; qid 2041 net/sched/sch_api.c portid = TC_H_MAKE(qid, portid); qid 2043 net/sched/sch_api.c if (qid == 0) qid 2044 net/sched/sch_api.c qid = dev->qdisc->handle; qid 2048 net/sched/sch_api.c q = qdisc_lookup(dev, qid); qid 2060 net/sched/sch_api.c clid = qid; qid 2062 net/sched/sch_api.c clid = TC_H_MAKE(qid, clid); qid 616 net/xdp/xsk.c u32 flags, qid; qid 647 net/xdp/xsk.c qid = sxdp->sxdp_queue_id; qid 678 net/xdp/xsk.c if (umem_xs->dev != dev || umem_xs->queue_id != qid) { qid 697 net/xdp/xsk.c err = xdp_umem_assign_dev(xs->umem, dev, qid, flags); qid 706 net/xdp/xsk.c xs->queue_id = qid;