/linux-4.1.27/drivers/md/ |
D | dm-cache-policy-mq.c | 495 static void hash_insert(struct mq_policy *mq, struct entry *e) in hash_insert() argument 497 unsigned h = hash_64(from_oblock(e->oblock), mq->hash_bits); in hash_insert() 499 hlist_add_head(&e->hlist, mq->table + h); in hash_insert() 502 static struct entry *hash_lookup(struct mq_policy *mq, dm_oblock_t oblock) in hash_lookup() argument 504 unsigned h = hash_64(from_oblock(oblock), mq->hash_bits); in hash_lookup() 505 struct hlist_head *bucket = mq->table + h; in hash_lookup() 525 static bool any_free_cblocks(struct mq_policy *mq) in any_free_cblocks() argument 527 return !epool_empty(&mq->cache_pool); in any_free_cblocks() 530 static bool any_clean_cblocks(struct mq_policy *mq) in any_clean_cblocks() argument 532 return !queue_empty(&mq->cache_clean); in any_clean_cblocks() [all …]
|
D | Makefile | 15 dm-cache-mq-y += dm-cache-policy-mq.o 56 obj-$(CONFIG_DM_CACHE_MQ) += dm-cache-mq.o
|
D | Kconfig | 216 bool "request-based DM: use blk-mq I/O path by default" 219 This option enables the blk-mq based I/O path for request-based
|
/linux-4.1.27/drivers/scsi/arm/ |
D | msgqueue.c | 27 struct msgqueue_entry *mq; in mqe_alloc() local 29 if ((mq = msgq->free) != NULL) in mqe_alloc() 30 msgq->free = mq->next; in mqe_alloc() 32 return mq; in mqe_alloc() 41 static void mqe_free(MsgQueue_t *msgq, struct msgqueue_entry *mq) in mqe_free() argument 43 if (mq) { in mqe_free() 44 mq->next = msgq->free; in mqe_free() 45 msgq->free = mq; in mqe_free() 85 struct msgqueue_entry *mq = msgq->qe; in msgqueue_msglength() local 88 for (mq = msgq->qe; mq; mq = mq->next) in msgqueue_msglength() [all …]
|
/linux-4.1.27/drivers/mmc/card/ |
D | queue.c | 31 struct mmc_queue *mq = q->queuedata; in mmc_prep_request() local 41 if (mq && (mmc_card_removed(mq->card) || mmc_access_rpmb(mq))) in mmc_prep_request() 51 struct mmc_queue *mq = d; in mmc_queue_thread() local 52 struct request_queue *q = mq->queue; in mmc_queue_thread() 56 down(&mq->thread_sem); in mmc_queue_thread() 65 mq->mqrq_cur->req = req; in mmc_queue_thread() 68 if (req || mq->mqrq_prev->req) { in mmc_queue_thread() 71 mq->issue_fn(mq, req); in mmc_queue_thread() 72 if (mq->flags & MMC_QUEUE_NEW_REQUEST) { in mmc_queue_thread() 73 mq->flags &= ~MMC_QUEUE_NEW_REQUEST; in mmc_queue_thread() [all …]
|
D | block.c | 1034 int mmc_access_rpmb(struct mmc_queue *mq) in mmc_access_rpmb() argument 1036 struct mmc_blk_data *md = mq->data; in mmc_access_rpmb() 1046 static int mmc_blk_issue_discard_rq(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_discard_rq() argument 1048 struct mmc_blk_data *md = mq->data; in mmc_blk_issue_discard_rq() 1089 static int mmc_blk_issue_secdiscard_rq(struct mmc_queue *mq, in mmc_blk_issue_secdiscard_rq() argument 1092 struct mmc_blk_data *md = mq->data; in mmc_blk_issue_secdiscard_rq() 1156 static int mmc_blk_issue_flush(struct mmc_queue *mq, struct request *req) in mmc_blk_issue_flush() argument 1158 struct mmc_blk_data *md = mq->data; in mmc_blk_issue_flush() 1360 struct mmc_queue *mq) in mmc_blk_rw_rq_prep() argument 1365 struct mmc_blk_data *md = mq->data; in mmc_blk_rw_rq_prep() [all …]
|
/linux-4.1.27/drivers/infiniband/hw/amso1100/ |
D | c2_cq.c | 92 q = &cq->mq; in c2_cq_clean() 138 ce = c2_mq_consume(&cq->mq); in c2_poll_one() 150 c2_mq_free(&cq->mq); in c2_poll_one() 151 ce = c2_mq_consume(&cq->mq); in c2_poll_one() 196 c2_mq_free(&cq->mq); in c2_poll_one() 230 shared = cq->mq.peer; in c2_arm_cq() 250 ret = !c2_mq_empty(&cq->mq); in c2_arm_cq() 257 static void c2_free_cq_buf(struct c2_dev *c2dev, struct c2_mq *mq) in c2_free_cq_buf() argument 259 dma_free_coherent(&c2dev->pcidev->dev, mq->q_size * mq->msg_size, in c2_free_cq_buf() 260 mq->msg_pool.host, dma_unmap_addr(mq, mapping)); in c2_free_cq_buf() [all …]
|
D | c2_ae.c | 149 struct c2_mq *mq = c2dev->qptr_array[mq_index]; in c2_ae_event() local 164 wr = c2_mq_consume(mq); in c2_ae_event() 326 c2_mq_free(mq); in c2_ae_event()
|
D | c2_provider.h | 100 struct c2_mq mq; member
|
/linux-4.1.27/drivers/sh/maple/ |
D | maple.c | 122 void (*callback) (struct mapleq *mq), in maple_getcond_callback() 140 struct mapleq *mq; in maple_release_device() local 143 mq = mdev->mq; in maple_release_device() 144 kmem_cache_free(maple_queue_cache, mq->recvbuf); in maple_release_device() 145 kfree(mq); in maple_release_device() 172 mdev->mq->command = command; in maple_add_packet() 173 mdev->mq->length = length; in maple_add_packet() 176 mdev->mq->sendbuf = sendbuf; in maple_add_packet() 179 list_add_tail(&mdev->mq->list, &maple_waitq); in maple_add_packet() 188 struct mapleq *mq; in maple_allocq() local [all …]
|
/linux-4.1.27/drivers/isdn/capi/ |
D | capilib.c | 47 struct capilib_msgidqueue *mq; in mq_enqueue() local 48 if ((mq = np->msgidfree) == NULL) in mq_enqueue() 50 np->msgidfree = mq->next; in mq_enqueue() 51 mq->msgid = msgid; in mq_enqueue() 52 mq->next = NULL; in mq_enqueue() 54 np->msgidlast->next = mq; in mq_enqueue() 55 np->msgidlast = mq; in mq_enqueue() 57 np->msgidqueue = mq; in mq_enqueue() 67 struct capilib_msgidqueue *mq = *pp; in mq_dequeue() local 68 *pp = mq->next; in mq_dequeue() [all …]
|
/linux-4.1.27/drivers/misc/sgi-xp/ |
D | xpc_uv.c | 109 xpc_get_gru_mq_irq_uv(struct xpc_gru_mq_uv *mq, int cpu, char *irq_name) in xpc_get_gru_mq_irq_uv() argument 111 int mmr_pnode = uv_blade_to_pnode(mq->mmr_blade); in xpc_get_gru_mq_irq_uv() 114 mq->irq = uv_setup_irq(irq_name, cpu, mq->mmr_blade, mq->mmr_offset, in xpc_get_gru_mq_irq_uv() 116 if (mq->irq < 0) in xpc_get_gru_mq_irq_uv() 117 return mq->irq; in xpc_get_gru_mq_irq_uv() 119 mq->mmr_value = uv_read_global_mmr64(mmr_pnode, mq->mmr_offset); in xpc_get_gru_mq_irq_uv() 123 mq->irq = SGI_XPC_ACTIVATE; in xpc_get_gru_mq_irq_uv() 125 mq->irq = SGI_XPC_NOTIFY; in xpc_get_gru_mq_irq_uv() 129 mq->mmr_value = (unsigned long)cpu_physical_id(cpu) << 32 | mq->irq; in xpc_get_gru_mq_irq_uv() 130 uv_write_global_mmr64(mmr_pnode, mq->mmr_offset, mq->mmr_value); in xpc_get_gru_mq_irq_uv() [all …]
|
/linux-4.1.27/drivers/misc/sgi-gru/ |
D | grukservices.c | 146 #define HSTATUS(mq, h) ((mq) + offsetof(struct message_queue, hstatus[h])) argument 555 struct message_queue *mq = p; in gru_create_message_queue() local 559 memset(mq, 0, bytes); in gru_create_message_queue() 560 mq->start = &mq->data; in gru_create_message_queue() 561 mq->start2 = &mq->data + (qlines / 2 - 1) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue() 562 mq->next = &mq->data; in gru_create_message_queue() 563 mq->limit = &mq->data + (qlines - 2) * GRU_CACHE_LINE_BYTES; in gru_create_message_queue() 564 mq->qlines = qlines; in gru_create_message_queue() 565 mq->hstatus[0] = 0; in gru_create_message_queue() 566 mq->hstatus[1] = 1; in gru_create_message_queue() [all …]
|
D | grukservices.h | 45 void *mq; /* message queue vaddress */ member
|
/linux-4.1.27/drivers/mailbox/ |
D | omap-mailbox.c | 305 struct omap_mbox_queue *mq = in mbox_rx_work() local 310 while (kfifo_len(&mq->fifo) >= sizeof(msg)) { in mbox_rx_work() 311 len = kfifo_out(&mq->fifo, (unsigned char *)&msg, sizeof(msg)); in mbox_rx_work() 314 mbox_chan_received_data(mq->mbox->chan, (void *)msg); in mbox_rx_work() 315 spin_lock_irq(&mq->lock); in mbox_rx_work() 316 if (mq->full) { in mbox_rx_work() 317 mq->full = false; in mbox_rx_work() 318 _omap_mbox_enable_irq(mq->mbox, IRQ_RX); in mbox_rx_work() 320 spin_unlock_irq(&mq->lock); in mbox_rx_work() 336 struct omap_mbox_queue *mq = mbox->rxq; in __mbox_rx_interrupt() local [all …]
|
/linux-4.1.27/block/ |
D | Makefile | 8 blk-iopoll.o blk-lib.o blk-mq.o blk-mq-tag.o \ 9 blk-mq-sysfs.o blk-mq-cpu.o blk-mq-cpumap.o ioctl.o \
|
/linux-4.1.27/include/linux/ |
D | maple.h | 69 struct mapleq *mq; member 70 void (*callback) (struct mapleq * mq); 89 void (*callback) (struct mapleq * mq),
|
/linux-4.1.27/drivers/mtd/maps/ |
D | vmu-flash.c | 91 static void vmu_blockread(struct mapleq *mq) in vmu_blockread() argument 96 mdev = mq->dev; in vmu_blockread() 103 memcpy(card->blockread, mq->recvbuf->buf + 12, in vmu_blockread() 193 list_del_init(&(mdev->mq->list)); in maple_vmu_read_block() 194 kfree(mdev->mq->sendbuf); in maple_vmu_read_block() 195 mdev->mq->sendbuf = NULL; in maple_vmu_read_block() 285 kfree(mdev->mq->sendbuf); in maple_vmu_write_block() 286 mdev->mq->sendbuf = NULL; in maple_vmu_write_block() 287 list_del_init(&(mdev->mq->list)); in maple_vmu_write_block() 501 static void vmu_queryblocks(struct mapleq *mq) in vmu_queryblocks() argument [all …]
|
/linux-4.1.27/drivers/input/mouse/ |
D | maplemouse.c | 26 static void dc_mouse_callback(struct mapleq *mq) in dc_mouse_callback() argument 29 struct maple_device *mapledev = mq->dev; in dc_mouse_callback() 32 unsigned char *res = mq->recvbuf->buf; in dc_mouse_callback()
|
/linux-4.1.27/Documentation/devicetree/bindings/powerpc/4xx/ |
D | ppc440spe-adma.txt | 84 - compatible : "ibm,mq-440spe"; 89 MQ0: mq { 90 compatible = "ibm,mq-440spe";
|
/linux-4.1.27/drivers/net/wireless/iwlwifi/dvm/ |
D | main.c | 1952 int mq = priv->queue_to_mac80211[queue]; in iwl_stop_sw_queue() local 1954 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_stop_sw_queue() 1957 if (atomic_inc_return(&priv->queue_stop_count[mq]) > 1) { in iwl_stop_sw_queue() 1960 queue, mq); in iwl_stop_sw_queue() 1964 set_bit(mq, &priv->transport_queue_stop); in iwl_stop_sw_queue() 1965 ieee80211_stop_queue(priv->hw, mq); in iwl_stop_sw_queue() 1971 int mq = priv->queue_to_mac80211[queue]; in iwl_wake_sw_queue() local 1973 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_wake_sw_queue() 1976 if (atomic_dec_return(&priv->queue_stop_count[mq]) > 0) { in iwl_wake_sw_queue() 1979 queue, mq); in iwl_wake_sw_queue() [all …]
|
D | tx.c | 483 static int iwlagn_alloc_agg_txq(struct iwl_priv *priv, int mq) in iwlagn_alloc_agg_txq() argument 490 priv->queue_to_mac80211[q] = mq; in iwlagn_alloc_agg_txq()
|
/linux-4.1.27/drivers/input/joystick/ |
D | maplecontrol.c | 26 static void dc_pad_callback(struct mapleq *mq) in dc_pad_callback() argument 29 struct maple_device *mapledev = mq->dev; in dc_pad_callback() 32 unsigned char *res = mq->recvbuf->buf; in dc_pad_callback()
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
D | ocrdma_hw.c | 107 (dev->mq.cq.va + (dev->mq.cq.tail * sizeof(struct ocrdma_mcqe))); in ocrdma_get_mcqe() 116 dev->mq.cq.tail = (dev->mq.cq.tail + 1) & (OCRDMA_MQ_CQ_LEN - 1); in ocrdma_mcq_inc_tail() 121 return dev->mq.sq.va + (dev->mq.sq.head * sizeof(struct ocrdma_mqe)); in ocrdma_get_mqe() 126 dev->mq.sq.head = (dev->mq.sq.head + 1) & (OCRDMA_MQ_LEN - 1); in ocrdma_mq_inc_head() 131 return dev->mq.sq.va + (dev->mqe_ctx.tag * sizeof(struct ocrdma_mqe)); in ocrdma_get_mqe_rsp() 305 val |= dev->mq.sq.id & OCRDMA_MQ_ID_MASK; in ocrdma_ring_mq_db() 546 struct ocrdma_queue_info *mq, in ocrdma_mbx_create_mq() argument 555 num_pages = PAGES_4K_SPANNED(mq->va, mq->size); in ocrdma_mbx_create_mq() 568 cmd->async_cqid_ringsize |= (ocrdma_encoded_q_len(mq->len) << in ocrdma_mbx_create_mq() 573 ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K); in ocrdma_mbx_create_mq() [all …]
|
D | ocrdma.h | 258 struct ocrdma_mq mq; member
|
/linux-4.1.27/drivers/net/wireless/iwlwifi/mvm/ |
D | ops.c | 768 int mq = mvm->queue_to_mac80211[queue]; in iwl_mvm_stop_sw_queue() local 770 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_mvm_stop_sw_queue() 773 if (atomic_inc_return(&mvm->mac80211_queue_stop_count[mq]) > 1) { in iwl_mvm_stop_sw_queue() 776 queue, mq); in iwl_mvm_stop_sw_queue() 780 ieee80211_stop_queue(mvm->hw, mq); in iwl_mvm_stop_sw_queue() 786 int mq = mvm->queue_to_mac80211[queue]; in iwl_mvm_wake_sw_queue() local 788 if (WARN_ON_ONCE(mq == IWL_INVALID_MAC80211_QUEUE)) in iwl_mvm_wake_sw_queue() 791 if (atomic_dec_return(&mvm->mac80211_queue_stop_count[mq]) > 0) { in iwl_mvm_wake_sw_queue() 794 queue, mq); in iwl_mvm_wake_sw_queue() 798 ieee80211_wake_queue(mvm->hw, mq); in iwl_mvm_wake_sw_queue()
|
/linux-4.1.27/Documentation/ABI/testing/ |
D | sysfs-block-dm | 45 Description: Request-based Device-mapper blk-mq I/O path mode. 46 Contains the value 1 if the device is using blk-mq.
|
/linux-4.1.27/drivers/input/keyboard/ |
D | maple_keyb.c | 139 static void dc_kbd_callback(struct mapleq *mq) in dc_kbd_callback() argument 141 struct maple_device *mapledev = mq->dev; in dc_kbd_callback() 143 unsigned long *buf = (unsigned long *)(mq->recvbuf->buf); in dc_kbd_callback()
|
/linux-4.1.27/arch/powerpc/kernel/ |
D | ppc32.h | 28 unsigned int mq; /* 601 only (not used at present) */ member
|
D | process.c | 1258 regs->mq = 0; in start_thread()
|
D | asm-offsets.c | 328 DEFINE(_MQ, STACK_FRAME_OVERHEAD+offsetof(struct pt_regs, mq)); in main()
|
D | ptrace.c | 108 REG_OFFSET_NAME(mq),
|
/linux-4.1.27/arch/powerpc/include/uapi/asm/ |
D | ptrace.h | 43 unsigned long mq; /* 601 only (not used at present) */ member
|
/linux-4.1.27/Documentation/device-mapper/ |
D | cache-policies.txt | 67 Internally the mq policy determines a promotion threshold. If the hit 95 /dev/sdd 512 0 mq 4 sequential_threshold 1024 random_threshold 8"
|
D | cache.txt | 146 For instance, the 'mq' policy, which is currently the default policy, 298 mq 4 sequential_threshold 1024 random_threshold 8'
|
/linux-4.1.27/drivers/scsi/qla2xxx/ |
D | qla_dbg.c | 629 struct qla2xxx_mq_chain *mq = ptr; in qla25xx_copy_mq() local 635 mq = ptr; in qla25xx_copy_mq() 636 *last_chain = &mq->type; in qla25xx_copy_mq() 637 mq->type = __constant_htonl(DUMP_CHAIN_MQ); in qla25xx_copy_mq() 638 mq->chain_size = __constant_htonl(sizeof(struct qla2xxx_mq_chain)); in qla25xx_copy_mq() 642 mq->count = htonl(que_cnt); in qla25xx_copy_mq() 646 mq->qregs[que_idx] = in qla25xx_copy_mq() 648 mq->qregs[que_idx+1] = in qla25xx_copy_mq() 650 mq->qregs[que_idx+2] = in qla25xx_copy_mq() 652 mq->qregs[que_idx+3] = in qla25xx_copy_mq()
|
/linux-4.1.27/drivers/scsi/ |
D | scsi_lib.c | 584 static void scsi_free_sgtable(struct scsi_data_buffer *sdb, bool mq) in scsi_free_sgtable() argument 586 if (mq && sdb->table.orig_nents <= SCSI_MAX_SG_SEGMENTS) in scsi_free_sgtable() 588 __sg_free_table(&sdb->table, SCSI_MAX_SG_SEGMENTS, mq, scsi_sg_free); in scsi_free_sgtable() 591 static int scsi_alloc_sgtable(struct scsi_data_buffer *sdb, int nents, bool mq) in scsi_alloc_sgtable() argument 598 if (mq) { in scsi_alloc_sgtable() 610 scsi_free_sgtable(sdb, mq); in scsi_alloc_sgtable()
|
D | Kconfig | 49 bool "SCSI: use blk-mq I/O path by default" 52 This option enables the new blk-mq based I/O path for SCSI 707 This is equivalent to the "eata=mq:8" boot option. 1426 This is equivalent to the "u14-34f=mq:8" boot option.
|
/linux-4.1.27/arch/ia64/include/asm/sn/ |
D | sn_sal.h | 1201 sn_mq_watchlist_alloc(int blade, void *mq, unsigned int mq_size, in sn_mq_watchlist_alloc() argument 1209 addr = (unsigned long)mq; in sn_mq_watchlist_alloc()
|
/linux-4.1.27/arch/powerpc/boot/dts/ |
D | icon.dts | 108 MQ0: mq { 109 compatible = "ibm,mq-440spe";
|
D | katmai.dts | 112 MQ0: mq { 113 compatible = "ibm,mq-440spe";
|
/linux-4.1.27/drivers/scsi/lpfc/ |
D | lpfc_sli.c | 13214 lpfc_mq_create_fb_init(struct lpfc_hba *phba, struct lpfc_queue *mq, in lpfc_mq_create_fb_init() argument 13228 mq->page_count); in lpfc_mq_create_fb_init() 13232 switch (mq->entry_count) { in lpfc_mq_create_fb_init() 13250 list_for_each_entry(dmabuf, &mq->page_list, list) { in lpfc_mq_create_fb_init() 13280 lpfc_mq_create(struct lpfc_hba *phba, struct lpfc_queue *mq, in lpfc_mq_create() argument 13293 if (!mq || !cq) in lpfc_mq_create() 13310 &mq_create_ext->u.request, mq->page_count); in lpfc_mq_create() 13330 switch (mq->entry_count) { in lpfc_mq_create() 13334 mq->entry_count); in lpfc_mq_create() 13335 if (mq->entry_count < 16) { in lpfc_mq_create() [all …]
|