/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_adminq.c | 60 hw->aq.arq.tail = I40E_VF_ARQT1; i40e_adminq_init_regs() 61 hw->aq.arq.head = I40E_VF_ARQH1; i40e_adminq_init_regs() 62 hw->aq.arq.len = I40E_VF_ARQLEN1; i40e_adminq_init_regs() 63 hw->aq.arq.bal = I40E_VF_ARQBAL1; i40e_adminq_init_regs() 64 hw->aq.arq.bah = I40E_VF_ARQBAH1; i40e_adminq_init_regs() 71 hw->aq.arq.tail = I40E_PF_ARQT; i40e_adminq_init_regs() 72 hw->aq.arq.head = I40E_PF_ARQH; i40e_adminq_init_regs() 73 hw->aq.arq.len = I40E_PF_ARQLEN; i40e_adminq_init_regs() 74 hw->aq.arq.bal = I40E_PF_ARQBAL; i40e_adminq_init_regs() 75 hw->aq.arq.bah = I40E_PF_ARQBAH; i40e_adminq_init_regs() 114 ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf, i40e_alloc_adminq_arq_ring() 144 i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf); i40e_free_adminq_arq() 163 ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head, i40e_alloc_arq_bufs() 167 hw->aq.arq.r.arq_bi = (struct i40e_dma_mem *)hw->aq.arq.dma_head.va; i40e_alloc_arq_bufs() 171 bi = &hw->aq.arq.r.arq_bi[i]; i40e_alloc_arq_bufs() 180 desc = I40E_ADMINQ_DESC(hw->aq.arq, i); i40e_alloc_arq_bufs() 208 i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]); i40e_alloc_arq_bufs() 209 i40e_free_virt_mem(hw, &hw->aq.arq.dma_head); i40e_alloc_arq_bufs() 264 i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]); i40e_free_arq_bufs() 267 i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf); i40e_free_arq_bufs() 270 i40e_free_virt_mem(hw, &hw->aq.arq.dma_head); i40e_free_arq_bufs() 337 wr32(hw, hw->aq.arq.head, 0); i40e_config_arq_regs() 338 wr32(hw, hw->aq.arq.tail, 0); i40e_config_arq_regs() 341 wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries | i40e_config_arq_regs() 343 wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs() 344 wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs() 347 wr32(hw, hw->aq.arq.tail, hw->aq.num_arq_entries - 1); i40e_config_arq_regs() 350 reg = rd32(hw, hw->aq.arq.bal); i40e_config_arq_regs() 351 if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa)) i40e_config_arq_regs() 433 if (hw->aq.arq.count > 0) { i40e_init_arq() 446 hw->aq.arq.next_to_use = 0; i40e_init_arq() 447 hw->aq.arq.next_to_clean = 0; i40e_init_arq() 448 hw->aq.arq.count = hw->aq.num_arq_entries; i40e_init_arq() 518 if (hw->aq.arq.count == 0) i40e_shutdown_arq() 522 wr32(hw, hw->aq.arq.head, 0); i40e_shutdown_arq() 523 wr32(hw, hw->aq.arq.tail, 0); i40e_shutdown_arq() 524 wr32(hw, hw->aq.arq.len, 0); i40e_shutdown_arq() 525 wr32(hw, hw->aq.arq.bal, 0); i40e_shutdown_arq() 526 wr32(hw, hw->aq.arq.bah, 0); i40e_shutdown_arq() 531 hw->aq.arq.count = 0; /* to indicate uninitialized queue */ i40e_shutdown_arq() 938 u16 ntc = hw->aq.arq.next_to_clean; i40e_clean_arq_element() 950 ntu = (rd32(hw, hw->aq.arq.head) & I40E_PF_ARQH_ARQH_MASK); i40e_clean_arq_element() 958 desc = I40E_ADMINQ_DESC(hw->aq.arq, ntc); i40e_clean_arq_element() 976 memcpy(e->msg_buf, hw->aq.arq.r.arq_bi[desc_idx].va, i40e_clean_arq_element() 987 bi = &hw->aq.arq.r.arq_bi[ntc]; i40e_clean_arq_element() 998 wr32(hw, hw->aq.arq.tail, ntc); i40e_clean_arq_element() 1003 hw->aq.arq.next_to_clean = ntc; i40e_clean_arq_element() 1004 hw->aq.arq.next_to_use = ntu; i40e_clean_arq_element() 1009 *pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc); i40e_clean_arq_element() 1030 hw->aq.arq.next_to_use = 0; i40e_resume_aq() 1031 hw->aq.arq.next_to_clean = 0; i40e_resume_aq()
|
H A D | i40e_adminq.h | 87 struct i40e_adminq_ring arq; /* receive queue */ member in struct:i40e_adminq_info
|
H A D | i40e_debugfs.c | 199 memcpy(p, pf->hw.aq.arq.desc_buf.va, len); i40e_dbg_dump_write() 755 ring = &(hw->aq.arq); i40e_dbg_dump_aq_desc()
|
H A D | i40e_virtchnl_pf.c | 1858 * called from the common aeq/arq handler to
|
H A D | i40e_main.c | 5842 val = rd32(&pf->hw, pf->hw.aq.arq.len); i40e_clean_adminq_subtask() 5857 wr32(&pf->hw, pf->hw.aq.arq.len, val); i40e_clean_adminq_subtask()
|
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/ |
H A D | i40e_adminq.c | 58 hw->aq.arq.tail = I40E_VF_ARQT1; i40e_adminq_init_regs() 59 hw->aq.arq.head = I40E_VF_ARQH1; i40e_adminq_init_regs() 60 hw->aq.arq.len = I40E_VF_ARQLEN1; i40e_adminq_init_regs() 61 hw->aq.arq.bal = I40E_VF_ARQBAL1; i40e_adminq_init_regs() 62 hw->aq.arq.bah = I40E_VF_ARQBAH1; i40e_adminq_init_regs() 69 hw->aq.arq.tail = I40E_PF_ARQT; i40e_adminq_init_regs() 70 hw->aq.arq.head = I40E_PF_ARQH; i40e_adminq_init_regs() 71 hw->aq.arq.len = I40E_PF_ARQLEN; i40e_adminq_init_regs() 72 hw->aq.arq.bal = I40E_PF_ARQBAL; i40e_adminq_init_regs() 73 hw->aq.arq.bah = I40E_PF_ARQBAH; i40e_adminq_init_regs() 112 ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf, i40e_alloc_adminq_arq_ring() 142 i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf); i40e_free_adminq_arq() 161 ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head, i40e_alloc_arq_bufs() 165 hw->aq.arq.r.arq_bi = (struct i40e_dma_mem *)hw->aq.arq.dma_head.va; i40e_alloc_arq_bufs() 169 bi = &hw->aq.arq.r.arq_bi[i]; i40e_alloc_arq_bufs() 178 desc = I40E_ADMINQ_DESC(hw->aq.arq, i); i40e_alloc_arq_bufs() 206 i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]); i40e_alloc_arq_bufs() 207 i40e_free_virt_mem(hw, &hw->aq.arq.dma_head); i40e_alloc_arq_bufs() 262 i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]); i40e_free_arq_bufs() 265 i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf); i40e_free_arq_bufs() 268 i40e_free_virt_mem(hw, &hw->aq.arq.dma_head); i40e_free_arq_bufs() 335 wr32(hw, hw->aq.arq.head, 0); i40e_config_arq_regs() 336 wr32(hw, hw->aq.arq.tail, 0); i40e_config_arq_regs() 339 wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries | i40e_config_arq_regs() 341 wr32(hw, hw->aq.arq.bal, lower_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs() 342 wr32(hw, hw->aq.arq.bah, upper_32_bits(hw->aq.arq.desc_buf.pa)); i40e_config_arq_regs() 345 wr32(hw, hw->aq.arq.tail, hw->aq.num_arq_entries - 1); i40e_config_arq_regs() 348 reg = rd32(hw, hw->aq.arq.bal); i40e_config_arq_regs() 349 if (reg != lower_32_bits(hw->aq.arq.desc_buf.pa)) i40e_config_arq_regs() 431 if (hw->aq.arq.count > 0) { i40e_init_arq() 444 hw->aq.arq.next_to_use = 0; i40e_init_arq() 445 hw->aq.arq.next_to_clean = 0; i40e_init_arq() 446 hw->aq.arq.count = hw->aq.num_arq_entries; i40e_init_arq() 516 if (hw->aq.arq.count == 0) i40e_shutdown_arq() 520 wr32(hw, hw->aq.arq.head, 0); i40e_shutdown_arq() 521 wr32(hw, hw->aq.arq.tail, 0); i40e_shutdown_arq() 522 wr32(hw, hw->aq.arq.len, 0); i40e_shutdown_arq() 523 wr32(hw, hw->aq.arq.bal, 0); i40e_shutdown_arq() 524 wr32(hw, hw->aq.arq.bah, 0); i40e_shutdown_arq() 529 hw->aq.arq.count = 0; /* to indicate uninitialized queue */ i40e_shutdown_arq() 890 u16 ntc = hw->aq.arq.next_to_clean; i40evf_clean_arq_element() 902 ntu = (rd32(hw, hw->aq.arq.head) & I40E_PF_ARQH_ARQH_MASK); i40evf_clean_arq_element() 910 desc = I40E_ADMINQ_DESC(hw->aq.arq, ntc); i40evf_clean_arq_element() 928 memcpy(e->msg_buf, hw->aq.arq.r.arq_bi[desc_idx].va, i40evf_clean_arq_element() 939 bi = &hw->aq.arq.r.arq_bi[ntc]; i40evf_clean_arq_element() 950 wr32(hw, hw->aq.arq.tail, ntc); i40evf_clean_arq_element() 955 hw->aq.arq.next_to_clean = ntc; i40evf_clean_arq_element() 956 hw->aq.arq.next_to_use = ntu; i40evf_clean_arq_element() 961 *pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc); i40evf_clean_arq_element() 975 hw->aq.arq.next_to_use = 0; i40evf_resume_aq() 976 hw->aq.arq.next_to_clean = 0; i40evf_resume_aq()
|
H A D | i40e_adminq.h | 87 struct i40e_adminq_ring arq; /* receive queue */ member in struct:i40e_adminq_info
|
H A D | i40evf_main.c | 1671 val = rd32(hw, hw->aq.arq.len); i40evf_adminq_task() 1686 wr32(hw, hw->aq.arq.len, val); i40evf_adminq_task()
|
/linux-4.1.27/drivers/crypto/ |
H A D | mxs-dcp.c | 249 static int mxs_dcp_aes_block_crypt(struct crypto_async_request *arq) mxs_dcp_aes_block_crypt() argument 253 struct ablkcipher_request *req = ablkcipher_request_cast(arq); mxs_dcp_aes_block_crypt() 254 struct dcp_async_ctx *actx = crypto_tfm_ctx(arq->tfm); mxs_dcp_aes_block_crypt() 348 struct crypto_async_request *arq; dcp_chan_thread_aes() local 357 arq = crypto_dequeue_request(&sdcp->queue[chan]); dcp_chan_thread_aes() 363 if (arq) { dcp_chan_thread_aes() 364 ret = mxs_dcp_aes_block_crypt(arq); dcp_chan_thread_aes() 365 arq->complete(arq, ret); dcp_chan_thread_aes() 398 struct crypto_async_request *arq = &req->base; mxs_dcp_aes_enqueue() local 399 struct dcp_async_ctx *actx = crypto_tfm_ctx(arq->tfm); mxs_dcp_aes_enqueue() 559 static int dcp_sha_req_to_buf(struct crypto_async_request *arq) dcp_sha_req_to_buf() argument 563 struct ahash_request *req = ahash_request_cast(arq); dcp_sha_req_to_buf() 641 struct crypto_async_request *arq; dcp_chan_thread_sha() local 653 arq = crypto_dequeue_request(&sdcp->queue[chan]); dcp_chan_thread_sha() 659 if (arq) { dcp_chan_thread_sha() 660 req = ahash_request_cast(arq); dcp_chan_thread_sha() 663 ret = dcp_sha_req_to_buf(arq); dcp_chan_thread_sha() 665 arq->complete(arq, ret); dcp_chan_thread_sha()
|
/linux-4.1.27/net/netfilter/ |
H A D | nf_conntrack_h323_main.c | 1535 unsigned char **data, AdmissionRequest *arq) process_arq() 1546 if ((arq->options & eAdmissionRequest_destCallSignalAddress) && process_arq() 1547 get_h225_addr(ct, *data, &arq->destCallSignalAddress, process_arq() 1555 &arq->destCallSignalAddress, process_arq() 1560 if ((arq->options & eAdmissionRequest_srcCallSignalAddress) && process_arq() 1561 get_h225_addr(ct, *data, &arq->srcCallSignalAddress, process_arq() 1568 &arq->srcCallSignalAddress, process_arq() 1532 process_arq(struct sk_buff *skb, struct nf_conn *ct, enum ip_conntrack_info ctinfo, unsigned int protoff, unsigned char **data, AdmissionRequest *arq) process_arq() argument
|
/linux-4.1.27/drivers/scsi/ |
H A D | qlogicpti.c | 296 * disconnect, parity, arq, reneg on reset, and, oddly enough qlogicpti_set_hostdev_defaults()
|