Home
last modified time | relevance | path

Searched refs:wc (Results 1 – 115 of 115) sorted by relevance

/linux-4.1.27/include/math-emu/
Dop-common.h27 #define _FP_DECL(wc, X) \ argument
29 _FP_FRAC_DECL_##wc(X)
36 #define _FP_UNPACK_CANONICAL(fs, wc, X) \ argument
42 _FP_FRAC_SLL_##wc(X, _FP_WORKBITS); \
48 if (_FP_FRAC_ZEROP_##wc(X)) \
54 _FP_FRAC_CLZ_##wc(_shift, X); \
56 _FP_FRAC_SLL_##wc(X, (_shift+_FP_WORKBITS)); \
69 if (_FP_FRAC_ZEROP_##wc(X)) \
89 #define _FP_PACK_CANONICAL(fs, wc, X) \ argument
97 _FP_ROUND(wc, X); \
[all …]
Dsoft-fp.h135 #define _FP_ROUND_NEAREST(wc, X) \ argument
137 if ((_FP_FRAC_LOW_##wc(X) & 15) != _FP_WORK_ROUND) \
138 _FP_FRAC_ADDI_##wc(X, _FP_WORK_ROUND); \
141 #define _FP_ROUND_ZERO(wc, X) 0 argument
143 #define _FP_ROUND_PINF(wc, X) \ argument
145 if (!X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \
146 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \
149 #define _FP_ROUND_MINF(wc, X) \ argument
151 if (X##_s && (_FP_FRAC_LOW_##wc(X) & 7)) \
152 _FP_FRAC_ADDI_##wc(X, _FP_WORK_LSB); \
[all …]
/linux-4.1.27/drivers/infiniband/hw/cxgb3/
Diwch_cq.c45 struct ib_wc *wc) in iwch_poll_cq_one() argument
81 wc->wr_id = cookie; in iwch_poll_cq_one()
82 wc->qp = &qhp->ibqp; in iwch_poll_cq_one()
83 wc->vendor_err = CQE_STATUS(cqe); in iwch_poll_cq_one()
84 wc->wc_flags = 0; in iwch_poll_cq_one()
94 wc->byte_len = CQE_LEN(cqe); in iwch_poll_cq_one()
96 wc->byte_len = 0; in iwch_poll_cq_one()
97 wc->opcode = IB_WC_RECV; in iwch_poll_cq_one()
100 wc->ex.invalidate_rkey = CQE_WRID_STAG(cqe); in iwch_poll_cq_one()
101 wc->wc_flags |= IB_WC_WITH_INVALIDATE; in iwch_poll_cq_one()
[all …]
Diwch_provider.h334 int iwch_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_cq.c50 struct ipath_cq_wc *wc; in ipath_cq_enter() local
61 wc = cq->queue; in ipath_cq_enter()
62 head = wc->head; in ipath_cq_enter()
68 if (unlikely(next == wc->tail)) { in ipath_cq_enter()
81 wc->uqueue[head].wr_id = entry->wr_id; in ipath_cq_enter()
82 wc->uqueue[head].status = entry->status; in ipath_cq_enter()
83 wc->uqueue[head].opcode = entry->opcode; in ipath_cq_enter()
84 wc->uqueue[head].vendor_err = entry->vendor_err; in ipath_cq_enter()
85 wc->uqueue[head].byte_len = entry->byte_len; in ipath_cq_enter()
86 wc->uqueue[head].ex.imm_data = (__u32 __force) entry->ex.imm_data; in ipath_cq_enter()
[all …]
Dipath_ud.c63 struct ib_wc wc; in ipath_ud_loopback() local
93 memset(&wc, 0, sizeof wc); in ipath_ud_loopback()
94 wc.byte_len = length + sizeof(struct ib_grh); in ipath_ud_loopback()
97 wc.wc_flags = IB_WC_WITH_IMM; in ipath_ud_loopback()
98 wc.ex.imm_data = swqe->wr.ex.imm_data; in ipath_ud_loopback()
141 if (wc.byte_len > rlen) { in ipath_ud_loopback()
149 wc.wr_id = wqe->wr_id; in ipath_ud_loopback()
181 wc.wc_flags |= IB_WC_GRH; in ipath_ud_loopback()
213 wc.status = IB_WC_SUCCESS; in ipath_ud_loopback()
214 wc.opcode = IB_WC_RECV; in ipath_ud_loopback()
[all …]
Dipath_ruc.c127 struct ib_wc wc; in ipath_init_sge() local
145 memset(&wc, 0, sizeof(wc)); in ipath_init_sge()
146 wc.wr_id = wqe->wr_id; in ipath_init_sge()
147 wc.status = IB_WC_LOC_PROT_ERR; in ipath_init_sge()
148 wc.opcode = IB_WC_RECV; in ipath_init_sge()
149 wc.qp = &qp->ibqp; in ipath_init_sge()
151 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in ipath_init_sge()
267 struct ib_wc wc; in ipath_ruc_loopback() local
325 memset(&wc, 0, sizeof wc); in ipath_ruc_loopback()
334 wc.wc_flags = IB_WC_WITH_IMM; in ipath_ruc_loopback()
[all …]
Dipath_uc.c248 struct ib_wc wc; in ipath_uc_rcv() local
285 memset(&wc, 0, sizeof wc); in ipath_uc_rcv()
382 wc.ex.imm_data = *(__be32 *) data; in ipath_uc_rcv()
386 wc.ex.imm_data = ohdr->u.imm_data; in ipath_uc_rcv()
389 wc.wc_flags = IB_WC_WITH_IMM; in ipath_uc_rcv()
404 wc.byte_len = tlen + qp->r_rcv_len; in ipath_uc_rcv()
405 if (unlikely(wc.byte_len > qp->r_len)) { in ipath_uc_rcv()
410 wc.opcode = IB_WC_RECV; in ipath_uc_rcv()
413 wc.wr_id = qp->r_wr_id; in ipath_uc_rcv()
414 wc.status = IB_WC_SUCCESS; in ipath_uc_rcv()
[all …]
Dipath_rc.c861 struct ib_wc wc; in do_rc_ack() local
951 memset(&wc, 0, sizeof wc); in do_rc_ack()
952 wc.wr_id = wqe->wr.wr_id; in do_rc_ack()
953 wc.status = IB_WC_SUCCESS; in do_rc_ack()
954 wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; in do_rc_ack()
955 wc.byte_len = wqe->length; in do_rc_ack()
956 wc.qp = &qp->ibqp; in do_rc_ack()
957 wc.src_qp = qp->remote_qpn; in do_rc_ack()
958 wc.slid = qp->remote_ah_attr.dlid; in do_rc_ack()
959 wc.sl = qp->remote_ah_attr.sl; in do_rc_ack()
[all …]
Dipath_qp.c380 struct ib_wc wc; in ipath_error_qp() local
399 memset(&wc, 0, sizeof(wc)); in ipath_error_qp()
400 wc.qp = &qp->ibqp; in ipath_error_qp()
401 wc.opcode = IB_WC_RECV; in ipath_error_qp()
404 wc.wr_id = qp->r_wr_id; in ipath_error_qp()
405 wc.status = err; in ipath_error_qp()
406 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in ipath_error_qp()
408 wc.status = IB_WC_WR_FLUSH_ERR; in ipath_error_qp()
426 wc.wr_id = get_rwqe_ptr(&qp->r_rq, tail)->wr_id; in ipath_error_qp()
429 ipath_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in ipath_error_qp()
/linux-4.1.27/drivers/infiniband/hw/mlx4/
Dcq.c496 struct ib_wc *wc) in mlx4_ib_handle_error_cqe() argument
510 wc->status = IB_WC_LOC_LEN_ERR; in mlx4_ib_handle_error_cqe()
513 wc->status = IB_WC_LOC_QP_OP_ERR; in mlx4_ib_handle_error_cqe()
516 wc->status = IB_WC_LOC_PROT_ERR; in mlx4_ib_handle_error_cqe()
519 wc->status = IB_WC_WR_FLUSH_ERR; in mlx4_ib_handle_error_cqe()
522 wc->status = IB_WC_MW_BIND_ERR; in mlx4_ib_handle_error_cqe()
525 wc->status = IB_WC_BAD_RESP_ERR; in mlx4_ib_handle_error_cqe()
528 wc->status = IB_WC_LOC_ACCESS_ERR; in mlx4_ib_handle_error_cqe()
531 wc->status = IB_WC_REM_INV_REQ_ERR; in mlx4_ib_handle_error_cqe()
534 wc->status = IB_WC_REM_ACCESS_ERR; in mlx4_ib_handle_error_cqe()
[all …]
Dmad.c462 enum ib_qp_type dest_qpt, struct ib_wc *wc, in mlx4_ib_send_to_slave() argument
497 ret = ib_get_cached_pkey(&dev->ib_dev, port, wc->pkey_index, &cached_pkey); in mlx4_ib_send_to_slave()
552 tun_mad->hdr.flags_src_qp = cpu_to_be32(wc->src_qp & 0xFFFFFF); in mlx4_ib_send_to_slave()
553 tun_mad->hdr.g_ml_path = (grh && (wc->wc_flags & IB_WC_GRH)) ? 0x80 : 0; in mlx4_ib_send_to_slave()
560 if (vlan != wc->vlan_id) in mlx4_ib_send_to_slave()
571 vlan = wc->vlan_id; in mlx4_ib_send_to_slave()
575 memcpy((char *)&tun_mad->hdr.mac_31_0, &(wc->smac[0]), 4); in mlx4_ib_send_to_slave()
576 memcpy((char *)&tun_mad->hdr.slid_mac_47_32, &(wc->smac[4]), 2); in mlx4_ib_send_to_slave()
578 tun_mad->hdr.sl_vid = cpu_to_be16(((u16)(wc->sl)) << 12); in mlx4_ib_send_to_slave()
579 tun_mad->hdr.slid_mac_47_32 = cpu_to_be16(wc->slid); in mlx4_ib_send_to_slave()
[all …]
Dmcg.c229 struct ib_wc wc; in send_mad_to_slave() local
238 if (ib_find_cached_pkey(&dev->ib_dev, ctx->port, IB_DEFAULT_PKEY_FULL, &wc.pkey_index)) in send_mad_to_slave()
240 wc.sl = 0; in send_mad_to_slave()
241 wc.dlid_path_bits = 0; in send_mad_to_slave()
242 wc.port_num = ctx->port; in send_mad_to_slave()
243 wc.slid = ah_attr.dlid; /* opensm lid */ in send_mad_to_slave()
244 wc.src_qp = 1; in send_mad_to_slave()
245 return mlx4_ib_send_to_slave(dev, slave, ctx->port, IB_QPT_GSI, &wc, NULL, mad); in send_mad_to_slave()
Dmlx4_ib.h675 int mlx4_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
763 enum ib_qp_type qpt, struct ib_wc *wc,
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_cq.c53 struct qib_cq_wc *wc; in qib_cq_enter() local
64 wc = cq->queue; in qib_cq_enter()
65 head = wc->head; in qib_cq_enter()
71 if (unlikely(next == wc->tail)) { in qib_cq_enter()
84 wc->uqueue[head].wr_id = entry->wr_id; in qib_cq_enter()
85 wc->uqueue[head].status = entry->status; in qib_cq_enter()
86 wc->uqueue[head].opcode = entry->opcode; in qib_cq_enter()
87 wc->uqueue[head].vendor_err = entry->vendor_err; in qib_cq_enter()
88 wc->uqueue[head].byte_len = entry->byte_len; in qib_cq_enter()
89 wc->uqueue[head].ex.imm_data = in qib_cq_enter()
[all …]
Dqib_ud.c58 struct ib_wc wc; in qib_ud_loopback() local
130 memset(&wc, 0, sizeof(wc)); in qib_ud_loopback()
131 wc.byte_len = length + sizeof(struct ib_grh); in qib_ud_loopback()
134 wc.wc_flags = IB_WC_WITH_IMM; in qib_ud_loopback()
135 wc.ex.imm_data = swqe->wr.ex.imm_data; in qib_ud_loopback()
160 if (unlikely(wc.byte_len > qp->r_len)) { in qib_ud_loopback()
169 wc.wc_flags |= IB_WC_GRH; in qib_ud_loopback()
207 wc.wr_id = qp->r_wr_id; in qib_ud_loopback()
208 wc.status = IB_WC_SUCCESS; in qib_ud_loopback()
209 wc.opcode = IB_WC_RECV; in qib_ud_loopback()
[all …]
Dqib_ruc.c84 struct ib_wc wc; in qib_init_sge() local
116 memset(&wc, 0, sizeof(wc)); in qib_init_sge()
117 wc.wr_id = wqe->wr_id; in qib_init_sge()
118 wc.status = IB_WC_LOC_PROT_ERR; in qib_init_sge()
119 wc.opcode = IB_WC_RECV; in qib_init_sge()
120 wc.qp = &qp->ibqp; in qib_init_sge()
122 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in qib_init_sge()
362 struct ib_wc wc; in qib_ruc_loopback() local
423 memset(&wc, 0, sizeof(wc)); in qib_ruc_loopback()
433 wc.wc_flags = IB_WC_WITH_IMM; in qib_ruc_loopback()
[all …]
Dqib_uc.c250 struct ib_wc wc; in qib_uc_rcv() local
381 wc.ex.imm_data = ohdr->u.imm_data; in qib_uc_rcv()
383 wc.wc_flags = IB_WC_WITH_IMM; in qib_uc_rcv()
387 wc.ex.imm_data = 0; in qib_uc_rcv()
388 wc.wc_flags = 0; in qib_uc_rcv()
398 wc.byte_len = tlen + qp->r_rcv_len; in qib_uc_rcv()
399 if (unlikely(wc.byte_len > qp->r_len)) in qib_uc_rcv()
401 wc.opcode = IB_WC_RECV; in qib_uc_rcv()
405 wc.wr_id = qp->r_wr_id; in qib_uc_rcv()
406 wc.status = IB_WC_SUCCESS; in qib_uc_rcv()
[all …]
Dqib_rc.c973 struct ib_wc wc; in qib_rc_send_complete() local
1020 memset(&wc, 0, sizeof(wc)); in qib_rc_send_complete()
1021 wc.wr_id = wqe->wr.wr_id; in qib_rc_send_complete()
1022 wc.status = IB_WC_SUCCESS; in qib_rc_send_complete()
1023 wc.opcode = ib_qib_wc_opcode[wqe->wr.opcode]; in qib_rc_send_complete()
1024 wc.byte_len = wqe->length; in qib_rc_send_complete()
1025 wc.qp = &qp->ibqp; in qib_rc_send_complete()
1026 qib_cq_enter(to_icq(qp->ibqp.send_cq), &wc, 0); in qib_rc_send_complete()
1058 struct ib_wc wc; in do_rc_completion() local
1076 memset(&wc, 0, sizeof(wc)); in do_rc_completion()
[all …]
Dqib_qp.c477 struct ib_wc wc; in qib_error_qp() local
518 memset(&wc, 0, sizeof(wc)); in qib_error_qp()
519 wc.qp = &qp->ibqp; in qib_error_qp()
520 wc.opcode = IB_WC_RECV; in qib_error_qp()
523 wc.wr_id = qp->r_wr_id; in qib_error_qp()
524 wc.status = err; in qib_error_qp()
525 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in qib_error_qp()
527 wc.status = IB_WC_WR_FLUSH_ERR; in qib_error_qp()
545 wc.wr_id = get_rwqe_ptr(&qp->r_rq, tail)->wr_id; in qib_error_qp()
548 qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, 1); in qib_error_qp()
/linux-4.1.27/drivers/net/ethernet/brocade/bna/
Dbfa_cs.h95 bfa_wc_up(struct bfa_wc *wc) in bfa_wc_up() argument
97 wc->wc_count++; in bfa_wc_up()
101 bfa_wc_down(struct bfa_wc *wc) in bfa_wc_down() argument
103 wc->wc_count--; in bfa_wc_down()
104 if (wc->wc_count == 0) in bfa_wc_down()
105 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down()
110 bfa_wc_init(struct bfa_wc *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument
112 wc->wc_resume = wc_resume; in bfa_wc_init()
113 wc->wc_cbarg = wc_cbarg; in bfa_wc_init()
114 wc->wc_count = 0; in bfa_wc_init()
[all …]
/linux-4.1.27/fs/ocfs2/
Daops.c1251 static void ocfs2_unlock_pages(struct ocfs2_write_ctxt *wc) in ocfs2_unlock_pages() argument
1260 if (wc->w_target_locked) { in ocfs2_unlock_pages()
1261 BUG_ON(!wc->w_target_page); in ocfs2_unlock_pages()
1262 for (i = 0; i < wc->w_num_pages; i++) { in ocfs2_unlock_pages()
1263 if (wc->w_target_page == wc->w_pages[i]) { in ocfs2_unlock_pages()
1264 wc->w_pages[i] = NULL; in ocfs2_unlock_pages()
1268 mark_page_accessed(wc->w_target_page); in ocfs2_unlock_pages()
1269 page_cache_release(wc->w_target_page); in ocfs2_unlock_pages()
1271 ocfs2_unlock_and_free_pages(wc->w_pages, wc->w_num_pages); in ocfs2_unlock_pages()
1274 static void ocfs2_free_write_ctxt(struct ocfs2_write_ctxt *wc) in ocfs2_free_write_ctxt() argument
[all …]
/linux-4.1.27/drivers/infiniband/hw/mlx5/
Dcq.c120 static void handle_good_req(struct ib_wc *wc, struct mlx5_cqe64 *cqe, in handle_good_req() argument
123 wc->wc_flags = 0; in handle_good_req()
126 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req()
128 wc->opcode = IB_WC_RDMA_WRITE; in handle_good_req()
131 wc->wc_flags |= IB_WC_WITH_IMM; in handle_good_req()
134 wc->opcode = IB_WC_SEND; in handle_good_req()
137 wc->opcode = IB_WC_RDMA_READ; in handle_good_req()
138 wc->byte_len = be32_to_cpu(cqe->byte_cnt); in handle_good_req()
141 wc->opcode = IB_WC_COMP_SWAP; in handle_good_req()
142 wc->byte_len = 8; in handle_good_req()
[all …]
Dmr.c732 struct ib_wc wc; in mlx5_umr_cq_handler() local
736 err = ib_poll_cq(cq, 1, &wc); in mlx5_umr_cq_handler()
744 context = (struct mlx5_ib_umr_context *) (unsigned long) wc.wr_id; in mlx5_umr_cq_handler()
745 context->status = wc.status; in mlx5_umr_cq_handler()
Dmlx5_ib.h563 int mlx5_ib_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
/linux-4.1.27/fs/ntfs/
Dunistr.c264 wchar_t wc; in ntfs_nlstoucs() local
273 &wc); in ntfs_nlstoucs()
276 if (likely(wc)) { in ntfs_nlstoucs()
277 ucs[o++] = cpu_to_le16(wc); in ntfs_nlstoucs()
341 int i, o, ns_len, wc; in ntfs_ucstonls() local
348 wc = -ENAMETOOLONG; in ntfs_ucstonls()
358 retry: wc = nls->uni2char(le16_to_cpu(ins[i]), ns + o, in ntfs_ucstonls()
360 if (wc > 0) { in ntfs_ucstonls()
361 o += wc; in ntfs_ucstonls()
363 } else if (!wc) in ntfs_ucstonls()
[all …]
/linux-4.1.27/arch/c6x/platforms/
Dcache.c141 unsigned int wc = 0; in cache_block_operation() local
143 for (; wcnt; wcnt -= wc, start += wc) { in cache_block_operation()
163 wc = 0xffff; in cache_block_operation()
165 wc = wcnt; in cache_block_operation()
168 imcr_set(wc_reg, wc & 0xffff); in cache_block_operation()
186 unsigned int wc = 0; in cache_block_operation_nowait() local
188 for (; wcnt; wcnt -= wc, start += wc) { in cache_block_operation_nowait()
195 wc = 0xffff; in cache_block_operation_nowait()
197 wc = wcnt; in cache_block_operation_nowait()
200 imcr_set(wc_reg, wc & 0xffff); in cache_block_operation_nowait()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_reqs.c625 static inline int ehca_poll_cq_one(struct ib_cq *cq, struct ib_wc *wc) in ehca_poll_cq_one() argument
705 wc->qp = &my_qp->ib_qp; in ehca_poll_cq_one()
743 wc->wr_id = replace_wr_id(cqe->work_request_id, qmap_entry->app_wr_id); in ehca_poll_cq_one()
758 wc->opcode = ib_wc_opcode[cqe->optype]-1; in ehca_poll_cq_one()
759 if (unlikely(wc->opcode == -1)) { in ehca_poll_cq_one()
773 map_ib_wc_status(cqe->status, &wc->status); in ehca_poll_cq_one()
774 wc->vendor_err = wc->status; in ehca_poll_cq_one()
776 wc->status = IB_WC_SUCCESS; in ehca_poll_cq_one()
778 wc->byte_len = cqe->nr_bytes_transferred; in ehca_poll_cq_one()
779 wc->pkey_index = cqe->pkey_index; in ehca_poll_cq_one()
[all …]
Dehca_iverbs.h137 int ehca_poll_cq(struct ib_cq *cq, int num_entries, struct ib_wc *wc);
/linux-4.1.27/drivers/scsi/bfa/
Dbfa_cs.h293 bfa_wc_up(struct bfa_wc_s *wc) in bfa_wc_up() argument
295 wc->wc_count++; in bfa_wc_up()
299 bfa_wc_down(struct bfa_wc_s *wc) in bfa_wc_down() argument
301 wc->wc_count--; in bfa_wc_down()
302 if (wc->wc_count == 0) in bfa_wc_down()
303 wc->wc_resume(wc->wc_cbarg); in bfa_wc_down()
310 bfa_wc_init(struct bfa_wc_s *wc, bfa_wc_resume_t wc_resume, void *wc_cbarg) in bfa_wc_init() argument
312 wc->wc_resume = wc_resume; in bfa_wc_init()
313 wc->wc_cbarg = wc_cbarg; in bfa_wc_init()
314 wc->wc_count = 0; in bfa_wc_init()
[all …]
Dbfa_fcs.c128 bfa_wc_init(&fcs->wc, bfa_fcs_exit_comp, fcs); in bfa_fcs_stop()
129 bfa_wc_up(&fcs->wc); in bfa_fcs_stop()
131 bfa_wc_wait(&fcs->wc); in bfa_fcs_stop()
185 bfa_wc_init(&fcs->wc, bfa_fcs_exit_comp, fcs); in bfa_fcs_exit()
193 bfa_wc_up(&fcs->wc); in bfa_fcs_exit()
198 bfa_wc_wait(&fcs->wc); in bfa_fcs_exit()
726 bfa_wc_down(&fabric->fcs->wc); in bfa_fcs_fabric_sm_deleting()
792 bfa_wc_down(&(fabric->fcs)->wc); in bfa_fcs_fabric_sm_cleanup()
1106 bfa_wc_wait(&fabric->wc); in bfa_fcs_fabric_delete()
1153 bfa_wc_init(&fabric->wc, bfa_fcs_fabric_delete_comp, fabric); in bfa_fcs_fabric_attach()
[all …]
Dbfa_fcpim.h213 struct bfa_wc_s wc; /* waiting counter */ member
239 struct bfa_wc_s wc; /* waiting counter */ member
Dbfa_fcs.h159 struct bfa_wc_s wc; /* waiting counter for events */ member
206 struct bfa_wc_s wc; /* wait counter for delete */ member
706 struct bfa_wc_s wc; /* waiting counter */ member
Dbfa_fcpim.c1025 bfa_wc_init(&itnim->wc, bfa_itnim_cleanp_comp, itnim); in bfa_itnim_cleanup()
1037 bfa_wc_up(&itnim->wc); in bfa_itnim_cleanup()
1043 bfa_wc_up(&itnim->wc); in bfa_itnim_cleanup()
1047 bfa_wc_wait(&itnim->wc); in bfa_itnim_cleanup()
1095 bfa_wc_down(&itnim->wc); in bfa_itnim_iodone()
1101 bfa_wc_down(&itnim->wc); in bfa_itnim_tskdone()
2698 bfa_wc_down(&ioim->iosp->tskim->wc); in bfa_ioim_notify_cleanup()
3375 bfa_wc_init(&tskim->wc, bfa_tskim_cleanp_comp, tskim); in bfa_tskim_cleanup_ios()
3379 bfa_wc_up(&tskim->wc); in bfa_tskim_cleanup_ios()
3383 bfa_wc_wait(&tskim->wc); in bfa_tskim_cleanup_ios()
[all …]
Dbfa_fcs_lport.c757 bfa_wc_down(&port->fabric->wc); in bfa_fcs_lport_deleted()
/linux-4.1.27/sound/isa/wavefront/
Dwavefront_synth.c1495 wavefront_control *wc) in wavefront_synth_control() argument
1503 "cmd 0x%x\n", wc->cmd); in wavefront_synth_control()
1507 switch (wc->cmd) { in wavefront_synth_control()
1522 wc->rbuf[0] = dev->interrupts_are_midi; in wavefront_synth_control()
1526 dev->rom_samples_rdonly = wc->wbuf[0]; in wavefront_synth_control()
1527 wc->status = 0; in wavefront_synth_control()
1531 i = wc->wbuf[0] | (wc->wbuf[1] << 7); in wavefront_synth_control()
1535 wc->status = EINVAL; in wavefront_synth_control()
1538 wc->rbuf[0] = dev->sample_status[i]; in wavefront_synth_control()
1539 wc->status = 0; in wavefront_synth_control()
[all …]
/linux-4.1.27/tools/testing/selftests/ftrace/
Dftracetest261 prlog "# of passed: " `echo $PASSED_CASES | wc -w`
262 prlog "# of failed: " `echo $FAILED_CASES | wc -w`
263 prlog "# of unresolved: " `echo $UNRESOLVED_CASES | wc -w`
264 prlog "# of untested: " `echo $UNTESTED_CASES | wc -w`
265 prlog "# of unsupported: " `echo $UNSUPPORTED_CASES | wc -w`
266 prlog "# of xfailed: " `echo $XFAILED_CASES | wc -w`
267 prlog "# of undefined(test bug): " `echo $UNDEFINED_CASES | wc -w`
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
Dcq.c684 static int c4iw_poll_cq_one(struct c4iw_cq *chp, struct ib_wc *wc) in c4iw_poll_cq_one() argument
710 wc->wr_id = cookie; in c4iw_poll_cq_one()
711 wc->qp = &qhp->ibqp; in c4iw_poll_cq_one()
712 wc->vendor_err = CQE_STATUS(&cqe); in c4iw_poll_cq_one()
713 wc->wc_flags = 0; in c4iw_poll_cq_one()
722 wc->byte_len = CQE_LEN(&cqe); in c4iw_poll_cq_one()
724 wc->byte_len = 0; in c4iw_poll_cq_one()
725 wc->opcode = IB_WC_RECV; in c4iw_poll_cq_one()
728 wc->ex.invalidate_rkey = CQE_WRID_STAG(&cqe); in c4iw_poll_cq_one()
729 wc->wc_flags |= IB_WC_WITH_INVALIDATE; in c4iw_poll_cq_one()
[all …]
Diw_cxgb4.h957 int c4iw_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc);
/linux-4.1.27/tools/testing/selftests/ftrace/test.d/ftrace/
Dfgraph-filter-stack.tc61 count=`cat trace | grep '()' | grep -v schedule | wc -l`
68 count=`cat trace | grep 'schedule()' | wc -l`
78 count=`cat trace | grep '()' | grep -v schedule | wc -l`
84 count=`cat trace | grep 'schedule()' | wc -l`
Dfgraph-filter.tc39 count=`cat trace | grep '()' | grep -v schedule | wc -l`
45 count=`cat trace | grep 'schedule()' | wc -l`
/linux-4.1.27/drivers/infiniband/ulp/ipoib/
Dipoib_ib.c176 static void ipoib_ib_handle_rx_wc(struct net_device *dev, struct ib_wc *wc) in ipoib_ib_handle_rx_wc() argument
179 unsigned int wr_id = wc->wr_id & ~IPOIB_OP_RECV; in ipoib_ib_handle_rx_wc()
185 wr_id, wc->status); in ipoib_ib_handle_rx_wc()
195 if (unlikely(wc->status != IB_WC_SUCCESS)) { in ipoib_ib_handle_rx_wc()
196 if (wc->status != IB_WC_WR_FLUSH_ERR) in ipoib_ib_handle_rx_wc()
199 wc->status, wr_id, wc->vendor_err); in ipoib_ib_handle_rx_wc()
210 if (wc->slid == priv->local_lid && wc->src_qp == priv->qp->qp_num) in ipoib_ib_handle_rx_wc()
226 wc->byte_len, wc->slid); in ipoib_ib_handle_rx_wc()
230 skb_put(skb, wc->byte_len); in ipoib_ib_handle_rx_wc()
235 if (!(wc->wc_flags & IB_WC_GRH) || dgid->raw[0] != 0xff) in ipoib_ib_handle_rx_wc()
[all …]
Dipoib_cm.c558 void ipoib_cm_handle_rx_wc(struct net_device *dev, struct ib_wc *wc) in ipoib_cm_handle_rx_wc() argument
562 unsigned int wr_id = wc->wr_id & ~(IPOIB_OP_CM | IPOIB_OP_RECV); in ipoib_cm_handle_rx_wc()
572 wr_id, wc->status); in ipoib_cm_handle_rx_wc()
587 p = wc->qp->qp_context; in ipoib_cm_handle_rx_wc()
594 if (unlikely(wc->status != IB_WC_SUCCESS)) { in ipoib_cm_handle_rx_wc()
597 wc->status, wr_id, wc->vendor_err); in ipoib_cm_handle_rx_wc()
624 if (wc->byte_len < IPOIB_CM_COPYBREAK) { in ipoib_cm_handle_rx_wc()
625 int dlen = wc->byte_len; in ipoib_cm_handle_rx_wc()
641 frags = PAGE_ALIGN(wc->byte_len - min(wc->byte_len, in ipoib_cm_handle_rx_wc()
660 wc->byte_len, wc->slid); in ipoib_cm_handle_rx_wc()
[all …]
Dipoib.h616 void ipoib_cm_handle_rx_wc(struct net_device *dev, struct ib_wc *wc);
617 void ipoib_cm_handle_tx_wc(struct net_device *dev, struct ib_wc *wc);
714 static inline void ipoib_cm_handle_rx_wc(struct net_device *dev, struct ib_wc *wc) in ipoib_cm_handle_rx_wc() argument
718 static inline void ipoib_cm_handle_tx_wc(struct net_device *dev, struct ib_wc *wc) in ipoib_cm_handle_tx_wc() argument
/linux-4.1.27/drivers/infiniband/core/
Dmad.c704 struct ib_wc *wc) in build_smp_wc() argument
706 memset(wc, 0, sizeof *wc); in build_smp_wc()
707 wc->wr_id = wr_id; in build_smp_wc()
708 wc->status = IB_WC_SUCCESS; in build_smp_wc()
709 wc->opcode = IB_WC_RECV; in build_smp_wc()
710 wc->pkey_index = pkey_index; in build_smp_wc()
711 wc->byte_len = sizeof(struct ib_mad) + sizeof(struct ib_grh); in build_smp_wc()
712 wc->src_qp = IB_QP0; in build_smp_wc()
713 wc->qp = qp; in build_smp_wc()
714 wc->slid = slid; in build_smp_wc()
[all …]
Dmad_rmpp.c140 msg = ib_create_send_mad(&rmpp_recv->agent->agent, recv_wc->wc->src_qp, in ack_recv()
141 recv_wc->wc->pkey_index, 1, hdr_len, in ack_recv()
160 ah = ib_create_ah_from_wc(agent->qp->pd, recv_wc->wc, in alloc_response_msg()
166 msg = ib_create_send_mad(agent, recv_wc->wc->src_qp, in alloc_response_msg()
167 recv_wc->wc->pkey_index, 1, in alloc_response_msg()
291 mad_recv_wc->wc, in create_rmpp_recv()
314 rmpp_recv->src_qp = mad_recv_wc->wc->src_qp; in create_rmpp_recv()
315 rmpp_recv->slid = mad_recv_wc->wc->slid; in create_rmpp_recv()
334 rmpp_recv->src_qp == mad_recv_wc->wc->src_qp && in find_rmpp_recv()
335 rmpp_recv->slid == mad_recv_wc->wc->slid && in find_rmpp_recv()
[all …]
Dagent.c82 struct ib_wc *wc, struct ib_device *device, in agent_send_response() argument
102 ah = ib_create_ah_from_wc(agent->qp->pd, wc, grh, port_num); in agent_send_response()
109 send_buf = ib_create_send_mad(agent, wc->src_qp, wc->pkey_index, 0, in agent_send_response()
Dverbs.c195 int ib_init_ah_from_wc(struct ib_device *device, u8 port_num, struct ib_wc *wc, in ib_init_ah_from_wc() argument
206 if (!(wc->wc_flags & IB_WC_GRH)) in ib_init_ah_from_wc()
209 if (wc->wc_flags & IB_WC_WITH_SMAC && in ib_init_ah_from_wc()
210 wc->wc_flags & IB_WC_WITH_VLAN) { in ib_init_ah_from_wc()
211 memcpy(ah_attr->dmac, wc->smac, ETH_ALEN); in ib_init_ah_from_wc()
212 ah_attr->vlan_id = wc->vlan_id; in ib_init_ah_from_wc()
223 ah_attr->dlid = wc->slid; in ib_init_ah_from_wc()
224 ah_attr->sl = wc->sl; in ib_init_ah_from_wc()
225 ah_attr->src_path_bits = wc->dlid_path_bits; in ib_init_ah_from_wc()
228 if (wc->wc_flags & IB_WC_GRH) { in ib_init_ah_from_wc()
[all …]
Dagent.h48 struct ib_wc *wc, struct ib_device *device,
Duser_mad.c219 if (mad_recv_wc->wc->status != IB_WC_SUCCESS) in recv_handler()
231 packet->mad.hdr.qpn = cpu_to_be32(mad_recv_wc->wc->src_qp); in recv_handler()
232 packet->mad.hdr.lid = cpu_to_be16(mad_recv_wc->wc->slid); in recv_handler()
233 packet->mad.hdr.sl = mad_recv_wc->wc->sl; in recv_handler()
234 packet->mad.hdr.path_bits = mad_recv_wc->wc->dlid_path_bits; in recv_handler()
235 packet->mad.hdr.pkey_index = mad_recv_wc->wc->pkey_index; in recv_handler()
236 packet->mad.hdr.grh_present = !!(mad_recv_wc->wc->wc_flags & IB_WC_GRH); in recv_handler()
241 mad_recv_wc->wc, mad_recv_wc->recv_buf.grh, in recv_handler()
Duverbs_cmd.c1471 static int copy_wc_to_user(void __user *dest, struct ib_wc *wc) in copy_wc_to_user() argument
1475 tmp.wr_id = wc->wr_id; in copy_wc_to_user()
1476 tmp.status = wc->status; in copy_wc_to_user()
1477 tmp.opcode = wc->opcode; in copy_wc_to_user()
1478 tmp.vendor_err = wc->vendor_err; in copy_wc_to_user()
1479 tmp.byte_len = wc->byte_len; in copy_wc_to_user()
1480 tmp.ex.imm_data = (__u32 __force) wc->ex.imm_data; in copy_wc_to_user()
1481 tmp.qp_num = wc->qp->qp_num; in copy_wc_to_user()
1482 tmp.src_qp = wc->src_qp; in copy_wc_to_user()
1483 tmp.wc_flags = wc->wc_flags; in copy_wc_to_user()
[all …]
Dcm.c293 ah = ib_create_ah_from_wc(port->mad_agent->qp->pd, mad_recv_wc->wc, in cm_alloc_response_msg()
298 m = ib_create_send_mad(port->mad_agent, 1, mad_recv_wc->wc->pkey_index, in cm_alloc_response_msg()
343 static void cm_init_av_for_response(struct cm_port *port, struct ib_wc *wc, in cm_init_av_for_response() argument
347 av->pkey_index = wc->pkey_index; in cm_init_av_for_response()
348 ib_init_ah_from_wc(port->cm_dev->ib_device, port->port_num, wc, in cm_init_av_for_response()
1502 static void cm_process_routed_req(struct cm_req_msg *req_msg, struct ib_wc *wc) in cm_process_routed_req() argument
1506 req_msg->primary_local_lid = cpu_to_be16(wc->slid); in cm_process_routed_req()
1507 cm_req_set_primary_sl(req_msg, wc->sl); in cm_process_routed_req()
1511 req_msg->primary_remote_lid = cpu_to_be16(wc->dlid_path_bits); in cm_process_routed_req()
1516 req_msg->alt_local_lid = cpu_to_be16(wc->slid); in cm_process_routed_req()
[all …]
Dmad_priv.h72 struct ib_wc wc; member
Dsa_query.c1136 mad_buf = (void *) (unsigned long) mad_recv_wc->wc->wr_id; in recv_handler()
1140 if (mad_recv_wc->wc->status == IB_WC_SUCCESS) in recv_handler()
/linux-4.1.27/drivers/net/wireless/hostap/
Dhostap_plx.c173 u8 *buf, int wc) in hfa384x_outsw_debug() argument
183 prism2_io_debug_add(dev, PRISM2_IO_DEBUG_CMD_OUTSW, a, wc); in hfa384x_outsw_debug()
184 outsw(dev->base_addr + a, buf, wc); in hfa384x_outsw_debug()
189 u8 *buf, int wc) in hfa384x_insw_debug() argument
199 prism2_io_debug_add(dev, PRISM2_IO_DEBUG_CMD_INSW, a, wc); in hfa384x_insw_debug()
200 insw(dev->base_addr + a, buf, wc); in hfa384x_insw_debug()
208 #define HFA384X_OUTSW(a, buf, wc) hfa384x_outsw_debug(dev, (a), (buf), (wc)) argument
209 #define HFA384X_INSW(a, buf, wc) hfa384x_insw_debug(dev, (a), (buf), (wc)) argument
217 #define HFA384X_INSW(a, buf, wc) insw(dev->base_addr + (a), buf, wc) argument
218 #define HFA384X_OUTSW(a, buf, wc) outsw(dev->base_addr + (a), buf, wc) argument
Dhostap_cs.c107 u8 *buf, int wc) in hfa384x_outsw_debug() argument
116 prism2_io_debug_add(dev, PRISM2_IO_DEBUG_CMD_OUTSW, a, wc); in hfa384x_outsw_debug()
117 outsw(dev->base_addr + a, buf, wc); in hfa384x_outsw_debug()
122 u8 *buf, int wc) in hfa384x_insw_debug() argument
131 prism2_io_debug_add(dev, PRISM2_IO_DEBUG_CMD_INSW, a, wc); in hfa384x_insw_debug()
132 insw(dev->base_addr + a, buf, wc); in hfa384x_insw_debug()
140 #define HFA384X_OUTSW(a, buf, wc) hfa384x_outsw_debug(dev, (a), (buf), (wc)) argument
141 #define HFA384X_INSW(a, buf, wc) hfa384x_insw_debug(dev, (a), (buf), (wc)) argument
149 #define HFA384X_INSW(a, buf, wc) insw(dev->base_addr + (a), buf, wc) argument
150 #define HFA384X_OUTSW(a, buf, wc) outsw(dev->base_addr + (a), buf, wc) argument
/linux-4.1.27/tools/testing/selftests/ftrace/test.d/event/
Devent-enable.tc31 count=`cat trace | grep sched_switch | wc -l`
42 count=`cat trace | grep sched_switch | wc -l`
53 count=`cat trace | grep sched_switch | wc -l`
/linux-4.1.27/net/rds/
Diw_send.c200 struct ib_wc wc; in rds_iw_send_cq_comp_handler() local
213 while (ib_poll_cq(cq, 1, &wc) > 0) { in rds_iw_send_cq_comp_handler()
215 (unsigned long long)wc.wr_id, wc.status, wc.byte_len, in rds_iw_send_cq_comp_handler()
216 be32_to_cpu(wc.ex.imm_data)); in rds_iw_send_cq_comp_handler()
219 if (wc.status != IB_WC_SUCCESS) { in rds_iw_send_cq_comp_handler()
220 printk(KERN_ERR "WC Error: status = %d opcode = %d\n", wc.status, wc.opcode); in rds_iw_send_cq_comp_handler()
224 if (wc.opcode == IB_WC_LOCAL_INV && wc.wr_id == RDS_IW_LOCAL_INV_WR_ID) { in rds_iw_send_cq_comp_handler()
229 if (wc.opcode == IB_WC_FAST_REG_MR && wc.wr_id == RDS_IW_FAST_REG_WR_ID) { in rds_iw_send_cq_comp_handler()
234 if (wc.wr_id == RDS_IW_ACK_WR_ID) { in rds_iw_send_cq_comp_handler()
243 completed = rds_iw_ring_completed(&ic->i_send_ring, wc.wr_id, oldest); in rds_iw_send_cq_comp_handler()
[all …]
Dib_send.c279 struct ib_wc wc; in rds_ib_send_cq_comp_handler() local
293 while (ib_poll_cq(cq, 1, &wc) > 0) { in rds_ib_send_cq_comp_handler()
295 (unsigned long long)wc.wr_id, wc.status, in rds_ib_send_cq_comp_handler()
296 rds_ib_wc_status_str(wc.status), wc.byte_len, in rds_ib_send_cq_comp_handler()
297 be32_to_cpu(wc.ex.imm_data)); in rds_ib_send_cq_comp_handler()
300 if (wc.wr_id == RDS_IB_ACK_WR_ID) { in rds_ib_send_cq_comp_handler()
309 completed = rds_ib_ring_completed(&ic->i_send_ring, wc.wr_id, oldest); in rds_ib_send_cq_comp_handler()
316 rm = rds_ib_send_unmap_op(ic, send, wc.status); in rds_ib_send_cq_comp_handler()
343 if (wc.status != IB_WC_SUCCESS && rds_conn_up(conn)) { in rds_ib_send_cq_comp_handler()
346 &conn->c_faddr, wc.status, in rds_ib_send_cq_comp_handler()
[all …]
Dib_recv.c953 struct ib_wc wc; in rds_poll_cq() local
956 while (ib_poll_cq(ic->i_recv_cq, 1, &wc) > 0) { in rds_poll_cq()
958 (unsigned long long)wc.wr_id, wc.status, in rds_poll_cq()
959 rds_ib_wc_status_str(wc.status), wc.byte_len, in rds_poll_cq()
960 be32_to_cpu(wc.ex.imm_data)); in rds_poll_cq()
972 if (wc.status == IB_WC_SUCCESS) { in rds_poll_cq()
973 rds_ib_process_recv(conn, recv, wc.byte_len, state); in rds_poll_cq()
980 wc.status, in rds_poll_cq()
981 rds_ib_wc_status_str(wc.status)); in rds_poll_cq()
Diw_recv.c784 struct ib_wc wc; in rds_poll_cq() local
787 while (ib_poll_cq(ic->i_recv_cq, 1, &wc) > 0) { in rds_poll_cq()
789 (unsigned long long)wc.wr_id, wc.status, wc.byte_len, in rds_poll_cq()
790 be32_to_cpu(wc.ex.imm_data)); in rds_poll_cq()
804 if (wc.status == IB_WC_SUCCESS) { in rds_poll_cq()
805 rds_iw_process_recv(conn, recv, wc.byte_len, state); in rds_poll_cq()
810 wc.status); in rds_poll_cq()
/linux-4.1.27/drivers/infiniband/ulp/iser/
Diser_verbs.c1123 struct ib_wc *wc) in iser_handle_comp_error() argument
1125 void *wr_id = (void *)(uintptr_t)wc->wr_id; in iser_handle_comp_error()
1129 if (wc->status != IB_WC_WR_FLUSH_ERR) in iser_handle_comp_error()
1134 if (wc->wr_id == ISER_FASTREG_LI_WRID) in iser_handle_comp_error()
1155 static void iser_handle_wc(struct ib_wc *wc) in iser_handle_wc() argument
1161 ib_conn = wc->qp->qp_context; in iser_handle_wc()
1162 if (likely(wc->status == IB_WC_SUCCESS)) { in iser_handle_wc()
1163 if (wc->opcode == IB_WC_RECV) { in iser_handle_wc()
1164 rx_desc = (struct iser_rx_desc *)(uintptr_t)wc->wr_id; in iser_handle_wc()
1165 iser_rcv_completion(rx_desc, wc->byte_len, in iser_handle_wc()
[all …]
/linux-4.1.27/arch/sh/include/asm/
Dsfp-machine.h57 #define _FP_CHOOSENAN(fs, wc, R, X, Y, OP) \ argument
63 _FP_FRAC_COPY_##wc(R,Y); \
68 _FP_FRAC_COPY_##wc(R,X); \
/linux-4.1.27/arch/sparc/include/asm/
Dsfp-machine_64.h59 #define _FP_CHOOSENAN(fs, wc, R, X, Y, OP) \ argument
65 _FP_FRAC_COPY_##wc(R,X); \
70 _FP_FRAC_COPY_##wc(R,Y); \
Dsfp-machine_32.h61 #define _FP_CHOOSENAN(fs, wc, R, X, Y, OP) \ argument
67 _FP_FRAC_COPY_##wc(R,X); \
72 _FP_FRAC_COPY_##wc(R,Y); \
/linux-4.1.27/net/sunrpc/xprtrdma/
Dverbs.c205 rpcrdma_sendcq_process_wc(struct ib_wc *wc) in rpcrdma_sendcq_process_wc() argument
208 if (wc->wr_id == RPCRDMA_IGNORE_COMPLETION) { in rpcrdma_sendcq_process_wc()
209 if (wc->status != IB_WC_SUCCESS && in rpcrdma_sendcq_process_wc()
210 wc->status != IB_WC_WR_FLUSH_ERR) in rpcrdma_sendcq_process_wc()
212 __func__, COMPLETION_MSG(wc->status)); in rpcrdma_sendcq_process_wc()
216 r = (struct rpcrdma_mw *)(unsigned long)wc->wr_id; in rpcrdma_sendcq_process_wc()
217 r->mw_sendcompletion(wc); in rpcrdma_sendcq_process_wc()
277 rpcrdma_recvcq_process_wc(struct ib_wc *wc, struct list_head *sched_list) in rpcrdma_recvcq_process_wc() argument
280 (struct rpcrdma_rep *)(unsigned long)wc->wr_id; in rpcrdma_recvcq_process_wc()
283 if (wc->status != IB_WC_SUCCESS) in rpcrdma_recvcq_process_wc()
[all …]
Dsvc_rdma_transport.c290 struct ib_wc wc; in rq_cq_reap() local
299 while ((ret = ib_poll_cq(xprt->sc_rq_cq, 1, &wc)) > 0) { in rq_cq_reap()
300 ctxt = (struct svc_rdma_op_ctxt *)(unsigned long)wc.wr_id; in rq_cq_reap()
301 ctxt->wc_status = wc.status; in rq_cq_reap()
302 ctxt->byte_len = wc.byte_len; in rq_cq_reap()
304 if (wc.status != IB_WC_SUCCESS) { in rq_cq_reap()
388 struct ib_wc *wc; in sq_cq_reap() local
403 wc = &wc_a[i]; in sq_cq_reap()
404 if (wc->status != IB_WC_SUCCESS) { in sq_cq_reap()
406 wc->status); in sq_cq_reap()
[all …]
Dfrwr_ops.c122 frwr_sendcompletion(struct ib_wc *wc) in frwr_sendcompletion() argument
126 if (likely(wc->status == IB_WC_SUCCESS)) in frwr_sendcompletion()
130 r = (struct rpcrdma_mw *)(unsigned long)wc->wr_id; in frwr_sendcompletion()
132 __func__, r, wc->status); in frwr_sendcompletion()
/linux-4.1.27/sound/pci/ctxfi/
Dcttimer.c55 unsigned int wc; /* current wallclock */ member
188 unsigned int wc, diff; in ct_xfitimer_reprogram() local
196 wc = ct_xfitimer_get_wc(atimer); in ct_xfitimer_reprogram()
197 diff = wc - atimer->wc; in ct_xfitimer_reprogram()
198 atimer->wc = wc; in ct_xfitimer_reprogram()
300 atimer->wc = ct_xfitimer_get_wc(atimer); in ct_xfitimer_start()
/linux-4.1.27/fs/btrfs/
Dextent-tree.c7685 struct walk_control *wc, in reada_walk_down() argument
7700 if (path->slots[wc->level] < wc->reada_slot) { in reada_walk_down()
7701 wc->reada_count = wc->reada_count * 2 / 3; in reada_walk_down()
7702 wc->reada_count = max(wc->reada_count, 2); in reada_walk_down()
7704 wc->reada_count = wc->reada_count * 3 / 2; in reada_walk_down()
7705 wc->reada_count = min_t(int, wc->reada_count, in reada_walk_down()
7709 eb = path->nodes[wc->level]; in reada_walk_down()
7713 for (slot = path->slots[wc->level]; slot < nritems; slot++) { in reada_walk_down()
7714 if (nread >= wc->reada_count) in reada_walk_down()
7721 if (slot == path->slots[wc->level]) in reada_walk_down()
[all …]
Dtree-log.c290 struct walk_control *wc, u64 gen);
298 struct walk_control *wc, u64 gen) in process_one_buffer() argument
312 if (wc->pin) in process_one_buffer()
317 if (wc->pin && btrfs_header_level(eb) == 0) in process_one_buffer()
319 if (wc->write) in process_one_buffer()
321 if (wc->wait) in process_one_buffer()
2178 struct walk_control *wc, u64 gen) in replay_one_buffer() argument
2182 struct btrfs_root *root = wc->replay_dest; in replay_one_buffer()
2207 wc->stage == LOG_WALK_REPLAY_INODES) { in replay_one_buffer()
2213 ret = replay_xattr_deletes(wc->trans, root, log, in replay_one_buffer()
[all …]
/linux-4.1.27/fs/logfs/
Dreadwrite.c1341 struct write_control *wc) in logfs_write_i0() argument
1349 if (wc->ofs == 0) in logfs_write_i0()
1353 shadow = alloc_shadow(inode, bix, level, wc->ofs); in logfs_write_i0()
1354 if (wc->flags & WF_WRITE) in logfs_write_i0()
1356 if (wc->flags & WF_DELETE) in logfs_write_i0()
1370 wc->ofs = shadow->new_ofs; in logfs_write_i0()
1371 if (wc->ofs && full) in logfs_write_i0()
1372 wc->ofs |= LOGFS_FULLY_POPULATED; in logfs_write_i0()
1380 struct write_control wc = { in logfs_write_direct() local
1388 err = logfs_write_i0(inode, page, &wc); in logfs_write_direct()
[all …]
/linux-4.1.27/tools/lib/lockdep/
Drun_tests.sh9 if [ $(timeout 1 ./tests/$testname | wc -l) -gt 0 ]; then
21 if [ $(timeout 1 ./lockdep ./tests/$testname | wc -l) -gt 0 ]; then
/linux-4.1.27/arch/s390/include/asm/
Dsfp-machine.h58 #define _FP_CHOOSENAN(fs, wc, R, X, Y, OP) \ argument
64 _FP_FRAC_COPY_##wc(R,Y); \
69 _FP_FRAC_COPY_##wc(R,X); \
/linux-4.1.27/net/9p/
Dtrans_rdma.c349 struct ib_wc wc; in cq_comp_handler() local
352 while ((ret = ib_poll_cq(cq, 1, &wc)) > 0) { in cq_comp_handler()
353 struct p9_rdma_context *c = (void *) (unsigned long) wc.wr_id; in cq_comp_handler()
357 handle_recv(client, rdma, c, wc.status, wc.byte_len); in cq_comp_handler()
362 handle_send(client, rdma, c, wc.status, wc.byte_len); in cq_comp_handler()
368 c->wc_op, wc.opcode, wc.status); in cq_comp_handler()
/linux-4.1.27/include/uapi/linux/netfilter/
Dxt_osf.h53 __u32 wc; member
63 struct xt_osf_wc wc; member
/linux-4.1.27/arch/alpha/include/asm/
Dsfp-machine.h55 #define _FP_CHOOSENAN(fs, wc, R, X, Y, OP) \ argument
58 _FP_FRAC_COPY_##wc(R,X); \
/linux-4.1.27/fs/ocfs2/cluster/
Dheartbeat.c347 static inline void o2hb_bio_wait_init(struct o2hb_bio_wait_ctxt *wc) in o2hb_bio_wait_init() argument
349 atomic_set(&wc->wc_num_reqs, 1); in o2hb_bio_wait_init()
350 init_completion(&wc->wc_io_complete); in o2hb_bio_wait_init()
351 wc->wc_error = 0; in o2hb_bio_wait_init()
355 static inline void o2hb_bio_wait_dec(struct o2hb_bio_wait_ctxt *wc, in o2hb_bio_wait_dec() argument
361 if (atomic_dec_and_test(&wc->wc_num_reqs)) { in o2hb_bio_wait_dec()
363 complete(&wc->wc_io_complete); in o2hb_bio_wait_dec()
369 struct o2hb_bio_wait_ctxt *wc) in o2hb_wait_on_io() argument
371 o2hb_bio_wait_dec(wc, 1); in o2hb_wait_on_io()
372 wait_for_completion(&wc->wc_io_complete); in o2hb_wait_on_io()
[all …]
/linux-4.1.27/scripts/
Ddecodecode95 faultlinenum=$(( $(wc -l $T.oo | cut -d" " -f1) - \
96 $(wc -l $T.aa | cut -d" " -f1) + 3))
/linux-4.1.27/Documentation/x86/
Dpat.txt118 set_memory_[uc|wc] and set_memory_wb should be used in pairs, where driver will
119 first make a region uc or wc and switch it back to wb after use.
124 Drivers should use ioremap_[uc|wc] to access PCI BARs with [uc|wc] access
127 Drivers should use set_memory_[uc|wc] to set access type for RAM ranges.
/linux-4.1.27/arch/powerpc/include/asm/
Dsfp-machine.h144 #define _FP_CHOOSENAN(fs, wc, R, X, Y, OP) \ argument
150 _FP_FRAC_COPY_##wc(R,X); \
155 _FP_FRAC_COPY_##wc(R,Y); \
/linux-4.1.27/arch/powerpc/
Drelocs_check.sh51 num_bad=$(echo "$bad_relocs" | wc -l)
/linux-4.1.27/tools/testing/selftests/rcutorture/bin/
Dcpus2use.sh25 ncpus=`grep '^processor' /proc/cpuinfo | wc -l`
Dparse-torture.sh99 …echo $title no success message, `grep --binary-files=text 'ver:' $file | wc -l` successful version…
/linux-4.1.27/arch/x86/kernel/cpu/
Dmkcapflags.sh16 PFX_SZ=$(echo $PFX | wc -c)
/linux-4.1.27/drivers/infiniband/ulp/srpt/
Dib_srpt.c469 ah = ib_create_ah_from_wc(mad_agent->qp->pd, mad_wc->wc, in srpt_mad_recv_handler()
476 rsp = ib_create_send_mad(mad_agent, mad_wc->wc->src_qp, in srpt_mad_recv_handler()
477 mad_wc->wc->pkey_index, 0, in srpt_mad_recv_handler()
1879 struct ib_wc *wc) in srpt_process_rcv_completion() argument
1885 index = idx_from_wr_id(wc->wr_id); in srpt_process_rcv_completion()
1886 if (wc->status == IB_WC_SUCCESS) { in srpt_process_rcv_completion()
1896 index, wc->status); in srpt_process_rcv_completion()
1917 struct ib_wc *wc) in srpt_process_send_completion() argument
1923 index = idx_from_wr_id(wc->wr_id); in srpt_process_send_completion()
1924 opcode = opcode_from_wr_id(wc->wr_id); in srpt_process_send_completion()
[all …]
Dib_srpt.h321 struct ib_wc wc[16]; member
/linux-4.1.27/include/rdma/
Dib_mad.h418 struct ib_wc *wc; member
604 struct ib_wc *wc);
Dib_verbs.h1574 struct ib_wc *wc);
1802 int ib_init_ah_from_wc(struct ib_device *device, u8 port_num, struct ib_wc *wc,
1817 struct ib_ah *ib_create_ah_from_wc(struct ib_pd *pd, struct ib_wc *wc,
2062 struct ib_wc *wc) in ib_poll_cq() argument
2064 return cq->device->poll_cq(cq, num_entries, wc); in ib_poll_cq()
/linux-4.1.27/drivers/infiniband/hw/usnic/
Dusnic_ib_verbs.h68 struct ib_wc *wc);
Dusnic_ib_verbs.c747 struct ib_wc *wc) in usnic_ib_poll_cq() argument
/linux-4.1.27/drivers/infiniband/ulp/isert/
Dib_isert.c2070 isert_cq_comp_err(struct isert_conn *isert_conn, struct ib_wc *wc) in isert_cq_comp_err() argument
2072 if (wc->wr_id == ISER_BEACON_WRID) { in isert_cq_comp_err()
2076 } else if (is_isert_tx_desc(isert_conn, (void *)(uintptr_t)wc->wr_id)) { in isert_cq_comp_err()
2081 desc = (struct iser_tx_desc *)(uintptr_t)wc->wr_id; in isert_cq_comp_err()
2096 isert_handle_wc(struct ib_wc *wc) in isert_handle_wc() argument
2102 isert_conn = wc->qp->qp_context; in isert_handle_wc()
2103 if (likely(wc->status == IB_WC_SUCCESS)) { in isert_handle_wc()
2104 if (wc->opcode == IB_WC_RECV) { in isert_handle_wc()
2105 rx_desc = (struct iser_rx_desc *)(uintptr_t)wc->wr_id; in isert_handle_wc()
2106 isert_rcv_completion(rx_desc, isert_conn, wc->byte_len); in isert_handle_wc()
[all …]
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
Docrdma_verbs.h36 int ocrdma_poll_cq(struct ib_cq *, int num_entries, struct ib_wc *wc);
Docrdma_verbs.c2908 int ocrdma_poll_cq(struct ib_cq *ibcq, int num_entries, struct ib_wc *wc) in ocrdma_poll_cq() argument
2919 num_os_cqe = ocrdma_poll_hwcq(cq, cqes_to_poll, wc); in ocrdma_poll_cq()
2924 wc = wc + num_os_cqe; in ocrdma_poll_cq()
2934 err_cqes = ocrdma_add_err_cqe(cq, cqes_to_poll, qp, wc); in ocrdma_poll_cq()
2937 wc = wc + err_cqes; in ocrdma_poll_cq()
/linux-4.1.27/net/netfilter/
Dxt_osf.c244 if (f->wss.wc >= OSF_WSS_MAX) in xt_osf_match_packet()
258 check_WSS = f->wss.wc; in xt_osf_match_packet()
/linux-4.1.27/arch/arm/mach-ixp4xx/
Dixp4xx_npe.c219 u32 wc; in npe_debug_instr() local
243 wc = __raw_readl(&npe->regs->watch_count); in npe_debug_instr()
250 if (wc != __raw_readl(&npe->regs->watch_count)) in npe_debug_instr()
/linux-4.1.27/drivers/infiniband/ulp/srp/
Dib_srp.c1847 static void srp_handle_recv(struct srp_rdma_ch *ch, struct ib_wc *wc) in srp_handle_recv() argument
1851 struct srp_iu *iu = (struct srp_iu *) (uintptr_t) wc->wr_id; in srp_handle_recv()
1864 iu->buf, wc->byte_len, true); in srp_handle_recv()
1950 struct ib_wc wc; in srp_recv_completion() local
1953 while (ib_poll_cq(cq, 1, &wc) > 0) { in srp_recv_completion()
1954 if (likely(wc.status == IB_WC_SUCCESS)) { in srp_recv_completion()
1955 srp_handle_recv(ch, &wc); in srp_recv_completion()
1957 srp_handle_qp_err(wc.wr_id, wc.status, false, ch); in srp_recv_completion()
1965 struct ib_wc wc; in srp_send_completion() local
1968 while (ib_poll_cq(cq, 1, &wc) > 0) { in srp_send_completion()
[all …]
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/o2iblnd/
Do2iblnd_cb.c3240 kiblnd_complete(struct ib_wc *wc) in kiblnd_complete() argument
3242 switch (kiblnd_wreqid2type(wc->wr_id)) { in kiblnd_complete()
3254 kiblnd_wreqid2ptr(wc->wr_id), wc->status); in kiblnd_complete()
3258 kiblnd_tx_complete(kiblnd_wreqid2ptr(wc->wr_id), wc->status); in kiblnd_complete()
3262 kiblnd_rx_complete(kiblnd_wreqid2ptr(wc->wr_id), wc->status, in kiblnd_complete()
3263 wc->byte_len); in kiblnd_complete()
3317 struct ib_wc wc; in kiblnd_scheduler() local
3358 rc = ib_poll_cq(conn->ibc_cq, 1, &wc); in kiblnd_scheduler()
3372 rc = ib_poll_cq(conn->ibc_cq, 1, &wc); in kiblnd_scheduler()
3403 kiblnd_complete(&wc); in kiblnd_scheduler()
/linux-4.1.27/arch/arm/boot/dts/
Dstih415-clock.dtsi437 "", /* clk-m-mpestac-wc */
451 clock-output-names = "", /* clk-m-mpevtacrx0-wc */
452 "", /* clk-m-mpevtacrx1-wc */
Dstih416-clock.dtsi439 "", /* clk-m-mpestac-wc */
453 clock-output-names = "", /* clk-m-mpevtacrx0-wc */
454 "", /* clk-m-mpevtacrx1-wc */
/linux-4.1.27/include/xen/interface/
Dxen.h581 struct pvclock_wall_clock wc; member
/linux-4.1.27/arch/arm/mm/
Dproc-arm740.S57 bic ip, ip, #0x0000000c @ ............wc..
/linux-4.1.27/drivers/isdn/hardware/mISDN/
Dhfcmulti.c746 vpm_init(struct hfc_multi *wc) in vpm_init() argument
756 ver = vpm_in(wc, x, 0x1a0); in vpm_init()
761 vpm_out(wc, x, 0x1a8 + y, 0x00); /* GPIO out */ in vpm_init()
762 vpm_out(wc, x, 0x1ac + y, 0x00); /* GPIO dir */ in vpm_init()
763 vpm_out(wc, x, 0x1b0 + y, 0x00); /* GPIO sel */ in vpm_init()
767 reg = vpm_in(wc, x, 0x1a3); /* misc_con */ in vpm_init()
768 vpm_out(wc, x, 0x1a3, reg & ~2); in vpm_init()
771 vpm_out(wc, x, 0x022, 1); in vpm_init()
772 vpm_out(wc, x, 0x023, 0xff); in vpm_init()
775 vpm_out(wc, x, 0x02f, 0x00); in vpm_init()
[all …]
/linux-4.1.27/Documentation/DocBook/media/
Dbayer.png.b6461 kauYEP0bslw4c9MD9xPIA9d/5wc/JJH+uWUlaunL6Di3P1GzPxhMaMfV920N0q8qcvVO27/34/80
/linux-4.1.27/arch/x86/xen/
Dtime.c175 struct pvclock_wall_clock *wall_clock = &(s->wc); in xen_read_wallclock()
/linux-4.1.27/kernel/debug/kdb/
Dkdb_main.c1518 } wc; in kdb_md_line() local
1521 cp = wc.c + 8 - bytesperword; in kdb_md_line()
1523 cp = wc.c; in kdb_md_line()
1525 wc.word = word; in kdb_md_line()
/linux-4.1.27/fs/fat/
Ddir.c210 wchar_t wc; in fat_short2lower_uni() local
212 charlen = t->char2uni(c, clen, &wc); in fat_short2lower_uni()
228 *uni = wc; in fat_short2lower_uni()
/linux-4.1.27/arch/arm/mach-sa1100/
DKconfig96 <http://h10025.www1.hp.com/ewfrf/wc/product?product=61677&cc=us&lc=en&dlc=en&product=61677#>
/linux-4.1.27/Documentation/
Ddynamic-debug-howto.txt72 nullarbor:~ # grep -i rdma <debugfs>/dynamic_debug/control | wc -l
75 nullarbor:~ # grep -i tcp <debugfs>/dynamic_debug/control | wc -l
/linux-4.1.27/drivers/gpu/drm/msm/dsi/
Ddsi_host.c807 u32 wc; in dsi_timing_setup() local
831 wc = mode->hdisplay * dsi_get_bpp(msm_host->format) / 8 + 1; in dsi_timing_setup()
834 DSI_CMD_MDP_STREAM_CTRL_WORD_COUNT(wc) | in dsi_timing_setup()
/linux-4.1.27/drivers/scsi/qla2xxx/
Dqla_gs.c718 uint16_t wc; in qla2x00_prep_sns_cmd() local
724 wc = data_size / 2; /* Size in 16bit words. */ in qla2x00_prep_sns_cmd()
725 sns_cmd->p.cmd.buffer_length = cpu_to_le16(wc); in qla2x00_prep_sns_cmd()
730 wc = (data_size - 16) / 4; /* Size in 32bit words. */ in qla2x00_prep_sns_cmd()
731 sns_cmd->p.cmd.size = cpu_to_le16(wc); in qla2x00_prep_sns_cmd()
/linux-4.1.27/include/uapi/rdma/
Dib_user_verbs.h404 struct ib_uverbs_wc wc[0]; member
/linux-4.1.27/arch/x86/kvm/
Dx86.c1113 struct pvclock_wall_clock wc; in kvm_write_wall_clock() local
1142 wc.sec = boot.tv_sec; in kvm_write_wall_clock()
1143 wc.nsec = boot.tv_nsec; in kvm_write_wall_clock()
1144 wc.version = version; in kvm_write_wall_clock()
1146 kvm_write_guest(kvm, wall_clock, &wc, sizeof(wc)); in kvm_write_wall_clock()
/linux-4.1.27/
DMakefile1004 if [ `echo -n "$(KERNELRELEASE)" | wc -c ` -gt $(uts_len) ]; then \
/linux-4.1.27/drivers/video/fbdev/omap2/dss/
Ddsi.c4340 int bl, wc, pps, tot; in print_dsi_vm() local
4342 wc = DIV_ROUND_UP(t->hact * t->bitspp, 8); in print_dsi_vm()
4343 pps = DIV_ROUND_UP(wc + 6, t->ndl); /* pixel packet size */ in print_dsi_vm()