/linux-4.4.14/net/rds/ |
D | iw_cm.c | 87 rds_iwdev = ib_get_client_data(ic->i_cm_id->device, &rds_iw_client); in rds_iw_cm_connect_complete() 88 err = rds_iw_update_cm_id(rds_iwdev, ic->i_cm_id); in rds_iw_cm_connect_complete() 156 rdma_notify(ic->i_cm_id, IB_EVENT_COMM_EST); in rds_iw_qp_event_handler() 253 struct ib_device *dev = ic->i_cm_id->device; in rds_iw_setup_qp() 288 ret = rdma_create_qp(ic->i_cm_id, ic->i_pd, &attr); in rds_iw_setup_qp() 443 BUG_ON(ic->i_cm_id); in rds_iw_cm_handle_connect() 445 ic->i_cm_id = cm_id; in rds_iw_cm_handle_connect() 512 if (ic->i_cm_id == cm_id) in rds_iw_cm_initiate_connect() 527 ic->i_cm_id = rdma_create_id(&init_net, rds_rdma_cm_event_handler, conn, in rds_iw_conn_connect() 529 if (IS_ERR(ic->i_cm_id)) { in rds_iw_conn_connect() [all …]
|
D | ib_cm.c | 86 ret = ib_modify_qp(ic->i_cm_id->qp, attr, IB_QP_MIN_RNR_TIMER); in rds_ib_tune_rnr() 144 err = ib_modify_qp(ic->i_cm_id->qp, &qp_attr, IB_QP_STATE); in rds_ib_cm_connect_complete() 319 rdma_notify(ic->i_cm_id, IB_EVENT_COMM_EST); in rds_ib_qp_event_handler() 350 struct ib_device *dev = ic->i_cm_id->device; in rds_ib_setup_qp() 428 ret = rdma_create_qp(ic->i_cm_id, ic->i_pd, &attr); in rds_ib_setup_qp() 594 BUG_ON(ic->i_cm_id); in rds_ib_cm_handle_connect() 596 ic->i_cm_id = cm_id; in rds_ib_cm_handle_connect() 657 if (ic->i_cm_id == cm_id) in rds_ib_cm_initiate_connect() 671 ic->i_cm_id = rdma_create_id(&init_net, rds_rdma_cm_event_handler, conn, in rds_ib_conn_connect() 673 if (IS_ERR(ic->i_cm_id)) { in rds_ib_conn_connect() [all …]
|
D | iw_rdma.c | 95 struct rds_iw_cm_id *i_cm_id; local 102 list_for_each_entry(i_cm_id, &iwdev->cm_id_list, list) { 105 src_addr = (struct sockaddr_in *)&i_cm_id->cm_id->route.addr.src_addr; 106 dst_addr = (struct sockaddr_in *)&i_cm_id->cm_id->route.addr.dst_addr; 136 *cm_id = i_cm_id->cm_id; 148 struct rds_iw_cm_id *i_cm_id; local 150 i_cm_id = kmalloc(sizeof *i_cm_id, GFP_KERNEL); 151 if (!i_cm_id) 154 i_cm_id->cm_id = cm_id; 157 list_add_tail(&i_cm_id->list, &rds_iwdev->cm_id_list); [all …]
|
D | iw.c | 131 struct rds_iw_cm_id *i_cm_id, *next; in rds_iw_remove_one() local 137 list_for_each_entry_safe(i_cm_id, next, &rds_iwdev->cm_id_list, list) { in rds_iw_remove_one() 138 list_del(&i_cm_id->list); in rds_iw_remove_one() 139 kfree(i_cm_id); in rds_iw_remove_one() 183 dev_addr = &ic->i_cm_id->route.addr.dev_addr; in rds_iw_conn_info_visitor() 188 rds_iwdev = ib_get_client_data(ic->i_cm_id->device, &rds_iw_client); in rds_iw_conn_info_visitor()
|
D | ib_send.c | 76 ib_dma_unmap_sg(ic->i_cm_id->device, in rds_ib_send_unmap_data() 86 ib_dma_unmap_sg(ic->i_cm_id->device, in rds_ib_send_unmap_rdma() 127 ib_dma_unmap_sg(ic->i_cm_id->device, op->op_sg, 1, in rds_ib_send_unmap_atomic() 378 struct rds_connection *conn = ic->i_cm_id->context; in rds_ib_send_grab_credits() 487 struct ib_device *dev = ic->i_cm_id->device; in rds_ib_xmit() 725 ret = ib_post_send(ic->i_cm_id->qp, &first->s_wr, &failed_wr); in rds_ib_xmit() 765 rds_ibdev = ib_get_client_data(ic->i_cm_id->device, &rds_ib_client); in rds_ib_xmit_atomic() 801 ret = ib_dma_map_sg(ic->i_cm_id->device, op->op_sg, 1, DMA_FROM_DEVICE); in rds_ib_xmit_atomic() 811 send->s_sge[0].addr = ib_sg_dma_address(ic->i_cm_id->device, op->op_sg); in rds_ib_xmit_atomic() 812 send->s_sge[0].length = ib_sg_dma_len(ic->i_cm_id->device, op->op_sg); in rds_ib_xmit_atomic() [all …]
|
D | iw_send.c | 70 ib_dma_unmap_sg(ic->i_cm_id->device, in rds_iw_send_unmap_rdma() 85 ib_dma_unmap_sg(ic->i_cm_id->device, in rds_iw_send_unmap_rm() 364 struct rds_connection *conn = ic->i_cm_id->context; in rds_iw_send_grab_credits() 492 struct ib_device *dev = ic->i_cm_id->device; in rds_iw_xmit() 743 ret = ib_post_send(ic->i_cm_id->qp, &first->s_send_wr, &failed_wr); in rds_iw_xmit() 806 rds_iwdev = ib_get_client_data(ic->i_cm_id->device, &rds_iw_client); in rds_iw_xmit_rdma() 810 op->op_count = ib_dma_map_sg(ic->i_cm_id->device, in rds_iw_xmit_rdma() 899 len = ib_sg_dma_len(ic->i_cm_id->device, scat); in rds_iw_xmit_rdma() 904 send->s_sge[j].addr = ib_sg_dma_address(ic->i_cm_id->device, scat); in rds_iw_xmit_rdma() 959 ret = ib_post_send(ic->i_cm_id->qp, &first->s_rdma_wr.wr, &failed_wr); in rds_iw_xmit_rdma()
|
D | ib_recv.c | 227 ib_dma_unmap_sg(ic->i_cm_id->device, &recv->r_frag->f_sg, 1, DMA_FROM_DEVICE); in rds_ib_recv_clear_one() 333 ret = ib_dma_map_sg(ic->i_cm_id->device, &recv->r_frag->f_sg, in rds_ib_recv_refill_one() 342 sge->addr = ib_sg_dma_address(ic->i_cm_id->device, &recv->r_frag->f_sg); in rds_ib_recv_refill_one() 343 sge->length = ib_sg_dma_len(ic->i_cm_id->device, &recv->r_frag->f_sg); in rds_ib_recv_refill_one() 407 ret = ib_post_recv(ic->i_cm_id->qp, &recv->r_wr, &failed_wr); in rds_ib_recv_refill() 411 ic->i_cm_id->device, in rds_ib_recv_refill() 659 ret = ib_post_send(ic->i_cm_id->qp, &ic->i_ack_wr, &failed_wr); in rds_ib_send_ack() 970 ib_dma_unmap_sg(ic->i_cm_id->device, &recv->r_frag->f_sg, 1, in rds_ib_recv_cqe_handler()
|
D | iw_recv.c | 76 ib_dma_unmap_page(ic->i_cm_id->device, in rds_iw_recv_unmap_page() 176 dma_addr = ib_dma_map_page(ic->i_cm_id->device, in rds_iw_recv_refill_one() 181 if (ib_dma_mapping_error(ic->i_cm_id->device, dma_addr)) in rds_iw_recv_refill_one() 251 ret = ib_post_recv(ic->i_cm_id->qp, &recv->r_wr, &failed_wr); in rds_iw_recv_refill() 448 ret = ib_post_send(ic->i_cm_id->qp, &ic->i_ack_wr, &failed_wr); in rds_iw_send_ack()
|
D | iw.h | 112 struct rdma_cm_id *i_cm_id; member 270 return ic->i_dma_local_lkey ? ic->i_cm_id->device->local_dma_lkey : ic->i_mr->lkey; in rds_iw_local_dma_lkey()
|
D | rdma_transport.c | 93 if (ibic && ibic->i_cm_id == cm_id) in rds_rdma_cm_event_handler()
|
D | ib.c | 296 dev_addr = &ic->i_cm_id->route.addr.dev_addr; in rds_ib_conn_info_visitor()
|
D | ib.h | 125 struct rdma_cm_id *i_cm_id; member
|