op_sg 78 net/rds/ib_send.c op->op_sg, op->op_nents, op_sg 88 net/rds/ib_send.c op->op_sg, op->op_nents, op_sg 128 net/rds/ib_send.c ib_dma_unmap_sg(ic->i_cm_id->device, op->op_sg, 1, op_sg 512 net/rds/ib_send.c scat = &rm->data.op_sg[sg]; op_sg 551 net/rds/ib_send.c rm->data.op_sg, op_sg 621 net/rds/ib_send.c scat = &ic->i_data_op->op_sg[rm->data.op_dmasg]; op_sg 642 net/rds/ib_send.c && scat != &rm->data.op_sg[rm->data.op_count]) { op_sg 695 net/rds/ib_send.c && scat != &rm->data.op_sg[rm->data.op_count]); op_sg 703 net/rds/ib_send.c if (scat == &rm->data.op_sg[rm->data.op_count]) { op_sg 797 net/rds/ib_send.c ret = ib_dma_map_sg(ic->i_cm_id->device, op->op_sg, 1, DMA_FROM_DEVICE); op_sg 807 net/rds/ib_send.c send->s_sge[0].addr = sg_dma_address(op->op_sg); op_sg 808 net/rds/ib_send.c send->s_sge[0].length = sg_dma_len(op->op_sg); op_sg 862 net/rds/ib_send.c op->op_sg, op->op_nents, (op->op_write) ? op_sg 891 net/rds/ib_send.c scat = &op->op_sg[0]; op_sg 895 net/rds/ib_send.c for (i = 0; i < work_alloc && scat != &op->op_sg[op->op_count]; i++) { op_sg 921 net/rds/ib_send.c scat != &op->op_sg[op->op_count]; j++) { op_sg 945 net/rds/ib_send.c if (scat == &op->op_sg[op->op_count]) { op_sg 79 net/rds/loop.c struct scatterlist *sgp = &rm->data.op_sg[sg]; op_sg 156 net/rds/message.c __free_page(sg_page(&rm->data.op_sg[i])); op_sg 158 net/rds/message.c put_page(sg_page(&rm->data.op_sg[i])); op_sg 348 net/rds/message.c rm->data.op_sg = rds_message_alloc_sgs(rm, num_sgs); op_sg 349 net/rds/message.c if (IS_ERR(rm->data.op_sg)) { op_sg 351 net/rds/message.c return ERR_CAST(rm->data.op_sg); op_sg 355 net/rds/message.c sg_set_page(&rm->data.op_sg[i], op_sg 376 net/rds/message.c sg = rm->data.op_sg; op_sg 400 net/rds/message.c put_page(sg_page(&rm->data.op_sg[i])); op_sg 432 net/rds/message.c sg = rm->data.op_sg; op_sg 479 net/rds/message.c sg = rm->data.op_sg; op_sg 460 net/rds/rdma.c struct page *page = sg_page(&ro->op_sg[i]); op_sg 479 net/rds/rdma.c struct page *page = sg_page(ao->op_sg); op_sg 627 net/rds/rdma.c op->op_sg = rds_message_alloc_sgs(rm, nr_pages); op_sg 628 net/rds/rdma.c if (IS_ERR(op->op_sg)) { op_sg 629 net/rds/rdma.c ret = PTR_ERR(op->op_sg); op_sg 691 net/rds/rdma.c sg = &op->op_sg[op->op_nents + j]; op_sg 833 net/rds/rdma.c rm->atomic.op_sg = rds_message_alloc_sgs(rm, 1); op_sg 834 net/rds/rdma.c if (IS_ERR(rm->atomic.op_sg)) { op_sg 835 net/rds/rdma.c ret = PTR_ERR(rm->atomic.op_sg); op_sg 850 net/rds/rdma.c sg_set_page(rm->atomic.op_sg, page, 8, offset_in_page(args->local_addr)); op_sg 459 net/rds/rds.h struct scatterlist *op_sg; op_sg 477 net/rds/rds.h struct scatterlist *op_sg; op_sg 489 net/rds/rds.h struct scatterlist *op_sg; op_sg 382 net/rds/send.c sg = &rm->data.op_sg[cp->cp_xmit_sg]; op_sg 1277 net/rds/send.c rm->data.op_sg = rds_message_alloc_sgs(rm, num_sgs); op_sg 1278 net/rds/send.c if (IS_ERR(rm->data.op_sg)) { op_sg 1279 net/rds/send.c ret = PTR_ERR(rm->data.op_sg); op_sg 124 net/rds/tcp_send.c sg_page(&rm->data.op_sg[sg]), op_sg 125 net/rds/tcp_send.c rm->data.op_sg[sg].offset + off, op_sg 126 net/rds/tcp_send.c rm->data.op_sg[sg].length - off, op_sg 128 net/rds/tcp_send.c rdsdebug("tcp sendpage %p:%u:%u ret %d\n", (void *)sg_page(&rm->data.op_sg[sg]), op_sg 129 net/rds/tcp_send.c rm->data.op_sg[sg].offset + off, rm->data.op_sg[sg].length - off, op_sg 136 net/rds/tcp_send.c if (off == rm->data.op_sg[sg].length) {