Lines Matching refs:rcd
101 struct qib_ctxtdata *rcd = ctxt_fp(fp); in qib_get_base_info() local
104 struct qib_devdata *dd = rcd->dd; in qib_get_base_info()
105 struct qib_pportdata *ppd = rcd->ppd; in qib_get_base_info()
110 subctxt_cnt = rcd->subctxt_cnt; in qib_get_base_info()
135 ret = dd->f_get_base_info(rcd, kinfo); in qib_get_base_info()
141 kinfo->spi_tidegrcnt = rcd->rcvegrcnt; in qib_get_base_info()
147 rcd->rcvegrbuf_chunks * rcd->rcvegrbuf_size; in qib_get_base_info()
148 kinfo->spi_rcv_egrperchunk = rcd->rcvegrbufs_perchunk; in qib_get_base_info()
150 rcd->rcvegrbuf_chunks; in qib_get_base_info()
184 kinfo->spi_rcvhdr_base = (u64) rcd->rcvhdrq_phys; in qib_get_base_info()
185 kinfo->spi_rcvhdr_tailaddr = (u64) rcd->rcvhdrqtailaddr_phys; in qib_get_base_info()
187 kinfo->spi_rcv_egrbufs = (u64) rcd->rcvegr_phys; in qib_get_base_info()
193 kinfo->spi_uregbase = (u64) dd->uregbase + dd->ureg_align * rcd->ctxt; in qib_get_base_info()
195 kinfo->spi_piocnt = rcd->piocnt; in qib_get_base_info()
196 kinfo->spi_piobufbase = (u64) rcd->piobufs; in qib_get_base_info()
197 kinfo->spi_sendbuf_status = cvt_kvaddr(rcd->user_event_mask); in qib_get_base_info()
199 kinfo->spi_piocnt = (rcd->piocnt / subctxt_cnt) + in qib_get_base_info()
200 (rcd->piocnt % subctxt_cnt); in qib_get_base_info()
202 kinfo->spi_piobufbase = (u64) rcd->piobufs + in qib_get_base_info()
204 (rcd->piocnt - kinfo->spi_piocnt); in qib_get_base_info()
208 kinfo->spi_piocnt = rcd->piocnt / subctxt_cnt; in qib_get_base_info()
209 kinfo->spi_piobufbase = (u64) rcd->piobufs + in qib_get_base_info()
215 cvt_kvaddr(&rcd->user_event_mask[subctxt_fp(fp)]); in qib_get_base_info()
217 kinfo->spi_subctxt_uregbase = cvt_kvaddr(rcd->subctxt_uregbase); in qib_get_base_info()
220 cvt_kvaddr(rcd->subctxt_rcvegrbuf); in qib_get_base_info()
222 cvt_kvaddr(rcd->subctxt_rcvhdr_base); in qib_get_base_info()
242 kinfo->spi_ctxt = rcd->ctxt; in qib_get_base_info()
284 static int qib_tid_update(struct qib_ctxtdata *rcd, struct file *fp, in qib_tid_update() argument
290 struct qib_devdata *dd = rcd->dd; in qib_tid_update()
308 ctxttid = rcd->ctxt * dd->rcvtidcnt; in qib_tid_update()
309 if (!rcd->subctxt_cnt) { in qib_tid_update()
311 tid = rcd->tidcursor; in qib_tid_update()
314 tidcnt = (dd->rcvtidcnt / rcd->subctxt_cnt) + in qib_tid_update()
315 (dd->rcvtidcnt % rcd->subctxt_cnt); in qib_tid_update()
320 tidcnt = dd->rcvtidcnt / rcd->subctxt_cnt; in qib_tid_update()
332 pagep = (struct page **) rcd->tid_pg_list; in qib_tid_update()
453 if (!rcd->subctxt_cnt) in qib_tid_update()
454 rcd->tidcursor = tid; in qib_tid_update()
479 static int qib_tid_free(struct qib_ctxtdata *rcd, unsigned subctxt, in qib_tid_free() argument
484 struct qib_devdata *dd = rcd->dd; in qib_tid_free()
499 ctxttid = rcd->ctxt * dd->rcvtidcnt; in qib_tid_free()
500 if (!rcd->subctxt_cnt) in qib_tid_free()
503 tidcnt = (dd->rcvtidcnt / rcd->subctxt_cnt) + in qib_tid_free()
504 (dd->rcvtidcnt % rcd->subctxt_cnt); in qib_tid_free()
507 tidcnt = dd->rcvtidcnt / rcd->subctxt_cnt; in qib_tid_free()
568 static int qib_set_part_key(struct qib_ctxtdata *rcd, u16 key) in qib_set_part_key() argument
570 struct qib_pportdata *ppd = rcd->ppd; in qib_set_part_key()
594 for (i = 0; i < ARRAY_SIZE(rcd->pkeys); i++) { in qib_set_part_key()
595 if (!rcd->pkeys[i] && pidx == -1) in qib_set_part_key()
597 if (rcd->pkeys[i] == key) { in qib_set_part_key()
615 rcd->pkeys[pidx] = key; in qib_set_part_key()
643 rcd->pkeys[pidx] = key; in qib_set_part_key()
666 static int qib_manage_rcvq(struct qib_ctxtdata *rcd, unsigned subctxt, in qib_manage_rcvq() argument
669 struct qib_devdata *dd = rcd->dd; in qib_manage_rcvq()
684 if (rcd->rcvhdrtail_kvaddr) in qib_manage_rcvq()
685 qib_clear_rcvhdrtail(rcd); in qib_manage_rcvq()
689 dd->f_rcvctrl(rcd->ppd, rcvctrl_op, rcd->ctxt); in qib_manage_rcvq()
695 static void qib_clean_part_key(struct qib_ctxtdata *rcd, in qib_clean_part_key() argument
700 struct qib_pportdata *ppd = rcd->ppd; in qib_clean_part_key()
708 for (i = 0; i < ARRAY_SIZE(rcd->pkeys); i++) { in qib_clean_part_key()
709 if (!rcd->pkeys[i]) in qib_clean_part_key()
714 (rcd->pkeys[i] & 0x7fff)) in qib_clean_part_key()
722 rcd->pkeys[i] = 0; in qib_clean_part_key()
729 static int qib_mmap_mem(struct vm_area_struct *vma, struct qib_ctxtdata *rcd, in qib_mmap_mem() argument
732 struct qib_devdata *dd = rcd->dd; in qib_mmap_mem()
766 what, rcd->ctxt, pfn, len, ret); in qib_mmap_mem()
804 struct qib_ctxtdata *rcd, in mmap_piobufs() argument
852 struct qib_ctxtdata *rcd) in mmap_rcvegrbufs() argument
854 struct qib_devdata *dd = rcd->dd; in mmap_rcvegrbufs()
860 size = rcd->rcvegrbuf_size; in mmap_rcvegrbufs()
861 total_size = rcd->rcvegrbuf_chunks * size; in mmap_rcvegrbufs()
883 for (i = 0; i < rcd->rcvegrbuf_chunks; i++, start += size) { in mmap_rcvegrbufs()
884 pfn = virt_to_phys(rcd->rcvegrbuf[i]) >> PAGE_SHIFT; in mmap_rcvegrbufs()
918 struct qib_ctxtdata *rcd, unsigned subctxt) in mmap_kvaddr() argument
920 struct qib_devdata *dd = rcd->dd; in mmap_kvaddr()
927 subctxt_cnt = rcd->subctxt_cnt; in mmap_kvaddr()
928 size = rcd->rcvegrbuf_chunks * rcd->rcvegrbuf_size; in mmap_kvaddr()
935 if (pgaddr == cvt_kvaddr(rcd->subctxt_uregbase)) { in mmap_kvaddr()
936 addr = rcd->subctxt_uregbase; in mmap_kvaddr()
938 } else if (pgaddr == cvt_kvaddr(rcd->subctxt_rcvhdr_base)) { in mmap_kvaddr()
939 addr = rcd->subctxt_rcvhdr_base; in mmap_kvaddr()
940 size = rcd->rcvhdrq_size * subctxt_cnt; in mmap_kvaddr()
941 } else if (pgaddr == cvt_kvaddr(rcd->subctxt_rcvegrbuf)) { in mmap_kvaddr()
942 addr = rcd->subctxt_rcvegrbuf; in mmap_kvaddr()
944 } else if (pgaddr == cvt_kvaddr(rcd->subctxt_uregbase + in mmap_kvaddr()
946 addr = rcd->subctxt_uregbase + PAGE_SIZE * subctxt; in mmap_kvaddr()
948 } else if (pgaddr == cvt_kvaddr(rcd->subctxt_rcvhdr_base + in mmap_kvaddr()
949 rcd->rcvhdrq_size * subctxt)) { in mmap_kvaddr()
950 addr = rcd->subctxt_rcvhdr_base + in mmap_kvaddr()
951 rcd->rcvhdrq_size * subctxt; in mmap_kvaddr()
952 size = rcd->rcvhdrq_size; in mmap_kvaddr()
953 } else if (pgaddr == cvt_kvaddr(&rcd->user_event_mask[subctxt])) { in mmap_kvaddr()
954 addr = rcd->user_event_mask; in mmap_kvaddr()
956 } else if (pgaddr == cvt_kvaddr(rcd->subctxt_rcvegrbuf + in mmap_kvaddr()
958 addr = rcd->subctxt_rcvegrbuf + size * subctxt; in mmap_kvaddr()
1001 struct qib_ctxtdata *rcd; in qib_mmapf() local
1007 rcd = ctxt_fp(fp); in qib_mmapf()
1008 if (!rcd || !(vma->vm_flags & VM_SHARED)) { in qib_mmapf()
1012 dd = rcd->dd; in qib_mmapf()
1039 ret = mmap_kvaddr(vma, pgaddr, rcd, subctxt_fp(fp)); in qib_mmapf()
1046 ureg = dd->uregbase + dd->ureg_align * rcd->ctxt; in qib_mmapf()
1047 if (!rcd->subctxt_cnt) { in qib_mmapf()
1049 piocnt = rcd->piocnt; in qib_mmapf()
1050 piobufs = rcd->piobufs; in qib_mmapf()
1053 piocnt = (rcd->piocnt / rcd->subctxt_cnt) + in qib_mmapf()
1054 (rcd->piocnt % rcd->subctxt_cnt); in qib_mmapf()
1055 piobufs = rcd->piobufs + in qib_mmapf()
1056 dd->palign * (rcd->piocnt - piocnt); in qib_mmapf()
1061 piocnt = rcd->piocnt / rcd->subctxt_cnt; in qib_mmapf()
1062 piobufs = rcd->piobufs + dd->palign * piocnt * slave; in qib_mmapf()
1068 ret = mmap_piobufs(vma, dd, rcd, piobufs, piocnt); in qib_mmapf()
1071 ret = qib_mmap_mem(vma, rcd, PAGE_SIZE, in qib_mmapf()
1074 else if (pgaddr == rcd->rcvegr_phys) in qib_mmapf()
1075 ret = mmap_rcvegrbufs(vma, rcd); in qib_mmapf()
1076 else if (pgaddr == (u64) rcd->rcvhdrq_phys) in qib_mmapf()
1082 ret = qib_mmap_mem(vma, rcd, rcd->rcvhdrq_size, in qib_mmapf()
1083 rcd->rcvhdrq, 1, "rcvhdrq"); in qib_mmapf()
1084 else if (pgaddr == (u64) rcd->rcvhdrqtailaddr_phys) in qib_mmapf()
1086 ret = qib_mmap_mem(vma, rcd, PAGE_SIZE, in qib_mmapf()
1087 rcd->rcvhdrtail_kvaddr, 0, in qib_mmapf()
1105 static unsigned int qib_poll_urgent(struct qib_ctxtdata *rcd, in qib_poll_urgent() argument
1109 struct qib_devdata *dd = rcd->dd; in qib_poll_urgent()
1112 poll_wait(fp, &rcd->wait, pt); in qib_poll_urgent()
1115 if (rcd->urgent != rcd->urgent_poll) { in qib_poll_urgent()
1117 rcd->urgent_poll = rcd->urgent; in qib_poll_urgent()
1120 set_bit(QIB_CTXT_WAITING_URG, &rcd->flag); in qib_poll_urgent()
1127 static unsigned int qib_poll_next(struct qib_ctxtdata *rcd, in qib_poll_next() argument
1131 struct qib_devdata *dd = rcd->dd; in qib_poll_next()
1134 poll_wait(fp, &rcd->wait, pt); in qib_poll_next()
1137 if (dd->f_hdrqempty(rcd)) { in qib_poll_next()
1138 set_bit(QIB_CTXT_WAITING_RCV, &rcd->flag); in qib_poll_next()
1139 dd->f_rcvctrl(rcd->ppd, QIB_RCVCTRL_INTRAVAIL_ENB, rcd->ctxt); in qib_poll_next()
1150 struct qib_ctxtdata *rcd; in qib_poll() local
1153 rcd = ctxt_fp(fp); in qib_poll()
1154 if (!rcd) in qib_poll()
1156 else if (rcd->poll_type == QIB_POLL_TYPE_URGENT) in qib_poll()
1157 pollflag = qib_poll_urgent(rcd, fp, pt); in qib_poll()
1158 else if (rcd->poll_type == QIB_POLL_TYPE_ANYRCV) in qib_poll()
1159 pollflag = qib_poll_next(rcd, fp, pt); in qib_poll()
1240 struct qib_ctxtdata *rcd, in init_subctxts() argument
1270 rcd->subctxt_uregbase = vmalloc_user(PAGE_SIZE * num_subctxts); in init_subctxts()
1271 if (!rcd->subctxt_uregbase) { in init_subctxts()
1278 rcd->subctxt_rcvhdr_base = vmalloc_user(size); in init_subctxts()
1279 if (!rcd->subctxt_rcvhdr_base) { in init_subctxts()
1284 rcd->subctxt_rcvegrbuf = vmalloc_user(rcd->rcvegrbuf_chunks * in init_subctxts()
1285 rcd->rcvegrbuf_size * in init_subctxts()
1287 if (!rcd->subctxt_rcvegrbuf) { in init_subctxts()
1292 rcd->subctxt_cnt = uinfo->spu_subctxt_cnt; in init_subctxts()
1293 rcd->subctxt_id = uinfo->spu_subctxt_id; in init_subctxts()
1294 rcd->active_slaves = 1; in init_subctxts()
1295 rcd->redirect_seq_cnt = 1; in init_subctxts()
1296 set_bit(QIB_CTXT_MASTER_UNINIT, &rcd->flag); in init_subctxts()
1300 vfree(rcd->subctxt_rcvhdr_base); in init_subctxts()
1302 vfree(rcd->subctxt_uregbase); in init_subctxts()
1303 rcd->subctxt_uregbase = NULL; in init_subctxts()
1313 struct qib_ctxtdata *rcd; in setup_ctxt() local
1324 rcd = qib_create_ctxtdata(ppd, ctxt, numa_id); in setup_ctxt()
1330 if (rcd) in setup_ctxt()
1335 if (!rcd || !ptmp) { in setup_ctxt()
1341 rcd->userversion = uinfo->spu_userversion; in setup_ctxt()
1342 ret = init_subctxts(dd, rcd, uinfo); in setup_ctxt()
1345 rcd->tid_pg_list = ptmp; in setup_ctxt()
1346 rcd->pid = current->pid; in setup_ctxt()
1347 init_waitqueue_head(&dd->rcd[ctxt]->wait); in setup_ctxt()
1348 strlcpy(rcd->comm, current->comm, sizeof(rcd->comm)); in setup_ctxt()
1349 ctxt_fp(fp) = rcd; in setup_ctxt()
1359 dd->rcd[ctxt] = NULL; in setup_ctxt()
1360 kfree(rcd); in setup_ctxt()
1391 for (ctxt = dd->first_user_ctxt; ctxt < dd->cfgctxts && dd->rcd[ctxt]; in choose_port_ctxt()
1467 if (dd->rcd[ctxt]) in get_a_ctxt()
1514 struct qib_ctxtdata *rcd = dd->rcd[i]; in find_shared_ctxt() local
1517 if (!rcd || !rcd->cnt) in find_shared_ctxt()
1520 if (rcd->subctxt_id != uinfo->spu_subctxt_id) in find_shared_ctxt()
1523 if (rcd->subctxt_cnt != uinfo->spu_subctxt_cnt || in find_shared_ctxt()
1524 rcd->userversion != uinfo->spu_userversion || in find_shared_ctxt()
1525 rcd->cnt >= rcd->subctxt_cnt) { in find_shared_ctxt()
1529 ctxt_fp(fp) = rcd; in find_shared_ctxt()
1530 subctxt_fp(fp) = rcd->cnt++; in find_shared_ctxt()
1531 rcd->subpid[subctxt_fp(fp)] = current->pid; in find_shared_ctxt()
1533 rcd->active_slaves |= 1 << subctxt_fp(fp); in find_shared_ctxt()
1589 struct qib_ctxtdata *rcd = fd->rcd; in do_qib_user_sdma_queue_create() local
1590 struct qib_devdata *dd = rcd->dd; in do_qib_user_sdma_queue_create()
1596 rcd->ctxt, in do_qib_user_sdma_queue_create()
1678 struct qib_ctxtdata *rcd = ctxt_fp(fp); in qib_do_user_init() local
1684 ret = wait_event_interruptible(rcd->wait, in qib_do_user_init()
1685 !test_bit(QIB_CTXT_MASTER_UNINIT, &rcd->flag)); in qib_do_user_init()
1689 dd = rcd->dd; in qib_do_user_init()
1692 uctxt = rcd->ctxt - dd->first_user_ctxt; in qib_do_user_init()
1694 rcd->piocnt = dd->pbufsctxt + 1; in qib_do_user_init()
1695 rcd->pio_base = rcd->piocnt * uctxt; in qib_do_user_init()
1697 rcd->piocnt = dd->pbufsctxt; in qib_do_user_init()
1698 rcd->pio_base = rcd->piocnt * uctxt + in qib_do_user_init()
1708 if ((rcd->pio_base + rcd->piocnt) > dd->piobcnt2k) { in qib_do_user_init()
1709 if (rcd->pio_base >= dd->piobcnt2k) { in qib_do_user_init()
1712 dd->unit, rcd->ctxt); in qib_do_user_init()
1716 rcd->piocnt = dd->piobcnt2k - rcd->pio_base; in qib_do_user_init()
1718 rcd->ctxt, rcd->piocnt); in qib_do_user_init()
1721 rcd->piobufs = dd->pio2k_bufbase + rcd->pio_base * dd->palign; in qib_do_user_init()
1722 qib_chg_pioavailkernel(dd, rcd->pio_base, rcd->piocnt, in qib_do_user_init()
1723 TXCHK_CHG_TYPE_USER, rcd); in qib_do_user_init()
1742 ret = qib_create_rcvhdrq(dd, rcd); in qib_do_user_init()
1744 ret = qib_setup_eagerbufs(rcd); in qib_do_user_init()
1748 rcd->tidcursor = 0; /* start at beginning after open */ in qib_do_user_init()
1751 rcd->urgent = 0; in qib_do_user_init()
1752 rcd->urgent_poll = 0; in qib_do_user_init()
1765 if (rcd->rcvhdrtail_kvaddr) in qib_do_user_init()
1766 qib_clear_rcvhdrtail(rcd); in qib_do_user_init()
1768 dd->f_rcvctrl(rcd->ppd, QIB_RCVCTRL_CTXT_ENB | QIB_RCVCTRL_TIDFLOW_ENB, in qib_do_user_init()
1769 rcd->ctxt); in qib_do_user_init()
1772 if (rcd->subctxt_cnt) { in qib_do_user_init()
1773 clear_bit(QIB_CTXT_MASTER_UNINIT, &rcd->flag); in qib_do_user_init()
1774 wake_up(&rcd->wait); in qib_do_user_init()
1779 qib_chg_pioavailkernel(dd, rcd->pio_base, rcd->piocnt, in qib_do_user_init()
1780 TXCHK_CHG_TYPE_KERN, rcd); in qib_do_user_init()
1792 static void unlock_expected_tids(struct qib_ctxtdata *rcd) in unlock_expected_tids() argument
1794 struct qib_devdata *dd = rcd->dd; in unlock_expected_tids()
1795 int ctxt_tidbase = rcd->ctxt * dd->rcvtidcnt; in unlock_expected_tids()
1819 struct qib_ctxtdata *rcd; in qib_close() local
1829 rcd = fd->rcd; in qib_close()
1830 if (!rcd) { in qib_close()
1835 dd = rcd->dd; in qib_close()
1842 qib_user_sdma_queue_drain(rcd->ppd, fd->pq); in qib_close()
1849 if (--rcd->cnt) { in qib_close()
1855 rcd->active_slaves &= ~(1 << fd->subctxt); in qib_close()
1856 rcd->subpid[fd->subctxt] = 0; in qib_close()
1863 ctxt = rcd->ctxt; in qib_close()
1864 dd->rcd[ctxt] = NULL; in qib_close()
1865 pid = rcd->pid; in qib_close()
1866 rcd->pid = 0; in qib_close()
1869 if (rcd->rcvwait_to || rcd->piowait_to || in qib_close()
1870 rcd->rcvnowait || rcd->pionowait) { in qib_close()
1871 rcd->rcvwait_to = 0; in qib_close()
1872 rcd->piowait_to = 0; in qib_close()
1873 rcd->rcvnowait = 0; in qib_close()
1874 rcd->pionowait = 0; in qib_close()
1876 if (rcd->flag) in qib_close()
1877 rcd->flag = 0; in qib_close()
1881 dd->f_rcvctrl(rcd->ppd, QIB_RCVCTRL_CTXT_DIS | in qib_close()
1885 qib_clean_part_key(rcd, dd); in qib_close()
1886 qib_disarm_piobufs(dd, rcd->pio_base, rcd->piocnt); in qib_close()
1887 qib_chg_pioavailkernel(dd, rcd->pio_base, in qib_close()
1888 rcd->piocnt, TXCHK_CHG_TYPE_KERN, NULL); in qib_close()
1890 dd->f_clear_tids(dd, rcd); in qib_close()
1893 unlock_expected_tids(rcd); in qib_close()
1899 qib_free_ctxtdata(dd, rcd); /* after releasing the mutex */ in qib_close()
1911 struct qib_ctxtdata *rcd = ctxt_fp(fp); in qib_ctxt_info() local
1917 info.unit = rcd->dd->unit; in qib_ctxt_info()
1918 info.port = rcd->ppd->port; in qib_ctxt_info()
1919 info.ctxt = rcd->ctxt; in qib_ctxt_info()
1922 info.num_ctxts = rcd->dd->cfgctxts - rcd->dd->first_user_ctxt; in qib_ctxt_info()
1923 info.num_subctxts = rcd->subctxt_cnt; in qib_ctxt_info()
1969 static int disarm_req_delay(struct qib_ctxtdata *rcd) in disarm_req_delay() argument
1973 if (!usable(rcd->ppd)) { in disarm_req_delay()
1981 if (rcd->user_event_mask) { in disarm_req_delay()
1987 &rcd->user_event_mask[0]); in disarm_req_delay()
1988 for (i = 1; i < rcd->subctxt_cnt; i++) in disarm_req_delay()
1990 &rcd->user_event_mask[i]); in disarm_req_delay()
1992 for (i = 0; !usable(rcd->ppd) && i < 300; i++) in disarm_req_delay()
2006 struct qib_ctxtdata *rcd; in qib_set_uevent_bits() local
2014 rcd = ppd->dd->rcd[ctxt]; in qib_set_uevent_bits()
2015 if (!rcd) in qib_set_uevent_bits()
2017 if (rcd->user_event_mask) { in qib_set_uevent_bits()
2023 set_bit(evtbit, &rcd->user_event_mask[0]); in qib_set_uevent_bits()
2024 for (i = 1; i < rcd->subctxt_cnt; i++) in qib_set_uevent_bits()
2025 set_bit(evtbit, &rcd->user_event_mask[i]); in qib_set_uevent_bits()
2044 static int qib_user_event_ack(struct qib_ctxtdata *rcd, int subctxt, in qib_user_event_ack() argument
2053 (void)qib_disarm_piobufs_ifneeded(rcd); in qib_user_event_ack()
2054 ret = disarm_req_delay(rcd); in qib_user_event_ack()
2056 clear_bit(i, &rcd->user_event_mask[subctxt]); in qib_user_event_ack()
2065 struct qib_ctxtdata *rcd; in qib_write() local
2176 rcd = ctxt_fp(fp); in qib_write()
2177 if (!rcd && cmd.type != QIB_CMD_ASSIGN_CTXT) { in qib_write()
2199 ret = qib_manage_rcvq(rcd, subctxt_fp(fp), cmd.cmd.recv_ctrl); in qib_write()
2208 ret = qib_tid_update(rcd, fp, &cmd.cmd.tid_info); in qib_write()
2212 ret = qib_tid_free(rcd, subctxt_fp(fp), &cmd.cmd.tid_info); in qib_write()
2216 ret = qib_set_part_key(rcd, cmd.cmd.part_key); in qib_write()
2220 (void)qib_disarm_piobufs_ifneeded(rcd); in qib_write()
2221 ret = disarm_req_delay(rcd); in qib_write()
2225 qib_force_pio_avail_update(rcd->dd); in qib_write()
2229 rcd->poll_type = cmd.cmd.poll_type; in qib_write()
2233 rcd->dd->f_set_armlaunch(rcd->dd, cmd.cmd.armlaunch_ctrl); in qib_write()
2243 ret = qib_sdma_get_complete(rcd->ppd, in qib_write()
2250 ret = qib_user_event_ack(rcd, subctxt_fp(fp), in qib_write()
2265 struct qib_ctxtdata *rcd = ctxt_fp(iocb->ki_filp); in qib_write_iter() local
2271 return qib_user_sdma_writev(rcd, pq, from->iov, from->nr_segs); in qib_write_iter()