Lines Matching refs:ppd
105 struct qib_pportdata *ppd = rcd->ppd; in qib_get_base_info() local
161 kinfo->spi_port = ppd->port; in qib_get_base_info()
191 (char *) ppd->statusp - in qib_get_base_info()
241 kinfo->spi_mtu = ppd->ibmaxlen; /* maxlen, not ibmtu */ in qib_get_base_info()
570 struct qib_pportdata *ppd = rcd->ppd; in qib_set_part_key() local
606 for (any = i = 0; i < ARRAY_SIZE(ppd->pkeys); i++) { in qib_set_part_key()
607 if (!ppd->pkeys[i]) { in qib_set_part_key()
611 if (ppd->pkeys[i] == key) { in qib_set_part_key()
612 atomic_t *pkrefs = &ppd->pkeyrefs[i]; in qib_set_part_key()
626 if ((ppd->pkeys[i] & 0x7FFF) == lkey) { in qib_set_part_key()
640 for (any = i = 0; i < ARRAY_SIZE(ppd->pkeys); i++) { in qib_set_part_key()
641 if (!ppd->pkeys[i] && in qib_set_part_key()
642 atomic_inc_return(&ppd->pkeyrefs[i]) == 1) { in qib_set_part_key()
644 ppd->pkeys[i] = key; in qib_set_part_key()
645 (void) ppd->dd->f_set_ib_cfg(ppd, QIB_IB_CFG_PKEYS, 0); in qib_set_part_key()
689 dd->f_rcvctrl(rcd->ppd, rcvctrl_op, rcd->ctxt); in qib_manage_rcvq()
700 struct qib_pportdata *ppd = rcd->ppd; in qib_clean_part_key() local
703 oldpkey = (u64) ppd->pkeys[0] | in qib_clean_part_key()
704 ((u64) ppd->pkeys[1] << 16) | in qib_clean_part_key()
705 ((u64) ppd->pkeys[2] << 32) | in qib_clean_part_key()
706 ((u64) ppd->pkeys[3] << 48); in qib_clean_part_key()
711 for (j = 0; j < ARRAY_SIZE(ppd->pkeys); j++) { in qib_clean_part_key()
713 if ((ppd->pkeys[j] & 0x7fff) != in qib_clean_part_key()
716 if (atomic_dec_and_test(&ppd->pkeyrefs[j])) { in qib_clean_part_key()
717 ppd->pkeys[j] = 0; in qib_clean_part_key()
725 (void) ppd->dd->f_set_ib_cfg(ppd, QIB_IB_CFG_PKEYS, 0); in qib_clean_part_key()
1139 dd->f_rcvctrl(rcd->ppd, QIB_RCVCTRL_INTRAVAIL_ENB, rcd->ctxt); in qib_poll_next()
1308 static int setup_ctxt(struct qib_pportdata *ppd, int ctxt, in setup_ctxt() argument
1312 struct qib_devdata *dd = ppd->dd; in setup_ctxt()
1324 rcd = qib_create_ctxtdata(ppd, ctxt, numa_id); in setup_ctxt()
1366 static inline int usable(struct qib_pportdata *ppd) in usable() argument
1368 struct qib_devdata *dd = ppd->dd; in usable()
1370 return dd && (dd->flags & QIB_PRESENT) && dd->kregbase && ppd->lid && in usable()
1371 (ppd->lflags & QIBL_LINKACTIVE); in usable()
1381 struct qib_pportdata *ppd = NULL; in choose_port_ctxt() local
1389 ppd = dd->pport + port - 1; in choose_port_ctxt()
1398 if (!ppd) { in choose_port_ctxt()
1402 ppd = dd->pport + pidx; in choose_port_ctxt()
1404 for (pidx = 0; pidx < dd->num_pports && !ppd; in choose_port_ctxt()
1407 ppd = dd->pport + pidx; in choose_port_ctxt()
1410 ret = ppd ? setup_ctxt(ppd, ctxt, fp, uinfo) : -ENETDOWN; in choose_port_ctxt()
1768 dd->f_rcvctrl(rcd->ppd, QIB_RCVCTRL_CTXT_ENB | QIB_RCVCTRL_TIDFLOW_ENB, in qib_do_user_init()
1842 qib_user_sdma_queue_drain(rcd->ppd, fd->pq); in qib_close()
1881 dd->f_rcvctrl(rcd->ppd, QIB_RCVCTRL_CTXT_DIS | in qib_close()
1918 info.port = rcd->ppd->port; in qib_ctxt_info()
1948 static int qib_sdma_get_complete(struct qib_pportdata *ppd, in qib_sdma_get_complete() argument
1958 err = qib_user_sdma_make_progress(ppd, pq); in qib_sdma_get_complete()
1973 if (!usable(rcd->ppd)) { in disarm_req_delay()
1992 for (i = 0; !usable(rcd->ppd) && i < 300; i++) in disarm_req_delay()
2004 int qib_set_uevent_bits(struct qib_pportdata *ppd, const int evtbit) in qib_set_uevent_bits() argument
2011 spin_lock_irqsave(&ppd->dd->uctxt_lock, flags); in qib_set_uevent_bits()
2012 for (ctxt = ppd->dd->first_user_ctxt; ctxt < ppd->dd->cfgctxts; in qib_set_uevent_bits()
2014 rcd = ppd->dd->rcd[ctxt]; in qib_set_uevent_bits()
2030 spin_unlock_irqrestore(&ppd->dd->uctxt_lock, flags); in qib_set_uevent_bits()
2243 ret = qib_sdma_get_complete(rcd->ppd, in qib_write()