Lines Matching refs:dd

98 	struct qib_devdata *dd;  in qib_count_active_units()  local
104 list_for_each_entry(dd, &qib_dev_list, list) { in qib_count_active_units()
105 if (!(dd->flags & QIB_PRESENT) || !dd->kregbase) in qib_count_active_units()
107 for (pidx = 0; pidx < dd->num_pports; ++pidx) { in qib_count_active_units()
108 ppd = dd->pport + pidx; in qib_count_active_units()
128 struct qib_devdata *dd; in qib_count_units() local
135 list_for_each_entry(dd, &qib_dev_list, list) { in qib_count_units()
137 if ((dd->flags & QIB_PRESENT) && dd->kregbase) in qib_count_units()
139 for (pidx = 0; pidx < dd->num_pports; ++pidx) { in qib_count_units()
140 ppd = dd->pport + pidx; in qib_count_units()
201 struct qib_devdata *dd = ppd->dd; in qib_set_linkstate() local
206 dd->f_set_ib_cfg(ppd, QIB_IB_CFG_LSTATE, in qib_set_linkstate()
213 dd->f_set_ib_cfg(ppd, QIB_IB_CFG_LSTATE, in qib_set_linkstate()
220 dd->f_set_ib_cfg(ppd, QIB_IB_CFG_LSTATE, in qib_set_linkstate()
227 dd->f_set_ib_cfg(ppd, QIB_IB_CFG_LSTATE, in qib_set_linkstate()
251 dd->f_set_ib_cfg(ppd, QIB_IB_CFG_LSTATE, in qib_set_linkstate()
265 dd->f_set_ib_cfg(ppd, QIB_IB_CFG_LSTATE, in qib_set_linkstate()
289 return rcd->rcvegrbuf[chunk] + (idx << rcd->dd->rcvegrbufsize_shift); in qib_get_egrbuf()
448 struct qib_devdata *dd = rcd->dd; in qib_kreceive() local
452 const u32 rsize = dd->rcvhdrentsize; /* words */ in qib_kreceive()
453 const u32 maxcnt = dd->rcvhdrcnt * rsize; /* words */ in qib_kreceive()
462 rhf_addr = (__le32 *) rcd->rcvhdrq + l + dd->rhf_offset; in qib_kreceive()
463 if (dd->flags & QIB_NODMA_RTAIL) { in qib_kreceive()
477 hdr = dd->f_get_msgheader(dd, rhf_addr); in qib_kreceive()
483 if ((dd->flags & QIB_NODMA_RTAIL) ? in qib_kreceive()
504 tlen > (dd->rcvhdrentsize - 2 + 1 - in qib_kreceive()
530 rhf_addr = (__le32 *) rcd->rcvhdrq + l + dd->rhf_offset; in qib_kreceive()
531 if (dd->flags & QIB_NODMA_RTAIL) { in qib_kreceive()
548 dd->f_update_usrhead(rcd, lval, updegr, etail, i); in qib_kreceive()
597 lval = (u64)rcd->head | dd->rhdrhead_intr_off; in qib_kreceive()
598 dd->f_update_usrhead(rcd, lval, updegr, etail, i); in qib_kreceive()
645 ppd->dd->f_set_ib_cfg(ppd, QIB_IB_CFG_MTU, 0); in qib_set_mtu()
655 struct qib_devdata *dd = ppd->dd; in qib_set_lid() local
660 dd->f_set_ib_cfg(ppd, QIB_IB_CFG_LIDLMC, in qib_set_lid()
663 qib_devinfo(dd->pcidev, "IB%u:%u got a lid: 0x%x\n", in qib_set_lid()
664 dd->unit, ppd->port, lid); in qib_set_lid()
684 struct qib_devdata *dd = ppd->dd; in qib_run_led_override() local
688 if (!(dd->flags & QIB_INITTED)) in qib_run_led_override()
695 dd->f_setextled(ppd, 1); in qib_run_led_override()
706 struct qib_devdata *dd = ppd->dd; in qib_set_led_override() local
709 if (!(dd->flags & QIB_INITTED)) in qib_set_led_override()
758 struct qib_devdata *dd = qib_lookup(unit); in qib_reset_device() local
763 if (!dd) { in qib_reset_device()
768 qib_devinfo(dd->pcidev, "Reset on unit %u requested\n", unit); in qib_reset_device()
770 if (!dd->kregbase || !(dd->flags & QIB_PRESENT)) { in qib_reset_device()
771 qib_devinfo(dd->pcidev, in qib_reset_device()
778 spin_lock_irqsave(&dd->uctxt_lock, flags); in qib_reset_device()
779 if (dd->rcd) in qib_reset_device()
780 for (i = dd->first_user_ctxt; i < dd->cfgctxts; i++) { in qib_reset_device()
781 if (!dd->rcd[i] || !dd->rcd[i]->cnt) in qib_reset_device()
783 spin_unlock_irqrestore(&dd->uctxt_lock, flags); in qib_reset_device()
787 spin_unlock_irqrestore(&dd->uctxt_lock, flags); in qib_reset_device()
789 for (pidx = 0; pidx < dd->num_pports; ++pidx) { in qib_reset_device()
790 ppd = dd->pport + pidx; in qib_reset_device()
799 dd->f_setextled(ppd, 0); in qib_reset_device()
800 if (dd->flags & QIB_HAS_SEND_DMA) in qib_reset_device()
804 ret = dd->f_reset(dd); in qib_reset_device()
806 ret = qib_init(dd, 1); in qib_reset_device()
810 qib_dev_err(dd, in qib_reset_device()
814 qib_devinfo(dd->pcidev, in qib_reset_device()