Lines Matching refs:lldi

70 	struct cxgb4_lld_info lldi;  member
127 le.poll_sge_ts = cxgb4_read_sge_timestamp(wq->rdev->lldi.ports[0]); in c4iw_log_wr_stats()
155 #define ts2ns(ts) div64_u64((ts) * dev->rdev.lldi.cclk_ps, 1000) in wr_log_show()
373 ret = cxgb4_read_tpte(stagd->devp->rdev.lldi.ports[0], (u32)id<<8, in dump_stag()
376 dev_err(&stagd->devp->rdev.lldi.pdev->dev, in dump_stag()
780 if (rdev->lldi.udb_density != rdev->lldi.ucq_density) { in c4iw_rdev_open()
782 pci_name(rdev->lldi.pdev), rdev->lldi.udb_density, in c4iw_rdev_open()
783 rdev->lldi.ucq_density); in c4iw_rdev_open()
787 if (rdev->lldi.vr->qp.start != rdev->lldi.vr->cq.start || in c4iw_rdev_open()
788 rdev->lldi.vr->qp.size != rdev->lldi.vr->cq.size) { in c4iw_rdev_open()
791 pci_name(rdev->lldi.pdev), rdev->lldi.vr->qp.start, in c4iw_rdev_open()
792 rdev->lldi.vr->qp.size, rdev->lldi.vr->cq.size, in c4iw_rdev_open()
793 rdev->lldi.vr->cq.size); in c4iw_rdev_open()
798 rdev->qpmask = rdev->lldi.udb_density - 1; in c4iw_rdev_open()
799 rdev->cqmask = rdev->lldi.ucq_density - 1; in c4iw_rdev_open()
803 __func__, pci_name(rdev->lldi.pdev), rdev->lldi.vr->stag.start, in c4iw_rdev_open()
804 rdev->lldi.vr->stag.size, c4iw_num_stags(rdev), in c4iw_rdev_open()
805 rdev->lldi.vr->pbl.start, in c4iw_rdev_open()
806 rdev->lldi.vr->pbl.size, rdev->lldi.vr->rq.start, in c4iw_rdev_open()
807 rdev->lldi.vr->rq.size, in c4iw_rdev_open()
808 rdev->lldi.vr->qp.start, in c4iw_rdev_open()
809 rdev->lldi.vr->qp.size, in c4iw_rdev_open()
810 rdev->lldi.vr->cq.start, in c4iw_rdev_open()
811 rdev->lldi.vr->cq.size); in c4iw_rdev_open()
814 (unsigned)pci_resource_len(rdev->lldi.pdev, 2), in c4iw_rdev_open()
815 (void *)pci_resource_start(rdev->lldi.pdev, 2), in c4iw_rdev_open()
816 rdev->lldi.db_reg, rdev->lldi.gts_reg, in c4iw_rdev_open()
825 rdev->stats.stag.total = rdev->lldi.vr->stag.size; in c4iw_rdev_open()
826 rdev->stats.pbl.total = rdev->lldi.vr->pbl.size; in c4iw_rdev_open()
827 rdev->stats.rqt.total = rdev->lldi.vr->rq.size; in c4iw_rdev_open()
828 rdev->stats.ocqp.total = rdev->lldi.vr->ocq.size; in c4iw_rdev_open()
829 rdev->stats.qid.total = rdev->lldi.vr->qp.size; in c4iw_rdev_open()
941 devp->rdev.lldi = *infop; in c4iw_alloc()
945 __func__, devp->rdev.lldi.sge_ingpadboundary, in c4iw_alloc()
946 devp->rdev.lldi.sge_egrstatuspagesize); in c4iw_alloc()
949 devp->rdev.lldi.sge_ingpadboundary > 64 ? 2 : 1; in c4iw_alloc()
962 devp->rdev.lldi.sge_egrstatuspagesize; in c4iw_alloc()
969 devp->rdev.bar2_pa = pci_resource_start(devp->rdev.lldi.pdev, 2); in c4iw_alloc()
970 if (!is_t4(devp->rdev.lldi.adapter_type)) { in c4iw_alloc()
972 pci_resource_len(devp->rdev.lldi.pdev, 2)); in c4iw_alloc()
980 pci_resource_start(devp->rdev.lldi.pdev, 2) + in c4iw_alloc()
981 pci_resource_len(devp->rdev.lldi.pdev, 2) - in c4iw_alloc()
982 roundup_pow_of_two(devp->rdev.lldi.vr->ocq.size); in c4iw_alloc()
984 devp->rdev.lldi.vr->ocq.size); in c4iw_alloc()
994 devp->rdev.lldi.vr->ocq.start, devp->rdev.lldi.vr->ocq.size, in c4iw_alloc()
1014 devp->avail_ird = devp->rdev.lldi.max_ird_adapter; in c4iw_alloc()
1018 pci_name(devp->rdev.lldi.pdev), in c4iw_alloc()
1042 ctx->lldi = *infop; in c4iw_uld_add()
1045 __func__, pci_name(ctx->lldi.pdev), in c4iw_uld_add()
1046 ctx->lldi.nchan, ctx->lldi.nrxq, in c4iw_uld_add()
1047 ctx->lldi.ntxq, ctx->lldi.nports); in c4iw_uld_add()
1053 for (i = 0; i < ctx->lldi.nrxq; i++) in c4iw_uld_add()
1054 PDBG("rxqid[%u] %u\n", i, ctx->lldi.rxq_ids[i]); in c4iw_uld_add()
1106 skb = copy_gl_to_skb_pkt(gl , rsp, dev->rdev.lldi.sge_pktshift); in recv_rx_pkt()
1151 pci_name(ctx->lldi.pdev), gl->va, in c4iw_uld_rx_handler()
1185 printk(KERN_INFO MOD "%s: Up\n", pci_name(ctx->lldi.pdev)); in c4iw_uld_state_change()
1189 ctx->dev = c4iw_alloc(&ctx->lldi); in c4iw_uld_state_change()
1193 pci_name(ctx->lldi.pdev), in c4iw_uld_state_change()
1202 pci_name(ctx->lldi.pdev), ret); in c4iw_uld_state_change()
1209 pci_name(ctx->lldi.pdev)); in c4iw_uld_state_change()
1215 pci_name(ctx->lldi.pdev)); in c4iw_uld_state_change()
1229 pci_name(ctx->lldi.pdev)); in c4iw_uld_state_change()
1311 if (cxgb4_dbfifo_count(ctx->dev->rdev.lldi.ports[0], 1) in resume_queues()
1312 < (ctx->dev->rdev.lldi.dbfifo_int_thresh << in resume_queues()
1374 ret = cxgb4_sync_txq_pidx(qp->rhp->rdev.lldi.ports[0], in recover_lost_dbs()
1382 pci_name(ctx->lldi.pdev), qp->wq.sq.qid); in recover_lost_dbs()
1389 ret = cxgb4_sync_txq_pidx(qp->rhp->rdev.lldi.ports[0], in recover_lost_dbs()
1398 pci_name(ctx->lldi.pdev), qp->wq.rq.qid); in recover_lost_dbs()
1408 while (cxgb4_dbfifo_count(qp->rhp->rdev.lldi.ports[0], 1) > 0) { in recover_lost_dbs()
1426 ret = cxgb4_flush_eq_cache(ctx->dev->rdev.lldi.ports[0]); in recover_queues()
1429 pci_name(ctx->lldi.pdev)); in recover_queues()
1442 pci_name(ctx->lldi.pdev)); in recover_queues()
1489 pci_name(ctx->lldi.pdev), control); in c4iw_uld_control()