Lines Matching refs:nr_grefs

77 		   grant_ref_t *gnt_refs, unsigned int nr_grefs,
470 unsigned int nr_grefs, void **vaddr) in xenbus_map_ring_valloc() argument
472 return ring_ops->map(dev, gnt_refs, nr_grefs, vaddr); in xenbus_map_ring_valloc()
481 unsigned int nr_grefs, in __xenbus_map_ring() argument
492 if (nr_grefs > XENBUS_MAX_RING_GRANTS) in __xenbus_map_ring()
495 for (i = 0; i < nr_grefs; i++) { in __xenbus_map_ring()
504 for (i = 0; i < nr_grefs; i++) { in __xenbus_map_ring()
518 for (i = j = 0; i < nr_grefs; i++) { in __xenbus_map_ring()
543 unsigned int nr_grefs, in xenbus_map_ring_valloc_pv() argument
556 if (nr_grefs > XENBUS_MAX_RING_GRANTS) in xenbus_map_ring_valloc_pv()
563 area = alloc_vm_area(XEN_PAGE_SIZE * nr_grefs, ptes); in xenbus_map_ring_valloc_pv()
569 for (i = 0; i < nr_grefs; i++) in xenbus_map_ring_valloc_pv()
572 err = __xenbus_map_ring(dev, gnt_refs, nr_grefs, node->handles, in xenbus_map_ring_valloc_pv()
579 node->nr_handles = nr_grefs; in xenbus_map_ring_valloc_pv()
593 pr_alert("leaking VM area %p size %u page(s)", area, nr_grefs); in xenbus_map_ring_valloc_pv()
624 unsigned int nr_grefs, in xenbus_map_ring_valloc_hvm() argument
634 unsigned int nr_pages = XENBUS_PAGES(nr_grefs); in xenbus_map_ring_valloc_hvm()
636 if (nr_grefs > XENBUS_MAX_RING_GRANTS) in xenbus_map_ring_valloc_hvm()
649 gnttab_foreach_grant(node->hvm.pages, nr_grefs, in xenbus_map_ring_valloc_hvm()
653 err = __xenbus_map_ring(dev, gnt_ref, nr_grefs, node->handles, in xenbus_map_ring_valloc_hvm()
655 node->nr_handles = nr_grefs; in xenbus_map_ring_valloc_hvm()
678 xenbus_unmap_ring(dev, node->handles, nr_grefs, info.addrs); in xenbus_map_ring_valloc_hvm()
711 unsigned int nr_grefs, grant_handle_t *handles, in xenbus_map_ring() argument
717 if (nr_grefs > XENBUS_MAX_RING_GRANTS) in xenbus_map_ring()
720 for (i = 0; i < nr_grefs; i++) in xenbus_map_ring()
723 return __xenbus_map_ring(dev, gnt_refs, nr_grefs, handles, in xenbus_map_ring()