io_tlb_nslabs 77 kernel/dma/swiotlb.c static unsigned long io_tlb_nslabs; io_tlb_nslabs 115 kernel/dma/swiotlb.c io_tlb_nslabs = simple_strtoul(str, &str, 0); io_tlb_nslabs 117 kernel/dma/swiotlb.c io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); io_tlb_nslabs 125 kernel/dma/swiotlb.c io_tlb_nslabs = 1; io_tlb_nslabs 136 kernel/dma/swiotlb.c return unlikely(no_iotlb_memory) ? 0 : io_tlb_nslabs; io_tlb_nslabs 160 kernel/dma/swiotlb.c size = io_tlb_nslabs << IO_TLB_SHIFT; io_tlb_nslabs 167 kernel/dma/swiotlb.c unsigned long bytes = io_tlb_nslabs << IO_TLB_SHIFT; io_tlb_nslabs 195 kernel/dma/swiotlb.c bytes = PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT); io_tlb_nslabs 207 kernel/dma/swiotlb.c io_tlb_nslabs = nslabs; io_tlb_nslabs 216 kernel/dma/swiotlb.c alloc_size = PAGE_ALIGN(io_tlb_nslabs * sizeof(int)); io_tlb_nslabs 222 kernel/dma/swiotlb.c alloc_size = PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t)); io_tlb_nslabs 228 kernel/dma/swiotlb.c for (i = 0; i < io_tlb_nslabs; i++) { io_tlb_nslabs 237 kernel/dma/swiotlb.c swiotlb_set_max_segment(io_tlb_nslabs << IO_TLB_SHIFT); io_tlb_nslabs 252 kernel/dma/swiotlb.c if (!io_tlb_nslabs) { io_tlb_nslabs 253 kernel/dma/swiotlb.c io_tlb_nslabs = (default_size >> IO_TLB_SHIFT); io_tlb_nslabs 254 kernel/dma/swiotlb.c io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); io_tlb_nslabs 257 kernel/dma/swiotlb.c bytes = io_tlb_nslabs << IO_TLB_SHIFT; io_tlb_nslabs 261 kernel/dma/swiotlb.c if (vstart && !swiotlb_init_with_tbl(vstart, io_tlb_nslabs, verbose)) io_tlb_nslabs 266 kernel/dma/swiotlb.c PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT)); io_tlb_nslabs 279 kernel/dma/swiotlb.c unsigned long bytes, req_nslabs = io_tlb_nslabs; io_tlb_nslabs 284 kernel/dma/swiotlb.c if (!io_tlb_nslabs) { io_tlb_nslabs 285 kernel/dma/swiotlb.c io_tlb_nslabs = (default_size >> IO_TLB_SHIFT); io_tlb_nslabs 286 kernel/dma/swiotlb.c io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); io_tlb_nslabs 292 kernel/dma/swiotlb.c order = get_order(io_tlb_nslabs << IO_TLB_SHIFT); io_tlb_nslabs 293 kernel/dma/swiotlb.c io_tlb_nslabs = SLABS_PER_PAGE << order; io_tlb_nslabs 294 kernel/dma/swiotlb.c bytes = io_tlb_nslabs << IO_TLB_SHIFT; io_tlb_nslabs 305 kernel/dma/swiotlb.c io_tlb_nslabs = req_nslabs; io_tlb_nslabs 311 kernel/dma/swiotlb.c io_tlb_nslabs = SLABS_PER_PAGE << order; io_tlb_nslabs 313 kernel/dma/swiotlb.c rc = swiotlb_late_init_with_tbl(vstart, io_tlb_nslabs); io_tlb_nslabs 324 kernel/dma/swiotlb.c io_tlb_nslabs = 0; io_tlb_nslabs 335 kernel/dma/swiotlb.c io_tlb_nslabs = nslabs; io_tlb_nslabs 348 kernel/dma/swiotlb.c get_order(io_tlb_nslabs * sizeof(int))); io_tlb_nslabs 354 kernel/dma/swiotlb.c get_order(io_tlb_nslabs * io_tlb_nslabs 359 kernel/dma/swiotlb.c for (i = 0; i < io_tlb_nslabs; i++) { io_tlb_nslabs 369 kernel/dma/swiotlb.c swiotlb_set_max_segment(io_tlb_nslabs << IO_TLB_SHIFT); io_tlb_nslabs 374 kernel/dma/swiotlb.c free_pages((unsigned long)io_tlb_list, get_order(io_tlb_nslabs * io_tlb_nslabs 389 kernel/dma/swiotlb.c get_order(io_tlb_nslabs * sizeof(phys_addr_t))); io_tlb_nslabs 390 kernel/dma/swiotlb.c free_pages((unsigned long)io_tlb_list, get_order(io_tlb_nslabs * io_tlb_nslabs 393 kernel/dma/swiotlb.c get_order(io_tlb_nslabs << IO_TLB_SHIFT)); io_tlb_nslabs 396 kernel/dma/swiotlb.c PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); io_tlb_nslabs 398 kernel/dma/swiotlb.c PAGE_ALIGN(io_tlb_nslabs * sizeof(int))); io_tlb_nslabs 400 kernel/dma/swiotlb.c PAGE_ALIGN(io_tlb_nslabs << IO_TLB_SHIFT)); io_tlb_nslabs 505 kernel/dma/swiotlb.c if (unlikely(nslots > io_tlb_nslabs - io_tlb_used)) io_tlb_nslabs 509 kernel/dma/swiotlb.c if (index >= io_tlb_nslabs) io_tlb_nslabs 517 kernel/dma/swiotlb.c if (index >= io_tlb_nslabs) io_tlb_nslabs 541 kernel/dma/swiotlb.c io_tlb_index = ((index + nslots) < io_tlb_nslabs io_tlb_nslabs 547 kernel/dma/swiotlb.c if (index >= io_tlb_nslabs) io_tlb_nslabs 557 kernel/dma/swiotlb.c alloc_size, io_tlb_nslabs, tmp_io_tlb_used); io_tlb_nslabs 711 kernel/dma/swiotlb.c debugfs_create_ulong("io_tlb_nslabs", 0400, root, &io_tlb_nslabs);