IO_TLB_SEGSIZE    240 arch/mips/cavium-octeon/dma-octeon.c 	swiotlb_nslabs = ALIGN(swiotlb_nslabs, IO_TLB_SEGSIZE);
IO_TLB_SEGSIZE   4196 drivers/mmc/host/sdhci.c 						IO_TLB_SEGSIZE;
IO_TLB_SEGSIZE    127 drivers/xen/swiotlb-xen.c 	dma_bits = get_order(IO_TLB_SEGSIZE << IO_TLB_SHIFT) + PAGE_SHIFT;
IO_TLB_SEGSIZE    131 drivers/xen/swiotlb-xen.c 		int slabs = min(nslabs - i, (unsigned long)IO_TLB_SEGSIZE);
IO_TLB_SEGSIZE    150 drivers/xen/swiotlb-xen.c 		xen_io_tlb_nslabs = ALIGN(xen_io_tlb_nslabs, IO_TLB_SEGSIZE);
IO_TLB_SEGSIZE    117 kernel/dma/swiotlb.c 		io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE);
IO_TLB_SEGSIZE    229 kernel/dma/swiotlb.c 		io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE);
IO_TLB_SEGSIZE    254 kernel/dma/swiotlb.c 		io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE);
IO_TLB_SEGSIZE    286 kernel/dma/swiotlb.c 		io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE);
IO_TLB_SEGSIZE    360 kernel/dma/swiotlb.c 		io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE);
IO_TLB_SEGSIZE    533 kernel/dma/swiotlb.c 			for (i = index - 1; (OFFSET(i, IO_TLB_SEGSIZE) != IO_TLB_SEGSIZE - 1) && io_tlb_list[i]; i--)
IO_TLB_SEGSIZE    605 kernel/dma/swiotlb.c 		count = ((index + nslots) < ALIGN(index + 1, IO_TLB_SEGSIZE) ?
IO_TLB_SEGSIZE    619 kernel/dma/swiotlb.c 		for (i = index - 1; (OFFSET(i, IO_TLB_SEGSIZE) != IO_TLB_SEGSIZE -1) && io_tlb_list[i]; i--)
IO_TLB_SEGSIZE    692 kernel/dma/swiotlb.c 	return ((size_t)1 << IO_TLB_SHIFT) * IO_TLB_SEGSIZE;