IO_TLB_SEGSIZE 240 arch/mips/cavium-octeon/dma-octeon.c swiotlb_nslabs = ALIGN(swiotlb_nslabs, IO_TLB_SEGSIZE); IO_TLB_SEGSIZE 4196 drivers/mmc/host/sdhci.c IO_TLB_SEGSIZE; IO_TLB_SEGSIZE 127 drivers/xen/swiotlb-xen.c dma_bits = get_order(IO_TLB_SEGSIZE << IO_TLB_SHIFT) + PAGE_SHIFT; IO_TLB_SEGSIZE 131 drivers/xen/swiotlb-xen.c int slabs = min(nslabs - i, (unsigned long)IO_TLB_SEGSIZE); IO_TLB_SEGSIZE 150 drivers/xen/swiotlb-xen.c xen_io_tlb_nslabs = ALIGN(xen_io_tlb_nslabs, IO_TLB_SEGSIZE); IO_TLB_SEGSIZE 117 kernel/dma/swiotlb.c io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); IO_TLB_SEGSIZE 229 kernel/dma/swiotlb.c io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE); IO_TLB_SEGSIZE 254 kernel/dma/swiotlb.c io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); IO_TLB_SEGSIZE 286 kernel/dma/swiotlb.c io_tlb_nslabs = ALIGN(io_tlb_nslabs, IO_TLB_SEGSIZE); IO_TLB_SEGSIZE 360 kernel/dma/swiotlb.c io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE); IO_TLB_SEGSIZE 533 kernel/dma/swiotlb.c for (i = index - 1; (OFFSET(i, IO_TLB_SEGSIZE) != IO_TLB_SEGSIZE - 1) && io_tlb_list[i]; i--) IO_TLB_SEGSIZE 605 kernel/dma/swiotlb.c count = ((index + nslots) < ALIGN(index + 1, IO_TLB_SEGSIZE) ? IO_TLB_SEGSIZE 619 kernel/dma/swiotlb.c for (i = index - 1; (OFFSET(i, IO_TLB_SEGSIZE) != IO_TLB_SEGSIZE -1) && io_tlb_list[i]; i--) IO_TLB_SEGSIZE 692 kernel/dma/swiotlb.c return ((size_t)1 << IO_TLB_SHIFT) * IO_TLB_SEGSIZE;