Lines Matching refs:size

102 	arena->size = window_size;  in iommu_arena_new_node()
140 nent = arena->size >> PAGE_SHIFT; in iommu_arena_find_pages()
251 pci_map_single_1(struct pci_dev *pdev, void *cpu_addr, size_t size, in pci_map_single_1() argument
267 if (paddr + size + __direct_map_base - 1 <= max_dma in pci_map_single_1()
268 && paddr + size <= __direct_map_size) { in pci_map_single_1()
272 cpu_addr, size, ret, __builtin_return_address(0)); in pci_map_single_1()
283 cpu_addr, size, ret, __builtin_return_address(0)); in pci_map_single_1()
297 if (!arena || arena->dma_base + arena->size - 1 > max_dma) in pci_map_single_1()
300 npages = iommu_num_pages(paddr, size, PAGE_SIZE); in pci_map_single_1()
320 cpu_addr, size, npages, ret, __builtin_return_address(0)); in pci_map_single_1()
350 unsigned long offset, size_t size, in alpha_pci_map_page() argument
361 size, dac_allowed); in alpha_pci_map_page()
371 size_t size, enum dma_data_direction dir, in alpha_pci_unmap_page() argument
387 dma_addr, size, __builtin_return_address(0)); in alpha_pci_unmap_page()
394 dma_addr, size, __builtin_return_address(0)); in alpha_pci_unmap_page()
403 if (dma_ofs * PAGE_SIZE >= arena->size) { in alpha_pci_unmap_page()
406 dma_addr, arena->dma_base, arena->size); in alpha_pci_unmap_page()
411 npages = iommu_num_pages(dma_addr, size, PAGE_SIZE); in alpha_pci_unmap_page()
421 alpha_mv.mv_pci_tbi(hose, dma_addr, dma_addr + size - 1); in alpha_pci_unmap_page()
426 dma_addr, size, npages, __builtin_return_address(0)); in alpha_pci_unmap_page()
434 static void *alpha_pci_alloc_coherent(struct device *dev, size_t size, in alpha_pci_alloc_coherent() argument
440 long order = get_order(size); in alpha_pci_alloc_coherent()
454 memset(cpu_addr, 0, size); in alpha_pci_alloc_coherent()
456 *dma_addrp = pci_map_single_1(pdev, cpu_addr, size, 0); in alpha_pci_alloc_coherent()
468 size, cpu_addr, *dma_addrp, __builtin_return_address(0)); in alpha_pci_alloc_coherent()
479 static void alpha_pci_free_coherent(struct device *dev, size_t size, in alpha_pci_free_coherent() argument
484 pci_unmap_single(pdev, dma_addr, size, PCI_DMA_BIDIRECTIONAL); in alpha_pci_free_coherent()
485 free_pages((unsigned long)cpu_addr, get_order(size)); in alpha_pci_free_coherent()
488 dma_addr, size, __builtin_return_address(0)); in alpha_pci_free_coherent()
559 long size = leader->dma_length; in sg_fill() local
568 && paddr + size + __direct_map_base - 1 <= max_dma in sg_fill()
569 && paddr + size <= __direct_map_size) { in sg_fill()
571 out->dma_length = size; in sg_fill()
574 __va(paddr), size, out->dma_address); in sg_fill()
583 out->dma_length = size; in sg_fill()
586 __va(paddr), size, out->dma_address); in sg_fill()
595 npages = iommu_num_pages(paddr, size, PAGE_SIZE); in sg_fill()
609 out->dma_length = size; in sg_fill()
612 __va(paddr), size, out->dma_address, npages); in sg_fill()
623 size = sg->length; in sg_fill()
627 size += sg[1].length; in sg_fill()
631 npages = iommu_num_pages(paddr, size, PAGE_SIZE); in sg_fill()
687 if (!arena || arena->dma_base + arena->size - 1 > max_dma) in alpha_pci_map_sg()
750 if (!arena || arena->dma_base + arena->size - 1 > max_dma) in alpha_pci_unmap_sg()
759 size_t size; in alpha_pci_unmap_sg() local
764 size = sg->dma_length; in alpha_pci_unmap_sg()
765 if (!size) in alpha_pci_unmap_sg()
771 sg - end + nents, addr, size); in alpha_pci_unmap_sg()
779 sg - end + nents, addr, size); in alpha_pci_unmap_sg()
784 sg - end + nents, addr, size); in alpha_pci_unmap_sg()
786 npages = iommu_num_pages(addr, size, PAGE_SIZE); in alpha_pci_unmap_sg()
790 tend = addr + size - 1; in alpha_pci_unmap_sg()
826 if (arena && arena->dma_base + arena->size - 1 <= mask) in alpha_pci_supported()
829 if (arena && arena->dma_base + arena->size - 1 <= mask) in alpha_pci_supported()