Lines Matching refs:r

226 static int ps3_mm_get_repository_highmem(struct mem_region *r)  in ps3_mm_get_repository_highmem()  argument
232 result = ps3_repository_read_highmem_info(0, &r->base, &r->size); in ps3_mm_get_repository_highmem()
237 if (!r->base || !r->size) { in ps3_mm_get_repository_highmem()
242 r->offset = r->base - map.rm.size; in ps3_mm_get_repository_highmem()
245 __func__, __LINE__, r->base, r->size); in ps3_mm_get_repository_highmem()
252 r->size = r->base = r->offset = 0; in ps3_mm_get_repository_highmem()
256 static int ps3_mm_set_repository_highmem(const struct mem_region *r) in ps3_mm_set_repository_highmem() argument
260 return r ? ps3_repository_write_highmem_info(0, r->base, r->size) : in ps3_mm_set_repository_highmem()
273 static int ps3_mm_region_create(struct mem_region *r, unsigned long size) in ps3_mm_region_create() argument
278 r->size = _ALIGN_DOWN(size, 1 << PAGE_SHIFT_16M); in ps3_mm_region_create()
281 DBG("%s:%d actual %llxh\n", __func__, __LINE__, r->size); in ps3_mm_region_create()
283 size - r->size, (size - r->size) / 1024 / 1024); in ps3_mm_region_create()
285 if (r->size == 0) { in ps3_mm_region_create()
291 result = lv1_allocate_memory(r->size, PAGE_SHIFT_16M, 0, in ps3_mm_region_create()
292 ALLOCATE_MEMORY_TRY_ALT_UNIT, &r->base, &muid); in ps3_mm_region_create()
294 if (result || r->base < map.rm.size) { in ps3_mm_region_create()
300 r->destroy = 1; in ps3_mm_region_create()
301 r->offset = r->base - map.rm.size; in ps3_mm_region_create()
305 r->size = r->base = r->offset = 0; in ps3_mm_region_create()
314 static void ps3_mm_region_destroy(struct mem_region *r) in ps3_mm_region_destroy() argument
318 if (!r->destroy) { in ps3_mm_region_destroy()
320 __func__, __LINE__, r->base, r->size); in ps3_mm_region_destroy()
324 DBG("%s:%d: r->base = %llxh\n", __func__, __LINE__, r->base); in ps3_mm_region_destroy()
326 if (r->base) { in ps3_mm_region_destroy()
327 result = lv1_release_memory(r->base); in ps3_mm_region_destroy()
329 r->size = r->base = r->offset = 0; in ps3_mm_region_destroy()
345 static unsigned long dma_sb_lpar_to_bus(struct ps3_dma_region *r, in dma_sb_lpar_to_bus() argument
350 BUG_ON(lpar_addr < r->offset); in dma_sb_lpar_to_bus()
351 BUG_ON(lpar_addr >= r->offset + r->len); in dma_sb_lpar_to_bus()
352 return r->bus_addr + lpar_addr - r->offset; in dma_sb_lpar_to_bus()
356 static void __maybe_unused _dma_dump_region(const struct ps3_dma_region *r, in _dma_dump_region() argument
359 DBG("%s:%d: dev %llu:%llu\n", func, line, r->dev->bus_id, in _dma_dump_region()
360 r->dev->dev_id); in _dma_dump_region()
361 DBG("%s:%d: page_size %u\n", func, line, r->page_size); in _dma_dump_region()
362 DBG("%s:%d: bus_addr %lxh\n", func, line, r->bus_addr); in _dma_dump_region()
363 DBG("%s:%d: len %lxh\n", func, line, r->len); in _dma_dump_region()
364 DBG("%s:%d: offset %lxh\n", func, line, r->offset); in _dma_dump_region()
405 static struct dma_chunk * dma_find_chunk(struct ps3_dma_region *r, in dma_find_chunk() argument
409 unsigned long aligned_bus = _ALIGN_DOWN(bus_addr, 1 << r->page_size); in dma_find_chunk()
411 1 << r->page_size); in dma_find_chunk()
413 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk()
434 static struct dma_chunk *dma_find_chunk_lpar(struct ps3_dma_region *r, in dma_find_chunk_lpar() argument
438 unsigned long aligned_lpar = _ALIGN_DOWN(lpar_addr, 1 << r->page_size); in dma_find_chunk_lpar()
440 1 << r->page_size); in dma_find_chunk_lpar()
442 list_for_each_entry(c, &r->chunk_list.head, link) { in dma_find_chunk_lpar()
484 struct ps3_dma_region *r = c->region; in dma_ioc0_free_chunk() local
487 for (iopage = 0; iopage < (c->len >> r->page_size); iopage++) { in dma_ioc0_free_chunk()
488 offset = (1 << r->page_size) * iopage; in dma_ioc0_free_chunk()
493 r->ioid, in dma_ioc0_free_chunk()
498 r->ioid); in dma_ioc0_free_chunk()
521 static int dma_sb_map_pages(struct ps3_dma_region *r, unsigned long phys_addr, in dma_sb_map_pages() argument
534 c->region = r; in dma_sb_map_pages()
536 c->bus_addr = dma_sb_lpar_to_bus(r, c->lpar_addr); in dma_sb_map_pages()
549 list_add(&c->link, &r->chunk_list.head); in dma_sb_map_pages()
562 static int dma_ioc0_map_pages(struct ps3_dma_region *r, unsigned long phys_addr, in dma_ioc0_map_pages() argument
580 c->region = r; in dma_ioc0_map_pages()
584 if (list_empty(&r->chunk_list.head)) { in dma_ioc0_map_pages()
586 c->bus_addr = r->bus_addr; in dma_ioc0_map_pages()
589 last = list_entry(r->chunk_list.head.next, in dma_ioc0_map_pages()
599 pages = len >> r->page_size; in dma_ioc0_map_pages()
601 r->page_size, r->len, pages, iopte_flag); in dma_ioc0_map_pages()
603 offset = (1 << r->page_size) * iopage; in dma_ioc0_map_pages()
607 r->ioid, in dma_ioc0_map_pages()
616 r->ioid); in dma_ioc0_map_pages()
620 list_add(&c->link, &r->chunk_list.head); in dma_ioc0_map_pages()
631 r->ioid, in dma_ioc0_map_pages()
648 static int dma_sb_region_create(struct ps3_dma_region *r) in dma_sb_region_create() argument
655 BUG_ON(!r); in dma_sb_region_create()
657 if (!r->dev->bus_id) { in dma_sb_region_create()
659 r->dev->bus_id, r->dev->dev_id); in dma_sb_region_create()
664 __LINE__, r->len, r->page_size, r->offset); in dma_sb_region_create()
666 BUG_ON(!r->len); in dma_sb_region_create()
667 BUG_ON(!r->page_size); in dma_sb_region_create()
668 BUG_ON(!r->region_ops); in dma_sb_region_create()
670 INIT_LIST_HEAD(&r->chunk_list.head); in dma_sb_region_create()
671 spin_lock_init(&r->chunk_list.lock); in dma_sb_region_create()
673 result = lv1_allocate_device_dma_region(r->dev->bus_id, r->dev->dev_id, in dma_sb_region_create()
674 roundup_pow_of_two(r->len), r->page_size, r->region_type, in dma_sb_region_create()
676 r->bus_addr = bus_addr; in dma_sb_region_create()
681 r->len = r->bus_addr = 0; in dma_sb_region_create()
687 static int dma_ioc0_region_create(struct ps3_dma_region *r) in dma_ioc0_region_create() argument
692 INIT_LIST_HEAD(&r->chunk_list.head); in dma_ioc0_region_create()
693 spin_lock_init(&r->chunk_list.lock); in dma_ioc0_region_create()
696 r->len, in dma_ioc0_region_create()
697 r->page_size, in dma_ioc0_region_create()
699 r->bus_addr = bus_addr; in dma_ioc0_region_create()
703 r->len = r->bus_addr = 0; in dma_ioc0_region_create()
706 r->len, r->page_size, r->bus_addr); in dma_ioc0_region_create()
718 static int dma_sb_region_free(struct ps3_dma_region *r) in dma_sb_region_free() argument
724 BUG_ON(!r); in dma_sb_region_free()
726 if (!r->dev->bus_id) { in dma_sb_region_free()
728 r->dev->bus_id, r->dev->dev_id); in dma_sb_region_free()
732 list_for_each_entry_safe(c, tmp, &r->chunk_list.head, link) { in dma_sb_region_free()
737 result = lv1_free_device_dma_region(r->dev->bus_id, r->dev->dev_id, in dma_sb_region_free()
738 r->bus_addr); in dma_sb_region_free()
744 r->bus_addr = 0; in dma_sb_region_free()
749 static int dma_ioc0_region_free(struct ps3_dma_region *r) in dma_ioc0_region_free() argument
755 list_for_each_entry_safe(c, n, &r->chunk_list.head, link) { in dma_ioc0_region_free()
760 result = lv1_release_io_segment(0, r->bus_addr); in dma_ioc0_region_free()
766 r->bus_addr = 0; in dma_ioc0_region_free()
783 static int dma_sb_map_area(struct ps3_dma_region *r, unsigned long virt_addr, in dma_sb_map_area() argument
792 unsigned long aligned_phys = _ALIGN_DOWN(phys_addr, 1 << r->page_size); in dma_sb_map_area()
794 1 << r->page_size); in dma_sb_map_area()
795 *bus_addr = dma_sb_lpar_to_bus(r, ps3_mm_phys_to_lpar(phys_addr)); in dma_sb_map_area()
811 spin_lock_irqsave(&r->chunk_list.lock, flags); in dma_sb_map_area()
812 c = dma_find_chunk(r, *bus_addr, len); in dma_sb_map_area()
818 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_sb_map_area()
822 result = dma_sb_map_pages(r, aligned_phys, aligned_len, &c, iopte_flag); in dma_sb_map_area()
828 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_sb_map_area()
834 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_sb_map_area()
838 static int dma_ioc0_map_area(struct ps3_dma_region *r, unsigned long virt_addr, in dma_ioc0_map_area() argument
847 unsigned long aligned_phys = _ALIGN_DOWN(phys_addr, 1 << r->page_size); in dma_ioc0_map_area()
849 1 << r->page_size); in dma_ioc0_map_area()
856 spin_lock_irqsave(&r->chunk_list.lock, flags); in dma_ioc0_map_area()
857 c = dma_find_chunk_lpar(r, ps3_mm_phys_to_lpar(phys_addr), len); in dma_ioc0_map_area()
864 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_ioc0_map_area()
868 result = dma_ioc0_map_pages(r, aligned_phys, aligned_len, &c, in dma_ioc0_map_area()
875 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_ioc0_map_area()
883 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_ioc0_map_area()
896 static int dma_sb_unmap_area(struct ps3_dma_region *r, dma_addr_t bus_addr, in dma_sb_unmap_area() argument
902 spin_lock_irqsave(&r->chunk_list.lock, flags); in dma_sb_unmap_area()
903 c = dma_find_chunk(r, bus_addr, len); in dma_sb_unmap_area()
907 1 << r->page_size); in dma_sb_unmap_area()
909 - aligned_bus, 1 << r->page_size); in dma_sb_unmap_area()
928 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_sb_unmap_area()
932 static int dma_ioc0_unmap_area(struct ps3_dma_region *r, in dma_ioc0_unmap_area() argument
939 spin_lock_irqsave(&r->chunk_list.lock, flags); in dma_ioc0_unmap_area()
940 c = dma_find_chunk(r, bus_addr, len); in dma_ioc0_unmap_area()
944 1 << r->page_size); in dma_ioc0_unmap_area()
947 1 << r->page_size); in dma_ioc0_unmap_area()
966 spin_unlock_irqrestore(&r->chunk_list.lock, flags); in dma_ioc0_unmap_area()
979 static int dma_sb_region_create_linear(struct ps3_dma_region *r) in dma_sb_region_create_linear() argument
985 if (r->len > 16*1024*1024) { /* FIXME: need proper fix */ in dma_sb_region_create_linear()
987 if (r->page_size != PS3_DMA_16M) { in dma_sb_region_create_linear()
990 r->page_size = PS3_DMA_16M; in dma_sb_region_create_linear()
991 r->len = _ALIGN_UP(r->len, 1 << r->page_size); in dma_sb_region_create_linear()
995 result = dma_sb_region_create(r); in dma_sb_region_create_linear()
998 if (r->offset < map.rm.size) { in dma_sb_region_create_linear()
1000 virt_addr = map.rm.base + r->offset; in dma_sb_region_create_linear()
1001 len = map.rm.size - r->offset; in dma_sb_region_create_linear()
1002 if (len > r->len) in dma_sb_region_create_linear()
1003 len = r->len; in dma_sb_region_create_linear()
1004 result = dma_sb_map_area(r, virt_addr, len, &tmp, in dma_sb_region_create_linear()
1010 if (r->offset + r->len > map.rm.size) { in dma_sb_region_create_linear()
1013 len = r->len; in dma_sb_region_create_linear()
1014 if (r->offset >= map.rm.size) in dma_sb_region_create_linear()
1015 virt_addr += r->offset - map.rm.size; in dma_sb_region_create_linear()
1017 len -= map.rm.size - r->offset; in dma_sb_region_create_linear()
1018 result = dma_sb_map_area(r, virt_addr, len, &tmp, in dma_sb_region_create_linear()
1034 static int dma_sb_region_free_linear(struct ps3_dma_region *r) in dma_sb_region_free_linear() argument
1040 if (r->offset < map.rm.size) { in dma_sb_region_free_linear()
1042 lpar_addr = map.rm.base + r->offset; in dma_sb_region_free_linear()
1043 len = map.rm.size - r->offset; in dma_sb_region_free_linear()
1044 if (len > r->len) in dma_sb_region_free_linear()
1045 len = r->len; in dma_sb_region_free_linear()
1046 bus_addr = dma_sb_lpar_to_bus(r, lpar_addr); in dma_sb_region_free_linear()
1047 result = dma_sb_unmap_area(r, bus_addr, len); in dma_sb_region_free_linear()
1051 if (r->offset + r->len > map.rm.size) { in dma_sb_region_free_linear()
1054 len = r->len; in dma_sb_region_free_linear()
1055 if (r->offset >= map.rm.size) in dma_sb_region_free_linear()
1056 lpar_addr += r->offset - map.rm.size; in dma_sb_region_free_linear()
1058 len -= map.rm.size - r->offset; in dma_sb_region_free_linear()
1059 bus_addr = dma_sb_lpar_to_bus(r, lpar_addr); in dma_sb_region_free_linear()
1060 result = dma_sb_unmap_area(r, bus_addr, len); in dma_sb_region_free_linear()
1064 result = dma_sb_region_free(r); in dma_sb_region_free_linear()
1082 static int dma_sb_map_area_linear(struct ps3_dma_region *r, in dma_sb_map_area_linear() argument
1088 *bus_addr = dma_sb_lpar_to_bus(r, ps3_mm_phys_to_lpar(phys_addr)); in dma_sb_map_area_linear()
1101 static int dma_sb_unmap_area_linear(struct ps3_dma_region *r, in dma_sb_unmap_area_linear() argument
1129 struct ps3_dma_region *r, enum ps3_dma_page_size page_size, in ps3_dma_region_init() argument
1136 r->dev = dev; in ps3_dma_region_init()
1137 r->page_size = page_size; in ps3_dma_region_init()
1138 r->region_type = region_type; in ps3_dma_region_init()
1139 r->offset = lpar_addr; in ps3_dma_region_init()
1140 if (r->offset >= map.rm.size) in ps3_dma_region_init()
1141 r->offset -= map.r1.offset; in ps3_dma_region_init()
1142 r->len = len ? len : _ALIGN_UP(map.total, 1 << r->page_size); in ps3_dma_region_init()
1146 r->region_ops = (USE_DYNAMIC_DMA) in ps3_dma_region_init()
1151 r->region_ops = &ps3_dma_ioc0_region_ops; in ps3_dma_region_init()
1161 int ps3_dma_region_create(struct ps3_dma_region *r) in ps3_dma_region_create() argument
1163 BUG_ON(!r); in ps3_dma_region_create()
1164 BUG_ON(!r->region_ops); in ps3_dma_region_create()
1165 BUG_ON(!r->region_ops->create); in ps3_dma_region_create()
1166 return r->region_ops->create(r); in ps3_dma_region_create()
1170 int ps3_dma_region_free(struct ps3_dma_region *r) in ps3_dma_region_free() argument
1172 BUG_ON(!r); in ps3_dma_region_free()
1173 BUG_ON(!r->region_ops); in ps3_dma_region_free()
1174 BUG_ON(!r->region_ops->free); in ps3_dma_region_free()
1175 return r->region_ops->free(r); in ps3_dma_region_free()
1179 int ps3_dma_map(struct ps3_dma_region *r, unsigned long virt_addr, in ps3_dma_map() argument
1183 return r->region_ops->map(r, virt_addr, len, bus_addr, iopte_flag); in ps3_dma_map()
1186 int ps3_dma_unmap(struct ps3_dma_region *r, dma_addr_t bus_addr, in ps3_dma_unmap() argument
1189 return r->region_ops->unmap(r, bus_addr, len); in ps3_dma_unmap()