/linux-4.1.27/sound/synth/ |
H A D | Makefile | 6 snd-util-mem-objs := util_mem.o 9 obj-$(CONFIG_SND_EMU10K1) += snd-util-mem.o 10 obj-$(CONFIG_SND_TRIDENT) += snd-util-mem.o 11 obj-$(CONFIG_SND_SBAWE_SEQ) += snd-util-mem.o
|
/linux-4.1.27/tools/testing/selftests/memory-hotplug/ |
H A D | Makefile | 5 TEST_PROGS := mem-on-off-test.sh 6 override RUN_TESTS := ./mem-on-off-test.sh -r 2 || echo "selftests: memory-hotplug [FAIL]" 10 @/bin/bash ./mem-on-off-test.sh || echo "memory-hotplug selftests: [FAIL]"
|
/linux-4.1.27/drivers/infiniband/hw/mlx5/ |
H A D | Makefile | 3 mlx5_ib-y := main.o cq.o doorbell.o qp.o mem.o srq.o mr.o ah.o mad.o
|
/linux-4.1.27/drivers/base/ |
H A D | dma-coherent.c | 22 struct dma_coherent_mem **mem) dma_init_coherent_memory() 52 *mem = dma_mem; dma_init_coherent_memory() 66 static void dma_release_coherent_memory(struct dma_coherent_mem *mem) dma_release_coherent_memory() argument 68 if (!mem) dma_release_coherent_memory() 70 iounmap(mem->virt_base); dma_release_coherent_memory() 71 kfree(mem->bitmap); dma_release_coherent_memory() 72 kfree(mem); dma_release_coherent_memory() 76 struct dma_coherent_mem *mem) dma_assign_coherent_memory() 81 dev->dma_mem = mem; dma_assign_coherent_memory() 90 struct dma_coherent_mem *mem; dma_declare_coherent_memory() local 94 &mem); dma_declare_coherent_memory() 98 if (dma_assign_coherent_memory(dev, mem) == 0) dma_declare_coherent_memory() 101 dma_release_coherent_memory(mem); dma_declare_coherent_memory() 108 struct dma_coherent_mem *mem = dev->dma_mem; dma_release_declared_memory() local 110 if (!mem) dma_release_declared_memory() 112 dma_release_coherent_memory(mem); dma_release_declared_memory() 120 struct dma_coherent_mem *mem = dev->dma_mem; dma_mark_declared_memory_occupied() local 126 if (!mem) dma_mark_declared_memory_occupied() 129 spin_lock_irqsave(&mem->spinlock, flags); dma_mark_declared_memory_occupied() 130 pos = (device_addr - mem->device_base) >> PAGE_SHIFT; dma_mark_declared_memory_occupied() 131 err = bitmap_allocate_region(mem->bitmap, pos, get_order(size)); dma_mark_declared_memory_occupied() 132 spin_unlock_irqrestore(&mem->spinlock, flags); dma_mark_declared_memory_occupied() 136 return mem->virt_base + (pos << PAGE_SHIFT); dma_mark_declared_memory_occupied() 158 struct dma_coherent_mem *mem; dma_alloc_from_coherent() local 165 mem = dev->dma_mem; dma_alloc_from_coherent() 166 if (!mem) dma_alloc_from_coherent() 170 spin_lock_irqsave(&mem->spinlock, flags); dma_alloc_from_coherent() 172 if (unlikely(size > (mem->size << PAGE_SHIFT))) dma_alloc_from_coherent() 175 pageno = bitmap_find_free_region(mem->bitmap, mem->size, order); dma_alloc_from_coherent() 182 *dma_handle = mem->device_base + (pageno << PAGE_SHIFT); dma_alloc_from_coherent() 183 *ret = mem->virt_base + (pageno << PAGE_SHIFT); dma_alloc_from_coherent() 185 spin_unlock_irqrestore(&mem->spinlock, flags); dma_alloc_from_coherent() 190 spin_unlock_irqrestore(&mem->spinlock, flags); dma_alloc_from_coherent() 196 return mem->flags & DMA_MEMORY_EXCLUSIVE; dma_alloc_from_coherent() 215 struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL; dma_release_from_coherent() local 217 if (mem && vaddr >= mem->virt_base && vaddr < dma_release_from_coherent() 218 (mem->virt_base + (mem->size << PAGE_SHIFT))) { dma_release_from_coherent() 219 int page = (vaddr - mem->virt_base) >> PAGE_SHIFT; dma_release_from_coherent() 222 spin_lock_irqsave(&mem->spinlock, flags); dma_release_from_coherent() 223 bitmap_release_region(mem->bitmap, page, order); dma_release_from_coherent() 224 spin_unlock_irqrestore(&mem->spinlock, flags); dma_release_from_coherent() 249 struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL; dma_mmap_from_coherent() local 251 if (mem && vaddr >= mem->virt_base && vaddr + size <= dma_mmap_from_coherent() 252 (mem->virt_base + (mem->size << PAGE_SHIFT))) { dma_mmap_from_coherent() 254 int start = (vaddr - mem->virt_base) >> PAGE_SHIFT; dma_mmap_from_coherent() 260 unsigned long pfn = mem->pfn_base + start + off; dma_mmap_from_coherent() 281 struct dma_coherent_mem *mem = rmem->priv; rmem_dma_device_init() local 283 if (!mem && rmem_dma_device_init() 286 &mem) != DMA_MEMORY_MAP) { rmem_dma_device_init() 291 rmem->priv = mem; rmem_dma_device_init() 292 dma_assign_coherent_memory(dev, mem); rmem_dma_device_init() 20 dma_init_coherent_memory(phys_addr_t phys_addr, dma_addr_t device_addr, size_t size, int flags, struct dma_coherent_mem **mem) dma_init_coherent_memory() argument 75 dma_assign_coherent_memory(struct device *dev, struct dma_coherent_mem *mem) dma_assign_coherent_memory() argument
|
H A D | memory.c | 81 struct memory_block *mem = to_memory_block(dev); memory_block_release() local 83 kfree(mem); memory_block_release() 114 struct memory_block *mem = to_memory_block(dev); show_mem_start_phys_index() local 117 phys_index = mem->start_section_nr / sections_per_block; show_mem_start_phys_index() 129 struct memory_block *mem = to_memory_block(dev); show_mem_removable() local 132 if (!present_section_nr(mem->start_section_nr + i)) show_mem_removable() 134 pfn = section_nr_to_pfn(mem->start_section_nr + i); show_mem_removable() 147 struct memory_block *mem = to_memory_block(dev); show_mem_state() local 154 switch (mem->state) { show_mem_state() 166 mem->state); show_mem_state() 254 static int memory_block_change_state(struct memory_block *mem, memory_block_change_state() argument 259 if (mem->state != from_state_req) memory_block_change_state() 263 mem->state = MEM_GOING_OFFLINE; memory_block_change_state() 265 ret = memory_block_action(mem->start_section_nr, to_state, memory_block_change_state() 266 mem->online_type); memory_block_change_state() 268 mem->state = ret ? from_state_req : to_state; memory_block_change_state() 276 struct memory_block *mem = to_memory_block(dev); memory_subsys_online() local 279 if (mem->state == MEM_ONLINE) memory_subsys_online() 287 if (mem->online_type < 0) memory_subsys_online() 288 mem->online_type = MMOP_ONLINE_KEEP; memory_subsys_online() 291 ret = memory_block_change_state(mem, MEM_ONLINE, MEM_OFFLINE); memory_subsys_online() 294 mem->online_type = -1; memory_subsys_online() 301 struct memory_block *mem = to_memory_block(dev); memory_subsys_offline() local 303 if (mem->state == MEM_OFFLINE) memory_subsys_offline() 306 return memory_block_change_state(mem, MEM_OFFLINE, MEM_ONLINE); memory_subsys_offline() 313 struct memory_block *mem = to_memory_block(dev); store_mem_state() local 346 mem->online_type = online_type; store_mem_state() 347 ret = device_online(&mem->dev); store_mem_state() 350 ret = device_offline(&mem->dev); store_mem_state() 377 struct memory_block *mem = to_memory_block(dev); show_phys_device() local 378 return sprintf(buf, "%d\n", mem->phys_device); show_phys_device() 385 struct memory_block *mem = to_memory_block(dev); show_valid_zones() local 391 start_pfn = section_nr_to_pfn(mem->start_section_nr); show_valid_zones() 402 /*The mem block is the last memoryblock of this zone.*/ show_valid_zones() 409 /*The mem block is the first memoryblock of ZONE_MOVABLE.*/ show_valid_zones() 603 struct memory_block *mem; init_memory_block() local 608 mem = kzalloc(sizeof(*mem), GFP_KERNEL); init_memory_block() 609 if (!mem) init_memory_block() 613 mem->start_section_nr = init_memory_block() 615 mem->end_section_nr = mem->start_section_nr + sections_per_block - 1; init_memory_block() 616 mem->state = state; init_memory_block() 617 mem->section_count++; init_memory_block() 618 start_pfn = section_nr_to_pfn(mem->start_section_nr); init_memory_block() 619 mem->phys_device = arch_get_memory_phys_device(start_pfn); init_memory_block() 621 ret = register_memory(mem); init_memory_block() 623 *memory = mem; init_memory_block() 629 struct memory_block *mem; add_memory_block() local 644 ret = init_memory_block(&mem, __nr_to_section(section_nr), MEM_ONLINE); add_memory_block() 647 mem->section_count = section_count; add_memory_block() 659 struct memory_block *mem; register_new_memory() local 663 mem = find_memory_block(section); register_new_memory() 664 if (mem) { register_new_memory() 665 mem->section_count++; register_new_memory() 666 put_device(&mem->dev); register_new_memory() 668 ret = init_memory_block(&mem, section, MEM_OFFLINE); register_new_memory() 673 if (mem->section_count == sections_per_block) register_new_memory() 674 ret = register_mem_sect_under_node(mem, nid); register_new_memory() 694 struct memory_block *mem; remove_memory_block() local 697 mem = find_memory_block(section); remove_memory_block() 698 unregister_mem_sect_under_nodes(mem, __section_nr(section)); remove_memory_block() 700 mem->section_count--; remove_memory_block() 701 if (mem->section_count == 0) remove_memory_block() 702 unregister_memory(mem); remove_memory_block() 704 put_device(&mem->dev); remove_memory_block() 720 bool is_memblock_offlined(struct memory_block *mem) is_memblock_offlined() argument 722 return mem->state == MEM_OFFLINE; is_memblock_offlined()
|
/linux-4.1.27/include/linux/ |
H A D | kbuild.h | 9 #define OFFSET(sym, str, mem) \ 10 DEFINE(sym, offsetof(struct str, mem))
|
H A D | trace_seq.h | 86 extern void trace_seq_putmem(struct trace_seq *s, const void *mem, unsigned int len); 87 extern void trace_seq_putmem_hex(struct trace_seq *s, const void *mem, 125 trace_seq_putmem(struct trace_seq *s, const void *mem, unsigned int len) trace_seq_putmem() argument 128 static inline void trace_seq_putmem_hex(struct trace_seq *s, const void *mem, trace_seq_putmem_hex() argument
|
/linux-4.1.27/drivers/media/v4l2-core/ |
H A D | videobuf-dma-contig.c | 41 struct videobuf_dma_contig_memory *mem, __videobuf_dc_alloc() 44 mem->size = size; __videobuf_dc_alloc() 45 mem->vaddr = dma_alloc_coherent(dev, mem->size, __videobuf_dc_alloc() 46 &mem->dma_handle, flags); __videobuf_dc_alloc() 48 if (!mem->vaddr) { __videobuf_dc_alloc() 49 dev_err(dev, "memory alloc size %ld failed\n", mem->size); __videobuf_dc_alloc() 53 dev_dbg(dev, "dma mapped data is at %p (%ld)\n", mem->vaddr, mem->size); __videobuf_dc_alloc() 59 struct videobuf_dma_contig_memory *mem) __videobuf_dc_free() 61 dma_free_coherent(dev, mem->size, mem->vaddr, mem->dma_handle); __videobuf_dc_free() 63 mem->vaddr = NULL; __videobuf_dc_free() 87 struct videobuf_dma_contig_memory *mem; videobuf_vm_close() local 103 mem = q->bufs[i]->priv; videobuf_vm_close() 104 if (mem) { videobuf_vm_close() 111 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); videobuf_vm_close() 117 i, mem->vaddr); videobuf_vm_close() 119 __videobuf_dc_free(q->dev, mem); videobuf_vm_close() 120 mem->vaddr = NULL; videobuf_vm_close() 140 * @mem: per-buffer private videobuf-dma-contig data 144 static void videobuf_dma_contig_user_put(struct videobuf_dma_contig_memory *mem) videobuf_dma_contig_user_put() argument 146 mem->dma_handle = 0; videobuf_dma_contig_user_put() 147 mem->size = 0; videobuf_dma_contig_user_put() 152 * @mem: per-buffer private videobuf-dma-contig data 160 static int videobuf_dma_contig_user_get(struct videobuf_dma_contig_memory *mem, videobuf_dma_contig_user_get() argument 171 mem->size = PAGE_ALIGN(vb->size + offset); videobuf_dma_contig_user_get() 180 if ((vb->baddr + mem->size) > vma->vm_end) videobuf_dma_contig_user_get() 187 while (pages_done < (mem->size >> PAGE_SHIFT)) { videobuf_dma_contig_user_get() 193 mem->dma_handle = (this_pfn << PAGE_SHIFT) + offset; videobuf_dma_contig_user_get() 213 struct videobuf_dma_contig_memory *mem; __videobuf_alloc() local 216 vb = kzalloc(size + sizeof(*mem), GFP_KERNEL); __videobuf_alloc() 219 mem = vb->priv; __videobuf_alloc() 220 mem->magic = MAGIC_DC_MEM; __videobuf_alloc() 228 struct videobuf_dma_contig_memory *mem = buf->priv; __videobuf_to_vaddr() local 230 BUG_ON(!mem); __videobuf_to_vaddr() 231 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); __videobuf_to_vaddr() 233 return mem->vaddr; __videobuf_to_vaddr() 240 struct videobuf_dma_contig_memory *mem = vb->priv; __videobuf_iolock() local 242 BUG_ON(!mem); __videobuf_iolock() 243 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); __videobuf_iolock() 250 if (!mem->vaddr) { __videobuf_iolock() 260 return videobuf_dma_contig_user_get(mem, vb); __videobuf_iolock() 263 if (__videobuf_dc_alloc(q->dev, mem, PAGE_ALIGN(vb->size), __videobuf_iolock() 280 struct videobuf_dma_contig_memory *mem; __videobuf_mmap_mapper() local 297 mem = buf->priv; __videobuf_mmap_mapper() 298 BUG_ON(!mem); __videobuf_mmap_mapper() 299 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); __videobuf_mmap_mapper() 301 if (__videobuf_dc_alloc(q->dev, mem, PAGE_ALIGN(buf->bsize), __videobuf_mmap_mapper() 317 retval = vm_iomap_memory(vma, mem->dma_handle, size); __videobuf_mmap_mapper() 321 dma_free_coherent(q->dev, mem->size, __videobuf_mmap_mapper() 322 mem->vaddr, mem->dma_handle); __videobuf_mmap_mapper() 368 struct videobuf_dma_contig_memory *mem = buf->priv; videobuf_to_dma_contig() local 370 BUG_ON(!mem); videobuf_to_dma_contig() 371 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); videobuf_to_dma_contig() 373 return mem->dma_handle; videobuf_to_dma_contig() 380 struct videobuf_dma_contig_memory *mem = buf->priv; videobuf_dma_contig_free() local 391 if (!mem) videobuf_dma_contig_free() 394 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); videobuf_dma_contig_free() 398 videobuf_dma_contig_user_put(mem); videobuf_dma_contig_free() 403 if (mem->vaddr) { videobuf_dma_contig_free() 404 __videobuf_dc_free(q->dev, mem); videobuf_dma_contig_free() 405 mem->vaddr = NULL; videobuf_dma_contig_free() 40 __videobuf_dc_alloc(struct device *dev, struct videobuf_dma_contig_memory *mem, unsigned long size, gfp_t flags) __videobuf_dc_alloc() argument 58 __videobuf_dc_free(struct device *dev, struct videobuf_dma_contig_memory *mem) __videobuf_dc_free() argument
|
H A D | videobuf-vmalloc.c | 75 struct videobuf_vmalloc_memory *mem; videobuf_vm_close() local 91 mem = q->bufs[i]->priv; videobuf_vm_close() 92 if (mem) { videobuf_vm_close() 99 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); videobuf_vm_close() 105 __func__, i, mem->vaddr); videobuf_vm_close() 107 vfree(mem->vaddr); videobuf_vm_close() 108 mem->vaddr = NULL; videobuf_vm_close() 140 struct videobuf_vmalloc_memory *mem; __videobuf_alloc_vb() local 143 vb = kzalloc(size + sizeof(*mem), GFP_KERNEL); __videobuf_alloc_vb() 147 mem = vb->priv = ((char *)vb) + size; __videobuf_alloc_vb() 148 mem->magic = MAGIC_VMAL_MEM; __videobuf_alloc_vb() 152 mem, (long)sizeof(*mem)); __videobuf_alloc_vb() 161 struct videobuf_vmalloc_memory *mem = vb->priv; __videobuf_iolock() local 164 BUG_ON(!mem); __videobuf_iolock() 166 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); __videobuf_iolock() 173 if (!mem->vaddr) { __videobuf_iolock() 192 mem->vaddr = vmalloc_user(pages); __videobuf_iolock() 193 if (!mem->vaddr) { __videobuf_iolock() 198 mem->vaddr, pages); __videobuf_iolock() 209 The code below won't work, since mem->vma = NULL __videobuf_iolock() 212 rc = remap_vmalloc_range(mem->vma, (void *)vb->baddr, 0); __videobuf_iolock() 236 struct videobuf_vmalloc_memory *mem; __videobuf_mmap_mapper() local 252 mem = buf->priv; __videobuf_mmap_mapper() 253 BUG_ON(!mem); __videobuf_mmap_mapper() 254 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); __videobuf_mmap_mapper() 257 mem->vaddr = vmalloc_user(pages); __videobuf_mmap_mapper() 258 if (!mem->vaddr) { __videobuf_mmap_mapper() 262 dprintk(1, "vmalloc is at addr %p (%d pages)\n", mem->vaddr, pages); __videobuf_mmap_mapper() 265 retval = remap_vmalloc_range(vma, mem->vaddr, 0); __videobuf_mmap_mapper() 268 vfree(mem->vaddr); __videobuf_mmap_mapper() 286 mem = NULL; __videobuf_mmap_mapper() 317 struct videobuf_vmalloc_memory *mem = buf->priv; videobuf_to_vmalloc() local 318 BUG_ON(!mem); videobuf_to_vmalloc() 319 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); videobuf_to_vmalloc() 321 return mem->vaddr; videobuf_to_vmalloc() 327 struct videobuf_vmalloc_memory *mem = buf->priv; videobuf_vmalloc_free() local 338 if (!mem) videobuf_vmalloc_free() 341 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); videobuf_vmalloc_free() 343 vfree(mem->vaddr); videobuf_vmalloc_free() 344 mem->vaddr = NULL; videobuf_vmalloc_free()
|
H A D | videobuf-dma-sg.c | 139 struct videobuf_dma_sg_memory *mem = buf->priv; videobuf_to_dma() local 140 BUG_ON(!mem); videobuf_to_dma() 142 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); videobuf_to_dma() 144 return &mem->dma; videobuf_to_dma() 398 struct videobuf_dma_sg_memory *mem; videobuf_vm_close() local 411 mem = q->bufs[i]->priv; videobuf_vm_close() 412 if (!mem) videobuf_vm_close() 415 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); videobuf_vm_close() 470 struct videobuf_dma_sg_memory *mem; __videobuf_alloc_vb() local 473 vb = kzalloc(size + sizeof(*mem), GFP_KERNEL); __videobuf_alloc_vb() 477 mem = vb->priv = ((char *)vb) + size; __videobuf_alloc_vb() 478 mem->magic = MAGIC_SG_MEM; __videobuf_alloc_vb() 480 videobuf_dma_init(&mem->dma); __videobuf_alloc_vb() 484 mem, (long)sizeof(*mem)); __videobuf_alloc_vb() 491 struct videobuf_dma_sg_memory *mem = buf->priv; __videobuf_to_vaddr() local 492 BUG_ON(!mem); __videobuf_to_vaddr() 494 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); __videobuf_to_vaddr() 496 return mem->dma.vaddr; __videobuf_to_vaddr() 505 struct videobuf_dma_sg_memory *mem = vb->priv; __videobuf_iolock() local 506 BUG_ON(!mem); __videobuf_iolock() 508 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); __videobuf_iolock() 510 if (!mem->dma.dev) __videobuf_iolock() 511 mem->dma.dev = q->dev; __videobuf_iolock() 513 WARN_ON(mem->dma.dev != q->dev); __videobuf_iolock() 521 err = videobuf_dma_init_kernel(&mem->dma, __videobuf_iolock() 528 err = videobuf_dma_init_user(&mem->dma, __videobuf_iolock() 539 err = videobuf_dma_init_user_locked(&mem->dma, __videobuf_iolock() 557 err = videobuf_dma_init_overlay(&mem->dma, DMA_FROM_DEVICE, __videobuf_iolock() 565 err = videobuf_dma_map(q->dev, &mem->dma); __videobuf_iolock() 575 struct videobuf_dma_sg_memory *mem = buf->priv; __videobuf_sync() local 576 BUG_ON(!mem || !mem->dma.sglen); __videobuf_sync() 578 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); __videobuf_sync() 579 MAGIC_CHECK(mem->dma.magic, MAGIC_DMABUF); __videobuf_sync() 581 dma_sync_sg_for_cpu(q->dev, mem->dma.sglist, __videobuf_sync() 582 mem->dma.sglen, mem->dma.direction); __videobuf_sync() 591 struct videobuf_dma_sg_memory *mem = buf->priv; __videobuf_mmap_mapper() local 598 BUG_ON(!mem); __videobuf_mmap_mapper() 599 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); __videobuf_mmap_mapper()
|
/linux-4.1.27/drivers/infiniband/hw/cxgb4/ |
H A D | Makefile | 5 iw_cxgb4-y := device.o cm.o provider.o mem.o cq.o qp.o resource.o ev.o id_table.o
|
/linux-4.1.27/drivers/uio/ |
H A D | uio_mf624.c | 48 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; mf624_disable_interrupt() 76 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; mf624_enable_interrupt() 103 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; mf624_irq_handler() 150 info->mem[0].name = "PCI chipset, interrupts, status " mf624_pci_probe() 152 info->mem[0].addr = pci_resource_start(dev, 0); mf624_pci_probe() 153 if (!info->mem[0].addr) mf624_pci_probe() 155 info->mem[0].size = pci_resource_len(dev, 0); mf624_pci_probe() 156 info->mem[0].memtype = UIO_MEM_PHYS; mf624_pci_probe() 157 info->mem[0].internal_addr = pci_ioremap_bar(dev, 0); mf624_pci_probe() 158 if (!info->mem[0].internal_addr) mf624_pci_probe() 162 info->mem[1].name = "ADC, DAC, DIO"; mf624_pci_probe() 163 info->mem[1].addr = pci_resource_start(dev, 2); mf624_pci_probe() 164 if (!info->mem[1].addr) mf624_pci_probe() 166 info->mem[1].size = pci_resource_len(dev, 2); mf624_pci_probe() 167 info->mem[1].memtype = UIO_MEM_PHYS; mf624_pci_probe() 168 info->mem[1].internal_addr = pci_ioremap_bar(dev, 2); mf624_pci_probe() 169 if (!info->mem[1].internal_addr) mf624_pci_probe() 173 info->mem[2].name = "Counter/timer chip"; mf624_pci_probe() 174 info->mem[2].addr = pci_resource_start(dev, 4); mf624_pci_probe() 175 if (!info->mem[2].addr) mf624_pci_probe() 177 info->mem[2].size = pci_resource_len(dev, 4); mf624_pci_probe() 178 info->mem[2].memtype = UIO_MEM_PHYS; mf624_pci_probe() 179 info->mem[2].internal_addr = pci_ioremap_bar(dev, 4); mf624_pci_probe() 180 if (!info->mem[2].internal_addr) mf624_pci_probe() 197 iounmap(info->mem[2].internal_addr); mf624_pci_probe() 199 iounmap(info->mem[1].internal_addr); mf624_pci_probe() 201 iounmap(info->mem[0].internal_addr); mf624_pci_probe() 224 iounmap(info->mem[0].internal_addr); mf624_pci_remove() 225 iounmap(info->mem[1].internal_addr); mf624_pci_remove() 226 iounmap(info->mem[2].internal_addr); mf624_pci_remove()
|
H A D | uio_cif.c | 31 void __iomem *plx_intscr = dev_info->mem[0].internal_addr hilscher_handler() 58 info->mem[0].addr = pci_resource_start(dev, 0); hilscher_pci_probe() 59 if (!info->mem[0].addr) hilscher_pci_probe() 61 info->mem[0].internal_addr = pci_ioremap_bar(dev, 0); hilscher_pci_probe() 62 if (!info->mem[0].internal_addr) hilscher_pci_probe() 65 info->mem[0].size = pci_resource_len(dev, 0); hilscher_pci_probe() 66 info->mem[0].memtype = UIO_MEM_PHYS; hilscher_pci_probe() 67 info->mem[1].addr = pci_resource_start(dev, 2); hilscher_pci_probe() 68 info->mem[1].size = pci_resource_len(dev, 2); hilscher_pci_probe() 69 info->mem[1].memtype = UIO_MEM_PHYS; hilscher_pci_probe() 92 iounmap(info->mem[0].internal_addr); hilscher_pci_probe() 109 iounmap(info->mem[0].internal_addr); hilscher_pci_remove()
|
H A D | uio_netx.c | 35 void __iomem *int_enable_reg = dev_info->mem[0].internal_addr netx_handler() 37 void __iomem *int_status_reg = dev_info->mem[0].internal_addr netx_handler() 82 info->mem[0].addr = pci_resource_start(dev, bar); netx_pci_probe() 83 if (!info->mem[0].addr) netx_pci_probe() 85 info->mem[0].internal_addr = ioremap(pci_resource_start(dev, bar), netx_pci_probe() 88 if (!info->mem[0].internal_addr) netx_pci_probe() 91 info->mem[0].size = pci_resource_len(dev, bar); netx_pci_probe() 92 info->mem[0].memtype = UIO_MEM_PHYS; netx_pci_probe() 99 iowrite32(0, info->mem[0].internal_addr + DPM_HOST_INT_EN0); netx_pci_probe() 111 iounmap(info->mem[0].internal_addr); netx_pci_probe() 126 iowrite32(0, info->mem[0].internal_addr + DPM_HOST_INT_EN0); netx_pci_remove() 130 iounmap(info->mem[0].internal_addr); netx_pci_remove()
|
H A D | uio_sercos3.c | 54 void __iomem *ier0 = info->mem[3].internal_addr + IER0_OFFSET; sercos3_disable_interrupts() 67 void __iomem *ier0 = info->mem[3].internal_addr + IER0_OFFSET; sercos3_enable_interrupts() 77 void __iomem *isr0 = info->mem[3].internal_addr + ISR0_OFFSET; sercos3_handler() 78 void __iomem *ier0 = info->mem[3].internal_addr + IER0_OFFSET; sercos3_handler() 107 info->mem[n].addr = pci_resource_start(dev, pci_bar); sercos3_setup_iomem() 108 if (!info->mem[n].addr) sercos3_setup_iomem() 110 info->mem[n].internal_addr = ioremap(pci_resource_start(dev, pci_bar), sercos3_setup_iomem() 112 if (!info->mem[n].internal_addr) sercos3_setup_iomem() 114 info->mem[n].size = pci_resource_len(dev, pci_bar); sercos3_setup_iomem() 115 info->mem[n].memtype = UIO_MEM_PHYS; sercos3_setup_iomem() 170 if (info->mem[i].internal_addr) sercos3_pci_probe() 171 iounmap(info->mem[i].internal_addr); sercos3_pci_probe() 192 if (info->mem[i].internal_addr) sercos3_pci_remove() 193 iounmap(info->mem[i].internal_addr); sercos3_pci_remove()
|
H A D | uio.c | 45 struct uio_mem *mem; member in struct:uio_map 49 static ssize_t map_name_show(struct uio_mem *mem, char *buf) map_name_show() argument 51 if (unlikely(!mem->name)) map_name_show() 52 mem->name = ""; map_name_show() 54 return sprintf(buf, "%s\n", mem->name); map_name_show() 57 static ssize_t map_addr_show(struct uio_mem *mem, char *buf) map_addr_show() argument 59 return sprintf(buf, "%pa\n", &mem->addr); map_addr_show() 62 static ssize_t map_size_show(struct uio_mem *mem, char *buf) map_size_show() argument 64 return sprintf(buf, "%pa\n", &mem->size); map_size_show() 67 static ssize_t map_offset_show(struct uio_mem *mem, char *buf) map_offset_show() argument 69 return sprintf(buf, "0x%llx\n", (unsigned long long)mem->addr & ~PAGE_MASK); map_offset_show() 105 struct uio_mem *mem = map->mem; map_type_show() local 113 return entry->show(mem, buf); map_type_show() 261 struct uio_mem *mem; uio_dev_add_attributes() local 267 mem = &idev->info->mem[mi]; uio_dev_add_attributes() 268 if (mem->size == 0) uio_dev_add_attributes() 281 map->mem = mem; uio_dev_add_attributes() 282 mem->map = map; uio_dev_add_attributes() 332 mem = &idev->info->mem[mi]; uio_dev_add_attributes() 333 map = mem->map; uio_dev_add_attributes() 344 struct uio_mem *mem; uio_dev_del_attributes() local 348 mem = &idev->info->mem[i]; uio_dev_del_attributes() 349 if (mem->size == 0) uio_dev_del_attributes() 351 kobject_put(&mem->map->kobj); uio_dev_del_attributes() 584 if (idev->info->mem[vma->vm_pgoff].size == 0) uio_find_mem_index() 604 * to use mem[N]. uio_vma_fault() 608 addr = (void *)(unsigned long)idev->info->mem[mi].addr + offset; uio_vma_fault() 609 if (idev->info->mem[mi].memtype == UIO_MEM_LOGICAL) uio_vma_fault() 639 struct uio_mem *mem; uio_mmap_physical() local 642 mem = idev->info->mem + mi; uio_mmap_physical() 644 if (mem->addr & ~PAGE_MASK) uio_mmap_physical() 646 if (vma->vm_end - vma->vm_start > mem->size) uio_mmap_physical() 663 mem->addr >> PAGE_SHIFT, uio_mmap_physical() 686 actual_pages = ((idev->info->mem[mi].addr & ~PAGE_MASK) uio_mmap() 687 + idev->info->mem[mi].size + PAGE_SIZE -1) >> PAGE_SHIFT; uio_mmap() 696 switch (idev->info->mem[mi].memtype) { uio_mmap()
|
H A D | uio_pruss.c | 188 p->mem[0].addr = regs_prussio->start; pruss_probe() 189 p->mem[0].size = resource_size(regs_prussio); pruss_probe() 190 p->mem[0].memtype = UIO_MEM_PHYS; pruss_probe() 192 p->mem[1].addr = gdev->sram_paddr; pruss_probe() 193 p->mem[1].size = sram_pool_sz; pruss_probe() 194 p->mem[1].memtype = UIO_MEM_PHYS; pruss_probe() 196 p->mem[2].addr = gdev->ddr_paddr; pruss_probe() 197 p->mem[2].size = extram_pool_sz; pruss_probe() 198 p->mem[2].memtype = UIO_MEM_PHYS; pruss_probe()
|
H A D | uio_dmem_genirq.c | 53 uiomem = &priv->uioinfo->mem[priv->dmem_region_start]; uio_dmem_genirq_open() 56 while (!priv->refcnt && uiomem < &priv->uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_open() 86 uiomem = &priv->uioinfo->mem[priv->dmem_region_start]; uio_dmem_genirq_release() 91 while (!priv->refcnt && uiomem < &priv->uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_release() 211 uiomem = &uioinfo->mem[0]; uio_dmem_genirq_probe() 219 if (uiomem >= &uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_probe() 236 if (uiomem >= &uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_probe() 248 while (uiomem < &uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_probe()
|
H A D | uio_fsl_elbc_gpcm.c | 152 void __iomem *reg_int_en = info->mem[0].internal_addr + netx5152_irq_handler() 155 void __iomem *reg_int_stat = info->mem[0].internal_addr + netx5152_irq_handler() 184 iowrite32(0, info->mem[0].internal_addr + win0_offset + netx5152_init() 194 iowrite32(0, info->mem[0].internal_addr + DPM_HOST_WIN0_OFFSET + netx5152_shutdown() 376 info->mem[0].internal_addr = ioremap(res.start, resource_size(&res)); uio_fsl_elbc_gpcm_probe() 377 if (!info->mem[0].internal_addr) { uio_fsl_elbc_gpcm_probe() 385 info->mem[0].name = kstrdup(node->name, GFP_KERNEL); uio_fsl_elbc_gpcm_probe() 386 info->mem[0].addr = res.start; uio_fsl_elbc_gpcm_probe() 387 info->mem[0].size = resource_size(&res); uio_fsl_elbc_gpcm_probe() 388 info->mem[0].memtype = UIO_MEM_PHYS; uio_fsl_elbc_gpcm_probe() 438 iounmap(info->mem[0].internal_addr); uio_fsl_elbc_gpcm_probe() 440 kfree(info->mem[0].name); uio_fsl_elbc_gpcm_probe() 459 iounmap(info->mem[0].internal_addr); uio_fsl_elbc_gpcm_remove() 460 kfree(info->mem[0].name); uio_fsl_elbc_gpcm_remove()
|
/linux-4.1.27/mm/ |
H A D | cma_debug.c | 71 static void cma_add_to_cma_mem_list(struct cma *cma, struct cma_mem *mem) cma_add_to_cma_mem_list() argument 74 hlist_add_head(&mem->node, &cma->mem_head); cma_add_to_cma_mem_list() 80 struct cma_mem *mem = NULL; cma_get_entry_from_list() local 84 mem = hlist_entry(cma->mem_head.first, struct cma_mem, node); cma_get_entry_from_list() 85 hlist_del_init(&mem->node); cma_get_entry_from_list() 89 return mem; cma_get_entry_from_list() 94 struct cma_mem *mem = NULL; cma_free_mem() local 97 mem = cma_get_entry_from_list(cma); cma_free_mem() 98 if (mem == NULL) cma_free_mem() 101 if (mem->n <= count) { cma_free_mem() 102 cma_release(cma, mem->p, mem->n); cma_free_mem() 103 count -= mem->n; cma_free_mem() 104 kfree(mem); cma_free_mem() 106 cma_release(cma, mem->p, count); cma_free_mem() 107 mem->p += count; cma_free_mem() 108 mem->n -= count; cma_free_mem() 110 cma_add_to_cma_mem_list(cma, mem); cma_free_mem() 113 cma_add_to_cma_mem_list(cma, mem); cma_free_mem() 133 struct cma_mem *mem; cma_alloc_mem() local 136 mem = kzalloc(sizeof(*mem), GFP_KERNEL); cma_alloc_mem() 137 if (!mem) cma_alloc_mem() 142 kfree(mem); cma_alloc_mem() 146 mem->p = p; cma_alloc_mem() 147 mem->n = count; cma_alloc_mem() 149 cma_add_to_cma_mem_list(cma, mem); cma_alloc_mem()
|
H A D | debug-pagealloc.c | 80 static void check_poison_mem(unsigned char *mem, size_t bytes) check_poison_mem() argument 86 start = memchr_inv(mem, PAGE_POISON, bytes); check_poison_mem() 90 for (end = mem + bytes - 1; end > start; end--) { check_poison_mem()
|
/linux-4.1.27/arch/um/os-Linux/skas/ |
H A D | Makefile | 6 obj-y := mem.o process.o
|
/linux-4.1.27/arch/alpha/include/uapi/asm/ |
H A D | compiler.h | 93 #define __kernel_ldbu(mem) (mem) 94 #define __kernel_ldwu(mem) (mem) 95 #define __kernel_stb(val,mem) ((mem) = (val)) 96 #define __kernel_stw(val,mem) ((mem) = (val)) 98 #define __kernel_ldbu(mem) \ 101 ldbu %0,%1" : "=r"(__kir) : "m"(mem)); \ 103 #define __kernel_ldwu(mem) \ 106 ldwu %0,%1" : "=r"(__kir) : "m"(mem)); \ 108 #define __kernel_stb(val,mem) \ 110 stb %1,%0" : "=m"(mem) : "r"(val)) 111 #define __kernel_stw(val,mem) \ 113 stw %1,%0" : "=m"(mem) : "r"(val))
|
/linux-4.1.27/tools/perf/bench/ |
H A D | mem-memcpy-arch.h | 7 #include "mem-memcpy-x86-64-asm-def.h"
|
H A D | mem-memset-arch.h | 7 #include "mem-memset-x86-64-asm-def.h"
|
H A D | numa.c | 202 "perf bench numa mem [<options>]", 660 printf("# NOTE: %d tasks mem-bound, %d tasks unbound\n", t, g->p.nr_tasks - t); parse_setup_node_list() 1066 printf("# thread %2d / %2d global mem: %p, process mem: %p, thread mem: %p\n", worker_thread() 1237 printf(" # process %2d global mem: %p, process mem: %p\n", worker_process() 1274 printf(" # %5dx %5ldMB global shared mem operations\n", print_summary() 1276 printf(" # %5dx %5ldMB process shared mem operations\n", print_summary() 1278 printf(" # %5dx %5ldMB thread local mem operations\n", print_summary() 1645 { "RAM-bw-local,", "mem", "-p", "1", "-t", "1", "-P", "1024", 1648 "mem", "-p", "1", "-t", "1", "-P", "1024", 1650 { "RAM-bw-remote,", "mem", "-p", "1", "-t", "1", "-P", "1024", 1654 { "RAM-bw-local-2x,", "mem", "-p", "2", "-t", "1", "-P", "1024", 1656 { "RAM-bw-remote-2x,", "mem", "-p", "2", "-t", "1", "-P", "1024", 1660 { "RAM-bw-cross,", "mem", "-p", "2", "-t", "1", "-P", "1024", 1664 { " 1x3-convergence,", "mem", "-p", "1", "-t", "3", "-P", "512", OPT_CONV }, 1665 { " 1x4-convergence,", "mem", "-p", "1", "-t", "4", "-P", "512", OPT_CONV }, 1666 { " 1x6-convergence,", "mem", "-p", "1", "-t", "6", "-P", "1020", OPT_CONV }, 1667 { " 2x3-convergence,", "mem", "-p", "3", "-t", "3", "-P", "1020", OPT_CONV }, 1668 { " 3x3-convergence,", "mem", "-p", "3", "-t", "3", "-P", "1020", OPT_CONV }, 1669 { " 4x4-convergence,", "mem", "-p", "4", "-t", "4", "-P", "512", OPT_CONV }, 1671 "mem", "-p", "4", "-t", "4", "-P", "512", OPT_CONV_NOTHP }, 1672 { " 4x6-convergence,", "mem", "-p", "4", "-t", "6", "-P", "1020", OPT_CONV }, 1673 { " 4x8-convergence,", "mem", "-p", "4", "-t", "8", "-P", "512", OPT_CONV }, 1674 { " 8x4-convergence,", "mem", "-p", "8", "-t", "4", "-P", "512", OPT_CONV }, 1676 "mem", "-p", "8", "-t", "4", "-P", "512", OPT_CONV_NOTHP }, 1677 { " 3x1-convergence,", "mem", "-p", "3", "-t", "1", "-P", "512", OPT_CONV }, 1678 { " 4x1-convergence,", "mem", "-p", "4", "-t", "1", "-P", "512", OPT_CONV }, 1679 { " 8x1-convergence,", "mem", "-p", "8", "-t", "1", "-P", "512", OPT_CONV }, 1680 { "16x1-convergence,", "mem", "-p", "16", "-t", "1", "-P", "256", OPT_CONV }, 1681 { "32x1-convergence,", "mem", "-p", "32", "-t", "1", "-P", "128", OPT_CONV }, 1684 { " 2x1-bw-process,", "mem", "-p", "2", "-t", "1", "-P", "1024", OPT_BW }, 1685 { " 3x1-bw-process,", "mem", "-p", "3", "-t", "1", "-P", "1024", OPT_BW }, 1686 { " 4x1-bw-process,", "mem", "-p", "4", "-t", "1", "-P", "1024", OPT_BW }, 1687 { " 8x1-bw-process,", "mem", "-p", "8", "-t", "1", "-P", " 512", OPT_BW }, 1689 "mem", "-p", "8", "-t", "1", "-P", " 512", OPT_BW_NOTHP }, 1690 { "16x1-bw-process,", "mem", "-p", "16", "-t", "1", "-P", "256", OPT_BW }, 1692 { " 4x1-bw-thread,", "mem", "-p", "1", "-t", "4", "-T", "256", OPT_BW }, 1693 { " 8x1-bw-thread,", "mem", "-p", "1", "-t", "8", "-T", "256", OPT_BW }, 1694 { "16x1-bw-thread,", "mem", "-p", "1", "-t", "16", "-T", "128", OPT_BW }, 1695 { "32x1-bw-thread,", "mem", "-p", "1", "-t", "32", "-T", "64", OPT_BW }, 1697 { " 2x3-bw-thread,", "mem", "-p", "2", "-t", "3", "-P", "512", OPT_BW }, 1698 { " 4x4-bw-thread,", "mem", "-p", "4", "-t", "4", "-P", "512", OPT_BW }, 1699 { " 4x6-bw-thread,", "mem", "-p", "4", "-t", "6", "-P", "512", OPT_BW }, 1700 { " 4x8-bw-thread,", "mem", "-p", "4", "-t", "8", "-P", "512", OPT_BW }, 1702 "mem", "-p", "4", "-t", "8", "-P", "512", OPT_BW_NOTHP }, 1703 { " 3x3-bw-thread,", "mem", "-p", "3", "-t", "3", "-P", "512", OPT_BW }, 1704 { " 5x5-bw-thread,", "mem", "-p", "5", "-t", "5", "-P", "512", OPT_BW }, 1706 { "2x16-bw-thread,", "mem", "-p", "2", "-t", "16", "-P", "512", OPT_BW }, 1707 { "1x32-bw-thread,", "mem", "-p", "1", "-t", "32", "-P", "2048", OPT_BW }, 1709 { "numa02-bw,", "mem", "-p", "1", "-t", "32", "-T", "32", OPT_BW }, 1710 { "numa02-bw-NOTHP,", "mem", "-p", "1", "-t", "32", "-T", "32", OPT_BW_NOTHP }, 1711 { "numa01-bw-thread,", "mem", "-p", "2", "-t", "16", "-T", "192", OPT_BW }, 1713 "mem", "-p", "2", "-t", "16", "-T", "192", OPT_BW_NOTHP },
|
/linux-4.1.27/tools/perf/util/include/asm/ |
H A D | dwarf2.h | 5 /* dwarf2.h ... dummy header file for including arch/x86/lib/mem{cpy,set}_64.S */
|
/linux-4.1.27/tools/perf/ |
H A D | builtin-mem.c | 24 static int __cmd_record(int argc, const char **argv, struct perf_mem *mem) __cmd_record() argument 37 if (mem->operation & MEM_OPERATION_LOAD) __cmd_record() 42 if (mem->operation & MEM_OPERATION_LOAD) { __cmd_record() 44 rec_argv[i++] = "cpu/mem-loads/pp"; __cmd_record() 47 if (mem->operation & MEM_OPERATION_STORE) { __cmd_record() 49 rec_argv[i++] = "cpu/mem-stores/pp"; __cmd_record() 66 struct perf_mem *mem = container_of(tool, struct perf_mem, tool); dump_raw_samples() local 76 if (al.filtered || (mem->hide_unresolved && al.sym == NULL)) dump_raw_samples() 119 static int report_raw_events(struct perf_mem *mem) report_raw_events() argument 124 .force = mem->force, report_raw_events() 129 &mem->tool); report_raw_events() 134 if (mem->cpu_list) { report_raw_events() 135 ret = perf_session__cpu_bitmap(session, mem->cpu_list, report_raw_events() 136 mem->cpu_bitmap); report_raw_events() 157 static int report_events(int argc, const char **argv, struct perf_mem *mem) report_events() argument 162 if (mem->dump_raw) report_events() 163 return report_raw_events(mem); report_events() 171 rep_argv[i++] = "--mem-mode"; report_events() 178 if (!(mem->operation & MEM_OPERATION_LOAD)) report_events() 179 rep_argv[i++] = "--sort=mem,sym,dso,symbol_daddr," report_events() 262 struct perf_mem mem = { cmd_mem() local 280 OPT_CALLBACK('t', "type", &mem.operation, cmd_mem() 283 OPT_BOOLEAN('D', "dump-raw-samples", &mem.dump_raw, cmd_mem() 285 OPT_BOOLEAN('U', "hide-unresolved", &mem.hide_unresolved, cmd_mem() 289 OPT_STRING('C', "cpu", &mem.cpu_list, "cpu", cmd_mem() 295 OPT_BOOLEAN('f', "force", &mem.force, "don't complain, do it"), cmd_mem() 308 if (!argc || !(strncmp(argv[0], "rec", 3) || mem.operation)) cmd_mem() 311 if (!mem.input_name || !strlen(mem.input_name)) { cmd_mem() 313 mem.input_name = "-"; cmd_mem() 315 mem.input_name = "perf.data"; cmd_mem() 319 return __cmd_record(argc, argv, &mem); cmd_mem() 321 return report_events(argc, argv, &mem); cmd_mem()
|
/linux-4.1.27/drivers/isdn/pcbit/ |
H A D | module.c | 25 static int mem[MAX_PCBIT_CARDS]; variable 28 module_param_array(mem, int, NULL, 0); 44 if (mem[0] || irq[0]) pcbit_init() 46 for (board = 0; board < MAX_PCBIT_CARDS && mem[board] && irq[board]; board++) pcbit_init() 48 if (!mem[board]) pcbit_init() 49 mem[board] = 0xD0000; pcbit_init() 53 if (pcbit_init_dev(board, mem[board], irq[board]) == 0) pcbit_init() 108 mem[i] = ints[j]; pcbit_setup()
|
/linux-4.1.27/arch/x86/um/shared/sysdep/ |
H A D | kernel-offsets.h | 12 #define OFFSET(sym, str, mem) \ 13 DEFINE(sym, offsetof(struct str, mem));
|
/linux-4.1.27/drivers/firmware/efi/ |
H A D | cper.c | 210 static int cper_mem_err_location(struct cper_mem_err_compact *mem, char *msg) cper_mem_err_location() argument 219 if (mem->validation_bits & CPER_MEM_VALID_NODE) cper_mem_err_location() 220 n += scnprintf(msg + n, len - n, "node: %d ", mem->node); cper_mem_err_location() 221 if (mem->validation_bits & CPER_MEM_VALID_CARD) cper_mem_err_location() 222 n += scnprintf(msg + n, len - n, "card: %d ", mem->card); cper_mem_err_location() 223 if (mem->validation_bits & CPER_MEM_VALID_MODULE) cper_mem_err_location() 224 n += scnprintf(msg + n, len - n, "module: %d ", mem->module); cper_mem_err_location() 225 if (mem->validation_bits & CPER_MEM_VALID_RANK_NUMBER) cper_mem_err_location() 226 n += scnprintf(msg + n, len - n, "rank: %d ", mem->rank); cper_mem_err_location() 227 if (mem->validation_bits & CPER_MEM_VALID_BANK) cper_mem_err_location() 228 n += scnprintf(msg + n, len - n, "bank: %d ", mem->bank); cper_mem_err_location() 229 if (mem->validation_bits & CPER_MEM_VALID_DEVICE) cper_mem_err_location() 230 n += scnprintf(msg + n, len - n, "device: %d ", mem->device); cper_mem_err_location() 231 if (mem->validation_bits & CPER_MEM_VALID_ROW) cper_mem_err_location() 232 n += scnprintf(msg + n, len - n, "row: %d ", mem->row); cper_mem_err_location() 233 if (mem->validation_bits & CPER_MEM_VALID_COLUMN) cper_mem_err_location() 234 n += scnprintf(msg + n, len - n, "column: %d ", mem->column); cper_mem_err_location() 235 if (mem->validation_bits & CPER_MEM_VALID_BIT_POSITION) cper_mem_err_location() 237 mem->bit_pos); cper_mem_err_location() 238 if (mem->validation_bits & CPER_MEM_VALID_REQUESTOR_ID) cper_mem_err_location() 240 mem->requestor_id); cper_mem_err_location() 241 if (mem->validation_bits & CPER_MEM_VALID_RESPONDER_ID) cper_mem_err_location() 243 mem->responder_id); cper_mem_err_location() 244 if (mem->validation_bits & CPER_MEM_VALID_TARGET_ID) cper_mem_err_location() 246 mem->target_id); cper_mem_err_location() 252 static int cper_dimm_err_location(struct cper_mem_err_compact *mem, char *msg) cper_dimm_err_location() argument 257 if (!msg || !(mem->validation_bits & CPER_MEM_VALID_MODULE_HANDLE)) cper_dimm_err_location() 262 dmi_memdev_name(mem->mem_dev_handle, &bank, &device); cper_dimm_err_location() 268 mem->mem_dev_handle); cper_dimm_err_location() 274 void cper_mem_err_pack(const struct cper_sec_mem_err *mem, cper_mem_err_pack() argument 277 cmem->validation_bits = mem->validation_bits; cper_mem_err_pack() 278 cmem->node = mem->node; cper_mem_err_pack() 279 cmem->card = mem->card; cper_mem_err_pack() 280 cmem->module = mem->module; cper_mem_err_pack() 281 cmem->bank = mem->bank; cper_mem_err_pack() 282 cmem->device = mem->device; cper_mem_err_pack() 283 cmem->row = mem->row; cper_mem_err_pack() 284 cmem->column = mem->column; cper_mem_err_pack() 285 cmem->bit_pos = mem->bit_pos; cper_mem_err_pack() 286 cmem->requestor_id = mem->requestor_id; cper_mem_err_pack() 287 cmem->responder_id = mem->responder_id; cper_mem_err_pack() 288 cmem->target_id = mem->target_id; cper_mem_err_pack() 289 cmem->rank = mem->rank; cper_mem_err_pack() 290 cmem->mem_array_handle = mem->mem_array_handle; cper_mem_err_pack() 291 cmem->mem_dev_handle = mem->mem_dev_handle; cper_mem_err_pack() 308 static void cper_print_mem(const char *pfx, const struct cper_sec_mem_err *mem, cper_print_mem() argument 315 (mem->validation_bits & ~(CPER_MEM_VALID_RANK_NUMBER - 1))) { cper_print_mem() 319 if (mem->validation_bits & CPER_MEM_VALID_ERROR_STATUS) cper_print_mem() 320 printk("%s""error_status: 0x%016llx\n", pfx, mem->error_status); cper_print_mem() 321 if (mem->validation_bits & CPER_MEM_VALID_PA) cper_print_mem() 323 pfx, mem->physical_addr); cper_print_mem() 324 if (mem->validation_bits & CPER_MEM_VALID_PA_MASK) cper_print_mem() 326 pfx, mem->physical_addr_mask); cper_print_mem() 327 cper_mem_err_pack(mem, &cmem); cper_print_mem() 330 if (mem->validation_bits & CPER_MEM_VALID_ERROR_TYPE) { cper_print_mem() 331 u8 etype = mem->error_type; cper_print_mem()
|
/linux-4.1.27/drivers/gpu/drm/ttm/ |
H A D | ttm_agp_backend.c | 47 struct agp_memory *mem; member in struct:ttm_agp_backend 55 struct agp_memory *mem; ttm_agp_bind() local 59 mem = agp_allocate_memory(agp_be->bridge, ttm->num_pages, AGP_USER_MEMORY); ttm_agp_bind() 60 if (unlikely(mem == NULL)) ttm_agp_bind() 63 mem->page_count = 0; ttm_agp_bind() 70 mem->pages[mem->page_count++] = page; ttm_agp_bind() 72 agp_be->mem = mem; ttm_agp_bind() 74 mem->is_flushed = 1; ttm_agp_bind() 75 mem->type = (cached) ? AGP_USER_CACHED_MEMORY : AGP_USER_MEMORY; ttm_agp_bind() 77 ret = agp_bind_memory(mem, node->start); ttm_agp_bind() 88 if (agp_be->mem) { ttm_agp_unbind() 89 if (agp_be->mem->is_bound) ttm_agp_unbind() 90 return agp_unbind_memory(agp_be->mem); ttm_agp_unbind() 91 agp_free_memory(agp_be->mem); ttm_agp_unbind() 92 agp_be->mem = NULL; ttm_agp_unbind() 101 if (agp_be->mem) ttm_agp_destroy() 124 agp_be->mem = NULL; ttm_agp_tt_create()
|
H A D | ttm_bo_util.c | 44 ttm_bo_mem_put(bo, &bo->mem); ttm_bo_free_old_node() 52 struct ttm_mem_reg *old_mem = &bo->mem; ttm_bo_move_ttm() 120 struct ttm_mem_reg *mem) ttm_mem_io_reserve() 122 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ttm_mem_io_reserve() 128 return bdev->driver->io_mem_reserve(bdev, mem); ttm_mem_io_reserve() 131 mem->bus.io_reserved_count++ == 0) { ttm_mem_io_reserve() 133 ret = bdev->driver->io_mem_reserve(bdev, mem); ttm_mem_io_reserve() 145 struct ttm_mem_reg *mem) ttm_mem_io_free() 147 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ttm_mem_io_free() 153 --mem->bus.io_reserved_count == 0 && ttm_mem_io_free() 155 bdev->driver->io_mem_free(bdev, mem); ttm_mem_io_free() 162 struct ttm_mem_reg *mem = &bo->mem; ttm_mem_io_reserve_vm() local 165 if (!mem->bus.io_reserved_vm) { ttm_mem_io_reserve_vm() 167 &bo->bdev->man[mem->mem_type]; ttm_mem_io_reserve_vm() 169 ret = ttm_mem_io_reserve(bo->bdev, mem); ttm_mem_io_reserve_vm() 172 mem->bus.io_reserved_vm = true; ttm_mem_io_reserve_vm() 182 struct ttm_mem_reg *mem = &bo->mem; ttm_mem_io_free_vm() local 184 if (mem->bus.io_reserved_vm) { ttm_mem_io_free_vm() 185 mem->bus.io_reserved_vm = false; ttm_mem_io_free_vm() 187 ttm_mem_io_free(bo->bdev, mem); ttm_mem_io_free_vm() 191 static int ttm_mem_reg_ioremap(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem, ttm_mem_reg_ioremap() argument 194 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ttm_mem_reg_ioremap() 200 ret = ttm_mem_io_reserve(bdev, mem); ttm_mem_reg_ioremap() 202 if (ret || !mem->bus.is_iomem) ttm_mem_reg_ioremap() 205 if (mem->bus.addr) { ttm_mem_reg_ioremap() 206 addr = mem->bus.addr; ttm_mem_reg_ioremap() 208 if (mem->placement & TTM_PL_FLAG_WC) ttm_mem_reg_ioremap() 209 addr = ioremap_wc(mem->bus.base + mem->bus.offset, mem->bus.size); ttm_mem_reg_ioremap() 211 addr = ioremap_nocache(mem->bus.base + mem->bus.offset, mem->bus.size); ttm_mem_reg_ioremap() 214 ttm_mem_io_free(bdev, mem); ttm_mem_reg_ioremap() 223 static void ttm_mem_reg_iounmap(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem, ttm_mem_reg_iounmap() argument 228 man = &bdev->man[mem->mem_type]; ttm_mem_reg_iounmap() 230 if (virtual && mem->bus.addr == NULL) ttm_mem_reg_iounmap() 233 ttm_mem_io_free(bdev, mem); ttm_mem_reg_iounmap() 330 struct ttm_mem_reg *old_mem = &bo->mem; ttm_bo_move_memcpy() 511 struct ttm_mem_reg *mem = &bo->mem; ttm_bo_ioremap() local 513 if (bo->mem.bus.addr) { ttm_bo_ioremap() 515 map->virtual = (void *)(((u8 *)bo->mem.bus.addr) + offset); ttm_bo_ioremap() 518 if (mem->placement & TTM_PL_FLAG_WC) ttm_bo_ioremap() 519 map->virtual = ioremap_wc(bo->mem.bus.base + bo->mem.bus.offset + offset, ttm_bo_ioremap() 522 map->virtual = ioremap_nocache(bo->mem.bus.base + bo->mem.bus.offset + offset, ttm_bo_ioremap() 533 struct ttm_mem_reg *mem = &bo->mem; pgprot_t prot; ttm_bo_kmap_ttm() local 545 if (num_pages == 1 && (mem->placement & TTM_PL_FLAG_CACHED)) { ttm_bo_kmap_ttm() 559 prot = ttm_io_prot(mem->placement, PAGE_KERNEL); ttm_bo_kmap_ttm() 572 &bo->bdev->man[bo->mem.mem_type]; ttm_bo_kmap() 588 ret = ttm_mem_io_reserve(bo->bdev, &bo->mem); ttm_bo_kmap() 592 if (!bo->mem.bus.is_iomem) { ttm_bo_kmap() 606 &bo->bdev->man[bo->mem.mem_type]; ttm_bo_kunmap() 626 ttm_mem_io_free(map->bo->bdev, &map->bo->mem); ttm_bo_kunmap() 641 struct ttm_mem_reg *old_mem = &bo->mem; ttm_bo_move_accel_cleanup() 119 ttm_mem_io_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ttm_mem_io_reserve() argument 144 ttm_mem_io_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ttm_mem_io_free() argument
|
H A D | ttm_bo.c | 91 bo, bo->mem.num_pages, bo->mem.size >> 10, ttm_bo_mem_space_debug() 92 bo->mem.size >> 20); ttm_bo_mem_space_debug() 146 BUG_ON(bo->mem.mm_node != NULL); ttm_bo_release_list() 171 if (!(bo->mem.placement & TTM_PL_FLAG_NO_EVICT)) { ttm_bo_add_to_lru() 175 man = &bdev->man[bo->mem.mem_type]; ttm_bo_add_to_lru() 277 struct ttm_mem_reg *mem, ttm_bo_handle_move_mem() 282 bool old_is_pci = ttm_mem_reg_is_pci(bdev, &bo->mem); ttm_bo_handle_move_mem() 283 bool new_is_pci = ttm_mem_reg_is_pci(bdev, mem); ttm_bo_handle_move_mem() 284 struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem.mem_type]; ttm_bo_handle_move_mem() 285 struct ttm_mem_type_manager *new_man = &bdev->man[mem->mem_type]; ttm_bo_handle_move_mem() 289 ((mem->placement & bo->mem.placement & TTM_PL_MASK_CACHING) == 0)) { ttm_bo_handle_move_mem() 309 ret = ttm_tt_set_placement_caching(bo->ttm, mem->placement); ttm_bo_handle_move_mem() 313 if (mem->mem_type != TTM_PL_SYSTEM) { ttm_bo_handle_move_mem() 314 ret = ttm_tt_bind(bo->ttm, mem); ttm_bo_handle_move_mem() 319 if (bo->mem.mem_type == TTM_PL_SYSTEM) { ttm_bo_handle_move_mem() 321 bdev->driver->move_notify(bo, mem); ttm_bo_handle_move_mem() 322 bo->mem = *mem; ttm_bo_handle_move_mem() 323 mem->mm_node = NULL; ttm_bo_handle_move_mem() 329 bdev->driver->move_notify(bo, mem); ttm_bo_handle_move_mem() 333 ret = ttm_bo_move_ttm(bo, evict, no_wait_gpu, mem); ttm_bo_handle_move_mem() 336 no_wait_gpu, mem); ttm_bo_handle_move_mem() 338 ret = ttm_bo_move_memcpy(bo, evict, no_wait_gpu, mem); ttm_bo_handle_move_mem() 342 struct ttm_mem_reg tmp_mem = *mem; ttm_bo_handle_move_mem() 343 *mem = bo->mem; ttm_bo_handle_move_mem() 344 bo->mem = tmp_mem; ttm_bo_handle_move_mem() 345 bdev->driver->move_notify(bo, mem); ttm_bo_handle_move_mem() 346 bo->mem = *mem; ttm_bo_handle_move_mem() 347 *mem = tmp_mem; ttm_bo_handle_move_mem() 356 ret = bdev->driver->invalidate_caches(bdev, bo->mem.placement); ttm_bo_handle_move_mem() 363 if (bo->mem.mm_node) { ttm_bo_handle_move_mem() 364 bo->offset = (bo->mem.start << PAGE_SHIFT) + ttm_bo_handle_move_mem() 365 bdev->man[bo->mem.mem_type].gpu_offset; ttm_bo_handle_move_mem() 366 bo->cur_placement = bo->mem.placement; ttm_bo_handle_move_mem() 373 new_man = &bdev->man[bo->mem.mem_type]; ttm_bo_handle_move_mem() 401 ttm_bo_mem_put(bo, &bo->mem); ttm_bo_cleanup_memtype_use() 454 if (bo->mem.placement & TTM_PL_FLAG_NO_EVICT) { ttm_bo_cleanup_refs_or_queue() 455 bo->mem.placement &= ~TTM_PL_FLAG_NO_EVICT; ttm_bo_cleanup_refs_or_queue() 626 struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; ttm_bo_release() 678 evict_mem = bo->mem; ttm_bo_evict() 729 if (place->fpfn >= (bo->mem.start + bo->mem.size) || ttm_mem_evict_first() 730 (place->lpfn && place->lpfn <= bo->mem.start)) { ttm_mem_evict_first() 769 void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) ttm_bo_mem_put() argument 771 struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; ttm_bo_mem_put() 773 if (mem->mm_node) ttm_bo_mem_put() 774 (*man->func->put_node)(man, mem); ttm_bo_mem_put() 785 struct ttm_mem_reg *mem, ttm_bo_mem_force_space() 794 ret = (*man->func->get_node)(man, bo, place, mem); ttm_bo_mem_force_space() 797 if (mem->mm_node) ttm_bo_mem_force_space() 804 if (mem->mm_node == NULL) ttm_bo_mem_force_space() 806 mem->mem_type = mem_type; ttm_bo_mem_force_space() 855 * Creates space for memory region @mem according to its type. 864 struct ttm_mem_reg *mem, ttm_bo_mem_space() 877 mem->mm_node = NULL; ttm_bo_mem_space() 892 cur_flags = ttm_bo_select_caching(man, bo->mem.placement, ttm_bo_mem_space() 906 ret = (*man->func->get_node)(man, bo, place, mem); ttm_bo_mem_space() 910 if (mem->mm_node) ttm_bo_mem_space() 914 if ((type_ok && (mem_type == TTM_PL_SYSTEM)) || mem->mm_node) { ttm_bo_mem_space() 915 mem->mem_type = mem_type; ttm_bo_mem_space() 916 mem->placement = cur_flags; ttm_bo_mem_space() 935 cur_flags = ttm_bo_select_caching(man, bo->mem.placement, ttm_bo_mem_space() 945 mem->mem_type = mem_type; ttm_bo_mem_space() 946 mem->placement = cur_flags; ttm_bo_mem_space() 947 mem->mm_node = NULL; ttm_bo_mem_space() 951 ret = ttm_bo_mem_force_space(bo, mem_type, place, mem, ttm_bo_mem_space() 953 if (ret == 0 && mem->mm_node) { ttm_bo_mem_space() 954 mem->placement = cur_flags; ttm_bo_mem_space() 971 struct ttm_mem_reg mem; ttm_bo_move_buffer() local 983 mem.num_pages = bo->num_pages; ttm_bo_move_buffer() 984 mem.size = mem.num_pages << PAGE_SHIFT; ttm_bo_move_buffer() 985 mem.page_alignment = bo->mem.page_alignment; ttm_bo_move_buffer() 986 mem.bus.io_reserved_vm = false; ttm_bo_move_buffer() 987 mem.bus.io_reserved_count = 0; ttm_bo_move_buffer() 991 ret = ttm_bo_mem_space(bo, placement, &mem, ttm_bo_move_buffer() 995 ret = ttm_bo_handle_move_mem(bo, &mem, false, ttm_bo_move_buffer() 998 if (ret && mem.mm_node) ttm_bo_move_buffer() 999 ttm_bo_mem_put(bo, &mem); ttm_bo_move_buffer() 1004 struct ttm_mem_reg *mem, ttm_bo_mem_compat() 1011 if (mem->mm_node && ttm_bo_mem_compat() 1012 (mem->start < heap->fpfn || ttm_bo_mem_compat() 1013 (heap->lpfn != 0 && (mem->start + mem->num_pages) > heap->lpfn))) ttm_bo_mem_compat() 1017 if ((*new_flags & mem->placement & TTM_PL_MASK_CACHING) && ttm_bo_mem_compat() 1018 (*new_flags & mem->placement & TTM_PL_MASK_MEM)) ttm_bo_mem_compat() 1024 if (mem->mm_node && ttm_bo_mem_compat() 1025 (mem->start < heap->fpfn || ttm_bo_mem_compat() 1026 (heap->lpfn != 0 && (mem->start + mem->num_pages) > heap->lpfn))) ttm_bo_mem_compat() 1030 if ((*new_flags & mem->placement & TTM_PL_MASK_CACHING) && ttm_bo_mem_compat() 1031 (*new_flags & mem->placement & TTM_PL_MASK_MEM)) ttm_bo_mem_compat() 1050 if (!ttm_bo_mem_compat(placement, &bo->mem, &new_flags)) { ttm_bo_validate() 1060 ttm_flag_masked(&bo->mem.placement, new_flags, ttm_bo_validate() 1066 if (bo->mem.mem_type == TTM_PL_SYSTEM && bo->ttm == NULL) { ttm_bo_validate() 1127 bo->mem.size = num_pages << PAGE_SHIFT; ttm_bo_init() 1128 bo->mem.mem_type = TTM_PL_SYSTEM; ttm_bo_init() 1129 bo->mem.num_pages = bo->num_pages; ttm_bo_init() 1130 bo->mem.mm_node = NULL; ttm_bo_init() 1131 bo->mem.page_alignment = page_alignment; ttm_bo_init() 1132 bo->mem.bus.io_reserved_vm = false; ttm_bo_init() 1133 bo->mem.bus.io_reserved_count = 0; ttm_bo_init() 1135 bo->mem.placement = (TTM_PL_FLAG_SYSTEM | TTM_PL_FLAG_CACHED); ttm_bo_init() 1156 bo->mem.num_pages); ttm_bo_init() 1496 bool ttm_mem_reg_is_pci(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ttm_mem_reg_is_pci() argument 1498 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ttm_mem_reg_is_pci() 1501 if (mem->mem_type == TTM_PL_SYSTEM) ttm_mem_reg_is_pci() 1507 if (mem->placement & TTM_PL_FLAG_CACHED) ttm_mem_reg_is_pci() 1524 struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; ttm_bo_unmap_virtual() 1655 if ((bo->mem.placement & swap_placement) != swap_placement) { ttm_bo_swapout() 1658 evict_mem = bo->mem; ttm_bo_swapout() 276 ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem, bool evict, bool interruptible, bool no_wait_gpu) ttm_bo_handle_move_mem() argument 782 ttm_bo_mem_force_space(struct ttm_buffer_object *bo, uint32_t mem_type, const struct ttm_place *place, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) ttm_bo_mem_force_space() argument 862 ttm_bo_mem_space(struct ttm_buffer_object *bo, struct ttm_placement *placement, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) ttm_bo_mem_space() argument 1003 ttm_bo_mem_compat(struct ttm_placement *placement, struct ttm_mem_reg *mem, uint32_t *new_flags) ttm_bo_mem_compat() argument
|
H A D | ttm_bo_manager.c | 53 struct ttm_mem_reg *mem) ttm_bo_man_get_node() 77 ret = drm_mm_insert_node_in_range_generic(mm, node, mem->num_pages, ttm_bo_man_get_node() 78 mem->page_alignment, 0, ttm_bo_man_get_node() 86 mem->mm_node = node; ttm_bo_man_get_node() 87 mem->start = node->start; ttm_bo_man_get_node() 94 struct ttm_mem_reg *mem) ttm_bo_man_put_node() 98 if (mem->mm_node) { ttm_bo_man_put_node() 100 drm_mm_remove_node(mem->mm_node); ttm_bo_man_put_node() 103 kfree(mem->mm_node); ttm_bo_man_put_node() 104 mem->mm_node = NULL; ttm_bo_man_put_node() 50 ttm_bo_man_get_node(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) ttm_bo_man_get_node() argument 93 ttm_bo_man_put_node(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) ttm_bo_man_put_node() argument
|
H A D | ttm_memory.c | 249 uint64_t mem; ttm_mem_init_kernel_zone() local 255 mem = si->totalram - si->totalhigh; ttm_mem_init_kernel_zone() 256 mem *= si->mem_unit; ttm_mem_init_kernel_zone() 259 zone->zone_mem = mem; ttm_mem_init_kernel_zone() 260 zone->max_mem = mem >> 1; ttm_mem_init_kernel_zone() 261 zone->emer_mem = (mem >> 1) + (mem >> 2); ttm_mem_init_kernel_zone() 262 zone->swap_limit = zone->max_mem - (mem >> 3); ttm_mem_init_kernel_zone() 281 uint64_t mem; ttm_mem_init_highmem_zone() local 291 mem = si->totalram; ttm_mem_init_highmem_zone() 292 mem *= si->mem_unit; ttm_mem_init_highmem_zone() 295 zone->zone_mem = mem; ttm_mem_init_highmem_zone() 296 zone->max_mem = mem >> 1; ttm_mem_init_highmem_zone() 297 zone->emer_mem = (mem >> 1) + (mem >> 2); ttm_mem_init_highmem_zone() 298 zone->swap_limit = zone->max_mem - (mem >> 3); ttm_mem_init_highmem_zone() 317 uint64_t mem; ttm_mem_init_dma32_zone() local 323 mem = si->totalram; ttm_mem_init_dma32_zone() 324 mem *= si->mem_unit; ttm_mem_init_dma32_zone() 330 if (mem <= ((uint64_t) 1ULL << 32)) { ttm_mem_init_dma32_zone() 341 mem = ((uint64_t) 1ULL << 32); ttm_mem_init_dma32_zone() 343 zone->zone_mem = mem; ttm_mem_init_dma32_zone() 344 zone->max_mem = mem >> 1; ttm_mem_init_dma32_zone() 345 zone->emer_mem = (mem >> 1) + (mem >> 2); ttm_mem_init_dma32_zone() 346 zone->swap_limit = zone->max_mem - (mem >> 3); ttm_mem_init_dma32_zone()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/subdev/bar/ |
H A D | gf100.c | 32 struct nvkm_gpuobj *mem; member in struct:gf100_bar_priv_vm 44 gf100_bar_kmap(struct nvkm_bar *bar, struct nvkm_mem *mem, u32 flags, gf100_bar_kmap() argument 50 ret = nvkm_vm_get(priv->bar[0].vm, mem->size << 12, 12, flags, vma); gf100_bar_kmap() 54 nvkm_vm_map(vma, mem); gf100_bar_kmap() 59 gf100_bar_umap(struct nvkm_bar *bar, struct nvkm_mem *mem, u32 flags, gf100_bar_umap() argument 65 ret = nvkm_vm_get(priv->bar[1].vm, mem->size << 12, gf100_bar_umap() 66 mem->page_shift, flags, vma); gf100_bar_umap() 70 nvkm_vm_map(vma, mem); gf100_bar_umap() 91 &bar_vm->mem); gf100_bar_ctor_vm() 126 nv_wo32(bar_vm->mem, 0x0200, lower_32_bits(bar_vm->pgd->addr)); gf100_bar_ctor_vm() 127 nv_wo32(bar_vm->mem, 0x0204, upper_32_bits(bar_vm->pgd->addr)); gf100_bar_ctor_vm() 128 nv_wo32(bar_vm->mem, 0x0208, lower_32_bits(bar_len - 1)); gf100_bar_ctor_vm() 129 nv_wo32(bar_vm->mem, 0x020c, upper_32_bits(bar_len - 1)); gf100_bar_ctor_vm() 178 nvkm_gpuobj_ref(NULL, &priv->bar[1].mem); gf100_bar_dtor() 185 nvkm_gpuobj_ref(NULL, &priv->bar[0].mem); gf100_bar_dtor() 203 nv_wr32(priv, 0x001704, 0x80000000 | priv->bar[1].mem->addr >> 12); gf100_bar_init() 204 if (priv->bar[0].mem) gf100_bar_init() 206 0xc0000000 | priv->bar[0].mem->addr >> 12); gf100_bar_init()
|
H A D | nv50.c | 35 struct nvkm_gpuobj *mem; member in struct:nv50_bar_priv 45 nv50_bar_kmap(struct nvkm_bar *bar, struct nvkm_mem *mem, u32 flags, nv50_bar_kmap() argument 51 ret = nvkm_vm_get(priv->bar3_vm, mem->size << 12, 12, flags, vma); nv50_bar_kmap() 55 nvkm_vm_map(vma, mem); nv50_bar_kmap() 60 nv50_bar_umap(struct nvkm_bar *bar, struct nvkm_mem *mem, u32 flags, nv50_bar_umap() argument 66 ret = nvkm_vm_get(priv->bar1_vm, mem->size << 12, 12, flags, vma); nv50_bar_umap() 70 nvkm_vm_map(vma, mem); nv50_bar_umap() 123 NVOBJ_FLAG_HEAP, &priv->mem); nv50_bar_ctor() 124 heap = nv_object(priv->mem); nv50_bar_ctor() 224 nvkm_gpuobj_ref(NULL, &priv->mem); nv50_bar_dtor() 246 nv_wr32(priv, 0x001704, 0x00000000 | priv->mem->addr >> 12); nv50_bar_init() 247 nv_wr32(priv, 0x001704, 0x40000000 | priv->mem->addr >> 12); nv50_bar_init()
|
H A D | base.c | 43 struct nvkm_mem *mem = data; nvkm_barobj_ctor() local 52 ret = bar->kmap(bar, mem, NV_MEM_ACCESS_RW, &barobj->vma); nvkm_barobj_ctor() 57 (u32)barobj->vma.offset, mem->size << 12); nvkm_barobj_ctor() 107 struct nvkm_mem *mem, struct nvkm_object **pobject) nvkm_bar_alloc() 111 &nvkm_barobj_oclass, mem, 0, &gpuobj); nvkm_bar_alloc() 106 nvkm_bar_alloc(struct nvkm_bar *bar, struct nvkm_object *parent, struct nvkm_mem *mem, struct nvkm_object **pobject) nvkm_bar_alloc() argument
|
/linux-4.1.27/arch/ia64/kernel/ |
H A D | minstate.h | 99 .mem.offset 0,0; st8.spill [r16]=r8,16; \ 100 .mem.offset 8,0; st8.spill [r17]=r9,16; \ 102 .mem.offset 0,0; st8.spill [r16]=r10,24; \ 103 .mem.offset 8,0; st8.spill [r17]=r11,24; \ 129 .mem.offset 0,0; st8.spill [r16]=r20,16; /* save original r1 */ \ 130 .mem.offset 8,0; st8.spill [r17]=r12,16; \ 133 .mem.offset 0,0; st8.spill [r16]=r13,16; \ 134 .mem.offset 8,0; st8.spill [r17]=r21,16; /* save ar.fpsr */ \ 137 .mem.offset 0,0; st8.spill [r16]=r15,16; \ 138 .mem.offset 8,0; st8.spill [r17]=r14,16; \ 140 .mem.offset 0,0; st8.spill [r16]=r2,16; \ 141 .mem.offset 8,0; st8.spill [r17]=r3,16; \ 167 .mem.offset 0,0; st8.spill [r2]=r16,16; \ 168 .mem.offset 8,0; st8.spill [r3]=r17,16; \ 170 .mem.offset 0,0; st8.spill [r2]=r18,16; \ 171 .mem.offset 8,0; st8.spill [r3]=r19,16; \ 173 .mem.offset 0,0; st8.spill [r2]=r20,16; \ 174 .mem.offset 8,0; st8.spill [r3]=r21,16; \ 177 .mem.offset 0,0; st8.spill [r2]=r22,16; \ 178 .mem.offset 8,0; st8.spill [r3]=r23,16; \ 181 .mem.offset 0,0; st8.spill [r2]=r24,16; \ 182 .mem.offset 8,0; st8.spill [r3]=r25,16; \ 184 .mem.offset 0,0; st8.spill [r2]=r26,16; \ 185 .mem.offset 8,0; st8.spill [r3]=r27,16; \ 187 .mem.offset 0,0; st8.spill [r2]=r28,16; \ 188 .mem.offset 8,0; st8.spill [r3]=r29,16; \ 190 .mem.offset 0,0; st8.spill [r2]=r30,16; \ 191 .mem.offset 8,0; st8.spill [r3]=r31,32; \
|
/linux-4.1.27/drivers/char/hw_random/ |
H A D | intel-rng.c | 155 static inline u8 hwstatus_get(void __iomem *mem) hwstatus_get() argument 157 return readb(mem + INTEL_RNG_HW_STATUS); hwstatus_get() 160 static inline u8 hwstatus_set(void __iomem *mem, hwstatus_set() argument 163 writeb(hw_status, mem + INTEL_RNG_HW_STATUS); hwstatus_set() 164 return hwstatus_get(mem); hwstatus_set() 169 void __iomem *mem = (void __iomem *)rng->priv; intel_rng_data_present() local 173 data = !!(readb(mem + INTEL_RNG_STATUS) & intel_rng_data_present() 184 void __iomem *mem = (void __iomem *)rng->priv; intel_rng_data_read() local 186 *data = readb(mem + INTEL_RNG_DATA); intel_rng_data_read() 193 void __iomem *mem = (void __iomem *)rng->priv; intel_rng_init() local 197 hw_status = hwstatus_get(mem); intel_rng_init() 200 hw_status = hwstatus_set(mem, hw_status | INTEL_RNG_ENABLED); intel_rng_init() 212 void __iomem *mem = (void __iomem *)rng->priv; intel_rng_cleanup() local 215 hw_status = hwstatus_get(mem); intel_rng_cleanup() 217 hwstatus_set(mem, hw_status & ~INTEL_RNG_ENABLED); intel_rng_cleanup() 233 void __iomem *mem; member in struct:intel_rng_hw 258 writeb(INTEL_FWH_RESET_CMD, intel_rng_hw->mem); intel_rng_hw_init() 259 writeb(INTEL_FWH_READ_ID_CMD, intel_rng_hw->mem); intel_rng_hw_init() 260 mfc = readb(intel_rng_hw->mem + INTEL_FWH_MANUFACTURER_CODE_ADDRESS); intel_rng_hw_init() 261 dvc = readb(intel_rng_hw->mem + INTEL_FWH_DEVICE_CODE_ADDRESS); intel_rng_hw_init() 262 writeb(INTEL_FWH_RESET_CMD, intel_rng_hw->mem); intel_rng_hw_init() 320 intel_rng_hw->mem = ioremap_nocache(INTEL_FWH_ADDR, INTEL_FWH_ADDR_LEN); intel_init_hw_struct() 321 if (intel_rng_hw->mem == NULL) intel_init_hw_struct() 333 void __iomem *mem = mem; mod_init() local 374 iounmap(intel_rng_hw->mem); mod_init() 381 mem = ioremap(INTEL_RNG_ADDR, INTEL_RNG_ADDR_LEN); mod_init() 382 if (!mem) mod_init() 384 intel_rng.priv = (unsigned long)mem; mod_init() 388 hw_status = hwstatus_get(mem); mod_init() 390 iounmap(mem); mod_init() 399 iounmap(mem); mod_init() 408 void __iomem *mem = (void __iomem *)intel_rng.priv; mod_exit() local 411 iounmap(mem); mod_exit()
|
H A D | mxc-rnga.c | 65 void __iomem *mem; member in struct:mxc_rng 76 int level = (__raw_readl(mxc_rng->mem + RNGA_STATUS) & mxc_rnga_data_present() 92 *data = __raw_readl(mxc_rng->mem + RNGA_OUTPUT_FIFO); mxc_rnga_data_read() 95 err = __raw_readl(mxc_rng->mem + RNGA_STATUS) & RNGA_STATUS_ERROR_INT; mxc_rnga_data_read() 100 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); mxc_rnga_data_read() 102 mxc_rng->mem + RNGA_CONTROL); mxc_rnga_data_read() 114 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); mxc_rnga_init() 115 __raw_writel(ctrl & ~RNGA_CONTROL_SLEEP, mxc_rng->mem + RNGA_CONTROL); mxc_rnga_init() 118 osc = __raw_readl(mxc_rng->mem + RNGA_STATUS); mxc_rnga_init() 125 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); mxc_rnga_init() 126 __raw_writel(ctrl | RNGA_CONTROL_GO, mxc_rng->mem + RNGA_CONTROL); mxc_rnga_init() 136 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); mxc_rnga_cleanup() 139 __raw_writel(ctrl & ~RNGA_CONTROL_GO, mxc_rng->mem + RNGA_CONTROL); mxc_rnga_cleanup() 172 mxc_rng->mem = devm_ioremap_resource(&pdev->dev, res); mxc_rnga_probe() 173 if (IS_ERR(mxc_rng->mem)) { mxc_rnga_probe() 174 err = PTR_ERR(mxc_rng->mem); mxc_rnga_probe()
|
H A D | geode-rng.c | 57 void __iomem *mem = (void __iomem *)rng->priv; geode_rng_data_read() local 59 *data = readl(mem + GEODE_RNG_DATA_REG); geode_rng_data_read() 66 void __iomem *mem = (void __iomem *)rng->priv; geode_rng_data_present() local 70 data = !!(readl(mem + GEODE_RNG_STATUS_REG)); geode_rng_data_present() 91 void __iomem *mem; mod_init() local 107 mem = ioremap(rng_base, 0x58); 108 if (!mem) 110 geode_rng.priv = (unsigned long)mem; 123 iounmap(mem); 129 void __iomem *mem = (void __iomem *)geode_rng.priv; mod_exit() local 132 iounmap(mem); mod_exit()
|
H A D | exynos-rng.c | 42 void __iomem *mem; member in struct:exynos_rng 48 return __raw_readl(rng->mem + offset); exynos_rng_readl() 53 __raw_writel(val, rng->mem + offset); exynos_rng_writel() 124 exynos_rng->mem = devm_ioremap_resource(&pdev->dev, res); exynos_rng_probe() 125 if (IS_ERR(exynos_rng->mem)) exynos_rng_probe() 126 return PTR_ERR(exynos_rng->mem); exynos_rng_probe()
|
/linux-4.1.27/drivers/net/wireless/ti/wlcore/ |
H A D | io.c | 83 if ((addr >= part->mem.start) && wlcore_translate_addr() 84 (addr < part->mem.start + part->mem.size)) wlcore_translate_addr() 85 return addr - part->mem.start; wlcore_translate_addr() 88 return addr - part->reg.start + part->mem.size; wlcore_translate_addr() 91 return addr - part->mem2.start + part->mem.size + wlcore_translate_addr() 95 return addr - part->mem3.start + part->mem.size + wlcore_translate_addr() 117 * ...+----+--> mem.start 121 * 00000000 <--+----+... ...+----+--> mem.start + mem.size 125 * mem.size <--+----+... | | {unused area) 128 * mem.size | | ... | | 146 p->mem.start, p->mem.size); wlcore_set_partition() 154 ret = wlcore_raw_write32(wl, HW_PART0_START_ADDR, p->mem.start); wlcore_set_partition() 158 ret = wlcore_raw_write32(wl, HW_PART0_SIZE_ADDR, p->mem.size); wlcore_set_partition()
|
/linux-4.1.27/drivers/iio/adc/ |
H A D | men_z188_adc.c | 28 struct resource *mem; member in struct:z188_adc 109 struct resource *mem; men_z188_probe() local 123 mem = mcb_request_mem(dev, "z188-adc"); men_z188_probe() 124 if (IS_ERR(mem)) men_z188_probe() 125 return PTR_ERR(mem); men_z188_probe() 127 adc->base = ioremap(mem->start, resource_size(mem)); men_z188_probe() 133 adc->mem = mem; men_z188_probe() 139 mcb_release_mem(mem); men_z188_probe() 150 mcb_release_mem(adc->mem); men_z188_remove()
|
/linux-4.1.27/arch/um/os-Linux/ |
H A D | Makefile | 6 obj-y = aio.o execvp.o file.o helper.o irq.o main.o mem.o process.o \ 13 main.o mem.o process.o registers.o sigio.o signal.o start_up.o time.o \
|
/linux-4.1.27/arch/avr32/kernel/ |
H A D | setup.c | 80 struct resource *mem, *res; resource_init() local 85 for (mem = system_ram; mem; mem = mem->sibling) { resource_init() 87 memcpy(new, mem, sizeof(struct resource)); resource_init() 92 mem->start, mem->end); resource_init() 175 find_free_region(const struct resource *mem, resource_size_t size, find_free_region() argument 181 target = ALIGN(mem->start, align); find_free_region() 189 if ((target + size) > (mem->end + 1)) find_free_region() 190 return mem->end + 1; find_free_region() 199 struct resource *mem; alloc_reserved_region() local 203 for (mem = system_ram; mem; mem = mem->sibling) { alloc_reserved_region() 204 target = find_free_region(mem, size, align); alloc_reserved_region() 205 if (target <= mem->end) { alloc_reserved_region() 287 * Pick out the memory size. We look for mem=size@start, 303 early_param("mem", early_mem); 339 struct tag_mem_range *mem = &tag->u.mem_range; parse_tag_rdimg() local 348 ret = add_reserved_region(mem->addr, mem->addr + mem->size - 1, parse_tag_rdimg() 356 initrd_start = (unsigned long)__va(mem->addr); parse_tag_rdimg() 357 initrd_end = initrd_start + mem->size; parse_tag_rdimg() 369 struct tag_mem_range *mem = &tag->u.mem_range; parse_tag_rsvd_mem() local 371 return add_reserved_region(mem->addr, mem->addr + mem->size - 1, parse_tag_rsvd_mem() 444 find_bootmap_pfn(const struct resource *mem) find_bootmap_pfn() argument 447 unsigned long node_pages = PFN_UP(resource_size(mem)); find_bootmap_pfn() 462 bootmap_start = find_free_region(mem, bootmap_len, PAGE_SIZE); find_bootmap_pfn()
|
/linux-4.1.27/drivers/net/can/sja1000/ |
H A D | sja1000_isa.c | 43 static unsigned long mem[MAXDEV]; variable 54 module_param_array(mem, ulong, NULL, S_IRUGO); 55 MODULE_PARM_DESC(mem, "I/O memory address"); 136 dev_dbg(&pdev->dev, "probing idx=%d: port=%#lx, mem=%#lx, irq=%d\n", sja1000_isa_probe() 137 idx, port[idx], mem[idx], irq[idx]); sja1000_isa_probe() 139 if (mem[idx]) { sja1000_isa_probe() 140 if (!request_mem_region(mem[idx], iosize, DRV_NAME)) { sja1000_isa_probe() 144 base = ioremap_nocache(mem[idx], iosize); sja1000_isa_probe() 168 if (mem[idx]) { sja1000_isa_probe() 170 dev->base_addr = mem[idx]; sja1000_isa_probe() 224 if (mem[idx]) sja1000_isa_probe() 227 if (mem[idx]) sja1000_isa_probe() 228 release_mem_region(mem[idx], iosize); sja1000_isa_probe() 243 if (mem[idx]) { sja1000_isa_remove() 245 release_mem_region(mem[idx], SJA1000_IOSIZE); sja1000_isa_remove() 270 if ((port[idx] || mem[idx]) && irq[idx]) { sja1000_isa_init() 282 pr_debug("%s: platform device %d: port=%#lx, mem=%#lx, " sja1000_isa_init() 284 DRV_NAME, idx, port[idx], mem[idx], irq[idx]); sja1000_isa_init() 285 } else if (idx == 0 || port[idx] || mem[idx]) { sja1000_isa_init()
|
/linux-4.1.27/drivers/char/agp/ |
H A D | sgi-agp.c | 59 * Flush GART tlb's. Cannot selectively flush based on memory so the mem 63 static void sgi_tioca_tlbflush(struct agp_memory *mem) sgi_tioca_tlbflush() argument 65 tioca_tlbflush(mem->bridge->dev_private_data); sgi_tioca_tlbflush() 127 static int sgi_tioca_insert_memory(struct agp_memory *mem, off_t pg_start, sgi_tioca_insert_memory() argument 137 bridge = mem->bridge; sgi_tioca_insert_memory() 169 if (type != 0 || mem->type != 0) { sgi_tioca_insert_memory() 173 if ((pg_start + mem->page_count) > num_entries) sgi_tioca_insert_memory() 178 while (j < (pg_start + mem->page_count)) { sgi_tioca_insert_memory() 184 if (!mem->is_flushed) { sgi_tioca_insert_memory() 186 mem->is_flushed = true; sgi_tioca_insert_memory() 189 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { sgi_tioca_insert_memory() 192 page_to_phys(mem->pages[i]), sgi_tioca_insert_memory() 193 mem->type); sgi_tioca_insert_memory() 196 bridge->driver->tlb_flush(mem); sgi_tioca_insert_memory() 200 static int sgi_tioca_remove_memory(struct agp_memory *mem, off_t pg_start, sgi_tioca_remove_memory() argument 207 bridge = mem->bridge; sgi_tioca_remove_memory() 211 if (type != 0 || mem->type != 0) { sgi_tioca_remove_memory() 217 for (i = pg_start; i < (mem->page_count + pg_start); i++) { sgi_tioca_remove_memory() 221 bridge->driver->tlb_flush(mem); sgi_tioca_remove_memory()
|
H A D | ati-agp.c | 170 static void ati_tlbflush(struct agp_memory * mem) ati_tlbflush() argument 266 static int ati_insert_memory(struct agp_memory * mem, ati_insert_memory() argument 276 mask_type = agp_generic_type_to_mask_type(mem->bridge, type); ati_insert_memory() 277 if (mask_type != 0 || type != mem->type) ati_insert_memory() 280 if (mem->page_count == 0) ati_insert_memory() 283 if ((pg_start + mem->page_count) > num_entries) ati_insert_memory() 287 while (j < (pg_start + mem->page_count)) { ati_insert_memory() 295 if (!mem->is_flushed) { ati_insert_memory() 298 mem->is_flushed = true; ati_insert_memory() 301 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { ati_insert_memory() 305 page_to_phys(mem->pages[i]), ati_insert_memory() 306 mem->type), ati_insert_memory() 310 agp_bridge->driver->tlb_flush(mem); ati_insert_memory() 314 static int ati_remove_memory(struct agp_memory * mem, off_t pg_start, ati_remove_memory() argument 322 mask_type = agp_generic_type_to_mask_type(mem->bridge, type); ati_remove_memory() 323 if (mask_type != 0 || type != mem->type) ati_remove_memory() 326 if (mem->page_count == 0) ati_remove_memory() 329 for (i = pg_start; i < (mem->page_count + pg_start); i++) { ati_remove_memory() 336 agp_bridge->driver->tlb_flush(mem); ati_remove_memory()
|
H A D | alpha-agp.c | 67 static void alpha_core_agp_tlbflush(struct agp_memory *mem) alpha_core_agp_tlbflush() argument 86 static int alpha_core_agp_insert_memory(struct agp_memory *mem, off_t pg_start, alpha_core_agp_insert_memory() argument 93 if (type >= AGP_USER_TYPES || mem->type >= AGP_USER_TYPES) alpha_core_agp_insert_memory() 98 if ((pg_start + mem->page_count) > num_entries) alpha_core_agp_insert_memory() 101 status = agp->ops->bind(agp, pg_start, mem); alpha_core_agp_insert_memory() 103 alpha_core_agp_tlbflush(mem); alpha_core_agp_insert_memory() 108 static int alpha_core_agp_remove_memory(struct agp_memory *mem, off_t pg_start, alpha_core_agp_remove_memory() argument 114 status = agp->ops->unbind(agp, pg_start, mem); alpha_core_agp_remove_memory() 115 alpha_core_agp_tlbflush(mem); alpha_core_agp_remove_memory()
|
H A D | nvidia-agp.c | 201 static int nvidia_insert_memory(struct agp_memory *mem, off_t pg_start, int type) nvidia_insert_memory() argument 206 mask_type = agp_generic_type_to_mask_type(mem->bridge, type); nvidia_insert_memory() 207 if (mask_type != 0 || type != mem->type) nvidia_insert_memory() 210 if (mem->page_count == 0) nvidia_insert_memory() 213 if ((pg_start + mem->page_count) > nvidia_insert_memory() 217 for (j = pg_start; j < (pg_start + mem->page_count); j++) { nvidia_insert_memory() 222 if (!mem->is_flushed) { nvidia_insert_memory() 224 mem->is_flushed = true; nvidia_insert_memory() 226 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { nvidia_insert_memory() 228 page_to_phys(mem->pages[i]), mask_type), nvidia_insert_memory() 235 agp_bridge->driver->tlb_flush(mem); nvidia_insert_memory() 240 static int nvidia_remove_memory(struct agp_memory *mem, off_t pg_start, int type) nvidia_remove_memory() argument 246 mask_type = agp_generic_type_to_mask_type(mem->bridge, type); nvidia_remove_memory() 247 if (mask_type != 0 || type != mem->type) nvidia_remove_memory() 250 if (mem->page_count == 0) nvidia_remove_memory() 253 for (i = pg_start; i < (mem->page_count + pg_start); i++) nvidia_remove_memory() 256 agp_bridge->driver->tlb_flush(mem); nvidia_remove_memory() 261 static void nvidia_tlbflush(struct agp_memory *mem) nvidia_tlbflush() argument
|
H A D | parisc-agp.c | 93 parisc_agp_tlbflush(struct agp_memory *mem) parisc_agp_tlbflush() argument 125 parisc_agp_insert_memory(struct agp_memory *mem, off_t pg_start, int type) parisc_agp_insert_memory() argument 132 if (type != mem->type || parisc_agp_insert_memory() 138 io_pg_count = info->io_pages_per_kpage * mem->page_count; parisc_agp_insert_memory() 150 if (!mem->is_flushed) { parisc_agp_insert_memory() 152 mem->is_flushed = true; parisc_agp_insert_memory() 155 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { parisc_agp_insert_memory() 158 paddr = page_to_phys(mem->pages[i]); parisc_agp_insert_memory() 168 agp_bridge->driver->tlb_flush(mem); parisc_agp_insert_memory() 174 parisc_agp_remove_memory(struct agp_memory *mem, off_t pg_start, int type) parisc_agp_remove_memory() argument 179 if (type != mem->type || parisc_agp_remove_memory() 185 io_pg_count = info->io_pages_per_kpage * mem->page_count; parisc_agp_remove_memory() 190 agp_bridge->driver->tlb_flush(mem); parisc_agp_remove_memory()
|
H A D | i460-agp.c | 170 static void i460_tlb_flush (struct agp_memory *mem) i460_tlb_flush() argument 293 static int i460_insert_memory_small_io_page (struct agp_memory *mem, i460_insert_memory_small_io_page() argument 300 pr_debug("i460_insert_memory_small_io_page(mem=%p, pg_start=%ld, type=%d, paddr0=0x%lx)\n", i460_insert_memory_small_io_page() 301 mem, pg_start, type, page_to_phys(mem->pages[0])); i460_insert_memory_small_io_page() 303 if (type >= AGP_USER_TYPES || mem->type >= AGP_USER_TYPES) i460_insert_memory_small_io_page() 311 if ((io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count) > num_entries) { i460_insert_memory_small_io_page() 317 while (j < (io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count)) { i460_insert_memory_small_io_page() 327 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { i460_insert_memory_small_io_page() 328 paddr = page_to_phys(mem->pages[i]); i460_insert_memory_small_io_page() 330 WR_GATT(j, i460_mask_memory(agp_bridge, paddr, mem->type)); i460_insert_memory_small_io_page() 336 static int i460_remove_memory_small_io_page(struct agp_memory *mem, i460_remove_memory_small_io_page() argument 341 pr_debug("i460_remove_memory_small_io_page(mem=%p, pg_start=%ld, type=%d)\n", i460_remove_memory_small_io_page() 342 mem, pg_start, type); i460_remove_memory_small_io_page() 346 for (i = pg_start; i < (pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count); i++) i460_remove_memory_small_io_page() 400 static int i460_insert_memory_large_io_page (struct agp_memory *mem, i460_insert_memory_large_io_page() argument 407 if (type >= AGP_USER_TYPES || mem->type >= AGP_USER_TYPES) i460_insert_memory_large_io_page() 415 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; i460_insert_memory_large_io_page() 417 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; i460_insert_memory_large_io_page() 453 mem->pages[i] = lp->page; i460_insert_memory_large_io_page() 461 static int i460_remove_memory_large_io_page (struct agp_memory *mem, i460_remove_memory_large_io_page() argument 473 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; i460_remove_memory_large_io_page() 475 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; i460_remove_memory_large_io_page() 482 mem->pages[i] = NULL; i460_remove_memory_large_io_page() 500 static int i460_insert_memory (struct agp_memory *mem, i460_insert_memory() argument 504 return i460_insert_memory_small_io_page(mem, pg_start, type); i460_insert_memory() 506 return i460_insert_memory_large_io_page(mem, pg_start, type); i460_insert_memory() 509 static int i460_remove_memory (struct agp_memory *mem, i460_remove_memory() argument 513 return i460_remove_memory_small_io_page(mem, pg_start, type); i460_remove_memory() 515 return i460_remove_memory_large_io_page(mem, pg_start, type); i460_remove_memory()
|
H A D | sworks-agp.c | 319 static int serverworks_insert_memory(struct agp_memory *mem, serverworks_insert_memory() argument 328 if (type != 0 || mem->type != 0) { serverworks_insert_memory() 331 if ((pg_start + mem->page_count) > num_entries) { serverworks_insert_memory() 336 while (j < (pg_start + mem->page_count)) { serverworks_insert_memory() 344 if (!mem->is_flushed) { serverworks_insert_memory() 346 mem->is_flushed = true; serverworks_insert_memory() 349 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { serverworks_insert_memory() 353 page_to_phys(mem->pages[i]), mem->type), serverworks_insert_memory() 356 serverworks_tlbflush(mem); serverworks_insert_memory() 360 static int serverworks_remove_memory(struct agp_memory *mem, off_t pg_start, serverworks_remove_memory() argument 367 if (type != 0 || mem->type != 0) { serverworks_remove_memory() 372 serverworks_tlbflush(mem); serverworks_remove_memory() 374 for (i = pg_start; i < (mem->page_count + pg_start); i++) { serverworks_remove_memory() 380 serverworks_tlbflush(mem); serverworks_remove_memory()
|
H A D | efficeon-agp.c | 109 static void efficeon_tlbflush(struct agp_memory * mem) efficeon_tlbflush() argument 238 static int efficeon_insert_memory(struct agp_memory * mem, off_t pg_start, int type) efficeon_insert_memory() argument 240 int i, count = mem->page_count, num_entries; efficeon_insert_memory() 248 if ((pg_start + mem->page_count) > num_entries) efficeon_insert_memory() 250 if (type != 0 || mem->type != 0) efficeon_insert_memory() 253 if (!mem->is_flushed) { efficeon_insert_memory() 255 mem->is_flushed = true; efficeon_insert_memory() 261 unsigned long insert = efficeon_mask_memory(mem->pages[i]); efficeon_insert_memory() 283 agp_bridge->driver->tlb_flush(mem); efficeon_insert_memory() 287 static int efficeon_remove_memory(struct agp_memory * mem, off_t pg_start, int type) efficeon_remove_memory() argument 289 int i, count = mem->page_count, num_entries; efficeon_remove_memory() 295 if ((pg_start + mem->page_count) > num_entries) efficeon_remove_memory() 297 if (type != 0 || mem->type != 0) efficeon_remove_memory() 309 agp_bridge->driver->tlb_flush(mem); efficeon_remove_memory()
|
H A D | uninorth-agp.c | 73 static void uninorth_tlbflush(struct agp_memory *mem) uninorth_tlbflush() argument 83 if (!mem && uninorth_rev <= 0x30) { uninorth_tlbflush() 149 static int uninorth_insert_memory(struct agp_memory *mem, off_t pg_start, int type) uninorth_insert_memory() argument 156 if (type != mem->type) uninorth_insert_memory() 165 if (mem->page_count == 0) uninorth_insert_memory() 171 if ((pg_start + mem->page_count) > num_entries) uninorth_insert_memory() 175 for (i = 0; i < mem->page_count; ++i) { uninorth_insert_memory() 184 for (i = 0; i < mem->page_count; i++) { uninorth_insert_memory() 186 gp[i] = (page_to_phys(mem->pages[i]) >> PAGE_SHIFT) | 0x80000000UL; uninorth_insert_memory() 188 gp[i] = cpu_to_le32((page_to_phys(mem->pages[i]) & 0xFFFFF000UL) | uninorth_insert_memory() 190 flush_dcache_range((unsigned long)__va(page_to_phys(mem->pages[i])), uninorth_insert_memory() 191 (unsigned long)__va(page_to_phys(mem->pages[i]))+0x1000); uninorth_insert_memory() 194 uninorth_tlbflush(mem); uninorth_insert_memory() 199 int uninorth_remove_memory(struct agp_memory *mem, off_t pg_start, int type) uninorth_remove_memory() argument 205 if (type != mem->type) uninorth_remove_memory() 214 if (mem->page_count == 0) uninorth_remove_memory() 218 for (i = 0; i < mem->page_count; ++i) { uninorth_remove_memory() 222 uninorth_tlbflush(mem); uninorth_remove_memory()
|
H A D | generic.c | 89 void agp_alloc_page_array(size_t size, struct agp_memory *mem) agp_alloc_page_array() argument 91 mem->pages = NULL; agp_alloc_page_array() 94 mem->pages = kmalloc(size, GFP_KERNEL | __GFP_NOWARN); agp_alloc_page_array() 95 if (mem->pages == NULL) { agp_alloc_page_array() 96 mem->pages = vmalloc(size); agp_alloc_page_array() 1033 int agp_generic_insert_memory(struct agp_memory * mem, off_t pg_start, int type) agp_generic_insert_memory() argument 1042 bridge = mem->bridge; agp_generic_insert_memory() 1046 if (mem->page_count == 0) agp_generic_insert_memory() 1075 if (type != mem->type) agp_generic_insert_memory() 1084 if (((pg_start + mem->page_count) > num_entries) || agp_generic_insert_memory() 1085 ((pg_start + mem->page_count) < pg_start)) agp_generic_insert_memory() 1090 while (j < (pg_start + mem->page_count)) { agp_generic_insert_memory() 1096 if (!mem->is_flushed) { agp_generic_insert_memory() 1098 mem->is_flushed = true; agp_generic_insert_memory() 1101 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { agp_generic_insert_memory() 1103 page_to_phys(mem->pages[i]), agp_generic_insert_memory() 1109 bridge->driver->tlb_flush(mem); agp_generic_insert_memory() 1115 int agp_generic_remove_memory(struct agp_memory *mem, off_t pg_start, int type) agp_generic_remove_memory() argument 1121 bridge = mem->bridge; agp_generic_remove_memory() 1125 if (mem->page_count == 0) agp_generic_remove_memory() 1128 if (type != mem->type) agp_generic_remove_memory() 1132 if (((pg_start + mem->page_count) > num_entries) || agp_generic_remove_memory() 1133 ((pg_start + mem->page_count) < pg_start)) agp_generic_remove_memory() 1143 for (i = pg_start; i < (mem->page_count + pg_start); i++) { agp_generic_remove_memory() 1148 bridge->driver->tlb_flush(mem); agp_generic_remove_memory() 1195 int agp_generic_alloc_pages(struct agp_bridge_data *bridge, struct agp_memory *mem, size_t num_pages) agp_generic_alloc_pages() argument 1212 mem->pages[i] = page; agp_generic_alloc_pages() 1213 mem->page_count++; agp_generic_alloc_pages() 1217 set_pages_array_uc(mem->pages, num_pages); agp_generic_alloc_pages() 1241 void agp_generic_destroy_pages(struct agp_memory *mem) agp_generic_destroy_pages() argument 1246 if (!mem) agp_generic_destroy_pages() 1250 set_pages_array_wb(mem->pages, mem->page_count); agp_generic_destroy_pages() 1253 for (i = 0; i < mem->page_count; i++) { agp_generic_destroy_pages() 1254 page = mem->pages[i]; agp_generic_destroy_pages() 1262 mem->pages[i] = NULL; agp_generic_destroy_pages() 1371 void agp3_generic_tlbflush(struct agp_memory *mem) agp3_generic_tlbflush() argument
|
H A D | amd-k7-agp.c | 283 static int amd_insert_memory(struct agp_memory *mem, off_t pg_start, int type) amd_insert_memory() argument 291 if (type != mem->type || amd_insert_memory() 295 if ((pg_start + mem->page_count) > num_entries) amd_insert_memory() 299 while (j < (pg_start + mem->page_count)) { amd_insert_memory() 307 if (!mem->is_flushed) { amd_insert_memory() 309 mem->is_flushed = true; amd_insert_memory() 312 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { amd_insert_memory() 316 page_to_phys(mem->pages[i]), amd_insert_memory() 317 mem->type), amd_insert_memory() 321 amd_irongate_tlbflush(mem); amd_insert_memory() 325 static int amd_remove_memory(struct agp_memory *mem, off_t pg_start, int type) amd_remove_memory() argument 331 if (type != mem->type || amd_remove_memory() 335 for (i = pg_start; i < (mem->page_count + pg_start); i++) { amd_remove_memory() 342 amd_irongate_tlbflush(mem); amd_remove_memory()
|
H A D | hp-agp.c | 282 hp_zx1_tlbflush (struct agp_memory *mem) hp_zx1_tlbflush() argument 332 hp_zx1_insert_memory (struct agp_memory *mem, off_t pg_start, int type) hp_zx1_insert_memory() argument 339 if (type != mem->type || hp_zx1_insert_memory() 345 io_pg_count = hp->io_pages_per_kpage * mem->page_count; hp_zx1_insert_memory() 358 if (!mem->is_flushed) { hp_zx1_insert_memory() 360 mem->is_flushed = true; hp_zx1_insert_memory() 363 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { hp_zx1_insert_memory() 366 paddr = page_to_phys(mem->pages[i]); hp_zx1_insert_memory() 374 agp_bridge->driver->tlb_flush(mem); hp_zx1_insert_memory() 379 hp_zx1_remove_memory (struct agp_memory *mem, off_t pg_start, int type) hp_zx1_remove_memory() argument 384 if (type != mem->type || hp_zx1_remove_memory() 390 io_pg_count = hp->io_pages_per_kpage * mem->page_count; hp_zx1_remove_memory() 395 agp_bridge->driver->tlb_flush(mem); hp_zx1_remove_memory()
|
H A D | intel-gtt.c | 126 DBG("try unmapping %lu pages\n", (unsigned long)mem->page_count); intel_gtt_unmap_memory() 211 static int i810_insert_dcache_entries(struct agp_memory *mem, off_t pg_start, i810_insert_dcache_entries() argument 216 if ((pg_start + mem->page_count) i810_insert_dcache_entries() 220 if (!mem->is_flushed) i810_insert_dcache_entries() 223 for (i = pg_start; i < (pg_start + mem->page_count); i++) { i810_insert_dcache_entries() 352 return 0; /* no stolen mem on i81x */ intel_gtt_stolen_size() 881 static int intel_fake_agp_insert_entries(struct agp_memory *mem, intel_fake_agp_insert_entries() argument 894 return i810_insert_dcache_entries(mem, pg_start, type); intel_fake_agp_insert_entries() 896 if (mem->page_count == 0) intel_fake_agp_insert_entries() 899 if (pg_start + mem->page_count > intel_private.gtt_total_entries) intel_fake_agp_insert_entries() 902 if (type != mem->type) intel_fake_agp_insert_entries() 908 if (!mem->is_flushed) intel_fake_agp_insert_entries() 914 ret = intel_gtt_map_memory(mem->pages, mem->page_count, &st); intel_fake_agp_insert_entries() 919 mem->sg_list = st.sgl; intel_fake_agp_insert_entries() 920 mem->num_sg = st.nents; intel_fake_agp_insert_entries() 922 intel_gtt_insert_pages(pg_start, mem->page_count, mem->pages, intel_fake_agp_insert_entries() 928 mem->is_flushed = true; intel_fake_agp_insert_entries() 946 static int intel_fake_agp_remove_entries(struct agp_memory *mem, intel_fake_agp_remove_entries() argument 949 if (mem->page_count == 0) intel_fake_agp_remove_entries() 952 intel_gtt_clear_range(pg_start, mem->page_count); intel_fake_agp_remove_entries() 955 intel_gtt_unmap_memory(mem->sg_list, mem->num_sg); intel_fake_agp_remove_entries() 956 mem->sg_list = NULL; intel_fake_agp_remove_entries() 957 mem->num_sg = 0; intel_fake_agp_remove_entries() 1189 /* i945 is the last gpu to need phys mem (for overlay and cursors). */
|
/linux-4.1.27/drivers/gpu/drm/sis/ |
H A D | sis_mm.c | 86 drm_sis_mem_t *mem = data; sis_drm_alloc() local 108 mem->size = (mem->size + SIS_MM_ALIGN_MASK) >> SIS_MM_ALIGN_SHIFT; sis_drm_alloc() 112 mem->size, 0, sis_drm_alloc() 117 item->req.size = mem->size; sis_drm_alloc() 125 mem->size, 0, sis_drm_alloc() 141 mem->offset = ((pool == 0) ? sis_drm_alloc() 144 mem->free = user_key; sis_drm_alloc() 145 mem->size = mem->size << SIS_MM_ALIGN_SHIFT; sis_drm_alloc() 155 mem->offset = 0; sis_drm_alloc() 156 mem->size = 0; sis_drm_alloc() 157 mem->free = 0; sis_drm_alloc() 159 DRM_DEBUG("alloc %d, size = %ld, offset = %ld\n", pool, mem->size, sis_drm_alloc() 160 mem->offset); sis_drm_alloc() 168 drm_sis_mem_t *mem = data; sis_drm_free() local 172 obj = idr_find(&dev_priv->object_idr, mem->free); sis_drm_free() 178 idr_remove(&dev_priv->object_idr, mem->free); sis_drm_free() 188 DRM_DEBUG("free = 0x%lx\n", mem->free); sis_drm_free()
|
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/ |
H A D | icm.c | 59 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, mlx4_free_icm_pages() 63 __free_pages(sg_page(&chunk->mem[i]), mlx4_free_icm_pages() 64 get_order(chunk->mem[i].length)); mlx4_free_icm_pages() 73 chunk->mem[i].length, mlx4_free_icm_coherent() 74 lowmem_page_address(sg_page(&chunk->mem[i])), mlx4_free_icm_coherent() 75 sg_dma_address(&chunk->mem[i])); mlx4_free_icm_coherent() 97 static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order, mlx4_alloc_icm_pages() argument 109 sg_set_page(mem, page, PAGE_SIZE << order, 0); mlx4_alloc_icm_pages() 113 static int mlx4_alloc_icm_coherent(struct device *dev, struct scatterlist *mem, mlx4_alloc_icm_coherent() argument 117 &sg_dma_address(mem), gfp_mask); mlx4_alloc_icm_coherent() 121 sg_set_buf(mem, buf, PAGE_SIZE << order); mlx4_alloc_icm_coherent() 122 BUG_ON(mem->offset); mlx4_alloc_icm_coherent() 123 sg_dma_len(mem) = PAGE_SIZE << order; mlx4_alloc_icm_coherent() 167 sg_init_table(chunk->mem, MLX4_ICM_CHUNK_LEN); mlx4_alloc_icm() 178 &chunk->mem[chunk->npages], mlx4_alloc_icm() 181 ret = mlx4_alloc_icm_pages(&chunk->mem[chunk->npages], mlx4_alloc_icm() 197 chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->mem, mlx4_alloc_icm() 212 chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->mem, mlx4_alloc_icm() 330 if (sg_dma_len(&chunk->mem[i]) > dma_offset) mlx4_table_find() 331 *dma_handle = sg_dma_address(&chunk->mem[i]) + mlx4_table_find() 333 dma_offset -= sg_dma_len(&chunk->mem[i]); mlx4_table_find() 340 if (chunk->mem[i].length > offset) { mlx4_table_find() 341 page = sg_page(&chunk->mem[i]); mlx4_table_find() 344 offset -= chunk->mem[i].length; mlx4_table_find()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/ |
H A D | nouveau_sgdma.c | 27 nv04_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem) nv04_sgdma_bind() argument 30 struct nvkm_mem *node = mem->mm_node; nv04_sgdma_bind() 39 node->size = (mem->num_pages << PAGE_SHIFT) >> 12; nv04_sgdma_bind() 61 nv50_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem) nv50_sgdma_bind() argument 64 struct nvkm_mem *node = mem->mm_node; nv50_sgdma_bind() 74 node->size = (mem->num_pages << PAGE_SHIFT) >> 12; nv50_sgdma_bind()
|
H A D | nouveau_ttm.c | 64 struct ttm_mem_reg *mem) nouveau_vram_manager_del() 68 nvkm_mem_node_cleanup(mem->mm_node); nouveau_vram_manager_del() 69 pfb->ram->put(pfb, (struct nvkm_mem **)&mem->mm_node); nouveau_vram_manager_del() 76 struct ttm_mem_reg *mem) nouveau_vram_manager_new() 91 ret = pfb->ram->get(pfb, mem->num_pages << PAGE_SHIFT, nouveau_vram_manager_new() 92 mem->page_alignment << PAGE_SHIFT, size_nc, nouveau_vram_manager_new() 95 mem->mm_node = NULL; nouveau_vram_manager_new() 101 mem->mm_node = node; nouveau_vram_manager_new() 102 mem->start = node->offset >> PAGE_SHIFT; nouveau_vram_manager_new() 154 struct ttm_mem_reg *mem) nouveau_gart_manager_del() 156 nvkm_mem_node_cleanup(mem->mm_node); nouveau_gart_manager_del() 157 kfree(mem->mm_node); nouveau_gart_manager_del() 158 mem->mm_node = NULL; nouveau_gart_manager_del() 165 struct ttm_mem_reg *mem) nouveau_gart_manager_new() 190 mem->mm_node = node; nouveau_gart_manager_new() 191 mem->start = 0; nouveau_gart_manager_new() 232 nv04_gart_manager_del(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) nv04_gart_manager_del() argument 234 struct nvkm_mem *node = mem->mm_node; nv04_gart_manager_del() 237 kfree(mem->mm_node); nv04_gart_manager_del() 238 mem->mm_node = NULL; nv04_gart_manager_del() 245 struct ttm_mem_reg *mem) nv04_gart_manager_new() 256 ret = nvkm_vm_get(man->priv, mem->num_pages << 12, node->page_shift, nv04_gart_manager_new() 263 mem->mm_node = node; nv04_gart_manager_new() 264 mem->start = node->vma[0].offset >> PAGE_SHIFT; nv04_gart_manager_new() 63 nouveau_vram_manager_del(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) nouveau_vram_manager_del() argument 73 nouveau_vram_manager_new(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) nouveau_vram_manager_new() argument 153 nouveau_gart_manager_del(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) nouveau_gart_manager_del() argument 162 nouveau_gart_manager_new(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) nouveau_gart_manager_new() argument 242 nv04_gart_manager_new(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) nv04_gart_manager_new() argument
|
H A D | nouveau_bo.c | 227 nvbo->bo.mem.num_pages = size >> PAGE_SHIFT; nouveau_bo_new() 268 nvbo->bo.mem.num_pages < vram_pages / 4) { set_placement_range() 327 if (bo->mem.mem_type == TTM_PL_VRAM) { nouveau_bo_pin() 328 struct nvkm_mem *mem = bo->mem.mm_node; nouveau_bo_pin() local 329 if (!list_is_singular(&mem->regions)) nouveau_bo_pin() 338 if (!(memtype & (1 << bo->mem.mem_type)) || evict) { nouveau_bo_pin() 341 1 << bo->mem.mem_type, memtype); nouveau_bo_pin() 368 switch (bo->mem.mem_type) { nouveau_bo_pin() 370 drm->gem.vram_available -= bo->mem.size; nouveau_bo_pin() 373 drm->gem.gart_available -= bo->mem.size; nouveau_bo_pin() 402 nouveau_bo_placement_set(nvbo, bo->mem.placement, 0); nouveau_bo_unpin() 406 switch (bo->mem.mem_type) { nouveau_bo_unpin() 408 drm->gem.vram_available += bo->mem.size; nouveau_bo_unpin() 411 drm->gem.gart_available += bo->mem.size; nouveau_bo_unpin() 437 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.mem.num_pages, nouveau_bo_map() 515 _nouveau_bo_mem_index(struct nouveau_bo *nvbo, unsigned index, void *mem, u8 sz) _nouveau_bo_mem_index() argument 518 u8 *m = mem; _nouveau_bo_mem_index() 540 u16 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem); nouveau_bo_wr16() local 542 mem = nouveau_bo_mem_index(nvbo, index, mem); nouveau_bo_wr16() 545 iowrite16_native(val, (void __force __iomem *)mem); nouveau_bo_wr16() 547 *mem = val; nouveau_bo_wr16() 554 u32 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem); nouveau_bo_rd32() local 556 mem = nouveau_bo_mem_index(nvbo, index, mem); nouveau_bo_rd32() 559 return ioread32_native((void __force __iomem *)mem); nouveau_bo_rd32() 561 return *mem; nouveau_bo_rd32() 568 u32 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem); nouveau_bo_wr32() local 570 mem = nouveau_bo_mem_index(nvbo, index, mem); nouveau_bo_wr32() 573 iowrite32_native(val, (void __force __iomem *)mem); nouveau_bo_wr32() 575 *mem = val; nouveau_bo_wr32() 668 switch (bo->mem.mem_type) { nouveau_bo_evict_flags() 982 struct nouveau_channel *chan, struct ttm_mem_reg *mem) nouveau_bo_mem_ctxdma() 984 if (mem->mem_type == TTM_PL_TT) nouveau_bo_mem_ctxdma() 1037 struct ttm_mem_reg *mem) nouveau_bo_move_prep() 1039 struct nvkm_mem *old_node = bo->mem.mm_node; nouveau_bo_move_prep() 1040 struct nvkm_mem *new_node = mem->mm_node; nouveau_bo_move_prep() 1041 u64 size = (u64)mem->num_pages << PAGE_SHIFT; nouveau_bo_move_prep() 1084 ret = drm->ttm.move(chan, bo, &bo->mem, new_mem); nouveau_bo_move_m2mf() 1294 struct ttm_mem_reg *old_mem = &bo->mem; nouveau_bo_move() 1309 BUG_ON(bo->mem.mm_node != NULL); nouveau_bo_move() 1310 bo->mem = *new_mem; nouveau_bo_move() 1355 nouveau_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) nouveau_ttm_io_mem_reserve() argument 1357 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; nouveau_ttm_io_mem_reserve() 1359 struct nvkm_mem *node = mem->mm_node; nouveau_ttm_io_mem_reserve() 1362 mem->bus.addr = NULL; nouveau_ttm_io_mem_reserve() 1363 mem->bus.offset = 0; nouveau_ttm_io_mem_reserve() 1364 mem->bus.size = mem->num_pages << PAGE_SHIFT; nouveau_ttm_io_mem_reserve() 1365 mem->bus.base = 0; nouveau_ttm_io_mem_reserve() 1366 mem->bus.is_iomem = false; nouveau_ttm_io_mem_reserve() 1369 switch (mem->mem_type) { nouveau_ttm_io_mem_reserve() 1376 mem->bus.offset = mem->start << PAGE_SHIFT; nouveau_ttm_io_mem_reserve() 1377 mem->bus.base = drm->agp.base; nouveau_ttm_io_mem_reserve() 1378 mem->bus.is_iomem = !drm->dev->agp->cant_use_aperture; nouveau_ttm_io_mem_reserve() 1386 mem->bus.offset = mem->start << PAGE_SHIFT; nouveau_ttm_io_mem_reserve() 1387 mem->bus.base = nv_device_resource_start(nvxx_device(&drm->device), 1); nouveau_ttm_io_mem_reserve() 1388 mem->bus.is_iomem = true; nouveau_ttm_io_mem_reserve() 1397 mem->bus.offset = node->bar_vma.offset; nouveau_ttm_io_mem_reserve() 1407 nouveau_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) nouveau_ttm_io_mem_free() argument 1411 struct nvkm_mem *node = mem->mm_node; nouveau_ttm_io_mem_free() 1431 if (bo->mem.mem_type != TTM_PL_VRAM) { nouveau_ttm_fault_reserve_notify() 1436 if (bo->mem.mem_type == TTM_PL_SYSTEM) { nouveau_ttm_fault_reserve_notify() 1448 bo->mem.start + bo->mem.num_pages < mappable) nouveau_ttm_fault_reserve_notify() 1634 const u32 size = nvbo->bo.mem.num_pages << PAGE_SHIFT; nouveau_bo_vma_add() 1642 if ( nvbo->bo.mem.mem_type != TTM_PL_SYSTEM && nouveau_bo_vma_add() 1643 (nvbo->bo.mem.mem_type == TTM_PL_VRAM || nouveau_bo_vma_add() 1645 nvkm_vm_map(vma, nvbo->bo.mem.mm_node); nouveau_bo_vma_add() 1656 if (nvbo->bo.mem.mem_type != TTM_PL_SYSTEM) nouveau_bo_vma_del() 981 nouveau_bo_mem_ctxdma(struct ttm_buffer_object *bo, struct nouveau_channel *chan, struct ttm_mem_reg *mem) nouveau_bo_mem_ctxdma() argument 1036 nouveau_bo_move_prep(struct nouveau_drm *drm, struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) nouveau_bo_move_prep() argument
|
H A D | nv50_fence.c | 40 struct ttm_mem_reg *mem = &priv->bo->bo.mem; nv50_fence_context_new() local 41 u32 start = mem->start * PAGE_SIZE; nv50_fence_context_new() 42 u32 limit = start + mem->size - 1; nv50_fence_context_new() 66 u32 start = bo->bo.mem.start * PAGE_SIZE; nv50_fence_context_new() 67 u32 limit = start + bo->bo.mem.size - 1; nv50_fence_context_new()
|
H A D | nouveau_gem.c | 117 const bool mapped = nvbo->bo.mem.mem_type != TTM_PL_SYSTEM; nouveau_gem_object_unmap() 212 ret = drm_gem_object_init(dev, &nvbo->gem, nvbo->bo.mem.size); nouveau_gem_new() 232 else if (nvbo->bo.mem.mem_type == TTM_PL_TT) nouveau_gem_info() 245 rep->size = nvbo->bo.mem.num_pages << PAGE_SHIFT; nouveau_gem_info() 306 bo->mem.mem_type == TTM_PL_VRAM) nouveau_gem_set_domain() 310 bo->mem.mem_type == TTM_PL_TT) nouveau_gem_set_domain() 500 ((nvbo->bo.mem.mem_type == TTM_PL_VRAM && list_for_each_entry() 502 (nvbo->bo.mem.mem_type == TTM_PL_TT && list_for_each_entry() 506 if (nvbo->bo.mem.mem_type == TTM_PL_TT) list_for_each_entry() 568 void *mem; u_memcpya() local 573 mem = kmalloc(size, GFP_KERNEL | __GFP_NOWARN); u_memcpya() 574 if (!mem) u_memcpya() 575 mem = vmalloc(size); u_memcpya() 576 if (!mem) u_memcpya() 579 if (copy_from_user(mem, userptr, size)) { u_memcpya() 580 u_free(mem); u_memcpya() 584 return mem; u_memcpya() 624 nvbo->bo.mem.num_pages << PAGE_SHIFT)) { nouveau_gem_pushbuf_reloc_apply() 631 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.mem.num_pages, nouveau_gem_pushbuf_reloc_apply() 803 nvbo->bo.mem. nouveau_gem_ioctl_pushbuf()
|
H A D | nv17_fence.c | 78 struct ttm_mem_reg *mem = &priv->bo->bo.mem; nv17_fence_context_new() local 79 u32 start = mem->start * PAGE_SIZE; nv17_fence_context_new() 80 u32 limit = start + mem->size - 1; nv17_fence_context_new()
|
/linux-4.1.27/drivers/ata/ |
H A D | pata_palmld.c | 55 void __iomem *mem; palmld_pata_probe() local 66 mem = devm_ioremap(&pdev->dev, PALMLD_IDE_PHYS, 0x1000); palmld_pata_probe() 67 if (!mem) { palmld_pata_probe() 91 ap->ioaddr.cmd_addr = mem + 0x10; palmld_pata_probe() 92 ap->ioaddr.altstatus_addr = mem + 0xe; palmld_pata_probe() 93 ap->ioaddr.ctl_addr = mem + 0xe; palmld_pata_probe()
|
/linux-4.1.27/arch/powerpc/boot/ |
H A D | cuboot-pq2.c | 117 * the device tree. The prefetch mem region and non-prefetch mem region 120 * 32-bit PCI is supported. All three region types (prefetchable mem, 121 * non-prefetchable mem, and I/O) must be present. 125 struct pci_range *mem = NULL, *mmio = NULL, fixup_pci() local 165 mem = &pci_ranges_buf[i]; fixup_pci() 172 if (!mem || !mmio || !io) fixup_pci() 174 if (mem->size[1] != mmio->size[1]) fixup_pci() 176 if (mem->size[1] & (mem->size[1] - 1)) fixup_pci() 181 if (mem->phys_addr + mem->size[1] == mmio->phys_addr) fixup_pci() 182 mem_base = mem; fixup_pci() 183 else if (mmio->phys_addr + mmio->size[1] == mem->phys_addr) fixup_pci() 189 out_be32(&pci_regs[2][0], ~(mem->size[1] + mmio->size[1] - 1)); fixup_pci() 194 out_le32(&pci_regs[0][0], mem->pci_addr[1] >> 12); fixup_pci() 195 out_le32(&pci_regs[0][2], mem->phys_addr >> 12); fixup_pci() 196 out_le32(&pci_regs[0][4], (~(mem->size[1] - 1) >> 12) | 0xa0000000); fixup_pci()
|
H A D | wii.c | 106 void *mem; platform_fixups() local 112 mem = finddevice("/memory"); platform_fixups() 113 if (!mem) platform_fixups() 117 len = getprop(mem, "reg", reg, sizeof(reg)); platform_fixups() 133 setprop(mem, "reg", reg, sizeof(reg)); platform_fixups()
|
/linux-4.1.27/tools/power/cpupower/debug/i386/ |
H A D | dump_psb.c | 154 char *mem = NULL; main() local 178 fd = open("/dev/mem", O_RDONLY); main() 180 printf ("Couldn't open /dev/mem. Are you root?\n"); main() 184 mem = mmap(mem, 0x100000 - 0xc0000, PROT_READ, MAP_SHARED, fd, 0xc0000); main() 187 for (p = mem; p - mem < LEN; p+=16) { main() 194 munmap(mem, LEN); main()
|
/linux-4.1.27/arch/mips/bcm47xx/ |
H A D | prom.c | 57 unsigned long mem; prom_init_mem() local 81 for (mem = 1 << 20; mem < max; mem += 1 << 20) { prom_init_mem() 83 if (off + mem >= max) { prom_init_mem() 84 mem = max; prom_init_mem() 88 if (!memcmp(prom_init, prom_init + mem, 32)) prom_init_mem() 91 lowmem = mem; prom_init_mem() 98 if (c->cputype == CPU_74K && (mem == (128 << 20))) prom_init_mem() 99 mem -= 0x1000; prom_init_mem() 100 add_memory_region(0, mem, BOOT_MEM_RAM); prom_init_mem()
|
/linux-4.1.27/net/core/ |
H A D | sock_diag.c | 54 u32 mem[SK_MEMINFO_VARS]; sock_diag_put_meminfo() local 56 mem[SK_MEMINFO_RMEM_ALLOC] = sk_rmem_alloc_get(sk); sock_diag_put_meminfo() 57 mem[SK_MEMINFO_RCVBUF] = sk->sk_rcvbuf; sock_diag_put_meminfo() 58 mem[SK_MEMINFO_WMEM_ALLOC] = sk_wmem_alloc_get(sk); sock_diag_put_meminfo() 59 mem[SK_MEMINFO_SNDBUF] = sk->sk_sndbuf; sock_diag_put_meminfo() 60 mem[SK_MEMINFO_FWD_ALLOC] = sk->sk_forward_alloc; sock_diag_put_meminfo() 61 mem[SK_MEMINFO_WMEM_QUEUED] = sk->sk_wmem_queued; sock_diag_put_meminfo() 62 mem[SK_MEMINFO_OPTMEM] = atomic_read(&sk->sk_omem_alloc); sock_diag_put_meminfo() 63 mem[SK_MEMINFO_BACKLOG] = sk->sk_backlog.len; sock_diag_put_meminfo() 65 return nla_put(skb, attrtype, sizeof(mem), &mem); sock_diag_put_meminfo()
|
/linux-4.1.27/drivers/net/can/cc770/ |
H A D | cc770_isa.c | 25 * insmod cc770_isa.ko mem=0xd1000,0xd1000 irq=7,11 43 * insmod cc770_isa.ko mem=0xd1000,0xd1000 irq=7,11 clk=24000000 47 * insmod cc770_isa.ko mem=0xd1000,0xd1000 irq=7,11 clk=24000000,24000000 77 static unsigned long mem[MAXDEV]; variable 88 module_param_array(mem, ulong, NULL, S_IRUGO); 89 MODULE_PARM_DESC(mem, "I/O memory address"); 179 dev_dbg(&pdev->dev, "probing idx=%d: port=%#lx, mem=%#lx, irq=%d\n", cc770_isa_probe() 180 idx, port[idx], mem[idx], irq[idx]); cc770_isa_probe() 181 if (mem[idx]) { cc770_isa_probe() 182 if (!request_mem_region(mem[idx], iosize, KBUILD_MODNAME)) { cc770_isa_probe() 186 base = ioremap_nocache(mem[idx], iosize); cc770_isa_probe() 210 if (mem[idx]) { cc770_isa_probe() 212 dev->base_addr = mem[idx]; cc770_isa_probe() 283 if (mem[idx]) cc770_isa_probe() 286 if (mem[idx]) cc770_isa_probe() 287 release_mem_region(mem[idx], iosize); cc770_isa_probe() 302 if (mem[idx]) { cc770_isa_remove() 304 release_mem_region(mem[idx], CC770_IOSIZE); cc770_isa_remove() 329 if ((port[idx] || mem[idx]) && irq[idx]) { cc770_isa_init() 341 pr_debug("platform device %d: port=%#lx, mem=%#lx, " cc770_isa_init() 343 idx, port[idx], mem[idx], irq[idx]); cc770_isa_init() 344 } else if (idx == 0 || port[idx] || mem[idx]) { cc770_isa_init()
|
H A D | cc770_platform.c | 171 struct resource *mem; cc770_platform_probe() local 176 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); cc770_platform_probe() 178 if (!mem || irq <= 0) cc770_platform_probe() 181 mem_size = resource_size(mem); cc770_platform_probe() 182 if (!request_mem_region(mem->start, mem_size, pdev->name)) cc770_platform_probe() 185 base = ioremap(mem->start, mem_size); cc770_platform_probe() 236 release_mem_region(mem->start, mem_size); cc770_platform_probe() 245 struct resource *mem; cc770_platform_remove() local 251 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); cc770_platform_remove() 252 release_mem_region(mem->start, resource_size(mem)); cc770_platform_remove()
|
/linux-4.1.27/drivers/gpu/drm/qxl/ |
H A D | qxl_ttm.c | 217 struct ttm_mem_reg *mem) qxl_ttm_io_mem_reserve() 219 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; qxl_ttm_io_mem_reserve() 222 mem->bus.addr = NULL; qxl_ttm_io_mem_reserve() 223 mem->bus.offset = 0; qxl_ttm_io_mem_reserve() 224 mem->bus.size = mem->num_pages << PAGE_SHIFT; qxl_ttm_io_mem_reserve() 225 mem->bus.base = 0; qxl_ttm_io_mem_reserve() 226 mem->bus.is_iomem = false; qxl_ttm_io_mem_reserve() 229 switch (mem->mem_type) { qxl_ttm_io_mem_reserve() 234 mem->bus.is_iomem = true; qxl_ttm_io_mem_reserve() 235 mem->bus.base = qdev->vram_base; qxl_ttm_io_mem_reserve() 236 mem->bus.offset = mem->start << PAGE_SHIFT; qxl_ttm_io_mem_reserve() 239 mem->bus.is_iomem = true; qxl_ttm_io_mem_reserve() 240 mem->bus.base = qdev->surfaceram_base; qxl_ttm_io_mem_reserve() 241 mem->bus.offset = mem->start << PAGE_SHIFT; qxl_ttm_io_mem_reserve() 250 struct ttm_mem_reg *mem) qxl_ttm_io_mem_free() 340 struct ttm_mem_reg *old_mem = &bo->mem; qxl_move_null() 352 struct ttm_mem_reg *old_mem = &bo->mem; qxl_bo_move() 371 if (bo->mem.mem_type == TTM_PL_PRIV0 && qbo->surface_id) qxl_bo_move_notify() 216 qxl_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) qxl_ttm_io_mem_reserve() argument 249 qxl_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) qxl_ttm_io_mem_free() argument
|
H A D | qxl_object.c | 147 struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; qxl_bo_kmap_atomic_page() 152 if (bo->tbo.mem.mem_type == TTM_PL_VRAM) qxl_bo_kmap_atomic_page() 154 else if (bo->tbo.mem.mem_type == TTM_PL_PRIV0) qxl_bo_kmap_atomic_page() 160 ret = ttm_mem_io_reserve(bo->tbo.bdev, &bo->tbo.mem); qxl_bo_kmap_atomic_page() 163 return io_mapping_map_atomic_wc(map, bo->tbo.mem.bus.offset + page_offset); qxl_bo_kmap_atomic_page() 189 struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; qxl_bo_kunmap_atomic_page() 192 if (bo->tbo.mem.mem_type == TTM_PL_VRAM) qxl_bo_kunmap_atomic_page() 194 else if (bo->tbo.mem.mem_type == TTM_PL_PRIV0) qxl_bo_kunmap_atomic_page() 202 ttm_mem_io_free(bo->tbo.bdev, &bo->tbo.mem); qxl_bo_kunmap_atomic_page()
|
/linux-4.1.27/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_alloc.h | 47 struct i40e_dma_mem *mem, 51 struct i40e_dma_mem *mem); 53 struct i40e_virt_mem *mem, 56 struct i40e_virt_mem *mem);
|
H A D | i40e_hmc.c | 51 struct i40e_dma_mem mem; i40e_add_sd_table_entry() local 78 ret_code = i40e_allocate_dma_mem(hw, &mem, mem_type, alloc_len, i40e_add_sd_table_entry() 92 sd_entry->u.pd_table.pd_page_addr = mem; i40e_add_sd_table_entry() 94 sd_entry->u.bp.addr = mem; i40e_add_sd_table_entry() 109 i40e_free_dma_mem(hw, &mem); i40e_add_sd_table_entry() 137 struct i40e_dma_mem mem; i40e_add_pd_table_entry() local 159 ret_code = i40e_allocate_dma_mem(hw, &mem, i40e_mem_bp, i40e_add_pd_table_entry() 165 pd_entry->bp.addr = mem; i40e_add_pd_table_entry() 169 page_desc = mem.pa | 0x1; i40e_add_pd_table_entry()
|
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/ |
H A D | i40e_alloc.h | 47 struct i40e_dma_mem *mem, 51 struct i40e_dma_mem *mem); 53 struct i40e_virt_mem *mem, 56 struct i40e_virt_mem *mem);
|
/linux-4.1.27/drivers/gpu/drm/mgag200/ |
H A D | mgag200_main.c | 79 static int mga_probe_vram(struct mga_device *mdev, void __iomem *mem) mga_probe_vram() argument 87 orig = ioread16(mem); mga_probe_vram() 88 iowrite16(0, mem); mga_probe_vram() 91 orig1 = ioread8(mem + offset); mga_probe_vram() 92 orig2 = ioread8(mem + offset + 0x100); mga_probe_vram() 94 iowrite16(0xaa55, mem + offset); mga_probe_vram() 95 iowrite16(0xaa55, mem + offset + 0x100); mga_probe_vram() 97 test1 = ioread16(mem + offset); mga_probe_vram() 98 test2 = ioread16(mem); mga_probe_vram() 100 iowrite16(orig1, mem + offset); mga_probe_vram() 101 iowrite16(orig2, mem + offset + 0x100); mga_probe_vram() 112 iowrite16(orig, mem); mga_probe_vram() 119 void __iomem *mem; mga_vram_init() local 140 mem = pci_iomap(mdev->dev->pdev, 0, 0); mga_vram_init() 142 mdev->mc.vram_size = mga_probe_vram(mdev, mem); mga_vram_init() 144 pci_iounmap(mdev->dev->pdev, mem); mga_vram_init()
|
H A D | mgag200_ttm.c | 157 struct ttm_mem_reg *mem) mgag200_ttm_io_mem_reserve() 159 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; mgag200_ttm_io_mem_reserve() 162 mem->bus.addr = NULL; mgag200_ttm_io_mem_reserve() 163 mem->bus.offset = 0; mgag200_ttm_io_mem_reserve() 164 mem->bus.size = mem->num_pages << PAGE_SHIFT; mgag200_ttm_io_mem_reserve() 165 mem->bus.base = 0; mgag200_ttm_io_mem_reserve() 166 mem->bus.is_iomem = false; mgag200_ttm_io_mem_reserve() 169 switch (mem->mem_type) { mgag200_ttm_io_mem_reserve() 174 mem->bus.offset = mem->start << PAGE_SHIFT; mgag200_ttm_io_mem_reserve() 175 mem->bus.base = pci_resource_start(mdev->dev->pdev, 0); mgag200_ttm_io_mem_reserve() 176 mem->bus.is_iomem = true; mgag200_ttm_io_mem_reserve() 185 static void mgag200_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) mgag200_ttm_io_mem_free() argument 156 mgag200_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) mgag200_ttm_io_mem_reserve() argument
|
/linux-4.1.27/drivers/usb/host/ |
H A D | ohci-sm501.c | 91 struct resource *res, *mem; ohci_hcd_sm501_drv_probe() local 99 mem = platform_get_resource(pdev, IORESOURCE_MEM, 1); ohci_hcd_sm501_drv_probe() 100 if (mem == NULL) { ohci_hcd_sm501_drv_probe() 106 if (!request_mem_region(mem->start, resource_size(mem), pdev->name)) { ohci_hcd_sm501_drv_probe() 126 if (!dma_declare_coherent_memory(dev, mem->start, ohci_hcd_sm501_drv_probe() 127 mem->start - mem->parent->start, ohci_hcd_sm501_drv_probe() 128 resource_size(mem), ohci_hcd_sm501_drv_probe() 188 release_mem_region(mem->start, resource_size(mem)); ohci_hcd_sm501_drv_probe() 196 struct resource *mem; ohci_hcd_sm501_drv_remove() local 202 mem = platform_get_resource(pdev, IORESOURCE_MEM, 1); ohci_hcd_sm501_drv_remove() 203 if (mem) ohci_hcd_sm501_drv_remove() 204 release_mem_region(mem->start, resource_size(mem)); ohci_hcd_sm501_drv_remove()
|
/linux-4.1.27/arch/arm/mach-sa1100/ |
H A D | pci-nanoengine.c | 76 pci 0000:00:00.0: reg 10: [mem 0x00021000-0x00021fff] 78 pci 0000:00:00.0: reg 18: [mem 0x00000000-0x0001ffff] 79 pci 0000:00:00.0: reg 30: [mem 0x00000000-0x000fffff pref] 84 pci 0000:00:00.0: BAR 6: can't assign mem pref (size 0x100000) 85 pci 0000:00:00.0: BAR 2: assigned [mem 0x18600000-0x1861ffff] 86 pci 0000:00:00.0: BAR 2: set to [mem 0x18600000-0x1861ffff] (PCI address [0x0-0x1ffff]) 87 pci 0000:00:00.0: BAR 0: assigned [mem 0x18620000-0x18620fff] 88 pci 0000:00:00.0: BAR 0: set to [mem 0x18620000-0x18620fff] (PCI address [0x20000-0x20fff]) 102 pci 0000:00:00.0: reg 10: [mem 0x00021000-0x00021fff] 104 pci 0000:00:00.0: reg 18: [mem 0x00000000-0x0001ffff] 105 pci 0000:00:00.0: reg 30: [mem 0x00000000-0x000fffff pref] 110 pci 0000:00:00.0: BAR 6: assigned [mem 0x78000000-0x780fffff pref] 111 pci 0000:00:00.0: BAR 2: assigned [mem 0x18600000-0x1861ffff] 112 pci 0000:00:00.0: BAR 2: set to [mem 0x18600000-0x1861ffff] (PCI address [0x0-0x1ffff]) 113 pci 0000:00:00.0: BAR 0: assigned [mem 0x18620000-0x18620fff] 114 pci 0000:00:00.0: BAR 0: set to [mem 0x18620000-0x18620fff] (PCI address [0x20000-0x20fff])
|
/linux-4.1.27/drivers/infiniband/ulp/iser/ |
H A D | iser_memory.c | 391 iser_reg_dma(struct iser_device *device, struct iser_data_buf *mem, iser_reg_dma() argument 394 struct scatterlist *sg = mem->sg; iser_reg_dma() 409 struct iser_data_buf *mem, fall_to_bounce_buf() 418 aligned_len, mem->size); fall_to_bounce_buf() 421 iser_data_buf_dump(mem, device->ib_device); fall_to_bounce_buf() 424 iser_dma_unmap_task_data(iser_task, mem, fall_to_bounce_buf() 430 if (iser_start_rdma_unaligned_sg(iser_task, mem, cmd_dir) != 0) fall_to_bounce_buf() 443 struct iser_data_buf *mem, iser_reg_page_vec() 452 plen = iser_sg_to_page_vec(mem, device->ib_device, iser_reg_page_vec() 459 iser_data_buf_dump(mem, device->ib_device); iser_reg_page_vec() 530 struct iser_data_buf *mem = &iser_task->data[cmd_dir]; iser_reg_rdma_mem_fmr() local 538 aligned_len = iser_data_buf_aligned_len(mem, ibdev); iser_reg_rdma_mem_fmr() 539 if (aligned_len != mem->dma_nents) { iser_reg_rdma_mem_fmr() 540 err = fall_to_bounce_buf(iser_task, mem, iser_reg_rdma_mem_fmr() 549 if (mem->dma_nents == 1) { iser_reg_rdma_mem_fmr() 550 return iser_reg_dma(device, mem, mem_reg); iser_reg_rdma_mem_fmr() 552 err = iser_reg_page_vec(iser_task, mem, ib_conn->fmr.page_vec, iser_reg_rdma_mem_fmr() 555 iser_data_buf_dump(mem, ibdev); iser_reg_rdma_mem_fmr() 556 iser_err("mem->dma_nents = %d (dlength = 0x%x)\n", iser_reg_rdma_mem_fmr() 557 mem->dma_nents, iser_reg_rdma_mem_fmr() 597 sig_attrs->mem.sig_type = IB_SIG_TYPE_NONE; iser_set_sig_attrs() 604 iser_set_dif_domain(sc, sig_attrs, &sig_attrs->mem); iser_set_sig_attrs() 605 sig_attrs->mem.sig.dif.bg_type = sc->prot_flags & SCSI_PROT_IP_CHECKSUM ? iser_set_sig_attrs() 612 iser_set_dif_domain(sc, sig_attrs, &sig_attrs->mem); iser_set_sig_attrs() 613 sig_attrs->mem.sig.dif.bg_type = sc->prot_flags & SCSI_PROT_IP_CHECKSUM ? iser_set_sig_attrs() 713 struct iser_data_buf *mem, iser_fast_reg_mr() 727 if (mem->dma_nents == 1) iser_fast_reg_mr() 728 return iser_reg_dma(device, mem, reg); iser_fast_reg_mr() 738 plen = iser_sg_to_page_vec(mem, device->ib_device, frpl->page_list, iser_fast_reg_mr() 796 struct iser_data_buf *mem = &iser_task->data[cmd_dir]; iser_reg_rdma_mem_fastreg() local 801 aligned_len = iser_data_buf_aligned_len(mem, ibdev); iser_reg_rdma_mem_fastreg() 802 if (aligned_len != mem->dma_nents) { iser_reg_rdma_mem_fastreg() 803 err = fall_to_bounce_buf(iser_task, mem, iser_reg_rdma_mem_fastreg() 811 if (mem->dma_nents != 1 || iser_reg_rdma_mem_fastreg() 817 err = iser_fast_reg_mr(iser_task, mem, desc, iser_reg_rdma_mem_fastreg() 827 mem = &iser_task->prot[cmd_dir]; iser_reg_rdma_mem_fastreg() 828 aligned_len = iser_data_buf_aligned_len(mem, ibdev); iser_reg_rdma_mem_fastreg() 829 if (aligned_len != mem->dma_nents) { iser_reg_rdma_mem_fastreg() 830 err = fall_to_bounce_buf(iser_task, mem, iser_reg_rdma_mem_fastreg() 838 err = iser_fast_reg_mr(iser_task, mem, desc, iser_reg_rdma_mem_fastreg() 408 fall_to_bounce_buf(struct iscsi_iser_task *iser_task, struct iser_data_buf *mem, enum iser_data_dir cmd_dir, int aligned_len) fall_to_bounce_buf() argument 442 iser_reg_page_vec(struct iscsi_iser_task *iser_task, struct iser_data_buf *mem, struct iser_page_vec *page_vec, struct iser_mem_reg *mem_reg) iser_reg_page_vec() argument 712 iser_fast_reg_mr(struct iscsi_iser_task *iser_task, struct iser_data_buf *mem, struct fast_reg_descriptor *desc, enum iser_reg_indicator ind, struct iser_mem_reg *reg) iser_fast_reg_mr() argument
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/ |
H A D | nv50.c | 36 struct nvkm_mem *mem; member in struct:nv50_instobj_priv 49 u64 base = (node->mem->offset + offset) & 0xffffff00000ULL; nv50_instobj_rd32() 50 u64 addr = (node->mem->offset + offset) & 0x000000fffffULL; nv50_instobj_rd32() 69 u64 base = (node->mem->offset + offset) & 0xffffff00000ULL; nv50_instobj_wr32() 70 u64 addr = (node->mem->offset + offset) & 0x000000fffffULL; nv50_instobj_wr32() 86 pfb->ram->put(pfb, &node->mem); nv50_instobj_dtor() 108 ret = pfb->ram->get(pfb, args->size, args->align, 0, 0x800, &node->mem); nv50_instobj_ctor() 112 node->base.addr = node->mem->offset; nv50_instobj_ctor() 113 node->base.size = node->mem->size << 12; nv50_instobj_ctor() 114 node->mem->page_shift = 12; nv50_instobj_ctor()
|
H A D | nv04.h | 29 struct nvkm_mm_node *mem; member in struct:nv04_instobj_priv
|
H A D | gk20a.c | 56 struct nvkm_mem *mem; member in struct:gk20a_instobj_priv 57 /* Pointed by mem */ 78 /* array of base.mem->size pages */ 112 u64 base = (node->mem->offset + offset) & 0xffffff00000ULL; gk20a_instobj_rd32() 113 u64 addr = (node->mem->offset + offset) & 0x000000fffffULL; gk20a_instobj_rd32() 132 u64 base = (node->mem->offset + offset) & 0xffffff00000ULL; gk20a_instobj_wr32() 133 u64 addr = (node->mem->offset + offset) & 0x000000fffffULL; gk20a_instobj_wr32() 154 dma_free_attrs(dev, _node->mem->size << PAGE_SHIFT, node->cpuaddr, gk20a_instobj_dtor_dma() 166 if (unlikely(list_empty(&_node->mem->regions))) gk20a_instobj_dtor_iommu() 169 r = list_first_entry(&_node->mem->regions, struct nvkm_mm_node, gk20a_instobj_dtor_iommu() 176 for (i = 0; i < _node->mem->size; i++) { gk20a_instobj_dtor_iommu() 351 node->mem = &node->_mem; gk20a_instobj_ctor() 354 node->mem->size = size >> 12; gk20a_instobj_ctor() 355 node->mem->memtype = 0; gk20a_instobj_ctor() 356 node->mem->page_shift = 12; gk20a_instobj_ctor() 358 node->base.addr = node->mem->offset; gk20a_instobj_ctor() 362 size, align, node->mem->offset); gk20a_instobj_ctor()
|
H A D | nv04.c | 37 return nv_ro32(priv, node->mem->offset + addr); nv04_instobj_rd32() 45 nv_wo32(priv, node->mem->offset + addr, data); nv04_instobj_wr32() 56 nvkm_mm_free(&priv->heap, &node->mem); nv04_instobj_dtor() 83 args->align, &node->mem); nv04_instobj_ctor() 88 node->base.addr = node->mem->offset; nv04_instobj_ctor() 89 node->base.size = node->mem->length; nv04_instobj_ctor()
|
/linux-4.1.27/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_gmrid_manager.c | 50 struct ttm_mem_reg *mem) vmw_gmrid_man_get_node() 57 mem->mm_node = NULL; vmw_gmrid_man_get_node() 84 mem->mm_node = gman; vmw_gmrid_man_get_node() 85 mem->start = id; vmw_gmrid_man_get_node() 86 mem->num_pages = bo->num_pages; vmw_gmrid_man_get_node() 102 struct ttm_mem_reg *mem) vmw_gmrid_man_put_node() 107 if (mem->mm_node) { vmw_gmrid_man_put_node() 109 ida_remove(&gman->gmr_ida, mem->start); vmw_gmrid_man_put_node() 110 gman->used_gmr_pages -= mem->num_pages; vmw_gmrid_man_put_node() 112 mem->mm_node = NULL; vmw_gmrid_man_put_node() 47 vmw_gmrid_man_get_node(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) vmw_gmrid_man_get_node() argument 101 vmw_gmrid_man_put_node(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) vmw_gmrid_man_put_node() argument
|
H A D | vmwgfx_buffer.c | 767 static int vmw_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) vmw_ttm_io_mem_reserve() argument 769 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; vmw_ttm_io_mem_reserve() 772 mem->bus.addr = NULL; vmw_ttm_io_mem_reserve() 773 mem->bus.is_iomem = false; vmw_ttm_io_mem_reserve() 774 mem->bus.offset = 0; vmw_ttm_io_mem_reserve() 775 mem->bus.size = mem->num_pages << PAGE_SHIFT; vmw_ttm_io_mem_reserve() 776 mem->bus.base = 0; vmw_ttm_io_mem_reserve() 779 switch (mem->mem_type) { vmw_ttm_io_mem_reserve() 785 mem->bus.offset = mem->start << PAGE_SHIFT; vmw_ttm_io_mem_reserve() 786 mem->bus.base = dev_priv->vram_start; vmw_ttm_io_mem_reserve() 787 mem->bus.is_iomem = true; vmw_ttm_io_mem_reserve() 795 static void vmw_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) vmw_ttm_io_mem_free() argument 808 * @mem: The truct ttm_mem_reg indicating to what memory 815 struct ttm_mem_reg *mem) vmw_move_notify() 817 vmw_resource_move_notify(bo, mem); vmw_move_notify() 814 vmw_move_notify(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) vmw_move_notify() argument
|
/linux-4.1.27/drivers/pnp/ |
H A D | support.c | 67 return "mem"; pnp_resource_type_name() 109 struct pnp_mem *mem; dbg_pnp_show_option() local 133 mem = &option->u.mem; dbg_pnp_show_option() 134 len += scnprintf(buf + len, sizeof(buf) - len, "mem min %#llx " dbg_pnp_show_option() 136 (unsigned long long) mem->min, dbg_pnp_show_option() 137 (unsigned long long) mem->max, dbg_pnp_show_option() 138 (unsigned long long) mem->align, dbg_pnp_show_option() 139 (unsigned long long) mem->size, mem->flags); dbg_pnp_show_option()
|
H A D | interface.c | 152 struct pnp_mem *mem) pnp_print_mem() 157 space, (unsigned long long) mem->min, pnp_print_mem() 158 (unsigned long long) mem->max, pnp_print_mem() 159 (unsigned long long) mem->align, pnp_print_mem() 160 (unsigned long long) mem->size); pnp_print_mem() 161 if (mem->flags & IORESOURCE_MEM_WRITEABLE) pnp_print_mem() 163 if (mem->flags & IORESOURCE_MEM_CACHEABLE) pnp_print_mem() 165 if (mem->flags & IORESOURCE_MEM_RANGELENGTH) pnp_print_mem() 167 if (mem->flags & IORESOURCE_MEM_SHADOWABLE) pnp_print_mem() 169 if (mem->flags & IORESOURCE_MEM_EXPANSIONROM) pnp_print_mem() 171 switch (mem->flags & IORESOURCE_MEM_TYPE_MASK) { pnp_print_mem() 195 pnp_print_mem(buffer, space, &option->u.mem); pnp_print_option() 401 } else if (!strncasecmp(buf, "mem", 3)) { resources_store() 151 pnp_print_mem(pnp_info_buffer_t * buffer, char *space, struct pnp_mem *mem) pnp_print_mem() argument
|
/linux-4.1.27/include/net/ |
H A D | inet_frag.h | 7 /* The percpu_counter "mem" need to be cacheline aligned. 8 * mem.count must not share cacheline with other writers 10 struct percpu_counter mem ____cacheline_aligned_in_smp; 131 * fragmentation mem acct sizes. 132 * The mem size of a 64K fragment is approx: 139 return percpu_counter_read(&nf->mem); frag_mem_limit() 144 __percpu_counter_add(&q->net->mem, -i, frag_percpu_counter_batch); sub_frag_mem_limit() 149 __percpu_counter_add(&q->net->mem, i, frag_percpu_counter_batch); add_frag_mem_limit() 154 percpu_counter_init(&nf->mem, 0, GFP_KERNEL); init_frag_mem_limit() 162 res = percpu_counter_sum_positive(&nf->mem); sum_frag_mem_limit()
|
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
H A D | ramnv1a.c | 36 u32 mem, mib; nv1a_ram_create() local 51 pci_read_config_dword(bridge, 0x7c, &mem); nv1a_ram_create() 52 mib = ((mem >> 6) & 31) + 1; nv1a_ram_create() 54 pci_read_config_dword(bridge, 0x84, &mem); nv1a_ram_create() 55 mib = ((mem >> 4) & 127) + 1; nv1a_ram_create()
|
H A D | ramnv50.c | 222 __nv50_ram_put(struct nvkm_fb *pfb, struct nvkm_mem *mem) __nv50_ram_put() argument 226 while (!list_empty(&mem->regions)) { __nv50_ram_put() 227 this = list_first_entry(&mem->regions, typeof(*this), rl_entry); __nv50_ram_put() 233 nvkm_mm_free(&pfb->tags, &mem->tag); __nv50_ram_put() 239 struct nvkm_mem *mem = *pmem; nv50_ram_put() local 242 if (unlikely(mem == NULL)) nv50_ram_put() 246 __nv50_ram_put(pfb, mem); nv50_ram_put() 249 kfree(mem); nv50_ram_put() 259 struct nvkm_mem *mem; nv50_ram_get() local 269 mem = kzalloc(sizeof(*mem), GFP_KERNEL); nv50_ram_get() 270 if (!mem) nv50_ram_get() 278 ret = nvkm_mm_head(tags, 0, 1, n, n, 1, &mem->tag); nv50_ram_get() 280 mem->tag = NULL; nv50_ram_get() 283 if (unlikely(!mem->tag)) nv50_ram_get() 287 INIT_LIST_HEAD(&mem->regions); nv50_ram_get() 288 mem->memtype = (comp << 7) | type; nv50_ram_get() 289 mem->size = max; nv50_ram_get() 299 pfb->ram->put(pfb, &mem); nv50_ram_get() 303 list_add_tail(&r->rl_entry, &mem->regions); nv50_ram_get() 308 r = list_first_entry(&mem->regions, struct nvkm_mm_node, rl_entry); nv50_ram_get() 309 mem->offset = (u64)r->offset << 12; nv50_ram_get() 310 *pmem = mem; nv50_ram_get()
|
H A D | ramgf100.c | 428 struct nvkm_mem *mem = *pmem; gf100_ram_put() local 431 if (unlikely(mem == NULL)) gf100_ram_put() 435 if (mem->tag) gf100_ram_put() 436 ltc->tags_free(ltc, &mem->tag); gf100_ram_put() 437 __nv50_ram_put(pfb, mem); gf100_ram_put() 440 kfree(mem); gf100_ram_put() 449 struct nvkm_mem *mem; gf100_ram_get() local 461 mem = kzalloc(sizeof(*mem), GFP_KERNEL); gf100_ram_get() 462 if (!mem) gf100_ram_get() 465 INIT_LIST_HEAD(&mem->regions); gf100_ram_get() 466 mem->size = size; gf100_ram_get() 475 ltc->tags_alloc(ltc, n, &mem->tag); gf100_ram_get() 478 if (unlikely(!mem->tag)) gf100_ram_get() 481 mem->memtype = type; gf100_ram_get() 490 pfb->ram->put(pfb, &mem); gf100_ram_get() 494 list_add_tail(&r->rl_entry, &mem->regions); gf100_ram_get() 499 r = list_first_entry(&mem->regions, struct nvkm_mm_node, rl_entry); gf100_ram_get() 500 mem->offset = (u64)r->offset << 12; gf100_ram_get() 501 *pmem = mem; gf100_ram_get()
|
/linux-4.1.27/arch/mips/include/asm/mach-jazz/ |
H A D | floppy.h | 107 unsigned long mem; fd_dma_mem_alloc() local 109 mem = __get_dma_pages(GFP_KERNEL, get_order(size)); fd_dma_mem_alloc() 110 if(!mem) fd_dma_mem_alloc() 112 vdma_alloc(CPHYSADDR(mem), size); /* XXX error checking */ fd_dma_mem_alloc() 114 return mem; fd_dma_mem_alloc()
|
/linux-4.1.27/drivers/rtc/ |
H A D | rtc-pcf8583.c | 135 static int pcf8583_read_mem(struct i2c_client *client, struct rtc_mem *mem) pcf8583_read_mem() argument 147 .len = mem->nr, pcf8583_read_mem() 148 .buf = mem->data, pcf8583_read_mem() 152 if (mem->loc < 8) pcf8583_read_mem() 155 addr[0] = mem->loc; pcf8583_read_mem() 160 static int pcf8583_write_mem(struct i2c_client *client, struct rtc_mem *mem) pcf8583_write_mem() argument 165 if (mem->loc < 8 || mem->nr > 8) pcf8583_write_mem() 168 buf[0] = mem->loc; pcf8583_write_mem() 169 memcpy(buf + 1, mem->data, mem->nr); pcf8583_write_mem() 171 ret = i2c_master_send(client, buf, mem->nr + 1); pcf8583_write_mem() 172 return ret == mem->nr + 1 ? 0 : -EIO; pcf8583_write_mem() 179 struct rtc_mem mem = { pcf8583_rtc_read_time() local 202 pcf8583_read_mem(client, &mem)) pcf8583_rtc_read_time()
|
/linux-4.1.27/drivers/gpu/drm/via/ |
H A D | via_mm.c | 113 drm_via_mem_t *mem = data; via_mem_alloc() local 120 if (mem->type > VIA_MEM_AGP) { via_mem_alloc() 125 if (0 == ((mem->type == VIA_MEM_VIDEO) ? dev_priv->vram_initialized : via_mem_alloc() 139 tmpSize = (mem->size + VIA_MM_ALIGN_MASK) >> VIA_MM_ALIGN_SHIFT; via_mem_alloc() 140 if (mem->type == VIA_MEM_AGP) via_mem_alloc() 159 mem->offset = ((mem->type == VIA_MEM_VIDEO) ? via_mem_alloc() 162 mem->index = user_key; via_mem_alloc() 172 mem->offset = 0; via_mem_alloc() 173 mem->size = 0; via_mem_alloc() 174 mem->index = 0; via_mem_alloc() 183 drm_via_mem_t *mem = data; via_mem_free() local 187 obj = idr_find(&dev_priv->object_idr, mem->index); via_mem_free() 193 idr_remove(&dev_priv->object_idr, mem->index); via_mem_free() 199 DRM_DEBUG("free = 0x%lx\n", mem->index); via_mem_free()
|
/linux-4.1.27/drivers/irqchip/ |
H A D | irq-tb10x.c | 111 struct resource mem; of_tb10x_init_irq() local 116 if (of_address_to_resource(ictl, 0, &mem)) { of_tb10x_init_irq() 122 if (!request_mem_region(mem.start, resource_size(&mem), of_tb10x_init_irq() 124 pr_err("%s: Request mem region failed.\n", ictl->name); of_tb10x_init_irq() 128 reg_base = ioremap(mem.start, resource_size(&mem)); of_tb10x_init_irq() 192 release_mem_region(mem.start, resource_size(&mem)); of_tb10x_init_irq()
|
/linux-4.1.27/include/ras/ |
H A D | ras_event.h | 26 TP_PROTO(struct cper_sec_mem_err *mem, 32 TP_ARGS(mem, err_seq, fru_id, fru_text, sev), 47 if (mem->validation_bits & CPER_MEM_VALID_ERROR_TYPE) 48 __entry->etype = mem->error_type; 52 if (mem->validation_bits & CPER_MEM_VALID_PA) 53 __entry->pa = mem->physical_addr; 57 if (mem->validation_bits & CPER_MEM_VALID_PA_MASK) 58 __entry->pa_mask_lsb = (u8)__ffs64(mem->physical_addr_mask); 63 cper_mem_err_pack(mem, &__entry->data);
|
/linux-4.1.27/arch/alpha/include/asm/ |
H A D | atomic.h | 139 "1: ldl_l %[old],%[mem]\n" __atomic_add_unless() 143 " stl_c %[new],%[mem]\n" __atomic_add_unless() 150 : [mem] "m"(*v), [a] "rI"(a), [u] "rI"((long)u) __atomic_add_unless() 171 "1: ldq_l %[tmp],%[mem]\n" atomic64_add_unless() 175 " stq_c %[tmp],%[mem]\n" atomic64_add_unless() 182 : [mem] "m"(*v), [a] "rI"(a), [u] "rI"(u) atomic64_add_unless() 200 "1: ldq_l %[old],%[mem]\n" atomic64_dec_if_positive() 203 " stq_c %[tmp],%[mem]\n" atomic64_dec_if_positive() 210 : [mem] "m"(*v) atomic64_dec_if_positive()
|
/linux-4.1.27/Documentation/ia64/ |
H A D | aliasing-test.c | 2 * Exercise /dev/mem mmap cases that have been troublesome in the past 213 if (map_mem("/dev/mem", 0, 0xA0000, 1) == 0) main() 214 fprintf(stderr, "PASS: /dev/mem 0x0-0xa0000 is readable\n"); main() 216 fprintf(stderr, "FAIL: /dev/mem 0x0-0xa0000 not accessible\n"); main() 225 if (map_mem("/dev/mem", 0xA0000, 0x20000, 0) == 0) main() 226 fprintf(stderr, "PASS: /dev/mem 0xa0000-0xc0000 is mappable\n"); main() 228 fprintf(stderr, "FAIL: /dev/mem 0xa0000-0xc0000 not accessible\n"); main() 230 if (map_mem("/dev/mem", 0xC0000, 0x40000, 1) == 0) main() 231 fprintf(stderr, "PASS: /dev/mem 0xc0000-0x100000 is readable\n"); main() 233 fprintf(stderr, "FAIL: /dev/mem 0xc0000-0x100000 not accessible\n"); main() 242 rc = map_mem("/dev/mem", 0, 1024*1024, 0); main() 244 fprintf(stderr, "PASS: /dev/mem 0x0-0x100000 is mappable\n"); main() 246 fprintf(stderr, "PASS: /dev/mem 0x0-0x100000 not mappable\n"); main() 248 fprintf(stderr, "FAIL: /dev/mem 0x0-0x100000 not accessible\n"); main()
|
/linux-4.1.27/arch/x86/realmode/ |
H A D | init.c | 13 phys_addr_t mem; reserve_real_mode() local 18 mem = memblock_find_in_range(0, 1<<20, size, PAGE_SIZE); reserve_real_mode() 19 if (!mem) reserve_real_mode() 22 base = __va(mem); reserve_real_mode() 23 memblock_reserve(mem, size); reserve_real_mode() 26 base, (unsigned long long)mem, size); reserve_real_mode()
|
/linux-4.1.27/arch/ia64/hp/sim/boot/ |
H A D | bootloader.c | 57 static char mem[4096]; start_bootloader() local 108 req.len = sizeof(mem); start_bootloader() 109 req.addr = (long) mem; start_bootloader() 113 elf = (struct elfhdr *) mem; start_bootloader() 137 req.addr = (long) mem; start_bootloader() 146 elf_phdr = (struct elf_phdr *) mem; start_bootloader()
|
/linux-4.1.27/drivers/gpu/drm/radeon/ |
H A D | radeon_kfd.c | 203 struct kgd_mem **mem = (struct kgd_mem **) mem_obj; alloc_gtt_mem() local 210 *mem = kmalloc(sizeof(struct kgd_mem), GFP_KERNEL); alloc_gtt_mem() 211 if ((*mem) == NULL) alloc_gtt_mem() 215 RADEON_GEM_GTT_WC, NULL, NULL, &(*mem)->bo); alloc_gtt_mem() 223 r = radeon_bo_reserve((*mem)->bo, true); alloc_gtt_mem() 229 r = radeon_bo_pin((*mem)->bo, RADEON_GEM_DOMAIN_GTT, alloc_gtt_mem() 230 &(*mem)->gpu_addr); alloc_gtt_mem() 235 *gpu_addr = (*mem)->gpu_addr; alloc_gtt_mem() 237 r = radeon_bo_kmap((*mem)->bo, &(*mem)->cpu_ptr); alloc_gtt_mem() 243 *cpu_ptr = (*mem)->cpu_ptr; alloc_gtt_mem() 245 radeon_bo_unreserve((*mem)->bo); alloc_gtt_mem() 250 radeon_bo_unpin((*mem)->bo); alloc_gtt_mem() 252 radeon_bo_unreserve((*mem)->bo); alloc_gtt_mem() 254 radeon_bo_unref(&(*mem)->bo); alloc_gtt_mem() 261 struct kgd_mem *mem = (struct kgd_mem *) mem_obj; free_gtt_mem() local 263 BUG_ON(mem == NULL); free_gtt_mem() 265 radeon_bo_reserve(mem->bo, true); free_gtt_mem() 266 radeon_bo_kunmap(mem->bo); free_gtt_mem() 267 radeon_bo_unpin(mem->bo); free_gtt_mem() 268 radeon_bo_unreserve(mem->bo); free_gtt_mem() 269 radeon_bo_unref(&(mem->bo)); free_gtt_mem() 270 kfree(mem); free_gtt_mem()
|
H A D | radeon_ttm.c | 197 switch (bo->mem.mem_type) { radeon_evict_flags() 202 bo->mem.start < (rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT)) { radeon_evict_flags() 246 struct ttm_mem_reg *old_mem = &bo->mem; radeon_move_null() 315 struct ttm_mem_reg *old_mem = &bo->mem; radeon_move_vram_ram() 362 struct ttm_mem_reg *old_mem = &bo->mem; radeon_move_ram_vram() 402 struct ttm_mem_reg *old_mem = &bo->mem; radeon_bo_move() 449 static int radeon_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) radeon_ttm_io_mem_reserve() argument 451 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; radeon_ttm_io_mem_reserve() 454 mem->bus.addr = NULL; radeon_ttm_io_mem_reserve() 455 mem->bus.offset = 0; radeon_ttm_io_mem_reserve() 456 mem->bus.size = mem->num_pages << PAGE_SHIFT; radeon_ttm_io_mem_reserve() 457 mem->bus.base = 0; radeon_ttm_io_mem_reserve() 458 mem->bus.is_iomem = false; radeon_ttm_io_mem_reserve() 461 switch (mem->mem_type) { radeon_ttm_io_mem_reserve() 469 mem->bus.offset = mem->start << PAGE_SHIFT; radeon_ttm_io_mem_reserve() 470 mem->bus.base = rdev->mc.agp_base; radeon_ttm_io_mem_reserve() 471 mem->bus.is_iomem = !rdev->ddev->agp->cant_use_aperture; radeon_ttm_io_mem_reserve() 476 mem->bus.offset = mem->start << PAGE_SHIFT; radeon_ttm_io_mem_reserve() 478 if ((mem->bus.offset + mem->bus.size) > rdev->mc.visible_vram_size) radeon_ttm_io_mem_reserve() 480 mem->bus.base = rdev->mc.aper_base; radeon_ttm_io_mem_reserve() 481 mem->bus.is_iomem = true; radeon_ttm_io_mem_reserve() 487 if (mem->placement & TTM_PL_FLAG_WC) radeon_ttm_io_mem_reserve() 488 mem->bus.addr = radeon_ttm_io_mem_reserve() 489 ioremap_wc(mem->bus.base + mem->bus.offset, radeon_ttm_io_mem_reserve() 490 mem->bus.size); radeon_ttm_io_mem_reserve() 492 mem->bus.addr = radeon_ttm_io_mem_reserve() 493 ioremap_nocache(mem->bus.base + mem->bus.offset, radeon_ttm_io_mem_reserve() 494 mem->bus.size); radeon_ttm_io_mem_reserve() 502 mem->bus.base = (mem->bus.base & 0x0ffffffffUL) + radeon_ttm_io_mem_reserve() 512 static void radeon_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) radeon_ttm_io_mem_free() argument
|
/linux-4.1.27/drivers/infiniband/hw/mthca/ |
H A D | mthca_memfree.c | 59 struct scatterlist mem; member in struct:mthca_user_db_table::__anon4937 69 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, mthca_free_icm_pages() 73 __free_pages(sg_page(&chunk->mem[i]), mthca_free_icm_pages() 74 get_order(chunk->mem[i].length)); mthca_free_icm_pages() 82 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, mthca_free_icm_coherent() 83 lowmem_page_address(sg_page(&chunk->mem[i])), mthca_free_icm_coherent() 84 sg_dma_address(&chunk->mem[i])); mthca_free_icm_coherent() 107 static int mthca_alloc_icm_pages(struct scatterlist *mem, int order, gfp_t gfp_mask) mthca_alloc_icm_pages() argument 119 sg_set_page(mem, page, PAGE_SIZE << order, 0); mthca_alloc_icm_pages() 123 static int mthca_alloc_icm_coherent(struct device *dev, struct scatterlist *mem, mthca_alloc_icm_coherent() argument 126 void *buf = dma_alloc_coherent(dev, PAGE_SIZE << order, &sg_dma_address(mem), mthca_alloc_icm_coherent() 131 sg_set_buf(mem, buf, PAGE_SIZE << order); mthca_alloc_icm_coherent() 132 BUG_ON(mem->offset); mthca_alloc_icm_coherent() 133 sg_dma_len(mem) = PAGE_SIZE << order; mthca_alloc_icm_coherent() 164 sg_init_table(chunk->mem, MTHCA_ICM_CHUNK_LEN); mthca_alloc_icm() 175 &chunk->mem[chunk->npages], mthca_alloc_icm() 178 ret = mthca_alloc_icm_pages(&chunk->mem[chunk->npages], mthca_alloc_icm() 187 chunk->nsg = pci_map_sg(dev->pdev, chunk->mem, mthca_alloc_icm() 207 chunk->nsg = pci_map_sg(dev->pdev, chunk->mem, mthca_alloc_icm() 300 if (sg_dma_len(&chunk->mem[i]) > dma_offset) mthca_table_find() 301 *dma_handle = sg_dma_address(&chunk->mem[i]) + mthca_table_find() 303 dma_offset -= sg_dma_len(&chunk->mem[i]); mthca_table_find() 308 if (chunk->mem[i].length > offset) { mthca_table_find() 309 page = sg_page(&chunk->mem[i]); mthca_table_find() 312 offset -= chunk->mem[i].length; mthca_table_find() 480 sg_set_page(&db_tab->page[i].mem, pages[0], MTHCA_ICM_PAGE_SIZE, mthca_map_user_db() 483 ret = pci_map_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE); mthca_map_user_db() 489 ret = mthca_MAP_ICM_page(dev, sg_dma_address(&db_tab->page[i].mem), mthca_map_user_db() 492 pci_unmap_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE); mthca_map_user_db() 493 put_page(sg_page(&db_tab->page[i].mem)); mthca_map_user_db() 541 sg_init_table(&db_tab->page[i].mem, 1); mthca_init_user_db_tab() 558 pci_unmap_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE); mthca_cleanup_user_db_tab() 559 put_page(sg_page(&db_tab->page[i].mem)); mthca_cleanup_user_db_tab()
|
H A D | mthca_mr.c | 583 mr->attr.max_pages * sizeof *mr->mem.arbel.mtts > PAGE_SIZE) mthca_fmr_alloc() 601 mr->mem.arbel.mpt = mthca_table_find(dev->mr_table.mpt_table, key, NULL); mthca_fmr_alloc() 602 BUG_ON(!mr->mem.arbel.mpt); mthca_fmr_alloc() 604 mr->mem.tavor.mpt = dev->mr_table.tavor_fmr.mpt_base + mthca_fmr_alloc() 605 sizeof *(mr->mem.tavor.mpt) * idx; mthca_fmr_alloc() 616 mr->mem.arbel.mtts = mthca_table_find(dev->mr_table.mtt_table, mthca_fmr_alloc() 618 &mr->mem.arbel.dma_handle); mthca_fmr_alloc() 619 BUG_ON(!mr->mem.arbel.mtts); mthca_fmr_alloc() 621 mr->mem.tavor.mtts = dev->mr_table.tavor_fmr.mtt_base + mtt_seg; mthca_fmr_alloc() 736 writeb(MTHCA_MPT_STATUS_SW, fmr->mem.tavor.mpt); mthca_tavor_map_phys_fmr() 741 mthca_write64_raw(mtt_entry, fmr->mem.tavor.mtts + i); mthca_tavor_map_phys_fmr() 748 __raw_writel((__force u32) mpt_entry.lkey, &fmr->mem.tavor.mpt->key); mthca_tavor_map_phys_fmr() 749 memcpy_toio(&fmr->mem.tavor.mpt->start, &mpt_entry.start, mthca_tavor_map_phys_fmr() 753 writeb(MTHCA_MPT_STATUS_HW, fmr->mem.tavor.mpt); mthca_tavor_map_phys_fmr() 779 *(u8 *) fmr->mem.arbel.mpt = MTHCA_MPT_STATUS_SW; mthca_arbel_map_phys_fmr() 783 dma_sync_single_for_cpu(&dev->pdev->dev, fmr->mem.arbel.dma_handle, mthca_arbel_map_phys_fmr() 787 fmr->mem.arbel.mtts[i] = cpu_to_be64(page_list[i] | mthca_arbel_map_phys_fmr() 790 dma_sync_single_for_device(&dev->pdev->dev, fmr->mem.arbel.dma_handle, mthca_arbel_map_phys_fmr() 793 fmr->mem.arbel.mpt->key = cpu_to_be32(key); mthca_arbel_map_phys_fmr() 794 fmr->mem.arbel.mpt->lkey = cpu_to_be32(key); mthca_arbel_map_phys_fmr() 795 fmr->mem.arbel.mpt->length = cpu_to_be64(list_len * (1ull << fmr->attr.page_shift)); mthca_arbel_map_phys_fmr() 796 fmr->mem.arbel.mpt->start = cpu_to_be64(iova); mthca_arbel_map_phys_fmr() 800 *(u8 *) fmr->mem.arbel.mpt = MTHCA_MPT_STATUS_HW; mthca_arbel_map_phys_fmr() 814 writeb(MTHCA_MPT_STATUS_SW, fmr->mem.tavor.mpt); mthca_tavor_fmr_unmap() 824 *(u8 *) fmr->mem.arbel.mpt = MTHCA_MPT_STATUS_SW; mthca_arbel_fmr_unmap()
|
/linux-4.1.27/arch/unicore32/kernel/ |
H A D | ksyms.c | 35 /* string / mem functions */ 43 /* user mem (segment) */
|
/linux-4.1.27/arch/hexagon/include/asm/ |
H A D | fixmap.h | 25 * A lot of the fixmap info is already in mem-layout.h 27 #include <asm/mem-layout.h>
|
/linux-4.1.27/drivers/staging/octeon/ |
H A D | Makefile | 16 octeon-ethernet-y += ethernet-mem.o
|
/linux-4.1.27/drivers/i2c/busses/ |
H A D | i2c-puv3.c | 189 struct resource *mem; puv3_i2c_probe() local 192 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); puv3_i2c_probe() 193 if (!mem) puv3_i2c_probe() 196 if (!request_mem_region(mem->start, resource_size(mem), "puv3_i2c")) puv3_i2c_probe() 206 mem->start); puv3_i2c_probe() 227 release_mem_region(mem->start, resource_size(mem)); puv3_i2c_probe() 235 struct resource *mem; puv3_i2c_remove() local 241 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); puv3_i2c_remove() 242 release_mem_region(mem->start, resource_size(mem)); puv3_i2c_remove()
|
/linux-4.1.27/drivers/power/reset/ |
H A D | msm-poweroff.c | 49 struct resource *mem; msm_restart_probe() local 51 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); msm_restart_probe() 52 msm_ps_hold = devm_ioremap_resource(dev, mem); msm_restart_probe()
|
/linux-4.1.27/arch/x86/kernel/ |
H A D | crash.c | 67 struct crash_mem mem; member in struct:crash_elf_data 244 static int exclude_mem_range(struct crash_mem *mem, exclude_mem_range() argument 251 for (i = 0; i < mem->nr_ranges; i++) { exclude_mem_range() 252 start = mem->ranges[i].start; exclude_mem_range() 253 end = mem->ranges[i].end; exclude_mem_range() 266 mem->ranges[i].start = 0; exclude_mem_range() 267 mem->ranges[i].end = 0; exclude_mem_range() 268 if (i < mem->nr_ranges - 1) { exclude_mem_range() 270 for (j = i; j < mem->nr_ranges - 1; j++) { exclude_mem_range() 271 mem->ranges[j].start = exclude_mem_range() 272 mem->ranges[j+1].start; exclude_mem_range() 273 mem->ranges[j].end = exclude_mem_range() 274 mem->ranges[j+1].end; exclude_mem_range() 277 mem->nr_ranges--; exclude_mem_range() 283 mem->ranges[i].end = mstart - 1; exclude_mem_range() 287 mem->ranges[i].end = mstart - 1; exclude_mem_range() 289 mem->ranges[i].start = mend + 1; exclude_mem_range() 305 if (j < mem->nr_ranges) { exclude_mem_range() 307 for (i = mem->nr_ranges - 1; i >= j; i--) exclude_mem_range() 308 mem->ranges[i + 1] = mem->ranges[i]; exclude_mem_range() 311 mem->ranges[j].start = temp_range.start; exclude_mem_range() 312 mem->ranges[j].end = temp_range.end; exclude_mem_range() 313 mem->nr_ranges++; exclude_mem_range() 319 * might lead to split and split ranges are put in ced->mem.ranges[] array 324 struct crash_mem *cmem = &ced->mem; elf_header_exclude_ranges() 366 /* Exclude unwanted mem ranges */ prepare_elf64_ram_headers_callback() 371 /* Go through all the ranges in ced->mem.ranges[] and prepare phdr */ prepare_elf64_ram_headers_callback() 372 cmem = &ced->mem; prepare_elf64_ram_headers_callback()
|
/linux-4.1.27/drivers/spi/ |
H A D | spi-dw-mmio.c | 36 struct resource *mem; dw_spi_mmio_probe() local 48 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); dw_spi_mmio_probe() 49 if (!mem) { dw_spi_mmio_probe() 50 dev_err(&pdev->dev, "no mem resource?\n"); dw_spi_mmio_probe() 54 dws->regs = devm_ioremap_resource(&pdev->dev, mem); dw_spi_mmio_probe()
|
H A D | spi-xtensa-xtfpga.c | 86 struct resource *mem; xtfpga_spi_probe() local 104 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); xtfpga_spi_probe() 105 if (!mem) { xtfpga_spi_probe() 110 xspi->regs = devm_ioremap_resource(&pdev->dev, mem); xtfpga_spi_probe()
|
/linux-4.1.27/drivers/pcmcia/ |
H A D | vrc4171_card.c | 422 static int pccard_set_mem_map(struct pcmcia_socket *sock, struct pccard_mem_map *mem) pccard_set_mem_map() argument 430 mem == NULL || mem->map >= MEM_MAX_MAPS || pccard_set_mem_map() 431 mem->res->start < CARD_MEM_START || mem->res->start > CARD_MEM_END || pccard_set_mem_map() 432 mem->res->end < CARD_MEM_START || mem->res->end > CARD_MEM_END || pccard_set_mem_map() 433 mem->res->start > mem->res->end || pccard_set_mem_map() 434 mem->card_start > CARD_MAX_MEM_OFFSET || pccard_set_mem_map() 435 mem->speed > CARD_MAX_MEM_SPEED) pccard_set_mem_map() 439 map = mem->map; pccard_set_mem_map() 447 start = (mem->res->start >> 12) & 0x3fff; pccard_set_mem_map() 448 if (mem->flags & MAP_16BIT) pccard_set_mem_map() 452 stop = (mem->res->end >> 12) & 0x3fff; pccard_set_mem_map() 453 switch (mem->speed) { pccard_set_mem_map() 468 offset = (mem->card_start >> 12) & 0x3fff; pccard_set_mem_map() 469 if (mem->flags & MAP_ATTRIB) pccard_set_mem_map() 471 if (mem->flags & MAP_WRPROT) pccard_set_mem_map() 475 if (mem->flags & MAP_ACTIVE) { pccard_set_mem_map()
|
H A D | tcic.c | 733 static int tcic_set_mem_map(struct pcmcia_socket *sock, struct pccard_mem_map *mem) tcic_set_mem_map() argument 740 "%#llx-%#llx, %#x)\n", psock, mem->map, mem->flags, tcic_set_mem_map() 741 mem->speed, (unsigned long long)mem->res->start, tcic_set_mem_map() 742 (unsigned long long)mem->res->end, mem->card_start); tcic_set_mem_map() 743 if ((mem->map > 3) || (mem->card_start > 0x3ffffff) || tcic_set_mem_map() 744 (mem->res->start > 0xffffff) || (mem->res->end > 0xffffff) || tcic_set_mem_map() 745 (mem->res->start > mem->res->end) || (mem->speed > 1000)) tcic_set_mem_map() 748 addr = TCIC_MWIN(psock, mem->map); tcic_set_mem_map() 750 base = mem->res->start; len = mem->res->end - mem->res->start; tcic_set_mem_map() 759 mmap = mem->card_start - mem->res->start; tcic_set_mem_map() 761 if (mem->flags & MAP_ATTRIB) mmap |= TCIC_MMAP_REG; tcic_set_mem_map() 766 ctl |= to_cycles(mem->speed) & TCIC_MCTL_WSCNT_MASK; tcic_set_mem_map() 767 ctl |= (mem->flags & MAP_16BIT) ? 0 : TCIC_MCTL_B8; tcic_set_mem_map() 768 ctl |= (mem->flags & MAP_WRPROT) ? TCIC_MCTL_WP : 0; tcic_set_mem_map() 769 ctl |= (mem->flags & MAP_ACTIVE) ? TCIC_MCTL_ENA : 0; tcic_set_mem_map() 783 pccard_mem_map mem = { .res = &res, }; tcic_init() local 790 mem.map = i; tcic_init() 791 tcic_set_mem_map(s, &mem); tcic_init()
|
H A D | i82092.c | 400 pccard_mem_map mem = { .res = &res, }; i82092aa_init() local 409 mem.map = i; i82092aa_init() 410 i82092aa_set_mem_map(sock, &mem); i82092aa_init() 601 static int i82092aa_set_mem_map(struct pcmcia_socket *socket, struct pccard_mem_map *mem) i82092aa_set_mem_map() argument 611 pcibios_resource_to_bus(sock_info->dev->bus, ®ion, mem->res); i82092aa_set_mem_map() 613 map = mem->map; i82092aa_set_mem_map() 620 if ( (mem->card_start > 0x3ffffff) || (region.start > region.end) || i82092aa_set_mem_map() 621 (mem->speed > 1000) ) { i82092aa_set_mem_map() 623 printk("invalid mem map for socket %i: %llx to %llx with a " i82092aa_set_mem_map() 628 mem->card_start); i82092aa_set_mem_map() 637 /* printk("set_mem_map: Setting map %i range to %x - %x on socket %i, speed is %i, active = %i \n",map, region.start,region.end,sock,mem->speed,mem->flags & MAP_ACTIVE); */ i82092aa_set_mem_map() 642 if (mem->flags & MAP_16BIT) i82092aa_set_mem_map() 644 if (mem->flags & MAP_0WS) i82092aa_set_mem_map() 651 switch (to_cycles(mem->speed)) { i82092aa_set_mem_map() 669 i = ((mem->card_start - region.start) >> 12) & 0x3fff; i82092aa_set_mem_map() 670 if (mem->flags & MAP_WRPROT) i82092aa_set_mem_map() 672 if (mem->flags & MAP_ATTRIB) { i82092aa_set_mem_map() 681 if (mem->flags & MAP_ACTIVE) i82092aa_set_mem_map()
|
H A D | vrc4173_cardu.c | 305 static int cardu_get_mem_map(unsigned int sock, struct pccard_mem_map *mem) cardu_get_mem_map() argument 312 map = mem->map; cardu_get_mem_map() 317 mem->flags = (window & MEM_WIN_EN(map)) ? MAP_ACTIVE : 0; cardu_get_mem_map() 320 mem->flags |= (start & MEM_WIN_DSIZE) ? MAP_16BIT : 0; cardu_get_mem_map() 327 mem->flags |= (offset & MEM_WIN_WP) ? MAP_WRPROT : 0; cardu_get_mem_map() 328 mem->flags |= (offset & MEM_WIN_REGSET) ? MAP_ATTRIB : 0; cardu_get_mem_map() 330 mem->card_start = offset & 0x03ffffff; cardu_get_mem_map() 333 mem->sys_start = start + page; cardu_get_mem_map() 334 mem->sys_stop = start + page; cardu_get_mem_map() 339 static int cardu_set_mem_map(unsigned int sock, struct pccard_mem_map *mem) cardu_set_mem_map() argument 347 map = mem->map; cardu_set_mem_map() 348 sys_start = mem->sys_start; cardu_set_mem_map() 349 sys_stop = mem->sys_stop; cardu_set_mem_map() 350 card_start = mem->card_start; cardu_set_mem_map() 366 if (mem->flags & MAP_16BIT) value |= MEM_WIN_DSIZE; cardu_set_mem_map() 373 if (mem->flags & MAP_WRPROT) value |= MEM_WIN_WP; cardu_set_mem_map() 374 if (mem->flags & MAP_ATTRIB) value |= MEM_WIN_REGSET; cardu_set_mem_map() 377 if (mem->flags & MAP_ACTIVE) cardu_set_mem_map()
|
H A D | m32r_pcc.c | 490 static int _pcc_set_mem_map(u_short sock, struct pccard_mem_map *mem) argument 493 u_char map = mem->map; 504 "%#llx, %#x)\n", sock, map, mem->flags, 505 mem->speed, (unsigned long long)mem->static_start, 506 mem->card_start); 511 if ((map > MAX_WIN) || (mem->card_start > 0x3ffffff)){ 518 if ((mem->flags & MAP_ACTIVE) == 0) { 531 if (mem->flags & MAP_ATTRIB) { 543 addr = t->mapaddr + (mem->card_start & M32R_PCC_MAPMASK); 546 mem->static_start = addr + mem->card_start; 639 static int pcc_set_mem_map(struct pcmcia_socket *s, struct pccard_mem_map *mem) pcc_set_mem_map() argument 645 LOCKED(_pcc_set_mem_map(sock, mem)); pcc_set_mem_map()
|
H A D | pd6729.c | 467 struct pccard_mem_map *mem) pd6729_set_mem_map() 474 map = mem->map; pd6729_set_mem_map() 480 if ((mem->res->start > mem->res->end) || (mem->speed > 1000)) { pd6729_set_mem_map() 491 i = (mem->res->start >> 12) & 0x0fff; pd6729_set_mem_map() 492 if (mem->flags & MAP_16BIT) pd6729_set_mem_map() 494 if (mem->flags & MAP_0WS) pd6729_set_mem_map() 500 i = (mem->res->end >> 12) & 0x0fff; pd6729_set_mem_map() 501 switch (to_cycles(mem->speed)) { pd6729_set_mem_map() 519 indirect_write(socket, PD67_EXT_DATA, mem->res->start >> 24); pd6729_set_mem_map() 523 i = ((mem->card_start - mem->res->start) >> 12) & 0x3fff; pd6729_set_mem_map() 524 if (mem->flags & MAP_WRPROT) pd6729_set_mem_map() 526 if (mem->flags & MAP_ATTRIB) { pd6729_set_mem_map() 537 if (mem->flags & MAP_ACTIVE) pd6729_set_mem_map() 548 pccard_mem_map mem = { .res = &res, }; pd6729_init() local 556 mem.map = i; pd6729_init() 557 pd6729_set_mem_map(sock, &mem); pd6729_init() 466 pd6729_set_mem_map(struct pcmcia_socket *sock, struct pccard_mem_map *mem) pd6729_set_mem_map() argument
|
H A D | i82365.c | 1061 static int i365_set_mem_map(u_short sock, struct pccard_mem_map *mem) i365_set_mem_map() argument 1067 "%#x)\n", sock, mem->map, mem->flags, mem->speed, i365_set_mem_map() 1068 (unsigned long long)mem->res->start, i365_set_mem_map() 1069 (unsigned long long)mem->res->end, mem->card_start); i365_set_mem_map() 1071 map = mem->map; i365_set_mem_map() 1072 if ((map > 4) || (mem->card_start > 0x3ffffff) || i365_set_mem_map() 1073 (mem->res->start > mem->res->end) || (mem->speed > 1000)) i365_set_mem_map() 1075 if ((mem->res->start > 0xffffff) || (mem->res->end > 0xffffff)) i365_set_mem_map() 1083 i = (mem->res->start >> 12) & 0x0fff; i365_set_mem_map() 1084 if (mem->flags & MAP_16BIT) i |= I365_MEM_16BIT; i365_set_mem_map() 1085 if (mem->flags & MAP_0WS) i |= I365_MEM_0WS; i365_set_mem_map() 1088 i = (mem->res->end >> 12) & 0x0fff; i365_set_mem_map() 1089 switch (to_cycles(mem->speed)) { i365_set_mem_map() 1097 i = ((mem->card_start - mem->res->start) >> 12) & 0x3fff; i365_set_mem_map() 1098 if (mem->flags & MAP_WRPROT) i |= I365_MEM_WRPROT; i365_set_mem_map() 1099 if (mem->flags & MAP_ATTRIB) i |= I365_MEM_REG; i365_set_mem_map() 1103 if (mem->flags & MAP_ACTIVE) i365_set_mem_map() 1195 static int pcic_set_mem_map(struct pcmcia_socket *s, struct pccard_mem_map *mem) pcic_set_mem_map() argument 1201 LOCKED(i365_set_mem_map(sock, mem)); pcic_set_mem_map() 1209 pccard_mem_map mem = { .res = &res, }; pcic_init() local 1216 mem.map = i; pcic_init() 1217 pcic_set_mem_map(s, &mem); pcic_init()
|
H A D | electra_cf.c | 191 struct resource mem, io; electra_cf_probe() local 197 err = of_address_to_resource(np, 0, &mem); electra_cf_probe() 213 cf->mem_phys = mem.start; electra_cf_probe() 214 cf->mem_size = PAGE_ALIGN(resource_size(&mem)); electra_cf_probe() 239 cf->iomem.end = (unsigned long)cf->mem_base + (mem.end - mem.start); electra_cf_probe() 302 dev_info(device, "at mem 0x%lx io 0x%llx irq %d\n", electra_cf_probe()
|
H A D | cistpl.c | 90 pccard_mem_map *mem = &s->cis_mem; set_cis_map() local 93 if (!(s->features & SS_CAP_STATIC_MAP) && (mem->res == NULL)) { set_cis_map() 94 mem->res = pcmcia_find_mem_region(0, s->map_size, set_cis_map() 96 if (mem->res == NULL) { set_cis_map() 105 s->cis_virt = ioremap(mem->res->start, s->map_size); set_cis_map() 107 mem->card_start = card_offset; set_cis_map() 108 mem->flags = flags; set_cis_map() 110 ret = s->ops->set_mem_map(s, mem); set_cis_map() 120 s->cis_virt = ioremap(mem->static_start, s->map_size); set_cis_map() 173 "attempt to read CIS mem at addr %#x", addr); pcmcia_read_cis_mem() 1025 static u_char *parse_mem(u_char *p, u_char *q, cistpl_mem_t *mem) parse_mem() argument 1033 mem->nwin = (*p & 0x07) + 1; parse_mem() 1040 for (i = 0; i < mem->nwin; i++) { parse_mem() 1058 mem->win[i].len = len << 8; parse_mem() 1059 mem->win[i].card_addr = ca << 8; parse_mem() 1060 mem->win[i].host_addr = ha << 8; parse_mem() 1161 entry->mem.nwin = 0; parse_cftable_entry() 1164 entry->mem.nwin = 1; parse_cftable_entry() 1165 entry->mem.win[0].len = get_unaligned_le16(p) << 8; parse_cftable_entry() 1166 entry->mem.win[0].card_addr = 0; parse_cftable_entry() 1167 entry->mem.win[0].host_addr = 0; parse_cftable_entry() 1173 entry->mem.nwin = 1; parse_cftable_entry() 1174 entry->mem.win[0].len = get_unaligned_le16(p) << 8; parse_cftable_entry() 1175 entry->mem.win[0].card_addr = get_unaligned_le16(p + 2) << 8; parse_cftable_entry() 1176 entry->mem.win[0].host_addr = 0; parse_cftable_entry() 1182 p = parse_mem(p, q, &entry->mem); parse_cftable_entry()
|
H A D | i82092aa.h | 32 static int i82092aa_set_mem_map(struct pcmcia_socket *socket, struct pccard_mem_map *mem);
|
/linux-4.1.27/tools/virtio/ |
H A D | virtio_test.c | 42 struct vhost_memory *mem; member in struct:vdev_info 125 dev->mem = malloc(offsetof(struct vhost_memory, regions) + vdev_info_init() 126 sizeof dev->mem->regions[0]); vdev_info_init() 127 assert(dev->mem); vdev_info_init() 128 memset(dev->mem, 0, offsetof(struct vhost_memory, regions) + vdev_info_init() 129 sizeof dev->mem->regions[0]); vdev_info_init() 130 dev->mem->nregions = 1; vdev_info_init() 131 dev->mem->regions[0].guest_phys_addr = (long)dev->buf; vdev_info_init() 132 dev->mem->regions[0].userspace_addr = (long)dev->buf; vdev_info_init() 133 dev->mem->regions[0].memory_size = dev->buf_size; vdev_info_init() 134 r = ioctl(dev->control, VHOST_SET_MEM_TABLE, dev->mem); vdev_info_init()
|
/linux-4.1.27/drivers/pci/hotplug/ |
H A D | ibmphp_pci.c | 335 } else if (cur_func->mem[i]) { ibmphp_configure_card() 336 ibmphp_remove_resource (cur_func->mem[i]); ibmphp_configure_card() 337 cur_func->mem[i] = NULL; ibmphp_configure_card() 364 struct resource_node *mem[6]; configure_device() local 469 debug ("there's no pfmem... going into mem.\n"); configure_device() 496 debug ("inside the mem 64 case, count %d\n", count); configure_device() 510 mem[count] = kzalloc(sizeof(struct resource_node), GFP_KERNEL); configure_device() 511 if (!mem[count]) { configure_device() 515 mem[count]->type = MEM; configure_device() 516 mem[count]->busno = func->busno; configure_device() 517 mem[count]->devfunc = PCI_DEVFN(func->device, configure_device() 519 mem[count]->len = len[count]; configure_device() 520 if (ibmphp_check_resource (mem[count], 0) == 0) { configure_device() 521 ibmphp_add_resource (mem[count]); configure_device() 522 func->mem[count] = mem[count]; configure_device() 524 err ("cannot allocate requested mem for bus %x, device %x, len %x\n", configure_device() 526 kfree (mem[count]); configure_device() 529 pci_bus_write_config_dword (ibmphp_pci_bus, devfn, address[count], func->mem[count]->start); configure_device() 531 debug ("b4 writing, start address is %x\n", func->mem[count]->start); configure_device() 538 debug ("inside mem 64 case, reg. mem, count %d\n", count); configure_device() 544 } /* end of mem */ configure_device() 584 struct resource_node *mem = NULL; configure_bridge() local 653 /* First we need to allocate mem/io for the bridge itself in case it needs it */ configure_bridge() 777 func->mem[count] = bus_mem[count]; configure_bridge() 779 err ("cannot allocate requested mem for bus %x, device %x, len %x\n", configure_bridge() 785 pci_bus_write_config_dword (ibmphp_pci_bus, devfn, address[count], func->mem[count]->start); configure_bridge() 795 } /* end of mem */ configure_bridge() 807 debug ("amount_needed->mem = %x\n", amount_needed->mem); configure_bridge() 822 func->mem[count] = NULL; configure_bridge() 852 if (!amount_needed->mem) { configure_bridge() 856 debug ("it wants %x memory behind the bridge\n", amount_needed->mem); configure_bridge() 857 mem = kzalloc(sizeof(*mem), GFP_KERNEL); configure_bridge() 858 if (!mem) { configure_bridge() 863 mem->type = MEM; configure_bridge() 864 mem->busno = func->busno; configure_bridge() 865 mem->devfunc = PCI_DEVFN(func->device, func->function); configure_bridge() 866 mem->len = amount_needed->mem; configure_bridge() 867 if (ibmphp_check_resource (mem, 1) == 0) { configure_bridge() 868 ibmphp_add_resource (mem); configure_bridge() 870 debug ("were we able to add mem\n"); configure_bridge() 875 debug ("it doesn't want n.e.pfmem mem?\n"); configure_bridge() 933 rc = add_new_bus (bus, io, mem, pfmem, func->busno); configure_bridge() 935 rc = add_new_bus (bus, io, mem, pfmem, 0xFF); configure_bridge() 1055 if (mem) configure_bridge() 1056 ibmphp_remove_resource (mem); configure_bridge() 1066 func->mem[i] = NULL; configure_bridge() 1182 amount->mem += len[count]; scan_behind_bridge() 1200 if ((amount->mem) && (amount->mem < MEMBRIDGE)) scan_behind_bridge() 1201 amount->mem = MEMBRIDGE; scan_behind_bridge() 1228 struct resource_node *mem; unconfigure_boot_device() local 1309 debug ("start address of mem is %x\n", start_address); unconfigure_boot_device() 1312 if (ibmphp_find_resource (bus, start_address, &mem, MEM) < 0) { unconfigure_boot_device() 1316 if (mem) { unconfigure_boot_device() 1317 debug ("mem->start = %x\n", mem->start); unconfigure_boot_device() 1319 ibmphp_remove_resource(mem); unconfigure_boot_device() 1326 } /* end of mem */ unconfigure_boot_device() 1339 struct resource_node *mem = NULL; unconfigure_boot_bridge() local 1424 if (ibmphp_find_resource (bus, start_address, &mem, MEM) < 0) { unconfigure_boot_bridge() 1428 if (mem) { unconfigure_boot_bridge() 1429 debug ("mem->start = %x\n", mem->start); unconfigure_boot_bridge() 1431 ibmphp_remove_resource(mem); unconfigure_boot_bridge() 1438 } /* end of mem */ unconfigure_boot_bridge() 1599 if (cur_func->mem[i]) { ibmphp_unconfigure_card() 1600 debug ("mem[%d] exists\n", i); ibmphp_unconfigure_card() 1602 ibmphp_remove_resource (cur_func->mem[i]); ibmphp_unconfigure_card() 1603 cur_func->mem[i] = NULL; ibmphp_unconfigure_card() 1633 static int add_new_bus (struct bus_node *bus, struct resource_node *io, struct resource_node *mem, struct resource_node *pfmem, u8 parent_busno) add_new_bus() argument 1662 if (mem) { add_new_bus() 1668 mem_range->start = mem->start; add_new_bus() 1669 mem_range->end = mem->end; add_new_bus()
|
/linux-4.1.27/drivers/net/wireless/ath/ath9k/ |
H A D | ahb.c | 85 void __iomem *mem; ath_ahb_probe() local 106 mem = devm_ioremap_nocache(&pdev->dev, res->start, resource_size(res)); ath_ahb_probe() 107 if (mem == NULL) { ath_ahb_probe() 133 sc->mem = mem; ath_ahb_probe() 150 wiphy_info(hw->wiphy, "%s mem=0x%lx, irq=%d\n", ath_ahb_probe() 151 hw_name, (unsigned long)mem, irq); ath_ahb_probe()
|
/linux-4.1.27/arch/arc/mm/ |
H A D | init.c | 27 /* Default tot mem from .config */ 30 /* User can over-ride above with "mem=nnn[KkMm]" in cmdline */ setup_mem_sz() 36 pr_info("\"mem=%s\": mem sz set to %ldM\n", str, TO_MB(arc_mem_sz)); setup_mem_sz() 40 early_param("mem", setup_mem_sz); 84 * even if "mem=xxx" cmline over-ride is given and/or setup_arch_memory() 95 /* Last usable page of low mem (no HIGHMEM yet for ARC port) */ setup_arch_memory()
|
/linux-4.1.27/drivers/gpu/drm/ast/ |
H A D | ast_ttm.c | 157 struct ttm_mem_reg *mem) ast_ttm_io_mem_reserve() 159 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ast_ttm_io_mem_reserve() 162 mem->bus.addr = NULL; ast_ttm_io_mem_reserve() 163 mem->bus.offset = 0; ast_ttm_io_mem_reserve() 164 mem->bus.size = mem->num_pages << PAGE_SHIFT; ast_ttm_io_mem_reserve() 165 mem->bus.base = 0; ast_ttm_io_mem_reserve() 166 mem->bus.is_iomem = false; ast_ttm_io_mem_reserve() 169 switch (mem->mem_type) { ast_ttm_io_mem_reserve() 174 mem->bus.offset = mem->start << PAGE_SHIFT; ast_ttm_io_mem_reserve() 175 mem->bus.base = pci_resource_start(ast->dev->pdev, 0); ast_ttm_io_mem_reserve() 176 mem->bus.is_iomem = true; ast_ttm_io_mem_reserve() 185 static void ast_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ast_ttm_io_mem_free() argument 156 ast_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ast_ttm_io_mem_reserve() argument
|
/linux-4.1.27/drivers/gpu/drm/bochs/ |
H A D | bochs_mm.c | 135 struct ttm_mem_reg *mem) bochs_ttm_io_mem_reserve() 137 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; bochs_ttm_io_mem_reserve() 140 mem->bus.addr = NULL; bochs_ttm_io_mem_reserve() 141 mem->bus.offset = 0; bochs_ttm_io_mem_reserve() 142 mem->bus.size = mem->num_pages << PAGE_SHIFT; bochs_ttm_io_mem_reserve() 143 mem->bus.base = 0; bochs_ttm_io_mem_reserve() 144 mem->bus.is_iomem = false; bochs_ttm_io_mem_reserve() 147 switch (mem->mem_type) { bochs_ttm_io_mem_reserve() 152 mem->bus.offset = mem->start << PAGE_SHIFT; bochs_ttm_io_mem_reserve() 153 mem->bus.base = bochs->fb_base; bochs_ttm_io_mem_reserve() 154 mem->bus.is_iomem = true; bochs_ttm_io_mem_reserve() 164 struct ttm_mem_reg *mem) bochs_ttm_io_mem_free() 134 bochs_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) bochs_ttm_io_mem_reserve() argument 163 bochs_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) bochs_ttm_io_mem_free() argument
|
H A D | bochs_hw.c | 54 unsigned long addr, size, mem, ioaddr, iosize, qext_size; bochs_hw_init() local 81 mem = bochs_dispi_read(bochs, VBE_DISPI_INDEX_VIDEO_MEMORY_64K) bochs_hw_init() 94 if (size != mem) { bochs_hw_init() 96 size, mem); bochs_hw_init() 97 size = min(size, mem); bochs_hw_init()
|
/linux-4.1.27/drivers/gpu/drm/cirrus/ |
H A D | cirrus_ttm.c | 157 struct ttm_mem_reg *mem) cirrus_ttm_io_mem_reserve() 159 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; cirrus_ttm_io_mem_reserve() 162 mem->bus.addr = NULL; cirrus_ttm_io_mem_reserve() 163 mem->bus.offset = 0; cirrus_ttm_io_mem_reserve() 164 mem->bus.size = mem->num_pages << PAGE_SHIFT; cirrus_ttm_io_mem_reserve() 165 mem->bus.base = 0; cirrus_ttm_io_mem_reserve() 166 mem->bus.is_iomem = false; cirrus_ttm_io_mem_reserve() 169 switch (mem->mem_type) { cirrus_ttm_io_mem_reserve() 174 mem->bus.offset = mem->start << PAGE_SHIFT; cirrus_ttm_io_mem_reserve() 175 mem->bus.base = pci_resource_start(cirrus->dev->pdev, 0); cirrus_ttm_io_mem_reserve() 176 mem->bus.is_iomem = true; cirrus_ttm_io_mem_reserve() 185 static void cirrus_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) cirrus_ttm_io_mem_free() argument 156 cirrus_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) cirrus_ttm_io_mem_reserve() argument
|
/linux-4.1.27/drivers/usb/gadget/ |
H A D | config.c | 130 void *mem; usb_copy_descriptors() local 138 mem = kmalloc(bytes, GFP_KERNEL); usb_copy_descriptors() 139 if (!mem) usb_copy_descriptors() 143 * to descriptors copied starting at "mem"; usb_copy_descriptors() 146 tmp = mem; usb_copy_descriptors() 147 ret = mem; usb_copy_descriptors() 148 mem += (n_desc + 1) * sizeof(*tmp); usb_copy_descriptors() 150 memcpy(mem, *src, (*src)->bLength); usb_copy_descriptors() 151 *tmp = mem; usb_copy_descriptors() 153 mem += (*src)->bLength; usb_copy_descriptors()
|
/linux-4.1.27/drivers/usb/gadget/function/ |
H A D | f_uvc.c | 453 #define UVC_COPY_DESCRIPTOR(mem, dst, desc) \ 455 memcpy(mem, desc, (desc)->bLength); \ 456 *(dst)++ = mem; \ 457 mem += (desc)->bLength; \ 460 #define UVC_COPY_DESCRIPTORS(mem, dst, src) \ 464 memcpy(mem, *__src, (*__src)->bLength); \ 465 *dst++ = mem; \ 466 mem += (*__src)->bLength; \ 485 void *mem; uvc_copy_descriptors() local 555 mem = kmalloc((n_desc + 1) * sizeof(*src) + bytes, GFP_KERNEL); uvc_copy_descriptors() 556 if (mem == NULL) uvc_copy_descriptors() 559 hdr = mem; uvc_copy_descriptors() 560 dst = mem; uvc_copy_descriptors() 561 mem += (n_desc + 1) * sizeof(*src); uvc_copy_descriptors() 564 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_iad); uvc_copy_descriptors() 565 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_control_intf); uvc_copy_descriptors() 567 uvc_control_header = mem; uvc_copy_descriptors() 568 UVC_COPY_DESCRIPTORS(mem, dst, uvc_copy_descriptors() 574 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_control_ep); uvc_copy_descriptors() 576 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_ss_control_comp); uvc_copy_descriptors() 578 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_control_cs_ep); uvc_copy_descriptors() 579 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_streaming_intf_alt0); uvc_copy_descriptors() 581 uvc_streaming_header = mem; uvc_copy_descriptors() 582 UVC_COPY_DESCRIPTORS(mem, dst, uvc_copy_descriptors() 587 UVC_COPY_DESCRIPTORS(mem, dst, uvc_streaming_std); uvc_copy_descriptors()
|
H A D | uvc_video.c | 49 void *mem; uvc_video_encode_data() local 52 mem = buf->mem + queue->buf_used; uvc_video_encode_data() 55 memcpy(data, mem, nbytes); uvc_video_encode_data() 65 void *mem = req->buf; uvc_video_encode_bulk() local 71 ret = uvc_video_encode_header(video, buf, mem, len); uvc_video_encode_bulk() 73 mem += ret; uvc_video_encode_bulk() 79 ret = uvc_video_encode_data(video, buf, mem, len); uvc_video_encode_bulk() 105 void *mem = req->buf; uvc_video_encode_isoc() local 110 ret = uvc_video_encode_header(video, buf, mem, len); uvc_video_encode_isoc() 111 mem += ret; uvc_video_encode_isoc() 115 ret = uvc_video_encode_data(video, buf, mem, len); uvc_video_encode_isoc()
|
/linux-4.1.27/drivers/gpu/drm/amd/amdkfd/ |
H A D | kfd_topology.c | 174 static int kfd_parse_subtype_mem(struct crat_subtype_memory *mem) kfd_parse_subtype_mem() argument 180 BUG_ON(!mem); kfd_parse_subtype_mem() 183 mem->promixity_domain); kfd_parse_subtype_mem() 185 if (mem->promixity_domain == i) { kfd_parse_subtype_mem() 195 if (mem->flags & CRAT_MEM_FLAGS_HOT_PLUGGABLE) kfd_parse_subtype_mem() 197 if (mem->flags & CRAT_MEM_FLAGS_NON_VOLATILE) kfd_parse_subtype_mem() 201 ((uint64_t)mem->length_high << 32) + kfd_parse_subtype_mem() 202 mem->length_low; kfd_parse_subtype_mem() 203 props->width = mem->width; kfd_parse_subtype_mem() 321 struct crat_subtype_memory *mem; kfd_parse_subtype() local 334 mem = (struct crat_subtype_memory *)sub_type_hdr; kfd_parse_subtype() 335 ret = kfd_parse_subtype_mem(mem); kfd_parse_subtype() 367 struct kfd_mem_properties *mem; kfd_release_topology_device() local 376 mem = container_of(dev->mem_props.next, kfd_release_topology_device() 378 list_del(&mem->list); kfd_release_topology_device() 379 kfree(mem); kfd_release_topology_device() 569 struct kfd_mem_properties *mem; mem_show() local 574 mem = container_of(attr, struct kfd_mem_properties, attr); mem_show() 575 sysfs_show_32bit_prop(buffer, "heap_type", mem->heap_type); mem_show() 576 sysfs_show_64bit_prop(buffer, "size_in_bytes", mem->size_in_bytes); mem_show() 577 sysfs_show_32bit_prop(buffer, "flags", mem->flags); mem_show() 578 sysfs_show_32bit_prop(buffer, "width", mem->width); mem_show() 579 ret = sysfs_show_32bit_prop(buffer, "mem_clk_max", mem->mem_clk_max); mem_show() 764 struct kfd_mem_properties *mem; kfd_remove_sysfs_node_entry() local 793 list_for_each_entry(mem, &dev->mem_props, list) kfd_remove_sysfs_node_entry() 794 if (mem->kobj) { kfd_remove_sysfs_node_entry() 795 kfd_remove_sysfs_file(mem->kobj, &mem->attr); kfd_remove_sysfs_node_entry() 796 mem->kobj = NULL; kfd_remove_sysfs_node_entry() 818 struct kfd_mem_properties *mem; kfd_build_sysfs_node_entry() local 872 list_for_each_entry(mem, &dev->mem_props, list) { kfd_build_sysfs_node_entry() 873 mem->kobj = kzalloc(sizeof(struct kobject), GFP_KERNEL); kfd_build_sysfs_node_entry() 874 if (!mem->kobj) kfd_build_sysfs_node_entry() 876 ret = kobject_init_and_add(mem->kobj, &mem_type, kfd_build_sysfs_node_entry() 881 mem->attr.name = "properties"; kfd_build_sysfs_node_entry() 882 mem->attr.mode = KFD_SYSFS_FILE_MODE; kfd_build_sysfs_node_entry() 883 sysfs_attr_init(&mem->attr); kfd_build_sysfs_node_entry() 884 ret = sysfs_create_file(mem->kobj, &mem->attr); kfd_build_sysfs_node_entry()
|
/linux-4.1.27/drivers/mfd/ |
H A D | vexpress-sysreg.c | 165 struct resource *mem; vexpress_sysreg_probe() local 171 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); vexpress_sysreg_probe() 172 if (!mem) vexpress_sysreg_probe() 175 base = devm_ioremap(&pdev->dev, mem->start, resource_size(mem)); vexpress_sysreg_probe() 209 ARRAY_SIZE(vexpress_sysreg_cells), mem, 0, NULL); vexpress_sysreg_probe()
|
H A D | jz4740-adc.c | 56 struct resource *mem; member in struct:jz4740_adc 240 adc->mem = request_mem_region(mem_base->start, JZ_REG_ADC_STATUS, jz4740_adc_probe() 242 if (!adc->mem) { jz4740_adc_probe() 247 adc->base = ioremap_nocache(adc->mem->start, resource_size(adc->mem)); jz4740_adc_probe() 299 release_mem_region(adc->mem->start, resource_size(adc->mem)); jz4740_adc_probe() 315 release_mem_region(adc->mem->start, resource_size(adc->mem)); jz4740_adc_remove()
|
/linux-4.1.27/arch/um/kernel/skas/ |
H A D | uaccess.c | 218 int __clear_user(void __user *mem, int len) __clear_user() argument 220 return buffer_op((unsigned long) mem, len, 1, clear_chunk, NULL); __clear_user() 223 int clear_user(void __user *mem, int len) clear_user() argument 226 memset((__force void*)mem, 0, len); clear_user() 230 return access_ok(VERIFY_WRITE, mem, len) ? clear_user() 231 buffer_op((unsigned long) mem, len, 1, clear_chunk, NULL) : len; clear_user()
|
/linux-4.1.27/arch/powerpc/sysdev/ |
H A D | ppc4xx_hsta_msi.c | 129 struct resource *mem; hsta_msi_probe() local 132 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); hsta_msi_probe() 133 if (IS_ERR(mem)) { hsta_msi_probe() 145 ppc4xx_hsta_msi.address = mem->start; hsta_msi_probe() 146 ppc4xx_hsta_msi.data = ioremap(mem->start, resource_size(mem)); hsta_msi_probe()
|
/linux-4.1.27/arch/mips/include/asm/mach-generic/ |
H A D | floppy.h | 118 unsigned long mem; fd_dma_mem_alloc() local 120 mem = __get_dma_pages(GFP_KERNEL, get_order(size)); fd_dma_mem_alloc() 122 return mem; fd_dma_mem_alloc()
|
/linux-4.1.27/drivers/pinctrl/sh-pfc/ |
H A D | gpio.c | 36 struct sh_pfc_window *mem; member in struct:sh_pfc_chip 66 void __iomem *mem = address - chip->mem->phys + chip->mem->virt; gpio_read_data_reg() local 68 return sh_pfc_read_raw_reg(mem, dreg->reg_width); gpio_read_data_reg() 75 void __iomem *mem = address - chip->mem->phys + chip->mem->virt; gpio_write_data_reg() local 77 sh_pfc_write_raw_reg(mem, dreg->reg_width, value); gpio_write_data_reg() 315 struct sh_pfc_window *mem) sh_pfc_add_gpiochip() 324 chip->mem = mem; sh_pfc_add_gpiochip() 314 sh_pfc_add_gpiochip(struct sh_pfc *pfc, int(*setup)(struct sh_pfc_chip *), struct sh_pfc_window *mem) sh_pfc_add_gpiochip() argument
|
/linux-4.1.27/arch/m68k/kernel/ |
H A D | sys_m68k.c | 458 /* This syscall gets its arguments in A0 (mem), D2 (oldval) and 462 unsigned long __user * mem) sys_atomic_cmpxchg_32() 474 pgd = pgd_offset(mm, (unsigned long)mem); sys_atomic_cmpxchg_32() 477 pmd = pmd_offset(pgd, (unsigned long)mem); sys_atomic_cmpxchg_32() 480 pte = pte_offset_map_lock(mm, pmd, (unsigned long)mem, &ptl); sys_atomic_cmpxchg_32() 491 __get_user(mem_value, mem); sys_atomic_cmpxchg_32() 493 __put_user(newval, mem); sys_atomic_cmpxchg_32() 511 if (do_page_fault(fp, (unsigned long)mem, 3)) sys_atomic_cmpxchg_32() 531 /* This syscall gets its arguments in A0 (mem), D2 (oldval) and 535 unsigned long __user * mem) sys_atomic_cmpxchg_32() 542 mem_value = *mem; sys_atomic_cmpxchg_32() 544 *mem = newval; sys_atomic_cmpxchg_32() 461 sys_atomic_cmpxchg_32(unsigned long newval, int oldval, int d3, int d4, int d5, unsigned long __user * mem) sys_atomic_cmpxchg_32() argument 534 sys_atomic_cmpxchg_32(unsigned long newval, int oldval, int d3, int d4, int d5, unsigned long __user * mem) sys_atomic_cmpxchg_32() argument
|
/linux-4.1.27/drivers/mcb/ |
H A D | mcb-core.c | 350 struct resource *mem; mcb_request_mem() local 356 size = resource_size(&dev->mem); mcb_request_mem() 358 mem = request_mem_region(dev->mem.start, size, name); mcb_request_mem() 359 if (!mem) mcb_request_mem() 362 return mem; mcb_request_mem() 372 void mcb_release_mem(struct resource *mem) mcb_release_mem() argument 376 size = resource_size(mem); mcb_release_mem() 377 release_mem_region(mem->start, size); mcb_release_mem()
|
H A D | mcb-parse.c | 70 mdev->mem.start = mapbase + offset; chameleon_parse_gdd() 71 mdev->mem.end = mdev->mem.start + size - 1; chameleon_parse_gdd() 72 mdev->mem.flags = IORESOURCE_MEM; chameleon_parse_gdd()
|
/linux-4.1.27/drivers/media/usb/uvc/ |
H A D | uvc_isight.c | 49 __u8 *mem; isight_decode() local 87 mem = buf->mem + buf->bytesused; isight_decode() 89 memcpy(mem, data, nbytes); isight_decode()
|
/linux-4.1.27/drivers/clk/sirf/ |
H A D | clk-atlas6.c | 63 rtc, osc, pll1, pll2, pll3, mem, sys, security, dsp, gps, enumerator in enum:atlas6_clk_index 145 clk_register_clkdev(atlas6_clks[mem], NULL, "mem"); atlas6_clk_init() 146 clk_register_clkdev(atlas6_clks[mem], NULL, "osc"); atlas6_clk_init()
|
H A D | clk-prima2.c | 62 rtc, osc, pll1, pll2, pll3, mem, sys, security, dsp, gps, enumerator in enum:prima2_clk_index 144 clk_register_clkdev(prima2_clks[mem], NULL, "mem"); prima2_clk_init() 145 clk_register_clkdev(prima2_clks[mem], NULL, "osc"); prima2_clk_init()
|
/linux-4.1.27/drivers/tty/serial/cpm_uart/ |
H A D | cpm_uart.h | 117 u32 mem = (u32)pinfo->mem_addr; cpu2cpm_addr() local 119 if (likely(val >= mem && val < mem + pinfo->mem_size)) { cpu2cpm_addr() 120 offset = val - mem; cpu2cpm_addr()
|
/linux-4.1.27/drivers/video/fbdev/ |
H A D | s3c2410fb.h | 25 struct resource *mem; member in struct:s3c2410fb_info
|
H A D | metronomefb.c | 148 static u8 calc_cksum(int start, int end, u8 *mem) calc_cksum() argument 154 tmp += mem[i]; calc_cksum() 170 static int load_waveform(u8 *mem, size_t size, int m, int t, load_waveform() argument 195 wfm_hdr = (struct waveform_hdr *) mem; load_waveform() 205 cksum = calc_cksum(32, 47, mem); load_waveform() 228 if (mem[i] > t) { load_waveform() 238 cksum = calc_cksum(sizeof(*wfm_hdr), cksum_idx, mem); load_waveform() 239 if (cksum != mem[cksum_idx]) { load_waveform() 241 " %x != %x\n", cksum, mem[cksum_idx]); load_waveform() 250 cksum = calc_cksum(cksum_idx - 3, cksum_idx, mem); load_waveform() 251 if (cksum != mem[cksum_idx]) { load_waveform() 253 " %x != %x\n", cksum, mem[cksum_idx]); load_waveform() 258 tta = get_unaligned_le32(mem + wmta + m * 4) & 0x00FFFFFF; load_waveform() 262 cksum = calc_cksum(cksum_idx - 3, cksum_idx, mem); load_waveform() 263 if (cksum != mem[cksum_idx]) { load_waveform() 265 " %x != %x\n", cksum, mem[cksum_idx]); load_waveform() 271 wfm_idx = get_unaligned_le32(mem + tta + trn * 4) & 0x00FFFFFF; load_waveform() 277 v = mem[wfm_idx++]; load_waveform() 279 while (((v = mem[wfm_idx++]) != wfm_hdr->swtb) && load_waveform() 289 rl = mem[wfm_idx++]; load_waveform() 297 cksum = calc_cksum(owfm_idx, cksum_idx, mem); load_waveform() 298 if (cksum != mem[cksum_idx]) { load_waveform() 300 " %x != %x\n", cksum, mem[cksum_idx]); load_waveform()
|
H A D | nuc900fb.h | 37 struct resource *mem; member in struct:nuc900fb_info
|
/linux-4.1.27/arch/powerpc/mm/ |
H A D | Makefile | 9 obj-y := fault.o mem.o pgtable.o mmap.o \
|
/linux-4.1.27/arch/arm64/mm/ |
H A D | mmap.c | 98 * You really shouldn't be using read() or write() on /dev/mem. This might go 112 * Do not allow /dev/mem mappings beyond the supported physical range. 124 * devmem_is_allowed() checks to see if /dev/mem access to a certain address 127 * This effectively disable read()/write() on /dev/mem.
|
/linux-4.1.27/arch/metag/include/asm/ |
H A D | mmzone.h | 36 /* SoC specific mem init */
|
/linux-4.1.27/arch/arm/plat-iop/ |
H A D | setup.c | 23 { /* mem mapped registers */
|
/linux-4.1.27/drivers/net/can/softing/ |
H A D | softing_fw.c | 109 const uint8_t *mem; fw_parse() local 127 mem = *pmem; fw_parse() 128 *ptype = le16_to_cpup((void *)&mem[0]); fw_parse() 129 *paddr = le32_to_cpup((void *)&mem[2]); fw_parse() 130 *plen = le16_to_cpup((void *)&mem[6]); fw_parse() 131 *pdat = &mem[8]; fw_parse() 133 end = &mem[8 + *plen]; fw_parse() 135 for (checksum[1] = 0; mem < end; ++mem) fw_parse() 136 checksum[1] += *mem; fw_parse() 149 const uint8_t *mem, *end, *dat; softing_load_fw() local 164 mem = fw->data; softing_load_fw() 165 end = &mem[fw->size]; softing_load_fw() 167 ret = fw_parse(&mem, &type, &addr, &len, &dat); softing_load_fw() 177 while (mem < end) { softing_load_fw() 178 ret = fw_parse(&mem, &type, &addr, &len, &dat); softing_load_fw() 232 const uint8_t *mem, *end, *dat; softing_load_app_fw() local 248 mem = fw->data; softing_load_app_fw() 249 end = &mem[fw->size]; softing_load_app_fw() 251 ret = fw_parse(&mem, &type, &addr, &len, &dat); softing_load_app_fw() 266 while (mem < end) { softing_load_app_fw() 267 ret = fw_parse(&mem, &type, &addr, &len, &dat); softing_load_app_fw()
|
/linux-4.1.27/drivers/net/ethernet/8390/ |
H A D | ax88796.c | 119 void __iomem *ioaddr = ei_local->mem; ax_initial_check() 174 void __iomem *nic_base = ei_local->mem; ax_get_8390_hdr() 218 void __iomem *nic_base = ei_local->mem; ax_block_input() 255 void __iomem *nic_base = ei_local->mem; ax_block_output() 392 ei_outb(reg_gpoc, ei_local->mem + EI_SHIFT(0x17)); ax_phy_switch() 527 u8 reg = ei_inb(ei_local->mem + AX_MEMR); ax_eeprom_register_read() 538 u8 reg = ei_inb(ei_local->mem + AX_MEMR); ax_eeprom_register_write() 549 ei_outb(reg, ei_local->mem + AX_MEMR); ax_eeprom_register_write() 633 ax->addr_memr = ei_local->mem + AX_MEMR; ax_mii_init() 670 void __iomem *ioaddr = ei_local->mem; ax_initial_setup() 693 void __iomem *ioaddr = ei_local->mem; ax_init_dev() 743 ei_outb(ax->plat->dcr_val, ei_local->mem + EN0_DCFG); ax_init_dev() 754 ei_local->mem + E8390_CMD); /* 0x61 */ ax_init_dev() 815 struct resource *mem; ax_remove() local 820 iounmap(ei_local->mem); ax_remove() 821 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); ax_remove() 822 release_mem_region(mem->start, resource_size(mem)); ax_remove() 826 mem = platform_get_resource(pdev, IORESOURCE_MEM, 1); ax_remove() 827 release_mem_region(mem->start, resource_size(mem)); ax_remove() 847 struct resource *irq, *mem, *mem2; ax_probe() local 876 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); ax_probe() 877 if (!mem) { ax_probe() 883 mem_size = resource_size(mem); ax_probe() 897 if (!request_mem_region(mem->start, mem_size, pdev->name)) { ax_probe() 903 ei_local->mem = ioremap(mem->start, mem_size); ax_probe() 904 dev->base_addr = (unsigned long)ei_local->mem; ax_probe() 906 if (ei_local->mem == NULL) { ax_probe() 907 dev_err(&pdev->dev, "Cannot ioremap area %pR\n", mem); ax_probe() 936 ei_local->reg_offset[0x1f] = ax->map2 - ei_local->mem; ax_probe() 953 iounmap(ei_local->mem); ax_probe() 956 release_mem_region(mem->start, mem_size); ax_probe()
|
H A D | mcf8390.c | 412 struct resource *mem, *irq; mcf8390_probe() local 422 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); mcf8390_probe() 423 if (mem == NULL) { mcf8390_probe() 427 msize = resource_size(mem); mcf8390_probe() 428 if (!request_mem_region(mem->start, msize, pdev->name)) mcf8390_probe() 433 release_mem_region(mem->start, msize); mcf8390_probe() 443 dev->base_addr = mem->start; mcf8390_probe() 447 release_mem_region(mem->start, msize); mcf8390_probe() 457 struct resource *mem; mcf8390_remove() local 460 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); mcf8390_remove() 461 if (mem) mcf8390_remove() 462 release_mem_region(mem->start, resource_size(mem)); mcf8390_remove()
|
/linux-4.1.27/arch/tile/gxio/ |
H A D | mpipe.c | 163 buffer_size_enum, void *mem, size_t mem_size, gxio_mpipe_init_buffer_stack() 168 memset(mem, 0, mem_size); gxio_mpipe_init_buffer_stack() 170 result = gxio_mpipe_init_buffer_stack_aux(context, mem, mem_size, gxio_mpipe_init_buffer_stack() 186 void *mem, size_t mem_size, gxio_mpipe_init_notif_ring() 189 return gxio_mpipe_init_notif_ring_aux(context, mem, mem_size, gxio_mpipe_init_notif_ring() 237 void *mem, size_t mem_size, gxio_mpipe_init_edma_ring() 240 memset(mem, 0, mem_size); gxio_mpipe_init_edma_ring() 242 return gxio_mpipe_init_edma_ring_aux(context, mem, mem_size, mem_flags, gxio_mpipe_init_edma_ring() 392 void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_iqueue_init() 398 iqueue->idescs = (gxio_mpipe_idesc_t *)mem; gxio_mpipe_iqueue_init() 409 __gxio_mmio_write(mem, iqueue->head); gxio_mpipe_iqueue_init() 411 return gxio_mpipe_init_notif_ring(context, ring, mem, mem_size, gxio_mpipe_iqueue_init() 421 void *mem, unsigned int mem_size, gxio_mpipe_equeue_init() 431 mem, mem_size, mem_flags); gxio_mpipe_equeue_init() 446 equeue->edescs = mem; gxio_mpipe_equeue_init() 160 gxio_mpipe_init_buffer_stack(gxio_mpipe_context_t *context, unsigned int stack, gxio_mpipe_buffer_size_enum_t buffer_size_enum, void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_init_buffer_stack() argument 184 gxio_mpipe_init_notif_ring(gxio_mpipe_context_t *context, unsigned int ring, void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_init_notif_ring() argument 235 gxio_mpipe_init_edma_ring(gxio_mpipe_context_t *context, unsigned int ring, unsigned int channel, void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_init_edma_ring() argument 389 gxio_mpipe_iqueue_init(gxio_mpipe_iqueue_t *iqueue, gxio_mpipe_context_t *context, unsigned int ring, void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_iqueue_init() argument 417 gxio_mpipe_equeue_init(gxio_mpipe_equeue_t *equeue, gxio_mpipe_context_t *context, unsigned int ering, unsigned int channel, void *mem, unsigned int mem_size, unsigned int mem_flags) gxio_mpipe_equeue_init() argument
|
/linux-4.1.27/arch/tile/include/asm/ |
H A D | uaccess.h | 527 * @mem: Destination address, in user space. 535 extern unsigned long clear_user_asm(void __user *mem, unsigned long len); __clear_user() 537 void __user *mem, unsigned long len) __clear_user() 540 return clear_user_asm(mem, len); __clear_user() 543 void __user *mem, unsigned long len) clear_user() 545 if (access_ok(VERIFY_WRITE, mem, len)) clear_user() 546 return __clear_user(mem, len); clear_user() 552 * @mem: Destination address, in user space. 558 extern unsigned long flush_user_asm(void __user *mem, unsigned long len); __flush_user() 560 void __user *mem, unsigned long len) __flush_user() 565 retval = flush_user_asm(mem, len); __flush_user() 571 void __user *mem, unsigned long len) flush_user() 573 if (access_ok(VERIFY_WRITE, mem, len)) flush_user() 574 return __flush_user(mem, len); flush_user() 580 * @mem: Destination address, in user space. 586 extern unsigned long finv_user_asm(void __user *mem, unsigned long len); __finv_user() 588 void __user *mem, unsigned long len) __finv_user() 593 retval = finv_user_asm(mem, len); __finv_user() 598 void __user *mem, unsigned long len) finv_user() 600 if (access_ok(VERIFY_WRITE, mem, len)) finv_user() 601 return __finv_user(mem, len); finv_user() 536 __clear_user( void __user *mem, unsigned long len) __clear_user() argument 542 clear_user( void __user *mem, unsigned long len) clear_user() argument 559 __flush_user( void __user *mem, unsigned long len) __flush_user() argument 570 flush_user( void __user *mem, unsigned long len) flush_user() argument 587 __finv_user( void __user *mem, unsigned long len) __finv_user() argument 597 finv_user( void __user *mem, unsigned long len) finv_user() argument
|
/linux-4.1.27/arch/unicore32/include/asm/ |
H A D | io.h | 57 * devmem_is_allowed() checks to see if /dev/mem access to a certain 61 * on /dev/mem.
|
/linux-4.1.27/arch/mips/include/asm/dec/ |
H A D | ecc.h | 35 #define KN0X_ESR_CHKHI (0x7f<<24) /* check bits read from mem */ 39 #define KN0X_ESR_CHKLO (0x7f<<8) /* check bits read from mem */
|
/linux-4.1.27/drivers/usb/isp1760/ |
H A D | isp1760-hcd.h | 73 struct resource *mem, int irq, unsigned long irqflags, 81 void __iomem *regs, struct resource *mem, isp1760_hcd_register() 80 isp1760_hcd_register(struct isp1760_hcd *priv, void __iomem *regs, struct resource *mem, int irq, unsigned long irqflags, struct device *dev) isp1760_hcd_register() argument
|
/linux-4.1.27/include/video/ |
H A D | sh_mobile_meram.h | 44 unsigned long mem, size_t size); 65 unsigned long mem, size_t size) sh_mobile_meram_free() 64 sh_mobile_meram_free(struct sh_mobile_meram_info *meram_dev, unsigned long mem, size_t size) sh_mobile_meram_free() argument
|
/linux-4.1.27/arch/arm64/kernel/ |
H A D | arm64ksyms.c | 35 /* user mem (segment) */ 44 /* string / mem functions */
|
/linux-4.1.27/arch/arc/lib/ |
H A D | memset.S | 51 ; memzero: @r0 = mem, @r1 = size_t 52 ; memset: @r0 = mem, @r1 = char, @r2 = size_t
|
/linux-4.1.27/sound/soc/tegra/ |
H A D | tegra20_spdif.c | 268 struct resource *mem, *memregion, *dmareq; tegra20_spdif_platform_probe() local 288 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); tegra20_spdif_platform_probe() 289 if (!mem) { tegra20_spdif_platform_probe() 302 memregion = devm_request_mem_region(&pdev->dev, mem->start, tegra20_spdif_platform_probe() 303 resource_size(mem), DRV_NAME); tegra20_spdif_platform_probe() 310 regs = devm_ioremap(&pdev->dev, mem->start, resource_size(mem)); tegra20_spdif_platform_probe() 325 spdif->playback_dma_data.addr = mem->start + TEGRA20_SPDIF_DATA_OUT; tegra20_spdif_platform_probe()
|
H A D | tegra20_i2s.c | 342 struct resource *mem, *memregion; tegra20_i2s_platform_probe() local 364 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); tegra20_i2s_platform_probe() 365 if (!mem) { tegra20_i2s_platform_probe() 371 memregion = devm_request_mem_region(&pdev->dev, mem->start, tegra20_i2s_platform_probe() 372 resource_size(mem), DRV_NAME); tegra20_i2s_platform_probe() 379 regs = devm_ioremap(&pdev->dev, mem->start, resource_size(mem)); tegra20_i2s_platform_probe() 394 i2s->capture_dma_data.addr = mem->start + TEGRA20_I2S_FIFO2; tegra20_i2s_platform_probe() 398 i2s->playback_dma_data.addr = mem->start + TEGRA20_I2S_FIFO1; tegra20_i2s_platform_probe()
|
/linux-4.1.27/drivers/media/common/saa7146/ |
H A D | saa7146_core.c | 179 void *mem = vmalloc_32(length); saa7146_vmalloc_build_pgtable() local 182 if (NULL == mem) saa7146_vmalloc_build_pgtable() 185 if (!(pt->slist = vmalloc_to_sg(mem, pages))) saa7146_vmalloc_build_pgtable() 199 return mem; saa7146_vmalloc_build_pgtable() 209 vfree(mem); saa7146_vmalloc_build_pgtable() 214 void saa7146_vfree_destroy_pgtable(struct pci_dev *pci, void *mem, struct saa7146_pgtable *pt) saa7146_vfree_destroy_pgtable() argument 220 vfree(mem); saa7146_vfree_destroy_pgtable() 360 /* clear out mem for sure */ saa7146_init_one() 392 dev->mem = ioremap(pci_resource_start(pci, 0), saa7146_init_one() 394 if (!dev->mem) { saa7146_init_one() 444 pr_info("found saa7146 @ mem %p (revision %d, irq %d) (0x%04x,0x%04x)\n", saa7146_init_one() 445 dev->mem, dev->revision, pci->irq, saa7146_init_one() 497 iounmap(dev->mem); saa7146_init_one() 536 iounmap(dev->mem); saa7146_remove_one()
|
/linux-4.1.27/drivers/gpu/host1x/hw/ |
H A D | channel_hw.c | 36 void *mem = NULL; trace_write_gather() local 39 mem = host1x_bo_mmap(bo); trace_write_gather() 41 if (mem) { trace_write_gather() 53 mem); trace_write_gather() local 56 host1x_bo_munmap(bo, mem); trace_write_gather()
|
/linux-4.1.27/arch/nios2/kernel/ |
H A D | kgdb.c | 84 char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs) dbg_get_reg() argument 90 memcpy(mem, (void *)regs + dbg_reg_def[regno].offset, dbg_get_reg() 93 memset(mem, 0, dbg_reg_def[regno].size); dbg_get_reg() 98 int dbg_set_reg(int regno, void *mem, struct pt_regs *regs) dbg_set_reg() argument 104 memcpy((void *)regs + dbg_reg_def[regno].offset, mem, dbg_set_reg()
|
/linux-4.1.27/arch/arm/boot/compressed/ |
H A D | atags_to_fdt.c | 153 if (!atag->u.mem.size) for_each_tag() 164 cpu_to_fdt64(atag->u.mem.start); for_each_tag() 166 cpu_to_fdt64(atag->u.mem.size); for_each_tag() 169 cpu_to_fdt32(atag->u.mem.start); for_each_tag() 171 cpu_to_fdt32(atag->u.mem.size); for_each_tag()
|
/linux-4.1.27/sound/oss/ |
H A D | kahlua.c | 59 void __iomem *mem; probe_one() local 70 mem = ioremap(base, 128); probe_one() 71 if (!mem) probe_one() 73 map = readw(mem + 0x18); /* Read the SMI enables */ probe_one() 74 iounmap(mem); probe_one()
|
/linux-4.1.27/drivers/media/pci/cx23885/ |
H A D | cimax2.c | 160 int mem; netup_ci_get_mem() local 164 mem = cx_read(MC417_RWD); netup_ci_get_mem() 165 if ((mem & NETUP_ACK) == 0) netup_ci_get_mem() 174 return mem & 0xff; netup_ci_get_mem() 185 int mem; netup_ci_op_cam() local 228 mem = netup_ci_get_mem(dev); netup_ci_op_cam() 233 if (mem < 0) netup_ci_op_cam() 238 (flag == NETUP_CI_CTL) ? "ctl" : "mem", netup_ci_op_cam() 239 (read) ? mem : data); netup_ci_op_cam() 242 return mem; netup_ci_op_cam()
|
/linux-4.1.27/arch/s390/kvm/ |
H A D | priv.c | 475 static void handle_stsi_3_2_2(struct kvm_vcpu *vcpu, struct sysinfo_3_2_2 *mem) handle_stsi_3_2_2() argument 483 if (stsi(mem, 3, 2, 2)) handle_stsi_3_2_2() 484 mem->count = 0; handle_stsi_3_2_2() 485 if (mem->count < 8) handle_stsi_3_2_2() 486 mem->count++; handle_stsi_3_2_2() 487 for (n = mem->count - 1; n > 0 ; n--) handle_stsi_3_2_2() 488 memcpy(&mem->vm[n], &mem->vm[n - 1], sizeof(mem->vm[0])); handle_stsi_3_2_2() 490 memset(&mem->vm[0], 0, sizeof(mem->vm[0])); handle_stsi_3_2_2() 491 mem->vm[0].cpus_total = cpus; handle_stsi_3_2_2() 492 mem->vm[0].cpus_configured = cpus; handle_stsi_3_2_2() 493 mem->vm[0].cpus_standby = 0; handle_stsi_3_2_2() 494 mem->vm[0].cpus_reserved = 0; handle_stsi_3_2_2() 495 mem->vm[0].caf = 1000; handle_stsi_3_2_2() 496 memcpy(mem->vm[0].name, "KVMguest", 8); handle_stsi_3_2_2() 497 ASCEBC(mem->vm[0].name, 8); handle_stsi_3_2_2() 498 memcpy(mem->vm[0].cpi, "KVM/Linux ", 16); handle_stsi_3_2_2() 499 ASCEBC(mem->vm[0].cpi, 16); handle_stsi_3_2_2() 518 unsigned long mem = 0; handle_stsi() local 552 mem = get_zeroed_page(GFP_KERNEL); handle_stsi() 553 if (!mem) handle_stsi() 555 if (stsi((void *) mem, fc, sel1, sel2)) handle_stsi() 561 mem = get_zeroed_page(GFP_KERNEL); handle_stsi() 562 if (!mem) handle_stsi() 564 handle_stsi_3_2_2(vcpu, (void *) mem); handle_stsi() 568 rc = write_guest(vcpu, operand2, ar, (void *)mem, PAGE_SIZE); handle_stsi() 578 free_page(mem); handle_stsi() 585 free_page(mem); handle_stsi()
|
/linux-4.1.27/drivers/infiniband/hw/ocrdma/ |
H A D | ocrdma_stats.c | 54 struct stats_mem *mem = &dev->stats_mem; ocrdma_alloc_stats_mem() local 56 /* Alloc mbox command mem*/ ocrdma_alloc_stats_mem() 57 mem->size = max_t(u32, sizeof(struct ocrdma_rdma_stats_req), ocrdma_alloc_stats_mem() 60 mem->va = dma_alloc_coherent(&dev->nic_info.pdev->dev, mem->size, ocrdma_alloc_stats_mem() 61 &mem->pa, GFP_KERNEL); ocrdma_alloc_stats_mem() 62 if (!mem->va) { ocrdma_alloc_stats_mem() 67 memset(mem->va, 0, mem->size); ocrdma_alloc_stats_mem() 69 /* Alloc debugfs mem */ ocrdma_alloc_stats_mem() 70 mem->debugfs_mem = kzalloc(OCRDMA_MAX_DBGFS_MEM, GFP_KERNEL); ocrdma_alloc_stats_mem() 71 if (!mem->debugfs_mem) { ocrdma_alloc_stats_mem() 72 pr_err("%s: stats debugfs mem allocation failed\n", __func__); ocrdma_alloc_stats_mem() 81 struct stats_mem *mem = &dev->stats_mem; ocrdma_release_stats_mem() local 83 if (mem->va) ocrdma_release_stats_mem() 84 dma_free_coherent(&dev->nic_info.pdev->dev, mem->size, ocrdma_release_stats_mem() 85 mem->va, mem->pa); ocrdma_release_stats_mem() 86 kfree(mem->debugfs_mem); ocrdma_release_stats_mem()
|
/linux-4.1.27/drivers/gpu/host1x/ |
H A D | job.c | 41 void *mem; host1x_job_alloc() local 54 mem = job = kzalloc(total, GFP_KERNEL); host1x_job_alloc() 62 mem += sizeof(struct host1x_job); host1x_job_alloc() 63 job->relocarray = num_relocs ? mem : NULL; host1x_job_alloc() 64 mem += num_relocs * sizeof(struct host1x_reloc); host1x_job_alloc() 65 job->unpins = num_unpins ? mem : NULL; host1x_job_alloc() 66 mem += num_unpins * sizeof(struct host1x_job_unpin_data); host1x_job_alloc() 67 job->waitchk = num_waitchks ? mem : NULL; host1x_job_alloc() 68 mem += num_waitchks * sizeof(struct host1x_waitchk); host1x_job_alloc() 69 job->gathers = num_cmdbufs ? mem : NULL; host1x_job_alloc() 70 mem += num_cmdbufs * sizeof(struct host1x_job_gather); host1x_job_alloc() 71 job->addr_phys = num_unpins ? mem : NULL; host1x_job_alloc() 535 /* process each gather mem only once */ host1x_job_pin()
|
/linux-4.1.27/arch/mips/kernel/ |
H A D | kgdb.c | 130 int dbg_set_reg(int regno, void *mem, struct pt_regs *regs) dbg_set_reg() argument 138 memcpy((void *)regs + dbg_reg_def[regno].offset, mem, dbg_set_reg() 146 memcpy((void *)¤t->thread.fpu.fcr31, mem, dbg_set_reg() 154 memcpy((void *)¤t->thread.fpu.fpr[fp_reg], mem, dbg_set_reg() 163 char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs) dbg_get_reg() argument 172 memcpy(mem, (void *)regs + dbg_reg_def[regno].offset, dbg_get_reg() 181 memcpy(mem, (void *)¤t->thread.fpu.fcr31, dbg_get_reg() 186 memset(mem, 0, dbg_reg_def[regno].size); dbg_get_reg() 190 memcpy(mem, (void *)¤t->thread.fpu.fpr[fp_reg], dbg_get_reg()
|
/linux-4.1.27/drivers/isdn/hardware/eicon/ |
H A D | pc_maint.h | 131 POINTER_32BIT mem; member in struct:pc_maint 143 POINTER_32BIT mem; member in struct:pc_maint 155 POINTER_32BIT mem; member in struct:mi_pc_maint
|
/linux-4.1.27/drivers/media/usb/go7007/ |
H A D | go7007-fw.c | 376 u16 mem = 0x3e00; gen_mjpeghdr_to_package() local 398 code[off + 1] = __cpu_to_le16(0x8000 | mem); gen_mjpeghdr_to_package() 401 if (mem + chunk > 0x4000) gen_mjpeghdr_to_package() 402 chunk = 0x4000 - mem; gen_mjpeghdr_to_package() 409 mem = 0x3e00; gen_mjpeghdr_to_package() 413 mem += 28; gen_mjpeghdr_to_package() 643 u16 mem = 0x3e00; gen_mpeg1hdr_to_package() local 687 code[off + 1] = __cpu_to_le16(0x8000 | mem); gen_mpeg1hdr_to_package() 690 if (mem + chunk > 0x4000) gen_mpeg1hdr_to_package() 691 chunk = 0x4000 - mem; gen_mpeg1hdr_to_package() 698 if (mem + chunk == 0x4000) { gen_mpeg1hdr_to_package() 699 mem = 0x3e00; gen_mpeg1hdr_to_package() 705 mem += 28; gen_mpeg1hdr_to_package() 829 u16 mem = 0x3e00; gen_mpeg4hdr_to_package() local 855 code[off + 1] = __cpu_to_le16(0x8000 | mem); gen_mpeg4hdr_to_package() 858 if (mem + chunk > 0x4000) gen_mpeg4hdr_to_package() 859 chunk = 0x4000 - mem; gen_mpeg4hdr_to_package() 866 if (mem + chunk == 0x4000) { gen_mpeg4hdr_to_package() 867 mem = 0x3e00; gen_mpeg4hdr_to_package() 873 mem += 28; gen_mpeg4hdr_to_package() 879 mem = 0x3e00; gen_mpeg4hdr_to_package() 898 code[off + 1] = __cpu_to_le16(0x8000 | mem); gen_mpeg4hdr_to_package() 901 if (mem + chunk > 0x4000) gen_mpeg4hdr_to_package() 902 chunk = 0x4000 - mem; gen_mpeg4hdr_to_package() 909 if (mem + chunk == 0x4000) { gen_mpeg4hdr_to_package() 910 mem = 0x3e00; gen_mpeg4hdr_to_package() 916 mem += 28; gen_mpeg4hdr_to_package()
|
/linux-4.1.27/drivers/net/wireless/prism54/ |
H A D | islpci_mgt.c | 123 if (buf->mem == NULL) { islpci_mgmt_rx_fill() 124 buf->mem = kmalloc(MGMT_FRAME_SIZE, GFP_ATOMIC); islpci_mgmt_rx_fill() 125 if (!buf->mem) islpci_mgmt_rx_fill() 130 buf->pci_addr = pci_map_single(priv->pdev, buf->mem, islpci_mgmt_rx_fill() 189 p = buf.mem = kmalloc(frag_len, GFP_KERNEL); islpci_mgt_transmit() 190 if (!buf.mem) islpci_mgt_transmit() 206 pimfor_header_t *h = buf.mem; islpci_mgt_transmit() 218 buf.pci_addr = pci_map_single(priv->pdev, buf.mem, frag_len, islpci_mgt_transmit() 257 kfree(buf.mem); islpci_mgt_transmit() 321 header = pimfor_decode_header(buf->mem, frag_len); islpci_mgt_receive() 432 kfree(buf->mem); islpci_mgt_cleanup_transmit() 433 buf->mem = NULL; islpci_mgt_cleanup_transmit()
|
/linux-4.1.27/include/drm/ttm/ |
H A D | ttm_bo_driver.h | 188 * @mem: Pointer to a struct ttm_mem_reg to be filled in. 193 * @mem::mm_node should be set to a non-null value, and 194 * @mem::start should be set to a value identifying the beginning 196 * If the memory region accommodate the buffer object, @mem::mm_node 201 * Note that @mem::mm_node will only be dereferenced by 212 struct ttm_mem_reg *mem); 218 * @mem: Pointer to a struct ttm_mem_reg to be filled in. 221 * and that are identified by @mem::mm_node and @mem::start. May not 225 struct ttm_mem_reg *mem); 259 * @io_reserve_lru: Optional lru list for unreserving io mem regions. 378 * finished, they'll end up in bo->mem.flags 435 int (*io_mem_reserve)(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem); 436 void (*io_mem_free)(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem); 667 * @mem: A valid struct ttm_mem_reg. 669 * Returns true if the memory described by @mem is PCI memory, 673 struct ttm_mem_reg *mem); 681 * @mem: A struct ttm_mem_reg. 686 * the placement flags in @mem, potentially evicting other idle buffer objects. 696 struct ttm_mem_reg *mem, 701 struct ttm_mem_reg *mem); 703 struct ttm_mem_reg *mem); 917 if (!(bo->mem.placement & TTM_PL_FLAG_NO_EVICT)) { ttm_bo_unreserve() 943 struct ttm_mem_reg *mem); 945 struct ttm_mem_reg *mem); 957 * and update the (@bo)->mem placement flags. If unsuccessful, the old 979 * and update the (@bo)->mem placement flags. If unsuccessful, the old
|
/linux-4.1.27/drivers/scsi/mpt3sas/ |
H A D | mpt3sas_config.c | 181 * @mem: struct config_request 189 struct config_request *mem) _config_alloc_config_dma_memory() 193 if (mem->sz > ioc->config_page_sz) { _config_alloc_config_dma_memory() 194 mem->page = dma_alloc_coherent(&ioc->pdev->dev, mem->sz, _config_alloc_config_dma_memory() 195 &mem->page_dma, GFP_KERNEL); _config_alloc_config_dma_memory() 196 if (!mem->page) { _config_alloc_config_dma_memory() 199 ioc->name, __func__, mem->sz); _config_alloc_config_dma_memory() 203 mem->page = ioc->config_page; _config_alloc_config_dma_memory() 204 mem->page_dma = ioc->config_page_dma; _config_alloc_config_dma_memory() 212 * @mem: struct config_request 220 struct config_request *mem) _config_free_config_dma_memory() 222 if (mem->sz > ioc->config_page_sz) _config_free_config_dma_memory() 223 dma_free_coherent(&ioc->pdev->dev, mem->sz, mem->page, _config_free_config_dma_memory() 224 mem->page_dma); _config_free_config_dma_memory() 297 struct config_request mem; _config_request() local 309 memset(&mem, 0, sizeof(struct config_request)); _config_request() 322 mem.sz = mpi_request->Header.PageLength * 4; _config_request() 324 mem.sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; _config_request() 325 r = _config_alloc_config_dma_memory(ioc, &mem); _config_request() 333 MPT3_CONFIG_COMMON_WRITE_SGLFLAGS | mem.sz, _config_request() 334 mem.page_dma); _config_request() 335 memcpy(mem.page, config_page, min_t(u16, mem.sz, _config_request() 340 MPT3_CONFIG_COMMON_SGLFLAGS | mem.sz, mem.page_dma); _config_request() 341 memset(mem.page, 0, min_t(u16, mem.sz, config_page_sz)); _config_request() 452 u8 *p = (u8 *)mem.page; _config_request() 460 _debug_dump_config(p, min_t(u16, mem.sz, _config_request() 477 _debug_dump_config(p, min_t(u16, mem.sz, _config_request() 488 memcpy(config_page, mem.page, min_t(u16, mem.sz, _config_request() 494 _config_free_config_dma_memory(ioc, &mem); _config_request() 188 _config_alloc_config_dma_memory(struct MPT3SAS_ADAPTER *ioc, struct config_request *mem) _config_alloc_config_dma_memory() argument 219 _config_free_config_dma_memory(struct MPT3SAS_ADAPTER *ioc, struct config_request *mem) _config_free_config_dma_memory() argument
|
/linux-4.1.27/sound/soc/codecs/ |
H A D | wm_adsp.c | 337 if (dsp->mem[i].type == type) wm_adsp_find_region() 338 return &dsp->mem[i]; wm_adsp_find_region() 380 const struct wm_adsp_region *mem; wm_coeff_write_control() local 386 mem = wm_adsp_find_region(adsp, region->type); wm_coeff_write_control() 387 if (!mem) { wm_coeff_write_control() 394 reg = wm_adsp_region_to_reg(mem, reg); wm_coeff_write_control() 435 const struct wm_adsp_region *mem; wm_coeff_read_control() local 441 mem = wm_adsp_find_region(adsp, region->type); wm_coeff_read_control() 442 if (!mem) { wm_coeff_read_control() 449 reg = wm_adsp_region_to_reg(mem, reg); wm_coeff_read_control() 534 const struct wm_adsp_region *mem; wm_adsp_load() local 633 mem = wm_adsp_find_region(dsp, type); wm_adsp_load() 652 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load() 656 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load() 660 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load() 664 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load() 668 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load() 888 const struct wm_adsp_region *mem; wm_adsp_setup_algs() local 896 mem = wm_adsp_find_region(dsp, WMFW_ADSP1_DM); wm_adsp_setup_algs() 899 mem = wm_adsp_find_region(dsp, WMFW_ADSP2_XM); wm_adsp_setup_algs() 902 mem = NULL; wm_adsp_setup_algs() 906 if (WARN_ON(!mem)) wm_adsp_setup_algs() 911 ret = regmap_raw_read(regmap, mem->base, &adsp1_id, wm_adsp_setup_algs() 952 ret = regmap_raw_read(regmap, mem->base, &adsp2_id, wm_adsp_setup_algs() 1018 ret = regmap_raw_read(regmap, mem->base + term, &val, sizeof(val)); wm_adsp_setup_algs() 1033 ret = regmap_raw_read(regmap, mem->base + pos, alg, (term - pos) * 2); wm_adsp_setup_algs() 1181 const struct wm_adsp_region *mem; wm_adsp_load_coeff() local 1263 mem = wm_adsp_find_region(dsp, type); wm_adsp_load_coeff() 1264 if (!mem) { wm_adsp_load_coeff() 1268 reg = wm_adsp_region_to_reg(mem, 0); wm_adsp_load_coeff() 1284 mem = wm_adsp_find_region(dsp, type); wm_adsp_load_coeff() 1285 if (!mem) { wm_adsp_load_coeff() 1296 reg = wm_adsp_region_to_reg(mem, wm_adsp_load_coeff()
|
/linux-4.1.27/drivers/net/appletalk/ |
H A D | ltpc.h | 65 struct lt_mem mem; member in union:lt_command
|
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_minidump.c | 235 struct __mem mem; member in union:qlcnic_dump_entry::__anon7208 564 struct __mem *rom = &entry->region.mem; qlcnic_read_rom() 662 struct __mem *mem, __le32 *buffer, qlcnic_read_memory_test_agent() 668 reg_read = mem->size; qlcnic_read_memory_test_agent() 669 addr = mem->addr; qlcnic_read_memory_test_agent() 709 return mem->size; qlcnic_read_memory_test_agent() 721 struct __mem *mem) qlcnic_start_pex_dma() 734 ret = qlcnic_ind_wr(adapter, temp_addr, mem->desc_card_addr); qlcnic_start_pex_dma() 744 ret = qlcnic_ind_wr(adapter, temp_addr, mem->start_dma_cmd); qlcnic_start_pex_dma() 768 struct __mem *mem, qlcnic_read_memory_pexdma() 803 * dma_desc_cmd 16:19 = mem->dma_desc_cmd 0:3 qlcnic_read_memory_pexdma() 805 * dma_desc_cmd 24:31 = mem->dma_desc_cmd 8:15 qlcnic_read_memory_pexdma() 810 temp = mem->dma_desc_cmd & 0xff0f; qlcnic_read_memory_pexdma() 818 while (read_size < mem->size) { qlcnic_read_memory_pexdma() 819 if (mem->size - read_size >= QLC_PEX_DMA_READ_SIZE) qlcnic_read_memory_pexdma() 822 size = mem->size - read_size; qlcnic_read_memory_pexdma() 824 dma_descr->src_addr_low = mem->addr + read_size; qlcnic_read_memory_pexdma() 829 *ret = qlcnic_ms_mem_write128(adapter, mem->desc_card_addr, qlcnic_read_memory_pexdma() 833 mem->desc_card_addr); qlcnic_read_memory_pexdma() 837 *ret = qlcnic_start_pex_dma(adapter, mem); qlcnic_read_memory_pexdma() 859 struct __mem *mem = &entry->region.mem; qlcnic_read_memory() local 864 data_size = qlcnic_read_memory_pexdma(adapter, mem, buffer, qlcnic_read_memory() 874 data_size = qlcnic_read_memory_test_agent(adapter, mem, buffer, &ret); qlcnic_read_memory() 1021 struct __mem *rom = &entry->region.mem; qlcnic_83xx_dump_rom() 661 qlcnic_read_memory_test_agent(struct qlcnic_adapter *adapter, struct __mem *mem, __le32 *buffer, int *ret) qlcnic_read_memory_test_agent() argument 720 qlcnic_start_pex_dma(struct qlcnic_adapter *adapter, struct __mem *mem) qlcnic_start_pex_dma() argument 767 qlcnic_read_memory_pexdma(struct qlcnic_adapter *adapter, struct __mem *mem, __le32 *buffer, int *ret) qlcnic_read_memory_pexdma() argument
|
/linux-4.1.27/drivers/media/platform/omap3isp/ |
H A D | ispcsiphy.h | 32 /* mem resources - enums as defined in enum isp_mem_resources */
|
/linux-4.1.27/arch/sparc/include/asm/ |
H A D | string_32.h | 22 /* First the mem*() things. */
|
H A D | string_64.h | 20 /* First the mem*() things. */
|
/linux-4.1.27/arch/um/kernel/ |
H A D | Makefile | 13 obj-y = config.o exec.o exitcode.o irq.o ksyms.o mem.o \
|
/linux-4.1.27/drivers/char/ |
H A D | Makefile | 5 obj-y += mem.o random.o
|
/linux-4.1.27/arch/sh/include/asm/ |
H A D | mmzone.h | 37 /* Platform specific mem init */
|