Searched refs:mem (Results 1 - 200 of 1569) sorted by relevance

12345678

/linux-4.1.27/sound/synth/
H A DMakefile6 snd-util-mem-objs := util_mem.o
9 obj-$(CONFIG_SND_EMU10K1) += snd-util-mem.o
10 obj-$(CONFIG_SND_TRIDENT) += snd-util-mem.o
11 obj-$(CONFIG_SND_SBAWE_SEQ) += snd-util-mem.o
/linux-4.1.27/tools/testing/selftests/memory-hotplug/
H A DMakefile5 TEST_PROGS := mem-on-off-test.sh
6 override RUN_TESTS := ./mem-on-off-test.sh -r 2 || echo "selftests: memory-hotplug [FAIL]"
10 @/bin/bash ./mem-on-off-test.sh || echo "memory-hotplug selftests: [FAIL]"
/linux-4.1.27/drivers/infiniband/hw/mlx5/
H A DMakefile3 mlx5_ib-y := main.o cq.o doorbell.o qp.o mem.o srq.o mr.o ah.o mad.o
/linux-4.1.27/drivers/base/
H A Ddma-coherent.c22 struct dma_coherent_mem **mem) dma_init_coherent_memory()
52 *mem = dma_mem; dma_init_coherent_memory()
66 static void dma_release_coherent_memory(struct dma_coherent_mem *mem) dma_release_coherent_memory() argument
68 if (!mem) dma_release_coherent_memory()
70 iounmap(mem->virt_base); dma_release_coherent_memory()
71 kfree(mem->bitmap); dma_release_coherent_memory()
72 kfree(mem); dma_release_coherent_memory()
76 struct dma_coherent_mem *mem) dma_assign_coherent_memory()
81 dev->dma_mem = mem; dma_assign_coherent_memory()
90 struct dma_coherent_mem *mem; dma_declare_coherent_memory() local
94 &mem); dma_declare_coherent_memory()
98 if (dma_assign_coherent_memory(dev, mem) == 0) dma_declare_coherent_memory()
101 dma_release_coherent_memory(mem); dma_declare_coherent_memory()
108 struct dma_coherent_mem *mem = dev->dma_mem; dma_release_declared_memory() local
110 if (!mem) dma_release_declared_memory()
112 dma_release_coherent_memory(mem); dma_release_declared_memory()
120 struct dma_coherent_mem *mem = dev->dma_mem; dma_mark_declared_memory_occupied() local
126 if (!mem) dma_mark_declared_memory_occupied()
129 spin_lock_irqsave(&mem->spinlock, flags); dma_mark_declared_memory_occupied()
130 pos = (device_addr - mem->device_base) >> PAGE_SHIFT; dma_mark_declared_memory_occupied()
131 err = bitmap_allocate_region(mem->bitmap, pos, get_order(size)); dma_mark_declared_memory_occupied()
132 spin_unlock_irqrestore(&mem->spinlock, flags); dma_mark_declared_memory_occupied()
136 return mem->virt_base + (pos << PAGE_SHIFT); dma_mark_declared_memory_occupied()
158 struct dma_coherent_mem *mem; dma_alloc_from_coherent() local
165 mem = dev->dma_mem; dma_alloc_from_coherent()
166 if (!mem) dma_alloc_from_coherent()
170 spin_lock_irqsave(&mem->spinlock, flags); dma_alloc_from_coherent()
172 if (unlikely(size > (mem->size << PAGE_SHIFT))) dma_alloc_from_coherent()
175 pageno = bitmap_find_free_region(mem->bitmap, mem->size, order); dma_alloc_from_coherent()
182 *dma_handle = mem->device_base + (pageno << PAGE_SHIFT); dma_alloc_from_coherent()
183 *ret = mem->virt_base + (pageno << PAGE_SHIFT); dma_alloc_from_coherent()
185 spin_unlock_irqrestore(&mem->spinlock, flags); dma_alloc_from_coherent()
190 spin_unlock_irqrestore(&mem->spinlock, flags); dma_alloc_from_coherent()
196 return mem->flags & DMA_MEMORY_EXCLUSIVE; dma_alloc_from_coherent()
215 struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL; dma_release_from_coherent() local
217 if (mem && vaddr >= mem->virt_base && vaddr < dma_release_from_coherent()
218 (mem->virt_base + (mem->size << PAGE_SHIFT))) { dma_release_from_coherent()
219 int page = (vaddr - mem->virt_base) >> PAGE_SHIFT; dma_release_from_coherent()
222 spin_lock_irqsave(&mem->spinlock, flags); dma_release_from_coherent()
223 bitmap_release_region(mem->bitmap, page, order); dma_release_from_coherent()
224 spin_unlock_irqrestore(&mem->spinlock, flags); dma_release_from_coherent()
249 struct dma_coherent_mem *mem = dev ? dev->dma_mem : NULL; dma_mmap_from_coherent() local
251 if (mem && vaddr >= mem->virt_base && vaddr + size <= dma_mmap_from_coherent()
252 (mem->virt_base + (mem->size << PAGE_SHIFT))) { dma_mmap_from_coherent()
254 int start = (vaddr - mem->virt_base) >> PAGE_SHIFT; dma_mmap_from_coherent()
260 unsigned long pfn = mem->pfn_base + start + off; dma_mmap_from_coherent()
281 struct dma_coherent_mem *mem = rmem->priv; rmem_dma_device_init() local
283 if (!mem && rmem_dma_device_init()
286 &mem) != DMA_MEMORY_MAP) { rmem_dma_device_init()
291 rmem->priv = mem; rmem_dma_device_init()
292 dma_assign_coherent_memory(dev, mem); rmem_dma_device_init()
20 dma_init_coherent_memory(phys_addr_t phys_addr, dma_addr_t device_addr, size_t size, int flags, struct dma_coherent_mem **mem) dma_init_coherent_memory() argument
75 dma_assign_coherent_memory(struct device *dev, struct dma_coherent_mem *mem) dma_assign_coherent_memory() argument
H A Dmemory.c81 struct memory_block *mem = to_memory_block(dev); memory_block_release() local
83 kfree(mem); memory_block_release()
114 struct memory_block *mem = to_memory_block(dev); show_mem_start_phys_index() local
117 phys_index = mem->start_section_nr / sections_per_block; show_mem_start_phys_index()
129 struct memory_block *mem = to_memory_block(dev); show_mem_removable() local
132 if (!present_section_nr(mem->start_section_nr + i)) show_mem_removable()
134 pfn = section_nr_to_pfn(mem->start_section_nr + i); show_mem_removable()
147 struct memory_block *mem = to_memory_block(dev); show_mem_state() local
154 switch (mem->state) { show_mem_state()
166 mem->state); show_mem_state()
254 static int memory_block_change_state(struct memory_block *mem, memory_block_change_state() argument
259 if (mem->state != from_state_req) memory_block_change_state()
263 mem->state = MEM_GOING_OFFLINE; memory_block_change_state()
265 ret = memory_block_action(mem->start_section_nr, to_state, memory_block_change_state()
266 mem->online_type); memory_block_change_state()
268 mem->state = ret ? from_state_req : to_state; memory_block_change_state()
276 struct memory_block *mem = to_memory_block(dev); memory_subsys_online() local
279 if (mem->state == MEM_ONLINE) memory_subsys_online()
287 if (mem->online_type < 0) memory_subsys_online()
288 mem->online_type = MMOP_ONLINE_KEEP; memory_subsys_online()
291 ret = memory_block_change_state(mem, MEM_ONLINE, MEM_OFFLINE); memory_subsys_online()
294 mem->online_type = -1; memory_subsys_online()
301 struct memory_block *mem = to_memory_block(dev); memory_subsys_offline() local
303 if (mem->state == MEM_OFFLINE) memory_subsys_offline()
306 return memory_block_change_state(mem, MEM_OFFLINE, MEM_ONLINE); memory_subsys_offline()
313 struct memory_block *mem = to_memory_block(dev); store_mem_state() local
346 mem->online_type = online_type; store_mem_state()
347 ret = device_online(&mem->dev); store_mem_state()
350 ret = device_offline(&mem->dev); store_mem_state()
377 struct memory_block *mem = to_memory_block(dev); show_phys_device() local
378 return sprintf(buf, "%d\n", mem->phys_device); show_phys_device()
385 struct memory_block *mem = to_memory_block(dev); show_valid_zones() local
391 start_pfn = section_nr_to_pfn(mem->start_section_nr); show_valid_zones()
402 /*The mem block is the last memoryblock of this zone.*/ show_valid_zones()
409 /*The mem block is the first memoryblock of ZONE_MOVABLE.*/ show_valid_zones()
603 struct memory_block *mem; init_memory_block() local
608 mem = kzalloc(sizeof(*mem), GFP_KERNEL); init_memory_block()
609 if (!mem) init_memory_block()
613 mem->start_section_nr = init_memory_block()
615 mem->end_section_nr = mem->start_section_nr + sections_per_block - 1; init_memory_block()
616 mem->state = state; init_memory_block()
617 mem->section_count++; init_memory_block()
618 start_pfn = section_nr_to_pfn(mem->start_section_nr); init_memory_block()
619 mem->phys_device = arch_get_memory_phys_device(start_pfn); init_memory_block()
621 ret = register_memory(mem); init_memory_block()
623 *memory = mem; init_memory_block()
629 struct memory_block *mem; add_memory_block() local
644 ret = init_memory_block(&mem, __nr_to_section(section_nr), MEM_ONLINE); add_memory_block()
647 mem->section_count = section_count; add_memory_block()
659 struct memory_block *mem; register_new_memory() local
663 mem = find_memory_block(section); register_new_memory()
664 if (mem) { register_new_memory()
665 mem->section_count++; register_new_memory()
666 put_device(&mem->dev); register_new_memory()
668 ret = init_memory_block(&mem, section, MEM_OFFLINE); register_new_memory()
673 if (mem->section_count == sections_per_block) register_new_memory()
674 ret = register_mem_sect_under_node(mem, nid); register_new_memory()
694 struct memory_block *mem; remove_memory_block() local
697 mem = find_memory_block(section); remove_memory_block()
698 unregister_mem_sect_under_nodes(mem, __section_nr(section)); remove_memory_block()
700 mem->section_count--; remove_memory_block()
701 if (mem->section_count == 0) remove_memory_block()
702 unregister_memory(mem); remove_memory_block()
704 put_device(&mem->dev); remove_memory_block()
720 bool is_memblock_offlined(struct memory_block *mem) is_memblock_offlined() argument
722 return mem->state == MEM_OFFLINE; is_memblock_offlined()
/linux-4.1.27/include/linux/
H A Dkbuild.h9 #define OFFSET(sym, str, mem) \
10 DEFINE(sym, offsetof(struct str, mem))
H A Dtrace_seq.h86 extern void trace_seq_putmem(struct trace_seq *s, const void *mem, unsigned int len);
87 extern void trace_seq_putmem_hex(struct trace_seq *s, const void *mem,
125 trace_seq_putmem(struct trace_seq *s, const void *mem, unsigned int len) trace_seq_putmem() argument
128 static inline void trace_seq_putmem_hex(struct trace_seq *s, const void *mem, trace_seq_putmem_hex() argument
/linux-4.1.27/drivers/media/v4l2-core/
H A Dvideobuf-dma-contig.c41 struct videobuf_dma_contig_memory *mem, __videobuf_dc_alloc()
44 mem->size = size; __videobuf_dc_alloc()
45 mem->vaddr = dma_alloc_coherent(dev, mem->size, __videobuf_dc_alloc()
46 &mem->dma_handle, flags); __videobuf_dc_alloc()
48 if (!mem->vaddr) { __videobuf_dc_alloc()
49 dev_err(dev, "memory alloc size %ld failed\n", mem->size); __videobuf_dc_alloc()
53 dev_dbg(dev, "dma mapped data is at %p (%ld)\n", mem->vaddr, mem->size); __videobuf_dc_alloc()
59 struct videobuf_dma_contig_memory *mem) __videobuf_dc_free()
61 dma_free_coherent(dev, mem->size, mem->vaddr, mem->dma_handle); __videobuf_dc_free()
63 mem->vaddr = NULL; __videobuf_dc_free()
87 struct videobuf_dma_contig_memory *mem; videobuf_vm_close() local
103 mem = q->bufs[i]->priv; videobuf_vm_close()
104 if (mem) { videobuf_vm_close()
111 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); videobuf_vm_close()
117 i, mem->vaddr); videobuf_vm_close()
119 __videobuf_dc_free(q->dev, mem); videobuf_vm_close()
120 mem->vaddr = NULL; videobuf_vm_close()
140 * @mem: per-buffer private videobuf-dma-contig data
144 static void videobuf_dma_contig_user_put(struct videobuf_dma_contig_memory *mem) videobuf_dma_contig_user_put() argument
146 mem->dma_handle = 0; videobuf_dma_contig_user_put()
147 mem->size = 0; videobuf_dma_contig_user_put()
152 * @mem: per-buffer private videobuf-dma-contig data
160 static int videobuf_dma_contig_user_get(struct videobuf_dma_contig_memory *mem, videobuf_dma_contig_user_get() argument
171 mem->size = PAGE_ALIGN(vb->size + offset); videobuf_dma_contig_user_get()
180 if ((vb->baddr + mem->size) > vma->vm_end) videobuf_dma_contig_user_get()
187 while (pages_done < (mem->size >> PAGE_SHIFT)) { videobuf_dma_contig_user_get()
193 mem->dma_handle = (this_pfn << PAGE_SHIFT) + offset; videobuf_dma_contig_user_get()
213 struct videobuf_dma_contig_memory *mem; __videobuf_alloc() local
216 vb = kzalloc(size + sizeof(*mem), GFP_KERNEL); __videobuf_alloc()
219 mem = vb->priv; __videobuf_alloc()
220 mem->magic = MAGIC_DC_MEM; __videobuf_alloc()
228 struct videobuf_dma_contig_memory *mem = buf->priv; __videobuf_to_vaddr() local
230 BUG_ON(!mem); __videobuf_to_vaddr()
231 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); __videobuf_to_vaddr()
233 return mem->vaddr; __videobuf_to_vaddr()
240 struct videobuf_dma_contig_memory *mem = vb->priv; __videobuf_iolock() local
242 BUG_ON(!mem); __videobuf_iolock()
243 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); __videobuf_iolock()
250 if (!mem->vaddr) { __videobuf_iolock()
260 return videobuf_dma_contig_user_get(mem, vb); __videobuf_iolock()
263 if (__videobuf_dc_alloc(q->dev, mem, PAGE_ALIGN(vb->size), __videobuf_iolock()
280 struct videobuf_dma_contig_memory *mem; __videobuf_mmap_mapper() local
297 mem = buf->priv; __videobuf_mmap_mapper()
298 BUG_ON(!mem); __videobuf_mmap_mapper()
299 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); __videobuf_mmap_mapper()
301 if (__videobuf_dc_alloc(q->dev, mem, PAGE_ALIGN(buf->bsize), __videobuf_mmap_mapper()
317 retval = vm_iomap_memory(vma, mem->dma_handle, size); __videobuf_mmap_mapper()
321 dma_free_coherent(q->dev, mem->size, __videobuf_mmap_mapper()
322 mem->vaddr, mem->dma_handle); __videobuf_mmap_mapper()
368 struct videobuf_dma_contig_memory *mem = buf->priv; videobuf_to_dma_contig() local
370 BUG_ON(!mem); videobuf_to_dma_contig()
371 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); videobuf_to_dma_contig()
373 return mem->dma_handle; videobuf_to_dma_contig()
380 struct videobuf_dma_contig_memory *mem = buf->priv; videobuf_dma_contig_free() local
391 if (!mem) videobuf_dma_contig_free()
394 MAGIC_CHECK(mem->magic, MAGIC_DC_MEM); videobuf_dma_contig_free()
398 videobuf_dma_contig_user_put(mem); videobuf_dma_contig_free()
403 if (mem->vaddr) { videobuf_dma_contig_free()
404 __videobuf_dc_free(q->dev, mem); videobuf_dma_contig_free()
405 mem->vaddr = NULL; videobuf_dma_contig_free()
40 __videobuf_dc_alloc(struct device *dev, struct videobuf_dma_contig_memory *mem, unsigned long size, gfp_t flags) __videobuf_dc_alloc() argument
58 __videobuf_dc_free(struct device *dev, struct videobuf_dma_contig_memory *mem) __videobuf_dc_free() argument
H A Dvideobuf-vmalloc.c75 struct videobuf_vmalloc_memory *mem; videobuf_vm_close() local
91 mem = q->bufs[i]->priv; videobuf_vm_close()
92 if (mem) { videobuf_vm_close()
99 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); videobuf_vm_close()
105 __func__, i, mem->vaddr); videobuf_vm_close()
107 vfree(mem->vaddr); videobuf_vm_close()
108 mem->vaddr = NULL; videobuf_vm_close()
140 struct videobuf_vmalloc_memory *mem; __videobuf_alloc_vb() local
143 vb = kzalloc(size + sizeof(*mem), GFP_KERNEL); __videobuf_alloc_vb()
147 mem = vb->priv = ((char *)vb) + size; __videobuf_alloc_vb()
148 mem->magic = MAGIC_VMAL_MEM; __videobuf_alloc_vb()
152 mem, (long)sizeof(*mem)); __videobuf_alloc_vb()
161 struct videobuf_vmalloc_memory *mem = vb->priv; __videobuf_iolock() local
164 BUG_ON(!mem); __videobuf_iolock()
166 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); __videobuf_iolock()
173 if (!mem->vaddr) { __videobuf_iolock()
192 mem->vaddr = vmalloc_user(pages); __videobuf_iolock()
193 if (!mem->vaddr) { __videobuf_iolock()
198 mem->vaddr, pages); __videobuf_iolock()
209 The code below won't work, since mem->vma = NULL __videobuf_iolock()
212 rc = remap_vmalloc_range(mem->vma, (void *)vb->baddr, 0); __videobuf_iolock()
236 struct videobuf_vmalloc_memory *mem; __videobuf_mmap_mapper() local
252 mem = buf->priv; __videobuf_mmap_mapper()
253 BUG_ON(!mem); __videobuf_mmap_mapper()
254 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); __videobuf_mmap_mapper()
257 mem->vaddr = vmalloc_user(pages); __videobuf_mmap_mapper()
258 if (!mem->vaddr) { __videobuf_mmap_mapper()
262 dprintk(1, "vmalloc is at addr %p (%d pages)\n", mem->vaddr, pages); __videobuf_mmap_mapper()
265 retval = remap_vmalloc_range(vma, mem->vaddr, 0); __videobuf_mmap_mapper()
268 vfree(mem->vaddr); __videobuf_mmap_mapper()
286 mem = NULL; __videobuf_mmap_mapper()
317 struct videobuf_vmalloc_memory *mem = buf->priv; videobuf_to_vmalloc() local
318 BUG_ON(!mem); videobuf_to_vmalloc()
319 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); videobuf_to_vmalloc()
321 return mem->vaddr; videobuf_to_vmalloc()
327 struct videobuf_vmalloc_memory *mem = buf->priv; videobuf_vmalloc_free() local
338 if (!mem) videobuf_vmalloc_free()
341 MAGIC_CHECK(mem->magic, MAGIC_VMAL_MEM); videobuf_vmalloc_free()
343 vfree(mem->vaddr); videobuf_vmalloc_free()
344 mem->vaddr = NULL; videobuf_vmalloc_free()
H A Dvideobuf-dma-sg.c139 struct videobuf_dma_sg_memory *mem = buf->priv; videobuf_to_dma() local
140 BUG_ON(!mem); videobuf_to_dma()
142 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); videobuf_to_dma()
144 return &mem->dma; videobuf_to_dma()
398 struct videobuf_dma_sg_memory *mem; videobuf_vm_close() local
411 mem = q->bufs[i]->priv; videobuf_vm_close()
412 if (!mem) videobuf_vm_close()
415 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); videobuf_vm_close()
470 struct videobuf_dma_sg_memory *mem; __videobuf_alloc_vb() local
473 vb = kzalloc(size + sizeof(*mem), GFP_KERNEL); __videobuf_alloc_vb()
477 mem = vb->priv = ((char *)vb) + size; __videobuf_alloc_vb()
478 mem->magic = MAGIC_SG_MEM; __videobuf_alloc_vb()
480 videobuf_dma_init(&mem->dma); __videobuf_alloc_vb()
484 mem, (long)sizeof(*mem)); __videobuf_alloc_vb()
491 struct videobuf_dma_sg_memory *mem = buf->priv; __videobuf_to_vaddr() local
492 BUG_ON(!mem); __videobuf_to_vaddr()
494 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); __videobuf_to_vaddr()
496 return mem->dma.vaddr; __videobuf_to_vaddr()
505 struct videobuf_dma_sg_memory *mem = vb->priv; __videobuf_iolock() local
506 BUG_ON(!mem); __videobuf_iolock()
508 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); __videobuf_iolock()
510 if (!mem->dma.dev) __videobuf_iolock()
511 mem->dma.dev = q->dev; __videobuf_iolock()
513 WARN_ON(mem->dma.dev != q->dev); __videobuf_iolock()
521 err = videobuf_dma_init_kernel(&mem->dma, __videobuf_iolock()
528 err = videobuf_dma_init_user(&mem->dma, __videobuf_iolock()
539 err = videobuf_dma_init_user_locked(&mem->dma, __videobuf_iolock()
557 err = videobuf_dma_init_overlay(&mem->dma, DMA_FROM_DEVICE, __videobuf_iolock()
565 err = videobuf_dma_map(q->dev, &mem->dma); __videobuf_iolock()
575 struct videobuf_dma_sg_memory *mem = buf->priv; __videobuf_sync() local
576 BUG_ON(!mem || !mem->dma.sglen); __videobuf_sync()
578 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); __videobuf_sync()
579 MAGIC_CHECK(mem->dma.magic, MAGIC_DMABUF); __videobuf_sync()
581 dma_sync_sg_for_cpu(q->dev, mem->dma.sglist, __videobuf_sync()
582 mem->dma.sglen, mem->dma.direction); __videobuf_sync()
591 struct videobuf_dma_sg_memory *mem = buf->priv; __videobuf_mmap_mapper() local
598 BUG_ON(!mem); __videobuf_mmap_mapper()
599 MAGIC_CHECK(mem->magic, MAGIC_SG_MEM); __videobuf_mmap_mapper()
/linux-4.1.27/drivers/infiniband/hw/cxgb4/
H A DMakefile5 iw_cxgb4-y := device.o cm.o provider.o mem.o cq.o qp.o resource.o ev.o id_table.o
/linux-4.1.27/drivers/uio/
H A Duio_mf624.c48 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; mf624_disable_interrupt()
76 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; mf624_enable_interrupt()
103 void __iomem *INTCSR_reg = info->mem[0].internal_addr + INTCSR; mf624_irq_handler()
150 info->mem[0].name = "PCI chipset, interrupts, status " mf624_pci_probe()
152 info->mem[0].addr = pci_resource_start(dev, 0); mf624_pci_probe()
153 if (!info->mem[0].addr) mf624_pci_probe()
155 info->mem[0].size = pci_resource_len(dev, 0); mf624_pci_probe()
156 info->mem[0].memtype = UIO_MEM_PHYS; mf624_pci_probe()
157 info->mem[0].internal_addr = pci_ioremap_bar(dev, 0); mf624_pci_probe()
158 if (!info->mem[0].internal_addr) mf624_pci_probe()
162 info->mem[1].name = "ADC, DAC, DIO"; mf624_pci_probe()
163 info->mem[1].addr = pci_resource_start(dev, 2); mf624_pci_probe()
164 if (!info->mem[1].addr) mf624_pci_probe()
166 info->mem[1].size = pci_resource_len(dev, 2); mf624_pci_probe()
167 info->mem[1].memtype = UIO_MEM_PHYS; mf624_pci_probe()
168 info->mem[1].internal_addr = pci_ioremap_bar(dev, 2); mf624_pci_probe()
169 if (!info->mem[1].internal_addr) mf624_pci_probe()
173 info->mem[2].name = "Counter/timer chip"; mf624_pci_probe()
174 info->mem[2].addr = pci_resource_start(dev, 4); mf624_pci_probe()
175 if (!info->mem[2].addr) mf624_pci_probe()
177 info->mem[2].size = pci_resource_len(dev, 4); mf624_pci_probe()
178 info->mem[2].memtype = UIO_MEM_PHYS; mf624_pci_probe()
179 info->mem[2].internal_addr = pci_ioremap_bar(dev, 4); mf624_pci_probe()
180 if (!info->mem[2].internal_addr) mf624_pci_probe()
197 iounmap(info->mem[2].internal_addr); mf624_pci_probe()
199 iounmap(info->mem[1].internal_addr); mf624_pci_probe()
201 iounmap(info->mem[0].internal_addr); mf624_pci_probe()
224 iounmap(info->mem[0].internal_addr); mf624_pci_remove()
225 iounmap(info->mem[1].internal_addr); mf624_pci_remove()
226 iounmap(info->mem[2].internal_addr); mf624_pci_remove()
H A Duio_cif.c31 void __iomem *plx_intscr = dev_info->mem[0].internal_addr hilscher_handler()
58 info->mem[0].addr = pci_resource_start(dev, 0); hilscher_pci_probe()
59 if (!info->mem[0].addr) hilscher_pci_probe()
61 info->mem[0].internal_addr = pci_ioremap_bar(dev, 0); hilscher_pci_probe()
62 if (!info->mem[0].internal_addr) hilscher_pci_probe()
65 info->mem[0].size = pci_resource_len(dev, 0); hilscher_pci_probe()
66 info->mem[0].memtype = UIO_MEM_PHYS; hilscher_pci_probe()
67 info->mem[1].addr = pci_resource_start(dev, 2); hilscher_pci_probe()
68 info->mem[1].size = pci_resource_len(dev, 2); hilscher_pci_probe()
69 info->mem[1].memtype = UIO_MEM_PHYS; hilscher_pci_probe()
92 iounmap(info->mem[0].internal_addr); hilscher_pci_probe()
109 iounmap(info->mem[0].internal_addr); hilscher_pci_remove()
H A Duio_netx.c35 void __iomem *int_enable_reg = dev_info->mem[0].internal_addr netx_handler()
37 void __iomem *int_status_reg = dev_info->mem[0].internal_addr netx_handler()
82 info->mem[0].addr = pci_resource_start(dev, bar); netx_pci_probe()
83 if (!info->mem[0].addr) netx_pci_probe()
85 info->mem[0].internal_addr = ioremap(pci_resource_start(dev, bar), netx_pci_probe()
88 if (!info->mem[0].internal_addr) netx_pci_probe()
91 info->mem[0].size = pci_resource_len(dev, bar); netx_pci_probe()
92 info->mem[0].memtype = UIO_MEM_PHYS; netx_pci_probe()
99 iowrite32(0, info->mem[0].internal_addr + DPM_HOST_INT_EN0); netx_pci_probe()
111 iounmap(info->mem[0].internal_addr); netx_pci_probe()
126 iowrite32(0, info->mem[0].internal_addr + DPM_HOST_INT_EN0); netx_pci_remove()
130 iounmap(info->mem[0].internal_addr); netx_pci_remove()
H A Duio_sercos3.c54 void __iomem *ier0 = info->mem[3].internal_addr + IER0_OFFSET; sercos3_disable_interrupts()
67 void __iomem *ier0 = info->mem[3].internal_addr + IER0_OFFSET; sercos3_enable_interrupts()
77 void __iomem *isr0 = info->mem[3].internal_addr + ISR0_OFFSET; sercos3_handler()
78 void __iomem *ier0 = info->mem[3].internal_addr + IER0_OFFSET; sercos3_handler()
107 info->mem[n].addr = pci_resource_start(dev, pci_bar); sercos3_setup_iomem()
108 if (!info->mem[n].addr) sercos3_setup_iomem()
110 info->mem[n].internal_addr = ioremap(pci_resource_start(dev, pci_bar), sercos3_setup_iomem()
112 if (!info->mem[n].internal_addr) sercos3_setup_iomem()
114 info->mem[n].size = pci_resource_len(dev, pci_bar); sercos3_setup_iomem()
115 info->mem[n].memtype = UIO_MEM_PHYS; sercos3_setup_iomem()
170 if (info->mem[i].internal_addr) sercos3_pci_probe()
171 iounmap(info->mem[i].internal_addr); sercos3_pci_probe()
192 if (info->mem[i].internal_addr) sercos3_pci_remove()
193 iounmap(info->mem[i].internal_addr); sercos3_pci_remove()
H A Duio.c45 struct uio_mem *mem; member in struct:uio_map
49 static ssize_t map_name_show(struct uio_mem *mem, char *buf) map_name_show() argument
51 if (unlikely(!mem->name)) map_name_show()
52 mem->name = ""; map_name_show()
54 return sprintf(buf, "%s\n", mem->name); map_name_show()
57 static ssize_t map_addr_show(struct uio_mem *mem, char *buf) map_addr_show() argument
59 return sprintf(buf, "%pa\n", &mem->addr); map_addr_show()
62 static ssize_t map_size_show(struct uio_mem *mem, char *buf) map_size_show() argument
64 return sprintf(buf, "%pa\n", &mem->size); map_size_show()
67 static ssize_t map_offset_show(struct uio_mem *mem, char *buf) map_offset_show() argument
69 return sprintf(buf, "0x%llx\n", (unsigned long long)mem->addr & ~PAGE_MASK); map_offset_show()
105 struct uio_mem *mem = map->mem; map_type_show() local
113 return entry->show(mem, buf); map_type_show()
261 struct uio_mem *mem; uio_dev_add_attributes() local
267 mem = &idev->info->mem[mi]; uio_dev_add_attributes()
268 if (mem->size == 0) uio_dev_add_attributes()
281 map->mem = mem; uio_dev_add_attributes()
282 mem->map = map; uio_dev_add_attributes()
332 mem = &idev->info->mem[mi]; uio_dev_add_attributes()
333 map = mem->map; uio_dev_add_attributes()
344 struct uio_mem *mem; uio_dev_del_attributes() local
348 mem = &idev->info->mem[i]; uio_dev_del_attributes()
349 if (mem->size == 0) uio_dev_del_attributes()
351 kobject_put(&mem->map->kobj); uio_dev_del_attributes()
584 if (idev->info->mem[vma->vm_pgoff].size == 0) uio_find_mem_index()
604 * to use mem[N]. uio_vma_fault()
608 addr = (void *)(unsigned long)idev->info->mem[mi].addr + offset; uio_vma_fault()
609 if (idev->info->mem[mi].memtype == UIO_MEM_LOGICAL) uio_vma_fault()
639 struct uio_mem *mem; uio_mmap_physical() local
642 mem = idev->info->mem + mi; uio_mmap_physical()
644 if (mem->addr & ~PAGE_MASK) uio_mmap_physical()
646 if (vma->vm_end - vma->vm_start > mem->size) uio_mmap_physical()
663 mem->addr >> PAGE_SHIFT, uio_mmap_physical()
686 actual_pages = ((idev->info->mem[mi].addr & ~PAGE_MASK) uio_mmap()
687 + idev->info->mem[mi].size + PAGE_SIZE -1) >> PAGE_SHIFT; uio_mmap()
696 switch (idev->info->mem[mi].memtype) { uio_mmap()
H A Duio_pruss.c188 p->mem[0].addr = regs_prussio->start; pruss_probe()
189 p->mem[0].size = resource_size(regs_prussio); pruss_probe()
190 p->mem[0].memtype = UIO_MEM_PHYS; pruss_probe()
192 p->mem[1].addr = gdev->sram_paddr; pruss_probe()
193 p->mem[1].size = sram_pool_sz; pruss_probe()
194 p->mem[1].memtype = UIO_MEM_PHYS; pruss_probe()
196 p->mem[2].addr = gdev->ddr_paddr; pruss_probe()
197 p->mem[2].size = extram_pool_sz; pruss_probe()
198 p->mem[2].memtype = UIO_MEM_PHYS; pruss_probe()
H A Duio_dmem_genirq.c53 uiomem = &priv->uioinfo->mem[priv->dmem_region_start]; uio_dmem_genirq_open()
56 while (!priv->refcnt && uiomem < &priv->uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_open()
86 uiomem = &priv->uioinfo->mem[priv->dmem_region_start]; uio_dmem_genirq_release()
91 while (!priv->refcnt && uiomem < &priv->uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_release()
211 uiomem = &uioinfo->mem[0]; uio_dmem_genirq_probe()
219 if (uiomem >= &uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_probe()
236 if (uiomem >= &uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_probe()
248 while (uiomem < &uioinfo->mem[MAX_UIO_MAPS]) { uio_dmem_genirq_probe()
H A Duio_fsl_elbc_gpcm.c152 void __iomem *reg_int_en = info->mem[0].internal_addr + netx5152_irq_handler()
155 void __iomem *reg_int_stat = info->mem[0].internal_addr + netx5152_irq_handler()
184 iowrite32(0, info->mem[0].internal_addr + win0_offset + netx5152_init()
194 iowrite32(0, info->mem[0].internal_addr + DPM_HOST_WIN0_OFFSET + netx5152_shutdown()
376 info->mem[0].internal_addr = ioremap(res.start, resource_size(&res)); uio_fsl_elbc_gpcm_probe()
377 if (!info->mem[0].internal_addr) { uio_fsl_elbc_gpcm_probe()
385 info->mem[0].name = kstrdup(node->name, GFP_KERNEL); uio_fsl_elbc_gpcm_probe()
386 info->mem[0].addr = res.start; uio_fsl_elbc_gpcm_probe()
387 info->mem[0].size = resource_size(&res); uio_fsl_elbc_gpcm_probe()
388 info->mem[0].memtype = UIO_MEM_PHYS; uio_fsl_elbc_gpcm_probe()
438 iounmap(info->mem[0].internal_addr); uio_fsl_elbc_gpcm_probe()
440 kfree(info->mem[0].name); uio_fsl_elbc_gpcm_probe()
459 iounmap(info->mem[0].internal_addr); uio_fsl_elbc_gpcm_remove()
460 kfree(info->mem[0].name); uio_fsl_elbc_gpcm_remove()
/linux-4.1.27/mm/
H A Dcma_debug.c71 static void cma_add_to_cma_mem_list(struct cma *cma, struct cma_mem *mem) cma_add_to_cma_mem_list() argument
74 hlist_add_head(&mem->node, &cma->mem_head); cma_add_to_cma_mem_list()
80 struct cma_mem *mem = NULL; cma_get_entry_from_list() local
84 mem = hlist_entry(cma->mem_head.first, struct cma_mem, node); cma_get_entry_from_list()
85 hlist_del_init(&mem->node); cma_get_entry_from_list()
89 return mem; cma_get_entry_from_list()
94 struct cma_mem *mem = NULL; cma_free_mem() local
97 mem = cma_get_entry_from_list(cma); cma_free_mem()
98 if (mem == NULL) cma_free_mem()
101 if (mem->n <= count) { cma_free_mem()
102 cma_release(cma, mem->p, mem->n); cma_free_mem()
103 count -= mem->n; cma_free_mem()
104 kfree(mem); cma_free_mem()
106 cma_release(cma, mem->p, count); cma_free_mem()
107 mem->p += count; cma_free_mem()
108 mem->n -= count; cma_free_mem()
110 cma_add_to_cma_mem_list(cma, mem); cma_free_mem()
113 cma_add_to_cma_mem_list(cma, mem); cma_free_mem()
133 struct cma_mem *mem; cma_alloc_mem() local
136 mem = kzalloc(sizeof(*mem), GFP_KERNEL); cma_alloc_mem()
137 if (!mem) cma_alloc_mem()
142 kfree(mem); cma_alloc_mem()
146 mem->p = p; cma_alloc_mem()
147 mem->n = count; cma_alloc_mem()
149 cma_add_to_cma_mem_list(cma, mem); cma_alloc_mem()
H A Ddebug-pagealloc.c80 static void check_poison_mem(unsigned char *mem, size_t bytes) check_poison_mem() argument
86 start = memchr_inv(mem, PAGE_POISON, bytes); check_poison_mem()
90 for (end = mem + bytes - 1; end > start; end--) { check_poison_mem()
/linux-4.1.27/arch/um/os-Linux/skas/
H A DMakefile6 obj-y := mem.o process.o
/linux-4.1.27/arch/alpha/include/uapi/asm/
H A Dcompiler.h93 #define __kernel_ldbu(mem) (mem)
94 #define __kernel_ldwu(mem) (mem)
95 #define __kernel_stb(val,mem) ((mem) = (val))
96 #define __kernel_stw(val,mem) ((mem) = (val))
98 #define __kernel_ldbu(mem) \
101 ldbu %0,%1" : "=r"(__kir) : "m"(mem)); \
103 #define __kernel_ldwu(mem) \
106 ldwu %0,%1" : "=r"(__kir) : "m"(mem)); \
108 #define __kernel_stb(val,mem) \
110 stb %1,%0" : "=m"(mem) : "r"(val))
111 #define __kernel_stw(val,mem) \
113 stw %1,%0" : "=m"(mem) : "r"(val))
/linux-4.1.27/tools/perf/bench/
H A Dmem-memcpy-arch.h7 #include "mem-memcpy-x86-64-asm-def.h"
H A Dmem-memset-arch.h7 #include "mem-memset-x86-64-asm-def.h"
H A Dnuma.c202 "perf bench numa mem [<options>]",
660 printf("# NOTE: %d tasks mem-bound, %d tasks unbound\n", t, g->p.nr_tasks - t); parse_setup_node_list()
1066 printf("# thread %2d / %2d global mem: %p, process mem: %p, thread mem: %p\n", worker_thread()
1237 printf(" # process %2d global mem: %p, process mem: %p\n", worker_process()
1274 printf(" # %5dx %5ldMB global shared mem operations\n", print_summary()
1276 printf(" # %5dx %5ldMB process shared mem operations\n", print_summary()
1278 printf(" # %5dx %5ldMB thread local mem operations\n", print_summary()
1645 { "RAM-bw-local,", "mem", "-p", "1", "-t", "1", "-P", "1024",
1648 "mem", "-p", "1", "-t", "1", "-P", "1024",
1650 { "RAM-bw-remote,", "mem", "-p", "1", "-t", "1", "-P", "1024",
1654 { "RAM-bw-local-2x,", "mem", "-p", "2", "-t", "1", "-P", "1024",
1656 { "RAM-bw-remote-2x,", "mem", "-p", "2", "-t", "1", "-P", "1024",
1660 { "RAM-bw-cross,", "mem", "-p", "2", "-t", "1", "-P", "1024",
1664 { " 1x3-convergence,", "mem", "-p", "1", "-t", "3", "-P", "512", OPT_CONV },
1665 { " 1x4-convergence,", "mem", "-p", "1", "-t", "4", "-P", "512", OPT_CONV },
1666 { " 1x6-convergence,", "mem", "-p", "1", "-t", "6", "-P", "1020", OPT_CONV },
1667 { " 2x3-convergence,", "mem", "-p", "3", "-t", "3", "-P", "1020", OPT_CONV },
1668 { " 3x3-convergence,", "mem", "-p", "3", "-t", "3", "-P", "1020", OPT_CONV },
1669 { " 4x4-convergence,", "mem", "-p", "4", "-t", "4", "-P", "512", OPT_CONV },
1671 "mem", "-p", "4", "-t", "4", "-P", "512", OPT_CONV_NOTHP },
1672 { " 4x6-convergence,", "mem", "-p", "4", "-t", "6", "-P", "1020", OPT_CONV },
1673 { " 4x8-convergence,", "mem", "-p", "4", "-t", "8", "-P", "512", OPT_CONV },
1674 { " 8x4-convergence,", "mem", "-p", "8", "-t", "4", "-P", "512", OPT_CONV },
1676 "mem", "-p", "8", "-t", "4", "-P", "512", OPT_CONV_NOTHP },
1677 { " 3x1-convergence,", "mem", "-p", "3", "-t", "1", "-P", "512", OPT_CONV },
1678 { " 4x1-convergence,", "mem", "-p", "4", "-t", "1", "-P", "512", OPT_CONV },
1679 { " 8x1-convergence,", "mem", "-p", "8", "-t", "1", "-P", "512", OPT_CONV },
1680 { "16x1-convergence,", "mem", "-p", "16", "-t", "1", "-P", "256", OPT_CONV },
1681 { "32x1-convergence,", "mem", "-p", "32", "-t", "1", "-P", "128", OPT_CONV },
1684 { " 2x1-bw-process,", "mem", "-p", "2", "-t", "1", "-P", "1024", OPT_BW },
1685 { " 3x1-bw-process,", "mem", "-p", "3", "-t", "1", "-P", "1024", OPT_BW },
1686 { " 4x1-bw-process,", "mem", "-p", "4", "-t", "1", "-P", "1024", OPT_BW },
1687 { " 8x1-bw-process,", "mem", "-p", "8", "-t", "1", "-P", " 512", OPT_BW },
1689 "mem", "-p", "8", "-t", "1", "-P", " 512", OPT_BW_NOTHP },
1690 { "16x1-bw-process,", "mem", "-p", "16", "-t", "1", "-P", "256", OPT_BW },
1692 { " 4x1-bw-thread,", "mem", "-p", "1", "-t", "4", "-T", "256", OPT_BW },
1693 { " 8x1-bw-thread,", "mem", "-p", "1", "-t", "8", "-T", "256", OPT_BW },
1694 { "16x1-bw-thread,", "mem", "-p", "1", "-t", "16", "-T", "128", OPT_BW },
1695 { "32x1-bw-thread,", "mem", "-p", "1", "-t", "32", "-T", "64", OPT_BW },
1697 { " 2x3-bw-thread,", "mem", "-p", "2", "-t", "3", "-P", "512", OPT_BW },
1698 { " 4x4-bw-thread,", "mem", "-p", "4", "-t", "4", "-P", "512", OPT_BW },
1699 { " 4x6-bw-thread,", "mem", "-p", "4", "-t", "6", "-P", "512", OPT_BW },
1700 { " 4x8-bw-thread,", "mem", "-p", "4", "-t", "8", "-P", "512", OPT_BW },
1702 "mem", "-p", "4", "-t", "8", "-P", "512", OPT_BW_NOTHP },
1703 { " 3x3-bw-thread,", "mem", "-p", "3", "-t", "3", "-P", "512", OPT_BW },
1704 { " 5x5-bw-thread,", "mem", "-p", "5", "-t", "5", "-P", "512", OPT_BW },
1706 { "2x16-bw-thread,", "mem", "-p", "2", "-t", "16", "-P", "512", OPT_BW },
1707 { "1x32-bw-thread,", "mem", "-p", "1", "-t", "32", "-P", "2048", OPT_BW },
1709 { "numa02-bw,", "mem", "-p", "1", "-t", "32", "-T", "32", OPT_BW },
1710 { "numa02-bw-NOTHP,", "mem", "-p", "1", "-t", "32", "-T", "32", OPT_BW_NOTHP },
1711 { "numa01-bw-thread,", "mem", "-p", "2", "-t", "16", "-T", "192", OPT_BW },
1713 "mem", "-p", "2", "-t", "16", "-T", "192", OPT_BW_NOTHP },
/linux-4.1.27/tools/perf/util/include/asm/
H A Ddwarf2.h5 /* dwarf2.h ... dummy header file for including arch/x86/lib/mem{cpy,set}_64.S */
/linux-4.1.27/tools/perf/
H A Dbuiltin-mem.c24 static int __cmd_record(int argc, const char **argv, struct perf_mem *mem) __cmd_record() argument
37 if (mem->operation & MEM_OPERATION_LOAD) __cmd_record()
42 if (mem->operation & MEM_OPERATION_LOAD) { __cmd_record()
44 rec_argv[i++] = "cpu/mem-loads/pp"; __cmd_record()
47 if (mem->operation & MEM_OPERATION_STORE) { __cmd_record()
49 rec_argv[i++] = "cpu/mem-stores/pp"; __cmd_record()
66 struct perf_mem *mem = container_of(tool, struct perf_mem, tool); dump_raw_samples() local
76 if (al.filtered || (mem->hide_unresolved && al.sym == NULL)) dump_raw_samples()
119 static int report_raw_events(struct perf_mem *mem) report_raw_events() argument
124 .force = mem->force, report_raw_events()
129 &mem->tool); report_raw_events()
134 if (mem->cpu_list) { report_raw_events()
135 ret = perf_session__cpu_bitmap(session, mem->cpu_list, report_raw_events()
136 mem->cpu_bitmap); report_raw_events()
157 static int report_events(int argc, const char **argv, struct perf_mem *mem) report_events() argument
162 if (mem->dump_raw) report_events()
163 return report_raw_events(mem); report_events()
171 rep_argv[i++] = "--mem-mode"; report_events()
178 if (!(mem->operation & MEM_OPERATION_LOAD)) report_events()
179 rep_argv[i++] = "--sort=mem,sym,dso,symbol_daddr," report_events()
262 struct perf_mem mem = { cmd_mem() local
280 OPT_CALLBACK('t', "type", &mem.operation, cmd_mem()
283 OPT_BOOLEAN('D', "dump-raw-samples", &mem.dump_raw, cmd_mem()
285 OPT_BOOLEAN('U', "hide-unresolved", &mem.hide_unresolved, cmd_mem()
289 OPT_STRING('C', "cpu", &mem.cpu_list, "cpu", cmd_mem()
295 OPT_BOOLEAN('f', "force", &mem.force, "don't complain, do it"), cmd_mem()
308 if (!argc || !(strncmp(argv[0], "rec", 3) || mem.operation)) cmd_mem()
311 if (!mem.input_name || !strlen(mem.input_name)) { cmd_mem()
313 mem.input_name = "-"; cmd_mem()
315 mem.input_name = "perf.data"; cmd_mem()
319 return __cmd_record(argc, argv, &mem); cmd_mem()
321 return report_events(argc, argv, &mem); cmd_mem()
/linux-4.1.27/drivers/isdn/pcbit/
H A Dmodule.c25 static int mem[MAX_PCBIT_CARDS]; variable
28 module_param_array(mem, int, NULL, 0);
44 if (mem[0] || irq[0]) pcbit_init()
46 for (board = 0; board < MAX_PCBIT_CARDS && mem[board] && irq[board]; board++) pcbit_init()
48 if (!mem[board]) pcbit_init()
49 mem[board] = 0xD0000; pcbit_init()
53 if (pcbit_init_dev(board, mem[board], irq[board]) == 0) pcbit_init()
108 mem[i] = ints[j]; pcbit_setup()
/linux-4.1.27/arch/x86/um/shared/sysdep/
H A Dkernel-offsets.h12 #define OFFSET(sym, str, mem) \
13 DEFINE(sym, offsetof(struct str, mem));
/linux-4.1.27/drivers/firmware/efi/
H A Dcper.c210 static int cper_mem_err_location(struct cper_mem_err_compact *mem, char *msg) cper_mem_err_location() argument
219 if (mem->validation_bits & CPER_MEM_VALID_NODE) cper_mem_err_location()
220 n += scnprintf(msg + n, len - n, "node: %d ", mem->node); cper_mem_err_location()
221 if (mem->validation_bits & CPER_MEM_VALID_CARD) cper_mem_err_location()
222 n += scnprintf(msg + n, len - n, "card: %d ", mem->card); cper_mem_err_location()
223 if (mem->validation_bits & CPER_MEM_VALID_MODULE) cper_mem_err_location()
224 n += scnprintf(msg + n, len - n, "module: %d ", mem->module); cper_mem_err_location()
225 if (mem->validation_bits & CPER_MEM_VALID_RANK_NUMBER) cper_mem_err_location()
226 n += scnprintf(msg + n, len - n, "rank: %d ", mem->rank); cper_mem_err_location()
227 if (mem->validation_bits & CPER_MEM_VALID_BANK) cper_mem_err_location()
228 n += scnprintf(msg + n, len - n, "bank: %d ", mem->bank); cper_mem_err_location()
229 if (mem->validation_bits & CPER_MEM_VALID_DEVICE) cper_mem_err_location()
230 n += scnprintf(msg + n, len - n, "device: %d ", mem->device); cper_mem_err_location()
231 if (mem->validation_bits & CPER_MEM_VALID_ROW) cper_mem_err_location()
232 n += scnprintf(msg + n, len - n, "row: %d ", mem->row); cper_mem_err_location()
233 if (mem->validation_bits & CPER_MEM_VALID_COLUMN) cper_mem_err_location()
234 n += scnprintf(msg + n, len - n, "column: %d ", mem->column); cper_mem_err_location()
235 if (mem->validation_bits & CPER_MEM_VALID_BIT_POSITION) cper_mem_err_location()
237 mem->bit_pos); cper_mem_err_location()
238 if (mem->validation_bits & CPER_MEM_VALID_REQUESTOR_ID) cper_mem_err_location()
240 mem->requestor_id); cper_mem_err_location()
241 if (mem->validation_bits & CPER_MEM_VALID_RESPONDER_ID) cper_mem_err_location()
243 mem->responder_id); cper_mem_err_location()
244 if (mem->validation_bits & CPER_MEM_VALID_TARGET_ID) cper_mem_err_location()
246 mem->target_id); cper_mem_err_location()
252 static int cper_dimm_err_location(struct cper_mem_err_compact *mem, char *msg) cper_dimm_err_location() argument
257 if (!msg || !(mem->validation_bits & CPER_MEM_VALID_MODULE_HANDLE)) cper_dimm_err_location()
262 dmi_memdev_name(mem->mem_dev_handle, &bank, &device); cper_dimm_err_location()
268 mem->mem_dev_handle); cper_dimm_err_location()
274 void cper_mem_err_pack(const struct cper_sec_mem_err *mem, cper_mem_err_pack() argument
277 cmem->validation_bits = mem->validation_bits; cper_mem_err_pack()
278 cmem->node = mem->node; cper_mem_err_pack()
279 cmem->card = mem->card; cper_mem_err_pack()
280 cmem->module = mem->module; cper_mem_err_pack()
281 cmem->bank = mem->bank; cper_mem_err_pack()
282 cmem->device = mem->device; cper_mem_err_pack()
283 cmem->row = mem->row; cper_mem_err_pack()
284 cmem->column = mem->column; cper_mem_err_pack()
285 cmem->bit_pos = mem->bit_pos; cper_mem_err_pack()
286 cmem->requestor_id = mem->requestor_id; cper_mem_err_pack()
287 cmem->responder_id = mem->responder_id; cper_mem_err_pack()
288 cmem->target_id = mem->target_id; cper_mem_err_pack()
289 cmem->rank = mem->rank; cper_mem_err_pack()
290 cmem->mem_array_handle = mem->mem_array_handle; cper_mem_err_pack()
291 cmem->mem_dev_handle = mem->mem_dev_handle; cper_mem_err_pack()
308 static void cper_print_mem(const char *pfx, const struct cper_sec_mem_err *mem, cper_print_mem() argument
315 (mem->validation_bits & ~(CPER_MEM_VALID_RANK_NUMBER - 1))) { cper_print_mem()
319 if (mem->validation_bits & CPER_MEM_VALID_ERROR_STATUS) cper_print_mem()
320 printk("%s""error_status: 0x%016llx\n", pfx, mem->error_status); cper_print_mem()
321 if (mem->validation_bits & CPER_MEM_VALID_PA) cper_print_mem()
323 pfx, mem->physical_addr); cper_print_mem()
324 if (mem->validation_bits & CPER_MEM_VALID_PA_MASK) cper_print_mem()
326 pfx, mem->physical_addr_mask); cper_print_mem()
327 cper_mem_err_pack(mem, &cmem); cper_print_mem()
330 if (mem->validation_bits & CPER_MEM_VALID_ERROR_TYPE) { cper_print_mem()
331 u8 etype = mem->error_type; cper_print_mem()
/linux-4.1.27/drivers/gpu/drm/ttm/
H A Dttm_agp_backend.c47 struct agp_memory *mem; member in struct:ttm_agp_backend
55 struct agp_memory *mem; ttm_agp_bind() local
59 mem = agp_allocate_memory(agp_be->bridge, ttm->num_pages, AGP_USER_MEMORY); ttm_agp_bind()
60 if (unlikely(mem == NULL)) ttm_agp_bind()
63 mem->page_count = 0; ttm_agp_bind()
70 mem->pages[mem->page_count++] = page; ttm_agp_bind()
72 agp_be->mem = mem; ttm_agp_bind()
74 mem->is_flushed = 1; ttm_agp_bind()
75 mem->type = (cached) ? AGP_USER_CACHED_MEMORY : AGP_USER_MEMORY; ttm_agp_bind()
77 ret = agp_bind_memory(mem, node->start); ttm_agp_bind()
88 if (agp_be->mem) { ttm_agp_unbind()
89 if (agp_be->mem->is_bound) ttm_agp_unbind()
90 return agp_unbind_memory(agp_be->mem); ttm_agp_unbind()
91 agp_free_memory(agp_be->mem); ttm_agp_unbind()
92 agp_be->mem = NULL; ttm_agp_unbind()
101 if (agp_be->mem) ttm_agp_destroy()
124 agp_be->mem = NULL; ttm_agp_tt_create()
H A Dttm_bo_util.c44 ttm_bo_mem_put(bo, &bo->mem); ttm_bo_free_old_node()
52 struct ttm_mem_reg *old_mem = &bo->mem; ttm_bo_move_ttm()
120 struct ttm_mem_reg *mem) ttm_mem_io_reserve()
122 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ttm_mem_io_reserve()
128 return bdev->driver->io_mem_reserve(bdev, mem); ttm_mem_io_reserve()
131 mem->bus.io_reserved_count++ == 0) { ttm_mem_io_reserve()
133 ret = bdev->driver->io_mem_reserve(bdev, mem); ttm_mem_io_reserve()
145 struct ttm_mem_reg *mem) ttm_mem_io_free()
147 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ttm_mem_io_free()
153 --mem->bus.io_reserved_count == 0 && ttm_mem_io_free()
155 bdev->driver->io_mem_free(bdev, mem); ttm_mem_io_free()
162 struct ttm_mem_reg *mem = &bo->mem; ttm_mem_io_reserve_vm() local
165 if (!mem->bus.io_reserved_vm) { ttm_mem_io_reserve_vm()
167 &bo->bdev->man[mem->mem_type]; ttm_mem_io_reserve_vm()
169 ret = ttm_mem_io_reserve(bo->bdev, mem); ttm_mem_io_reserve_vm()
172 mem->bus.io_reserved_vm = true; ttm_mem_io_reserve_vm()
182 struct ttm_mem_reg *mem = &bo->mem; ttm_mem_io_free_vm() local
184 if (mem->bus.io_reserved_vm) { ttm_mem_io_free_vm()
185 mem->bus.io_reserved_vm = false; ttm_mem_io_free_vm()
187 ttm_mem_io_free(bo->bdev, mem); ttm_mem_io_free_vm()
191 static int ttm_mem_reg_ioremap(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem, ttm_mem_reg_ioremap() argument
194 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ttm_mem_reg_ioremap()
200 ret = ttm_mem_io_reserve(bdev, mem); ttm_mem_reg_ioremap()
202 if (ret || !mem->bus.is_iomem) ttm_mem_reg_ioremap()
205 if (mem->bus.addr) { ttm_mem_reg_ioremap()
206 addr = mem->bus.addr; ttm_mem_reg_ioremap()
208 if (mem->placement & TTM_PL_FLAG_WC) ttm_mem_reg_ioremap()
209 addr = ioremap_wc(mem->bus.base + mem->bus.offset, mem->bus.size); ttm_mem_reg_ioremap()
211 addr = ioremap_nocache(mem->bus.base + mem->bus.offset, mem->bus.size); ttm_mem_reg_ioremap()
214 ttm_mem_io_free(bdev, mem); ttm_mem_reg_ioremap()
223 static void ttm_mem_reg_iounmap(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem, ttm_mem_reg_iounmap() argument
228 man = &bdev->man[mem->mem_type]; ttm_mem_reg_iounmap()
230 if (virtual && mem->bus.addr == NULL) ttm_mem_reg_iounmap()
233 ttm_mem_io_free(bdev, mem); ttm_mem_reg_iounmap()
330 struct ttm_mem_reg *old_mem = &bo->mem; ttm_bo_move_memcpy()
511 struct ttm_mem_reg *mem = &bo->mem; ttm_bo_ioremap() local
513 if (bo->mem.bus.addr) { ttm_bo_ioremap()
515 map->virtual = (void *)(((u8 *)bo->mem.bus.addr) + offset); ttm_bo_ioremap()
518 if (mem->placement & TTM_PL_FLAG_WC) ttm_bo_ioremap()
519 map->virtual = ioremap_wc(bo->mem.bus.base + bo->mem.bus.offset + offset, ttm_bo_ioremap()
522 map->virtual = ioremap_nocache(bo->mem.bus.base + bo->mem.bus.offset + offset, ttm_bo_ioremap()
533 struct ttm_mem_reg *mem = &bo->mem; pgprot_t prot; ttm_bo_kmap_ttm() local
545 if (num_pages == 1 && (mem->placement & TTM_PL_FLAG_CACHED)) { ttm_bo_kmap_ttm()
559 prot = ttm_io_prot(mem->placement, PAGE_KERNEL); ttm_bo_kmap_ttm()
572 &bo->bdev->man[bo->mem.mem_type]; ttm_bo_kmap()
588 ret = ttm_mem_io_reserve(bo->bdev, &bo->mem); ttm_bo_kmap()
592 if (!bo->mem.bus.is_iomem) { ttm_bo_kmap()
606 &bo->bdev->man[bo->mem.mem_type]; ttm_bo_kunmap()
626 ttm_mem_io_free(map->bo->bdev, &map->bo->mem); ttm_bo_kunmap()
641 struct ttm_mem_reg *old_mem = &bo->mem; ttm_bo_move_accel_cleanup()
119 ttm_mem_io_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ttm_mem_io_reserve() argument
144 ttm_mem_io_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ttm_mem_io_free() argument
H A Dttm_bo.c91 bo, bo->mem.num_pages, bo->mem.size >> 10, ttm_bo_mem_space_debug()
92 bo->mem.size >> 20); ttm_bo_mem_space_debug()
146 BUG_ON(bo->mem.mm_node != NULL); ttm_bo_release_list()
171 if (!(bo->mem.placement & TTM_PL_FLAG_NO_EVICT)) { ttm_bo_add_to_lru()
175 man = &bdev->man[bo->mem.mem_type]; ttm_bo_add_to_lru()
277 struct ttm_mem_reg *mem, ttm_bo_handle_move_mem()
282 bool old_is_pci = ttm_mem_reg_is_pci(bdev, &bo->mem); ttm_bo_handle_move_mem()
283 bool new_is_pci = ttm_mem_reg_is_pci(bdev, mem); ttm_bo_handle_move_mem()
284 struct ttm_mem_type_manager *old_man = &bdev->man[bo->mem.mem_type]; ttm_bo_handle_move_mem()
285 struct ttm_mem_type_manager *new_man = &bdev->man[mem->mem_type]; ttm_bo_handle_move_mem()
289 ((mem->placement & bo->mem.placement & TTM_PL_MASK_CACHING) == 0)) { ttm_bo_handle_move_mem()
309 ret = ttm_tt_set_placement_caching(bo->ttm, mem->placement); ttm_bo_handle_move_mem()
313 if (mem->mem_type != TTM_PL_SYSTEM) { ttm_bo_handle_move_mem()
314 ret = ttm_tt_bind(bo->ttm, mem); ttm_bo_handle_move_mem()
319 if (bo->mem.mem_type == TTM_PL_SYSTEM) { ttm_bo_handle_move_mem()
321 bdev->driver->move_notify(bo, mem); ttm_bo_handle_move_mem()
322 bo->mem = *mem; ttm_bo_handle_move_mem()
323 mem->mm_node = NULL; ttm_bo_handle_move_mem()
329 bdev->driver->move_notify(bo, mem); ttm_bo_handle_move_mem()
333 ret = ttm_bo_move_ttm(bo, evict, no_wait_gpu, mem); ttm_bo_handle_move_mem()
336 no_wait_gpu, mem); ttm_bo_handle_move_mem()
338 ret = ttm_bo_move_memcpy(bo, evict, no_wait_gpu, mem); ttm_bo_handle_move_mem()
342 struct ttm_mem_reg tmp_mem = *mem; ttm_bo_handle_move_mem()
343 *mem = bo->mem; ttm_bo_handle_move_mem()
344 bo->mem = tmp_mem; ttm_bo_handle_move_mem()
345 bdev->driver->move_notify(bo, mem); ttm_bo_handle_move_mem()
346 bo->mem = *mem; ttm_bo_handle_move_mem()
347 *mem = tmp_mem; ttm_bo_handle_move_mem()
356 ret = bdev->driver->invalidate_caches(bdev, bo->mem.placement); ttm_bo_handle_move_mem()
363 if (bo->mem.mm_node) { ttm_bo_handle_move_mem()
364 bo->offset = (bo->mem.start << PAGE_SHIFT) + ttm_bo_handle_move_mem()
365 bdev->man[bo->mem.mem_type].gpu_offset; ttm_bo_handle_move_mem()
366 bo->cur_placement = bo->mem.placement; ttm_bo_handle_move_mem()
373 new_man = &bdev->man[bo->mem.mem_type]; ttm_bo_handle_move_mem()
401 ttm_bo_mem_put(bo, &bo->mem); ttm_bo_cleanup_memtype_use()
454 if (bo->mem.placement & TTM_PL_FLAG_NO_EVICT) { ttm_bo_cleanup_refs_or_queue()
455 bo->mem.placement &= ~TTM_PL_FLAG_NO_EVICT; ttm_bo_cleanup_refs_or_queue()
626 struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; ttm_bo_release()
678 evict_mem = bo->mem; ttm_bo_evict()
729 if (place->fpfn >= (bo->mem.start + bo->mem.size) || ttm_mem_evict_first()
730 (place->lpfn && place->lpfn <= bo->mem.start)) { ttm_mem_evict_first()
769 void ttm_bo_mem_put(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) ttm_bo_mem_put() argument
771 struct ttm_mem_type_manager *man = &bo->bdev->man[mem->mem_type]; ttm_bo_mem_put()
773 if (mem->mm_node) ttm_bo_mem_put()
774 (*man->func->put_node)(man, mem); ttm_bo_mem_put()
785 struct ttm_mem_reg *mem, ttm_bo_mem_force_space()
794 ret = (*man->func->get_node)(man, bo, place, mem); ttm_bo_mem_force_space()
797 if (mem->mm_node) ttm_bo_mem_force_space()
804 if (mem->mm_node == NULL) ttm_bo_mem_force_space()
806 mem->mem_type = mem_type; ttm_bo_mem_force_space()
855 * Creates space for memory region @mem according to its type.
864 struct ttm_mem_reg *mem, ttm_bo_mem_space()
877 mem->mm_node = NULL; ttm_bo_mem_space()
892 cur_flags = ttm_bo_select_caching(man, bo->mem.placement, ttm_bo_mem_space()
906 ret = (*man->func->get_node)(man, bo, place, mem); ttm_bo_mem_space()
910 if (mem->mm_node) ttm_bo_mem_space()
914 if ((type_ok && (mem_type == TTM_PL_SYSTEM)) || mem->mm_node) { ttm_bo_mem_space()
915 mem->mem_type = mem_type; ttm_bo_mem_space()
916 mem->placement = cur_flags; ttm_bo_mem_space()
935 cur_flags = ttm_bo_select_caching(man, bo->mem.placement, ttm_bo_mem_space()
945 mem->mem_type = mem_type; ttm_bo_mem_space()
946 mem->placement = cur_flags; ttm_bo_mem_space()
947 mem->mm_node = NULL; ttm_bo_mem_space()
951 ret = ttm_bo_mem_force_space(bo, mem_type, place, mem, ttm_bo_mem_space()
953 if (ret == 0 && mem->mm_node) { ttm_bo_mem_space()
954 mem->placement = cur_flags; ttm_bo_mem_space()
971 struct ttm_mem_reg mem; ttm_bo_move_buffer() local
983 mem.num_pages = bo->num_pages; ttm_bo_move_buffer()
984 mem.size = mem.num_pages << PAGE_SHIFT; ttm_bo_move_buffer()
985 mem.page_alignment = bo->mem.page_alignment; ttm_bo_move_buffer()
986 mem.bus.io_reserved_vm = false; ttm_bo_move_buffer()
987 mem.bus.io_reserved_count = 0; ttm_bo_move_buffer()
991 ret = ttm_bo_mem_space(bo, placement, &mem, ttm_bo_move_buffer()
995 ret = ttm_bo_handle_move_mem(bo, &mem, false, ttm_bo_move_buffer()
998 if (ret && mem.mm_node) ttm_bo_move_buffer()
999 ttm_bo_mem_put(bo, &mem); ttm_bo_move_buffer()
1004 struct ttm_mem_reg *mem, ttm_bo_mem_compat()
1011 if (mem->mm_node && ttm_bo_mem_compat()
1012 (mem->start < heap->fpfn || ttm_bo_mem_compat()
1013 (heap->lpfn != 0 && (mem->start + mem->num_pages) > heap->lpfn))) ttm_bo_mem_compat()
1017 if ((*new_flags & mem->placement & TTM_PL_MASK_CACHING) && ttm_bo_mem_compat()
1018 (*new_flags & mem->placement & TTM_PL_MASK_MEM)) ttm_bo_mem_compat()
1024 if (mem->mm_node && ttm_bo_mem_compat()
1025 (mem->start < heap->fpfn || ttm_bo_mem_compat()
1026 (heap->lpfn != 0 && (mem->start + mem->num_pages) > heap->lpfn))) ttm_bo_mem_compat()
1030 if ((*new_flags & mem->placement & TTM_PL_MASK_CACHING) && ttm_bo_mem_compat()
1031 (*new_flags & mem->placement & TTM_PL_MASK_MEM)) ttm_bo_mem_compat()
1050 if (!ttm_bo_mem_compat(placement, &bo->mem, &new_flags)) { ttm_bo_validate()
1060 ttm_flag_masked(&bo->mem.placement, new_flags, ttm_bo_validate()
1066 if (bo->mem.mem_type == TTM_PL_SYSTEM && bo->ttm == NULL) { ttm_bo_validate()
1127 bo->mem.size = num_pages << PAGE_SHIFT; ttm_bo_init()
1128 bo->mem.mem_type = TTM_PL_SYSTEM; ttm_bo_init()
1129 bo->mem.num_pages = bo->num_pages; ttm_bo_init()
1130 bo->mem.mm_node = NULL; ttm_bo_init()
1131 bo->mem.page_alignment = page_alignment; ttm_bo_init()
1132 bo->mem.bus.io_reserved_vm = false; ttm_bo_init()
1133 bo->mem.bus.io_reserved_count = 0; ttm_bo_init()
1135 bo->mem.placement = (TTM_PL_FLAG_SYSTEM | TTM_PL_FLAG_CACHED); ttm_bo_init()
1156 bo->mem.num_pages); ttm_bo_init()
1496 bool ttm_mem_reg_is_pci(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ttm_mem_reg_is_pci() argument
1498 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ttm_mem_reg_is_pci()
1501 if (mem->mem_type == TTM_PL_SYSTEM) ttm_mem_reg_is_pci()
1507 if (mem->placement & TTM_PL_FLAG_CACHED) ttm_mem_reg_is_pci()
1524 struct ttm_mem_type_manager *man = &bdev->man[bo->mem.mem_type]; ttm_bo_unmap_virtual()
1655 if ((bo->mem.placement & swap_placement) != swap_placement) { ttm_bo_swapout()
1658 evict_mem = bo->mem; ttm_bo_swapout()
276 ttm_bo_handle_move_mem(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem, bool evict, bool interruptible, bool no_wait_gpu) ttm_bo_handle_move_mem() argument
782 ttm_bo_mem_force_space(struct ttm_buffer_object *bo, uint32_t mem_type, const struct ttm_place *place, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) ttm_bo_mem_force_space() argument
862 ttm_bo_mem_space(struct ttm_buffer_object *bo, struct ttm_placement *placement, struct ttm_mem_reg *mem, bool interruptible, bool no_wait_gpu) ttm_bo_mem_space() argument
1003 ttm_bo_mem_compat(struct ttm_placement *placement, struct ttm_mem_reg *mem, uint32_t *new_flags) ttm_bo_mem_compat() argument
H A Dttm_bo_manager.c53 struct ttm_mem_reg *mem) ttm_bo_man_get_node()
77 ret = drm_mm_insert_node_in_range_generic(mm, node, mem->num_pages, ttm_bo_man_get_node()
78 mem->page_alignment, 0, ttm_bo_man_get_node()
86 mem->mm_node = node; ttm_bo_man_get_node()
87 mem->start = node->start; ttm_bo_man_get_node()
94 struct ttm_mem_reg *mem) ttm_bo_man_put_node()
98 if (mem->mm_node) { ttm_bo_man_put_node()
100 drm_mm_remove_node(mem->mm_node); ttm_bo_man_put_node()
103 kfree(mem->mm_node); ttm_bo_man_put_node()
104 mem->mm_node = NULL; ttm_bo_man_put_node()
50 ttm_bo_man_get_node(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) ttm_bo_man_get_node() argument
93 ttm_bo_man_put_node(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) ttm_bo_man_put_node() argument
H A Dttm_memory.c249 uint64_t mem; ttm_mem_init_kernel_zone() local
255 mem = si->totalram - si->totalhigh; ttm_mem_init_kernel_zone()
256 mem *= si->mem_unit; ttm_mem_init_kernel_zone()
259 zone->zone_mem = mem; ttm_mem_init_kernel_zone()
260 zone->max_mem = mem >> 1; ttm_mem_init_kernel_zone()
261 zone->emer_mem = (mem >> 1) + (mem >> 2); ttm_mem_init_kernel_zone()
262 zone->swap_limit = zone->max_mem - (mem >> 3); ttm_mem_init_kernel_zone()
281 uint64_t mem; ttm_mem_init_highmem_zone() local
291 mem = si->totalram; ttm_mem_init_highmem_zone()
292 mem *= si->mem_unit; ttm_mem_init_highmem_zone()
295 zone->zone_mem = mem; ttm_mem_init_highmem_zone()
296 zone->max_mem = mem >> 1; ttm_mem_init_highmem_zone()
297 zone->emer_mem = (mem >> 1) + (mem >> 2); ttm_mem_init_highmem_zone()
298 zone->swap_limit = zone->max_mem - (mem >> 3); ttm_mem_init_highmem_zone()
317 uint64_t mem; ttm_mem_init_dma32_zone() local
323 mem = si->totalram; ttm_mem_init_dma32_zone()
324 mem *= si->mem_unit; ttm_mem_init_dma32_zone()
330 if (mem <= ((uint64_t) 1ULL << 32)) { ttm_mem_init_dma32_zone()
341 mem = ((uint64_t) 1ULL << 32); ttm_mem_init_dma32_zone()
343 zone->zone_mem = mem; ttm_mem_init_dma32_zone()
344 zone->max_mem = mem >> 1; ttm_mem_init_dma32_zone()
345 zone->emer_mem = (mem >> 1) + (mem >> 2); ttm_mem_init_dma32_zone()
346 zone->swap_limit = zone->max_mem - (mem >> 3); ttm_mem_init_dma32_zone()
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/subdev/bar/
H A Dgf100.c32 struct nvkm_gpuobj *mem; member in struct:gf100_bar_priv_vm
44 gf100_bar_kmap(struct nvkm_bar *bar, struct nvkm_mem *mem, u32 flags, gf100_bar_kmap() argument
50 ret = nvkm_vm_get(priv->bar[0].vm, mem->size << 12, 12, flags, vma); gf100_bar_kmap()
54 nvkm_vm_map(vma, mem); gf100_bar_kmap()
59 gf100_bar_umap(struct nvkm_bar *bar, struct nvkm_mem *mem, u32 flags, gf100_bar_umap() argument
65 ret = nvkm_vm_get(priv->bar[1].vm, mem->size << 12, gf100_bar_umap()
66 mem->page_shift, flags, vma); gf100_bar_umap()
70 nvkm_vm_map(vma, mem); gf100_bar_umap()
91 &bar_vm->mem); gf100_bar_ctor_vm()
126 nv_wo32(bar_vm->mem, 0x0200, lower_32_bits(bar_vm->pgd->addr)); gf100_bar_ctor_vm()
127 nv_wo32(bar_vm->mem, 0x0204, upper_32_bits(bar_vm->pgd->addr)); gf100_bar_ctor_vm()
128 nv_wo32(bar_vm->mem, 0x0208, lower_32_bits(bar_len - 1)); gf100_bar_ctor_vm()
129 nv_wo32(bar_vm->mem, 0x020c, upper_32_bits(bar_len - 1)); gf100_bar_ctor_vm()
178 nvkm_gpuobj_ref(NULL, &priv->bar[1].mem); gf100_bar_dtor()
185 nvkm_gpuobj_ref(NULL, &priv->bar[0].mem); gf100_bar_dtor()
203 nv_wr32(priv, 0x001704, 0x80000000 | priv->bar[1].mem->addr >> 12); gf100_bar_init()
204 if (priv->bar[0].mem) gf100_bar_init()
206 0xc0000000 | priv->bar[0].mem->addr >> 12); gf100_bar_init()
H A Dnv50.c35 struct nvkm_gpuobj *mem; member in struct:nv50_bar_priv
45 nv50_bar_kmap(struct nvkm_bar *bar, struct nvkm_mem *mem, u32 flags, nv50_bar_kmap() argument
51 ret = nvkm_vm_get(priv->bar3_vm, mem->size << 12, 12, flags, vma); nv50_bar_kmap()
55 nvkm_vm_map(vma, mem); nv50_bar_kmap()
60 nv50_bar_umap(struct nvkm_bar *bar, struct nvkm_mem *mem, u32 flags, nv50_bar_umap() argument
66 ret = nvkm_vm_get(priv->bar1_vm, mem->size << 12, 12, flags, vma); nv50_bar_umap()
70 nvkm_vm_map(vma, mem); nv50_bar_umap()
123 NVOBJ_FLAG_HEAP, &priv->mem); nv50_bar_ctor()
124 heap = nv_object(priv->mem); nv50_bar_ctor()
224 nvkm_gpuobj_ref(NULL, &priv->mem); nv50_bar_dtor()
246 nv_wr32(priv, 0x001704, 0x00000000 | priv->mem->addr >> 12); nv50_bar_init()
247 nv_wr32(priv, 0x001704, 0x40000000 | priv->mem->addr >> 12); nv50_bar_init()
H A Dbase.c43 struct nvkm_mem *mem = data; nvkm_barobj_ctor() local
52 ret = bar->kmap(bar, mem, NV_MEM_ACCESS_RW, &barobj->vma); nvkm_barobj_ctor()
57 (u32)barobj->vma.offset, mem->size << 12); nvkm_barobj_ctor()
107 struct nvkm_mem *mem, struct nvkm_object **pobject) nvkm_bar_alloc()
111 &nvkm_barobj_oclass, mem, 0, &gpuobj); nvkm_bar_alloc()
106 nvkm_bar_alloc(struct nvkm_bar *bar, struct nvkm_object *parent, struct nvkm_mem *mem, struct nvkm_object **pobject) nvkm_bar_alloc() argument
/linux-4.1.27/arch/ia64/kernel/
H A Dminstate.h99 .mem.offset 0,0; st8.spill [r16]=r8,16; \
100 .mem.offset 8,0; st8.spill [r17]=r9,16; \
102 .mem.offset 0,0; st8.spill [r16]=r10,24; \
103 .mem.offset 8,0; st8.spill [r17]=r11,24; \
129 .mem.offset 0,0; st8.spill [r16]=r20,16; /* save original r1 */ \
130 .mem.offset 8,0; st8.spill [r17]=r12,16; \
133 .mem.offset 0,0; st8.spill [r16]=r13,16; \
134 .mem.offset 8,0; st8.spill [r17]=r21,16; /* save ar.fpsr */ \
137 .mem.offset 0,0; st8.spill [r16]=r15,16; \
138 .mem.offset 8,0; st8.spill [r17]=r14,16; \
140 .mem.offset 0,0; st8.spill [r16]=r2,16; \
141 .mem.offset 8,0; st8.spill [r17]=r3,16; \
167 .mem.offset 0,0; st8.spill [r2]=r16,16; \
168 .mem.offset 8,0; st8.spill [r3]=r17,16; \
170 .mem.offset 0,0; st8.spill [r2]=r18,16; \
171 .mem.offset 8,0; st8.spill [r3]=r19,16; \
173 .mem.offset 0,0; st8.spill [r2]=r20,16; \
174 .mem.offset 8,0; st8.spill [r3]=r21,16; \
177 .mem.offset 0,0; st8.spill [r2]=r22,16; \
178 .mem.offset 8,0; st8.spill [r3]=r23,16; \
181 .mem.offset 0,0; st8.spill [r2]=r24,16; \
182 .mem.offset 8,0; st8.spill [r3]=r25,16; \
184 .mem.offset 0,0; st8.spill [r2]=r26,16; \
185 .mem.offset 8,0; st8.spill [r3]=r27,16; \
187 .mem.offset 0,0; st8.spill [r2]=r28,16; \
188 .mem.offset 8,0; st8.spill [r3]=r29,16; \
190 .mem.offset 0,0; st8.spill [r2]=r30,16; \
191 .mem.offset 8,0; st8.spill [r3]=r31,32; \
/linux-4.1.27/drivers/char/hw_random/
H A Dintel-rng.c155 static inline u8 hwstatus_get(void __iomem *mem) hwstatus_get() argument
157 return readb(mem + INTEL_RNG_HW_STATUS); hwstatus_get()
160 static inline u8 hwstatus_set(void __iomem *mem, hwstatus_set() argument
163 writeb(hw_status, mem + INTEL_RNG_HW_STATUS); hwstatus_set()
164 return hwstatus_get(mem); hwstatus_set()
169 void __iomem *mem = (void __iomem *)rng->priv; intel_rng_data_present() local
173 data = !!(readb(mem + INTEL_RNG_STATUS) & intel_rng_data_present()
184 void __iomem *mem = (void __iomem *)rng->priv; intel_rng_data_read() local
186 *data = readb(mem + INTEL_RNG_DATA); intel_rng_data_read()
193 void __iomem *mem = (void __iomem *)rng->priv; intel_rng_init() local
197 hw_status = hwstatus_get(mem); intel_rng_init()
200 hw_status = hwstatus_set(mem, hw_status | INTEL_RNG_ENABLED); intel_rng_init()
212 void __iomem *mem = (void __iomem *)rng->priv; intel_rng_cleanup() local
215 hw_status = hwstatus_get(mem); intel_rng_cleanup()
217 hwstatus_set(mem, hw_status & ~INTEL_RNG_ENABLED); intel_rng_cleanup()
233 void __iomem *mem; member in struct:intel_rng_hw
258 writeb(INTEL_FWH_RESET_CMD, intel_rng_hw->mem); intel_rng_hw_init()
259 writeb(INTEL_FWH_READ_ID_CMD, intel_rng_hw->mem); intel_rng_hw_init()
260 mfc = readb(intel_rng_hw->mem + INTEL_FWH_MANUFACTURER_CODE_ADDRESS); intel_rng_hw_init()
261 dvc = readb(intel_rng_hw->mem + INTEL_FWH_DEVICE_CODE_ADDRESS); intel_rng_hw_init()
262 writeb(INTEL_FWH_RESET_CMD, intel_rng_hw->mem); intel_rng_hw_init()
320 intel_rng_hw->mem = ioremap_nocache(INTEL_FWH_ADDR, INTEL_FWH_ADDR_LEN); intel_init_hw_struct()
321 if (intel_rng_hw->mem == NULL) intel_init_hw_struct()
333 void __iomem *mem = mem; mod_init() local
374 iounmap(intel_rng_hw->mem); mod_init()
381 mem = ioremap(INTEL_RNG_ADDR, INTEL_RNG_ADDR_LEN); mod_init()
382 if (!mem) mod_init()
384 intel_rng.priv = (unsigned long)mem; mod_init()
388 hw_status = hwstatus_get(mem); mod_init()
390 iounmap(mem); mod_init()
399 iounmap(mem); mod_init()
408 void __iomem *mem = (void __iomem *)intel_rng.priv; mod_exit() local
411 iounmap(mem); mod_exit()
H A Dmxc-rnga.c65 void __iomem *mem; member in struct:mxc_rng
76 int level = (__raw_readl(mxc_rng->mem + RNGA_STATUS) & mxc_rnga_data_present()
92 *data = __raw_readl(mxc_rng->mem + RNGA_OUTPUT_FIFO); mxc_rnga_data_read()
95 err = __raw_readl(mxc_rng->mem + RNGA_STATUS) & RNGA_STATUS_ERROR_INT; mxc_rnga_data_read()
100 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); mxc_rnga_data_read()
102 mxc_rng->mem + RNGA_CONTROL); mxc_rnga_data_read()
114 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); mxc_rnga_init()
115 __raw_writel(ctrl & ~RNGA_CONTROL_SLEEP, mxc_rng->mem + RNGA_CONTROL); mxc_rnga_init()
118 osc = __raw_readl(mxc_rng->mem + RNGA_STATUS); mxc_rnga_init()
125 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); mxc_rnga_init()
126 __raw_writel(ctrl | RNGA_CONTROL_GO, mxc_rng->mem + RNGA_CONTROL); mxc_rnga_init()
136 ctrl = __raw_readl(mxc_rng->mem + RNGA_CONTROL); mxc_rnga_cleanup()
139 __raw_writel(ctrl & ~RNGA_CONTROL_GO, mxc_rng->mem + RNGA_CONTROL); mxc_rnga_cleanup()
172 mxc_rng->mem = devm_ioremap_resource(&pdev->dev, res); mxc_rnga_probe()
173 if (IS_ERR(mxc_rng->mem)) { mxc_rnga_probe()
174 err = PTR_ERR(mxc_rng->mem); mxc_rnga_probe()
H A Dgeode-rng.c57 void __iomem *mem = (void __iomem *)rng->priv; geode_rng_data_read() local
59 *data = readl(mem + GEODE_RNG_DATA_REG); geode_rng_data_read()
66 void __iomem *mem = (void __iomem *)rng->priv; geode_rng_data_present() local
70 data = !!(readl(mem + GEODE_RNG_STATUS_REG)); geode_rng_data_present()
91 void __iomem *mem; mod_init() local
107 mem = ioremap(rng_base, 0x58);
108 if (!mem)
110 geode_rng.priv = (unsigned long)mem;
123 iounmap(mem);
129 void __iomem *mem = (void __iomem *)geode_rng.priv; mod_exit() local
132 iounmap(mem); mod_exit()
H A Dexynos-rng.c42 void __iomem *mem; member in struct:exynos_rng
48 return __raw_readl(rng->mem + offset); exynos_rng_readl()
53 __raw_writel(val, rng->mem + offset); exynos_rng_writel()
124 exynos_rng->mem = devm_ioremap_resource(&pdev->dev, res); exynos_rng_probe()
125 if (IS_ERR(exynos_rng->mem)) exynos_rng_probe()
126 return PTR_ERR(exynos_rng->mem); exynos_rng_probe()
/linux-4.1.27/drivers/net/wireless/ti/wlcore/
H A Dio.c83 if ((addr >= part->mem.start) && wlcore_translate_addr()
84 (addr < part->mem.start + part->mem.size)) wlcore_translate_addr()
85 return addr - part->mem.start; wlcore_translate_addr()
88 return addr - part->reg.start + part->mem.size; wlcore_translate_addr()
91 return addr - part->mem2.start + part->mem.size + wlcore_translate_addr()
95 return addr - part->mem3.start + part->mem.size + wlcore_translate_addr()
117 * ...+----+--> mem.start
121 * 00000000 <--+----+... ...+----+--> mem.start + mem.size
125 * mem.size <--+----+... | | {unused area)
128 * mem.size | | ... | |
146 p->mem.start, p->mem.size); wlcore_set_partition()
154 ret = wlcore_raw_write32(wl, HW_PART0_START_ADDR, p->mem.start); wlcore_set_partition()
158 ret = wlcore_raw_write32(wl, HW_PART0_SIZE_ADDR, p->mem.size); wlcore_set_partition()
/linux-4.1.27/drivers/iio/adc/
H A Dmen_z188_adc.c28 struct resource *mem; member in struct:z188_adc
109 struct resource *mem; men_z188_probe() local
123 mem = mcb_request_mem(dev, "z188-adc"); men_z188_probe()
124 if (IS_ERR(mem)) men_z188_probe()
125 return PTR_ERR(mem); men_z188_probe()
127 adc->base = ioremap(mem->start, resource_size(mem)); men_z188_probe()
133 adc->mem = mem; men_z188_probe()
139 mcb_release_mem(mem); men_z188_probe()
150 mcb_release_mem(adc->mem); men_z188_remove()
/linux-4.1.27/arch/um/os-Linux/
H A DMakefile6 obj-y = aio.o execvp.o file.o helper.o irq.o main.o mem.o process.o \
13 main.o mem.o process.o registers.o sigio.o signal.o start_up.o time.o \
/linux-4.1.27/arch/avr32/kernel/
H A Dsetup.c80 struct resource *mem, *res; resource_init() local
85 for (mem = system_ram; mem; mem = mem->sibling) { resource_init()
87 memcpy(new, mem, sizeof(struct resource)); resource_init()
92 mem->start, mem->end); resource_init()
175 find_free_region(const struct resource *mem, resource_size_t size, find_free_region() argument
181 target = ALIGN(mem->start, align); find_free_region()
189 if ((target + size) > (mem->end + 1)) find_free_region()
190 return mem->end + 1; find_free_region()
199 struct resource *mem; alloc_reserved_region() local
203 for (mem = system_ram; mem; mem = mem->sibling) { alloc_reserved_region()
204 target = find_free_region(mem, size, align); alloc_reserved_region()
205 if (target <= mem->end) { alloc_reserved_region()
287 * Pick out the memory size. We look for mem=size@start,
303 early_param("mem", early_mem);
339 struct tag_mem_range *mem = &tag->u.mem_range; parse_tag_rdimg() local
348 ret = add_reserved_region(mem->addr, mem->addr + mem->size - 1, parse_tag_rdimg()
356 initrd_start = (unsigned long)__va(mem->addr); parse_tag_rdimg()
357 initrd_end = initrd_start + mem->size; parse_tag_rdimg()
369 struct tag_mem_range *mem = &tag->u.mem_range; parse_tag_rsvd_mem() local
371 return add_reserved_region(mem->addr, mem->addr + mem->size - 1, parse_tag_rsvd_mem()
444 find_bootmap_pfn(const struct resource *mem) find_bootmap_pfn() argument
447 unsigned long node_pages = PFN_UP(resource_size(mem)); find_bootmap_pfn()
462 bootmap_start = find_free_region(mem, bootmap_len, PAGE_SIZE); find_bootmap_pfn()
/linux-4.1.27/drivers/net/can/sja1000/
H A Dsja1000_isa.c43 static unsigned long mem[MAXDEV]; variable
54 module_param_array(mem, ulong, NULL, S_IRUGO);
55 MODULE_PARM_DESC(mem, "I/O memory address");
136 dev_dbg(&pdev->dev, "probing idx=%d: port=%#lx, mem=%#lx, irq=%d\n", sja1000_isa_probe()
137 idx, port[idx], mem[idx], irq[idx]); sja1000_isa_probe()
139 if (mem[idx]) { sja1000_isa_probe()
140 if (!request_mem_region(mem[idx], iosize, DRV_NAME)) { sja1000_isa_probe()
144 base = ioremap_nocache(mem[idx], iosize); sja1000_isa_probe()
168 if (mem[idx]) { sja1000_isa_probe()
170 dev->base_addr = mem[idx]; sja1000_isa_probe()
224 if (mem[idx]) sja1000_isa_probe()
227 if (mem[idx]) sja1000_isa_probe()
228 release_mem_region(mem[idx], iosize); sja1000_isa_probe()
243 if (mem[idx]) { sja1000_isa_remove()
245 release_mem_region(mem[idx], SJA1000_IOSIZE); sja1000_isa_remove()
270 if ((port[idx] || mem[idx]) && irq[idx]) { sja1000_isa_init()
282 pr_debug("%s: platform device %d: port=%#lx, mem=%#lx, " sja1000_isa_init()
284 DRV_NAME, idx, port[idx], mem[idx], irq[idx]); sja1000_isa_init()
285 } else if (idx == 0 || port[idx] || mem[idx]) { sja1000_isa_init()
/linux-4.1.27/drivers/char/agp/
H A Dsgi-agp.c59 * Flush GART tlb's. Cannot selectively flush based on memory so the mem
63 static void sgi_tioca_tlbflush(struct agp_memory *mem) sgi_tioca_tlbflush() argument
65 tioca_tlbflush(mem->bridge->dev_private_data); sgi_tioca_tlbflush()
127 static int sgi_tioca_insert_memory(struct agp_memory *mem, off_t pg_start, sgi_tioca_insert_memory() argument
137 bridge = mem->bridge; sgi_tioca_insert_memory()
169 if (type != 0 || mem->type != 0) { sgi_tioca_insert_memory()
173 if ((pg_start + mem->page_count) > num_entries) sgi_tioca_insert_memory()
178 while (j < (pg_start + mem->page_count)) { sgi_tioca_insert_memory()
184 if (!mem->is_flushed) { sgi_tioca_insert_memory()
186 mem->is_flushed = true; sgi_tioca_insert_memory()
189 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { sgi_tioca_insert_memory()
192 page_to_phys(mem->pages[i]), sgi_tioca_insert_memory()
193 mem->type); sgi_tioca_insert_memory()
196 bridge->driver->tlb_flush(mem); sgi_tioca_insert_memory()
200 static int sgi_tioca_remove_memory(struct agp_memory *mem, off_t pg_start, sgi_tioca_remove_memory() argument
207 bridge = mem->bridge; sgi_tioca_remove_memory()
211 if (type != 0 || mem->type != 0) { sgi_tioca_remove_memory()
217 for (i = pg_start; i < (mem->page_count + pg_start); i++) { sgi_tioca_remove_memory()
221 bridge->driver->tlb_flush(mem); sgi_tioca_remove_memory()
H A Dati-agp.c170 static void ati_tlbflush(struct agp_memory * mem) ati_tlbflush() argument
266 static int ati_insert_memory(struct agp_memory * mem, ati_insert_memory() argument
276 mask_type = agp_generic_type_to_mask_type(mem->bridge, type); ati_insert_memory()
277 if (mask_type != 0 || type != mem->type) ati_insert_memory()
280 if (mem->page_count == 0) ati_insert_memory()
283 if ((pg_start + mem->page_count) > num_entries) ati_insert_memory()
287 while (j < (pg_start + mem->page_count)) { ati_insert_memory()
295 if (!mem->is_flushed) { ati_insert_memory()
298 mem->is_flushed = true; ati_insert_memory()
301 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { ati_insert_memory()
305 page_to_phys(mem->pages[i]), ati_insert_memory()
306 mem->type), ati_insert_memory()
310 agp_bridge->driver->tlb_flush(mem); ati_insert_memory()
314 static int ati_remove_memory(struct agp_memory * mem, off_t pg_start, ati_remove_memory() argument
322 mask_type = agp_generic_type_to_mask_type(mem->bridge, type); ati_remove_memory()
323 if (mask_type != 0 || type != mem->type) ati_remove_memory()
326 if (mem->page_count == 0) ati_remove_memory()
329 for (i = pg_start; i < (mem->page_count + pg_start); i++) { ati_remove_memory()
336 agp_bridge->driver->tlb_flush(mem); ati_remove_memory()
H A Dalpha-agp.c67 static void alpha_core_agp_tlbflush(struct agp_memory *mem) alpha_core_agp_tlbflush() argument
86 static int alpha_core_agp_insert_memory(struct agp_memory *mem, off_t pg_start, alpha_core_agp_insert_memory() argument
93 if (type >= AGP_USER_TYPES || mem->type >= AGP_USER_TYPES) alpha_core_agp_insert_memory()
98 if ((pg_start + mem->page_count) > num_entries) alpha_core_agp_insert_memory()
101 status = agp->ops->bind(agp, pg_start, mem); alpha_core_agp_insert_memory()
103 alpha_core_agp_tlbflush(mem); alpha_core_agp_insert_memory()
108 static int alpha_core_agp_remove_memory(struct agp_memory *mem, off_t pg_start, alpha_core_agp_remove_memory() argument
114 status = agp->ops->unbind(agp, pg_start, mem); alpha_core_agp_remove_memory()
115 alpha_core_agp_tlbflush(mem); alpha_core_agp_remove_memory()
H A Dnvidia-agp.c201 static int nvidia_insert_memory(struct agp_memory *mem, off_t pg_start, int type) nvidia_insert_memory() argument
206 mask_type = agp_generic_type_to_mask_type(mem->bridge, type); nvidia_insert_memory()
207 if (mask_type != 0 || type != mem->type) nvidia_insert_memory()
210 if (mem->page_count == 0) nvidia_insert_memory()
213 if ((pg_start + mem->page_count) > nvidia_insert_memory()
217 for (j = pg_start; j < (pg_start + mem->page_count); j++) { nvidia_insert_memory()
222 if (!mem->is_flushed) { nvidia_insert_memory()
224 mem->is_flushed = true; nvidia_insert_memory()
226 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { nvidia_insert_memory()
228 page_to_phys(mem->pages[i]), mask_type), nvidia_insert_memory()
235 agp_bridge->driver->tlb_flush(mem); nvidia_insert_memory()
240 static int nvidia_remove_memory(struct agp_memory *mem, off_t pg_start, int type) nvidia_remove_memory() argument
246 mask_type = agp_generic_type_to_mask_type(mem->bridge, type); nvidia_remove_memory()
247 if (mask_type != 0 || type != mem->type) nvidia_remove_memory()
250 if (mem->page_count == 0) nvidia_remove_memory()
253 for (i = pg_start; i < (mem->page_count + pg_start); i++) nvidia_remove_memory()
256 agp_bridge->driver->tlb_flush(mem); nvidia_remove_memory()
261 static void nvidia_tlbflush(struct agp_memory *mem) nvidia_tlbflush() argument
H A Dparisc-agp.c93 parisc_agp_tlbflush(struct agp_memory *mem) parisc_agp_tlbflush() argument
125 parisc_agp_insert_memory(struct agp_memory *mem, off_t pg_start, int type) parisc_agp_insert_memory() argument
132 if (type != mem->type || parisc_agp_insert_memory()
138 io_pg_count = info->io_pages_per_kpage * mem->page_count; parisc_agp_insert_memory()
150 if (!mem->is_flushed) { parisc_agp_insert_memory()
152 mem->is_flushed = true; parisc_agp_insert_memory()
155 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { parisc_agp_insert_memory()
158 paddr = page_to_phys(mem->pages[i]); parisc_agp_insert_memory()
168 agp_bridge->driver->tlb_flush(mem); parisc_agp_insert_memory()
174 parisc_agp_remove_memory(struct agp_memory *mem, off_t pg_start, int type) parisc_agp_remove_memory() argument
179 if (type != mem->type || parisc_agp_remove_memory()
185 io_pg_count = info->io_pages_per_kpage * mem->page_count; parisc_agp_remove_memory()
190 agp_bridge->driver->tlb_flush(mem); parisc_agp_remove_memory()
H A Di460-agp.c170 static void i460_tlb_flush (struct agp_memory *mem) i460_tlb_flush() argument
293 static int i460_insert_memory_small_io_page (struct agp_memory *mem, i460_insert_memory_small_io_page() argument
300 pr_debug("i460_insert_memory_small_io_page(mem=%p, pg_start=%ld, type=%d, paddr0=0x%lx)\n", i460_insert_memory_small_io_page()
301 mem, pg_start, type, page_to_phys(mem->pages[0])); i460_insert_memory_small_io_page()
303 if (type >= AGP_USER_TYPES || mem->type >= AGP_USER_TYPES) i460_insert_memory_small_io_page()
311 if ((io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count) > num_entries) { i460_insert_memory_small_io_page()
317 while (j < (io_pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count)) { i460_insert_memory_small_io_page()
327 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { i460_insert_memory_small_io_page()
328 paddr = page_to_phys(mem->pages[i]); i460_insert_memory_small_io_page()
330 WR_GATT(j, i460_mask_memory(agp_bridge, paddr, mem->type)); i460_insert_memory_small_io_page()
336 static int i460_remove_memory_small_io_page(struct agp_memory *mem, i460_remove_memory_small_io_page() argument
341 pr_debug("i460_remove_memory_small_io_page(mem=%p, pg_start=%ld, type=%d)\n", i460_remove_memory_small_io_page()
342 mem, pg_start, type); i460_remove_memory_small_io_page()
346 for (i = pg_start; i < (pg_start + I460_IOPAGES_PER_KPAGE * mem->page_count); i++) i460_remove_memory_small_io_page()
400 static int i460_insert_memory_large_io_page (struct agp_memory *mem, i460_insert_memory_large_io_page() argument
407 if (type >= AGP_USER_TYPES || mem->type >= AGP_USER_TYPES) i460_insert_memory_large_io_page()
415 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; i460_insert_memory_large_io_page()
417 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; i460_insert_memory_large_io_page()
453 mem->pages[i] = lp->page; i460_insert_memory_large_io_page()
461 static int i460_remove_memory_large_io_page (struct agp_memory *mem, i460_remove_memory_large_io_page() argument
473 end = &i460.lp_desc[(pg_start + mem->page_count - 1) / I460_KPAGES_PER_IOPAGE]; i460_remove_memory_large_io_page()
475 end_offset = (pg_start + mem->page_count - 1) % I460_KPAGES_PER_IOPAGE; i460_remove_memory_large_io_page()
482 mem->pages[i] = NULL; i460_remove_memory_large_io_page()
500 static int i460_insert_memory (struct agp_memory *mem, i460_insert_memory() argument
504 return i460_insert_memory_small_io_page(mem, pg_start, type); i460_insert_memory()
506 return i460_insert_memory_large_io_page(mem, pg_start, type); i460_insert_memory()
509 static int i460_remove_memory (struct agp_memory *mem, i460_remove_memory() argument
513 return i460_remove_memory_small_io_page(mem, pg_start, type); i460_remove_memory()
515 return i460_remove_memory_large_io_page(mem, pg_start, type); i460_remove_memory()
H A Dsworks-agp.c319 static int serverworks_insert_memory(struct agp_memory *mem, serverworks_insert_memory() argument
328 if (type != 0 || mem->type != 0) { serverworks_insert_memory()
331 if ((pg_start + mem->page_count) > num_entries) { serverworks_insert_memory()
336 while (j < (pg_start + mem->page_count)) { serverworks_insert_memory()
344 if (!mem->is_flushed) { serverworks_insert_memory()
346 mem->is_flushed = true; serverworks_insert_memory()
349 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { serverworks_insert_memory()
353 page_to_phys(mem->pages[i]), mem->type), serverworks_insert_memory()
356 serverworks_tlbflush(mem); serverworks_insert_memory()
360 static int serverworks_remove_memory(struct agp_memory *mem, off_t pg_start, serverworks_remove_memory() argument
367 if (type != 0 || mem->type != 0) { serverworks_remove_memory()
372 serverworks_tlbflush(mem); serverworks_remove_memory()
374 for (i = pg_start; i < (mem->page_count + pg_start); i++) { serverworks_remove_memory()
380 serverworks_tlbflush(mem); serverworks_remove_memory()
H A Defficeon-agp.c109 static void efficeon_tlbflush(struct agp_memory * mem) efficeon_tlbflush() argument
238 static int efficeon_insert_memory(struct agp_memory * mem, off_t pg_start, int type) efficeon_insert_memory() argument
240 int i, count = mem->page_count, num_entries; efficeon_insert_memory()
248 if ((pg_start + mem->page_count) > num_entries) efficeon_insert_memory()
250 if (type != 0 || mem->type != 0) efficeon_insert_memory()
253 if (!mem->is_flushed) { efficeon_insert_memory()
255 mem->is_flushed = true; efficeon_insert_memory()
261 unsigned long insert = efficeon_mask_memory(mem->pages[i]); efficeon_insert_memory()
283 agp_bridge->driver->tlb_flush(mem); efficeon_insert_memory()
287 static int efficeon_remove_memory(struct agp_memory * mem, off_t pg_start, int type) efficeon_remove_memory() argument
289 int i, count = mem->page_count, num_entries; efficeon_remove_memory()
295 if ((pg_start + mem->page_count) > num_entries) efficeon_remove_memory()
297 if (type != 0 || mem->type != 0) efficeon_remove_memory()
309 agp_bridge->driver->tlb_flush(mem); efficeon_remove_memory()
H A Duninorth-agp.c73 static void uninorth_tlbflush(struct agp_memory *mem) uninorth_tlbflush() argument
83 if (!mem && uninorth_rev <= 0x30) { uninorth_tlbflush()
149 static int uninorth_insert_memory(struct agp_memory *mem, off_t pg_start, int type) uninorth_insert_memory() argument
156 if (type != mem->type) uninorth_insert_memory()
165 if (mem->page_count == 0) uninorth_insert_memory()
171 if ((pg_start + mem->page_count) > num_entries) uninorth_insert_memory()
175 for (i = 0; i < mem->page_count; ++i) { uninorth_insert_memory()
184 for (i = 0; i < mem->page_count; i++) { uninorth_insert_memory()
186 gp[i] = (page_to_phys(mem->pages[i]) >> PAGE_SHIFT) | 0x80000000UL; uninorth_insert_memory()
188 gp[i] = cpu_to_le32((page_to_phys(mem->pages[i]) & 0xFFFFF000UL) | uninorth_insert_memory()
190 flush_dcache_range((unsigned long)__va(page_to_phys(mem->pages[i])), uninorth_insert_memory()
191 (unsigned long)__va(page_to_phys(mem->pages[i]))+0x1000); uninorth_insert_memory()
194 uninorth_tlbflush(mem); uninorth_insert_memory()
199 int uninorth_remove_memory(struct agp_memory *mem, off_t pg_start, int type) uninorth_remove_memory() argument
205 if (type != mem->type) uninorth_remove_memory()
214 if (mem->page_count == 0) uninorth_remove_memory()
218 for (i = 0; i < mem->page_count; ++i) { uninorth_remove_memory()
222 uninorth_tlbflush(mem); uninorth_remove_memory()
H A Dgeneric.c89 void agp_alloc_page_array(size_t size, struct agp_memory *mem) agp_alloc_page_array() argument
91 mem->pages = NULL; agp_alloc_page_array()
94 mem->pages = kmalloc(size, GFP_KERNEL | __GFP_NOWARN); agp_alloc_page_array()
95 if (mem->pages == NULL) { agp_alloc_page_array()
96 mem->pages = vmalloc(size); agp_alloc_page_array()
1033 int agp_generic_insert_memory(struct agp_memory * mem, off_t pg_start, int type) agp_generic_insert_memory() argument
1042 bridge = mem->bridge; agp_generic_insert_memory()
1046 if (mem->page_count == 0) agp_generic_insert_memory()
1075 if (type != mem->type) agp_generic_insert_memory()
1084 if (((pg_start + mem->page_count) > num_entries) || agp_generic_insert_memory()
1085 ((pg_start + mem->page_count) < pg_start)) agp_generic_insert_memory()
1090 while (j < (pg_start + mem->page_count)) { agp_generic_insert_memory()
1096 if (!mem->is_flushed) { agp_generic_insert_memory()
1098 mem->is_flushed = true; agp_generic_insert_memory()
1101 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { agp_generic_insert_memory()
1103 page_to_phys(mem->pages[i]), agp_generic_insert_memory()
1109 bridge->driver->tlb_flush(mem); agp_generic_insert_memory()
1115 int agp_generic_remove_memory(struct agp_memory *mem, off_t pg_start, int type) agp_generic_remove_memory() argument
1121 bridge = mem->bridge; agp_generic_remove_memory()
1125 if (mem->page_count == 0) agp_generic_remove_memory()
1128 if (type != mem->type) agp_generic_remove_memory()
1132 if (((pg_start + mem->page_count) > num_entries) || agp_generic_remove_memory()
1133 ((pg_start + mem->page_count) < pg_start)) agp_generic_remove_memory()
1143 for (i = pg_start; i < (mem->page_count + pg_start); i++) { agp_generic_remove_memory()
1148 bridge->driver->tlb_flush(mem); agp_generic_remove_memory()
1195 int agp_generic_alloc_pages(struct agp_bridge_data *bridge, struct agp_memory *mem, size_t num_pages) agp_generic_alloc_pages() argument
1212 mem->pages[i] = page; agp_generic_alloc_pages()
1213 mem->page_count++; agp_generic_alloc_pages()
1217 set_pages_array_uc(mem->pages, num_pages); agp_generic_alloc_pages()
1241 void agp_generic_destroy_pages(struct agp_memory *mem) agp_generic_destroy_pages() argument
1246 if (!mem) agp_generic_destroy_pages()
1250 set_pages_array_wb(mem->pages, mem->page_count); agp_generic_destroy_pages()
1253 for (i = 0; i < mem->page_count; i++) { agp_generic_destroy_pages()
1254 page = mem->pages[i]; agp_generic_destroy_pages()
1262 mem->pages[i] = NULL; agp_generic_destroy_pages()
1371 void agp3_generic_tlbflush(struct agp_memory *mem) agp3_generic_tlbflush() argument
H A Damd-k7-agp.c283 static int amd_insert_memory(struct agp_memory *mem, off_t pg_start, int type) amd_insert_memory() argument
291 if (type != mem->type || amd_insert_memory()
295 if ((pg_start + mem->page_count) > num_entries) amd_insert_memory()
299 while (j < (pg_start + mem->page_count)) { amd_insert_memory()
307 if (!mem->is_flushed) { amd_insert_memory()
309 mem->is_flushed = true; amd_insert_memory()
312 for (i = 0, j = pg_start; i < mem->page_count; i++, j++) { amd_insert_memory()
316 page_to_phys(mem->pages[i]), amd_insert_memory()
317 mem->type), amd_insert_memory()
321 amd_irongate_tlbflush(mem); amd_insert_memory()
325 static int amd_remove_memory(struct agp_memory *mem, off_t pg_start, int type) amd_remove_memory() argument
331 if (type != mem->type || amd_remove_memory()
335 for (i = pg_start; i < (mem->page_count + pg_start); i++) { amd_remove_memory()
342 amd_irongate_tlbflush(mem); amd_remove_memory()
H A Dhp-agp.c282 hp_zx1_tlbflush (struct agp_memory *mem) hp_zx1_tlbflush() argument
332 hp_zx1_insert_memory (struct agp_memory *mem, off_t pg_start, int type) hp_zx1_insert_memory() argument
339 if (type != mem->type || hp_zx1_insert_memory()
345 io_pg_count = hp->io_pages_per_kpage * mem->page_count; hp_zx1_insert_memory()
358 if (!mem->is_flushed) { hp_zx1_insert_memory()
360 mem->is_flushed = true; hp_zx1_insert_memory()
363 for (i = 0, j = io_pg_start; i < mem->page_count; i++) { hp_zx1_insert_memory()
366 paddr = page_to_phys(mem->pages[i]); hp_zx1_insert_memory()
374 agp_bridge->driver->tlb_flush(mem); hp_zx1_insert_memory()
379 hp_zx1_remove_memory (struct agp_memory *mem, off_t pg_start, int type) hp_zx1_remove_memory() argument
384 if (type != mem->type || hp_zx1_remove_memory()
390 io_pg_count = hp->io_pages_per_kpage * mem->page_count; hp_zx1_remove_memory()
395 agp_bridge->driver->tlb_flush(mem); hp_zx1_remove_memory()
H A Dintel-gtt.c126 DBG("try unmapping %lu pages\n", (unsigned long)mem->page_count); intel_gtt_unmap_memory()
211 static int i810_insert_dcache_entries(struct agp_memory *mem, off_t pg_start, i810_insert_dcache_entries() argument
216 if ((pg_start + mem->page_count) i810_insert_dcache_entries()
220 if (!mem->is_flushed) i810_insert_dcache_entries()
223 for (i = pg_start; i < (pg_start + mem->page_count); i++) { i810_insert_dcache_entries()
352 return 0; /* no stolen mem on i81x */ intel_gtt_stolen_size()
881 static int intel_fake_agp_insert_entries(struct agp_memory *mem, intel_fake_agp_insert_entries() argument
894 return i810_insert_dcache_entries(mem, pg_start, type); intel_fake_agp_insert_entries()
896 if (mem->page_count == 0) intel_fake_agp_insert_entries()
899 if (pg_start + mem->page_count > intel_private.gtt_total_entries) intel_fake_agp_insert_entries()
902 if (type != mem->type) intel_fake_agp_insert_entries()
908 if (!mem->is_flushed) intel_fake_agp_insert_entries()
914 ret = intel_gtt_map_memory(mem->pages, mem->page_count, &st); intel_fake_agp_insert_entries()
919 mem->sg_list = st.sgl; intel_fake_agp_insert_entries()
920 mem->num_sg = st.nents; intel_fake_agp_insert_entries()
922 intel_gtt_insert_pages(pg_start, mem->page_count, mem->pages, intel_fake_agp_insert_entries()
928 mem->is_flushed = true; intel_fake_agp_insert_entries()
946 static int intel_fake_agp_remove_entries(struct agp_memory *mem, intel_fake_agp_remove_entries() argument
949 if (mem->page_count == 0) intel_fake_agp_remove_entries()
952 intel_gtt_clear_range(pg_start, mem->page_count); intel_fake_agp_remove_entries()
955 intel_gtt_unmap_memory(mem->sg_list, mem->num_sg); intel_fake_agp_remove_entries()
956 mem->sg_list = NULL; intel_fake_agp_remove_entries()
957 mem->num_sg = 0; intel_fake_agp_remove_entries()
1189 /* i945 is the last gpu to need phys mem (for overlay and cursors). */
/linux-4.1.27/drivers/gpu/drm/sis/
H A Dsis_mm.c86 drm_sis_mem_t *mem = data; sis_drm_alloc() local
108 mem->size = (mem->size + SIS_MM_ALIGN_MASK) >> SIS_MM_ALIGN_SHIFT; sis_drm_alloc()
112 mem->size, 0, sis_drm_alloc()
117 item->req.size = mem->size; sis_drm_alloc()
125 mem->size, 0, sis_drm_alloc()
141 mem->offset = ((pool == 0) ? sis_drm_alloc()
144 mem->free = user_key; sis_drm_alloc()
145 mem->size = mem->size << SIS_MM_ALIGN_SHIFT; sis_drm_alloc()
155 mem->offset = 0; sis_drm_alloc()
156 mem->size = 0; sis_drm_alloc()
157 mem->free = 0; sis_drm_alloc()
159 DRM_DEBUG("alloc %d, size = %ld, offset = %ld\n", pool, mem->size, sis_drm_alloc()
160 mem->offset); sis_drm_alloc()
168 drm_sis_mem_t *mem = data; sis_drm_free() local
172 obj = idr_find(&dev_priv->object_idr, mem->free); sis_drm_free()
178 idr_remove(&dev_priv->object_idr, mem->free); sis_drm_free()
188 DRM_DEBUG("free = 0x%lx\n", mem->free); sis_drm_free()
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx4/
H A Dicm.c59 pci_unmap_sg(dev->persist->pdev, chunk->mem, chunk->npages, mlx4_free_icm_pages()
63 __free_pages(sg_page(&chunk->mem[i]), mlx4_free_icm_pages()
64 get_order(chunk->mem[i].length)); mlx4_free_icm_pages()
73 chunk->mem[i].length, mlx4_free_icm_coherent()
74 lowmem_page_address(sg_page(&chunk->mem[i])), mlx4_free_icm_coherent()
75 sg_dma_address(&chunk->mem[i])); mlx4_free_icm_coherent()
97 static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order, mlx4_alloc_icm_pages() argument
109 sg_set_page(mem, page, PAGE_SIZE << order, 0); mlx4_alloc_icm_pages()
113 static int mlx4_alloc_icm_coherent(struct device *dev, struct scatterlist *mem, mlx4_alloc_icm_coherent() argument
117 &sg_dma_address(mem), gfp_mask); mlx4_alloc_icm_coherent()
121 sg_set_buf(mem, buf, PAGE_SIZE << order); mlx4_alloc_icm_coherent()
122 BUG_ON(mem->offset); mlx4_alloc_icm_coherent()
123 sg_dma_len(mem) = PAGE_SIZE << order; mlx4_alloc_icm_coherent()
167 sg_init_table(chunk->mem, MLX4_ICM_CHUNK_LEN); mlx4_alloc_icm()
178 &chunk->mem[chunk->npages], mlx4_alloc_icm()
181 ret = mlx4_alloc_icm_pages(&chunk->mem[chunk->npages], mlx4_alloc_icm()
197 chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->mem, mlx4_alloc_icm()
212 chunk->nsg = pci_map_sg(dev->persist->pdev, chunk->mem, mlx4_alloc_icm()
330 if (sg_dma_len(&chunk->mem[i]) > dma_offset) mlx4_table_find()
331 *dma_handle = sg_dma_address(&chunk->mem[i]) + mlx4_table_find()
333 dma_offset -= sg_dma_len(&chunk->mem[i]); mlx4_table_find()
340 if (chunk->mem[i].length > offset) { mlx4_table_find()
341 page = sg_page(&chunk->mem[i]); mlx4_table_find()
344 offset -= chunk->mem[i].length; mlx4_table_find()
/linux-4.1.27/drivers/gpu/drm/nouveau/
H A Dnouveau_sgdma.c27 nv04_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem) nv04_sgdma_bind() argument
30 struct nvkm_mem *node = mem->mm_node; nv04_sgdma_bind()
39 node->size = (mem->num_pages << PAGE_SHIFT) >> 12; nv04_sgdma_bind()
61 nv50_sgdma_bind(struct ttm_tt *ttm, struct ttm_mem_reg *mem) nv50_sgdma_bind() argument
64 struct nvkm_mem *node = mem->mm_node; nv50_sgdma_bind()
74 node->size = (mem->num_pages << PAGE_SHIFT) >> 12; nv50_sgdma_bind()
H A Dnouveau_ttm.c64 struct ttm_mem_reg *mem) nouveau_vram_manager_del()
68 nvkm_mem_node_cleanup(mem->mm_node); nouveau_vram_manager_del()
69 pfb->ram->put(pfb, (struct nvkm_mem **)&mem->mm_node); nouveau_vram_manager_del()
76 struct ttm_mem_reg *mem) nouveau_vram_manager_new()
91 ret = pfb->ram->get(pfb, mem->num_pages << PAGE_SHIFT, nouveau_vram_manager_new()
92 mem->page_alignment << PAGE_SHIFT, size_nc, nouveau_vram_manager_new()
95 mem->mm_node = NULL; nouveau_vram_manager_new()
101 mem->mm_node = node; nouveau_vram_manager_new()
102 mem->start = node->offset >> PAGE_SHIFT; nouveau_vram_manager_new()
154 struct ttm_mem_reg *mem) nouveau_gart_manager_del()
156 nvkm_mem_node_cleanup(mem->mm_node); nouveau_gart_manager_del()
157 kfree(mem->mm_node); nouveau_gart_manager_del()
158 mem->mm_node = NULL; nouveau_gart_manager_del()
165 struct ttm_mem_reg *mem) nouveau_gart_manager_new()
190 mem->mm_node = node; nouveau_gart_manager_new()
191 mem->start = 0; nouveau_gart_manager_new()
232 nv04_gart_manager_del(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) nv04_gart_manager_del() argument
234 struct nvkm_mem *node = mem->mm_node; nv04_gart_manager_del()
237 kfree(mem->mm_node); nv04_gart_manager_del()
238 mem->mm_node = NULL; nv04_gart_manager_del()
245 struct ttm_mem_reg *mem) nv04_gart_manager_new()
256 ret = nvkm_vm_get(man->priv, mem->num_pages << 12, node->page_shift, nv04_gart_manager_new()
263 mem->mm_node = node; nv04_gart_manager_new()
264 mem->start = node->vma[0].offset >> PAGE_SHIFT; nv04_gart_manager_new()
63 nouveau_vram_manager_del(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) nouveau_vram_manager_del() argument
73 nouveau_vram_manager_new(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) nouveau_vram_manager_new() argument
153 nouveau_gart_manager_del(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) nouveau_gart_manager_del() argument
162 nouveau_gart_manager_new(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) nouveau_gart_manager_new() argument
242 nv04_gart_manager_new(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) nv04_gart_manager_new() argument
H A Dnouveau_bo.c227 nvbo->bo.mem.num_pages = size >> PAGE_SHIFT; nouveau_bo_new()
268 nvbo->bo.mem.num_pages < vram_pages / 4) { set_placement_range()
327 if (bo->mem.mem_type == TTM_PL_VRAM) { nouveau_bo_pin()
328 struct nvkm_mem *mem = bo->mem.mm_node; nouveau_bo_pin() local
329 if (!list_is_singular(&mem->regions)) nouveau_bo_pin()
338 if (!(memtype & (1 << bo->mem.mem_type)) || evict) { nouveau_bo_pin()
341 1 << bo->mem.mem_type, memtype); nouveau_bo_pin()
368 switch (bo->mem.mem_type) { nouveau_bo_pin()
370 drm->gem.vram_available -= bo->mem.size; nouveau_bo_pin()
373 drm->gem.gart_available -= bo->mem.size; nouveau_bo_pin()
402 nouveau_bo_placement_set(nvbo, bo->mem.placement, 0); nouveau_bo_unpin()
406 switch (bo->mem.mem_type) { nouveau_bo_unpin()
408 drm->gem.vram_available += bo->mem.size; nouveau_bo_unpin()
411 drm->gem.gart_available += bo->mem.size; nouveau_bo_unpin()
437 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.mem.num_pages, nouveau_bo_map()
515 _nouveau_bo_mem_index(struct nouveau_bo *nvbo, unsigned index, void *mem, u8 sz) _nouveau_bo_mem_index() argument
518 u8 *m = mem; _nouveau_bo_mem_index()
540 u16 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem); nouveau_bo_wr16() local
542 mem = nouveau_bo_mem_index(nvbo, index, mem); nouveau_bo_wr16()
545 iowrite16_native(val, (void __force __iomem *)mem); nouveau_bo_wr16()
547 *mem = val; nouveau_bo_wr16()
554 u32 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem); nouveau_bo_rd32() local
556 mem = nouveau_bo_mem_index(nvbo, index, mem); nouveau_bo_rd32()
559 return ioread32_native((void __force __iomem *)mem); nouveau_bo_rd32()
561 return *mem; nouveau_bo_rd32()
568 u32 *mem = ttm_kmap_obj_virtual(&nvbo->kmap, &is_iomem); nouveau_bo_wr32() local
570 mem = nouveau_bo_mem_index(nvbo, index, mem); nouveau_bo_wr32()
573 iowrite32_native(val, (void __force __iomem *)mem); nouveau_bo_wr32()
575 *mem = val; nouveau_bo_wr32()
668 switch (bo->mem.mem_type) { nouveau_bo_evict_flags()
982 struct nouveau_channel *chan, struct ttm_mem_reg *mem) nouveau_bo_mem_ctxdma()
984 if (mem->mem_type == TTM_PL_TT) nouveau_bo_mem_ctxdma()
1037 struct ttm_mem_reg *mem) nouveau_bo_move_prep()
1039 struct nvkm_mem *old_node = bo->mem.mm_node; nouveau_bo_move_prep()
1040 struct nvkm_mem *new_node = mem->mm_node; nouveau_bo_move_prep()
1041 u64 size = (u64)mem->num_pages << PAGE_SHIFT; nouveau_bo_move_prep()
1084 ret = drm->ttm.move(chan, bo, &bo->mem, new_mem); nouveau_bo_move_m2mf()
1294 struct ttm_mem_reg *old_mem = &bo->mem; nouveau_bo_move()
1309 BUG_ON(bo->mem.mm_node != NULL); nouveau_bo_move()
1310 bo->mem = *new_mem; nouveau_bo_move()
1355 nouveau_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) nouveau_ttm_io_mem_reserve() argument
1357 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; nouveau_ttm_io_mem_reserve()
1359 struct nvkm_mem *node = mem->mm_node; nouveau_ttm_io_mem_reserve()
1362 mem->bus.addr = NULL; nouveau_ttm_io_mem_reserve()
1363 mem->bus.offset = 0; nouveau_ttm_io_mem_reserve()
1364 mem->bus.size = mem->num_pages << PAGE_SHIFT; nouveau_ttm_io_mem_reserve()
1365 mem->bus.base = 0; nouveau_ttm_io_mem_reserve()
1366 mem->bus.is_iomem = false; nouveau_ttm_io_mem_reserve()
1369 switch (mem->mem_type) { nouveau_ttm_io_mem_reserve()
1376 mem->bus.offset = mem->start << PAGE_SHIFT; nouveau_ttm_io_mem_reserve()
1377 mem->bus.base = drm->agp.base; nouveau_ttm_io_mem_reserve()
1378 mem->bus.is_iomem = !drm->dev->agp->cant_use_aperture; nouveau_ttm_io_mem_reserve()
1386 mem->bus.offset = mem->start << PAGE_SHIFT; nouveau_ttm_io_mem_reserve()
1387 mem->bus.base = nv_device_resource_start(nvxx_device(&drm->device), 1); nouveau_ttm_io_mem_reserve()
1388 mem->bus.is_iomem = true; nouveau_ttm_io_mem_reserve()
1397 mem->bus.offset = node->bar_vma.offset; nouveau_ttm_io_mem_reserve()
1407 nouveau_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) nouveau_ttm_io_mem_free() argument
1411 struct nvkm_mem *node = mem->mm_node; nouveau_ttm_io_mem_free()
1431 if (bo->mem.mem_type != TTM_PL_VRAM) { nouveau_ttm_fault_reserve_notify()
1436 if (bo->mem.mem_type == TTM_PL_SYSTEM) { nouveau_ttm_fault_reserve_notify()
1448 bo->mem.start + bo->mem.num_pages < mappable) nouveau_ttm_fault_reserve_notify()
1634 const u32 size = nvbo->bo.mem.num_pages << PAGE_SHIFT; nouveau_bo_vma_add()
1642 if ( nvbo->bo.mem.mem_type != TTM_PL_SYSTEM && nouveau_bo_vma_add()
1643 (nvbo->bo.mem.mem_type == TTM_PL_VRAM || nouveau_bo_vma_add()
1645 nvkm_vm_map(vma, nvbo->bo.mem.mm_node); nouveau_bo_vma_add()
1656 if (nvbo->bo.mem.mem_type != TTM_PL_SYSTEM) nouveau_bo_vma_del()
981 nouveau_bo_mem_ctxdma(struct ttm_buffer_object *bo, struct nouveau_channel *chan, struct ttm_mem_reg *mem) nouveau_bo_mem_ctxdma() argument
1036 nouveau_bo_move_prep(struct nouveau_drm *drm, struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) nouveau_bo_move_prep() argument
H A Dnv50_fence.c40 struct ttm_mem_reg *mem = &priv->bo->bo.mem; nv50_fence_context_new() local
41 u32 start = mem->start * PAGE_SIZE; nv50_fence_context_new()
42 u32 limit = start + mem->size - 1; nv50_fence_context_new()
66 u32 start = bo->bo.mem.start * PAGE_SIZE; nv50_fence_context_new()
67 u32 limit = start + bo->bo.mem.size - 1; nv50_fence_context_new()
H A Dnouveau_gem.c117 const bool mapped = nvbo->bo.mem.mem_type != TTM_PL_SYSTEM; nouveau_gem_object_unmap()
212 ret = drm_gem_object_init(dev, &nvbo->gem, nvbo->bo.mem.size); nouveau_gem_new()
232 else if (nvbo->bo.mem.mem_type == TTM_PL_TT) nouveau_gem_info()
245 rep->size = nvbo->bo.mem.num_pages << PAGE_SHIFT; nouveau_gem_info()
306 bo->mem.mem_type == TTM_PL_VRAM) nouveau_gem_set_domain()
310 bo->mem.mem_type == TTM_PL_TT) nouveau_gem_set_domain()
500 ((nvbo->bo.mem.mem_type == TTM_PL_VRAM && list_for_each_entry()
502 (nvbo->bo.mem.mem_type == TTM_PL_TT && list_for_each_entry()
506 if (nvbo->bo.mem.mem_type == TTM_PL_TT) list_for_each_entry()
568 void *mem; u_memcpya() local
573 mem = kmalloc(size, GFP_KERNEL | __GFP_NOWARN); u_memcpya()
574 if (!mem) u_memcpya()
575 mem = vmalloc(size); u_memcpya()
576 if (!mem) u_memcpya()
579 if (copy_from_user(mem, userptr, size)) { u_memcpya()
580 u_free(mem); u_memcpya()
584 return mem; u_memcpya()
624 nvbo->bo.mem.num_pages << PAGE_SHIFT)) { nouveau_gem_pushbuf_reloc_apply()
631 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.mem.num_pages, nouveau_gem_pushbuf_reloc_apply()
803 nvbo->bo.mem. nouveau_gem_ioctl_pushbuf()
H A Dnv17_fence.c78 struct ttm_mem_reg *mem = &priv->bo->bo.mem; nv17_fence_context_new() local
79 u32 start = mem->start * PAGE_SIZE; nv17_fence_context_new()
80 u32 limit = start + mem->size - 1; nv17_fence_context_new()
/linux-4.1.27/drivers/ata/
H A Dpata_palmld.c55 void __iomem *mem; palmld_pata_probe() local
66 mem = devm_ioremap(&pdev->dev, PALMLD_IDE_PHYS, 0x1000); palmld_pata_probe()
67 if (!mem) { palmld_pata_probe()
91 ap->ioaddr.cmd_addr = mem + 0x10; palmld_pata_probe()
92 ap->ioaddr.altstatus_addr = mem + 0xe; palmld_pata_probe()
93 ap->ioaddr.ctl_addr = mem + 0xe; palmld_pata_probe()
/linux-4.1.27/arch/powerpc/boot/
H A Dcuboot-pq2.c117 * the device tree. The prefetch mem region and non-prefetch mem region
120 * 32-bit PCI is supported. All three region types (prefetchable mem,
121 * non-prefetchable mem, and I/O) must be present.
125 struct pci_range *mem = NULL, *mmio = NULL, fixup_pci() local
165 mem = &pci_ranges_buf[i]; fixup_pci()
172 if (!mem || !mmio || !io) fixup_pci()
174 if (mem->size[1] != mmio->size[1]) fixup_pci()
176 if (mem->size[1] & (mem->size[1] - 1)) fixup_pci()
181 if (mem->phys_addr + mem->size[1] == mmio->phys_addr) fixup_pci()
182 mem_base = mem; fixup_pci()
183 else if (mmio->phys_addr + mmio->size[1] == mem->phys_addr) fixup_pci()
189 out_be32(&pci_regs[2][0], ~(mem->size[1] + mmio->size[1] - 1)); fixup_pci()
194 out_le32(&pci_regs[0][0], mem->pci_addr[1] >> 12); fixup_pci()
195 out_le32(&pci_regs[0][2], mem->phys_addr >> 12); fixup_pci()
196 out_le32(&pci_regs[0][4], (~(mem->size[1] - 1) >> 12) | 0xa0000000); fixup_pci()
H A Dwii.c106 void *mem; platform_fixups() local
112 mem = finddevice("/memory"); platform_fixups()
113 if (!mem) platform_fixups()
117 len = getprop(mem, "reg", reg, sizeof(reg)); platform_fixups()
133 setprop(mem, "reg", reg, sizeof(reg)); platform_fixups()
/linux-4.1.27/tools/power/cpupower/debug/i386/
H A Ddump_psb.c154 char *mem = NULL; main() local
178 fd = open("/dev/mem", O_RDONLY); main()
180 printf ("Couldn't open /dev/mem. Are you root?\n"); main()
184 mem = mmap(mem, 0x100000 - 0xc0000, PROT_READ, MAP_SHARED, fd, 0xc0000); main()
187 for (p = mem; p - mem < LEN; p+=16) { main()
194 munmap(mem, LEN); main()
/linux-4.1.27/arch/mips/bcm47xx/
H A Dprom.c57 unsigned long mem; prom_init_mem() local
81 for (mem = 1 << 20; mem < max; mem += 1 << 20) { prom_init_mem()
83 if (off + mem >= max) { prom_init_mem()
84 mem = max; prom_init_mem()
88 if (!memcmp(prom_init, prom_init + mem, 32)) prom_init_mem()
91 lowmem = mem; prom_init_mem()
98 if (c->cputype == CPU_74K && (mem == (128 << 20))) prom_init_mem()
99 mem -= 0x1000; prom_init_mem()
100 add_memory_region(0, mem, BOOT_MEM_RAM); prom_init_mem()
/linux-4.1.27/net/core/
H A Dsock_diag.c54 u32 mem[SK_MEMINFO_VARS]; sock_diag_put_meminfo() local
56 mem[SK_MEMINFO_RMEM_ALLOC] = sk_rmem_alloc_get(sk); sock_diag_put_meminfo()
57 mem[SK_MEMINFO_RCVBUF] = sk->sk_rcvbuf; sock_diag_put_meminfo()
58 mem[SK_MEMINFO_WMEM_ALLOC] = sk_wmem_alloc_get(sk); sock_diag_put_meminfo()
59 mem[SK_MEMINFO_SNDBUF] = sk->sk_sndbuf; sock_diag_put_meminfo()
60 mem[SK_MEMINFO_FWD_ALLOC] = sk->sk_forward_alloc; sock_diag_put_meminfo()
61 mem[SK_MEMINFO_WMEM_QUEUED] = sk->sk_wmem_queued; sock_diag_put_meminfo()
62 mem[SK_MEMINFO_OPTMEM] = atomic_read(&sk->sk_omem_alloc); sock_diag_put_meminfo()
63 mem[SK_MEMINFO_BACKLOG] = sk->sk_backlog.len; sock_diag_put_meminfo()
65 return nla_put(skb, attrtype, sizeof(mem), &mem); sock_diag_put_meminfo()
/linux-4.1.27/drivers/net/can/cc770/
H A Dcc770_isa.c25 * insmod cc770_isa.ko mem=0xd1000,0xd1000 irq=7,11
43 * insmod cc770_isa.ko mem=0xd1000,0xd1000 irq=7,11 clk=24000000
47 * insmod cc770_isa.ko mem=0xd1000,0xd1000 irq=7,11 clk=24000000,24000000
77 static unsigned long mem[MAXDEV]; variable
88 module_param_array(mem, ulong, NULL, S_IRUGO);
89 MODULE_PARM_DESC(mem, "I/O memory address");
179 dev_dbg(&pdev->dev, "probing idx=%d: port=%#lx, mem=%#lx, irq=%d\n", cc770_isa_probe()
180 idx, port[idx], mem[idx], irq[idx]); cc770_isa_probe()
181 if (mem[idx]) { cc770_isa_probe()
182 if (!request_mem_region(mem[idx], iosize, KBUILD_MODNAME)) { cc770_isa_probe()
186 base = ioremap_nocache(mem[idx], iosize); cc770_isa_probe()
210 if (mem[idx]) { cc770_isa_probe()
212 dev->base_addr = mem[idx]; cc770_isa_probe()
283 if (mem[idx]) cc770_isa_probe()
286 if (mem[idx]) cc770_isa_probe()
287 release_mem_region(mem[idx], iosize); cc770_isa_probe()
302 if (mem[idx]) { cc770_isa_remove()
304 release_mem_region(mem[idx], CC770_IOSIZE); cc770_isa_remove()
329 if ((port[idx] || mem[idx]) && irq[idx]) { cc770_isa_init()
341 pr_debug("platform device %d: port=%#lx, mem=%#lx, " cc770_isa_init()
343 idx, port[idx], mem[idx], irq[idx]); cc770_isa_init()
344 } else if (idx == 0 || port[idx] || mem[idx]) { cc770_isa_init()
H A Dcc770_platform.c171 struct resource *mem; cc770_platform_probe() local
176 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); cc770_platform_probe()
178 if (!mem || irq <= 0) cc770_platform_probe()
181 mem_size = resource_size(mem); cc770_platform_probe()
182 if (!request_mem_region(mem->start, mem_size, pdev->name)) cc770_platform_probe()
185 base = ioremap(mem->start, mem_size); cc770_platform_probe()
236 release_mem_region(mem->start, mem_size); cc770_platform_probe()
245 struct resource *mem; cc770_platform_remove() local
251 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); cc770_platform_remove()
252 release_mem_region(mem->start, resource_size(mem)); cc770_platform_remove()
/linux-4.1.27/drivers/gpu/drm/qxl/
H A Dqxl_ttm.c217 struct ttm_mem_reg *mem) qxl_ttm_io_mem_reserve()
219 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; qxl_ttm_io_mem_reserve()
222 mem->bus.addr = NULL; qxl_ttm_io_mem_reserve()
223 mem->bus.offset = 0; qxl_ttm_io_mem_reserve()
224 mem->bus.size = mem->num_pages << PAGE_SHIFT; qxl_ttm_io_mem_reserve()
225 mem->bus.base = 0; qxl_ttm_io_mem_reserve()
226 mem->bus.is_iomem = false; qxl_ttm_io_mem_reserve()
229 switch (mem->mem_type) { qxl_ttm_io_mem_reserve()
234 mem->bus.is_iomem = true; qxl_ttm_io_mem_reserve()
235 mem->bus.base = qdev->vram_base; qxl_ttm_io_mem_reserve()
236 mem->bus.offset = mem->start << PAGE_SHIFT; qxl_ttm_io_mem_reserve()
239 mem->bus.is_iomem = true; qxl_ttm_io_mem_reserve()
240 mem->bus.base = qdev->surfaceram_base; qxl_ttm_io_mem_reserve()
241 mem->bus.offset = mem->start << PAGE_SHIFT; qxl_ttm_io_mem_reserve()
250 struct ttm_mem_reg *mem) qxl_ttm_io_mem_free()
340 struct ttm_mem_reg *old_mem = &bo->mem; qxl_move_null()
352 struct ttm_mem_reg *old_mem = &bo->mem; qxl_bo_move()
371 if (bo->mem.mem_type == TTM_PL_PRIV0 && qbo->surface_id) qxl_bo_move_notify()
216 qxl_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) qxl_ttm_io_mem_reserve() argument
249 qxl_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) qxl_ttm_io_mem_free() argument
H A Dqxl_object.c147 struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; qxl_bo_kmap_atomic_page()
152 if (bo->tbo.mem.mem_type == TTM_PL_VRAM) qxl_bo_kmap_atomic_page()
154 else if (bo->tbo.mem.mem_type == TTM_PL_PRIV0) qxl_bo_kmap_atomic_page()
160 ret = ttm_mem_io_reserve(bo->tbo.bdev, &bo->tbo.mem); qxl_bo_kmap_atomic_page()
163 return io_mapping_map_atomic_wc(map, bo->tbo.mem.bus.offset + page_offset); qxl_bo_kmap_atomic_page()
189 struct ttm_mem_type_manager *man = &bo->tbo.bdev->man[bo->tbo.mem.mem_type]; qxl_bo_kunmap_atomic_page()
192 if (bo->tbo.mem.mem_type == TTM_PL_VRAM) qxl_bo_kunmap_atomic_page()
194 else if (bo->tbo.mem.mem_type == TTM_PL_PRIV0) qxl_bo_kunmap_atomic_page()
202 ttm_mem_io_free(bo->tbo.bdev, &bo->tbo.mem); qxl_bo_kunmap_atomic_page()
/linux-4.1.27/drivers/net/ethernet/intel/i40e/
H A Di40e_alloc.h47 struct i40e_dma_mem *mem,
51 struct i40e_dma_mem *mem);
53 struct i40e_virt_mem *mem,
56 struct i40e_virt_mem *mem);
H A Di40e_hmc.c51 struct i40e_dma_mem mem; i40e_add_sd_table_entry() local
78 ret_code = i40e_allocate_dma_mem(hw, &mem, mem_type, alloc_len, i40e_add_sd_table_entry()
92 sd_entry->u.pd_table.pd_page_addr = mem; i40e_add_sd_table_entry()
94 sd_entry->u.bp.addr = mem; i40e_add_sd_table_entry()
109 i40e_free_dma_mem(hw, &mem); i40e_add_sd_table_entry()
137 struct i40e_dma_mem mem; i40e_add_pd_table_entry() local
159 ret_code = i40e_allocate_dma_mem(hw, &mem, i40e_mem_bp, i40e_add_pd_table_entry()
165 pd_entry->bp.addr = mem; i40e_add_pd_table_entry()
169 page_desc = mem.pa | 0x1; i40e_add_pd_table_entry()
/linux-4.1.27/drivers/net/ethernet/intel/i40evf/
H A Di40e_alloc.h47 struct i40e_dma_mem *mem,
51 struct i40e_dma_mem *mem);
53 struct i40e_virt_mem *mem,
56 struct i40e_virt_mem *mem);
/linux-4.1.27/drivers/gpu/drm/mgag200/
H A Dmgag200_main.c79 static int mga_probe_vram(struct mga_device *mdev, void __iomem *mem) mga_probe_vram() argument
87 orig = ioread16(mem); mga_probe_vram()
88 iowrite16(0, mem); mga_probe_vram()
91 orig1 = ioread8(mem + offset); mga_probe_vram()
92 orig2 = ioread8(mem + offset + 0x100); mga_probe_vram()
94 iowrite16(0xaa55, mem + offset); mga_probe_vram()
95 iowrite16(0xaa55, mem + offset + 0x100); mga_probe_vram()
97 test1 = ioread16(mem + offset); mga_probe_vram()
98 test2 = ioread16(mem); mga_probe_vram()
100 iowrite16(orig1, mem + offset); mga_probe_vram()
101 iowrite16(orig2, mem + offset + 0x100); mga_probe_vram()
112 iowrite16(orig, mem); mga_probe_vram()
119 void __iomem *mem; mga_vram_init() local
140 mem = pci_iomap(mdev->dev->pdev, 0, 0); mga_vram_init()
142 mdev->mc.vram_size = mga_probe_vram(mdev, mem); mga_vram_init()
144 pci_iounmap(mdev->dev->pdev, mem); mga_vram_init()
H A Dmgag200_ttm.c157 struct ttm_mem_reg *mem) mgag200_ttm_io_mem_reserve()
159 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; mgag200_ttm_io_mem_reserve()
162 mem->bus.addr = NULL; mgag200_ttm_io_mem_reserve()
163 mem->bus.offset = 0; mgag200_ttm_io_mem_reserve()
164 mem->bus.size = mem->num_pages << PAGE_SHIFT; mgag200_ttm_io_mem_reserve()
165 mem->bus.base = 0; mgag200_ttm_io_mem_reserve()
166 mem->bus.is_iomem = false; mgag200_ttm_io_mem_reserve()
169 switch (mem->mem_type) { mgag200_ttm_io_mem_reserve()
174 mem->bus.offset = mem->start << PAGE_SHIFT; mgag200_ttm_io_mem_reserve()
175 mem->bus.base = pci_resource_start(mdev->dev->pdev, 0); mgag200_ttm_io_mem_reserve()
176 mem->bus.is_iomem = true; mgag200_ttm_io_mem_reserve()
185 static void mgag200_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) mgag200_ttm_io_mem_free() argument
156 mgag200_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) mgag200_ttm_io_mem_reserve() argument
/linux-4.1.27/drivers/usb/host/
H A Dohci-sm501.c91 struct resource *res, *mem; ohci_hcd_sm501_drv_probe() local
99 mem = platform_get_resource(pdev, IORESOURCE_MEM, 1); ohci_hcd_sm501_drv_probe()
100 if (mem == NULL) { ohci_hcd_sm501_drv_probe()
106 if (!request_mem_region(mem->start, resource_size(mem), pdev->name)) { ohci_hcd_sm501_drv_probe()
126 if (!dma_declare_coherent_memory(dev, mem->start, ohci_hcd_sm501_drv_probe()
127 mem->start - mem->parent->start, ohci_hcd_sm501_drv_probe()
128 resource_size(mem), ohci_hcd_sm501_drv_probe()
188 release_mem_region(mem->start, resource_size(mem)); ohci_hcd_sm501_drv_probe()
196 struct resource *mem; ohci_hcd_sm501_drv_remove() local
202 mem = platform_get_resource(pdev, IORESOURCE_MEM, 1); ohci_hcd_sm501_drv_remove()
203 if (mem) ohci_hcd_sm501_drv_remove()
204 release_mem_region(mem->start, resource_size(mem)); ohci_hcd_sm501_drv_remove()
/linux-4.1.27/arch/arm/mach-sa1100/
H A Dpci-nanoengine.c76 pci 0000:00:00.0: reg 10: [mem 0x00021000-0x00021fff]
78 pci 0000:00:00.0: reg 18: [mem 0x00000000-0x0001ffff]
79 pci 0000:00:00.0: reg 30: [mem 0x00000000-0x000fffff pref]
84 pci 0000:00:00.0: BAR 6: can't assign mem pref (size 0x100000)
85 pci 0000:00:00.0: BAR 2: assigned [mem 0x18600000-0x1861ffff]
86 pci 0000:00:00.0: BAR 2: set to [mem 0x18600000-0x1861ffff] (PCI address [0x0-0x1ffff])
87 pci 0000:00:00.0: BAR 0: assigned [mem 0x18620000-0x18620fff]
88 pci 0000:00:00.0: BAR 0: set to [mem 0x18620000-0x18620fff] (PCI address [0x20000-0x20fff])
102 pci 0000:00:00.0: reg 10: [mem 0x00021000-0x00021fff]
104 pci 0000:00:00.0: reg 18: [mem 0x00000000-0x0001ffff]
105 pci 0000:00:00.0: reg 30: [mem 0x00000000-0x000fffff pref]
110 pci 0000:00:00.0: BAR 6: assigned [mem 0x78000000-0x780fffff pref]
111 pci 0000:00:00.0: BAR 2: assigned [mem 0x18600000-0x1861ffff]
112 pci 0000:00:00.0: BAR 2: set to [mem 0x18600000-0x1861ffff] (PCI address [0x0-0x1ffff])
113 pci 0000:00:00.0: BAR 0: assigned [mem 0x18620000-0x18620fff]
114 pci 0000:00:00.0: BAR 0: set to [mem 0x18620000-0x18620fff] (PCI address [0x20000-0x20fff])
/linux-4.1.27/drivers/infiniband/ulp/iser/
H A Diser_memory.c391 iser_reg_dma(struct iser_device *device, struct iser_data_buf *mem, iser_reg_dma() argument
394 struct scatterlist *sg = mem->sg; iser_reg_dma()
409 struct iser_data_buf *mem, fall_to_bounce_buf()
418 aligned_len, mem->size); fall_to_bounce_buf()
421 iser_data_buf_dump(mem, device->ib_device); fall_to_bounce_buf()
424 iser_dma_unmap_task_data(iser_task, mem, fall_to_bounce_buf()
430 if (iser_start_rdma_unaligned_sg(iser_task, mem, cmd_dir) != 0) fall_to_bounce_buf()
443 struct iser_data_buf *mem, iser_reg_page_vec()
452 plen = iser_sg_to_page_vec(mem, device->ib_device, iser_reg_page_vec()
459 iser_data_buf_dump(mem, device->ib_device); iser_reg_page_vec()
530 struct iser_data_buf *mem = &iser_task->data[cmd_dir]; iser_reg_rdma_mem_fmr() local
538 aligned_len = iser_data_buf_aligned_len(mem, ibdev); iser_reg_rdma_mem_fmr()
539 if (aligned_len != mem->dma_nents) { iser_reg_rdma_mem_fmr()
540 err = fall_to_bounce_buf(iser_task, mem, iser_reg_rdma_mem_fmr()
549 if (mem->dma_nents == 1) { iser_reg_rdma_mem_fmr()
550 return iser_reg_dma(device, mem, mem_reg); iser_reg_rdma_mem_fmr()
552 err = iser_reg_page_vec(iser_task, mem, ib_conn->fmr.page_vec, iser_reg_rdma_mem_fmr()
555 iser_data_buf_dump(mem, ibdev); iser_reg_rdma_mem_fmr()
556 iser_err("mem->dma_nents = %d (dlength = 0x%x)\n", iser_reg_rdma_mem_fmr()
557 mem->dma_nents, iser_reg_rdma_mem_fmr()
597 sig_attrs->mem.sig_type = IB_SIG_TYPE_NONE; iser_set_sig_attrs()
604 iser_set_dif_domain(sc, sig_attrs, &sig_attrs->mem); iser_set_sig_attrs()
605 sig_attrs->mem.sig.dif.bg_type = sc->prot_flags & SCSI_PROT_IP_CHECKSUM ? iser_set_sig_attrs()
612 iser_set_dif_domain(sc, sig_attrs, &sig_attrs->mem); iser_set_sig_attrs()
613 sig_attrs->mem.sig.dif.bg_type = sc->prot_flags & SCSI_PROT_IP_CHECKSUM ? iser_set_sig_attrs()
713 struct iser_data_buf *mem, iser_fast_reg_mr()
727 if (mem->dma_nents == 1) iser_fast_reg_mr()
728 return iser_reg_dma(device, mem, reg); iser_fast_reg_mr()
738 plen = iser_sg_to_page_vec(mem, device->ib_device, frpl->page_list, iser_fast_reg_mr()
796 struct iser_data_buf *mem = &iser_task->data[cmd_dir]; iser_reg_rdma_mem_fastreg() local
801 aligned_len = iser_data_buf_aligned_len(mem, ibdev); iser_reg_rdma_mem_fastreg()
802 if (aligned_len != mem->dma_nents) { iser_reg_rdma_mem_fastreg()
803 err = fall_to_bounce_buf(iser_task, mem, iser_reg_rdma_mem_fastreg()
811 if (mem->dma_nents != 1 || iser_reg_rdma_mem_fastreg()
817 err = iser_fast_reg_mr(iser_task, mem, desc, iser_reg_rdma_mem_fastreg()
827 mem = &iser_task->prot[cmd_dir]; iser_reg_rdma_mem_fastreg()
828 aligned_len = iser_data_buf_aligned_len(mem, ibdev); iser_reg_rdma_mem_fastreg()
829 if (aligned_len != mem->dma_nents) { iser_reg_rdma_mem_fastreg()
830 err = fall_to_bounce_buf(iser_task, mem, iser_reg_rdma_mem_fastreg()
838 err = iser_fast_reg_mr(iser_task, mem, desc, iser_reg_rdma_mem_fastreg()
408 fall_to_bounce_buf(struct iscsi_iser_task *iser_task, struct iser_data_buf *mem, enum iser_data_dir cmd_dir, int aligned_len) fall_to_bounce_buf() argument
442 iser_reg_page_vec(struct iscsi_iser_task *iser_task, struct iser_data_buf *mem, struct iser_page_vec *page_vec, struct iser_mem_reg *mem_reg) iser_reg_page_vec() argument
712 iser_fast_reg_mr(struct iscsi_iser_task *iser_task, struct iser_data_buf *mem, struct fast_reg_descriptor *desc, enum iser_reg_indicator ind, struct iser_mem_reg *reg) iser_fast_reg_mr() argument
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/subdev/instmem/
H A Dnv50.c36 struct nvkm_mem *mem; member in struct:nv50_instobj_priv
49 u64 base = (node->mem->offset + offset) & 0xffffff00000ULL; nv50_instobj_rd32()
50 u64 addr = (node->mem->offset + offset) & 0x000000fffffULL; nv50_instobj_rd32()
69 u64 base = (node->mem->offset + offset) & 0xffffff00000ULL; nv50_instobj_wr32()
70 u64 addr = (node->mem->offset + offset) & 0x000000fffffULL; nv50_instobj_wr32()
86 pfb->ram->put(pfb, &node->mem); nv50_instobj_dtor()
108 ret = pfb->ram->get(pfb, args->size, args->align, 0, 0x800, &node->mem); nv50_instobj_ctor()
112 node->base.addr = node->mem->offset; nv50_instobj_ctor()
113 node->base.size = node->mem->size << 12; nv50_instobj_ctor()
114 node->mem->page_shift = 12; nv50_instobj_ctor()
H A Dnv04.h29 struct nvkm_mm_node *mem; member in struct:nv04_instobj_priv
H A Dgk20a.c56 struct nvkm_mem *mem; member in struct:gk20a_instobj_priv
57 /* Pointed by mem */
78 /* array of base.mem->size pages */
112 u64 base = (node->mem->offset + offset) & 0xffffff00000ULL; gk20a_instobj_rd32()
113 u64 addr = (node->mem->offset + offset) & 0x000000fffffULL; gk20a_instobj_rd32()
132 u64 base = (node->mem->offset + offset) & 0xffffff00000ULL; gk20a_instobj_wr32()
133 u64 addr = (node->mem->offset + offset) & 0x000000fffffULL; gk20a_instobj_wr32()
154 dma_free_attrs(dev, _node->mem->size << PAGE_SHIFT, node->cpuaddr, gk20a_instobj_dtor_dma()
166 if (unlikely(list_empty(&_node->mem->regions))) gk20a_instobj_dtor_iommu()
169 r = list_first_entry(&_node->mem->regions, struct nvkm_mm_node, gk20a_instobj_dtor_iommu()
176 for (i = 0; i < _node->mem->size; i++) { gk20a_instobj_dtor_iommu()
351 node->mem = &node->_mem; gk20a_instobj_ctor()
354 node->mem->size = size >> 12; gk20a_instobj_ctor()
355 node->mem->memtype = 0; gk20a_instobj_ctor()
356 node->mem->page_shift = 12; gk20a_instobj_ctor()
358 node->base.addr = node->mem->offset; gk20a_instobj_ctor()
362 size, align, node->mem->offset); gk20a_instobj_ctor()
H A Dnv04.c37 return nv_ro32(priv, node->mem->offset + addr); nv04_instobj_rd32()
45 nv_wo32(priv, node->mem->offset + addr, data); nv04_instobj_wr32()
56 nvkm_mm_free(&priv->heap, &node->mem); nv04_instobj_dtor()
83 args->align, &node->mem); nv04_instobj_ctor()
88 node->base.addr = node->mem->offset; nv04_instobj_ctor()
89 node->base.size = node->mem->length; nv04_instobj_ctor()
/linux-4.1.27/drivers/gpu/drm/vmwgfx/
H A Dvmwgfx_gmrid_manager.c50 struct ttm_mem_reg *mem) vmw_gmrid_man_get_node()
57 mem->mm_node = NULL; vmw_gmrid_man_get_node()
84 mem->mm_node = gman; vmw_gmrid_man_get_node()
85 mem->start = id; vmw_gmrid_man_get_node()
86 mem->num_pages = bo->num_pages; vmw_gmrid_man_get_node()
102 struct ttm_mem_reg *mem) vmw_gmrid_man_put_node()
107 if (mem->mm_node) { vmw_gmrid_man_put_node()
109 ida_remove(&gman->gmr_ida, mem->start); vmw_gmrid_man_put_node()
110 gman->used_gmr_pages -= mem->num_pages; vmw_gmrid_man_put_node()
112 mem->mm_node = NULL; vmw_gmrid_man_put_node()
47 vmw_gmrid_man_get_node(struct ttm_mem_type_manager *man, struct ttm_buffer_object *bo, const struct ttm_place *place, struct ttm_mem_reg *mem) vmw_gmrid_man_get_node() argument
101 vmw_gmrid_man_put_node(struct ttm_mem_type_manager *man, struct ttm_mem_reg *mem) vmw_gmrid_man_put_node() argument
H A Dvmwgfx_buffer.c767 static int vmw_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) vmw_ttm_io_mem_reserve() argument
769 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; vmw_ttm_io_mem_reserve()
772 mem->bus.addr = NULL; vmw_ttm_io_mem_reserve()
773 mem->bus.is_iomem = false; vmw_ttm_io_mem_reserve()
774 mem->bus.offset = 0; vmw_ttm_io_mem_reserve()
775 mem->bus.size = mem->num_pages << PAGE_SHIFT; vmw_ttm_io_mem_reserve()
776 mem->bus.base = 0; vmw_ttm_io_mem_reserve()
779 switch (mem->mem_type) { vmw_ttm_io_mem_reserve()
785 mem->bus.offset = mem->start << PAGE_SHIFT; vmw_ttm_io_mem_reserve()
786 mem->bus.base = dev_priv->vram_start; vmw_ttm_io_mem_reserve()
787 mem->bus.is_iomem = true; vmw_ttm_io_mem_reserve()
795 static void vmw_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) vmw_ttm_io_mem_free() argument
808 * @mem: The truct ttm_mem_reg indicating to what memory
815 struct ttm_mem_reg *mem) vmw_move_notify()
817 vmw_resource_move_notify(bo, mem); vmw_move_notify()
814 vmw_move_notify(struct ttm_buffer_object *bo, struct ttm_mem_reg *mem) vmw_move_notify() argument
/linux-4.1.27/drivers/pnp/
H A Dsupport.c67 return "mem"; pnp_resource_type_name()
109 struct pnp_mem *mem; dbg_pnp_show_option() local
133 mem = &option->u.mem; dbg_pnp_show_option()
134 len += scnprintf(buf + len, sizeof(buf) - len, "mem min %#llx " dbg_pnp_show_option()
136 (unsigned long long) mem->min, dbg_pnp_show_option()
137 (unsigned long long) mem->max, dbg_pnp_show_option()
138 (unsigned long long) mem->align, dbg_pnp_show_option()
139 (unsigned long long) mem->size, mem->flags); dbg_pnp_show_option()
H A Dinterface.c152 struct pnp_mem *mem) pnp_print_mem()
157 space, (unsigned long long) mem->min, pnp_print_mem()
158 (unsigned long long) mem->max, pnp_print_mem()
159 (unsigned long long) mem->align, pnp_print_mem()
160 (unsigned long long) mem->size); pnp_print_mem()
161 if (mem->flags & IORESOURCE_MEM_WRITEABLE) pnp_print_mem()
163 if (mem->flags & IORESOURCE_MEM_CACHEABLE) pnp_print_mem()
165 if (mem->flags & IORESOURCE_MEM_RANGELENGTH) pnp_print_mem()
167 if (mem->flags & IORESOURCE_MEM_SHADOWABLE) pnp_print_mem()
169 if (mem->flags & IORESOURCE_MEM_EXPANSIONROM) pnp_print_mem()
171 switch (mem->flags & IORESOURCE_MEM_TYPE_MASK) { pnp_print_mem()
195 pnp_print_mem(buffer, space, &option->u.mem); pnp_print_option()
401 } else if (!strncasecmp(buf, "mem", 3)) { resources_store()
151 pnp_print_mem(pnp_info_buffer_t * buffer, char *space, struct pnp_mem *mem) pnp_print_mem() argument
/linux-4.1.27/include/net/
H A Dinet_frag.h7 /* The percpu_counter "mem" need to be cacheline aligned.
8 * mem.count must not share cacheline with other writers
10 struct percpu_counter mem ____cacheline_aligned_in_smp;
131 * fragmentation mem acct sizes.
132 * The mem size of a 64K fragment is approx:
139 return percpu_counter_read(&nf->mem); frag_mem_limit()
144 __percpu_counter_add(&q->net->mem, -i, frag_percpu_counter_batch); sub_frag_mem_limit()
149 __percpu_counter_add(&q->net->mem, i, frag_percpu_counter_batch); add_frag_mem_limit()
154 percpu_counter_init(&nf->mem, 0, GFP_KERNEL); init_frag_mem_limit()
162 res = percpu_counter_sum_positive(&nf->mem); sum_frag_mem_limit()
/linux-4.1.27/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
H A Dramnv1a.c36 u32 mem, mib; nv1a_ram_create() local
51 pci_read_config_dword(bridge, 0x7c, &mem); nv1a_ram_create()
52 mib = ((mem >> 6) & 31) + 1; nv1a_ram_create()
54 pci_read_config_dword(bridge, 0x84, &mem); nv1a_ram_create()
55 mib = ((mem >> 4) & 127) + 1; nv1a_ram_create()
H A Dramnv50.c222 __nv50_ram_put(struct nvkm_fb *pfb, struct nvkm_mem *mem) __nv50_ram_put() argument
226 while (!list_empty(&mem->regions)) { __nv50_ram_put()
227 this = list_first_entry(&mem->regions, typeof(*this), rl_entry); __nv50_ram_put()
233 nvkm_mm_free(&pfb->tags, &mem->tag); __nv50_ram_put()
239 struct nvkm_mem *mem = *pmem; nv50_ram_put() local
242 if (unlikely(mem == NULL)) nv50_ram_put()
246 __nv50_ram_put(pfb, mem); nv50_ram_put()
249 kfree(mem); nv50_ram_put()
259 struct nvkm_mem *mem; nv50_ram_get() local
269 mem = kzalloc(sizeof(*mem), GFP_KERNEL); nv50_ram_get()
270 if (!mem) nv50_ram_get()
278 ret = nvkm_mm_head(tags, 0, 1, n, n, 1, &mem->tag); nv50_ram_get()
280 mem->tag = NULL; nv50_ram_get()
283 if (unlikely(!mem->tag)) nv50_ram_get()
287 INIT_LIST_HEAD(&mem->regions); nv50_ram_get()
288 mem->memtype = (comp << 7) | type; nv50_ram_get()
289 mem->size = max; nv50_ram_get()
299 pfb->ram->put(pfb, &mem); nv50_ram_get()
303 list_add_tail(&r->rl_entry, &mem->regions); nv50_ram_get()
308 r = list_first_entry(&mem->regions, struct nvkm_mm_node, rl_entry); nv50_ram_get()
309 mem->offset = (u64)r->offset << 12; nv50_ram_get()
310 *pmem = mem; nv50_ram_get()
H A Dramgf100.c428 struct nvkm_mem *mem = *pmem; gf100_ram_put() local
431 if (unlikely(mem == NULL)) gf100_ram_put()
435 if (mem->tag) gf100_ram_put()
436 ltc->tags_free(ltc, &mem->tag); gf100_ram_put()
437 __nv50_ram_put(pfb, mem); gf100_ram_put()
440 kfree(mem); gf100_ram_put()
449 struct nvkm_mem *mem; gf100_ram_get() local
461 mem = kzalloc(sizeof(*mem), GFP_KERNEL); gf100_ram_get()
462 if (!mem) gf100_ram_get()
465 INIT_LIST_HEAD(&mem->regions); gf100_ram_get()
466 mem->size = size; gf100_ram_get()
475 ltc->tags_alloc(ltc, n, &mem->tag); gf100_ram_get()
478 if (unlikely(!mem->tag)) gf100_ram_get()
481 mem->memtype = type; gf100_ram_get()
490 pfb->ram->put(pfb, &mem); gf100_ram_get()
494 list_add_tail(&r->rl_entry, &mem->regions); gf100_ram_get()
499 r = list_first_entry(&mem->regions, struct nvkm_mm_node, rl_entry); gf100_ram_get()
500 mem->offset = (u64)r->offset << 12; gf100_ram_get()
501 *pmem = mem; gf100_ram_get()
/linux-4.1.27/arch/mips/include/asm/mach-jazz/
H A Dfloppy.h107 unsigned long mem; fd_dma_mem_alloc() local
109 mem = __get_dma_pages(GFP_KERNEL, get_order(size)); fd_dma_mem_alloc()
110 if(!mem) fd_dma_mem_alloc()
112 vdma_alloc(CPHYSADDR(mem), size); /* XXX error checking */ fd_dma_mem_alloc()
114 return mem; fd_dma_mem_alloc()
/linux-4.1.27/drivers/rtc/
H A Drtc-pcf8583.c135 static int pcf8583_read_mem(struct i2c_client *client, struct rtc_mem *mem) pcf8583_read_mem() argument
147 .len = mem->nr, pcf8583_read_mem()
148 .buf = mem->data, pcf8583_read_mem()
152 if (mem->loc < 8) pcf8583_read_mem()
155 addr[0] = mem->loc; pcf8583_read_mem()
160 static int pcf8583_write_mem(struct i2c_client *client, struct rtc_mem *mem) pcf8583_write_mem() argument
165 if (mem->loc < 8 || mem->nr > 8) pcf8583_write_mem()
168 buf[0] = mem->loc; pcf8583_write_mem()
169 memcpy(buf + 1, mem->data, mem->nr); pcf8583_write_mem()
171 ret = i2c_master_send(client, buf, mem->nr + 1); pcf8583_write_mem()
172 return ret == mem->nr + 1 ? 0 : -EIO; pcf8583_write_mem()
179 struct rtc_mem mem = { pcf8583_rtc_read_time() local
202 pcf8583_read_mem(client, &mem)) pcf8583_rtc_read_time()
/linux-4.1.27/drivers/gpu/drm/via/
H A Dvia_mm.c113 drm_via_mem_t *mem = data; via_mem_alloc() local
120 if (mem->type > VIA_MEM_AGP) { via_mem_alloc()
125 if (0 == ((mem->type == VIA_MEM_VIDEO) ? dev_priv->vram_initialized : via_mem_alloc()
139 tmpSize = (mem->size + VIA_MM_ALIGN_MASK) >> VIA_MM_ALIGN_SHIFT; via_mem_alloc()
140 if (mem->type == VIA_MEM_AGP) via_mem_alloc()
159 mem->offset = ((mem->type == VIA_MEM_VIDEO) ? via_mem_alloc()
162 mem->index = user_key; via_mem_alloc()
172 mem->offset = 0; via_mem_alloc()
173 mem->size = 0; via_mem_alloc()
174 mem->index = 0; via_mem_alloc()
183 drm_via_mem_t *mem = data; via_mem_free() local
187 obj = idr_find(&dev_priv->object_idr, mem->index); via_mem_free()
193 idr_remove(&dev_priv->object_idr, mem->index); via_mem_free()
199 DRM_DEBUG("free = 0x%lx\n", mem->index); via_mem_free()
/linux-4.1.27/drivers/irqchip/
H A Dirq-tb10x.c111 struct resource mem; of_tb10x_init_irq() local
116 if (of_address_to_resource(ictl, 0, &mem)) { of_tb10x_init_irq()
122 if (!request_mem_region(mem.start, resource_size(&mem), of_tb10x_init_irq()
124 pr_err("%s: Request mem region failed.\n", ictl->name); of_tb10x_init_irq()
128 reg_base = ioremap(mem.start, resource_size(&mem)); of_tb10x_init_irq()
192 release_mem_region(mem.start, resource_size(&mem)); of_tb10x_init_irq()
/linux-4.1.27/include/ras/
H A Dras_event.h26 TP_PROTO(struct cper_sec_mem_err *mem,
32 TP_ARGS(mem, err_seq, fru_id, fru_text, sev),
47 if (mem->validation_bits & CPER_MEM_VALID_ERROR_TYPE)
48 __entry->etype = mem->error_type;
52 if (mem->validation_bits & CPER_MEM_VALID_PA)
53 __entry->pa = mem->physical_addr;
57 if (mem->validation_bits & CPER_MEM_VALID_PA_MASK)
58 __entry->pa_mask_lsb = (u8)__ffs64(mem->physical_addr_mask);
63 cper_mem_err_pack(mem, &__entry->data);
/linux-4.1.27/arch/alpha/include/asm/
H A Datomic.h139 "1: ldl_l %[old],%[mem]\n" __atomic_add_unless()
143 " stl_c %[new],%[mem]\n" __atomic_add_unless()
150 : [mem] "m"(*v), [a] "rI"(a), [u] "rI"((long)u) __atomic_add_unless()
171 "1: ldq_l %[tmp],%[mem]\n" atomic64_add_unless()
175 " stq_c %[tmp],%[mem]\n" atomic64_add_unless()
182 : [mem] "m"(*v), [a] "rI"(a), [u] "rI"(u) atomic64_add_unless()
200 "1: ldq_l %[old],%[mem]\n" atomic64_dec_if_positive()
203 " stq_c %[tmp],%[mem]\n" atomic64_dec_if_positive()
210 : [mem] "m"(*v) atomic64_dec_if_positive()
/linux-4.1.27/Documentation/ia64/
H A Daliasing-test.c2 * Exercise /dev/mem mmap cases that have been troublesome in the past
213 if (map_mem("/dev/mem", 0, 0xA0000, 1) == 0) main()
214 fprintf(stderr, "PASS: /dev/mem 0x0-0xa0000 is readable\n"); main()
216 fprintf(stderr, "FAIL: /dev/mem 0x0-0xa0000 not accessible\n"); main()
225 if (map_mem("/dev/mem", 0xA0000, 0x20000, 0) == 0) main()
226 fprintf(stderr, "PASS: /dev/mem 0xa0000-0xc0000 is mappable\n"); main()
228 fprintf(stderr, "FAIL: /dev/mem 0xa0000-0xc0000 not accessible\n"); main()
230 if (map_mem("/dev/mem", 0xC0000, 0x40000, 1) == 0) main()
231 fprintf(stderr, "PASS: /dev/mem 0xc0000-0x100000 is readable\n"); main()
233 fprintf(stderr, "FAIL: /dev/mem 0xc0000-0x100000 not accessible\n"); main()
242 rc = map_mem("/dev/mem", 0, 1024*1024, 0); main()
244 fprintf(stderr, "PASS: /dev/mem 0x0-0x100000 is mappable\n"); main()
246 fprintf(stderr, "PASS: /dev/mem 0x0-0x100000 not mappable\n"); main()
248 fprintf(stderr, "FAIL: /dev/mem 0x0-0x100000 not accessible\n"); main()
/linux-4.1.27/arch/x86/realmode/
H A Dinit.c13 phys_addr_t mem; reserve_real_mode() local
18 mem = memblock_find_in_range(0, 1<<20, size, PAGE_SIZE); reserve_real_mode()
19 if (!mem) reserve_real_mode()
22 base = __va(mem); reserve_real_mode()
23 memblock_reserve(mem, size); reserve_real_mode()
26 base, (unsigned long long)mem, size); reserve_real_mode()
/linux-4.1.27/arch/ia64/hp/sim/boot/
H A Dbootloader.c57 static char mem[4096]; start_bootloader() local
108 req.len = sizeof(mem); start_bootloader()
109 req.addr = (long) mem; start_bootloader()
113 elf = (struct elfhdr *) mem; start_bootloader()
137 req.addr = (long) mem; start_bootloader()
146 elf_phdr = (struct elf_phdr *) mem; start_bootloader()
/linux-4.1.27/drivers/gpu/drm/radeon/
H A Dradeon_kfd.c203 struct kgd_mem **mem = (struct kgd_mem **) mem_obj; alloc_gtt_mem() local
210 *mem = kmalloc(sizeof(struct kgd_mem), GFP_KERNEL); alloc_gtt_mem()
211 if ((*mem) == NULL) alloc_gtt_mem()
215 RADEON_GEM_GTT_WC, NULL, NULL, &(*mem)->bo); alloc_gtt_mem()
223 r = radeon_bo_reserve((*mem)->bo, true); alloc_gtt_mem()
229 r = radeon_bo_pin((*mem)->bo, RADEON_GEM_DOMAIN_GTT, alloc_gtt_mem()
230 &(*mem)->gpu_addr); alloc_gtt_mem()
235 *gpu_addr = (*mem)->gpu_addr; alloc_gtt_mem()
237 r = radeon_bo_kmap((*mem)->bo, &(*mem)->cpu_ptr); alloc_gtt_mem()
243 *cpu_ptr = (*mem)->cpu_ptr; alloc_gtt_mem()
245 radeon_bo_unreserve((*mem)->bo); alloc_gtt_mem()
250 radeon_bo_unpin((*mem)->bo); alloc_gtt_mem()
252 radeon_bo_unreserve((*mem)->bo); alloc_gtt_mem()
254 radeon_bo_unref(&(*mem)->bo); alloc_gtt_mem()
261 struct kgd_mem *mem = (struct kgd_mem *) mem_obj; free_gtt_mem() local
263 BUG_ON(mem == NULL); free_gtt_mem()
265 radeon_bo_reserve(mem->bo, true); free_gtt_mem()
266 radeon_bo_kunmap(mem->bo); free_gtt_mem()
267 radeon_bo_unpin(mem->bo); free_gtt_mem()
268 radeon_bo_unreserve(mem->bo); free_gtt_mem()
269 radeon_bo_unref(&(mem->bo)); free_gtt_mem()
270 kfree(mem); free_gtt_mem()
H A Dradeon_ttm.c197 switch (bo->mem.mem_type) { radeon_evict_flags()
202 bo->mem.start < (rbo->rdev->mc.visible_vram_size >> PAGE_SHIFT)) { radeon_evict_flags()
246 struct ttm_mem_reg *old_mem = &bo->mem; radeon_move_null()
315 struct ttm_mem_reg *old_mem = &bo->mem; radeon_move_vram_ram()
362 struct ttm_mem_reg *old_mem = &bo->mem; radeon_move_ram_vram()
402 struct ttm_mem_reg *old_mem = &bo->mem; radeon_bo_move()
449 static int radeon_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) radeon_ttm_io_mem_reserve() argument
451 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; radeon_ttm_io_mem_reserve()
454 mem->bus.addr = NULL; radeon_ttm_io_mem_reserve()
455 mem->bus.offset = 0; radeon_ttm_io_mem_reserve()
456 mem->bus.size = mem->num_pages << PAGE_SHIFT; radeon_ttm_io_mem_reserve()
457 mem->bus.base = 0; radeon_ttm_io_mem_reserve()
458 mem->bus.is_iomem = false; radeon_ttm_io_mem_reserve()
461 switch (mem->mem_type) { radeon_ttm_io_mem_reserve()
469 mem->bus.offset = mem->start << PAGE_SHIFT; radeon_ttm_io_mem_reserve()
470 mem->bus.base = rdev->mc.agp_base; radeon_ttm_io_mem_reserve()
471 mem->bus.is_iomem = !rdev->ddev->agp->cant_use_aperture; radeon_ttm_io_mem_reserve()
476 mem->bus.offset = mem->start << PAGE_SHIFT; radeon_ttm_io_mem_reserve()
478 if ((mem->bus.offset + mem->bus.size) > rdev->mc.visible_vram_size) radeon_ttm_io_mem_reserve()
480 mem->bus.base = rdev->mc.aper_base; radeon_ttm_io_mem_reserve()
481 mem->bus.is_iomem = true; radeon_ttm_io_mem_reserve()
487 if (mem->placement & TTM_PL_FLAG_WC) radeon_ttm_io_mem_reserve()
488 mem->bus.addr = radeon_ttm_io_mem_reserve()
489 ioremap_wc(mem->bus.base + mem->bus.offset, radeon_ttm_io_mem_reserve()
490 mem->bus.size); radeon_ttm_io_mem_reserve()
492 mem->bus.addr = radeon_ttm_io_mem_reserve()
493 ioremap_nocache(mem->bus.base + mem->bus.offset, radeon_ttm_io_mem_reserve()
494 mem->bus.size); radeon_ttm_io_mem_reserve()
502 mem->bus.base = (mem->bus.base & 0x0ffffffffUL) + radeon_ttm_io_mem_reserve()
512 static void radeon_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) radeon_ttm_io_mem_free() argument
/linux-4.1.27/drivers/infiniband/hw/mthca/
H A Dmthca_memfree.c59 struct scatterlist mem; member in struct:mthca_user_db_table::__anon4937
69 pci_unmap_sg(dev->pdev, chunk->mem, chunk->npages, mthca_free_icm_pages()
73 __free_pages(sg_page(&chunk->mem[i]), mthca_free_icm_pages()
74 get_order(chunk->mem[i].length)); mthca_free_icm_pages()
82 dma_free_coherent(&dev->pdev->dev, chunk->mem[i].length, mthca_free_icm_coherent()
83 lowmem_page_address(sg_page(&chunk->mem[i])), mthca_free_icm_coherent()
84 sg_dma_address(&chunk->mem[i])); mthca_free_icm_coherent()
107 static int mthca_alloc_icm_pages(struct scatterlist *mem, int order, gfp_t gfp_mask) mthca_alloc_icm_pages() argument
119 sg_set_page(mem, page, PAGE_SIZE << order, 0); mthca_alloc_icm_pages()
123 static int mthca_alloc_icm_coherent(struct device *dev, struct scatterlist *mem, mthca_alloc_icm_coherent() argument
126 void *buf = dma_alloc_coherent(dev, PAGE_SIZE << order, &sg_dma_address(mem), mthca_alloc_icm_coherent()
131 sg_set_buf(mem, buf, PAGE_SIZE << order); mthca_alloc_icm_coherent()
132 BUG_ON(mem->offset); mthca_alloc_icm_coherent()
133 sg_dma_len(mem) = PAGE_SIZE << order; mthca_alloc_icm_coherent()
164 sg_init_table(chunk->mem, MTHCA_ICM_CHUNK_LEN); mthca_alloc_icm()
175 &chunk->mem[chunk->npages], mthca_alloc_icm()
178 ret = mthca_alloc_icm_pages(&chunk->mem[chunk->npages], mthca_alloc_icm()
187 chunk->nsg = pci_map_sg(dev->pdev, chunk->mem, mthca_alloc_icm()
207 chunk->nsg = pci_map_sg(dev->pdev, chunk->mem, mthca_alloc_icm()
300 if (sg_dma_len(&chunk->mem[i]) > dma_offset) mthca_table_find()
301 *dma_handle = sg_dma_address(&chunk->mem[i]) + mthca_table_find()
303 dma_offset -= sg_dma_len(&chunk->mem[i]); mthca_table_find()
308 if (chunk->mem[i].length > offset) { mthca_table_find()
309 page = sg_page(&chunk->mem[i]); mthca_table_find()
312 offset -= chunk->mem[i].length; mthca_table_find()
480 sg_set_page(&db_tab->page[i].mem, pages[0], MTHCA_ICM_PAGE_SIZE, mthca_map_user_db()
483 ret = pci_map_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE); mthca_map_user_db()
489 ret = mthca_MAP_ICM_page(dev, sg_dma_address(&db_tab->page[i].mem), mthca_map_user_db()
492 pci_unmap_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE); mthca_map_user_db()
493 put_page(sg_page(&db_tab->page[i].mem)); mthca_map_user_db()
541 sg_init_table(&db_tab->page[i].mem, 1); mthca_init_user_db_tab()
558 pci_unmap_sg(dev->pdev, &db_tab->page[i].mem, 1, PCI_DMA_TODEVICE); mthca_cleanup_user_db_tab()
559 put_page(sg_page(&db_tab->page[i].mem)); mthca_cleanup_user_db_tab()
H A Dmthca_mr.c583 mr->attr.max_pages * sizeof *mr->mem.arbel.mtts > PAGE_SIZE) mthca_fmr_alloc()
601 mr->mem.arbel.mpt = mthca_table_find(dev->mr_table.mpt_table, key, NULL); mthca_fmr_alloc()
602 BUG_ON(!mr->mem.arbel.mpt); mthca_fmr_alloc()
604 mr->mem.tavor.mpt = dev->mr_table.tavor_fmr.mpt_base + mthca_fmr_alloc()
605 sizeof *(mr->mem.tavor.mpt) * idx; mthca_fmr_alloc()
616 mr->mem.arbel.mtts = mthca_table_find(dev->mr_table.mtt_table, mthca_fmr_alloc()
618 &mr->mem.arbel.dma_handle); mthca_fmr_alloc()
619 BUG_ON(!mr->mem.arbel.mtts); mthca_fmr_alloc()
621 mr->mem.tavor.mtts = dev->mr_table.tavor_fmr.mtt_base + mtt_seg; mthca_fmr_alloc()
736 writeb(MTHCA_MPT_STATUS_SW, fmr->mem.tavor.mpt); mthca_tavor_map_phys_fmr()
741 mthca_write64_raw(mtt_entry, fmr->mem.tavor.mtts + i); mthca_tavor_map_phys_fmr()
748 __raw_writel((__force u32) mpt_entry.lkey, &fmr->mem.tavor.mpt->key); mthca_tavor_map_phys_fmr()
749 memcpy_toio(&fmr->mem.tavor.mpt->start, &mpt_entry.start, mthca_tavor_map_phys_fmr()
753 writeb(MTHCA_MPT_STATUS_HW, fmr->mem.tavor.mpt); mthca_tavor_map_phys_fmr()
779 *(u8 *) fmr->mem.arbel.mpt = MTHCA_MPT_STATUS_SW; mthca_arbel_map_phys_fmr()
783 dma_sync_single_for_cpu(&dev->pdev->dev, fmr->mem.arbel.dma_handle, mthca_arbel_map_phys_fmr()
787 fmr->mem.arbel.mtts[i] = cpu_to_be64(page_list[i] | mthca_arbel_map_phys_fmr()
790 dma_sync_single_for_device(&dev->pdev->dev, fmr->mem.arbel.dma_handle, mthca_arbel_map_phys_fmr()
793 fmr->mem.arbel.mpt->key = cpu_to_be32(key); mthca_arbel_map_phys_fmr()
794 fmr->mem.arbel.mpt->lkey = cpu_to_be32(key); mthca_arbel_map_phys_fmr()
795 fmr->mem.arbel.mpt->length = cpu_to_be64(list_len * (1ull << fmr->attr.page_shift)); mthca_arbel_map_phys_fmr()
796 fmr->mem.arbel.mpt->start = cpu_to_be64(iova); mthca_arbel_map_phys_fmr()
800 *(u8 *) fmr->mem.arbel.mpt = MTHCA_MPT_STATUS_HW; mthca_arbel_map_phys_fmr()
814 writeb(MTHCA_MPT_STATUS_SW, fmr->mem.tavor.mpt); mthca_tavor_fmr_unmap()
824 *(u8 *) fmr->mem.arbel.mpt = MTHCA_MPT_STATUS_SW; mthca_arbel_fmr_unmap()
/linux-4.1.27/arch/unicore32/kernel/
H A Dksyms.c35 /* string / mem functions */
43 /* user mem (segment) */
/linux-4.1.27/arch/hexagon/include/asm/
H A Dfixmap.h25 * A lot of the fixmap info is already in mem-layout.h
27 #include <asm/mem-layout.h>
/linux-4.1.27/drivers/staging/octeon/
H A DMakefile16 octeon-ethernet-y += ethernet-mem.o
/linux-4.1.27/drivers/i2c/busses/
H A Di2c-puv3.c189 struct resource *mem; puv3_i2c_probe() local
192 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); puv3_i2c_probe()
193 if (!mem) puv3_i2c_probe()
196 if (!request_mem_region(mem->start, resource_size(mem), "puv3_i2c")) puv3_i2c_probe()
206 mem->start); puv3_i2c_probe()
227 release_mem_region(mem->start, resource_size(mem)); puv3_i2c_probe()
235 struct resource *mem; puv3_i2c_remove() local
241 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); puv3_i2c_remove()
242 release_mem_region(mem->start, resource_size(mem)); puv3_i2c_remove()
/linux-4.1.27/drivers/power/reset/
H A Dmsm-poweroff.c49 struct resource *mem; msm_restart_probe() local
51 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); msm_restart_probe()
52 msm_ps_hold = devm_ioremap_resource(dev, mem); msm_restart_probe()
/linux-4.1.27/arch/x86/kernel/
H A Dcrash.c67 struct crash_mem mem; member in struct:crash_elf_data
244 static int exclude_mem_range(struct crash_mem *mem, exclude_mem_range() argument
251 for (i = 0; i < mem->nr_ranges; i++) { exclude_mem_range()
252 start = mem->ranges[i].start; exclude_mem_range()
253 end = mem->ranges[i].end; exclude_mem_range()
266 mem->ranges[i].start = 0; exclude_mem_range()
267 mem->ranges[i].end = 0; exclude_mem_range()
268 if (i < mem->nr_ranges - 1) { exclude_mem_range()
270 for (j = i; j < mem->nr_ranges - 1; j++) { exclude_mem_range()
271 mem->ranges[j].start = exclude_mem_range()
272 mem->ranges[j+1].start; exclude_mem_range()
273 mem->ranges[j].end = exclude_mem_range()
274 mem->ranges[j+1].end; exclude_mem_range()
277 mem->nr_ranges--; exclude_mem_range()
283 mem->ranges[i].end = mstart - 1; exclude_mem_range()
287 mem->ranges[i].end = mstart - 1; exclude_mem_range()
289 mem->ranges[i].start = mend + 1; exclude_mem_range()
305 if (j < mem->nr_ranges) { exclude_mem_range()
307 for (i = mem->nr_ranges - 1; i >= j; i--) exclude_mem_range()
308 mem->ranges[i + 1] = mem->ranges[i]; exclude_mem_range()
311 mem->ranges[j].start = temp_range.start; exclude_mem_range()
312 mem->ranges[j].end = temp_range.end; exclude_mem_range()
313 mem->nr_ranges++; exclude_mem_range()
319 * might lead to split and split ranges are put in ced->mem.ranges[] array
324 struct crash_mem *cmem = &ced->mem; elf_header_exclude_ranges()
366 /* Exclude unwanted mem ranges */ prepare_elf64_ram_headers_callback()
371 /* Go through all the ranges in ced->mem.ranges[] and prepare phdr */ prepare_elf64_ram_headers_callback()
372 cmem = &ced->mem; prepare_elf64_ram_headers_callback()
/linux-4.1.27/drivers/spi/
H A Dspi-dw-mmio.c36 struct resource *mem; dw_spi_mmio_probe() local
48 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); dw_spi_mmio_probe()
49 if (!mem) { dw_spi_mmio_probe()
50 dev_err(&pdev->dev, "no mem resource?\n"); dw_spi_mmio_probe()
54 dws->regs = devm_ioremap_resource(&pdev->dev, mem); dw_spi_mmio_probe()
H A Dspi-xtensa-xtfpga.c86 struct resource *mem; xtfpga_spi_probe() local
104 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); xtfpga_spi_probe()
105 if (!mem) { xtfpga_spi_probe()
110 xspi->regs = devm_ioremap_resource(&pdev->dev, mem); xtfpga_spi_probe()
/linux-4.1.27/drivers/pcmcia/
H A Dvrc4171_card.c422 static int pccard_set_mem_map(struct pcmcia_socket *sock, struct pccard_mem_map *mem) pccard_set_mem_map() argument
430 mem == NULL || mem->map >= MEM_MAX_MAPS || pccard_set_mem_map()
431 mem->res->start < CARD_MEM_START || mem->res->start > CARD_MEM_END || pccard_set_mem_map()
432 mem->res->end < CARD_MEM_START || mem->res->end > CARD_MEM_END || pccard_set_mem_map()
433 mem->res->start > mem->res->end || pccard_set_mem_map()
434 mem->card_start > CARD_MAX_MEM_OFFSET || pccard_set_mem_map()
435 mem->speed > CARD_MAX_MEM_SPEED) pccard_set_mem_map()
439 map = mem->map; pccard_set_mem_map()
447 start = (mem->res->start >> 12) & 0x3fff; pccard_set_mem_map()
448 if (mem->flags & MAP_16BIT) pccard_set_mem_map()
452 stop = (mem->res->end >> 12) & 0x3fff; pccard_set_mem_map()
453 switch (mem->speed) { pccard_set_mem_map()
468 offset = (mem->card_start >> 12) & 0x3fff; pccard_set_mem_map()
469 if (mem->flags & MAP_ATTRIB) pccard_set_mem_map()
471 if (mem->flags & MAP_WRPROT) pccard_set_mem_map()
475 if (mem->flags & MAP_ACTIVE) { pccard_set_mem_map()
H A Dtcic.c733 static int tcic_set_mem_map(struct pcmcia_socket *sock, struct pccard_mem_map *mem) tcic_set_mem_map() argument
740 "%#llx-%#llx, %#x)\n", psock, mem->map, mem->flags, tcic_set_mem_map()
741 mem->speed, (unsigned long long)mem->res->start, tcic_set_mem_map()
742 (unsigned long long)mem->res->end, mem->card_start); tcic_set_mem_map()
743 if ((mem->map > 3) || (mem->card_start > 0x3ffffff) || tcic_set_mem_map()
744 (mem->res->start > 0xffffff) || (mem->res->end > 0xffffff) || tcic_set_mem_map()
745 (mem->res->start > mem->res->end) || (mem->speed > 1000)) tcic_set_mem_map()
748 addr = TCIC_MWIN(psock, mem->map); tcic_set_mem_map()
750 base = mem->res->start; len = mem->res->end - mem->res->start; tcic_set_mem_map()
759 mmap = mem->card_start - mem->res->start; tcic_set_mem_map()
761 if (mem->flags & MAP_ATTRIB) mmap |= TCIC_MMAP_REG; tcic_set_mem_map()
766 ctl |= to_cycles(mem->speed) & TCIC_MCTL_WSCNT_MASK; tcic_set_mem_map()
767 ctl |= (mem->flags & MAP_16BIT) ? 0 : TCIC_MCTL_B8; tcic_set_mem_map()
768 ctl |= (mem->flags & MAP_WRPROT) ? TCIC_MCTL_WP : 0; tcic_set_mem_map()
769 ctl |= (mem->flags & MAP_ACTIVE) ? TCIC_MCTL_ENA : 0; tcic_set_mem_map()
783 pccard_mem_map mem = { .res = &res, }; tcic_init() local
790 mem.map = i; tcic_init()
791 tcic_set_mem_map(s, &mem); tcic_init()
H A Di82092.c400 pccard_mem_map mem = { .res = &res, }; i82092aa_init() local
409 mem.map = i; i82092aa_init()
410 i82092aa_set_mem_map(sock, &mem); i82092aa_init()
601 static int i82092aa_set_mem_map(struct pcmcia_socket *socket, struct pccard_mem_map *mem) i82092aa_set_mem_map() argument
611 pcibios_resource_to_bus(sock_info->dev->bus, &region, mem->res); i82092aa_set_mem_map()
613 map = mem->map; i82092aa_set_mem_map()
620 if ( (mem->card_start > 0x3ffffff) || (region.start > region.end) || i82092aa_set_mem_map()
621 (mem->speed > 1000) ) { i82092aa_set_mem_map()
623 printk("invalid mem map for socket %i: %llx to %llx with a " i82092aa_set_mem_map()
628 mem->card_start); i82092aa_set_mem_map()
637 /* printk("set_mem_map: Setting map %i range to %x - %x on socket %i, speed is %i, active = %i \n",map, region.start,region.end,sock,mem->speed,mem->flags & MAP_ACTIVE); */ i82092aa_set_mem_map()
642 if (mem->flags & MAP_16BIT) i82092aa_set_mem_map()
644 if (mem->flags & MAP_0WS) i82092aa_set_mem_map()
651 switch (to_cycles(mem->speed)) { i82092aa_set_mem_map()
669 i = ((mem->card_start - region.start) >> 12) & 0x3fff; i82092aa_set_mem_map()
670 if (mem->flags & MAP_WRPROT) i82092aa_set_mem_map()
672 if (mem->flags & MAP_ATTRIB) { i82092aa_set_mem_map()
681 if (mem->flags & MAP_ACTIVE) i82092aa_set_mem_map()
H A Dvrc4173_cardu.c305 static int cardu_get_mem_map(unsigned int sock, struct pccard_mem_map *mem) cardu_get_mem_map() argument
312 map = mem->map; cardu_get_mem_map()
317 mem->flags = (window & MEM_WIN_EN(map)) ? MAP_ACTIVE : 0; cardu_get_mem_map()
320 mem->flags |= (start & MEM_WIN_DSIZE) ? MAP_16BIT : 0; cardu_get_mem_map()
327 mem->flags |= (offset & MEM_WIN_WP) ? MAP_WRPROT : 0; cardu_get_mem_map()
328 mem->flags |= (offset & MEM_WIN_REGSET) ? MAP_ATTRIB : 0; cardu_get_mem_map()
330 mem->card_start = offset & 0x03ffffff; cardu_get_mem_map()
333 mem->sys_start = start + page; cardu_get_mem_map()
334 mem->sys_stop = start + page; cardu_get_mem_map()
339 static int cardu_set_mem_map(unsigned int sock, struct pccard_mem_map *mem) cardu_set_mem_map() argument
347 map = mem->map; cardu_set_mem_map()
348 sys_start = mem->sys_start; cardu_set_mem_map()
349 sys_stop = mem->sys_stop; cardu_set_mem_map()
350 card_start = mem->card_start; cardu_set_mem_map()
366 if (mem->flags & MAP_16BIT) value |= MEM_WIN_DSIZE; cardu_set_mem_map()
373 if (mem->flags & MAP_WRPROT) value |= MEM_WIN_WP; cardu_set_mem_map()
374 if (mem->flags & MAP_ATTRIB) value |= MEM_WIN_REGSET; cardu_set_mem_map()
377 if (mem->flags & MAP_ACTIVE) cardu_set_mem_map()
H A Dm32r_pcc.c490 static int _pcc_set_mem_map(u_short sock, struct pccard_mem_map *mem) argument
493 u_char map = mem->map;
504 "%#llx, %#x)\n", sock, map, mem->flags,
505 mem->speed, (unsigned long long)mem->static_start,
506 mem->card_start);
511 if ((map > MAX_WIN) || (mem->card_start > 0x3ffffff)){
518 if ((mem->flags & MAP_ACTIVE) == 0) {
531 if (mem->flags & MAP_ATTRIB) {
543 addr = t->mapaddr + (mem->card_start & M32R_PCC_MAPMASK);
546 mem->static_start = addr + mem->card_start;
639 static int pcc_set_mem_map(struct pcmcia_socket *s, struct pccard_mem_map *mem) pcc_set_mem_map() argument
645 LOCKED(_pcc_set_mem_map(sock, mem)); pcc_set_mem_map()
H A Dpd6729.c467 struct pccard_mem_map *mem) pd6729_set_mem_map()
474 map = mem->map; pd6729_set_mem_map()
480 if ((mem->res->start > mem->res->end) || (mem->speed > 1000)) { pd6729_set_mem_map()
491 i = (mem->res->start >> 12) & 0x0fff; pd6729_set_mem_map()
492 if (mem->flags & MAP_16BIT) pd6729_set_mem_map()
494 if (mem->flags & MAP_0WS) pd6729_set_mem_map()
500 i = (mem->res->end >> 12) & 0x0fff; pd6729_set_mem_map()
501 switch (to_cycles(mem->speed)) { pd6729_set_mem_map()
519 indirect_write(socket, PD67_EXT_DATA, mem->res->start >> 24); pd6729_set_mem_map()
523 i = ((mem->card_start - mem->res->start) >> 12) & 0x3fff; pd6729_set_mem_map()
524 if (mem->flags & MAP_WRPROT) pd6729_set_mem_map()
526 if (mem->flags & MAP_ATTRIB) { pd6729_set_mem_map()
537 if (mem->flags & MAP_ACTIVE) pd6729_set_mem_map()
548 pccard_mem_map mem = { .res = &res, }; pd6729_init() local
556 mem.map = i; pd6729_init()
557 pd6729_set_mem_map(sock, &mem); pd6729_init()
466 pd6729_set_mem_map(struct pcmcia_socket *sock, struct pccard_mem_map *mem) pd6729_set_mem_map() argument
H A Di82365.c1061 static int i365_set_mem_map(u_short sock, struct pccard_mem_map *mem) i365_set_mem_map() argument
1067 "%#x)\n", sock, mem->map, mem->flags, mem->speed, i365_set_mem_map()
1068 (unsigned long long)mem->res->start, i365_set_mem_map()
1069 (unsigned long long)mem->res->end, mem->card_start); i365_set_mem_map()
1071 map = mem->map; i365_set_mem_map()
1072 if ((map > 4) || (mem->card_start > 0x3ffffff) || i365_set_mem_map()
1073 (mem->res->start > mem->res->end) || (mem->speed > 1000)) i365_set_mem_map()
1075 if ((mem->res->start > 0xffffff) || (mem->res->end > 0xffffff)) i365_set_mem_map()
1083 i = (mem->res->start >> 12) & 0x0fff; i365_set_mem_map()
1084 if (mem->flags & MAP_16BIT) i |= I365_MEM_16BIT; i365_set_mem_map()
1085 if (mem->flags & MAP_0WS) i |= I365_MEM_0WS; i365_set_mem_map()
1088 i = (mem->res->end >> 12) & 0x0fff; i365_set_mem_map()
1089 switch (to_cycles(mem->speed)) { i365_set_mem_map()
1097 i = ((mem->card_start - mem->res->start) >> 12) & 0x3fff; i365_set_mem_map()
1098 if (mem->flags & MAP_WRPROT) i |= I365_MEM_WRPROT; i365_set_mem_map()
1099 if (mem->flags & MAP_ATTRIB) i |= I365_MEM_REG; i365_set_mem_map()
1103 if (mem->flags & MAP_ACTIVE) i365_set_mem_map()
1195 static int pcic_set_mem_map(struct pcmcia_socket *s, struct pccard_mem_map *mem) pcic_set_mem_map() argument
1201 LOCKED(i365_set_mem_map(sock, mem)); pcic_set_mem_map()
1209 pccard_mem_map mem = { .res = &res, }; pcic_init() local
1216 mem.map = i; pcic_init()
1217 pcic_set_mem_map(s, &mem); pcic_init()
H A Delectra_cf.c191 struct resource mem, io; electra_cf_probe() local
197 err = of_address_to_resource(np, 0, &mem); electra_cf_probe()
213 cf->mem_phys = mem.start; electra_cf_probe()
214 cf->mem_size = PAGE_ALIGN(resource_size(&mem)); electra_cf_probe()
239 cf->iomem.end = (unsigned long)cf->mem_base + (mem.end - mem.start); electra_cf_probe()
302 dev_info(device, "at mem 0x%lx io 0x%llx irq %d\n", electra_cf_probe()
H A Dcistpl.c90 pccard_mem_map *mem = &s->cis_mem; set_cis_map() local
93 if (!(s->features & SS_CAP_STATIC_MAP) && (mem->res == NULL)) { set_cis_map()
94 mem->res = pcmcia_find_mem_region(0, s->map_size, set_cis_map()
96 if (mem->res == NULL) { set_cis_map()
105 s->cis_virt = ioremap(mem->res->start, s->map_size); set_cis_map()
107 mem->card_start = card_offset; set_cis_map()
108 mem->flags = flags; set_cis_map()
110 ret = s->ops->set_mem_map(s, mem); set_cis_map()
120 s->cis_virt = ioremap(mem->static_start, s->map_size); set_cis_map()
173 "attempt to read CIS mem at addr %#x", addr); pcmcia_read_cis_mem()
1025 static u_char *parse_mem(u_char *p, u_char *q, cistpl_mem_t *mem) parse_mem() argument
1033 mem->nwin = (*p & 0x07) + 1; parse_mem()
1040 for (i = 0; i < mem->nwin; i++) { parse_mem()
1058 mem->win[i].len = len << 8; parse_mem()
1059 mem->win[i].card_addr = ca << 8; parse_mem()
1060 mem->win[i].host_addr = ha << 8; parse_mem()
1161 entry->mem.nwin = 0; parse_cftable_entry()
1164 entry->mem.nwin = 1; parse_cftable_entry()
1165 entry->mem.win[0].len = get_unaligned_le16(p) << 8; parse_cftable_entry()
1166 entry->mem.win[0].card_addr = 0; parse_cftable_entry()
1167 entry->mem.win[0].host_addr = 0; parse_cftable_entry()
1173 entry->mem.nwin = 1; parse_cftable_entry()
1174 entry->mem.win[0].len = get_unaligned_le16(p) << 8; parse_cftable_entry()
1175 entry->mem.win[0].card_addr = get_unaligned_le16(p + 2) << 8; parse_cftable_entry()
1176 entry->mem.win[0].host_addr = 0; parse_cftable_entry()
1182 p = parse_mem(p, q, &entry->mem); parse_cftable_entry()
H A Di82092aa.h32 static int i82092aa_set_mem_map(struct pcmcia_socket *socket, struct pccard_mem_map *mem);
/linux-4.1.27/tools/virtio/
H A Dvirtio_test.c42 struct vhost_memory *mem; member in struct:vdev_info
125 dev->mem = malloc(offsetof(struct vhost_memory, regions) + vdev_info_init()
126 sizeof dev->mem->regions[0]); vdev_info_init()
127 assert(dev->mem); vdev_info_init()
128 memset(dev->mem, 0, offsetof(struct vhost_memory, regions) + vdev_info_init()
129 sizeof dev->mem->regions[0]); vdev_info_init()
130 dev->mem->nregions = 1; vdev_info_init()
131 dev->mem->regions[0].guest_phys_addr = (long)dev->buf; vdev_info_init()
132 dev->mem->regions[0].userspace_addr = (long)dev->buf; vdev_info_init()
133 dev->mem->regions[0].memory_size = dev->buf_size; vdev_info_init()
134 r = ioctl(dev->control, VHOST_SET_MEM_TABLE, dev->mem); vdev_info_init()
/linux-4.1.27/drivers/pci/hotplug/
H A Dibmphp_pci.c335 } else if (cur_func->mem[i]) { ibmphp_configure_card()
336 ibmphp_remove_resource (cur_func->mem[i]); ibmphp_configure_card()
337 cur_func->mem[i] = NULL; ibmphp_configure_card()
364 struct resource_node *mem[6]; configure_device() local
469 debug ("there's no pfmem... going into mem.\n"); configure_device()
496 debug ("inside the mem 64 case, count %d\n", count); configure_device()
510 mem[count] = kzalloc(sizeof(struct resource_node), GFP_KERNEL); configure_device()
511 if (!mem[count]) { configure_device()
515 mem[count]->type = MEM; configure_device()
516 mem[count]->busno = func->busno; configure_device()
517 mem[count]->devfunc = PCI_DEVFN(func->device, configure_device()
519 mem[count]->len = len[count]; configure_device()
520 if (ibmphp_check_resource (mem[count], 0) == 0) { configure_device()
521 ibmphp_add_resource (mem[count]); configure_device()
522 func->mem[count] = mem[count]; configure_device()
524 err ("cannot allocate requested mem for bus %x, device %x, len %x\n", configure_device()
526 kfree (mem[count]); configure_device()
529 pci_bus_write_config_dword (ibmphp_pci_bus, devfn, address[count], func->mem[count]->start); configure_device()
531 debug ("b4 writing, start address is %x\n", func->mem[count]->start); configure_device()
538 debug ("inside mem 64 case, reg. mem, count %d\n", count); configure_device()
544 } /* end of mem */ configure_device()
584 struct resource_node *mem = NULL; configure_bridge() local
653 /* First we need to allocate mem/io for the bridge itself in case it needs it */ configure_bridge()
777 func->mem[count] = bus_mem[count]; configure_bridge()
779 err ("cannot allocate requested mem for bus %x, device %x, len %x\n", configure_bridge()
785 pci_bus_write_config_dword (ibmphp_pci_bus, devfn, address[count], func->mem[count]->start); configure_bridge()
795 } /* end of mem */ configure_bridge()
807 debug ("amount_needed->mem = %x\n", amount_needed->mem); configure_bridge()
822 func->mem[count] = NULL; configure_bridge()
852 if (!amount_needed->mem) { configure_bridge()
856 debug ("it wants %x memory behind the bridge\n", amount_needed->mem); configure_bridge()
857 mem = kzalloc(sizeof(*mem), GFP_KERNEL); configure_bridge()
858 if (!mem) { configure_bridge()
863 mem->type = MEM; configure_bridge()
864 mem->busno = func->busno; configure_bridge()
865 mem->devfunc = PCI_DEVFN(func->device, func->function); configure_bridge()
866 mem->len = amount_needed->mem; configure_bridge()
867 if (ibmphp_check_resource (mem, 1) == 0) { configure_bridge()
868 ibmphp_add_resource (mem); configure_bridge()
870 debug ("were we able to add mem\n"); configure_bridge()
875 debug ("it doesn't want n.e.pfmem mem?\n"); configure_bridge()
933 rc = add_new_bus (bus, io, mem, pfmem, func->busno); configure_bridge()
935 rc = add_new_bus (bus, io, mem, pfmem, 0xFF); configure_bridge()
1055 if (mem) configure_bridge()
1056 ibmphp_remove_resource (mem); configure_bridge()
1066 func->mem[i] = NULL; configure_bridge()
1182 amount->mem += len[count]; scan_behind_bridge()
1200 if ((amount->mem) && (amount->mem < MEMBRIDGE)) scan_behind_bridge()
1201 amount->mem = MEMBRIDGE; scan_behind_bridge()
1228 struct resource_node *mem; unconfigure_boot_device() local
1309 debug ("start address of mem is %x\n", start_address); unconfigure_boot_device()
1312 if (ibmphp_find_resource (bus, start_address, &mem, MEM) < 0) { unconfigure_boot_device()
1316 if (mem) { unconfigure_boot_device()
1317 debug ("mem->start = %x\n", mem->start); unconfigure_boot_device()
1319 ibmphp_remove_resource(mem); unconfigure_boot_device()
1326 } /* end of mem */ unconfigure_boot_device()
1339 struct resource_node *mem = NULL; unconfigure_boot_bridge() local
1424 if (ibmphp_find_resource (bus, start_address, &mem, MEM) < 0) { unconfigure_boot_bridge()
1428 if (mem) { unconfigure_boot_bridge()
1429 debug ("mem->start = %x\n", mem->start); unconfigure_boot_bridge()
1431 ibmphp_remove_resource(mem); unconfigure_boot_bridge()
1438 } /* end of mem */ unconfigure_boot_bridge()
1599 if (cur_func->mem[i]) { ibmphp_unconfigure_card()
1600 debug ("mem[%d] exists\n", i); ibmphp_unconfigure_card()
1602 ibmphp_remove_resource (cur_func->mem[i]); ibmphp_unconfigure_card()
1603 cur_func->mem[i] = NULL; ibmphp_unconfigure_card()
1633 static int add_new_bus (struct bus_node *bus, struct resource_node *io, struct resource_node *mem, struct resource_node *pfmem, u8 parent_busno) add_new_bus() argument
1662 if (mem) { add_new_bus()
1668 mem_range->start = mem->start; add_new_bus()
1669 mem_range->end = mem->end; add_new_bus()
/linux-4.1.27/drivers/net/wireless/ath/ath9k/
H A Dahb.c85 void __iomem *mem; ath_ahb_probe() local
106 mem = devm_ioremap_nocache(&pdev->dev, res->start, resource_size(res)); ath_ahb_probe()
107 if (mem == NULL) { ath_ahb_probe()
133 sc->mem = mem; ath_ahb_probe()
150 wiphy_info(hw->wiphy, "%s mem=0x%lx, irq=%d\n", ath_ahb_probe()
151 hw_name, (unsigned long)mem, irq); ath_ahb_probe()
/linux-4.1.27/arch/arc/mm/
H A Dinit.c27 /* Default tot mem from .config */
30 /* User can over-ride above with "mem=nnn[KkMm]" in cmdline */ setup_mem_sz()
36 pr_info("\"mem=%s\": mem sz set to %ldM\n", str, TO_MB(arc_mem_sz)); setup_mem_sz()
40 early_param("mem", setup_mem_sz);
84 * even if "mem=xxx" cmline over-ride is given and/or setup_arch_memory()
95 /* Last usable page of low mem (no HIGHMEM yet for ARC port) */ setup_arch_memory()
/linux-4.1.27/drivers/gpu/drm/ast/
H A Dast_ttm.c157 struct ttm_mem_reg *mem) ast_ttm_io_mem_reserve()
159 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; ast_ttm_io_mem_reserve()
162 mem->bus.addr = NULL; ast_ttm_io_mem_reserve()
163 mem->bus.offset = 0; ast_ttm_io_mem_reserve()
164 mem->bus.size = mem->num_pages << PAGE_SHIFT; ast_ttm_io_mem_reserve()
165 mem->bus.base = 0; ast_ttm_io_mem_reserve()
166 mem->bus.is_iomem = false; ast_ttm_io_mem_reserve()
169 switch (mem->mem_type) { ast_ttm_io_mem_reserve()
174 mem->bus.offset = mem->start << PAGE_SHIFT; ast_ttm_io_mem_reserve()
175 mem->bus.base = pci_resource_start(ast->dev->pdev, 0); ast_ttm_io_mem_reserve()
176 mem->bus.is_iomem = true; ast_ttm_io_mem_reserve()
185 static void ast_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ast_ttm_io_mem_free() argument
156 ast_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) ast_ttm_io_mem_reserve() argument
/linux-4.1.27/drivers/gpu/drm/bochs/
H A Dbochs_mm.c135 struct ttm_mem_reg *mem) bochs_ttm_io_mem_reserve()
137 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; bochs_ttm_io_mem_reserve()
140 mem->bus.addr = NULL; bochs_ttm_io_mem_reserve()
141 mem->bus.offset = 0; bochs_ttm_io_mem_reserve()
142 mem->bus.size = mem->num_pages << PAGE_SHIFT; bochs_ttm_io_mem_reserve()
143 mem->bus.base = 0; bochs_ttm_io_mem_reserve()
144 mem->bus.is_iomem = false; bochs_ttm_io_mem_reserve()
147 switch (mem->mem_type) { bochs_ttm_io_mem_reserve()
152 mem->bus.offset = mem->start << PAGE_SHIFT; bochs_ttm_io_mem_reserve()
153 mem->bus.base = bochs->fb_base; bochs_ttm_io_mem_reserve()
154 mem->bus.is_iomem = true; bochs_ttm_io_mem_reserve()
164 struct ttm_mem_reg *mem) bochs_ttm_io_mem_free()
134 bochs_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) bochs_ttm_io_mem_reserve() argument
163 bochs_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) bochs_ttm_io_mem_free() argument
H A Dbochs_hw.c54 unsigned long addr, size, mem, ioaddr, iosize, qext_size; bochs_hw_init() local
81 mem = bochs_dispi_read(bochs, VBE_DISPI_INDEX_VIDEO_MEMORY_64K) bochs_hw_init()
94 if (size != mem) { bochs_hw_init()
96 size, mem); bochs_hw_init()
97 size = min(size, mem); bochs_hw_init()
/linux-4.1.27/drivers/gpu/drm/cirrus/
H A Dcirrus_ttm.c157 struct ttm_mem_reg *mem) cirrus_ttm_io_mem_reserve()
159 struct ttm_mem_type_manager *man = &bdev->man[mem->mem_type]; cirrus_ttm_io_mem_reserve()
162 mem->bus.addr = NULL; cirrus_ttm_io_mem_reserve()
163 mem->bus.offset = 0; cirrus_ttm_io_mem_reserve()
164 mem->bus.size = mem->num_pages << PAGE_SHIFT; cirrus_ttm_io_mem_reserve()
165 mem->bus.base = 0; cirrus_ttm_io_mem_reserve()
166 mem->bus.is_iomem = false; cirrus_ttm_io_mem_reserve()
169 switch (mem->mem_type) { cirrus_ttm_io_mem_reserve()
174 mem->bus.offset = mem->start << PAGE_SHIFT; cirrus_ttm_io_mem_reserve()
175 mem->bus.base = pci_resource_start(cirrus->dev->pdev, 0); cirrus_ttm_io_mem_reserve()
176 mem->bus.is_iomem = true; cirrus_ttm_io_mem_reserve()
185 static void cirrus_ttm_io_mem_free(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) cirrus_ttm_io_mem_free() argument
156 cirrus_ttm_io_mem_reserve(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem) cirrus_ttm_io_mem_reserve() argument
/linux-4.1.27/drivers/usb/gadget/
H A Dconfig.c130 void *mem; usb_copy_descriptors() local
138 mem = kmalloc(bytes, GFP_KERNEL); usb_copy_descriptors()
139 if (!mem) usb_copy_descriptors()
143 * to descriptors copied starting at "mem"; usb_copy_descriptors()
146 tmp = mem; usb_copy_descriptors()
147 ret = mem; usb_copy_descriptors()
148 mem += (n_desc + 1) * sizeof(*tmp); usb_copy_descriptors()
150 memcpy(mem, *src, (*src)->bLength); usb_copy_descriptors()
151 *tmp = mem; usb_copy_descriptors()
153 mem += (*src)->bLength; usb_copy_descriptors()
/linux-4.1.27/drivers/usb/gadget/function/
H A Df_uvc.c453 #define UVC_COPY_DESCRIPTOR(mem, dst, desc) \
455 memcpy(mem, desc, (desc)->bLength); \
456 *(dst)++ = mem; \
457 mem += (desc)->bLength; \
460 #define UVC_COPY_DESCRIPTORS(mem, dst, src) \
464 memcpy(mem, *__src, (*__src)->bLength); \
465 *dst++ = mem; \
466 mem += (*__src)->bLength; \
485 void *mem; uvc_copy_descriptors() local
555 mem = kmalloc((n_desc + 1) * sizeof(*src) + bytes, GFP_KERNEL); uvc_copy_descriptors()
556 if (mem == NULL) uvc_copy_descriptors()
559 hdr = mem; uvc_copy_descriptors()
560 dst = mem; uvc_copy_descriptors()
561 mem += (n_desc + 1) * sizeof(*src); uvc_copy_descriptors()
564 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_iad); uvc_copy_descriptors()
565 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_control_intf); uvc_copy_descriptors()
567 uvc_control_header = mem; uvc_copy_descriptors()
568 UVC_COPY_DESCRIPTORS(mem, dst, uvc_copy_descriptors()
574 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_control_ep); uvc_copy_descriptors()
576 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_ss_control_comp); uvc_copy_descriptors()
578 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_control_cs_ep); uvc_copy_descriptors()
579 UVC_COPY_DESCRIPTOR(mem, dst, &uvc_streaming_intf_alt0); uvc_copy_descriptors()
581 uvc_streaming_header = mem; uvc_copy_descriptors()
582 UVC_COPY_DESCRIPTORS(mem, dst, uvc_copy_descriptors()
587 UVC_COPY_DESCRIPTORS(mem, dst, uvc_streaming_std); uvc_copy_descriptors()
H A Duvc_video.c49 void *mem; uvc_video_encode_data() local
52 mem = buf->mem + queue->buf_used; uvc_video_encode_data()
55 memcpy(data, mem, nbytes); uvc_video_encode_data()
65 void *mem = req->buf; uvc_video_encode_bulk() local
71 ret = uvc_video_encode_header(video, buf, mem, len); uvc_video_encode_bulk()
73 mem += ret; uvc_video_encode_bulk()
79 ret = uvc_video_encode_data(video, buf, mem, len); uvc_video_encode_bulk()
105 void *mem = req->buf; uvc_video_encode_isoc() local
110 ret = uvc_video_encode_header(video, buf, mem, len); uvc_video_encode_isoc()
111 mem += ret; uvc_video_encode_isoc()
115 ret = uvc_video_encode_data(video, buf, mem, len); uvc_video_encode_isoc()
/linux-4.1.27/drivers/gpu/drm/amd/amdkfd/
H A Dkfd_topology.c174 static int kfd_parse_subtype_mem(struct crat_subtype_memory *mem) kfd_parse_subtype_mem() argument
180 BUG_ON(!mem); kfd_parse_subtype_mem()
183 mem->promixity_domain); kfd_parse_subtype_mem()
185 if (mem->promixity_domain == i) { kfd_parse_subtype_mem()
195 if (mem->flags & CRAT_MEM_FLAGS_HOT_PLUGGABLE) kfd_parse_subtype_mem()
197 if (mem->flags & CRAT_MEM_FLAGS_NON_VOLATILE) kfd_parse_subtype_mem()
201 ((uint64_t)mem->length_high << 32) + kfd_parse_subtype_mem()
202 mem->length_low; kfd_parse_subtype_mem()
203 props->width = mem->width; kfd_parse_subtype_mem()
321 struct crat_subtype_memory *mem; kfd_parse_subtype() local
334 mem = (struct crat_subtype_memory *)sub_type_hdr; kfd_parse_subtype()
335 ret = kfd_parse_subtype_mem(mem); kfd_parse_subtype()
367 struct kfd_mem_properties *mem; kfd_release_topology_device() local
376 mem = container_of(dev->mem_props.next, kfd_release_topology_device()
378 list_del(&mem->list); kfd_release_topology_device()
379 kfree(mem); kfd_release_topology_device()
569 struct kfd_mem_properties *mem; mem_show() local
574 mem = container_of(attr, struct kfd_mem_properties, attr); mem_show()
575 sysfs_show_32bit_prop(buffer, "heap_type", mem->heap_type); mem_show()
576 sysfs_show_64bit_prop(buffer, "size_in_bytes", mem->size_in_bytes); mem_show()
577 sysfs_show_32bit_prop(buffer, "flags", mem->flags); mem_show()
578 sysfs_show_32bit_prop(buffer, "width", mem->width); mem_show()
579 ret = sysfs_show_32bit_prop(buffer, "mem_clk_max", mem->mem_clk_max); mem_show()
764 struct kfd_mem_properties *mem; kfd_remove_sysfs_node_entry() local
793 list_for_each_entry(mem, &dev->mem_props, list) kfd_remove_sysfs_node_entry()
794 if (mem->kobj) { kfd_remove_sysfs_node_entry()
795 kfd_remove_sysfs_file(mem->kobj, &mem->attr); kfd_remove_sysfs_node_entry()
796 mem->kobj = NULL; kfd_remove_sysfs_node_entry()
818 struct kfd_mem_properties *mem; kfd_build_sysfs_node_entry() local
872 list_for_each_entry(mem, &dev->mem_props, list) { kfd_build_sysfs_node_entry()
873 mem->kobj = kzalloc(sizeof(struct kobject), GFP_KERNEL); kfd_build_sysfs_node_entry()
874 if (!mem->kobj) kfd_build_sysfs_node_entry()
876 ret = kobject_init_and_add(mem->kobj, &mem_type, kfd_build_sysfs_node_entry()
881 mem->attr.name = "properties"; kfd_build_sysfs_node_entry()
882 mem->attr.mode = KFD_SYSFS_FILE_MODE; kfd_build_sysfs_node_entry()
883 sysfs_attr_init(&mem->attr); kfd_build_sysfs_node_entry()
884 ret = sysfs_create_file(mem->kobj, &mem->attr); kfd_build_sysfs_node_entry()
/linux-4.1.27/drivers/mfd/
H A Dvexpress-sysreg.c165 struct resource *mem; vexpress_sysreg_probe() local
171 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); vexpress_sysreg_probe()
172 if (!mem) vexpress_sysreg_probe()
175 base = devm_ioremap(&pdev->dev, mem->start, resource_size(mem)); vexpress_sysreg_probe()
209 ARRAY_SIZE(vexpress_sysreg_cells), mem, 0, NULL); vexpress_sysreg_probe()
H A Djz4740-adc.c56 struct resource *mem; member in struct:jz4740_adc
240 adc->mem = request_mem_region(mem_base->start, JZ_REG_ADC_STATUS, jz4740_adc_probe()
242 if (!adc->mem) { jz4740_adc_probe()
247 adc->base = ioremap_nocache(adc->mem->start, resource_size(adc->mem)); jz4740_adc_probe()
299 release_mem_region(adc->mem->start, resource_size(adc->mem)); jz4740_adc_probe()
315 release_mem_region(adc->mem->start, resource_size(adc->mem)); jz4740_adc_remove()
/linux-4.1.27/arch/um/kernel/skas/
H A Duaccess.c218 int __clear_user(void __user *mem, int len) __clear_user() argument
220 return buffer_op((unsigned long) mem, len, 1, clear_chunk, NULL); __clear_user()
223 int clear_user(void __user *mem, int len) clear_user() argument
226 memset((__force void*)mem, 0, len); clear_user()
230 return access_ok(VERIFY_WRITE, mem, len) ? clear_user()
231 buffer_op((unsigned long) mem, len, 1, clear_chunk, NULL) : len; clear_user()
/linux-4.1.27/arch/powerpc/sysdev/
H A Dppc4xx_hsta_msi.c129 struct resource *mem; hsta_msi_probe() local
132 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); hsta_msi_probe()
133 if (IS_ERR(mem)) { hsta_msi_probe()
145 ppc4xx_hsta_msi.address = mem->start; hsta_msi_probe()
146 ppc4xx_hsta_msi.data = ioremap(mem->start, resource_size(mem)); hsta_msi_probe()
/linux-4.1.27/arch/mips/include/asm/mach-generic/
H A Dfloppy.h118 unsigned long mem; fd_dma_mem_alloc() local
120 mem = __get_dma_pages(GFP_KERNEL, get_order(size)); fd_dma_mem_alloc()
122 return mem; fd_dma_mem_alloc()
/linux-4.1.27/drivers/pinctrl/sh-pfc/
H A Dgpio.c36 struct sh_pfc_window *mem; member in struct:sh_pfc_chip
66 void __iomem *mem = address - chip->mem->phys + chip->mem->virt; gpio_read_data_reg() local
68 return sh_pfc_read_raw_reg(mem, dreg->reg_width); gpio_read_data_reg()
75 void __iomem *mem = address - chip->mem->phys + chip->mem->virt; gpio_write_data_reg() local
77 sh_pfc_write_raw_reg(mem, dreg->reg_width, value); gpio_write_data_reg()
315 struct sh_pfc_window *mem) sh_pfc_add_gpiochip()
324 chip->mem = mem; sh_pfc_add_gpiochip()
314 sh_pfc_add_gpiochip(struct sh_pfc *pfc, int(*setup)(struct sh_pfc_chip *), struct sh_pfc_window *mem) sh_pfc_add_gpiochip() argument
/linux-4.1.27/arch/m68k/kernel/
H A Dsys_m68k.c458 /* This syscall gets its arguments in A0 (mem), D2 (oldval) and
462 unsigned long __user * mem) sys_atomic_cmpxchg_32()
474 pgd = pgd_offset(mm, (unsigned long)mem); sys_atomic_cmpxchg_32()
477 pmd = pmd_offset(pgd, (unsigned long)mem); sys_atomic_cmpxchg_32()
480 pte = pte_offset_map_lock(mm, pmd, (unsigned long)mem, &ptl); sys_atomic_cmpxchg_32()
491 __get_user(mem_value, mem); sys_atomic_cmpxchg_32()
493 __put_user(newval, mem); sys_atomic_cmpxchg_32()
511 if (do_page_fault(fp, (unsigned long)mem, 3)) sys_atomic_cmpxchg_32()
531 /* This syscall gets its arguments in A0 (mem), D2 (oldval) and
535 unsigned long __user * mem) sys_atomic_cmpxchg_32()
542 mem_value = *mem; sys_atomic_cmpxchg_32()
544 *mem = newval; sys_atomic_cmpxchg_32()
461 sys_atomic_cmpxchg_32(unsigned long newval, int oldval, int d3, int d4, int d5, unsigned long __user * mem) sys_atomic_cmpxchg_32() argument
534 sys_atomic_cmpxchg_32(unsigned long newval, int oldval, int d3, int d4, int d5, unsigned long __user * mem) sys_atomic_cmpxchg_32() argument
/linux-4.1.27/drivers/mcb/
H A Dmcb-core.c350 struct resource *mem; mcb_request_mem() local
356 size = resource_size(&dev->mem); mcb_request_mem()
358 mem = request_mem_region(dev->mem.start, size, name); mcb_request_mem()
359 if (!mem) mcb_request_mem()
362 return mem; mcb_request_mem()
372 void mcb_release_mem(struct resource *mem) mcb_release_mem() argument
376 size = resource_size(mem); mcb_release_mem()
377 release_mem_region(mem->start, size); mcb_release_mem()
H A Dmcb-parse.c70 mdev->mem.start = mapbase + offset; chameleon_parse_gdd()
71 mdev->mem.end = mdev->mem.start + size - 1; chameleon_parse_gdd()
72 mdev->mem.flags = IORESOURCE_MEM; chameleon_parse_gdd()
/linux-4.1.27/drivers/media/usb/uvc/
H A Duvc_isight.c49 __u8 *mem; isight_decode() local
87 mem = buf->mem + buf->bytesused; isight_decode()
89 memcpy(mem, data, nbytes); isight_decode()
/linux-4.1.27/drivers/clk/sirf/
H A Dclk-atlas6.c63 rtc, osc, pll1, pll2, pll3, mem, sys, security, dsp, gps, enumerator in enum:atlas6_clk_index
145 clk_register_clkdev(atlas6_clks[mem], NULL, "mem"); atlas6_clk_init()
146 clk_register_clkdev(atlas6_clks[mem], NULL, "osc"); atlas6_clk_init()
H A Dclk-prima2.c62 rtc, osc, pll1, pll2, pll3, mem, sys, security, dsp, gps, enumerator in enum:prima2_clk_index
144 clk_register_clkdev(prima2_clks[mem], NULL, "mem"); prima2_clk_init()
145 clk_register_clkdev(prima2_clks[mem], NULL, "osc"); prima2_clk_init()
/linux-4.1.27/drivers/tty/serial/cpm_uart/
H A Dcpm_uart.h117 u32 mem = (u32)pinfo->mem_addr; cpu2cpm_addr() local
119 if (likely(val >= mem && val < mem + pinfo->mem_size)) { cpu2cpm_addr()
120 offset = val - mem; cpu2cpm_addr()
/linux-4.1.27/drivers/video/fbdev/
H A Ds3c2410fb.h25 struct resource *mem; member in struct:s3c2410fb_info
H A Dmetronomefb.c148 static u8 calc_cksum(int start, int end, u8 *mem) calc_cksum() argument
154 tmp += mem[i]; calc_cksum()
170 static int load_waveform(u8 *mem, size_t size, int m, int t, load_waveform() argument
195 wfm_hdr = (struct waveform_hdr *) mem; load_waveform()
205 cksum = calc_cksum(32, 47, mem); load_waveform()
228 if (mem[i] > t) { load_waveform()
238 cksum = calc_cksum(sizeof(*wfm_hdr), cksum_idx, mem); load_waveform()
239 if (cksum != mem[cksum_idx]) { load_waveform()
241 " %x != %x\n", cksum, mem[cksum_idx]); load_waveform()
250 cksum = calc_cksum(cksum_idx - 3, cksum_idx, mem); load_waveform()
251 if (cksum != mem[cksum_idx]) { load_waveform()
253 " %x != %x\n", cksum, mem[cksum_idx]); load_waveform()
258 tta = get_unaligned_le32(mem + wmta + m * 4) & 0x00FFFFFF; load_waveform()
262 cksum = calc_cksum(cksum_idx - 3, cksum_idx, mem); load_waveform()
263 if (cksum != mem[cksum_idx]) { load_waveform()
265 " %x != %x\n", cksum, mem[cksum_idx]); load_waveform()
271 wfm_idx = get_unaligned_le32(mem + tta + trn * 4) & 0x00FFFFFF; load_waveform()
277 v = mem[wfm_idx++]; load_waveform()
279 while (((v = mem[wfm_idx++]) != wfm_hdr->swtb) && load_waveform()
289 rl = mem[wfm_idx++]; load_waveform()
297 cksum = calc_cksum(owfm_idx, cksum_idx, mem); load_waveform()
298 if (cksum != mem[cksum_idx]) { load_waveform()
300 " %x != %x\n", cksum, mem[cksum_idx]); load_waveform()
H A Dnuc900fb.h37 struct resource *mem; member in struct:nuc900fb_info
/linux-4.1.27/arch/powerpc/mm/
H A DMakefile9 obj-y := fault.o mem.o pgtable.o mmap.o \
/linux-4.1.27/arch/arm64/mm/
H A Dmmap.c98 * You really shouldn't be using read() or write() on /dev/mem. This might go
112 * Do not allow /dev/mem mappings beyond the supported physical range.
124 * devmem_is_allowed() checks to see if /dev/mem access to a certain address
127 * This effectively disable read()/write() on /dev/mem.
/linux-4.1.27/arch/metag/include/asm/
H A Dmmzone.h36 /* SoC specific mem init */
/linux-4.1.27/arch/arm/plat-iop/
H A Dsetup.c23 { /* mem mapped registers */
/linux-4.1.27/drivers/net/can/softing/
H A Dsofting_fw.c109 const uint8_t *mem; fw_parse() local
127 mem = *pmem; fw_parse()
128 *ptype = le16_to_cpup((void *)&mem[0]); fw_parse()
129 *paddr = le32_to_cpup((void *)&mem[2]); fw_parse()
130 *plen = le16_to_cpup((void *)&mem[6]); fw_parse()
131 *pdat = &mem[8]; fw_parse()
133 end = &mem[8 + *plen]; fw_parse()
135 for (checksum[1] = 0; mem < end; ++mem) fw_parse()
136 checksum[1] += *mem; fw_parse()
149 const uint8_t *mem, *end, *dat; softing_load_fw() local
164 mem = fw->data; softing_load_fw()
165 end = &mem[fw->size]; softing_load_fw()
167 ret = fw_parse(&mem, &type, &addr, &len, &dat); softing_load_fw()
177 while (mem < end) { softing_load_fw()
178 ret = fw_parse(&mem, &type, &addr, &len, &dat); softing_load_fw()
232 const uint8_t *mem, *end, *dat; softing_load_app_fw() local
248 mem = fw->data; softing_load_app_fw()
249 end = &mem[fw->size]; softing_load_app_fw()
251 ret = fw_parse(&mem, &type, &addr, &len, &dat); softing_load_app_fw()
266 while (mem < end) { softing_load_app_fw()
267 ret = fw_parse(&mem, &type, &addr, &len, &dat); softing_load_app_fw()
/linux-4.1.27/drivers/net/ethernet/8390/
H A Dax88796.c119 void __iomem *ioaddr = ei_local->mem; ax_initial_check()
174 void __iomem *nic_base = ei_local->mem; ax_get_8390_hdr()
218 void __iomem *nic_base = ei_local->mem; ax_block_input()
255 void __iomem *nic_base = ei_local->mem; ax_block_output()
392 ei_outb(reg_gpoc, ei_local->mem + EI_SHIFT(0x17)); ax_phy_switch()
527 u8 reg = ei_inb(ei_local->mem + AX_MEMR); ax_eeprom_register_read()
538 u8 reg = ei_inb(ei_local->mem + AX_MEMR); ax_eeprom_register_write()
549 ei_outb(reg, ei_local->mem + AX_MEMR); ax_eeprom_register_write()
633 ax->addr_memr = ei_local->mem + AX_MEMR; ax_mii_init()
670 void __iomem *ioaddr = ei_local->mem; ax_initial_setup()
693 void __iomem *ioaddr = ei_local->mem; ax_init_dev()
743 ei_outb(ax->plat->dcr_val, ei_local->mem + EN0_DCFG); ax_init_dev()
754 ei_local->mem + E8390_CMD); /* 0x61 */ ax_init_dev()
815 struct resource *mem; ax_remove() local
820 iounmap(ei_local->mem); ax_remove()
821 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); ax_remove()
822 release_mem_region(mem->start, resource_size(mem)); ax_remove()
826 mem = platform_get_resource(pdev, IORESOURCE_MEM, 1); ax_remove()
827 release_mem_region(mem->start, resource_size(mem)); ax_remove()
847 struct resource *irq, *mem, *mem2; ax_probe() local
876 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); ax_probe()
877 if (!mem) { ax_probe()
883 mem_size = resource_size(mem); ax_probe()
897 if (!request_mem_region(mem->start, mem_size, pdev->name)) { ax_probe()
903 ei_local->mem = ioremap(mem->start, mem_size); ax_probe()
904 dev->base_addr = (unsigned long)ei_local->mem; ax_probe()
906 if (ei_local->mem == NULL) { ax_probe()
907 dev_err(&pdev->dev, "Cannot ioremap area %pR\n", mem); ax_probe()
936 ei_local->reg_offset[0x1f] = ax->map2 - ei_local->mem; ax_probe()
953 iounmap(ei_local->mem); ax_probe()
956 release_mem_region(mem->start, mem_size); ax_probe()
H A Dmcf8390.c412 struct resource *mem, *irq; mcf8390_probe() local
422 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); mcf8390_probe()
423 if (mem == NULL) { mcf8390_probe()
427 msize = resource_size(mem); mcf8390_probe()
428 if (!request_mem_region(mem->start, msize, pdev->name)) mcf8390_probe()
433 release_mem_region(mem->start, msize); mcf8390_probe()
443 dev->base_addr = mem->start; mcf8390_probe()
447 release_mem_region(mem->start, msize); mcf8390_probe()
457 struct resource *mem; mcf8390_remove() local
460 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); mcf8390_remove()
461 if (mem) mcf8390_remove()
462 release_mem_region(mem->start, resource_size(mem)); mcf8390_remove()
/linux-4.1.27/arch/tile/gxio/
H A Dmpipe.c163 buffer_size_enum, void *mem, size_t mem_size, gxio_mpipe_init_buffer_stack()
168 memset(mem, 0, mem_size); gxio_mpipe_init_buffer_stack()
170 result = gxio_mpipe_init_buffer_stack_aux(context, mem, mem_size, gxio_mpipe_init_buffer_stack()
186 void *mem, size_t mem_size, gxio_mpipe_init_notif_ring()
189 return gxio_mpipe_init_notif_ring_aux(context, mem, mem_size, gxio_mpipe_init_notif_ring()
237 void *mem, size_t mem_size, gxio_mpipe_init_edma_ring()
240 memset(mem, 0, mem_size); gxio_mpipe_init_edma_ring()
242 return gxio_mpipe_init_edma_ring_aux(context, mem, mem_size, mem_flags, gxio_mpipe_init_edma_ring()
392 void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_iqueue_init()
398 iqueue->idescs = (gxio_mpipe_idesc_t *)mem; gxio_mpipe_iqueue_init()
409 __gxio_mmio_write(mem, iqueue->head); gxio_mpipe_iqueue_init()
411 return gxio_mpipe_init_notif_ring(context, ring, mem, mem_size, gxio_mpipe_iqueue_init()
421 void *mem, unsigned int mem_size, gxio_mpipe_equeue_init()
431 mem, mem_size, mem_flags); gxio_mpipe_equeue_init()
446 equeue->edescs = mem; gxio_mpipe_equeue_init()
160 gxio_mpipe_init_buffer_stack(gxio_mpipe_context_t *context, unsigned int stack, gxio_mpipe_buffer_size_enum_t buffer_size_enum, void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_init_buffer_stack() argument
184 gxio_mpipe_init_notif_ring(gxio_mpipe_context_t *context, unsigned int ring, void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_init_notif_ring() argument
235 gxio_mpipe_init_edma_ring(gxio_mpipe_context_t *context, unsigned int ring, unsigned int channel, void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_init_edma_ring() argument
389 gxio_mpipe_iqueue_init(gxio_mpipe_iqueue_t *iqueue, gxio_mpipe_context_t *context, unsigned int ring, void *mem, size_t mem_size, unsigned int mem_flags) gxio_mpipe_iqueue_init() argument
417 gxio_mpipe_equeue_init(gxio_mpipe_equeue_t *equeue, gxio_mpipe_context_t *context, unsigned int ering, unsigned int channel, void *mem, unsigned int mem_size, unsigned int mem_flags) gxio_mpipe_equeue_init() argument
/linux-4.1.27/arch/tile/include/asm/
H A Duaccess.h527 * @mem: Destination address, in user space.
535 extern unsigned long clear_user_asm(void __user *mem, unsigned long len); __clear_user()
537 void __user *mem, unsigned long len) __clear_user()
540 return clear_user_asm(mem, len); __clear_user()
543 void __user *mem, unsigned long len) clear_user()
545 if (access_ok(VERIFY_WRITE, mem, len)) clear_user()
546 return __clear_user(mem, len); clear_user()
552 * @mem: Destination address, in user space.
558 extern unsigned long flush_user_asm(void __user *mem, unsigned long len); __flush_user()
560 void __user *mem, unsigned long len) __flush_user()
565 retval = flush_user_asm(mem, len); __flush_user()
571 void __user *mem, unsigned long len) flush_user()
573 if (access_ok(VERIFY_WRITE, mem, len)) flush_user()
574 return __flush_user(mem, len); flush_user()
580 * @mem: Destination address, in user space.
586 extern unsigned long finv_user_asm(void __user *mem, unsigned long len); __finv_user()
588 void __user *mem, unsigned long len) __finv_user()
593 retval = finv_user_asm(mem, len); __finv_user()
598 void __user *mem, unsigned long len) finv_user()
600 if (access_ok(VERIFY_WRITE, mem, len)) finv_user()
601 return __finv_user(mem, len); finv_user()
536 __clear_user( void __user *mem, unsigned long len) __clear_user() argument
542 clear_user( void __user *mem, unsigned long len) clear_user() argument
559 __flush_user( void __user *mem, unsigned long len) __flush_user() argument
570 flush_user( void __user *mem, unsigned long len) flush_user() argument
587 __finv_user( void __user *mem, unsigned long len) __finv_user() argument
597 finv_user( void __user *mem, unsigned long len) finv_user() argument
/linux-4.1.27/arch/unicore32/include/asm/
H A Dio.h57 * devmem_is_allowed() checks to see if /dev/mem access to a certain
61 * on /dev/mem.
/linux-4.1.27/arch/mips/include/asm/dec/
H A Decc.h35 #define KN0X_ESR_CHKHI (0x7f<<24) /* check bits read from mem */
39 #define KN0X_ESR_CHKLO (0x7f<<8) /* check bits read from mem */
/linux-4.1.27/drivers/usb/isp1760/
H A Disp1760-hcd.h73 struct resource *mem, int irq, unsigned long irqflags,
81 void __iomem *regs, struct resource *mem, isp1760_hcd_register()
80 isp1760_hcd_register(struct isp1760_hcd *priv, void __iomem *regs, struct resource *mem, int irq, unsigned long irqflags, struct device *dev) isp1760_hcd_register() argument
/linux-4.1.27/include/video/
H A Dsh_mobile_meram.h44 unsigned long mem, size_t size);
65 unsigned long mem, size_t size) sh_mobile_meram_free()
64 sh_mobile_meram_free(struct sh_mobile_meram_info *meram_dev, unsigned long mem, size_t size) sh_mobile_meram_free() argument
/linux-4.1.27/arch/arm64/kernel/
H A Darm64ksyms.c35 /* user mem (segment) */
44 /* string / mem functions */
/linux-4.1.27/arch/arc/lib/
H A Dmemset.S51 ; memzero: @r0 = mem, @r1 = size_t
52 ; memset: @r0 = mem, @r1 = char, @r2 = size_t
/linux-4.1.27/sound/soc/tegra/
H A Dtegra20_spdif.c268 struct resource *mem, *memregion, *dmareq; tegra20_spdif_platform_probe() local
288 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); tegra20_spdif_platform_probe()
289 if (!mem) { tegra20_spdif_platform_probe()
302 memregion = devm_request_mem_region(&pdev->dev, mem->start, tegra20_spdif_platform_probe()
303 resource_size(mem), DRV_NAME); tegra20_spdif_platform_probe()
310 regs = devm_ioremap(&pdev->dev, mem->start, resource_size(mem)); tegra20_spdif_platform_probe()
325 spdif->playback_dma_data.addr = mem->start + TEGRA20_SPDIF_DATA_OUT; tegra20_spdif_platform_probe()
H A Dtegra20_i2s.c342 struct resource *mem, *memregion; tegra20_i2s_platform_probe() local
364 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); tegra20_i2s_platform_probe()
365 if (!mem) { tegra20_i2s_platform_probe()
371 memregion = devm_request_mem_region(&pdev->dev, mem->start, tegra20_i2s_platform_probe()
372 resource_size(mem), DRV_NAME); tegra20_i2s_platform_probe()
379 regs = devm_ioremap(&pdev->dev, mem->start, resource_size(mem)); tegra20_i2s_platform_probe()
394 i2s->capture_dma_data.addr = mem->start + TEGRA20_I2S_FIFO2; tegra20_i2s_platform_probe()
398 i2s->playback_dma_data.addr = mem->start + TEGRA20_I2S_FIFO1; tegra20_i2s_platform_probe()
/linux-4.1.27/drivers/media/common/saa7146/
H A Dsaa7146_core.c179 void *mem = vmalloc_32(length); saa7146_vmalloc_build_pgtable() local
182 if (NULL == mem) saa7146_vmalloc_build_pgtable()
185 if (!(pt->slist = vmalloc_to_sg(mem, pages))) saa7146_vmalloc_build_pgtable()
199 return mem; saa7146_vmalloc_build_pgtable()
209 vfree(mem); saa7146_vmalloc_build_pgtable()
214 void saa7146_vfree_destroy_pgtable(struct pci_dev *pci, void *mem, struct saa7146_pgtable *pt) saa7146_vfree_destroy_pgtable() argument
220 vfree(mem); saa7146_vfree_destroy_pgtable()
360 /* clear out mem for sure */ saa7146_init_one()
392 dev->mem = ioremap(pci_resource_start(pci, 0), saa7146_init_one()
394 if (!dev->mem) { saa7146_init_one()
444 pr_info("found saa7146 @ mem %p (revision %d, irq %d) (0x%04x,0x%04x)\n", saa7146_init_one()
445 dev->mem, dev->revision, pci->irq, saa7146_init_one()
497 iounmap(dev->mem); saa7146_init_one()
536 iounmap(dev->mem); saa7146_remove_one()
/linux-4.1.27/drivers/gpu/host1x/hw/
H A Dchannel_hw.c36 void *mem = NULL; trace_write_gather() local
39 mem = host1x_bo_mmap(bo); trace_write_gather()
41 if (mem) { trace_write_gather()
53 mem); trace_write_gather() local
56 host1x_bo_munmap(bo, mem); trace_write_gather()
/linux-4.1.27/arch/nios2/kernel/
H A Dkgdb.c84 char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs) dbg_get_reg() argument
90 memcpy(mem, (void *)regs + dbg_reg_def[regno].offset, dbg_get_reg()
93 memset(mem, 0, dbg_reg_def[regno].size); dbg_get_reg()
98 int dbg_set_reg(int regno, void *mem, struct pt_regs *regs) dbg_set_reg() argument
104 memcpy((void *)regs + dbg_reg_def[regno].offset, mem, dbg_set_reg()
/linux-4.1.27/arch/arm/boot/compressed/
H A Datags_to_fdt.c153 if (!atag->u.mem.size) for_each_tag()
164 cpu_to_fdt64(atag->u.mem.start); for_each_tag()
166 cpu_to_fdt64(atag->u.mem.size); for_each_tag()
169 cpu_to_fdt32(atag->u.mem.start); for_each_tag()
171 cpu_to_fdt32(atag->u.mem.size); for_each_tag()
/linux-4.1.27/sound/oss/
H A Dkahlua.c59 void __iomem *mem; probe_one() local
70 mem = ioremap(base, 128); probe_one()
71 if (!mem) probe_one()
73 map = readw(mem + 0x18); /* Read the SMI enables */ probe_one()
74 iounmap(mem); probe_one()
/linux-4.1.27/drivers/media/pci/cx23885/
H A Dcimax2.c160 int mem; netup_ci_get_mem() local
164 mem = cx_read(MC417_RWD); netup_ci_get_mem()
165 if ((mem & NETUP_ACK) == 0) netup_ci_get_mem()
174 return mem & 0xff; netup_ci_get_mem()
185 int mem; netup_ci_op_cam() local
228 mem = netup_ci_get_mem(dev); netup_ci_op_cam()
233 if (mem < 0) netup_ci_op_cam()
238 (flag == NETUP_CI_CTL) ? "ctl" : "mem", netup_ci_op_cam()
239 (read) ? mem : data); netup_ci_op_cam()
242 return mem; netup_ci_op_cam()
/linux-4.1.27/arch/s390/kvm/
H A Dpriv.c475 static void handle_stsi_3_2_2(struct kvm_vcpu *vcpu, struct sysinfo_3_2_2 *mem) handle_stsi_3_2_2() argument
483 if (stsi(mem, 3, 2, 2)) handle_stsi_3_2_2()
484 mem->count = 0; handle_stsi_3_2_2()
485 if (mem->count < 8) handle_stsi_3_2_2()
486 mem->count++; handle_stsi_3_2_2()
487 for (n = mem->count - 1; n > 0 ; n--) handle_stsi_3_2_2()
488 memcpy(&mem->vm[n], &mem->vm[n - 1], sizeof(mem->vm[0])); handle_stsi_3_2_2()
490 memset(&mem->vm[0], 0, sizeof(mem->vm[0])); handle_stsi_3_2_2()
491 mem->vm[0].cpus_total = cpus; handle_stsi_3_2_2()
492 mem->vm[0].cpus_configured = cpus; handle_stsi_3_2_2()
493 mem->vm[0].cpus_standby = 0; handle_stsi_3_2_2()
494 mem->vm[0].cpus_reserved = 0; handle_stsi_3_2_2()
495 mem->vm[0].caf = 1000; handle_stsi_3_2_2()
496 memcpy(mem->vm[0].name, "KVMguest", 8); handle_stsi_3_2_2()
497 ASCEBC(mem->vm[0].name, 8); handle_stsi_3_2_2()
498 memcpy(mem->vm[0].cpi, "KVM/Linux ", 16); handle_stsi_3_2_2()
499 ASCEBC(mem->vm[0].cpi, 16); handle_stsi_3_2_2()
518 unsigned long mem = 0; handle_stsi() local
552 mem = get_zeroed_page(GFP_KERNEL); handle_stsi()
553 if (!mem) handle_stsi()
555 if (stsi((void *) mem, fc, sel1, sel2)) handle_stsi()
561 mem = get_zeroed_page(GFP_KERNEL); handle_stsi()
562 if (!mem) handle_stsi()
564 handle_stsi_3_2_2(vcpu, (void *) mem); handle_stsi()
568 rc = write_guest(vcpu, operand2, ar, (void *)mem, PAGE_SIZE); handle_stsi()
578 free_page(mem); handle_stsi()
585 free_page(mem); handle_stsi()
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
H A Docrdma_stats.c54 struct stats_mem *mem = &dev->stats_mem; ocrdma_alloc_stats_mem() local
56 /* Alloc mbox command mem*/ ocrdma_alloc_stats_mem()
57 mem->size = max_t(u32, sizeof(struct ocrdma_rdma_stats_req), ocrdma_alloc_stats_mem()
60 mem->va = dma_alloc_coherent(&dev->nic_info.pdev->dev, mem->size, ocrdma_alloc_stats_mem()
61 &mem->pa, GFP_KERNEL); ocrdma_alloc_stats_mem()
62 if (!mem->va) { ocrdma_alloc_stats_mem()
67 memset(mem->va, 0, mem->size); ocrdma_alloc_stats_mem()
69 /* Alloc debugfs mem */ ocrdma_alloc_stats_mem()
70 mem->debugfs_mem = kzalloc(OCRDMA_MAX_DBGFS_MEM, GFP_KERNEL); ocrdma_alloc_stats_mem()
71 if (!mem->debugfs_mem) { ocrdma_alloc_stats_mem()
72 pr_err("%s: stats debugfs mem allocation failed\n", __func__); ocrdma_alloc_stats_mem()
81 struct stats_mem *mem = &dev->stats_mem; ocrdma_release_stats_mem() local
83 if (mem->va) ocrdma_release_stats_mem()
84 dma_free_coherent(&dev->nic_info.pdev->dev, mem->size, ocrdma_release_stats_mem()
85 mem->va, mem->pa); ocrdma_release_stats_mem()
86 kfree(mem->debugfs_mem); ocrdma_release_stats_mem()
/linux-4.1.27/drivers/gpu/host1x/
H A Djob.c41 void *mem; host1x_job_alloc() local
54 mem = job = kzalloc(total, GFP_KERNEL); host1x_job_alloc()
62 mem += sizeof(struct host1x_job); host1x_job_alloc()
63 job->relocarray = num_relocs ? mem : NULL; host1x_job_alloc()
64 mem += num_relocs * sizeof(struct host1x_reloc); host1x_job_alloc()
65 job->unpins = num_unpins ? mem : NULL; host1x_job_alloc()
66 mem += num_unpins * sizeof(struct host1x_job_unpin_data); host1x_job_alloc()
67 job->waitchk = num_waitchks ? mem : NULL; host1x_job_alloc()
68 mem += num_waitchks * sizeof(struct host1x_waitchk); host1x_job_alloc()
69 job->gathers = num_cmdbufs ? mem : NULL; host1x_job_alloc()
70 mem += num_cmdbufs * sizeof(struct host1x_job_gather); host1x_job_alloc()
71 job->addr_phys = num_unpins ? mem : NULL; host1x_job_alloc()
535 /* process each gather mem only once */ host1x_job_pin()
/linux-4.1.27/arch/mips/kernel/
H A Dkgdb.c130 int dbg_set_reg(int regno, void *mem, struct pt_regs *regs) dbg_set_reg() argument
138 memcpy((void *)regs + dbg_reg_def[regno].offset, mem, dbg_set_reg()
146 memcpy((void *)&current->thread.fpu.fcr31, mem, dbg_set_reg()
154 memcpy((void *)&current->thread.fpu.fpr[fp_reg], mem, dbg_set_reg()
163 char *dbg_get_reg(int regno, void *mem, struct pt_regs *regs) dbg_get_reg() argument
172 memcpy(mem, (void *)regs + dbg_reg_def[regno].offset, dbg_get_reg()
181 memcpy(mem, (void *)&current->thread.fpu.fcr31, dbg_get_reg()
186 memset(mem, 0, dbg_reg_def[regno].size); dbg_get_reg()
190 memcpy(mem, (void *)&current->thread.fpu.fpr[fp_reg], dbg_get_reg()
/linux-4.1.27/drivers/isdn/hardware/eicon/
H A Dpc_maint.h131 POINTER_32BIT mem; member in struct:pc_maint
143 POINTER_32BIT mem; member in struct:pc_maint
155 POINTER_32BIT mem; member in struct:mi_pc_maint
/linux-4.1.27/drivers/media/usb/go7007/
H A Dgo7007-fw.c376 u16 mem = 0x3e00; gen_mjpeghdr_to_package() local
398 code[off + 1] = __cpu_to_le16(0x8000 | mem); gen_mjpeghdr_to_package()
401 if (mem + chunk > 0x4000) gen_mjpeghdr_to_package()
402 chunk = 0x4000 - mem; gen_mjpeghdr_to_package()
409 mem = 0x3e00; gen_mjpeghdr_to_package()
413 mem += 28; gen_mjpeghdr_to_package()
643 u16 mem = 0x3e00; gen_mpeg1hdr_to_package() local
687 code[off + 1] = __cpu_to_le16(0x8000 | mem); gen_mpeg1hdr_to_package()
690 if (mem + chunk > 0x4000) gen_mpeg1hdr_to_package()
691 chunk = 0x4000 - mem; gen_mpeg1hdr_to_package()
698 if (mem + chunk == 0x4000) { gen_mpeg1hdr_to_package()
699 mem = 0x3e00; gen_mpeg1hdr_to_package()
705 mem += 28; gen_mpeg1hdr_to_package()
829 u16 mem = 0x3e00; gen_mpeg4hdr_to_package() local
855 code[off + 1] = __cpu_to_le16(0x8000 | mem); gen_mpeg4hdr_to_package()
858 if (mem + chunk > 0x4000) gen_mpeg4hdr_to_package()
859 chunk = 0x4000 - mem; gen_mpeg4hdr_to_package()
866 if (mem + chunk == 0x4000) { gen_mpeg4hdr_to_package()
867 mem = 0x3e00; gen_mpeg4hdr_to_package()
873 mem += 28; gen_mpeg4hdr_to_package()
879 mem = 0x3e00; gen_mpeg4hdr_to_package()
898 code[off + 1] = __cpu_to_le16(0x8000 | mem); gen_mpeg4hdr_to_package()
901 if (mem + chunk > 0x4000) gen_mpeg4hdr_to_package()
902 chunk = 0x4000 - mem; gen_mpeg4hdr_to_package()
909 if (mem + chunk == 0x4000) { gen_mpeg4hdr_to_package()
910 mem = 0x3e00; gen_mpeg4hdr_to_package()
916 mem += 28; gen_mpeg4hdr_to_package()
/linux-4.1.27/drivers/net/wireless/prism54/
H A Dislpci_mgt.c123 if (buf->mem == NULL) { islpci_mgmt_rx_fill()
124 buf->mem = kmalloc(MGMT_FRAME_SIZE, GFP_ATOMIC); islpci_mgmt_rx_fill()
125 if (!buf->mem) islpci_mgmt_rx_fill()
130 buf->pci_addr = pci_map_single(priv->pdev, buf->mem, islpci_mgmt_rx_fill()
189 p = buf.mem = kmalloc(frag_len, GFP_KERNEL); islpci_mgt_transmit()
190 if (!buf.mem) islpci_mgt_transmit()
206 pimfor_header_t *h = buf.mem; islpci_mgt_transmit()
218 buf.pci_addr = pci_map_single(priv->pdev, buf.mem, frag_len, islpci_mgt_transmit()
257 kfree(buf.mem); islpci_mgt_transmit()
321 header = pimfor_decode_header(buf->mem, frag_len); islpci_mgt_receive()
432 kfree(buf->mem); islpci_mgt_cleanup_transmit()
433 buf->mem = NULL; islpci_mgt_cleanup_transmit()
/linux-4.1.27/include/drm/ttm/
H A Dttm_bo_driver.h188 * @mem: Pointer to a struct ttm_mem_reg to be filled in.
193 * @mem::mm_node should be set to a non-null value, and
194 * @mem::start should be set to a value identifying the beginning
196 * If the memory region accommodate the buffer object, @mem::mm_node
201 * Note that @mem::mm_node will only be dereferenced by
212 struct ttm_mem_reg *mem);
218 * @mem: Pointer to a struct ttm_mem_reg to be filled in.
221 * and that are identified by @mem::mm_node and @mem::start. May not
225 struct ttm_mem_reg *mem);
259 * @io_reserve_lru: Optional lru list for unreserving io mem regions.
378 * finished, they'll end up in bo->mem.flags
435 int (*io_mem_reserve)(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem);
436 void (*io_mem_free)(struct ttm_bo_device *bdev, struct ttm_mem_reg *mem);
667 * @mem: A valid struct ttm_mem_reg.
669 * Returns true if the memory described by @mem is PCI memory,
673 struct ttm_mem_reg *mem);
681 * @mem: A struct ttm_mem_reg.
686 * the placement flags in @mem, potentially evicting other idle buffer objects.
696 struct ttm_mem_reg *mem,
701 struct ttm_mem_reg *mem);
703 struct ttm_mem_reg *mem);
917 if (!(bo->mem.placement & TTM_PL_FLAG_NO_EVICT)) { ttm_bo_unreserve()
943 struct ttm_mem_reg *mem);
945 struct ttm_mem_reg *mem);
957 * and update the (@bo)->mem placement flags. If unsuccessful, the old
979 * and update the (@bo)->mem placement flags. If unsuccessful, the old
/linux-4.1.27/drivers/scsi/mpt3sas/
H A Dmpt3sas_config.c181 * @mem: struct config_request
189 struct config_request *mem) _config_alloc_config_dma_memory()
193 if (mem->sz > ioc->config_page_sz) { _config_alloc_config_dma_memory()
194 mem->page = dma_alloc_coherent(&ioc->pdev->dev, mem->sz, _config_alloc_config_dma_memory()
195 &mem->page_dma, GFP_KERNEL); _config_alloc_config_dma_memory()
196 if (!mem->page) { _config_alloc_config_dma_memory()
199 ioc->name, __func__, mem->sz); _config_alloc_config_dma_memory()
203 mem->page = ioc->config_page; _config_alloc_config_dma_memory()
204 mem->page_dma = ioc->config_page_dma; _config_alloc_config_dma_memory()
212 * @mem: struct config_request
220 struct config_request *mem) _config_free_config_dma_memory()
222 if (mem->sz > ioc->config_page_sz) _config_free_config_dma_memory()
223 dma_free_coherent(&ioc->pdev->dev, mem->sz, mem->page, _config_free_config_dma_memory()
224 mem->page_dma); _config_free_config_dma_memory()
297 struct config_request mem; _config_request() local
309 memset(&mem, 0, sizeof(struct config_request)); _config_request()
322 mem.sz = mpi_request->Header.PageLength * 4; _config_request()
324 mem.sz = le16_to_cpu(mpi_reply->ExtPageLength) * 4; _config_request()
325 r = _config_alloc_config_dma_memory(ioc, &mem); _config_request()
333 MPT3_CONFIG_COMMON_WRITE_SGLFLAGS | mem.sz, _config_request()
334 mem.page_dma); _config_request()
335 memcpy(mem.page, config_page, min_t(u16, mem.sz, _config_request()
340 MPT3_CONFIG_COMMON_SGLFLAGS | mem.sz, mem.page_dma); _config_request()
341 memset(mem.page, 0, min_t(u16, mem.sz, config_page_sz)); _config_request()
452 u8 *p = (u8 *)mem.page; _config_request()
460 _debug_dump_config(p, min_t(u16, mem.sz, _config_request()
477 _debug_dump_config(p, min_t(u16, mem.sz, _config_request()
488 memcpy(config_page, mem.page, min_t(u16, mem.sz, _config_request()
494 _config_free_config_dma_memory(ioc, &mem); _config_request()
188 _config_alloc_config_dma_memory(struct MPT3SAS_ADAPTER *ioc, struct config_request *mem) _config_alloc_config_dma_memory() argument
219 _config_free_config_dma_memory(struct MPT3SAS_ADAPTER *ioc, struct config_request *mem) _config_free_config_dma_memory() argument
/linux-4.1.27/sound/soc/codecs/
H A Dwm_adsp.c337 if (dsp->mem[i].type == type) wm_adsp_find_region()
338 return &dsp->mem[i]; wm_adsp_find_region()
380 const struct wm_adsp_region *mem; wm_coeff_write_control() local
386 mem = wm_adsp_find_region(adsp, region->type); wm_coeff_write_control()
387 if (!mem) { wm_coeff_write_control()
394 reg = wm_adsp_region_to_reg(mem, reg); wm_coeff_write_control()
435 const struct wm_adsp_region *mem; wm_coeff_read_control() local
441 mem = wm_adsp_find_region(adsp, region->type); wm_coeff_read_control()
442 if (!mem) { wm_coeff_read_control()
449 reg = wm_adsp_region_to_reg(mem, reg); wm_coeff_read_control()
534 const struct wm_adsp_region *mem; wm_adsp_load() local
633 mem = wm_adsp_find_region(dsp, type); wm_adsp_load()
652 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load()
656 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load()
660 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load()
664 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load()
668 reg = wm_adsp_region_to_reg(mem, offset); wm_adsp_load()
888 const struct wm_adsp_region *mem; wm_adsp_setup_algs() local
896 mem = wm_adsp_find_region(dsp, WMFW_ADSP1_DM); wm_adsp_setup_algs()
899 mem = wm_adsp_find_region(dsp, WMFW_ADSP2_XM); wm_adsp_setup_algs()
902 mem = NULL; wm_adsp_setup_algs()
906 if (WARN_ON(!mem)) wm_adsp_setup_algs()
911 ret = regmap_raw_read(regmap, mem->base, &adsp1_id, wm_adsp_setup_algs()
952 ret = regmap_raw_read(regmap, mem->base, &adsp2_id, wm_adsp_setup_algs()
1018 ret = regmap_raw_read(regmap, mem->base + term, &val, sizeof(val)); wm_adsp_setup_algs()
1033 ret = regmap_raw_read(regmap, mem->base + pos, alg, (term - pos) * 2); wm_adsp_setup_algs()
1181 const struct wm_adsp_region *mem; wm_adsp_load_coeff() local
1263 mem = wm_adsp_find_region(dsp, type); wm_adsp_load_coeff()
1264 if (!mem) { wm_adsp_load_coeff()
1268 reg = wm_adsp_region_to_reg(mem, 0); wm_adsp_load_coeff()
1284 mem = wm_adsp_find_region(dsp, type); wm_adsp_load_coeff()
1285 if (!mem) { wm_adsp_load_coeff()
1296 reg = wm_adsp_region_to_reg(mem, wm_adsp_load_coeff()
/linux-4.1.27/drivers/net/appletalk/
H A Dltpc.h65 struct lt_mem mem; member in union:lt_command
/linux-4.1.27/drivers/net/ethernet/qlogic/qlcnic/
H A Dqlcnic_minidump.c235 struct __mem mem; member in union:qlcnic_dump_entry::__anon7208
564 struct __mem *rom = &entry->region.mem; qlcnic_read_rom()
662 struct __mem *mem, __le32 *buffer, qlcnic_read_memory_test_agent()
668 reg_read = mem->size; qlcnic_read_memory_test_agent()
669 addr = mem->addr; qlcnic_read_memory_test_agent()
709 return mem->size; qlcnic_read_memory_test_agent()
721 struct __mem *mem) qlcnic_start_pex_dma()
734 ret = qlcnic_ind_wr(adapter, temp_addr, mem->desc_card_addr); qlcnic_start_pex_dma()
744 ret = qlcnic_ind_wr(adapter, temp_addr, mem->start_dma_cmd); qlcnic_start_pex_dma()
768 struct __mem *mem, qlcnic_read_memory_pexdma()
803 * dma_desc_cmd 16:19 = mem->dma_desc_cmd 0:3 qlcnic_read_memory_pexdma()
805 * dma_desc_cmd 24:31 = mem->dma_desc_cmd 8:15 qlcnic_read_memory_pexdma()
810 temp = mem->dma_desc_cmd & 0xff0f; qlcnic_read_memory_pexdma()
818 while (read_size < mem->size) { qlcnic_read_memory_pexdma()
819 if (mem->size - read_size >= QLC_PEX_DMA_READ_SIZE) qlcnic_read_memory_pexdma()
822 size = mem->size - read_size; qlcnic_read_memory_pexdma()
824 dma_descr->src_addr_low = mem->addr + read_size; qlcnic_read_memory_pexdma()
829 *ret = qlcnic_ms_mem_write128(adapter, mem->desc_card_addr, qlcnic_read_memory_pexdma()
833 mem->desc_card_addr); qlcnic_read_memory_pexdma()
837 *ret = qlcnic_start_pex_dma(adapter, mem); qlcnic_read_memory_pexdma()
859 struct __mem *mem = &entry->region.mem; qlcnic_read_memory() local
864 data_size = qlcnic_read_memory_pexdma(adapter, mem, buffer, qlcnic_read_memory()
874 data_size = qlcnic_read_memory_test_agent(adapter, mem, buffer, &ret); qlcnic_read_memory()
1021 struct __mem *rom = &entry->region.mem; qlcnic_83xx_dump_rom()
661 qlcnic_read_memory_test_agent(struct qlcnic_adapter *adapter, struct __mem *mem, __le32 *buffer, int *ret) qlcnic_read_memory_test_agent() argument
720 qlcnic_start_pex_dma(struct qlcnic_adapter *adapter, struct __mem *mem) qlcnic_start_pex_dma() argument
767 qlcnic_read_memory_pexdma(struct qlcnic_adapter *adapter, struct __mem *mem, __le32 *buffer, int *ret) qlcnic_read_memory_pexdma() argument
/linux-4.1.27/drivers/media/platform/omap3isp/
H A Dispcsiphy.h32 /* mem resources - enums as defined in enum isp_mem_resources */
/linux-4.1.27/arch/sparc/include/asm/
H A Dstring_32.h22 /* First the mem*() things. */
H A Dstring_64.h20 /* First the mem*() things. */
/linux-4.1.27/arch/um/kernel/
H A DMakefile13 obj-y = config.o exec.o exitcode.o irq.o ksyms.o mem.o \
/linux-4.1.27/drivers/char/
H A DMakefile5 obj-y += mem.o random.o
/linux-4.1.27/arch/sh/include/asm/
H A Dmmzone.h37 /* Platform specific mem init */

Completed in 8191 milliseconds

12345678