Home
last modified time | relevance | path

Searched refs:cma (Results 1 – 20 of 20) sorted by relevance

/linux-4.1.27/mm/
Dcma_debug.c37 struct cma *cma = data; in cma_used_get() local
40 mutex_lock(&cma->lock); in cma_used_get()
42 used = bitmap_weight(cma->bitmap, (int)cma->count); in cma_used_get()
43 mutex_unlock(&cma->lock); in cma_used_get()
44 *val = (u64)used << cma->order_per_bit; in cma_used_get()
52 struct cma *cma = data; in cma_maxchunk_get() local
56 mutex_lock(&cma->lock); in cma_maxchunk_get()
58 start = find_next_zero_bit(cma->bitmap, cma->count, end); in cma_maxchunk_get()
59 if (start >= cma->count) in cma_maxchunk_get()
61 end = find_next_bit(cma->bitmap, cma->count, start); in cma_maxchunk_get()
[all …]
Dcma.c42 struct cma cma_areas[MAX_CMA_AREAS];
46 phys_addr_t cma_get_base(const struct cma *cma) in cma_get_base() argument
48 return PFN_PHYS(cma->base_pfn); in cma_get_base()
51 unsigned long cma_get_size(const struct cma *cma) in cma_get_size() argument
53 return cma->count << PAGE_SHIFT; in cma_get_size()
56 static unsigned long cma_bitmap_aligned_mask(const struct cma *cma, in cma_bitmap_aligned_mask() argument
59 if (align_order <= cma->order_per_bit) in cma_bitmap_aligned_mask()
61 return (1UL << (align_order - cma->order_per_bit)) - 1; in cma_bitmap_aligned_mask()
68 static unsigned long cma_bitmap_aligned_offset(const struct cma *cma, in cma_bitmap_aligned_offset() argument
71 if (align_order <= cma->order_per_bit) in cma_bitmap_aligned_offset()
[all …]
Dcma.h4 struct cma { struct
16 extern struct cma cma_areas[MAX_CMA_AREAS]; argument
19 static unsigned long cma_bitmap_maxno(struct cma *cma) in cma_bitmap_maxno() argument
21 return cma->count >> cma->order_per_bit; in cma_bitmap_maxno()
DMakefile77 obj-$(CONFIG_CMA) += cma.o
/linux-4.1.27/include/linux/
Ddma-contiguous.h58 struct cma;
63 extern struct cma *dma_contiguous_default_area;
65 static inline struct cma *dev_get_cma_area(struct device *dev) in dev_get_cma_area()
72 static inline void dev_set_cma_area(struct device *dev, struct cma *cma) in dev_set_cma_area() argument
75 dev->cma_area = cma; in dev_set_cma_area()
78 static inline void dma_contiguous_set_default(struct cma *cma) in dma_contiguous_set_default() argument
80 dma_contiguous_default_area = cma; in dma_contiguous_set_default()
86 phys_addr_t limit, struct cma **res_cma,
105 struct cma *cma; in dma_declare_contiguous() local
107 ret = dma_contiguous_reserve_area(size, base, limit, &cma, true); in dma_declare_contiguous()
[all …]
Dcma.h16 struct cma;
19 extern phys_addr_t cma_get_base(const struct cma *cma);
20 extern unsigned long cma_get_size(const struct cma *cma);
25 bool fixed, struct cma **res_cma);
28 struct cma **res_cma);
29 extern struct page *cma_alloc(struct cma *cma, unsigned int count, unsigned int align);
30 extern bool cma_release(struct cma *cma, const struct page *pages, unsigned int count);
Ddevice.h768 struct cma *cma_area; /* contiguous memory area for dma
/linux-4.1.27/Documentation/cma/
Ddebugfs.txt4 Each CMA zone represents a directory under <debugfs>/cma/, indexed by the
7 <debugfs>/cma/cma-0
17 echo 5 > <debugfs>/cma/cma-2/alloc
19 would try to allocate 5 pages from the cma-2 area.
/linux-4.1.27/drivers/base/
Ddma-contiguous.c37 struct cma *dma_contiguous_default_area;
163 phys_addr_t limit, struct cma **res_cma, in dma_contiguous_reserve_area()
248 struct cma *cma; in rmem_cma_setup() local
260 err = cma_init_reserved_mem(rmem->base, rmem->size, 0, &cma); in rmem_cma_setup()
269 dma_contiguous_set_default(cma); in rmem_cma_setup()
272 rmem->priv = cma; in rmem_cma_setup()
279 RESERVEDMEM_OF_DECLARE(cma, "shared-dma-pool", rmem_cma_setup);
DKconfig258 You can disable CMA by specifying "cma=0" on the kernel's command
275 default, but it can be enabled by passing cma=size[MG] to the kernel.
287 enabled by passing cma=size[MG] to the kernel.
/linux-4.1.27/drivers/isdn/hardware/eicon/
Ddi.c570 byte *cma = NULL; in isdn_ind() local
609 (int)a->rx_stream[this->Id], (void **)&cma, &d); in isdn_ind()
611 cma = &a->stream_buffer[0]; in isdn_ind()
612 cma[0] = cma[1] = cma[2] = cma[3] = 0; in isdn_ind()
614 this->RLength = MLength = (word)*(dword *)cma; in isdn_ind()
615 cma += 4; in isdn_ind()
618 cma = &a->stream_buffer[0]; in isdn_ind()
621 cma, in isdn_ind()
631 memcpy(IoAdapter->RBuffer.P, cma, IoAdapter->RBuffer.length); in isdn_ind()
635 if (!cma) { in isdn_ind()
[all …]
/linux-4.1.27/Documentation/devicetree/bindings/reserved-memory/
Dreserved-memory.txt68 - If a "linux,cma-default" property is present, then Linux will use the
81 one default of all device drivers (named linux,cma@72000000 and 64MiB in size),
99 linux,cma {
104 linux,cma-default;
/linux-4.1.27/include/trace/events/
Dcma.h2 #define TRACE_SYSTEM cma
/linux-4.1.27/drivers/infiniband/core/
DMakefile24 rdma_cm-y := cma.o
/linux-4.1.27/fs/dlm/
Dconfig.c713 struct comm_attribute *cma = in show_comm() local
715 return cma->show ? cma->show(cm, buf) : 0; in show_comm()
722 struct comm_attribute *cma = in store_comm() local
724 return cma->store ? cma->store(cm, buf, len) : -EINVAL; in store_comm()
/linux-4.1.27/arch/arm/mach-shmobile/
Dsetup-rcar-gen2.c187 struct cma *rcar_gen2_dma_contiguous;
/linux-4.1.27/arch/powerpc/kvm/
Dbook3s_hv_builtin.c41 static struct cma *kvm_cma;
/linux-4.1.27/kernel/rcu/
Dtree.c3273 bool cma = false; in synchronize_sched_expedited() local
3310 cma = zalloc_cpumask_var(&cm, GFP_KERNEL); in synchronize_sched_expedited()
3311 if (cma) { in synchronize_sched_expedited()
3328 while (try_stop_cpus(cma ? cm : cpu_online_mask, in synchronize_sched_expedited()
/linux-4.1.27/Documentation/arm/OMAP/
DDSS290 allocator, and if CMA is enabled, you use "cma=" kernel parameter to increase
/linux-4.1.27/Documentation/
Dkernel-parameters.txt659 cma=nn[MG]@[start[MG][-end[MG]]]