Home
last modified time | relevance | path

Searched refs:order (Results 1 – 200 of 1179) sorted by relevance

123456

/linux-4.4.14/drivers/media/platform/soc_camera/
Dsoc_mediabus.c26 .order = SOC_MBUS_ORDER_LE,
36 .order = SOC_MBUS_ORDER_LE,
46 .order = SOC_MBUS_ORDER_LE,
56 .order = SOC_MBUS_ORDER_LE,
66 .order = SOC_MBUS_ORDER_LE,
76 .order = SOC_MBUS_ORDER_BE,
86 .order = SOC_MBUS_ORDER_LE,
96 .order = SOC_MBUS_ORDER_BE,
106 .order = SOC_MBUS_ORDER_LE,
115 .order = SOC_MBUS_ORDER_LE,
[all …]
/linux-4.4.14/include/trace/events/
Dvmscan.h58 TP_PROTO(int nid, int order),
60 TP_ARGS(nid, order),
64 __field( int, order )
69 __entry->order = order;
72 TP_printk("nid=%d order=%d", __entry->nid, __entry->order)
77 TP_PROTO(int nid, int zid, int order),
79 TP_ARGS(nid, zid, order),
84 __field( int, order )
90 __entry->order = order;
96 __entry->order)
[all …]
Dcompaction.h226 int order,
230 TP_ARGS(order, gfp_mask, mode),
233 __field(int, order)
239 __entry->order = order;
245 __entry->order,
253 int order,
256 TP_ARGS(zone, order, ret),
261 __field(int, order)
268 __entry->order = order;
275 __entry->order,
[all …]
Dkmem.h163 TP_PROTO(struct page *page, unsigned int order),
165 TP_ARGS(page, order),
181 __field( unsigned int, order )
186 __entry->order = order;
192 __entry->order)
219 TP_PROTO(struct page *page, unsigned int order,
222 TP_ARGS(page, order, gfp_flags, migratetype),
226 __field( unsigned int, order )
233 __entry->order = order;
241 __entry->order,
[all …]
/linux-4.4.14/arch/s390/mm/
Dpage-states.c54 static inline void set_page_unstable(struct page *page, int order) in set_page_unstable() argument
58 for (i = 0; i < (1 << order); i++) in set_page_unstable()
65 void arch_free_page(struct page *page, int order) in arch_free_page() argument
69 set_page_unstable(page, order); in arch_free_page()
72 static inline void set_page_stable(struct page *page, int order) in set_page_stable() argument
76 for (i = 0; i < (1 << order); i++) in set_page_stable()
83 void arch_alloc_page(struct page *page, int order) in arch_alloc_page() argument
87 set_page_stable(page, order); in arch_alloc_page()
92 unsigned long flags, order, t; in arch_set_page_states() local
103 for_each_migratetype_order(order, t) { in arch_set_page_states()
[all …]
Dinit.c51 unsigned int order; in setup_zero_pages() local
56 order = 7; in setup_zero_pages()
59 while (order > 2 && (totalram_pages >> 10) < (1UL << order)) in setup_zero_pages()
60 order--; in setup_zero_pages()
62 empty_zero_page = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order); in setup_zero_pages()
67 split_page(page, order); in setup_zero_pages()
68 for (i = 1 << order; i > 0; i--) { in setup_zero_pages()
73 zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; in setup_zero_pages()
/linux-4.4.14/include/linux/
Dcompaction.h40 extern int fragmentation_index(struct zone *zone, unsigned int order);
41 extern unsigned long try_to_compact_pages(gfp_t gfp_mask, unsigned int order,
44 extern void compact_pgdat(pg_data_t *pgdat, int order);
46 extern unsigned long compaction_suitable(struct zone *zone, int order,
49 extern void defer_compaction(struct zone *zone, int order);
50 extern bool compaction_deferred(struct zone *zone, int order);
51 extern void compaction_defer_reset(struct zone *zone, int order,
53 extern bool compaction_restarting(struct zone *zone, int order);
57 unsigned int order, int alloc_flags, in try_to_compact_pages() argument
64 static inline void compact_pgdat(pg_data_t *pgdat, int order) in compact_pgdat() argument
[all …]
Dgfp.h401 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument
404 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument
408 __alloc_pages_nodemask(gfp_t gfp_mask, unsigned int order,
412 __alloc_pages(gfp_t gfp_mask, unsigned int order, in __alloc_pages() argument
415 return __alloc_pages_nodemask(gfp_mask, order, zonelist, NULL); in __alloc_pages()
423 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument
428 return __alloc_pages(gfp_mask, order, node_zonelist(nid, gfp_mask)); in __alloc_pages_node()
437 unsigned int order) in alloc_pages_node() argument
442 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node()
446 extern struct page *alloc_pages_current(gfp_t gfp_mask, unsigned order);
[all …]
Dpage_owner.h8 extern void __reset_page_owner(struct page *page, unsigned int order);
10 unsigned int order, gfp_t gfp_mask);
13 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument
18 __reset_page_owner(page, order); in reset_page_owner()
22 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument
27 __set_page_owner(page, order, gfp_mask); in set_page_owner()
38 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument
42 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument
Dkmemcheck.h11 void kmemcheck_alloc_shadow(struct page *page, int order, gfp_t flags, int node);
12 void kmemcheck_free_shadow(struct page *page, int order);
17 void kmemcheck_pagealloc_alloc(struct page *p, unsigned int order,
93 kmemcheck_alloc_shadow(struct page *page, int order, gfp_t flags, int node) in kmemcheck_alloc_shadow() argument
98 kmemcheck_free_shadow(struct page *page, int order) in kmemcheck_free_shadow() argument
114 unsigned int order, gfp_t gfpflags) in kmemcheck_pagealloc_alloc() argument
Dbitops.h62 int order; in get_bitmask_order() local
64 order = fls(count); in get_bitmask_order()
65 return order; /* We could be slightly more clever with -1 here... */ in get_bitmask_order()
70 int order; in get_count_order() local
72 order = fls(count) - 1; in get_count_order()
74 order++; in get_count_order()
75 return order; in get_count_order()
Dkasan.h46 void kasan_alloc_pages(struct page *page, unsigned int order);
47 void kasan_free_pages(struct page *page, unsigned int order);
72 static inline void kasan_alloc_pages(struct page *page, unsigned int order) {} in kasan_alloc_pages() argument
73 static inline void kasan_free_pages(struct page *page, unsigned int order) {} in kasan_free_pages() argument
Dmemcontrol.h465 unsigned long mem_cgroup_soft_limit_reclaim(struct zone *zone, int order,
657 unsigned long mem_cgroup_soft_limit_reclaim(struct zone *zone, int order, in mem_cgroup_soft_limit_reclaim() argument
754 int __memcg_kmem_charge_memcg(struct page *page, gfp_t gfp, int order,
756 int __memcg_kmem_charge(struct page *page, gfp_t gfp, int order);
757 void __memcg_kmem_uncharge(struct page *page, int order);
792 gfp_t gfp, int order) in memcg_kmem_charge() argument
796 return __memcg_kmem_charge(page, gfp, order); in memcg_kmem_charge()
804 static __always_inline void memcg_kmem_uncharge(struct page *page, int order) in memcg_kmem_uncharge() argument
807 __memcg_kmem_uncharge(page, order); in memcg_kmem_uncharge()
844 static inline int memcg_kmem_charge(struct page *page, gfp_t gfp, int order) in memcg_kmem_charge() argument
[all …]
Dslab.h375 extern void *kmalloc_order(size_t size, gfp_t flags, unsigned int order) __assume_page_alignment;
378 extern void *kmalloc_order_trace(size_t size, gfp_t flags, unsigned int order) __assume_page_alignm…
381 kmalloc_order_trace(size_t size, gfp_t flags, unsigned int order) in kmalloc_order_trace() argument
383 return kmalloc_order(size, flags, order); in kmalloc_order_trace()
389 unsigned int order = get_order(size); in kmalloc_large() local
390 return kmalloc_order_trace(size, flags, order); in kmalloc_large()
Dhugetlb.h305 unsigned int order; member
344 void __init hugetlb_add_hstate(unsigned order);
383 return (unsigned long)PAGE_SIZE << h->order; in huge_page_size()
397 return h->order; in huge_page_order()
402 return h->order + PAGE_SHIFT; in huge_page_shift()
412 return 1 << h->order; in pages_per_huge_page()
438 return hstates[index].order + PAGE_SHIFT; in hstate_index_to_shift()
Dmempool.h69 static inline mempool_t *mempool_create_page_pool(int min_nr, int order) in mempool_create_page_pool() argument
72 (void *)(long)order); in mempool_create_page_pool()
Ddma-contiguous.h115 unsigned int order);
148 unsigned int order) in dma_alloc_from_contiguous() argument
Dmmzone.h72 #define for_each_migratetype_order(order, type) \ argument
73 for (order = 0; order < MAX_ORDER; order++) \
739 void wakeup_kswapd(struct zone *zone, int order, enum zone_type classzone_idx);
740 bool zone_watermark_ok(struct zone *z, unsigned int order,
742 bool zone_watermark_ok_safe(struct zone *z, unsigned int order,
Dbitmap.h163 extern int bitmap_find_free_region(unsigned long *bitmap, unsigned int bits, int order);
164 extern void bitmap_release_region(unsigned long *bitmap, unsigned int pos, int order);
165 extern int bitmap_allocate_region(unsigned long *bitmap, unsigned int pos, int order);
Di2c-pnx.h25 int order; /* RX Bytes to order via TX */ member
Dpage_ext.h45 unsigned int order; member
/linux-4.4.14/drivers/media/pci/cx18/
Dcx18-mailbox.c244 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument
253 mb = &order->mb; in epu_dma_done()
260 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done()
266 mdl_ack = order->mdl_ack; in epu_dma_done()
290 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done()
338 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument
341 char *str = order->str; in epu_debug()
343 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug()
349 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument
351 switch (order->rpu) { in epu_cmd()
[all …]
/linux-4.4.14/Documentation/trace/postprocess/
Dtrace-vmscan-postprocess.pl316 my $order = $1;
317 $perprocesspid{$process_pid}->{MM_VMSCAN_DIRECT_RECLAIM_BEGIN_PERORDER}[$order]++;
318 $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER} = $order;
327 my $order = $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER};
329 $perprocesspid{$process_pid}->{HIGH_DIRECT_RECLAIM_LATENCY}[$index] = "$order-$latency";
340 my $order = $2;
341 $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER} = $order;
346 $perprocesspid{$process_pid}->{MM_VMSCAN_KSWAPD_WAKE_PERORDER}[$order]++;
349 $perprocesspid{$process_pid}->{HIGH_KSWAPD_REWAKEUP_PERORDER}[$order]++;
359 my $order = $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER};
[all …]
/linux-4.4.14/mm/
Dpage_alloc.c187 static void __free_pages_ok(struct page *page, unsigned int order);
465 void prep_compound_page(struct page *page, unsigned int order) in prep_compound_page() argument
468 int nr_pages = 1 << order; in prep_compound_page()
471 set_compound_order(page, order); in prep_compound_page()
534 unsigned int order, int migratetype) in set_page_guard() argument
545 set_page_private(page, order); in set_page_guard()
547 __mod_zone_freepage_state(zone, -(1 << order), migratetype); in set_page_guard()
551 unsigned int order, int migratetype) in clear_page_guard() argument
563 __mod_zone_freepage_state(zone, (1 << order), migratetype); in clear_page_guard()
568 unsigned int order, int migratetype) {} in set_page_guard() argument
[all …]
Dcompaction.c127 void defer_compaction(struct zone *zone, int order) in defer_compaction() argument
132 if (order < zone->compact_order_failed) in defer_compaction()
133 zone->compact_order_failed = order; in defer_compaction()
138 trace_mm_compaction_defer_compaction(zone, order); in defer_compaction()
142 bool compaction_deferred(struct zone *zone, int order) in compaction_deferred() argument
146 if (order < zone->compact_order_failed) in compaction_deferred()
156 trace_mm_compaction_deferred(zone, order); in compaction_deferred()
166 void compaction_defer_reset(struct zone *zone, int order, in compaction_defer_reset() argument
173 if (order >= zone->compact_order_failed) in compaction_defer_reset()
174 zone->compact_order_failed = order + 1; in compaction_defer_reset()
[all …]
Dkmemcheck.c8 void kmemcheck_alloc_shadow(struct page *page, int order, gfp_t flags, int node) in kmemcheck_alloc_shadow() argument
14 pages = 1 << order; in kmemcheck_alloc_shadow()
20 shadow = alloc_pages_node(node, flags | __GFP_NOTRACK, order); in kmemcheck_alloc_shadow()
39 void kmemcheck_free_shadow(struct page *page, int order) in kmemcheck_free_shadow() argument
48 pages = 1 << order; in kmemcheck_free_shadow()
57 __free_pages(shadow, order); in kmemcheck_free_shadow()
100 void kmemcheck_pagealloc_alloc(struct page *page, unsigned int order, in kmemcheck_pagealloc_alloc() argument
108 pages = 1 << order; in kmemcheck_pagealloc_alloc()
117 kmemcheck_alloc_shadow(page, order, gfpflags, -1); in kmemcheck_pagealloc_alloc()
Dvmstat.c638 unsigned int order; in fill_contig_page_info() local
644 for (order = 0; order < MAX_ORDER; order++) { in fill_contig_page_info()
648 blocks = zone->free_area[order].nr_free; in fill_contig_page_info()
652 info->free_pages += blocks << order; in fill_contig_page_info()
655 if (order >= suitable_order) in fill_contig_page_info()
657 (order - suitable_order); in fill_contig_page_info()
668 static int __fragmentation_index(unsigned int order, struct contig_page_info *info) in __fragmentation_index() argument
670 unsigned long requested = 1UL << order; in __fragmentation_index()
689 int fragmentation_index(struct zone *zone, unsigned int order) in fragmentation_index() argument
693 fill_contig_page_info(zone, order, &info); in fragmentation_index()
[all …]
Dvmscan.c69 int order; member
1394 trace_mm_vmscan_lru_isolate(sc->order, nr_to_scan, scan, in isolate_lru_pages()
2302 if (IS_ENABLED(CONFIG_COMPACTION) && sc->order && in in_reclaim_compaction()
2303 (sc->order > PAGE_ALLOC_COSTLY_ORDER || in in_reclaim_compaction()
2356 pages_for_compaction = (2UL << sc->order); in should_continue_reclaim()
2365 switch (compaction_suitable(zone, sc->order, 0, 0)) { in should_continue_reclaim()
2466 static inline bool compaction_ready(struct zone *zone, int order) in compaction_ready() argument
2479 watermark = high_wmark_pages(zone) + balance_gap + (2UL << order); in compaction_ready()
2486 if (compaction_deferred(zone, order)) in compaction_ready()
2493 if (compaction_suitable(zone, order, 0, 0) == COMPACT_SKIPPED) in compaction_ready()
[all …]
Dpage_isolation.c81 unsigned int order; in unset_migratetype_isolate() local
99 order = page_order(page); in unset_migratetype_isolate()
100 if (order >= pageblock_order) { in unset_migratetype_isolate()
102 buddy_idx = __find_buddy_index(page_idx, order); in unset_migratetype_isolate()
107 __isolate_free_page(page, order); in unset_migratetype_isolate()
108 kernel_map_pages(page, (1 << order), 1); in unset_migratetype_isolate()
129 __free_pages(isolated_page, order); in unset_migratetype_isolate()
Dinternal.h172 __find_buddy_index(unsigned long page_idx, unsigned int order) in __find_buddy_index() argument
174 return page_idx ^ (1 << order); in __find_buddy_index()
177 extern int __isolate_free_page(struct page *page, unsigned int order);
179 unsigned int order);
180 extern void prep_compound_page(struct page *page, unsigned int order);
208 int order; /* order a direct compactor needs */ member
225 int find_suitable_fallback(struct free_area *area, unsigned int order,
Dpage_owner.c49 void __reset_page_owner(struct page *page, unsigned int order) in __reset_page_owner() argument
54 for (i = 0; i < (1 << order); i++) { in __reset_page_owner()
60 void __set_page_owner(struct page *page, unsigned int order, gfp_t gfp_mask) in __set_page_owner() argument
72 page_ext->order = order; in __set_page_owner()
104 page_ext->order, page_ext->gfp_mask); in print_page_owner()
Dnobootmem.c96 int order; in __free_pages_memory() local
99 order = min(MAX_ORDER - 1UL, __ffs(start)); in __free_pages_memory()
101 while (start + (1UL << order) > end) in __free_pages_memory()
102 order--; in __free_pages_memory()
104 __free_pages_bootmem(pfn_to_page(start), start, order); in __free_pages_memory()
106 start += (1UL << order); in __free_pages_memory()
Dmempool.c65 int order = (int)(long)pool->pool_data; in check_element() local
68 __check_element(pool, addr, 1UL << (PAGE_SHIFT + order)); in check_element()
89 int order = (int)(long)pool->pool_data; in poison_element() local
92 __poison_element(addr, 1UL << (PAGE_SHIFT + order)); in poison_element()
487 int order = (int)(long)pool_data; in mempool_alloc_pages() local
488 return alloc_pages(gfp_mask, order); in mempool_alloc_pages()
494 int order = (int)(long)pool_data; in mempool_free_pages() local
495 __free_pages(element, order); in mempool_free_pages()
Dslob.c190 static void *slob_new_pages(gfp_t gfp, int order, int node) in slob_new_pages() argument
196 page = __alloc_pages_node(node, gfp, order); in slob_new_pages()
199 page = alloc_pages(gfp, order); in slob_new_pages()
207 static void slob_free_pages(void *b, int order) in slob_free_pages() argument
210 current->reclaim_state->reclaimed_slab += 1 << order; in slob_free_pages()
211 free_pages((unsigned long)b, order); in slob_free_pages()
451 unsigned int order = get_order(size); in __do_kmalloc_node() local
453 if (likely(order)) in __do_kmalloc_node()
455 ret = slob_new_pages(gfp, order, node); in __do_kmalloc_node()
458 size, PAGE_SIZE << order, gfp, node); in __do_kmalloc_node()
Dslub.c311 static inline int order_objects(int order, unsigned long size, int reserved) in order_objects() argument
313 return ((PAGE_SIZE << order) - reserved) / size; in order_objects()
316 static inline struct kmem_cache_order_objects oo_make(int order, in oo_make() argument
320 (order << OO_SHIFT) + order_objects(order, size, reserved) in oo_make()
1378 int order = oo_order(oo); in alloc_slab_page() local
1383 page = alloc_pages(flags, order); in alloc_slab_page()
1385 page = __alloc_pages_node(node, flags, order); in alloc_slab_page()
1387 if (page && memcg_charge_slab(page, flags, order, s)) { in alloc_slab_page()
1388 __free_pages(page, order); in alloc_slab_page()
1401 int idx, order; in allocate_slab() local
[all …]
Dhugetlb.c1007 unsigned int order) in destroy_compound_gigantic_page() argument
1010 int nr_pages = 1 << order; in destroy_compound_gigantic_page()
1022 static void free_gigantic_page(struct page *page, unsigned int order) in free_gigantic_page() argument
1024 free_contig_range(page_to_pfn(page), 1 << order); in free_gigantic_page()
1066 static struct page *alloc_gigantic_page(int nid, unsigned int order) in alloc_gigantic_page() argument
1068 unsigned long nr_pages = 1 << order; in alloc_gigantic_page()
1102 static void prep_compound_gigantic_page(struct page *page, unsigned int order);
1135 static inline void free_gigantic_page(struct page *page, unsigned int order) { } in free_gigantic_page() argument
1137 unsigned int order) { } in destroy_compound_gigantic_page() argument
1263 static void prep_compound_gigantic_page(struct page *page, unsigned int order) in prep_compound_gigantic_page() argument
[all …]
Dslab.h240 gfp_t gfp, int order, in memcg_charge_slab() argument
247 return __memcg_kmem_charge_memcg(page, gfp, order, in memcg_charge_slab()
285 static inline int memcg_charge_slab(struct page *page, gfp_t gfp, int order, in memcg_charge_slab() argument
Dvmalloc.c818 static void *new_vmap_block(unsigned int order, gfp_t gfp_mask) in new_vmap_block() argument
853 BUG_ON(VMAP_BBMAP_BITS <= (1UL << order)); in new_vmap_block()
854 vb->free = VMAP_BBMAP_BITS - (1UL << order); in new_vmap_block()
939 unsigned int order; in vb_alloc() local
951 order = get_order(size); in vb_alloc()
959 if (vb->free < (1UL << order)) { in vb_alloc()
966 vb->free -= 1UL << order; in vb_alloc()
982 vaddr = new_vmap_block(order, gfp_mask); in vb_alloc()
991 unsigned int order; in vb_free() local
999 order = get_order(size); in vb_free()
[all …]
/linux-4.4.14/arch/c6x/mm/
Ddma-coherent.c45 static inline u32 __alloc_dma_pages(int order) in __alloc_dma_pages() argument
51 pos = bitmap_find_free_region(dma_bitmap, dma_pages, order); in __alloc_dma_pages()
57 static void __free_dma_pages(u32 addr, int order) in __free_dma_pages() argument
62 if (addr < dma_base || (pos + (1 << order)) >= dma_pages) { in __free_dma_pages()
68 bitmap_release_region(dma_bitmap, pos, order); in __free_dma_pages()
80 int order; in dma_alloc_coherent() local
85 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); in dma_alloc_coherent()
87 paddr = __alloc_dma_pages(order); in dma_alloc_coherent()
105 int order; in dma_free_coherent() local
110 order = get_count_order(((size - 1) >> PAGE_SHIFT) + 1); in dma_free_coherent()
[all …]
/linux-4.4.14/arch/arm/lib/
Dlib1funcs.S106 .macro ARM_DIV2_ORDER divisor, order
110 clz \order, \divisor
111 rsb \order, \order, #31
117 movhs \order, #16
118 movlo \order, #0
122 addhs \order, \order, #8
126 addhs \order, \order, #4
129 addhi \order, \order, #3
130 addls \order, \order, \divisor, lsr #1
137 .macro ARM_MOD_BODY dividend, divisor, order, spare
[all …]
/linux-4.4.14/drivers/staging/android/ion/
Dion_system_heap.c34 static int order_to_index(unsigned int order) in order_to_index() argument
39 if (order == orders[i]) in order_to_index()
45 static inline unsigned int order_to_size(int order) in order_to_size() argument
47 return PAGE_SIZE << order; in order_to_size()
57 unsigned long order) in alloc_buffer_page() argument
60 struct ion_page_pool *pool = heap->pools[order_to_index(order)]; in alloc_buffer_page()
68 if (order > 4) in alloc_buffer_page()
70 page = alloc_pages(gfp_flags | __GFP_COMP, order); in alloc_buffer_page()
73 ion_pages_sync_for_device(NULL, page, PAGE_SIZE << order, in alloc_buffer_page()
83 unsigned int order = compound_order(page); in free_buffer_page() local
[all …]
Dion_page_pool.c29 struct page *page = alloc_pages(pool->gfp_mask, pool->order); in ion_page_pool_alloc_pages()
33 ion_pages_sync_for_device(NULL, page, PAGE_SIZE << pool->order, in ion_page_pool_alloc_pages()
41 __free_pages(page, pool->order); in ion_page_pool_free_pages()
99 BUG_ON(pool->order != compound_order(page)); in ion_page_pool_free()
113 return count << pool->order; in ion_page_pool_total()
144 freed += (1 << pool->order); in ion_page_pool_shrink()
150 struct ion_page_pool *ion_page_pool_create(gfp_t gfp_mask, unsigned int order) in ion_page_pool_create() argument
161 pool->order = order; in ion_page_pool_create()
163 plist_node_init(&pool->list, order); in ion_page_pool_create()
Dion_priv.h376 unsigned int order; member
380 struct ion_page_pool *ion_page_pool_create(gfp_t gfp_mask, unsigned int order);
/linux-4.4.14/arch/powerpc/platforms/cell/
Dras.c103 int order; member
111 static int __init cbe_ptcal_enable_on_node(int nid, int order) in cbe_ptcal_enable_on_node() argument
125 area->order = order; in cbe_ptcal_enable_on_node()
128 area->order); in cbe_ptcal_enable_on_node()
159 __free_pages(area->pages, area->order); in cbe_ptcal_enable_on_node()
170 int order, found_mic = 0; in cbe_ptcal_enable() local
183 order = get_order(*size); in cbe_ptcal_enable()
188 cbe_ptcal_enable_on_node(of_node_to_nid(np), order); in cbe_ptcal_enable()
203 cbe_ptcal_enable_on_node(*nid, order); in cbe_ptcal_enable()
229 1 << (area->order + PAGE_SHIFT)); in cbe_ptcal_disable()
[all …]
/linux-4.4.14/drivers/staging/rdma/ehca/
Dipz_pt_fn.c129 int order = ilog2(queue->pagesize) - 9; in alloc_small_queue_page() local
135 if (!list_empty(&pd->free[order])) in alloc_small_queue_page()
136 page = list_entry(pd->free[order].next, in alloc_small_queue_page()
149 list_add(&page->list, &pd->free[order]); in alloc_small_queue_page()
152 bit = find_first_zero_bit(page->bitmap, IPZ_SPAGE_PER_KPAGE >> order); in alloc_small_queue_page()
156 if (page->fill == IPZ_SPAGE_PER_KPAGE >> order) in alloc_small_queue_page()
157 list_move(&page->list, &pd->full[order]); in alloc_small_queue_page()
161 queue->queue_pages[0] = (void *)(page->page | (bit << (order + 9))); in alloc_small_queue_page()
163 queue->offset = bit << (order + 9); in alloc_small_queue_page()
174 int order = ilog2(queue->pagesize) - 9; in free_small_queue_page() local
[all …]
/linux-4.4.14/arch/cris/arch-v32/drivers/pci/
Ddma.c23 int order = get_order(size); in dma_alloc_coherent() local
33 ret = (void *)__get_free_pages(gfp, order); in dma_alloc_coherent()
45 int order = get_order(size); in dma_free_coherent() local
47 if (!dma_release_from_coherent(dev, order, vaddr)) in dma_free_coherent()
48 free_pages((unsigned long)vaddr, order); in dma_free_coherent()
/linux-4.4.14/arch/tile/mm/
Dhomecache.c362 void homecache_change_page_home(struct page *page, int order, int home) in homecache_change_page_home() argument
364 int i, pages = (1 << order); in homecache_change_page_home()
385 unsigned int order, int home) in homecache_alloc_pages() argument
389 page = alloc_pages(gfp_mask, order); in homecache_alloc_pages()
391 homecache_change_page_home(page, order, home); in homecache_alloc_pages()
397 unsigned int order, int home) in homecache_alloc_pages_node() argument
401 page = alloc_pages_node(nid, gfp_mask, order); in homecache_alloc_pages_node()
403 homecache_change_page_home(page, order, home); in homecache_alloc_pages_node()
407 void __homecache_free_pages(struct page *page, unsigned int order) in __homecache_free_pages() argument
410 homecache_change_page_home(page, order, PAGE_HOME_HASH); in __homecache_free_pages()
[all …]
Dpgtable.c65 unsigned long flags, order, total = 0, largest_order = -1; in show_mem() local
71 for (order = 0; order < MAX_ORDER; order++) { in show_mem()
72 int nr = zone->free_area[order].nr_free; in show_mem()
73 total += nr << order; in show_mem()
75 largest_order = order; in show_mem()
232 int order) in pgtable_alloc_one() argument
252 for (i = 1; i < order; ++i) { in pgtable_alloc_one()
265 void pgtable_free(struct mm_struct *mm, struct page *p, int order) in pgtable_free() argument
272 for (i = 1; i < order; ++i) { in pgtable_free()
279 unsigned long address, int order) in __pgtable_free_tlb() argument
[all …]
Dinit.c665 int order = __ffs(pfn); in init_free_pfn_range() local
669 if (order >= MAX_ORDER) in init_free_pfn_range()
670 order = MAX_ORDER-1; in init_free_pfn_range()
671 count = 1 << order; in init_free_pfn_range()
674 --order; in init_free_pfn_range()
686 __free_pages(page, order); in init_free_pfn_range()
/linux-4.4.14/drivers/gpu/drm/
Ddrm_hashtab.c41 int drm_ht_create(struct drm_open_hash *ht, unsigned int order) in drm_ht_create() argument
43 unsigned int size = 1 << order; in drm_ht_create()
45 ht->order = order; in drm_ht_create()
66 hashed_key = hash_long(key, ht->order); in drm_ht_verbose_list()
80 hashed_key = hash_long(key, ht->order); in drm_ht_find_key()
98 hashed_key = hash_long(key, ht->order); in drm_ht_find_key_rcu()
117 hashed_key = hash_long(key, ht->order); in drm_ht_insert_item()
201 if ((PAGE_SIZE / sizeof(*ht->table)) >> ht->order) in drm_ht_remove()
Ddrm_bufs.c607 int order; in drm_legacy_addbufs_agp() local
620 order = order_base_2(request->size); in drm_legacy_addbufs_agp()
621 size = 1 << order; in drm_legacy_addbufs_agp()
625 page_order = order - PAGE_SHIFT > 0 ? order - PAGE_SHIFT : 0; in drm_legacy_addbufs_agp()
632 DRM_DEBUG("order: %d\n", order); in drm_legacy_addbufs_agp()
639 if (order < DRM_MIN_ORDER || order > DRM_MAX_ORDER) in drm_legacy_addbufs_agp()
664 entry = &dma->bufs[order]; in drm_legacy_addbufs_agp()
693 buf->order = order; in drm_legacy_addbufs_agp()
766 int order; in drm_legacy_addbufs_pci() local
791 order = order_base_2(request->size); in drm_legacy_addbufs_pci()
[all …]
/linux-4.4.14/arch/m68k/kernel/
Ddma.c27 int i, order; in dma_alloc_coherent() local
32 order = get_order(size); in dma_alloc_coherent()
34 page = alloc_pages(flag, order); in dma_alloc_coherent()
39 map = kmalloc(sizeof(struct page *) << order, flag & ~__GFP_DMA); in dma_alloc_coherent()
41 __free_pages(page, order); in dma_alloc_coherent()
44 split_page(page, order); in dma_alloc_coherent()
46 order = 1 << order; in dma_alloc_coherent()
51 for (; i < order; i++) in dma_alloc_coherent()
/linux-4.4.14/arch/alpha/include/asm/
Dagp.h13 #define alloc_gatt_pages(order) \ argument
14 ((char *)__get_free_pages(GFP_KERNEL, (order)))
15 #define free_gatt_pages(table, order) \ argument
16 free_pages((unsigned long)(table), (order))
/linux-4.4.14/arch/parisc/include/asm/
Dagp.h15 #define alloc_gatt_pages(order) \ argument
16 ((char *)__get_free_pages(GFP_KERNEL, (order)))
17 #define free_gatt_pages(table, order) \ argument
18 free_pages((unsigned long)(table), (order))
/linux-4.4.14/arch/ia64/include/asm/
Dagp.h21 #define alloc_gatt_pages(order) \ argument
22 ((char *)__get_free_pages(GFP_KERNEL, (order)))
23 #define free_gatt_pages(table, order) \ argument
24 free_pages((unsigned long)(table), (order))
Dpage.h159 long order; in get_order() local
161 order = ia64_getf_exp(d); in get_order()
162 order = order - PAGE_SHIFT - 0xffff + 1; in get_order()
163 if (order < 0) in get_order()
164 order = 0; in get_order()
165 return order; in get_order()
/linux-4.4.14/arch/sparc/include/asm/
Dagp.h11 #define alloc_gatt_pages(order) \ argument
12 ((char *)__get_free_pages(GFP_KERNEL, (order)))
13 #define free_gatt_pages(table, order) \ argument
14 free_pages((unsigned long)(table), (order))
/linux-4.4.14/drivers/iommu/
Ddma-iommu.c91 unsigned long order, base_pfn, end_pfn; in iommu_dma_init_domain() local
97 order = __ffs(domain->ops->pgsize_bitmap); in iommu_dma_init_domain()
98 base_pfn = max_t(unsigned long, 1, base >> order); in iommu_dma_init_domain()
99 end_pfn = (base + size - 1) >> order; in iommu_dma_init_domain()
110 domain->geometry.aperture_start >> order); in iommu_dma_init_domain()
112 domain->geometry.aperture_end >> order); in iommu_dma_init_domain()
117 if (1UL << order != iovad->granule || in iommu_dma_init_domain()
125 init_iova_domain(iovad, 1UL << order, base_pfn, end_pfn); in iommu_dma_init_domain()
197 unsigned int order = MAX_ORDER; in __iommu_dma_alloc_pages() local
218 for (order = min_t(unsigned int, order, __fls(count)); in __iommu_dma_alloc_pages()
[all …]
Dintel-svm.c40 int order; in intel_svm_alloc_pasid_tables() local
42 order = ecap_pss(iommu->ecap) + 7 - PAGE_SHIFT; in intel_svm_alloc_pasid_tables()
43 if (order < 0) in intel_svm_alloc_pasid_tables()
44 order = 0; in intel_svm_alloc_pasid_tables()
46 pages = alloc_pages(GFP_KERNEL | __GFP_ZERO, order); in intel_svm_alloc_pasid_tables()
53 pr_info("%s: Allocated order %d PASID table.\n", iommu->name, order); in intel_svm_alloc_pasid_tables()
56 pages = alloc_pages(GFP_KERNEL | __GFP_ZERO, order); in intel_svm_alloc_pasid_tables()
71 int order; in intel_svm_free_pasid_tables() local
73 order = ecap_pss(iommu->ecap) + 7 - PAGE_SHIFT; in intel_svm_free_pasid_tables()
74 if (order < 0) in intel_svm_free_pasid_tables()
[all …]
/linux-4.4.14/arch/powerpc/include/asm/
Dagp.h12 #define alloc_gatt_pages(order) \ argument
13 ((char *)__get_free_pages(GFP_KERNEL, (order)))
14 #define free_gatt_pages(table, order) \ argument
15 free_pages((unsigned long)(table), (order))
Dpage_32.h40 extern void clear_pages(void *page, int order);
/linux-4.4.14/drivers/net/ethernet/mellanox/mlx4/
Dmr.c46 static u32 mlx4_buddy_alloc(struct mlx4_buddy *buddy, int order) in mlx4_buddy_alloc() argument
54 for (o = order; o <= buddy->max_order; ++o) in mlx4_buddy_alloc()
69 while (o > order) { in mlx4_buddy_alloc()
78 seg <<= order; in mlx4_buddy_alloc()
83 static void mlx4_buddy_free(struct mlx4_buddy *buddy, u32 seg, int order) in mlx4_buddy_free() argument
85 seg >>= order; in mlx4_buddy_free()
89 while (test_bit(seg ^ 1, buddy->bits[order])) { in mlx4_buddy_free()
90 clear_bit(seg ^ 1, buddy->bits[order]); in mlx4_buddy_free()
91 --buddy->num_free[order]; in mlx4_buddy_free()
93 ++order; in mlx4_buddy_free()
[all …]
Dicm.c97 static int mlx4_alloc_icm_pages(struct scatterlist *mem, int order, in mlx4_alloc_icm_pages() argument
102 page = alloc_pages_node(node, gfp_mask, order); in mlx4_alloc_icm_pages()
104 page = alloc_pages(gfp_mask, order); in mlx4_alloc_icm_pages()
109 sg_set_page(mem, page, PAGE_SIZE << order, 0); in mlx4_alloc_icm_pages()
114 int order, gfp_t gfp_mask) in mlx4_alloc_icm_coherent() argument
116 void *buf = dma_alloc_coherent(dev, PAGE_SIZE << order, in mlx4_alloc_icm_coherent()
121 sg_set_buf(mem, buf, PAGE_SIZE << order); in mlx4_alloc_icm_coherent()
123 sg_dma_len(mem) = PAGE_SIZE << order; in mlx4_alloc_icm_coherent()
Dalloc.c702 struct mlx4_db *db, int order) in mlx4_alloc_db_from_pgdir() argument
707 for (o = order; o <= 1; ++o) { in mlx4_alloc_db_from_pgdir()
720 if (o > order) in mlx4_alloc_db_from_pgdir()
721 set_bit(i ^ 1, pgdir->bits[order]); in mlx4_alloc_db_from_pgdir()
727 db->order = order; in mlx4_alloc_db_from_pgdir()
732 int mlx4_db_alloc(struct mlx4_dev *dev, struct mlx4_db *db, int order, gfp_t gfp) in mlx4_db_alloc() argument
741 if (!mlx4_alloc_db_from_pgdir(pgdir, db, order)) in mlx4_db_alloc()
753 WARN_ON(mlx4_alloc_db_from_pgdir(pgdir, db, order)); in mlx4_db_alloc()
770 o = db->order; in mlx4_db_free()
773 if (db->order == 0 && test_bit(i ^ 1, db->u.pgdir->order0)) { in mlx4_db_free()
/linux-4.4.14/arch/x86/include/asm/
Dagp.h26 #define alloc_gatt_pages(order) \ argument
27 ((char *)__get_free_pages(GFP_KERNEL, (order)))
28 #define free_gatt_pages(table, order) \ argument
29 free_pages((unsigned long)(table), (order))
Dgart.h61 static inline void gart_set_size_and_enable(struct pci_dev *dev, u32 order) in gart_set_size_and_enable() argument
69 ctl = order << 1; in gart_set_size_and_enable()
/linux-4.4.14/arch/tile/include/asm/
Dhomecache.h68 extern void homecache_change_page_home(struct page *, int order, int home);
93 unsigned int order, int home);
95 unsigned int order, int home);
105 void __homecache_free_pages(struct page *, unsigned int order);
106 void homecache_free_pages(unsigned long addr, unsigned int order);
Dpgalloc.h72 int order);
73 extern void pgtable_free(struct mm_struct *mm, struct page *pte, int order);
101 unsigned long address, int order);
Dkexec.h50 struct page *kimage_alloc_pages_arch(gfp_t gfp_mask, unsigned int order);
/linux-4.4.14/include/asm-generic/
Dgetorder.h15 int order; in __get_order() local
20 order = fls(size); in __get_order()
22 order = fls64(size); in __get_order()
24 return order; in __get_order()
Ddma-coherent.h11 int dma_release_from_coherent(struct device *dev, int order, void *vaddr);
28 #define dma_release_from_coherent(dev, order, vaddr) (0) argument
29 #define dma_mmap_from_coherent(dev, vma, vaddr, order, ret) (0) argument
/linux-4.4.14/arch/frv/mm/
Ddma-alloc.c86 int order, err, i; in consistent_alloc() local
93 order = get_order(size); in consistent_alloc()
95 page = __get_free_pages(gfp, order); in consistent_alloc()
104 free_pages(page, order); in consistent_alloc()
116 if (order > 0) { in consistent_alloc()
118 split_page(rpage, order); in consistent_alloc()
/linux-4.4.14/arch/x86/kernel/
Daperture_64.c121 static u32 __init read_agp(int bus, int slot, int func, int cap, u32 *order) in read_agp() argument
139 old_order = *order; in read_agp()
146 *order = 7 - nbits; in read_agp()
147 if ((int)*order < 0) /* < 32MB */ in read_agp()
148 *order = 0; in read_agp()
161 if (aper + (32ULL<<(20 + *order)) > 0x100000000ULL) { in read_agp()
163 bus, slot, func, 32 << *order, apsizereg); in read_agp()
164 *order = old_order; in read_agp()
168 bus, slot, func, aper, aper + (32ULL << (*order + 20)) - 1, in read_agp()
169 32 << *order, apsizereg); in read_agp()
[all …]
/linux-4.4.14/arch/microblaze/mm/
Dconsistent.c64 unsigned long order, vaddr; in consistent_alloc() local
80 order = get_order(size); in consistent_alloc()
82 vaddr = __get_free_pages(gfp, order); in consistent_alloc()
113 free_pages(vaddr, order); in consistent_alloc()
130 end = page + (1 << order); in consistent_alloc()
132 split_page(page, order); in consistent_alloc()
151 free_pages(vaddr, order); in consistent_alloc()
/linux-4.4.14/arch/sh/mm/
Dconsistent.c40 int order = get_order(size); in dma_generic_alloc_coherent() local
44 ret = (void *)__get_free_pages(gfp, order); in dma_generic_alloc_coherent()
56 free_pages((unsigned long)ret, order); in dma_generic_alloc_coherent()
60 split_page(pfn_to_page(virt_to_phys(ret) >> PAGE_SHIFT), order); in dma_generic_alloc_coherent()
71 int order = get_order(size); in dma_generic_free_coherent() local
75 for (k = 0; k < (1 << order); k++) in dma_generic_free_coherent()
/linux-4.4.14/security/integrity/ima/
Dima_crypto.c47 int order; in param_set_bufsize() local
50 order = get_order(size); in param_set_bufsize()
51 if (order >= MAX_ORDER) in param_set_bufsize()
53 ima_maxorder = order; in param_set_bufsize()
54 ima_bufsize = PAGE_SIZE << order; in param_set_bufsize()
128 int order = ima_maxorder; in ima_alloc_pages() local
131 if (order) in ima_alloc_pages()
132 order = min(get_order(max_size), order); in ima_alloc_pages()
134 for (; order; order--) { in ima_alloc_pages()
135 ptr = (void *)__get_free_pages(gfp_mask, order); in ima_alloc_pages()
[all …]
/linux-4.4.14/drivers/infiniband/hw/mlx5/
Dmem.c46 int *ncont, int *order) in mlx5_ib_cont_pages() argument
66 if (order) in mlx5_ib_cont_pages()
67 *order = ilog2(roundup_pow_of_two(*count)); in mlx5_ib_cont_pages()
107 if (order) in mlx5_ib_cont_pages()
108 *order = ilog2(roundup_pow_of_two(i) >> m); in mlx5_ib_cont_pages()
114 if (order) in mlx5_ib_cont_pages()
115 *order = 0; in mlx5_ib_cont_pages()
Dmr.c70 static int order2idx(struct mlx5_ib_dev *dev, int order) in order2idx() argument
74 if (order < cache->ent[0].order) in order2idx()
77 return order - cache->ent[0].order; in order2idx()
85 int c = order2idx(dev, mr->order); in reg_mr_callback()
140 int npages = 1 << ent->order; in add_keys()
159 mr->order = ent->order; in add_keys()
228 c = order2idx(dev, ent->order); in size_write()
294 c = order2idx(dev, ent->order); in limit_write()
359 int i = order2idx(dev, ent->order); in __cache_work_func()
423 static struct mlx5_ib_mr *alloc_cached_mr(struct mlx5_ib_dev *dev, int order) in alloc_cached_mr() argument
[all …]
/linux-4.4.14/drivers/xen/
Dswiotlb-xen.c219 unsigned long bytes, order; in xen_swiotlb_init() local
227 order = get_order(xen_io_tlb_nslabs << IO_TLB_SHIFT); in xen_swiotlb_init()
236 while ((SLABS_PER_PAGE << order) > IO_TLB_MIN_SLABS) { in xen_swiotlb_init()
237 xen_io_tlb_start = (void *)xen_get_swiotlb_free_pages(order); in xen_swiotlb_init()
240 order--; in xen_swiotlb_init()
242 if (order != get_order(bytes)) { in xen_swiotlb_init()
244 (PAGE_SIZE << order) >> 20); in xen_swiotlb_init()
245 xen_io_tlb_nslabs = SLABS_PER_PAGE << order; in xen_swiotlb_init()
264 free_pages((unsigned long)xen_io_tlb_start, order); in xen_swiotlb_init()
291 free_pages((unsigned long)xen_io_tlb_start, order); in xen_swiotlb_init()
[all …]
/linux-4.4.14/lib/
Dgenalloc.c243 int order = pool->min_alloc_order; in gen_pool_destroy() local
250 end_bit = chunk_size(chunk) >> order; in gen_pool_destroy()
275 int order = pool->min_alloc_order; in gen_pool_alloc() local
285 nbits = (size + (1UL << order) - 1) >> order; in gen_pool_alloc()
291 end_bit = chunk_size(chunk) >> order; in gen_pool_alloc()
305 addr = chunk->start_addr + ((unsigned long)start_bit << order); in gen_pool_alloc()
306 size = nbits << order; in gen_pool_alloc()
357 int order = pool->min_alloc_order; in gen_pool_free() local
364 nbits = (size + (1UL << order) - 1) >> order; in gen_pool_free()
369 start_bit = (addr - chunk->start_addr) >> order; in gen_pool_free()
[all …]
Dbitmap.c946 static int __reg_op(unsigned long *bitmap, unsigned int pos, int order, int reg_op) in __reg_op() argument
961 nbits_reg = 1 << order; in __reg_op()
1012 int bitmap_find_free_region(unsigned long *bitmap, unsigned int bits, int order) in bitmap_find_free_region() argument
1016 for (pos = 0 ; (end = pos + (1U << order)) <= bits; pos = end) { in bitmap_find_free_region()
1017 if (!__reg_op(bitmap, pos, order, REG_OP_ISFREE)) in bitmap_find_free_region()
1019 __reg_op(bitmap, pos, order, REG_OP_ALLOC); in bitmap_find_free_region()
1037 void bitmap_release_region(unsigned long *bitmap, unsigned int pos, int order) in bitmap_release_region() argument
1039 __reg_op(bitmap, pos, order, REG_OP_RELEASE); in bitmap_release_region()
1054 int bitmap_allocate_region(unsigned long *bitmap, unsigned int pos, int order) in bitmap_allocate_region() argument
1056 if (!__reg_op(bitmap, pos, order, REG_OP_ISFREE)) in bitmap_allocate_region()
[all …]
Dswiotlb.c247 unsigned int order; in swiotlb_late_init_with_default_size() local
258 order = get_order(io_tlb_nslabs << IO_TLB_SHIFT); in swiotlb_late_init_with_default_size()
259 io_tlb_nslabs = SLABS_PER_PAGE << order; in swiotlb_late_init_with_default_size()
262 while ((SLABS_PER_PAGE << order) > IO_TLB_MIN_SLABS) { in swiotlb_late_init_with_default_size()
264 order); in swiotlb_late_init_with_default_size()
267 order--; in swiotlb_late_init_with_default_size()
274 if (order != get_order(bytes)) { in swiotlb_late_init_with_default_size()
276 "for software IO TLB\n", (PAGE_SIZE << order) >> 20); in swiotlb_late_init_with_default_size()
277 io_tlb_nslabs = SLABS_PER_PAGE << order; in swiotlb_late_init_with_default_size()
281 free_pages((unsigned long)vstart, order); in swiotlb_late_init_with_default_size()
[all …]
Dpercpu_ida.c289 unsigned i, cpu, order; in __percpu_ida_init() local
305 order = get_order(nr_tags * sizeof(unsigned)); in __percpu_ida_init()
306 pool->freelist = (void *) __get_free_pages(GFP_KERNEL, order); in __percpu_ida_init()
/linux-4.4.14/arch/avr32/mm/
Ddma-coherent.c44 int order; in __dma_alloc() local
54 order = get_order(size); in __dma_alloc()
56 page = alloc_pages(gfp, order); in __dma_alloc()
59 split_page(page, order); in __dma_alloc()
74 end = page + (1 << order); in __dma_alloc()
/linux-4.4.14/drivers/atm/
Deni.c173 1 << eni_dev->free_list[i].order); in dump_mem()
205 int len,order; in eni_put_free() local
217 for (order = 0; !(((unsigned long)start | size) & (1 << order)); order++); in eni_put_free()
218 if (MID_MIN_BUF_SIZE > (1 << order)) { in eni_put_free()
220 order); in eni_put_free()
224 list[len].order = order; in eni_put_free()
226 start += 1 << order; in eni_put_free()
227 size -= 1 << order; in eni_put_free()
238 int len,i,order,best_order,index; in eni_alloc_mem() local
244 for (order = 0; (1 << order) < *size; order++); in eni_alloc_mem()
[all …]
/linux-4.4.14/fs/ramfs/
Dfile-nommu.c70 unsigned order; in ramfs_nommu_expand_for_mapping() local
76 order = get_order(newsize); in ramfs_nommu_expand_for_mapping()
77 if (unlikely(order >= MAX_ORDER)) in ramfs_nommu_expand_for_mapping()
88 pages = alloc_pages(gfp, order); in ramfs_nommu_expand_for_mapping()
93 xpages = 1UL << order; in ramfs_nommu_expand_for_mapping()
96 split_page(pages, order); in ramfs_nommu_expand_for_mapping()
/linux-4.4.14/arch/mn10300/include/asm/
Dpage.h80 int order; in get_order() local
83 order = -1; in get_order()
86 order++; in get_order()
88 return order; in get_order()
/linux-4.4.14/Documentation/devicetree/bindings/display/
Dst,stih4xx.txt19 order.
37 order.
41 order.
50 the same order.
54 order.
62 the same order.
65 the same order
69 order.
78 the same order.
83 order.
[all …]
/linux-4.4.14/drivers/infiniband/hw/mthca/
Dmthca_mr.c43 int order; member
84 static u32 mthca_buddy_alloc(struct mthca_buddy *buddy, int order) in mthca_buddy_alloc() argument
92 for (o = order; o <= buddy->max_order; ++o) in mthca_buddy_alloc()
107 while (o > order) { in mthca_buddy_alloc()
116 seg <<= order; in mthca_buddy_alloc()
121 static void mthca_buddy_free(struct mthca_buddy *buddy, u32 seg, int order) in mthca_buddy_free() argument
123 seg >>= order; in mthca_buddy_free()
127 while (test_bit(seg ^ 1, buddy->bits[order])) { in mthca_buddy_free()
128 clear_bit(seg ^ 1, buddy->bits[order]); in mthca_buddy_free()
129 --buddy->num_free[order]; in mthca_buddy_free()
[all …]
Dmthca_memfree.c107 static int mthca_alloc_icm_pages(struct scatterlist *mem, int order, gfp_t gfp_mask) in mthca_alloc_icm_pages() argument
115 page = alloc_pages(gfp_mask | __GFP_ZERO, order); in mthca_alloc_icm_pages()
119 sg_set_page(mem, page, PAGE_SIZE << order, 0); in mthca_alloc_icm_pages()
124 int order, gfp_t gfp_mask) in mthca_alloc_icm_coherent() argument
126 void *buf = dma_alloc_coherent(dev, PAGE_SIZE << order, &sg_dma_address(mem), in mthca_alloc_icm_coherent()
131 sg_set_buf(mem, buf, PAGE_SIZE << order); in mthca_alloc_icm_coherent()
133 sg_dma_len(mem) = PAGE_SIZE << order; in mthca_alloc_icm_coherent()
/linux-4.4.14/arch/sparc/kernel/
Dpci_sun4v.c135 unsigned long flags, order, first_page, npages, n; in dma_4v_alloc_coherent() local
143 order = get_order(size); in dma_4v_alloc_coherent()
144 if (unlikely(order >= MAX_ORDER)) in dma_4v_alloc_coherent()
150 page = alloc_pages_node(nid, gfp, order); in dma_4v_alloc_coherent()
155 memset((char *)first_page, 0, PAGE_SIZE << order); in dma_4v_alloc_coherent()
193 free_pages(first_page, order); in dma_4v_alloc_coherent()
220 unsigned long order, npages, entry; in dma_4v_free_coherent() local
230 order = get_order(size); in dma_4v_free_coherent()
231 if (order < 10) in dma_4v_free_coherent()
232 free_pages((unsigned long)cpu, order); in dma_4v_free_coherent()
[all …]
Diommu.c96 unsigned long i, order, sz, num_tsb_entries; in iommu_table_init() local
132 order = get_order(tsbsize); in iommu_table_init()
133 page = alloc_pages_node(numa_node, GFP_KERNEL, order); in iommu_table_init()
201 unsigned long order, first_page; in dma_4u_alloc_coherent() local
209 order = get_order(size); in dma_4u_alloc_coherent()
210 if (order >= 10) in dma_4u_alloc_coherent()
214 page = alloc_pages_node(nid, gfp, order); in dma_4u_alloc_coherent()
219 memset((char *)first_page, 0, PAGE_SIZE << order); in dma_4u_alloc_coherent()
226 free_pages(first_page, order); in dma_4u_alloc_coherent()
251 unsigned long order, npages; in dma_4u_free_coherent() local
[all …]
Dpci_fire.c231 unsigned long pages, order, i; in pci_fire_msiq_alloc() local
233 order = get_order(512 * 1024); in pci_fire_msiq_alloc()
234 pages = __get_free_pages(GFP_KERNEL | __GFP_COMP, order); in pci_fire_msiq_alloc()
237 order); in pci_fire_msiq_alloc()
240 memset((char *)pages, 0, PAGE_SIZE << order); in pci_fire_msiq_alloc()
263 unsigned long pages, order; in pci_fire_msiq_free() local
265 order = get_order(512 * 1024); in pci_fire_msiq_free()
268 free_pages(pages, order); in pci_fire_msiq_free()
Dioport.c269 int order; in sbus_alloc_coherent() local
280 order = get_order(len_total); in sbus_alloc_coherent()
281 va = __get_free_pages(gfp, order); in sbus_alloc_coherent()
312 free_pages(va, order); in sbus_alloc_coherent()
437 int order; in pci32_alloc_coherent() local
446 order = get_order(len_total); in pci32_alloc_coherent()
447 va = (void *) __get_free_pages(gfp, order); in pci32_alloc_coherent()
471 free_pages((unsigned long)va, order); in pci32_alloc_coherent()
Dldc.c972 unsigned long size, order; in alloc_queue() local
976 order = get_order(size); in alloc_queue()
978 q = (void *) __get_free_pages(GFP_KERNEL, order); in alloc_queue()
981 "size=%lu order=%lu\n", name, size, order); in alloc_queue()
985 memset(q, 0, PAGE_SIZE << order); in alloc_queue()
995 unsigned long size, order; in free_queue() local
1001 order = get_order(size); in free_queue()
1003 free_pages((unsigned long)q, order); in free_queue()
1037 unsigned long sz, num_tsb_entries, tsbsize, order; in ldc_iommu_init() local
1060 order = get_order(tsbsize); in ldc_iommu_init()
[all …]
/linux-4.4.14/Documentation/devicetree/bindings/arm/msm/
Dtimer.txt13 optionally as well, 2 watchdog interrupts, in that order.
18 must appear in the same order as the clock names.
21 the same order as the clocks.
24 timer(s) in Hz in that order.
/linux-4.4.14/drivers/net/ethernet/ti/
Dnetcp.h211 typedef int netcp_hook_rtn(int order, void *data, struct netcp_packet *packet);
212 int netcp_register_txhook(struct netcp_intf *netcp_priv, int order,
214 int netcp_unregister_txhook(struct netcp_intf *netcp_priv, int order,
216 int netcp_register_rxhook(struct netcp_intf *netcp_priv, int order,
218 int netcp_unregister_rxhook(struct netcp_intf *netcp_priv, int order,
Dnetcp_core.c458 int order; member
461 int netcp_register_txhook(struct netcp_intf *netcp_priv, int order, in netcp_register_txhook() argument
474 entry->order = order; in netcp_register_txhook()
478 if (next->order > order) in netcp_register_txhook()
488 int netcp_unregister_txhook(struct netcp_intf *netcp_priv, int order, in netcp_unregister_txhook() argument
496 if ((next->order == order) && in netcp_unregister_txhook()
510 int netcp_register_rxhook(struct netcp_intf *netcp_priv, int order, in netcp_register_rxhook() argument
523 entry->order = order; in netcp_register_rxhook()
527 if (next->order > order) in netcp_register_rxhook()
536 int netcp_unregister_rxhook(struct netcp_intf *netcp_priv, int order, in netcp_unregister_rxhook() argument
[all …]
/linux-4.4.14/Documentation/vm/
Dzsmalloc.txt6 never attempts higher order page allocation which is very likely to
8 (0-order) pages, it would suffer from very high fragmentation --
12 To overcome these issues, zsmalloc allocates a bunch of 0-order pages
14 pages act as a single higher-order page i.e. an object can span 0-order
59 pages_per_zspage: the number of 0-order pages to make a zspage
Dslub.txt7 SLUB can enable debugging only for selected slabs in order to avoid
11 In order to switch debugging on one can add a option "slub_debug"
50 F.e. in order to boot just with sanity checks and red zoning one would specify:
65 Debugging options may require the minimum possible slab order to increase as
92 in order to reduce overhead and increase cache hotness of objects.
99 order to do so you must have the slabinfo tool. Then you can do
115 governed by the order of the allocation for each slab. The allocations
123 into one slab in order for the allocation order to be acceptable.
128 slub_min_order specifies a minim order of slabs. A similar effect like
131 slub_max_order specified the order at which slub_min_objects should no
[all …]
Didle_page_tracking.txt18 mapped to bit #i%64 of array element #i/64, byte order is native. When a bit is
33 /proc/kpageflags in order to correctly count idle huge pages.
40 That said, in order to estimate the amount of pages that are not used by a
60 The kernel internally keeps track of accesses to user memory pages in order to
97 result noticeably. In order not to stall scanning of the idle page bitmap,
/linux-4.4.14/drivers/char/agp/
Damd64-agp.c273 int order = 0; in fix_northbridge() local
295 order = 7 - hweight16(apsize); in fix_northbridge()
303 if (order >=0 && aper + (32ULL<<(20 + order)) > 0x100000000ULL) { in fix_northbridge()
305 32 << order); in fix_northbridge()
306 order = nb_order; in fix_northbridge()
309 if (nb_order >= order) { in fix_northbridge()
315 aper, 32 << order); in fix_northbridge()
316 if (order < 0 || !agp_aperture_valid(aper, (32*1024*1024)<<order)) in fix_northbridge()
319 gart_set_size_and_enable(nb, order); in fix_northbridge()
/linux-4.4.14/Documentation/filesystems/
Ddirectory-locking6 always acquire the locks in order by increasing address. We'll call
7 that "inode pointer" order in the following.
21 lock both, lock them in inode pointer order.
32 * lock parents in "ancestors first" order.
40 do so in inode pointer order.
61 (2) if cross-directory rename holds the lock on filesystem, order will not
64 the order until we had acquired all locks).
67 directory objects, and are acquired in inode pointer order.
70 target in inode pointer order in the case they are not directories.)
100 means that cross-directory rename is taking locks out of order. Due
[all …]
Dgfs2-glocks.txt13 workqueue) when it releases it in order to ensure any pending tasks
19 of the list. Locks are granted in strictly the order that they
69 grant for which we ignore remote demote requests. This is in order to
78 if possible, in order to try and speed up the fast path though the locking.
101 Glock locking order within GFS2:
107 lock number order)
113 itself (locking order as above), and the other, known as the iopen
124 super block stats are done on a per cpu basis in order to
/linux-4.4.14/Documentation/trace/
Devents-kmem.txt41 mm_page_alloc page=%p pfn=%lu order=%d migratetype=%d gfp_flags=%s
42 mm_page_alloc_zone_locked page=%p pfn=%lu order=%u migratetype=%d cpu=%d percpu_refill=%d
43 mm_page_free page=%p pfn=%lu order=%d
44 mm_page_free_batched page=%p pfn=%lu order=%d cold=%d
68 mm_page_alloc_zone_locked page=%p pfn=%lu order=%u migratetype=%d cpu=%d percpu_refill=%d
69 mm_page_pcpu_drain page=%p pfn=%lu order=%d cpu=%d migratetype=%d
72 for order-0 pages, reduces contention on the zone->lock and reduces the
97 External fragmentation affects whether a high-order allocation will be
104 high-order allocations will start failing at some time in the future. One
/linux-4.4.14/arch/powerpc/mm/
Ddma-noncoherent.c159 unsigned long order; in __dma_alloc_coherent() local
192 order = get_order(size); in __dma_alloc_coherent()
198 page = alloc_pages(gfp, order); in __dma_alloc_coherent()
219 struct page *end = page + (1 << order); in __dma_alloc_coherent()
221 split_page(page, order); in __dma_alloc_coherent()
248 __free_pages(page, order); in __dma_alloc_coherent()
/linux-4.4.14/kernel/events/
Dring_buffer.c435 static struct page *rb_alloc_aux_page(int node, int order) in rb_alloc_aux_page() argument
439 if (order > MAX_ORDER) in rb_alloc_aux_page()
440 order = MAX_ORDER; in rb_alloc_aux_page()
443 page = alloc_pages_node(node, PERF_AUX_GFP, order); in rb_alloc_aux_page()
444 } while (!page && order--); in rb_alloc_aux_page()
446 if (page && order) { in rb_alloc_aux_page()
453 split_page(page, order); in rb_alloc_aux_page()
455 set_page_private(page, order); in rb_alloc_aux_page()
507 int last, order; in rb_alloc_aux() local
509 order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages)); in rb_alloc_aux()
[all …]
/linux-4.4.14/tools/perf/
Dbuiltin-kmem.c263 int order; member
448 data->order = pstat->order; in __page_stat__findnew_page()
503 data->order = pstat->order; in __page_stat__findnew_alloc()
558 data->order = pstat->order; in __page_stat__findnew_caller()
762 unsigned int order = perf_evsel__intval(evsel, sample, "order"); in perf_evsel__process_page_alloc_event() local
766 u64 bytes = kmem_page_size << order; in perf_evsel__process_page_alloc_event()
770 .order = order, in perf_evsel__process_page_alloc_event()
826 order_stats[order][migrate_type]++; in perf_evsel__process_page_alloc_event()
835 unsigned int order = perf_evsel__intval(evsel, sample, "order"); in perf_evsel__process_page_free_event() local
836 u64 bytes = kmem_page_size << order; in perf_evsel__process_page_free_event()
[all …]
/linux-4.4.14/Documentation/devicetree/bindings/sound/
Dst,sti-asoc-card.txt18 - clock-names: name of the clocks listed in clocks property in the same order
21 order than the CPU_DAI properties.
24 the same order.
26 - interrupts: CPU_DAI interrupt line, listed in the same order than the
30 order than the CPU_DAI properties.
41 - clocks: CPU_DAI IP clock source, listed in the same order than the
/linux-4.4.14/kernel/
Dkexec_core.c282 static struct page *kimage_alloc_pages(gfp_t gfp_mask, unsigned int order) in kimage_alloc_pages() argument
286 pages = alloc_pages(gfp_mask, order); in kimage_alloc_pages()
291 set_page_private(pages, order); in kimage_alloc_pages()
292 count = 1 << order; in kimage_alloc_pages()
302 unsigned int order, count, i; in kimage_free_pages() local
304 order = page_private(page); in kimage_free_pages()
305 count = 1 << order; in kimage_free_pages()
308 __free_pages(page, order); in kimage_free_pages()
325 unsigned int order) in kimage_alloc_normal_control_pages() argument
344 count = 1 << order; in kimage_alloc_normal_control_pages()
[all …]
/linux-4.4.14/arch/arm/xen/
Dmm.c25 unsigned long xen_get_swiotlb_free_pages(unsigned int order) in xen_get_swiotlb_free_pages() argument
36 return __get_free_pages(flags, order); in xen_get_swiotlb_free_pages()
166 int xen_create_contiguous_region(phys_addr_t pstart, unsigned int order, in xen_create_contiguous_region() argument
179 void xen_destroy_contiguous_region(phys_addr_t pstart, unsigned int order) in xen_destroy_contiguous_region() argument
/linux-4.4.14/drivers/media/pci/solo6x10/
Dsolo6x10-p2m.c222 int order = get_order(size); in solo_p2m_test() local
224 wr_buf = (u32 *)__get_free_pages(GFP_KERNEL, order); in solo_p2m_test()
228 rd_buf = (u32 *)__get_free_pages(GFP_KERNEL, order); in solo_p2m_test()
230 free_pages((unsigned long)wr_buf, order); in solo_p2m_test()
256 free_pages((unsigned long)wr_buf, order); in solo_p2m_test()
257 free_pages((unsigned long)rd_buf, order); in solo_p2m_test()
/linux-4.4.14/arch/metag/kernel/
Ddma.c179 unsigned long order; in dma_alloc_coherent() local
200 order = get_order(size); in dma_alloc_coherent()
205 page = alloc_pages(gfp, order); in dma_alloc_coherent()
227 struct page *end = page + (1 << order); in dma_alloc_coherent()
230 split_page(page, order); in dma_alloc_coherent()
262 __free_pages(page, order); in dma_alloc_coherent()
/linux-4.4.14/drivers/media/platform/omap/
Domap_voutlib.c303 u32 order, size; in omap_vout_alloc_buffer() local
307 order = get_order(size); in omap_vout_alloc_buffer()
308 virt_addr = __get_free_pages(GFP_KERNEL, order); in omap_vout_alloc_buffer()
327 u32 order, size; in omap_vout_free_buffer() local
331 order = get_order(size); in omap_vout_free_buffer()
338 free_pages((unsigned long) virtaddr, order); in omap_vout_free_buffer()
/linux-4.4.14/arch/s390/include/asm/
Dsigp.h39 static inline int __pcpu_sigp(u16 addr, u8 order, unsigned long parm, in __pcpu_sigp() argument
49 : "=d" (cc), "+d" (reg1) : "d" (addr), "a" (order) : "cc"); in __pcpu_sigp()
Dpage.h128 void arch_free_page(struct page *page, int order);
129 void arch_alloc_page(struct page *page, int order);
/linux-4.4.14/drivers/net/ethernet/amd/xgbe/
Dxgbe-desc.c264 struct xgbe_page_alloc *pa, gfp_t gfp, int order) in xgbe_alloc_pages() argument
272 while (order >= 0) { in xgbe_alloc_pages()
273 pages = alloc_pages(gfp, order); in xgbe_alloc_pages()
277 order--; in xgbe_alloc_pages()
284 PAGE_SIZE << order, DMA_FROM_DEVICE); in xgbe_alloc_pages()
292 pa->pages_len = PAGE_SIZE << order; in xgbe_alloc_pages()
327 int order, ret; in xgbe_map_rx_buffer() local
336 order = max_t(int, PAGE_ALLOC_COSTLY_ORDER - 1, 0); in xgbe_map_rx_buffer()
338 order); in xgbe_map_rx_buffer()
/linux-4.4.14/drivers/media/v4l2-core/
Dvideobuf2-dma-sg.c69 int order; in vb2_dma_sg_alloc_compacted() local
72 order = get_order(size); in vb2_dma_sg_alloc_compacted()
74 if ((PAGE_SIZE << order) > size) in vb2_dma_sg_alloc_compacted()
75 order--; in vb2_dma_sg_alloc_compacted()
80 __GFP_NOWARN | gfp_flags, order); in vb2_dma_sg_alloc_compacted()
84 if (order == 0) { in vb2_dma_sg_alloc_compacted()
89 order--; in vb2_dma_sg_alloc_compacted()
92 split_page(pages, order); in vb2_dma_sg_alloc_compacted()
93 for (i = 0; i < (1 << order); i++) in vb2_dma_sg_alloc_compacted()
96 size -= PAGE_SIZE << order; in vb2_dma_sg_alloc_compacted()
/linux-4.4.14/drivers/s390/char/
Dtape_3590.c204 struct tape3592_kekl_query_order *order; in tape_3592_kekl_query() local
212 request = tape_alloc_request(2, sizeof(*order)); in tape_3592_kekl_query()
217 order = request->cpdata; in tape_3592_kekl_query()
218 memset(order,0,sizeof(*order)); in tape_3592_kekl_query()
219 order->code = 0xe2; in tape_3592_kekl_query()
220 order->max_count = 2; in tape_3592_kekl_query()
222 tape_ccw_cc(request->cpaddr, PERF_SUBSYS_FUNC, sizeof(*order), order); in tape_3592_kekl_query()
277 struct tape3592_kekl_set_order *order; in tape_3592_kekl_set() local
286 request = tape_alloc_request(1, sizeof(*order)); in tape_3592_kekl_set()
289 order = request->cpdata; in tape_3592_kekl_set()
[all …]
Dhmcdrv_ftp.c239 int order; in hmcdrv_ftp_cmd() local
247 order = get_order(ftp.len); in hmcdrv_ftp_cmd()
248 ftp.buf = (void *) __get_free_pages(GFP_KERNEL | GFP_DMA, order); in hmcdrv_ftp_cmd()
281 free_pages((unsigned long) ftp.buf, order); in hmcdrv_ftp_cmd()
/linux-4.4.14/arch/powerpc/kvm/
Dbook3s_64_mmu_hv.c56 long order = KVM_DEFAULT_HPT_ORDER; in kvmppc_alloc_hpt() local
59 order = *htab_orderp; in kvmppc_alloc_hpt()
60 if (order < PPC_MIN_HPT_ORDER) in kvmppc_alloc_hpt()
61 order = PPC_MIN_HPT_ORDER; in kvmppc_alloc_hpt()
65 page = kvm_alloc_hpt(1ul << (order - PAGE_SHIFT)); in kvmppc_alloc_hpt()
68 memset((void *)hpt, 0, (1ul << order)); in kvmppc_alloc_hpt()
74 while (!hpt && order > PPC_MIN_HPT_ORDER && !htab_orderp) { in kvmppc_alloc_hpt()
76 __GFP_NOWARN, order - PAGE_SHIFT); in kvmppc_alloc_hpt()
78 --order; in kvmppc_alloc_hpt()
85 kvm->arch.hpt_order = order; in kvmppc_alloc_hpt()
[all …]
/linux-4.4.14/tools/testing/fault-injection/
Dfailcmd.sh51 --ignore-gfp-highmem=value, --min-order=value
93 LONGOPTS=$LONGOPTS,ignore-gfp-wait:,ignore-gfp-highmem:,min-order:
193 --min-order)
194 echo $2 > $FAULTATTR/min-order
/linux-4.4.14/arch/parisc/kernel/
Dpci-dma.c420 int order; in pa11_dma_alloc_consistent() local
422 order = get_order(size); in pa11_dma_alloc_consistent()
423 size = 1 << (order + PAGE_SHIFT); in pa11_dma_alloc_consistent()
425 paddr = __get_free_pages(flag, order); in pa11_dma_alloc_consistent()
444 int order; in pa11_dma_free_consistent() local
446 order = get_order(size); in pa11_dma_free_consistent()
447 size = 1 << (order + PAGE_SHIFT); in pa11_dma_free_consistent()
450 free_pages((unsigned long)__va(dma_handle), order); in pa11_dma_free_consistent()
/linux-4.4.14/arch/mips/mm/
Dinit.c65 unsigned int order, i; in setup_zero_pages() local
69 order = 3; in setup_zero_pages()
71 order = 0; in setup_zero_pages()
73 empty_zero_page = __get_free_pages(GFP_KERNEL | __GFP_ZERO, order); in setup_zero_pages()
78 split_page(page, order); in setup_zero_pages()
79 for (i = 0; i < (1 << order); i++, page++) in setup_zero_pages()
82 zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; in setup_zero_pages()
/linux-4.4.14/arch/x86/um/
Dldt.c235 int i, size, k, order; in ldt_get_host_info() local
247 for (i = LDT_PAGES_MAX-1, order=0; i; i>>=1, order++) in ldt_get_host_info()
251 __get_free_pages(GFP_KERNEL|__GFP_ZERO, order); in ldt_get_host_info()
258 ret = modify_ldt(0, ldt, (1<<order)*PAGE_SIZE); in ldt_get_host_info()
294 free_pages((unsigned long)ldt, order); in ldt_get_host_info()
/linux-4.4.14/include/drm/
Ddrm_hashtab.h49 u8 order; member
52 extern int drm_ht_create(struct drm_open_hash *ht, unsigned int order);
/linux-4.4.14/Documentation/devicetree/bindings/arm/
Dfw-cfg.txt19 The selector register takes keys in big endian byte order.
24 such a word, in increasing address order, correspond to the bytes that would
25 have been transferred by byte-wide accesses in chronological order.
29 OS. For example, boot order of devices, ACPI tables, SMBIOS tables, kernel and
45 as a uint32_t value in little endian byte order. The current value
Darch_timer.txt18 hypervisor timers, in that order.
64 - interrupts : Interrupt list for physical and virtual timers in that order.
67 - reg : The first and second view base addresses in that order. The second view
/linux-4.4.14/drivers/usb/atm/
DKconfig25 modem. In order to use your modem you will need to install the
37 AccessRunner chipset. In order to use your modem you will need to
49 or eagle chipset. In order to use your modem you will need to
60 another USB DSL drivers. In order to use your modem you will need to
/linux-4.4.14/arch/score/mm/
Dinit.c101 #define __page_aligned(order) __attribute__((__aligned__(PAGE_SIZE<<order))) argument
/linux-4.4.14/sound/soc/
Dsoc-core.c1000 static void soc_remove_dai(struct snd_soc_dai *dai, int order) in soc_remove_dai() argument
1005 dai->driver->remove_order == order) { in soc_remove_dai()
1017 static void soc_remove_link_dais(struct snd_soc_card *card, int num, int order) in soc_remove_link_dais() argument
1030 soc_remove_dai(rtd->codec_dais[i], order); in soc_remove_link_dais()
1032 soc_remove_dai(rtd->cpu_dai, order); in soc_remove_link_dais()
1036 int order) in soc_remove_link_components() argument
1045 if (platform && platform->component.driver->remove_order == order) in soc_remove_link_components()
1051 if (component->driver->remove_order == order) in soc_remove_link_components()
1057 if (cpu_dai->component->driver->remove_order == order) in soc_remove_link_components()
1064 int dai, order; in soc_remove_dai_links() local
[all …]
/linux-4.4.14/arch/m68k/ifpsp060/src/
DREADME-SRC6 assembler, however it is being included in order to comply with the
9 You don't need to actually assemble these files in order to compile a
/linux-4.4.14/arch/arm/mm/
Ddma-mapping.c260 unsigned long order = get_order(size); in __dma_alloc_buffer() local
263 page = alloc_pages(gfp, order); in __dma_alloc_buffer()
270 split_page(page, order); in __dma_alloc_buffer()
271 for (p = page + (size >> PAGE_SHIFT), e = page + (1 << order); p < e; p++) in __dma_alloc_buffer()
535 unsigned long order = get_order(size); in __alloc_from_contiguous() local
540 page = dma_alloc_from_contiguous(dev, count, order); in __alloc_from_contiguous()
1035 unsigned int order = get_order(size); in __alloc_iova() local
1043 if (order > CONFIG_ARM_DMA_IOMMU_ALIGNMENT) in __alloc_iova()
1044 order = CONFIG_ARM_DMA_IOMMU_ALIGNMENT; in __alloc_iova()
1047 align = (1 << order) - 1; in __alloc_iova()
[all …]
/linux-4.4.14/drivers/s390/net/
Dfsm.c17 int nr_events, const fsm_node *tmpl, int tmpl_len, gfp_t order) in init_fsm() argument
24 this = kzalloc(sizeof(fsm_instance), order); in init_fsm()
33 f = kzalloc(sizeof(fsm), order); in init_fsm()
46 m = kcalloc(nr_states*nr_events, sizeof(fsm_function_t), order); in init_fsm()
/linux-4.4.14/tools/virtio/virtio-trace/
DREADME15 3) A controller thread does poll() for a start order of a host.
16 4) After the controller of the trace agent receives a start order from a host,
20 6) If the controller receives a stop order from a host, the read/write threads
103 read/write threads in the agent wait for start order from host. If you add -o
113 A host injects read start order to the guest via virtio-serial.
117 A host injects read stop order to the guest via virtio-serial.
/linux-4.4.14/include/xen/
Dxen-ops.h24 int xen_create_contiguous_region(phys_addr_t pstart, unsigned int order,
28 void xen_destroy_contiguous_region(phys_addr_t pstart, unsigned int order);
/linux-4.4.14/arch/um/include/shared/
Dkern_util.h22 extern unsigned long alloc_stack(int order, int atomic);
23 extern void free_stack(unsigned long stack, int order);
/linux-4.4.14/Documentation/video4linux/cx2341x/
Dfw-decoder-api.txt72 reverse order.
82 has to be set to the correct value in order to keep the timing correct.
208 Frame count by decode order
210 Video PTS bits 0:31 by display order
212 Video PTS bit 32 by display order
214 SCR bits 0:31 by display order
216 SCR bit 32 by display order
/linux-4.4.14/Documentation/security/
Dkeys-ecryptfs.txt10 the FEK is encrypted by 'ecryptfsd' with the help of external libraries in order
21 format 'ecryptfs' in order to be used in conjunction with the eCryptfs
26 In order to avoid known-plaintext attacks, the datablob obtained through
33 time after the unsealing of a 'trusted' key in order to perform the mount in a
/linux-4.4.14/drivers/staging/comedi/drivers/
Dmite.h153 unsigned order = 0; in MITE_IODWBSR_1_WSIZE_bits() local
156 order = ilog2(size); in MITE_IODWBSR_1_WSIZE_bits()
157 BUG_ON(order < 1); in MITE_IODWBSR_1_WSIZE_bits()
158 return (order - 1) & 0x1f; in MITE_IODWBSR_1_WSIZE_bits()
/linux-4.4.14/Documentation/devicetree/bindings/powerpc/4xx/
Dcpm.txt10 one of two different order for the CPM
12 in the following order (ER,FR,SR). The
13 others have them in the following order
/linux-4.4.14/Documentation/networking/caif/
Dspi_porting.txt6 Two extra GPIOs have been added in order to negotiate the transfers
26 In order to implement a SPI device you will, as a minimum,
51 This function is called by the CAIF SPI slave device in order to
54 This function can be called from IRQ context (recommended in order
60 This function is called by the CAIF SPI slave device in order to
63 This function can be called from IRQ context (recommended in order
83 In order to optimize throughput, a number of SPI padding options are provided.
123 * order not to add any delay. */
/linux-4.4.14/arch/nios2/mm/
Dinit.c97 #define __page_aligned(order) __aligned(PAGE_SIZE << (order)) argument
/linux-4.4.14/sound/drivers/pcsp/
Dpcsp.c46 int err, div, min_div, order; in snd_pcsp_create() local
69 order = fls(div) - 1; in snd_pcsp_create()
71 pcsp_chip.max_treble = min(order, PCSP_MAX_TREBLE); in snd_pcsp_create()
/linux-4.4.14/Documentation/devicetree/bindings/display/exynos/
Dexynos5433-decon.txt11 order: VSYNC, LCD_SYSTEM. The interrupt specifier format
14 in the same order as they were listed in the interrupts
18 - clock-names: list of clock names sorted in the same order as the clocks
Dexynos7-decon.txt16 order: FIFO Level, VSYNC, LCD_SYSTEM. The interrupt specifier
20 "lcd_sys", in the same order as they were listed in the interrupts
30 - clock-names: list of clock names sorted in the same order as the clocks
Dsamsung-fimd.txt23 order: FIFO Level, VSYNC, LCD_SYSTEM. The interrupt specifier
27 "lcd_sys", in the same order as they were listed in the interrupts
37 - clock-names: list of clock names sorted in the same order as the clocks
/linux-4.4.14/Documentation/ABI/testing/
Dsysfs-bus-iio-trigger-sysfs18 trigger. In order to associate the trigger with an IIO device
28 In order to achieve this, one should write a positive integer
38 created trigger from the list of available triggers. In order to
/linux-4.4.14/drivers/base/
Ddma-coherent.c159 int order = get_order(size); in dma_alloc_from_coherent() local
175 pageno = bitmap_find_free_region(mem->bitmap, mem->size, order); in dma_alloc_from_coherent()
213 int dma_release_from_coherent(struct device *dev, int order, void *vaddr) in dma_release_from_coherent() argument
223 bitmap_release_region(mem->bitmap, page, order); in dma_release_from_coherent()
Ddevres.c915 unsigned int order; member
930 free_pages(devres->addr, devres->order); in devm_pages_release()
947 gfp_t gfp_mask, unsigned int order) in devm_get_free_pages() argument
952 addr = __get_free_pages(gfp_mask, order); in devm_get_free_pages()
960 free_pages(addr, order); in devm_get_free_pages()
965 devres->order = order; in devm_get_free_pages()
/linux-4.4.14/Documentation/nios2/
DREADME6 In order to compile for Nios II, you need a version of GCC with support for the generic
17 Altera family of FPGAs. In order to support Linux, Nios II needs to be configured
/linux-4.4.14/arch/um/kernel/
Dprocess.c58 void free_stack(unsigned long stack, int order) in free_stack() argument
60 free_pages(stack, order); in free_stack()
63 unsigned long alloc_stack(int order, int atomic) in alloc_stack() argument
70 page = __get_free_pages(flags, order); in alloc_stack()
/linux-4.4.14/Documentation/devicetree/bindings/input/touchscreen/
Dti-tsc-adc.txt18 ti,wire-config: Different boards could have a different order for
35 in order to avoid false pen-up events. This value
47 ADC in the order of ti,adc-channels. The
54 of ADC in the order of ti,adc-channels.
/linux-4.4.14/fs/ext4/
Dmballoc.c441 static void *mb_find_buddy(struct ext4_buddy *e4b, int order, int *max) in mb_find_buddy() argument
448 if (order > e4b->bd_blkbits + 1) { in mb_find_buddy()
454 if (order == 0) { in mb_find_buddy()
459 bb = e4b->bd_buddy + EXT4_SB(e4b->bd_sb)->s_mb_offsets[order]; in mb_find_buddy()
460 *max = EXT4_SB(e4b->bd_sb)->s_mb_maxs[order]; in mb_find_buddy()
558 int order = e4b->bd_blkbits + 1; in __mb_check_buddy() local
578 while (order > 1) { in __mb_check_buddy()
579 buddy = mb_find_buddy(e4b, order, &max); in __mb_check_buddy()
581 buddy2 = mb_find_buddy(e4b, order - 1, &max2); in __mb_check_buddy()
605 for (j = 0; j < (1 << order); j++) { in __mb_check_buddy()
[all …]
/linux-4.4.14/Documentation/devicetree/bindings/power/reset/
Dkeystone-reset.txt22 in order to use mux block registers for all watchdogs.
32 in format: <0>, <2>; It can be in random order and
34 reset watchdogs and can be in random order.
/linux-4.4.14/arch/sh/
DKconfig.cpu8 endian byte order. These modes require different kernels.
27 This option must be set in order to enable the FPU.
50 This option must be set in order to enable the DSP.
/linux-4.4.14/Documentation/devicetree/bindings/hsi/
Domap-ssi.txt8 - reg-names: Contains the values "sys" and "gdd" (in this order).
30 - reg-names: Contains the values "tx" and "rx" (in this order).
35 0 and 1 (in this order).
/linux-4.4.14/drivers/media/platform/exynos4-is/
Dfimc-is-param.c704 isp->otf_input.order = OTF_INPUT_ORDER_BAYER_GR_BG; in fimc_is_set_initial_params()
715 isp->dma1_input.order = 0; in fimc_is_set_initial_params()
727 isp->dma2_input.order = 0; in fimc_is_set_initial_params()
780 isp->otf_output.order = 0; in fimc_is_set_initial_params()
790 isp->dma1_output.order = 0; in fimc_is_set_initial_params()
806 isp->dma2_output.order = 0; in fimc_is_set_initial_params()
833 drc->otf_input.order = 0; in fimc_is_set_initial_params()
842 drc->dma_input.order = 0; in fimc_is_set_initial_params()
856 drc->otf_output.order = 0; in fimc_is_set_initial_params()
872 fd->otf_input.order = 0; in fimc_is_set_initial_params()
[all …]
/linux-4.4.14/Documentation/devicetree/bindings/dma/
Dbrcm,bcm2835-dma.txt14 to the DMA channels in ascending order.
18 not used by the firmware in ascending order,
Dsnps-dma.txt11 - chan_allocation_order: order of allocation of channel, 0 (default): ascending,
46 The four cells in order are:
/linux-4.4.14/Documentation/devicetree/bindings/mmc/
Dexynos-dw-mshc.txt32 data rate mode operation. Refer notes below for the order of the cells and the
37 data rate mode operation. Refer notes below for the order of the cells and the
44 The order of the cells should be
63 no particular order. The format of the gpio specifier depends on the gpio
/linux-4.4.14/Documentation/
Dprintk-formats.txt153 specifier to use reversed byte order suitable for visual interpretation
154 of Bluetooth addresses which are in the little endian order.
169 host, network, big or little endian order addresses respectively. Where
170 no specifier is provided the default network/big endian order is used.
180 For printing IPv6 network-order 16-bit hex addresses. The 'I6' and 'i6'
232 'b' and 'B' specifiers are used to specify a little endian order in
233 lower ('l') or upper case ('L') hex characters - and big endian order
237 order with lower case hex characters will be printed.
/linux-4.4.14/arch/x86/xen/
Dmmu.c2511 static void xen_zap_pfn_range(unsigned long vaddr, unsigned int order, in xen_zap_pfn_range() argument
2519 for (i = 0; i < (1UL<<order); i++, vaddr += PAGE_SIZE) { in xen_zap_pfn_range()
2539 static void xen_remap_exchanged_ptes(unsigned long vaddr, int order, in xen_remap_exchanged_ptes() argument
2548 limit = 1u << order; in xen_remap_exchanged_ptes()
2562 if (order == 0) in xen_remap_exchanged_ptes()
2622 int xen_create_contiguous_region(phys_addr_t pstart, unsigned int order, in xen_create_contiguous_region() argument
2640 if (unlikely(order > MAX_CONTIG_ORDER)) in xen_create_contiguous_region()
2643 memset((void *) vstart, 0, PAGE_SIZE << order); in xen_create_contiguous_region()
2648 xen_zap_pfn_range(vstart, order, in_frames, NULL); in xen_create_contiguous_region()
2652 success = xen_exchange_memory(1UL << order, 0, in_frames, in xen_create_contiguous_region()
[all …]
/linux-4.4.14/drivers/misc/sgi-gru/
Dgrufile.c268 int order = get_order(sizeof(struct gru_blade_state)); in gru_init_tables() local
279 page = alloc_pages_node(nid, GFP_KERNEL, order); in gru_init_tables()
309 free_pages((unsigned long)gru_base[bid], order); in gru_init_tables()
316 int order = get_order(sizeof(struct gru_state) * in gru_free_tables() local
320 free_pages((unsigned long)gru_base[bid], order); in gru_free_tables()
/linux-4.4.14/drivers/pci/
Dsetup-bus.c974 int order; in calculate_mem_align() local
976 for (order = 0; order <= max_order; order++) { in calculate_mem_align()
979 align1 <<= (order + 20); in calculate_mem_align()
985 align += aligns[order]; in calculate_mem_align()
1019 int order, max_order; in pbus_size_mem() local
1064 order = __ffs(align) - 20; in pbus_size_mem()
1065 if (order < 0) in pbus_size_mem()
1066 order = 0; in pbus_size_mem()
1067 if (order >= ARRAY_SIZE(aligns)) { in pbus_size_mem()
1077 aligns[order] += align; in pbus_size_mem()
[all …]
/linux-4.4.14/Documentation/mtd/nand/
Dpxa3xx-nand.txt11 The controller has a 2176 bytes FIFO buffer. Therefore, in order to support
36 So, in order to achieve reading (for instance), we issue several READ0 commands
75 data area. In other words, in order to read the OOB (aka READOOB), the entire
78 In the same sense, in order to write to the spare OOB the driver has to write
/linux-4.4.14/Documentation/input/
Dappletouch.txt11 * appletouch can be interfaced with the synaptics X11 driver, in order
22 In order to use the touchpad in the basic mode, compile the driver and load
29 emulation, etc. In order to do this, make sure you're using a recent version of
/linux-4.4.14/drivers/mfd/
DKconfig60 additional drivers must be enabled in order to use the
82 additional drivers must be enabled in order to use the
171 order to use the functionality of the device.
183 order to use the functionality of the device.
195 Additional drivers must be enabled in order to use the functionality
210 Additional drivers must be enabled in order to use the functionality
222 Additional drivers must be enabled in order to use the functionality
234 Additional drivers must be enabled in order to use the specific
244 etc. must be enabled in order to use the functionality of
255 additional drivers must be enabled in order to use the
[all …]
/linux-4.4.14/tools/vm/
Dslabinfo.c34 int order, poison, reclaim_account, red_zone; member
339 return s->slabs * (page_size << s->order); in slab_size()
525 s->name, s->aliases, s->order, s->objects); in report()
539 s->slabs * (page_size << s->order)); in report()
544 page_size << s->order, s->partial, onoff(s->poison), in report()
545 s->slabs * (page_size << s->order) - s->objects * s->object_size); in report()
551 ((page_size << s->order) - s->objs_per_slab * s->slab_size) * in report()
623 s->order_fallback, s->order, s->cmpxchg_double_fail, in slabcache()
628 s->objs_per_slab, s->order, in slabcache()
631 (s->slabs * (page_size << s->order)) : 100, in slabcache()
[all …]
/linux-4.4.14/net/sctp/
Dprotocol.c1356 int order; in sctp_init() local
1419 order = get_order(goal); in sctp_init()
1428 order = min(order, max_entry_order); in sctp_init()
1431 sctp_assoc_hashsize = (1UL << order) * PAGE_SIZE / in sctp_init()
1433 if ((sctp_assoc_hashsize > (64 * 1024)) && order > 0) in sctp_init()
1436 __get_free_pages(GFP_ATOMIC|__GFP_NOWARN, order); in sctp_init()
1437 } while (!sctp_assoc_hashtable && --order > 0); in sctp_init()
1469 __get_free_pages(GFP_ATOMIC|__GFP_NOWARN, order); in sctp_init()
1470 } while (!sctp_port_hashtable && --order > 0); in sctp_init()
1481 num_entries = (1UL << order) * PAGE_SIZE / in sctp_init()
/linux-4.4.14/arch/xtensa/include/uapi/asm/
Dbyteorder.h9 # error processor byte order undefined!
Dmsgbuf.h37 # error processor byte order undefined!
/linux-4.4.14/drivers/net/wireless/iwlwifi/mvm/
Dfw.c216 int order, num_of_pages; in iwl_alloc_fw_paging_mem() local
241 order = get_order(FW_PAGING_SIZE); in iwl_alloc_fw_paging_mem()
242 block = alloc_pages(GFP_KERNEL, order); in iwl_alloc_fw_paging_mem()
254 PAGE_SIZE << order, DMA_BIDIRECTIONAL); in iwl_alloc_fw_paging_mem()
271 order); in iwl_alloc_fw_paging_mem()
279 order = get_order(PAGING_BLOCK_SIZE); in iwl_alloc_fw_paging_mem()
280 block = alloc_pages(GFP_KERNEL, order); in iwl_alloc_fw_paging_mem()
292 PAGE_SIZE << order, in iwl_alloc_fw_paging_mem()
311 order); in iwl_alloc_fw_paging_mem()
/linux-4.4.14/Documentation/cris/
DREADME9 In order to compile this you need a version of gcc with support for the
93 Dentry-cache hash table entries: 2048 (order: 1, 16384 bytes)
94 Buffer-cache hash table entries: 2048 (order: 0, 8192 bytes)
95 Page-cache hash table entries: 2048 (order: 0, 8192 bytes)
98 Inode-cache hash table entries: 1024 (order: 0, 8192 bytes)
/linux-4.4.14/Documentation/sound/alsa/
DJack-Controls.txt31 In order to keep compatibility, snd_jack_new() has been modified by
38 HDA jacks can set phantom_jack to true in order to create a phantom
/linux-4.4.14/Documentation/devicetree/bindings/display/tilcdc/
Dpanel.txt13 - raster-order: Raster Data Order Select: 1=Most-to-least 0=Least-to-most
47 raster-order = <0>;
/linux-4.4.14/Documentation/devicetree/bindings/ata/
Dsata_highbank.txt21 - calxeda,led-order : a u32 array that map port numbers to offsets within the
40 calxeda,led-order = <4 0 1 2 3>;
/linux-4.4.14/mm/kasan/
Dkasan.c303 void kasan_alloc_pages(struct page *page, unsigned int order) in kasan_alloc_pages() argument
306 kasan_unpoison_shadow(page_address(page), PAGE_SIZE << order); in kasan_alloc_pages()
309 void kasan_free_pages(struct page *page, unsigned int order) in kasan_free_pages() argument
313 PAGE_SIZE << order, in kasan_free_pages()
/linux-4.4.14/arch/powerpc/kernel/
Diommu.c716 unsigned int order; in iommu_free_table() local
741 order = get_order(bitmap_sz); in iommu_free_table()
742 free_pages((unsigned long) tbl->it_map, order); in iommu_free_table()
816 unsigned int order; in iommu_alloc_coherent() local
821 order = get_order(size); in iommu_alloc_coherent()
828 if (order >= IOMAP_MAX_ORDER) { in iommu_alloc_coherent()
838 page = alloc_pages_node(node, flag, order); in iommu_alloc_coherent()
850 free_pages((unsigned long)ret, order); in iommu_alloc_coherent()
Dfadump.c551 unsigned long order, count, i; in fadump_cpu_notes_buf_alloc() local
553 order = get_order(size); in fadump_cpu_notes_buf_alloc()
554 vaddr = (void *)__get_free_pages(GFP_KERNEL|__GFP_ZERO, order); in fadump_cpu_notes_buf_alloc()
558 count = 1 << order; in fadump_cpu_notes_buf_alloc()
568 unsigned long order, count, i; in fadump_cpu_notes_buf_free() local
570 order = get_order(size); in fadump_cpu_notes_buf_free()
571 count = 1 << order; in fadump_cpu_notes_buf_free()
575 __free_pages(page, order); in fadump_cpu_notes_buf_free()
/linux-4.4.14/arch/tile/kernel/
Dpci-dma.c42 int order = get_order(size); in tile_dma_alloc_coherent() local
61 pg = homecache_alloc_pages_node(node, gfp, order, PAGE_HOME_DMA); in tile_dma_alloc_coherent()
67 __homecache_free_pages(pg, order); in tile_dma_alloc_coherent()
349 int order = get_order(size); in tile_pci_dma_alloc_coherent() local
355 pg = homecache_alloc_pages_node(node, gfp, order, PAGE_HOME_DMA); in tile_pci_dma_alloc_coherent()
Dmachine_kexec.c215 struct page *kimage_alloc_pages_arch(gfp_t gfp_mask, unsigned int order) in kimage_alloc_pages_arch() argument
218 return alloc_pages_node(0, gfp_mask, order); in kimage_alloc_pages_arch()
/linux-4.4.14/drivers/irqchip/
Dirq-gic-v3-its.c839 int order = get_order(psz); in its_alloc_tables() local
863 order = max(get_order((1UL << ids) * entry_size), in its_alloc_tables()
864 order); in its_alloc_tables()
865 if (order >= MAX_ORDER) { in its_alloc_tables()
866 order = MAX_ORDER - 1; in its_alloc_tables()
868 node_name, order); in its_alloc_tables()
872 alloc_size = (1 << order) * PAGE_SIZE; in its_alloc_tables()
876 order = get_order(GITS_BASER_PAGES_MAX * psz); in its_alloc_tables()
878 node_name, order, alloc_pages); in its_alloc_tables()
881 base = (void *)__get_free_pages(GFP_KERNEL | __GFP_ZERO, order); in its_alloc_tables()
/linux-4.4.14/drivers/acpi/
Dpower.c63 u32 order; member
115 if (e->resource->order > resource->order) { in acpi_power_resources_list_add()
770 if (r->order > resource->order) { in acpi_power_add_resource_to_list()
814 resource->order = acpi_object.power_resource.resource_order; in acpi_add_power_resource()
/linux-4.4.14/net/ipv4/
Dfib_semantics.c429 static int fib_detect_death(struct fib_info *fi, int order, in fib_detect_death() argument
445 if ((state & NUD_VALID) && order != dflt) in fib_detect_death()
448 (*last_idx < 0 && order > dflt && state != NUD_INCOMPLETE)) { in fib_detect_death()
450 *last_idx = order; in fib_detect_death()
1432 int order = -1, last_idx = -1; in fib_select_default() local
1470 } else if (!fib_detect_death(fi, order, &last_resort, in fib_select_default()
1473 fa1->fa_default = order; in fib_select_default()
1477 order++; in fib_select_default()
1480 if (order <= 0 || !fi) { in fib_select_default()
1486 if (!fib_detect_death(fi, order, &last_resort, &last_idx, in fib_select_default()
[all …]
/linux-4.4.14/drivers/dma/ioat/
Ddma.c329 ioat_alloc_ring(struct dma_chan *c, int order, gfp_t flags) in ioat_alloc_ring() argument
332 int descs = 1 << order; in ioat_alloc_ring()
335 if (order > ioat_get_max_alloc_order()) in ioat_alloc_ring()
365 static bool reshape_ring(struct ioatdma_chan *ioat_chan, int order) in reshape_ring() argument
374 const u32 new_size = 1 << order; in reshape_ring()
378 if (order > ioat_get_max_alloc_order()) in reshape_ring()
468 ioat_chan->alloc_order = order; in reshape_ring()
/linux-4.4.14/Documentation/scsi/
Dncr53c8xx.txt23 8.4 Set order type for tagged command
102 chips in order to gain advantage of new features, as LOAD/STORE instructions
240 In order to really gain advantage of this feature, devices must have
310 checking. These features must be enabled in order to ensure safe data
333 In order to display profiling information, just enter:
476 8.4 Set order type for tagged command
478 setorder <order>
480 order: 3 possible values:
511 data transferred reaches 1000 GB in order to avoid overflow.
525 Do not specify any flag in order to reset the flag. For example:
[all …]
/linux-4.4.14/Documentation/misc-devices/mei/
Dmei-client-bus.txt13 In order to plug seamlessly into the kernel device driver model we add kernel virtual
50 In order for drivers to be notified of pending events waiting for them (e.g.
54 to call mei_recv() from the event handler in order to fetch the pending
/linux-4.4.14/drivers/media/i2c/soc_camera/
Dmt9t112.c83 u16 order; member
109 .order = 0,
114 .order = 1,
119 .order = 2,
124 .order = 3,
129 .order = 2,
134 .order = 2,
822 mt9t112_mcu_write(ret, client, VAR(26, 9), priv->format->order); in mt9t112_s_stream()
/linux-4.4.14/fs/cramfs/
DREADME25 The order of inode traversal is described as "width-first" (not to be
28 same order as `ls -AUR' (but without the /^\..*:$/ directory header
29 lines); put another way, the same order as `find -type d -exec
52 The order of <file_data>'s is a depth-first descent of the directory
53 tree, i.e. the same order as `find -size +0 \( -type f -o -type l \)
/linux-4.4.14/Documentation/devicetree/bindings/spmi/
Dqcom,spmi-pmic-arb.txt26 - reg : address + size pairs describing the PMIC arb register sets; order must
27 correspond with the order of entries in reg-names
/linux-4.4.14/Documentation/isdn/
DINTERFACE.fax10 In order to use fax, the LL provides the pointer to this struct with the
16 In send-mode the HL-driver has to handle the <DLE> codes and the bit-order
18 In receive-mode the LL-driver takes care of the bit-order conversion
/linux-4.4.14/Documentation/locking/
Drt-mutex.txt32 priority order. For same priorities FIFO order is chosen. For each
34 priority waiters list. This list too queues in priority order. Whenever
/linux-4.4.14/drivers/staging/lustre/lustre/ptlrpc/
Devents.c443 __u32 order; in ptlrpc_uuid_to_peer() local
453 dist = LNetDist(dst_nid, &src_nid, &order); in ptlrpc_uuid_to_peer()
465 (dist == best_dist && order < best_order)) { in ptlrpc_uuid_to_peer()
467 best_order = order; in ptlrpc_uuid_to_peer()
/linux-4.4.14/drivers/i2c/busses/
Di2c-pnx.c298 if (alg_data->mif.order) { in i2c_pnx_master_rcv()
303 if (alg_data->mif.order == 1) { in i2c_pnx_master_rcv()
323 alg_data->mif.order--; in i2c_pnx_master_rcv()
524 alg_data->mif.order = pmsg->len; in i2c_pnx_xfer()
577 alg_data->mif.order = 0; in i2c_pnx_xfer()
/linux-4.4.14/drivers/staging/wilc1000/
Dwilc_spi.c682 u8 cmd, order, crc[2] = {0}; in spi_data_write() local
702 order = 0x3; in spi_data_write()
704 order = 0x1; in spi_data_write()
707 order = 0x3; in spi_data_write()
709 order = 0x2; in spi_data_write()
711 cmd |= order; in spi_data_write()
/linux-4.4.14/drivers/block/
Drbd_types.h67 __u8 order; member
/linux-4.4.14/arch/s390/kernel/
Dsmp.c108 static inline int __pcpu_sigp_relax(u16 addr, u8 order, unsigned long parm, in __pcpu_sigp_relax() argument
114 cc = __pcpu_sigp(addr, order, parm, NULL); in __pcpu_sigp_relax()
121 static int pcpu_sigp_retry(struct pcpu *pcpu, u8 order, u32 parm) in pcpu_sigp_retry() argument
126 cc = __pcpu_sigp(pcpu->address, order, parm, NULL); in pcpu_sigp_retry()
169 int order; in pcpu_ec_call() local
173 order = pcpu_running(pcpu) ? SIGP_EXTERNAL_CALL : SIGP_EMERGENCY_SIGNAL; in pcpu_ec_call()
174 pcpu_sigp_retry(pcpu, order, 0); in pcpu_ec_call()
/linux-4.4.14/drivers/video/fbdev/vermilion/
Dvermilion.c112 va->order = max_order; in vmlfb_alloc_vram_area()
170 free_pages(va->logical, va->order); in vmlfb_free_vram_area()
202 int order; in vmlfb_alloc_vram() local
211 order = 0; in vmlfb_alloc_vram()
213 while (requested > (PAGE_SIZE << order) && order < MAX_ORDER) in vmlfb_alloc_vram()
214 order++; in vmlfb_alloc_vram()
216 err = vmlfb_alloc_vram_area(va, order, 0); in vmlfb_alloc_vram()
/linux-4.4.14/tools/power/cpupower/debug/kernel/
DMakefile16 - rm -rf .tmp_versions* Module.symvers modules.order

123456