Home
last modified time | relevance | path

Searched refs:batch (Results 1 – 86 of 86) sorted by relevance

/linux-4.4.14/arch/powerpc/mm/
Dtlb_hash64.c47 struct ppc64_tlb_batch *batch = &get_cpu_var(ppc64_tlb_batch); in hpte_need_flush() local
54 i = batch->index; in hpte_need_flush()
100 if (!batch->active) { in hpte_need_flush()
116 if (i != 0 && (mm != batch->mm || batch->psize != psize || in hpte_need_flush()
117 batch->ssize != ssize)) { in hpte_need_flush()
118 __flush_tlb_pending(batch); in hpte_need_flush()
122 batch->mm = mm; in hpte_need_flush()
123 batch->psize = psize; in hpte_need_flush()
124 batch->ssize = ssize; in hpte_need_flush()
126 batch->pte[i] = rpte; in hpte_need_flush()
[all …]
Dhash_native_64.c644 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in native_flush_hash_range() local
645 unsigned long psize = batch->psize; in native_flush_hash_range()
646 int ssize = batch->ssize; in native_flush_hash_range()
652 vpn = batch->vpn[i]; in native_flush_hash_range()
653 pte = batch->pte[i]; in native_flush_hash_range()
678 vpn = batch->vpn[i]; in native_flush_hash_range()
679 pte = batch->pte[i]; in native_flush_hash_range()
695 vpn = batch->vpn[i]; in native_flush_hash_range()
696 pte = batch->pte[i]; in native_flush_hash_range()
Dhugetlbpage.c475 struct hugepd_freelist *batch = in hugepd_free_rcu_callback() local
479 for (i = 0; i < batch->index; i++) in hugepd_free_rcu_callback()
480 kmem_cache_free(hugepte_cache, batch->ptes[i]); in hugepd_free_rcu_callback()
482 free_page((unsigned long)batch); in hugepd_free_rcu_callback()
Dhash_utils_64.c1410 struct ppc64_tlb_batch *batch = in flush_hash_range() local
1414 flush_hash_page(batch->vpn[i], batch->pte[i], in flush_hash_range()
1415 batch->psize, batch->ssize, local); in flush_hash_range()
/linux-4.4.14/drivers/gpu/drm/vmwgfx/
Dvmwgfx_mob.c238 struct vmw_otable_batch *batch) in vmw_otable_batch_setup() argument
242 struct vmw_otable *otables = batch->otables; in vmw_otable_batch_setup()
247 for (i = 0; i < batch->num_otables; ++i) { in vmw_otable_batch_setup()
260 &batch->otable_bo); in vmw_otable_batch_setup()
265 ret = ttm_bo_reserve(batch->otable_bo, false, true, false, NULL); in vmw_otable_batch_setup()
267 ret = vmw_bo_driver.ttm_tt_populate(batch->otable_bo->ttm); in vmw_otable_batch_setup()
270 ret = vmw_bo_map_dma(batch->otable_bo); in vmw_otable_batch_setup()
274 ttm_bo_unreserve(batch->otable_bo); in vmw_otable_batch_setup()
277 for (i = 0; i < batch->num_otables; ++i) { in vmw_otable_batch_setup()
278 if (!batch->otables[i].enabled) in vmw_otable_batch_setup()
[all …]
/linux-4.4.14/arch/powerpc/include/asm/
Dtlbflush.h104 extern void __flush_tlb_pending(struct ppc64_tlb_batch *batch);
110 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_enter_lazy_mmu_mode() local
112 batch->active = 1; in arch_enter_lazy_mmu_mode()
117 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_leave_lazy_mmu_mode() local
119 if (batch->index) in arch_leave_lazy_mmu_mode()
120 __flush_tlb_pending(batch); in arch_leave_lazy_mmu_mode()
121 batch->active = 0; in arch_leave_lazy_mmu_mode()
/linux-4.4.14/drivers/gpu/drm/i915/
Dintel_lrc.c1114 #define wa_ctx_emit(batch, index, cmd) \ argument
1120 batch[__index] = (cmd); \
1141 uint32_t *const batch, in gen8_emit_flush_coherentl3_wa() argument
1155 wa_ctx_emit(batch, index, (MI_STORE_REGISTER_MEM_GEN8 | in gen8_emit_flush_coherentl3_wa()
1157 wa_ctx_emit(batch, index, GEN8_L3SQCREG4); in gen8_emit_flush_coherentl3_wa()
1158 wa_ctx_emit(batch, index, ring->scratch.gtt_offset + 256); in gen8_emit_flush_coherentl3_wa()
1159 wa_ctx_emit(batch, index, 0); in gen8_emit_flush_coherentl3_wa()
1161 wa_ctx_emit(batch, index, MI_LOAD_REGISTER_IMM(1)); in gen8_emit_flush_coherentl3_wa()
1162 wa_ctx_emit(batch, index, GEN8_L3SQCREG4); in gen8_emit_flush_coherentl3_wa()
1163 wa_ctx_emit(batch, index, l3sqc4_flush); in gen8_emit_flush_coherentl3_wa()
[all …]
Di915_gem_render_state.c85 #define OUT_BATCH(batch, i, val) \ argument
91 (batch)[(i)++] = (val); \
110 u32 s = rodata->batch[i]; in render_state_setup()
117 rodata->batch[i + 1] != 0) { in render_state_setup()
Dintel_renderstate.h37 .batch = gen ## _g ## _null_state_batch, \
Di915_gem_render_state.h31 const u32 *batch; member
/linux-4.4.14/lib/
Dpercpu_counter.c75 void __percpu_counter_add(struct percpu_counter *fbc, s64 amount, s32 batch) in __percpu_counter_add() argument
81 if (count >= batch || count <= -batch) { in __percpu_counter_add()
200 int __percpu_counter_compare(struct percpu_counter *fbc, s64 rhs, s32 batch) in __percpu_counter_compare() argument
206 if (abs(count - rhs) > (batch * num_online_cpus())) { in __percpu_counter_compare()
/linux-4.4.14/mm/
Dmemory.c186 struct mmu_gather_batch *batch; in tlb_next_batch() local
188 batch = tlb->active; in tlb_next_batch()
189 if (batch->next) { in tlb_next_batch()
190 tlb->active = batch->next; in tlb_next_batch()
197 batch = (void *)__get_free_pages(GFP_NOWAIT | __GFP_NOWARN, 0); in tlb_next_batch()
198 if (!batch) in tlb_next_batch()
202 batch->next = NULL; in tlb_next_batch()
203 batch->nr = 0; in tlb_next_batch()
204 batch->max = MAX_GATHER_BATCH; in tlb_next_batch()
206 tlb->active->next = batch; in tlb_next_batch()
[all …]
Dmm_init.c152 s32 batch = max_t(s32, nr*2, 32); in mm_compute_batch() local
157 vm_committed_as_batch = max_t(s32, memsized_batch, batch); in mm_compute_batch()
Dpage_alloc.c1880 int to_drain, batch; in drain_zone_pages() local
1883 batch = READ_ONCE(pcp->batch); in drain_zone_pages()
1884 to_drain = min(pcp->count, batch); in drain_zone_pages()
2082 unsigned long batch = READ_ONCE(pcp->batch); in free_hot_cold_page() local
2083 free_pcppages_bulk(zone, batch, pcp); in free_hot_cold_page()
2084 pcp->count -= batch; in free_hot_cold_page()
2228 pcp->batch, list, in buffered_rmqueue()
4296 static void setup_pageset(struct per_cpu_pageset *p, unsigned long batch);
4557 int batch; in zone_batchsize() local
4565 batch = zone->managed_pages / 1024; in zone_batchsize()
[all …]
Dmemcontrol.c2030 unsigned int batch = max(CHARGE_BATCH, nr_pages); in try_charge() local
2045 page_counter_try_charge(&memcg->memsw, batch, &counter)) { in try_charge()
2046 if (page_counter_try_charge(&memcg->memory, batch, &counter)) in try_charge()
2049 page_counter_uncharge(&memcg->memsw, batch); in try_charge()
2056 if (batch > nr_pages) { in try_charge()
2057 batch = nr_pages; in try_charge()
2142 css_get_many(&memcg->css, batch); in try_charge()
2143 if (batch > nr_pages) in try_charge()
2144 refill_stock(memcg, batch - nr_pages); in try_charge()
2157 current->memcg_nr_pages_over_high += batch; in try_charge()
Dslab.c661 static void init_arraycache(struct array_cache *ac, int limit, int batch) in init_arraycache() argument
674 ac->batchcount = batch; in init_arraycache()
872 int batch, gfp_t gfp) in __alloc_alien_cache() argument
878 init_arraycache(&alc->ac, entries, batch); in __alloc_alien_cache()
Dvmscan.c278 long batch_size = shrinker->batch ? shrinker->batch in do_shrink_slab()
Dvmstat.c1252 pageset->pcp.batch); in zoneinfo_show_print()
Dzsmalloc.c1871 pool->shrinker.batch = 0; in zs_register_shrinker()
/linux-4.4.14/include/linux/
Dpercpu_counter.h42 void __percpu_counter_add(struct percpu_counter *fbc, s64 amount, s32 batch);
44 int __percpu_counter_compare(struct percpu_counter *fbc, s64 rhs, s32 batch);
125 __percpu_counter_compare(struct percpu_counter *fbc, s64 rhs, s32 batch) in __percpu_counter_compare() argument
139 __percpu_counter_add(struct percpu_counter *fbc, s64 amount, s32 batch) in __percpu_counter_add() argument
Dshrinker.h56 long batch; /* reclaim batch size, 0 = default */ member
Dmmzone.h255 int batch; /* chunk size for buddy add/remove */ member
/linux-4.4.14/drivers/staging/lustre/lnet/selftest/
Dconsole.c1140 lstcon_batch_t *batch; in lstcon_testrpc_condition() local
1148 batch = test->tes_batch; in lstcon_testrpc_condition()
1149 LASSERT(batch != NULL); in lstcon_testrpc_condition()
1159 hash = batch->bat_cli_hash; in lstcon_testrpc_condition()
1160 head = &batch->bat_cli_list; in lstcon_testrpc_condition()
1165 hash = batch->bat_srv_hash; in lstcon_testrpc_condition()
1166 head = &batch->bat_srv_list; in lstcon_testrpc_condition()
1232 lstcon_verify_batch(const char *name, lstcon_batch_t **batch) in lstcon_verify_batch() argument
1236 rc = lstcon_batch_find(name, batch); in lstcon_verify_batch()
1242 if ((*batch)->bat_state != LST_BATCH_IDLE) { in lstcon_verify_batch()
[all …]
Dconrpc.c658 lstcon_batch_t *batch; in lstcon_batrpc_prep() local
681 batch = (lstcon_batch_t *)tsb; in lstcon_batrpc_prep()
682 brq->bar_arg = batch->bat_arg; in lstcon_batrpc_prep()
Dframework.c662 sfw_batch_t *batch; in sfw_destroy_session() local
668 batch = list_entry(sn->sn_batches.next, in sfw_destroy_session()
670 list_del_init(&batch->bat_list); in sfw_destroy_session()
671 sfw_destroy_batch(batch); in sfw_destroy_session()
/linux-4.4.14/arch/s390/mm/
Dpgtable.c1060 struct mmu_table_batch *batch; in tlb_remove_table_rcu() local
1063 batch = container_of(head, struct mmu_table_batch, rcu); in tlb_remove_table_rcu()
1065 for (i = 0; i < batch->nr; i++) in tlb_remove_table_rcu()
1066 __tlb_remove_table(batch->tables[i]); in tlb_remove_table_rcu()
1068 free_page((unsigned long)batch); in tlb_remove_table_rcu()
1073 struct mmu_table_batch **batch = &tlb->batch; in tlb_table_flush() local
1075 if (*batch) { in tlb_table_flush()
1076 call_rcu_sched(&(*batch)->rcu, tlb_remove_table_rcu); in tlb_table_flush()
1077 *batch = NULL; in tlb_table_flush()
1083 struct mmu_table_batch **batch = &tlb->batch; in tlb_remove_table() local
[all …]
/linux-4.4.14/fs/xfs/
Dxfs_icache.c532 struct xfs_inode *batch[XFS_LOOKUP_BATCH]; in xfs_inode_ag_walk() local
540 (void **)batch, first_index, in xfs_inode_ag_walk()
545 (void **) batch, first_index, in xfs_inode_ag_walk()
558 struct xfs_inode *ip = batch[i]; in xfs_inode_ag_walk()
561 batch[i] = NULL; in xfs_inode_ag_walk()
586 if (!batch[i]) in xfs_inode_ag_walk()
588 error = execute(batch[i], flags, args); in xfs_inode_ag_walk()
589 IRELE(batch[i]); in xfs_inode_ag_walk()
1053 struct xfs_inode *batch[XFS_LOOKUP_BATCH]; in xfs_reclaim_inodes_ag() local
1059 (void **)batch, first_index, in xfs_reclaim_inodes_ag()
[all …]
Dxfs_mount.c1175 s32 batch; in xfs_mod_fdblocks() local
1211 batch = 1; in xfs_mod_fdblocks()
1213 batch = XFS_FDBLOCKS_BATCH; in xfs_mod_fdblocks()
1215 __percpu_counter_add(&mp->m_fdblocks, delta, batch); in xfs_mod_fdblocks()
Dxfs_qm.c81 struct xfs_dquot *batch[XFS_DQ_LOOKUP_BATCH]; in xfs_qm_dquot_walk() local
86 nr_found = radix_tree_gang_lookup(tree, (void **)batch, in xfs_qm_dquot_walk()
94 struct xfs_dquot *dqp = batch[i]; in xfs_qm_dquot_walk()
98 error = execute(batch[i], data); in xfs_qm_dquot_walk()
/linux-4.4.14/arch/s390/include/asm/
Dtlb.h33 struct mmu_table_batch *batch; member
59 tlb->batch = NULL; in tlb_gather_mmu()
/linux-4.4.14/drivers/xen/
Dgrant-table.c753 void gnttab_batch_map(struct gnttab_map_grant_ref *batch, unsigned count) in gnttab_batch_map() argument
757 if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, batch, count)) in gnttab_batch_map()
759 for (op = batch; op < batch + count; op++) in gnttab_batch_map()
766 void gnttab_batch_copy(struct gnttab_copy *batch, unsigned count) in gnttab_batch_copy() argument
770 if (HYPERVISOR_grant_table_op(GNTTABOP_copy, batch, count)) in gnttab_batch_copy()
772 for (op = batch; op < batch + count; op++) in gnttab_batch_copy()
/linux-4.4.14/Documentation/networking/timestamping/
Dtxtimestamp.c184 int batch = 0; in __recv_errmsg_cmsg() local
223 batch++; in __recv_errmsg_cmsg()
227 if (batch > 1) in __recv_errmsg_cmsg()
228 fprintf(stderr, "batched %d timestamps\n", batch); in __recv_errmsg_cmsg()
/linux-4.4.14/arch/powerpc/platforms/pseries/
Dlpar.c524 struct ppc64_tlb_batch *batch = this_cpu_ptr(&ppc64_tlb_batch); in pSeries_lpar_flush_hash_range() local
534 psize = batch->psize; in pSeries_lpar_flush_hash_range()
535 ssize = batch->ssize; in pSeries_lpar_flush_hash_range()
538 vpn = batch->vpn[i]; in pSeries_lpar_flush_hash_range()
539 pte = batch->pte[i]; in pSeries_lpar_flush_hash_range()
/linux-4.4.14/include/xen/
Dgrant_table.h217 void gnttab_batch_map(struct gnttab_map_grant_ref *batch, unsigned count);
218 void gnttab_batch_copy(struct gnttab_copy *batch, unsigned count);
/linux-4.4.14/virt/kvm/
DKconfig28 # Toggle to switch between direct notification and batch job
/linux-4.4.14/Documentation/features/vm/TLB/
Darch-support.txt2 # Feature name: batch-unmap-tlb-flush
/linux-4.4.14/arch/arm/include/asm/
Dtlb.h68 struct mmu_table_batch *batch; member
164 tlb->batch = NULL; in tlb_gather_mmu()
/linux-4.4.14/drivers/misc/
Dvmw_balloon.c182 static u64 vmballoon_batch_get_pa(struct vmballoon_batch_page *batch, int idx) in vmballoon_batch_get_pa() argument
184 return batch->pages[idx] & VMW_BALLOON_BATCH_PAGE_MASK; in vmballoon_batch_get_pa()
187 static int vmballoon_batch_get_status(struct vmballoon_batch_page *batch, in vmballoon_batch_get_status() argument
190 return (int)(batch->pages[idx] & VMW_BALLOON_BATCH_STATUS_MASK); in vmballoon_batch_get_status()
193 static void vmballoon_batch_set_pa(struct vmballoon_batch_page *batch, int idx, in vmballoon_batch_set_pa() argument
196 batch->pages[idx] = pa; in vmballoon_batch_set_pa()
/linux-4.4.14/tools/vm/
Dpage-types.c598 unsigned long batch; in walk_pfn() local
603 batch = min_t(unsigned long, count, KPAGEFLAGS_BATCH); in walk_pfn()
604 pages = kpageflags_read(buf, index, batch); in walk_pfn()
620 unsigned long batch; in walk_vma() local
626 batch = min_t(unsigned long, count, PAGEMAP_BATCH); in walk_vma()
627 pages = pagemap_read(buf, index, batch); in walk_vma()
/linux-4.4.14/arch/powerpc/kernel/
Dprocess.c769 struct ppc64_tlb_batch *batch; in __switch_to() local
884 batch = this_cpu_ptr(&ppc64_tlb_batch); in __switch_to()
885 if (batch->active) { in __switch_to()
887 if (batch->index) in __switch_to()
888 __flush_tlb_pending(batch); in __switch_to()
889 batch->active = 0; in __switch_to()
907 batch = this_cpu_ptr(&ppc64_tlb_batch); in __switch_to()
908 batch->active = 1; in __switch_to()
/linux-4.4.14/certs/
DMakefile53 -batch -x509 -config $(obj)/x509.genkey \
/linux-4.4.14/include/asm-generic/
Dtlb.h95 struct mmu_table_batch *batch; member
/linux-4.4.14/block/
DKconfig.iosched21 a new point in the service tree and doing a batch of IO from there
/linux-4.4.14/drivers/target/iscsi/
Discsi_target_erl1.c1073 int batch = 0; in iscsit_handle_ooo_cmdsn() local
1081 batch = 1; in iscsit_handle_ooo_cmdsn()
1086 batch = 1; in iscsit_handle_ooo_cmdsn()
1094 ooo_cmdsn->batch_count = (batch) ? in iscsit_handle_ooo_cmdsn()
/linux-4.4.14/Documentation/cgroups/
Dfreezer-subsystem.txt1 The cgroup freezer is useful to batch job management system which start
6 be started/stopped by the batch job management system. It also provides
Dcpuacct.txt49 due to the batch processing nature of percpu_counter.
Dcpusets.txt253 This enables batch managers monitoring jobs running in dedicated
264 This mechanism provides a very economical way for the batch manager
266 batch manager or other user code to decide what to do about it and
278 the system load imposed by a batch scheduler monitoring this
283 counter, a batch scheduler can detect memory pressure with a
288 the batch scheduler can obtain the key information, memory
/linux-4.4.14/Documentation/trace/
Devents-kmem.txt60 When pages are freed in batch, the also mm_page_free_batched is triggered.
62 freed in batch with a page list. Significant amounts of activity here could
/linux-4.4.14/Documentation/ABI/testing/
Dsysfs-fs-f2fs82 Controls the trimming rate in batch mode.
/linux-4.4.14/Documentation/block/
Ddeadline-iosched.txt39 maximum number of requests per batch.
Dbiodoc.txt983 The generic i/o scheduler algorithm attempts to sort/merge/batch requests for
1011 iii. Plugging the queue to batch requests in anticipation of opportunities for
/linux-4.4.14/drivers/block/xen-blkback/
Dblkback.c770 unsigned int batch = min(num, BLKIF_MAX_SEGMENTS_PER_REQUEST); in xen_blkbk_unmap() local
772 invcount = xen_blkbk_unmap_prepare(blkif, pages, batch, in xen_blkbk_unmap()
779 pages += batch; in xen_blkbk_unmap()
780 num -= batch; in xen_blkbk_unmap()
/linux-4.4.14/drivers/staging/android/ion/
Dion_heap.c314 heap->shrinker.batch = 0; in ion_heap_init_shrinker()
/linux-4.4.14/arch/x86/xen/
Dmmu.c2844 int batch = min(REMAP_BATCH_SIZE, nr); in do_remap_gfn() local
2845 int batch_left = batch; in do_remap_gfn()
2846 range = (unsigned long)batch << PAGE_SHIFT; in do_remap_gfn()
2882 nr -= batch; in do_remap_gfn()
2885 err_ptr += batch; in do_remap_gfn()
/linux-4.4.14/Documentation/scheduler/
Dsched-bwc.txt54 "silos" in a batch fashion. This greatly reduces global accounting pressure
Dsched-design-CFS.txt125 batch jobs.
/linux-4.4.14/Documentation/
Ddynamic-debug-howto.txt107 If your query set is big, you can batch them too:
109 ~# cat query-batch-file > <debugfs>/dynamic_debug/control
Dmodule-signing.txt163 openssl req -new -nodes -utf8 -sha256 -days 36500 -batch -x509 \
Dkprobes.txt43 There are also register_/unregister_*probes() functions for batch
Dkernel-parameters.txt3084 process in one batch.
3160 batch limiting is disabled.
3164 batch limiting is re-enabled.
Ddevices.txt67 in "batch mode", so there is likely additional registrations that
/linux-4.4.14/arch/arm/lib/
Dlib1funcs.S171 @ Do comparisons in batch of 4 first.
/linux-4.4.14/Documentation/DocBook/
Dgpu.xml.db595 API-i915-gem-batch-pool-init
596 API-i915-gem-batch-pool-fini
597 API-i915-gem-batch-pool-get
/linux-4.4.14/Documentation/vm/
Dpage_migration28 a process to a processor on a distant node. A batch scheduler or an
/linux-4.4.14/Documentation/crypto/
Dasync-tx-api.txt92 async_<operation> call. Offload engine drivers batch operations to
/linux-4.4.14/Documentation/filesystems/
Dbtrfs.txt109 initiate batch trims from userspace).
Df2fs.txt212 to be trimmed out in batch mode when FITRIM
Dext4.txt277 additional filesystem operations to be batch
Dvfs.txt318 appropriate scan batch sizes without having to worry about whether
319 implementations will cause holdoff problems due to large scan batch
Dxfs-delayed-logging-design.txt82 the current batch completes. It is now common for a single current CPU core to
/linux-4.4.14/fs/
Dsuper.c247 s->s_shrink.batch = 1024; in alloc_super()
/linux-4.4.14/drivers/staging/lustre/lustre/obdecho/
Decho_client.c1647 u64 batch, struct obd_trans_info *oti, in echo_client_prep_commit() argument
1662 npages = batch >> PAGE_CACHE_SHIFT; in echo_client_prep_commit()
/linux-4.4.14/arch/x86/kvm/
Dmmu.c4735 int batch = 0; in kvm_zap_obsolete_pages() local
4761 if (batch >= BATCH_ZAP_PAGES && in kvm_zap_obsolete_pages()
4763 batch = 0; in kvm_zap_obsolete_pages()
4769 batch += ret; in kvm_zap_obsolete_pages()
/linux-4.4.14/Documentation/locking/
Dww-mutex-design.txt16 there is no way to guarantee that buffers appear in a execbuf/batch in
/linux-4.4.14/Documentation/sysctl/
Dvm.txt712 The batch value of each per cpu pagelist is also updated as a result. It is
713 set to pcp->high/4. The upper limit of batch is (PAGE_SHIFT * 8)
/linux-4.4.14/drivers/md/
Draid5.c5745 struct stripe_head *batch[MAX_STRIPE_BATCH], *sh; in handle_active_stripes() local
5751 batch[batch_size++] = sh; in handle_active_stripes()
5777 handle_stripe(batch[i]); in handle_active_stripes()
5784 hash = batch[i]->hash_lock_index; in handle_active_stripes()
5785 __release_stripe(conf, batch[i], &temp_inactive_list[hash]); in handle_active_stripes()
6637 conf->shrinker.batch = 128; in setup_conf()
Ddm-bufio.c1674 c->shrinker.batch = 0; in dm_bufio_client_create()
/linux-4.4.14/Documentation/RCU/
Dchecklist.txt190 a single non-expedited primitive to cover the entire batch.
Dtrace.txt166 o "b" is the batch limit for this CPU. If more than this number
DRTFP.txt1030 Add per-cpu batch counter"
/linux-4.4.14/Documentation/networking/
Dcs89x0.txt89 build - batch file to compile cs89x0.c.
/linux-4.4.14/drivers/message/fusion/lsi/
Dmpi_history.txt230 * 06-22-00 01.00.03 Removed batch controls from LAN_0 page and updated the
/linux-4.4.14/fs/btrfs/
Dtree-log.c2757 int batch = atomic_read(&root->log_batch); in btrfs_sync_log() local
2766 if (batch == atomic_read(&root->log_batch)) in btrfs_sync_log()
/linux-4.4.14/drivers/md/bcache/
Dbtree.c810 c->shrink.batch = c->btree_pages * 2; in bch_btree_cache_alloc()
/linux-4.4.14/drivers/scsi/aic7xxx/
Daic7xxx.seq1078 * we can batch the clearing of HADDR with the fixup.
Daic79xx.seq376 * order is preserved even if we batch.