Lines Matching refs:avail

190 	unsigned int avail;  member
672 ac->avail = 0; in init_arraycache()
728 void *objp = ac->entry[--ac->avail]; in __ac_get_obj()
740 for (i = 0; i < ac->avail; i++) { in __ac_get_obj()
744 ac->entry[i] = ac->entry[ac->avail]; in __ac_get_obj()
745 ac->entry[ac->avail] = objp; in __ac_get_obj()
764 ac->avail++; in __ac_get_obj()
779 objp = ac->entry[--ac->avail]; in ac_get_obj()
803 ac->entry[ac->avail++] = objp; in ac_put_obj()
816 int nr = min3(from->avail, max, to->limit - to->avail); in transfer_objects()
821 memcpy(to->entry + to->avail, from->entry + from->avail -nr, in transfer_objects()
824 from->avail -= nr; in transfer_objects()
825 to->avail += nr; in transfer_objects()
926 if (ac->avail) { in __drain_alien_cache()
936 free_block(cachep, ac->entry, ac->avail, node, list); in __drain_alien_cache()
937 ac->avail = 0; in __drain_alien_cache()
955 if (ac->avail && spin_trylock_irq(&alc->lock)) { in reap_alien()
1002 if (unlikely(ac->avail == ac->limit)) { in __cache_free_alien()
1121 free_block(cachep, nc->entry, nc->avail, node, &list); in cpuup_canceled()
1122 nc->avail = 0; in cpuup_canceled()
1133 shared->avail, node, &list); in cpuup_canceled()
2043 cpu_cache_get(cachep)->avail = 0; in setup_cpu_cache()
2335 free_block(cachep, ac->entry, ac->avail, node, &list); in do_drain()
2338 ac->avail = 0; in do_drain()
2784 BUG_ON(ac->avail > 0 || !n); in cache_alloc_refill()
2833 n->free_objects -= ac->avail; in cache_alloc_refill()
2837 if (unlikely(!ac->avail)) { in cache_alloc_refill()
2847 if (!x && (ac->avail == 0 || force_refill)) in cache_alloc_refill()
2850 if (!ac->avail) /* objects refilled by interrupt? */ in cache_alloc_refill()
2937 if (likely(ac->avail)) { in ____cache_alloc()
2967 kmemleak_erase(&ac->entry[ac->avail]); in ____cache_alloc()
3318 BUG_ON(!batchcount || batchcount > ac->avail); in cache_flusharray()
3325 int max = shared_array->limit - shared_array->avail; in cache_flusharray()
3329 memcpy(&(shared_array->entry[shared_array->avail]), in cache_flusharray()
3331 shared_array->avail += batchcount; in cache_flusharray()
3358 ac->avail -= batchcount; in cache_flusharray()
3359 memmove(ac->entry, &(ac->entry[batchcount]), sizeof(void *)*ac->avail); in cache_flusharray()
3387 if (ac->avail < ac->limit) { in __cache_free()
3639 shared->avail, node, &list); in alloc_kmem_cache_node()
3722 free_block(cachep, ac->entry, ac->avail, node, &list); in __do_tune_cpucache()
3833 if (!ac || !ac->avail) in drain_array()
3839 if (ac->avail) { in drain_array()
3840 tofree = force ? ac->avail : (ac->limit + 4) / 5; in drain_array()
3841 if (tofree > ac->avail) in drain_array()
3842 tofree = (ac->avail + 1) / 2; in drain_array()
3844 ac->avail -= tofree; in drain_array()
3846 sizeof(void *) * ac->avail); in drain_array()
3962 shared_avail += n->shared->avail; in get_slabinfo()