Lines Matching refs:avail
190 unsigned int avail; member
671 ac->avail = 0; in init_arraycache()
727 void *objp = ac->entry[--ac->avail]; in __ac_get_obj()
739 for (i = 0; i < ac->avail; i++) { in __ac_get_obj()
743 ac->entry[i] = ac->entry[ac->avail]; in __ac_get_obj()
744 ac->entry[ac->avail] = objp; in __ac_get_obj()
763 ac->avail++; in __ac_get_obj()
778 objp = ac->entry[--ac->avail]; in ac_get_obj()
802 ac->entry[ac->avail++] = objp; in ac_put_obj()
815 int nr = min3(from->avail, max, to->limit - to->avail); in transfer_objects()
820 memcpy(to->entry + to->avail, from->entry + from->avail -nr, in transfer_objects()
823 from->avail -= nr; in transfer_objects()
824 to->avail += nr; in transfer_objects()
925 if (ac->avail) { in __drain_alien_cache()
935 free_block(cachep, ac->entry, ac->avail, node, list); in __drain_alien_cache()
936 ac->avail = 0; in __drain_alien_cache()
954 if (ac->avail && spin_trylock_irq(&alc->lock)) { in reap_alien()
1001 if (unlikely(ac->avail == ac->limit)) { in __cache_free_alien()
1120 free_block(cachep, nc->entry, nc->avail, node, &list); in cpuup_canceled()
1121 nc->avail = 0; in cpuup_canceled()
1132 shared->avail, node, &list); in cpuup_canceled()
2052 cpu_cache_get(cachep)->avail = 0; in setup_cpu_cache()
2344 free_block(cachep, ac->entry, ac->avail, node, &list); in do_drain()
2347 ac->avail = 0; in do_drain()
2793 BUG_ON(ac->avail > 0 || !n); in cache_alloc_refill()
2842 n->free_objects -= ac->avail; in cache_alloc_refill()
2846 if (unlikely(!ac->avail)) { in cache_alloc_refill()
2856 if (!x && (ac->avail == 0 || force_refill)) in cache_alloc_refill()
2859 if (!ac->avail) /* objects refilled by interrupt? */ in cache_alloc_refill()
2946 if (likely(ac->avail)) { in ____cache_alloc()
2976 kmemleak_erase(&ac->entry[ac->avail]); in ____cache_alloc()
3327 BUG_ON(!batchcount || batchcount > ac->avail); in cache_flusharray()
3334 int max = shared_array->limit - shared_array->avail; in cache_flusharray()
3338 memcpy(&(shared_array->entry[shared_array->avail]), in cache_flusharray()
3340 shared_array->avail += batchcount; in cache_flusharray()
3367 ac->avail -= batchcount; in cache_flusharray()
3368 memmove(ac->entry, &(ac->entry[batchcount]), sizeof(void *)*ac->avail); in cache_flusharray()
3396 if (ac->avail < ac->limit) { in __cache_free()
3635 shared->avail, node, &list); in alloc_kmem_cache_node()
3718 free_block(cachep, ac->entry, ac->avail, node, &list); in __do_tune_cpucache()
3829 if (!ac || !ac->avail) in drain_array()
3835 if (ac->avail) { in drain_array()
3836 tofree = force ? ac->avail : (ac->limit + 4) / 5; in drain_array()
3837 if (tofree > ac->avail) in drain_array()
3838 tofree = (ac->avail + 1) / 2; in drain_array()
3840 ac->avail -= tofree; in drain_array()
3842 sizeof(void *) * ac->avail); in drain_array()
3958 shared_avail += n->shared->avail; in get_slabinfo()