Lines Matching refs:atomic_read
195 mb_assert(!(ce->e_used || ce->e_queued || atomic_read(&ce->e_refcnt))); in __mb_cache_entry_forget()
215 if (!(ce->e_used || ce->e_queued || atomic_read(&ce->e_refcnt))) { in __mb_cache_entry_release()
263 if (ce->e_used || ce->e_queued || atomic_read(&ce->e_refcnt)) in mb_cache_shrink_scan()
270 if (ce->e_used || ce->e_queued || atomic_read(&ce->e_refcnt) || in mb_cache_shrink_scan()
299 atomic_read(&cache->c_entry_count)); in mb_cache_shrink_count()
300 count += atomic_read(&cache->c_entry_count); in mb_cache_shrink_count()
410 atomic_read(&ce->e_refcnt)) in mb_cache_shrink()
420 atomic_read(&ce->e_refcnt) || in mb_cache_shrink()
430 atomic_read(&ce->e_refcnt))); in mb_cache_shrink()
473 atomic_read(&ce->e_refcnt))); in mb_cache_destroy()
478 if (atomic_read(&cache->c_entry_count) > 0) { in mb_cache_destroy()
481 atomic_read(&cache->c_entry_count)); in mb_cache_destroy()
506 if (atomic_read(&cache->c_entry_count) >= cache->c_max_entries) { in mb_cache_entry_alloc()
517 atomic_read(&ce->e_refcnt)) in mb_cache_entry_alloc()
528 atomic_read(&ce->e_refcnt) || in mb_cache_entry_alloc()
538 atomic_read(&ce->e_refcnt))); in mb_cache_entry_alloc()