Lines Matching refs:cache_alloc
811 struct entry_alloc cache_alloc; member
988 return to_cblock(get_index(&mq->cache_alloc, e)); in infer_cblock()
1045 unsigned threshold_level = allocator_empty(&mq->cache_alloc) ? in update_promote_levels()
1155 nr_free = from_cblock(mq->cache_size) - mq->cache_alloc.nr_allocated; in free_target_met()
1238 if (allocator_empty(&mq->cache_alloc)) { in queue_promotion()
1255 e = alloc_entry(&mq->cache_alloc); in queue_promotion()
1263 free_entry(&mq->cache_alloc, e); in queue_promotion()
1286 if (!allocator_empty(&mq->cache_alloc) && fast_promote) in should_promote()
1464 struct entry *e = get_entry(&mq->cache_alloc, in __complete_background_work()
1477 free_entry(&mq->cache_alloc, e); in __complete_background_work()
1486 free_entry(&mq->cache_alloc, e); in __complete_background_work()
1521 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in __smq_set_clear_dirty()
1564 e = alloc_particular_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_load_mapping()
1582 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_invalidate_mapping()
1590 free_entry(&mq->cache_alloc, e); in smq_invalidate_mapping()
1597 struct entry *e = get_entry(&mq->cache_alloc, from_cblock(cblock)); in smq_get_hint()
1612 r = to_cblock(mq->cache_alloc.nr_allocated); in smq_residency()
1760 init_allocator(&mq->cache_alloc, &mq->es, in __smq_create()