/linux-6.1.9/include/trace/events/ |
D | oom.h | 36 unsigned long reclaimable, 42 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check), 48 __field( unsigned long, reclaimable) 59 __entry->reclaimable = reclaimable; 69 __entry->reclaimable, __entry->available, __entry->min_wmark,
|
/linux-6.1.9/drivers/net/ethernet/chelsio/cxgb4vf/ |
D | sge.c | 411 static inline int reclaimable(const struct sge_txq *tq) in reclaimable() function 414 int reclaimable = hw_cidx - tq->cidx; in reclaimable() local 415 if (reclaimable < 0) in reclaimable() 416 reclaimable += tq->size; in reclaimable() 417 return reclaimable; in reclaimable() 434 int avail = reclaimable(tq); in reclaim_completed_tx() 2133 if (reclaimable(&txq->q) && __netif_tx_trylock(txq->txq)) { in sge_tx_timer_cb() 2134 int avail = reclaimable(&txq->q); in sge_tx_timer_cb()
|
/linux-6.1.9/mm/ |
D | page-writeback.c | 1972 unsigned long reclaimable; in wb_over_bg_thresh() local 1988 reclaimable = wb_stat_sum(wb, WB_RECLAIMABLE); in wb_over_bg_thresh() 1990 reclaimable = wb_stat(wb, WB_RECLAIMABLE); in wb_over_bg_thresh() 1992 if (reclaimable > thresh) in wb_over_bg_thresh() 2008 reclaimable = wb_stat_sum(wb, WB_RECLAIMABLE); in wb_over_bg_thresh() 2010 reclaimable = wb_stat(wb, WB_RECLAIMABLE); in wb_over_bg_thresh() 2012 if (reclaimable > thresh) in wb_over_bg_thresh()
|
D | page_alloc.c | 4954 unsigned long reclaimable; in should_reclaim_retry() local 4958 available = reclaimable = zone_reclaimable_pages(zone); in should_reclaim_retry() 4967 trace_reclaim_retry_zone(z, order, reclaimable, in should_reclaim_retry() 5939 unsigned long reclaimable; in si_mem_available() local 5969 reclaimable = global_node_page_state_pages(NR_SLAB_RECLAIMABLE_B) + in si_mem_available() 5971 available += reclaimable - min(reclaimable / 2, wmark_low); in si_mem_available()
|
D | vmscan.c | 1084 int reclaimable = 0, write_pending = 0; in skip_throttle_noprogress() local 1105 reclaimable += zone_reclaimable_pages(zone); in skip_throttle_noprogress() 1109 if (2 * write_pending <= reclaimable) in skip_throttle_noprogress() 6129 bool reclaimable = false; in shrink_node() local 6155 reclaimable = true; in shrink_node() 6226 if (reclaimable) in shrink_node()
|
/linux-6.1.9/tools/testing/selftests/cgroup/ |
D | memcg_protection.m | 10 % reclaim) and then the reclaim starts, all memory is reclaimable, i.e. treated
|
/linux-6.1.9/Documentation/admin-guide/mm/ |
D | concepts.rst | 169 disk, are called `reclaimable`. The most notable categories of the 170 reclaimable pages are page cache and anonymous memory. 181 The process of freeing the reclaimable physical memory pages and
|
D | idle_page_tracking.rst | 65 are not reclaimable, he or she can filter them out using
|
/linux-6.1.9/fs/xfs/ |
D | xfs_icache.c | 1022 long reclaimable = 0; in xfs_reclaim_inodes_count() local 1026 reclaimable += pag->pag_ici_reclaimable; in xfs_reclaim_inodes_count() 1029 return reclaimable; in xfs_reclaim_inodes_count()
|
/linux-6.1.9/drivers/md/ |
D | raid5-cache.c | 1501 sector_t reclaimable; in r5l_do_reclaim() local 1514 reclaimable = r5l_reclaimable_space(log); in r5l_do_reclaim() 1515 if (reclaimable >= reclaim_target || in r5l_do_reclaim() 1524 r5l_reclaimable_space(log) > reclaimable, in r5l_do_reclaim() 1531 if (reclaimable == 0 || !write_super) in r5l_do_reclaim()
|
/linux-6.1.9/Documentation/admin-guide/sysctl/ |
D | vm.rst | 164 and reclaimable pages, the number of pages at which the background kernel 199 and reclaimable pages, the number of pages at which a process which is 231 reclaimable slab objects like dentries and inodes. Once dropped, their 238 To free reclaimable slab objects (includes dentries and inodes):: 333 And on large highmem machines this lack of reclaimable lowmem memory 489 than this percentage of pages in a zone are reclaimable slab pages.
|
/linux-6.1.9/include/net/ |
D | sock.h | 1660 int reclaimable; in sk_mem_reclaim() local 1665 reclaimable = sk->sk_forward_alloc - sk_unused_reserved_mem(sk); in sk_mem_reclaim() 1667 if (reclaimable >= (int)PAGE_SIZE) in sk_mem_reclaim() 1668 __sk_mem_reclaim(sk, reclaimable); in sk_mem_reclaim()
|
/linux-6.1.9/net/mptcp/ |
D | protocol.c | 184 int reclaimable; in mptcp_rmem_uncharge() local 187 reclaimable = msk->rmem_fwd_alloc - sk_unused_reserved_mem(sk); in mptcp_rmem_uncharge() 190 if (unlikely(reclaimable >= PAGE_SIZE)) in mptcp_rmem_uncharge() 191 __mptcp_rmem_reclaim(sk, reclaimable); in mptcp_rmem_uncharge()
|
/linux-6.1.9/Documentation/scheduler/ |
D | sched-deadline.rst | 216 - Umax is the maximum reclaimable utilization (subjected to RT throttling 220 - Uextra is the (per runqueue) extra reclaimable utilization
|
/linux-6.1.9/Documentation/ABI/testing/ |
D | sysfs-kernel-slab | 393 are reclaimable (and grouped by their mobility).
|
/linux-6.1.9/drivers/net/ethernet/chelsio/cxgb4/ |
D | sge.c | 341 static inline int reclaimable(const struct sge_txq *q) in reclaimable() function 362 int reclaim = reclaimable(q); in reclaim_completed_tx() 4304 avail = reclaimable(&q->q); in sge_tx_timer_cb()
|
/linux-6.1.9/Documentation/admin-guide/cgroup-v1/ |
D | memory.rst | 192 Some pages which are never reclaimable and will not be on the LRU 365 box can still run out of non-reclaimable memory.
|
/linux-6.1.9/Documentation/mm/ |
D | unevictable-lru.rst | 438 whole of the huge page, we want the rest of the page to be reclaimable.
|
/linux-6.1.9/Documentation/filesystems/ |
D | proc.rst | 1024 page cache to function well, and that not all reclaimable 1025 slab will be reclaimable, due to items being in use. The
|
/linux-6.1.9/Documentation/admin-guide/ |
D | cgroup-v2.rst | 1167 unprotected reclaimable memory available, OOM killer 1192 memory won't be reclaimed unless there is no reclaimable
|