Lines Matching refs:ipv4_dst_ops
143 struct dst_ops ipv4_dst_ops = { variable
283 unsigned int dst_entries = atomic_read(&ipv4_dst_ops.entries); in rt_cache_stat_get_info()
548 atomic_read(&ipv4_dst_ops.entries) < ip_rt_max_size) { in rt_garbage_collect()
554 goal = atomic_read(&ipv4_dst_ops.entries) - in rt_garbage_collect()
557 if (equilibrium < ipv4_dst_ops.gc_thresh) in rt_garbage_collect()
558 equilibrium = ipv4_dst_ops.gc_thresh; in rt_garbage_collect()
559 goal = atomic_read(&ipv4_dst_ops.entries) - equilibrium; in rt_garbage_collect()
562 goal = atomic_read(&ipv4_dst_ops.entries) - equilibrium; in rt_garbage_collect()
569 equilibrium = atomic_read(&ipv4_dst_ops.entries) - goal; in rt_garbage_collect()
625 atomic_read(&ipv4_dst_ops.entries), goal, i); in rt_garbage_collect()
628 if (atomic_read(&ipv4_dst_ops.entries) < ip_rt_max_size) in rt_garbage_collect()
632 if (atomic_read(&ipv4_dst_ops.entries) < ip_rt_max_size) in rt_garbage_collect()
642 atomic_read(&ipv4_dst_ops.entries) < ipv4_dst_ops.gc_thresh) in rt_garbage_collect()
646 atomic_read(&ipv4_dst_ops.entries), goal, rover); in rt_garbage_collect()
900 rt = dst_alloc(&ipv4_dst_ops); in ip_rt_redirect()
1331 rth = dst_alloc(&ipv4_dst_ops); in ip_route_input_mc()
1539 rth = dst_alloc(&ipv4_dst_ops); in ip_route_input_slow()
1614 rth = dst_alloc(&ipv4_dst_ops); in ip_route_input_slow()
1994 rth = dst_alloc(&ipv4_dst_ops); in ip_route_output_slow()
2379 data: &ipv4_dst_ops.gc_thresh,
2572 ipv4_dst_ops.kmem_cachep = kmem_cache_create("ip_dst_cache", in ip_rt_init()
2577 if (!ipv4_dst_ops.kmem_cachep) in ip_rt_init()
2610 ipv4_dst_ops.gc_thresh = (rt_hash_mask + 1); in ip_rt_init()