Lines Matching refs:allocated_stack
611 int allocated_slots = state->allocated_stack / BPF_REG_SIZE; in is_spi_bounds_valid()
909 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in print_verifier_state()
1035 size_t n = src->allocated_stack / BPF_REG_SIZE; in copy_stack_state()
1042 dst->allocated_stack = src->allocated_stack; in copy_stack_state()
1059 size_t old_n = state->allocated_stack / BPF_REG_SIZE, n = size / BPF_REG_SIZE; in grow_stack_state()
1068 state->allocated_stack = size; in grow_stack_state()
2728 for (j = 0; j < func->allocated_stack / BPF_REG_SIZE; j++) { in mark_all_scalars_precise()
2852 if (i >= func->allocated_stack / BPF_REG_SIZE) { in __mark_chain_precision()
4611 min_valid_off = -state->allocated_stack; in check_stack_slot_within_bounds()
5117 if (state->allocated_stack <= slot) in check_stack_range_initialized()
11252 for (i = 0; i < st->allocated_stack / BPF_REG_SIZE; i++) { in clean_func_state()
11459 for (i = 0; i < old->allocated_stack; i++) { in stacksafe()
11474 if (i >= cur->allocated_stack) in stacksafe()
11664 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE && in propagate_liveness()
11665 i < parent->allocated_stack / BPF_REG_SIZE; i++) { in propagate_liveness()
11700 for (i = 0; i < state->allocated_stack / BPF_REG_SIZE; i++) { in propagate_precision()
11949 for (i = 0; i < frame->allocated_stack / BPF_REG_SIZE; i++) { in is_state_visited()