Lines Matching refs:nres

91 	CLST svcn = le64_to_cpu(attr->nres.svcn);  in attr_load_runs()
92 CLST evcn = le64_to_cpu(attr->nres.evcn); in attr_load_runs()
103 run_off = le16_to_cpu(attr->nres.run_off); in attr_load_runs()
328 attr->nres.data_size = cpu_to_le64(rsize); in attr_make_nonresident()
329 attr->nres.valid_size = attr->nres.data_size; in attr_make_nonresident()
451 align <<= attr_b->nres.c_unit; in attr_set_size()
453 old_valid = le64_to_cpu(attr_b->nres.valid_size); in attr_set_size()
454 old_size = le64_to_cpu(attr_b->nres.data_size); in attr_set_size()
455 old_alloc = le64_to_cpu(attr_b->nres.alloc_size); in attr_set_size()
462 attr_b->nres.data_size = cpu_to_le64(new_size); in attr_set_size()
469 svcn = le64_to_cpu(attr_b->nres.svcn); in attr_set_size()
470 evcn = le64_to_cpu(attr_b->nres.evcn); in attr_set_size()
489 svcn = le64_to_cpu(attr->nres.svcn); in attr_set_size()
490 evcn = le64_to_cpu(attr->nres.evcn); in attr_set_size()
505 attr_b->nres.data_size = cpu_to_le64(new_size); in attr_set_size()
585 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_set_size()
587 attr_b->nres.alloc_size = cpu_to_le64(new_alloc_tmp); in attr_set_size()
592 attr_b->nres.data_size = cpu_to_le64(new_size); in attr_set_size()
600 attr_b->nres.data_size = attr_b->nres.alloc_size; in attr_set_size()
610 attr_b->nres.data_size = attr_b->nres.alloc_size; in attr_set_size()
612 attr_b->nres.valid_size = attr_b->nres.data_size; in attr_set_size()
647 svcn = le64_to_cpu(attr->nres.svcn); in attr_set_size()
648 evcn = le64_to_cpu(attr->nres.evcn); in attr_set_size()
662 attr_b->nres.alloc_size = cpu_to_le64((u64)vcn << cluster_bits); in attr_set_size()
663 attr_b->nres.data_size = attr_b->nres.alloc_size; in attr_set_size()
664 attr_b->nres.valid_size = attr_b->nres.alloc_size; in attr_set_size()
705 attr->nres.evcn = cpu_to_le64((u64)vcn - 1); in attr_set_size()
709 attr_b->nres.alloc_size = cpu_to_le64(new_alloc_tmp); in attr_set_size()
712 attr_b->nres.data_size = cpu_to_le64(new_size); in attr_set_size()
714 attr_b->nres.valid_size = in attr_set_size()
715 attr_b->nres.data_size; in attr_set_size()
718 le64_to_cpu(attr_b->nres.data_size)) in attr_set_size()
719 attr_b->nres.data_size = in attr_set_size()
720 attr_b->nres.alloc_size; in attr_set_size()
722 le64_to_cpu(attr_b->nres.valid_size)) in attr_set_size()
723 attr_b->nres.valid_size = in attr_set_size()
724 attr_b->nres.alloc_size; in attr_set_size()
728 le64_sub_cpu(&attr_b->nres.total_size, in attr_set_size()
769 if (attr_b->nres.valid_size != valid) { in attr_set_size()
770 attr_b->nres.valid_size = valid; in attr_set_size()
790 new_alloc = le64_to_cpu(attr_b->nres.alloc_size); in attr_set_size()
858 asize = le64_to_cpu(attr_b->nres.alloc_size) >> sbi->cluster_bits; in attr_data_get_block()
864 clst_per_frame = 1u << attr_b->nres.c_unit; in attr_data_get_block()
870 svcn = le64_to_cpu(attr_b->nres.svcn); in attr_data_get_block()
871 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_data_get_block()
884 svcn = le64_to_cpu(attr->nres.svcn); in attr_data_get_block()
885 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
941 total_size = le64_to_cpu(attr_b->nres.total_size) + in attr_data_get_block()
949 attr_b->nres.total_size = cpu_to_le64(total_size); in attr_data_get_block()
957 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
992 sbi, le64_to_cpu(attr_b->nres.alloc_size)); in attr_data_get_block()
993 CLST evcn = le64_to_cpu(attr->nres.evcn); in attr_data_get_block()
1023 svcn = le64_to_cpu(attr->nres.svcn); in attr_data_get_block()
1024 evcn = le64_to_cpu(attr->nres.evcn); in attr_data_get_block()
1035 attr->nres.svcn = cpu_to_le64(next_svcn); in attr_data_get_block()
1044 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_data_get_block()
1151 svcn = le64_to_cpu(attr->nres.svcn); in attr_load_runs_vcn()
1152 evcn = le64_to_cpu(attr->nres.evcn); in attr_load_runs_vcn()
1159 ro = le16_to_cpu(attr->nres.run_off); in attr_load_runs_vcn()
1252 wof_size = le64_to_cpu(attr->nres.data_size); in attr_wof_frame_info()
1362 clst_frame = 1u << attr->nres.c_unit; in attr_is_frame_compressed()
1390 alen = bytes_to_cluster(ni->mi.sbi, le64_to_cpu(attr->nres.alloc_size)); in attr_is_frame_compressed()
1471 total_size = le64_to_cpu(attr_b->nres.total_size); in attr_allocate_frame()
1473 svcn = le64_to_cpu(attr_b->nres.svcn); in attr_allocate_frame()
1474 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_allocate_frame()
1475 data_size = le64_to_cpu(attr_b->nres.data_size); in attr_allocate_frame()
1492 svcn = le64_to_cpu(attr->nres.svcn); in attr_allocate_frame()
1493 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_allocate_frame()
1550 attr_b->nres.total_size = cpu_to_le64(total_size); in attr_allocate_frame()
1557 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_allocate_frame()
1592 sbi, le64_to_cpu(attr_b->nres.alloc_size)); in attr_allocate_frame()
1593 CLST evcn = le64_to_cpu(attr->nres.evcn); in attr_allocate_frame()
1623 svcn = le64_to_cpu(attr->nres.svcn); in attr_allocate_frame()
1624 evcn = le64_to_cpu(attr->nres.evcn); in attr_allocate_frame()
1635 attr->nres.svcn = cpu_to_le64(next_svcn); in attr_allocate_frame()
1644 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_allocate_frame()
1660 valid_size = le64_to_cpu(attr_b->nres.valid_size); in attr_allocate_frame()
1662 attr_b->nres.valid_size = cpu_to_le64(valid_size); in attr_allocate_frame()
1699 data_size = le64_to_cpu(attr_b->nres.data_size); in attr_collapse_range()
1700 alloc_size = le64_to_cpu(attr_b->nres.alloc_size); in attr_collapse_range()
1704 total_size = le64_to_cpu(attr_b->nres.total_size); in attr_collapse_range()
1705 mask = (sbi->cluster_size << attr_b->nres.c_unit) - 1; in attr_collapse_range()
1744 svcn = le64_to_cpu(attr_b->nres.svcn); in attr_collapse_range()
1745 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_collapse_range()
1763 svcn = le64_to_cpu(attr->nres.svcn); in attr_collapse_range()
1764 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_collapse_range()
1770 attr->nres.svcn = cpu_to_le64(svcn - len); in attr_collapse_range()
1771 attr->nres.evcn = cpu_to_le64(evcn1 - 1 - len); in attr_collapse_range()
1773 le->vcn = attr->nres.svcn; in attr_collapse_range()
1799 attr->nres.svcn = cpu_to_le64(vcn); in attr_collapse_range()
1801 le->vcn = attr->nres.svcn; in attr_collapse_range()
1810 next_svcn = le64_to_cpu(attr->nres.evcn) + 1; in attr_collapse_range()
1833 u16 roff = le16_to_cpu(attr->nres.run_off); in attr_collapse_range()
1882 svcn = le64_to_cpu(attr->nres.svcn); in attr_collapse_range()
1883 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_collapse_range()
1903 attr_b->nres.alloc_size = cpu_to_le64(alloc_size - bytes); in attr_collapse_range()
1904 attr_b->nres.data_size = cpu_to_le64(data_size); in attr_collapse_range()
1905 attr_b->nres.valid_size = cpu_to_le64(min(valid_size, data_size)); in attr_collapse_range()
1908 attr_b->nres.total_size = cpu_to_le64(total_size); in attr_collapse_range()
1967 alloc_size = le64_to_cpu(attr_b->nres.alloc_size); in attr_punch_hole()
1968 total_size = le64_to_cpu(attr_b->nres.total_size); in attr_punch_hole()
1975 mask = (sbi->cluster_size << attr_b->nres.c_unit) - 1; in attr_punch_hole()
2002 svcn = le64_to_cpu(attr_b->nres.svcn); in attr_punch_hole()
2003 evcn1 = le64_to_cpu(attr_b->nres.evcn) + 1; in attr_punch_hole()
2021 svcn = le64_to_cpu(attr->nres.svcn); in attr_punch_hole()
2022 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_punch_hole()
2064 svcn = le64_to_cpu(attr->nres.svcn); in attr_punch_hole()
2065 evcn1 = le64_to_cpu(attr->nres.evcn) + 1; in attr_punch_hole()
2069 attr_b->nres.total_size = cpu_to_le64(total_size); in attr_punch_hole()