Searched refs:dev_sectors (Results 1 – 14 of 14) sorted by relevance
248 sector_t dev_sectors; member440 return rs->md.recovery_cp < rs->md.dev_sectors; in rs_is_recovering()694 rdev->sectors = mddev->dev_sectors; in rs_set_rdev_sectors()1614 if (ds < rs->md.dev_sectors) { in _check_data_dev_sectors()1628 sector_t array_sectors = sectors, dev_sectors = sectors; in rs_set_dev_and_array_sectors() local1649 dev_sectors *= rs->raid10_copies; in rs_set_dev_and_array_sectors()1650 if (sector_div(dev_sectors, data_stripes)) in rs_set_dev_and_array_sectors()1653 array_sectors = (data_stripes + delta_disks) * dev_sectors; in rs_set_dev_and_array_sectors()1657 } else if (sector_div(dev_sectors, data_stripes)) in rs_set_dev_and_array_sectors()1662 array_sectors = (data_stripes + delta_disks) * dev_sectors; in rs_set_dev_and_array_sectors()[all …]
62 sector_t dev_sectors; /* temp copy of member
1294 mddev->dev_sectors = ((sector_t)sb->size) * 2; in super_90_validate()1435 sb->size = mddev->dev_sectors / 2; in super_90_sync()1546 if (num_sectors && num_sectors < rdev->mddev->dev_sectors) in super_90_rdev_size_change()1804 mddev->dev_sectors = le64_to_cpu(sb->size); in super_1_validate()2013 sb->size = cpu_to_le64(mddev->dev_sectors); in super_1_sync()2178 if (num_sectors && num_sectors < rdev->mddev->dev_sectors) in super_1_rdev_size_change()2420 (mddev->dev_sectors == 0 || rdev->sectors < mddev->dev_sectors)) { in bind_rdev_to_array()2429 mddev->dev_sectors = rdev->sectors; in bind_rdev_to_array()2632 if ((mddev->dev_sectors != le64_to_cpu(sb->size)) || in does_sb_need_changing()3283 + mddev->dev_sectors > rdev->sectors) in new_offset_store()[all …]
3321 return mddev->dev_sectors - sector_nr; in raid10_sync_request()3329 max_sector = mddev->dev_sectors; in raid10_sync_request()3939 sectors = conf->dev_sectors; in raid10_size()3969 conf->dev_sectors = size << conf->geo.chunk_shift; in calc_sectors()4098 calc_sectors(conf, mddev->dev_sectors); in setup_conf()4112 conf->prev.stride = conf->dev_sectors; in setup_conf()4286 mddev->dev_sectors = conf->dev_sectors; in raid10_run()4379 if (sectors > mddev->dev_sectors && in raid10_resize()4385 mddev->dev_sectors = conf->dev_sectors; in raid10_resize()4411 mddev->dev_sectors = size; in raid10_takeover_raid0()
291 return mddev->dev_sectors; in faulty_size()
334 return mddev->dev_sectors; in multipath_size()
2683 max_sector = mddev->dev_sectors; in raid1_sync_request()2965 return mddev->dev_sectors; in raid1_size()3232 if (sectors > mddev->dev_sectors && in raid1_resize()3233 mddev->recovery_cp > mddev->dev_sectors) { in raid1_resize()3234 mddev->recovery_cp = mddev->dev_sectors; in raid1_resize()3237 mddev->dev_sectors = sectors; in raid1_resize()
346 sector_t dev_sectors; /* used size of member
6350 BUG_ON((mddev->dev_sectors & in reshape_request()6471 if (last_sector >= mddev->dev_sectors) in reshape_request()6472 last_sector = mddev->dev_sectors - 1; in reshape_request()6537 sector_t max_sector = mddev->dev_sectors; in raid5_sync_request()6578 sector_t rv = mddev->dev_sectors - sector_nr; in raid5_sync_request()7348 sectors = mddev->dev_sectors; in raid5_size()8003 mddev->dev_sectors &= ~((sector_t)mddev->chunk_sectors - 1); in raid5_run()8004 mddev->resync_max_sectors = mddev->dev_sectors; in raid5_run()8420 if (sectors > mddev->dev_sectors && in raid5_resize()8421 mddev->recovery_cp > mddev->dev_sectors) { in raid5_resize()[all …]
659 rdev->sectors = mddev->dev_sectors; in raid0_takeover_raid45()
253 sboff < (doff + mddev->dev_sectors + PAGE_SIZE / SECTOR_SIZE)) in __write_sb_page()262 if (doff + mddev->dev_sectors > sboff) in __write_sb_page()
345 __u64 dev_sectors; member
1406 u64 dev_sectors = qc->dev->n_sectors; in ata_scsi_verify_xlat() local1437 if (block >= dev_sectors) in ata_scsi_verify_xlat()1439 if ((block + n_block) > dev_sectors) in ata_scsi_verify_xlat()
223 return p->dev_sectors >> ilog2(p->chunk_sectors); in ublk_get_nr_zones()523 set_capacity(ub->ub_disk, p->dev_sectors); in ublk_dev_param_basic_apply()