forked from luck/tmp_suning_uos_patched
[PATCH] md: Only checkpoint expansion progress occasionally
Instead of checkpointing at each stripe, only checkpoint when a new write would overwrite uncheckpointed data. Block any write to the uncheckpointed area. Arbitrarily checkpoint at least every 3Meg. Signed-off-by: Neil Brown <neilb@suse.de> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
parent
f67055780c
commit
b578d55fdd
@ -1762,8 +1762,9 @@ static int make_request(request_queue_t *q, struct bio * bi)
|
||||
for (;logical_sector < last_sector; logical_sector += STRIPE_SECTORS) {
|
||||
DEFINE_WAIT(w);
|
||||
int disks;
|
||||
|
||||
|
||||
retry:
|
||||
prepare_to_wait(&conf->wait_for_overlap, &w, TASK_UNINTERRUPTIBLE);
|
||||
if (likely(conf->expand_progress == MaxSector))
|
||||
disks = conf->raid_disks;
|
||||
else {
|
||||
@ -1771,6 +1772,13 @@ static int make_request(request_queue_t *q, struct bio * bi)
|
||||
disks = conf->raid_disks;
|
||||
if (logical_sector >= conf->expand_progress)
|
||||
disks = conf->previous_raid_disks;
|
||||
else {
|
||||
if (logical_sector >= conf->expand_lo) {
|
||||
spin_unlock_irq(&conf->device_lock);
|
||||
schedule();
|
||||
goto retry;
|
||||
}
|
||||
}
|
||||
spin_unlock_irq(&conf->device_lock);
|
||||
}
|
||||
new_sector = raid5_compute_sector(logical_sector, disks, disks - 1,
|
||||
@ -1779,7 +1787,6 @@ static int make_request(request_queue_t *q, struct bio * bi)
|
||||
(unsigned long long)new_sector,
|
||||
(unsigned long long)logical_sector);
|
||||
|
||||
prepare_to_wait(&conf->wait_for_overlap, &w, TASK_UNINTERRUPTIBLE);
|
||||
sh = get_active_stripe(conf, new_sector, disks, pd_idx, (bi->bi_rw&RWA_MASK));
|
||||
if (sh) {
|
||||
if (unlikely(conf->expand_progress != MaxSector)) {
|
||||
@ -1877,6 +1884,7 @@ static sector_t sync_request(mddev_t *mddev, sector_t sector_nr, int *skipped, i
|
||||
*/
|
||||
int i;
|
||||
int dd_idx;
|
||||
sector_t writepos, safepos, gap;
|
||||
|
||||
if (sector_nr == 0 &&
|
||||
conf->expand_progress != 0) {
|
||||
@ -1887,15 +1895,36 @@ static sector_t sync_request(mddev_t *mddev, sector_t sector_nr, int *skipped, i
|
||||
return sector_nr;
|
||||
}
|
||||
|
||||
/* Cannot proceed until we've updated the superblock... */
|
||||
wait_event(conf->wait_for_overlap,
|
||||
atomic_read(&conf->reshape_stripes)==0);
|
||||
mddev->reshape_position = conf->expand_progress;
|
||||
/* we update the metadata when there is more than 3Meg
|
||||
* in the block range (that is rather arbitrary, should
|
||||
* probably be time based) or when the data about to be
|
||||
* copied would over-write the source of the data at
|
||||
* the front of the range.
|
||||
* i.e. one new_stripe forward from expand_progress new_maps
|
||||
* to after where expand_lo old_maps to
|
||||
*/
|
||||
writepos = conf->expand_progress +
|
||||
conf->chunk_size/512*(conf->raid_disks-1);
|
||||
sector_div(writepos, conf->raid_disks-1);
|
||||
safepos = conf->expand_lo;
|
||||
sector_div(safepos, conf->previous_raid_disks-1);
|
||||
gap = conf->expand_progress - conf->expand_lo;
|
||||
|
||||
mddev->sb_dirty = 1;
|
||||
md_wakeup_thread(mddev->thread);
|
||||
wait_event(mddev->sb_wait, mddev->sb_dirty == 0 ||
|
||||
kthread_should_stop());
|
||||
if (writepos >= safepos ||
|
||||
gap > (conf->raid_disks-1)*3000*2 /*3Meg*/) {
|
||||
/* Cannot proceed until we've updated the superblock... */
|
||||
wait_event(conf->wait_for_overlap,
|
||||
atomic_read(&conf->reshape_stripes)==0);
|
||||
mddev->reshape_position = conf->expand_progress;
|
||||
mddev->sb_dirty = 1;
|
||||
md_wakeup_thread(mddev->thread);
|
||||
wait_event(mddev->sb_wait, mddev->sb_dirty == 0 ||
|
||||
kthread_should_stop());
|
||||
spin_lock_irq(&conf->device_lock);
|
||||
conf->expand_lo = mddev->reshape_position;
|
||||
spin_unlock_irq(&conf->device_lock);
|
||||
wake_up(&conf->wait_for_overlap);
|
||||
}
|
||||
|
||||
for (i=0; i < conf->chunk_size/512; i+= STRIPE_SECTORS) {
|
||||
int j;
|
||||
@ -2322,6 +2351,7 @@ static int run(mddev_t *mddev)
|
||||
|
||||
if (conf->expand_progress != MaxSector) {
|
||||
printk("...ok start reshape thread\n");
|
||||
conf->expand_lo = conf->expand_progress;
|
||||
atomic_set(&conf->reshape_stripes, 0);
|
||||
clear_bit(MD_RECOVERY_SYNC, &mddev->recovery);
|
||||
clear_bit(MD_RECOVERY_CHECK, &mddev->recovery);
|
||||
@ -2610,6 +2640,7 @@ static int raid5_reshape(mddev_t *mddev, int raid_disks)
|
||||
conf->previous_raid_disks = conf->raid_disks;
|
||||
conf->raid_disks = raid_disks;
|
||||
conf->expand_progress = 0;
|
||||
conf->expand_lo = 0;
|
||||
spin_unlock_irq(&conf->device_lock);
|
||||
|
||||
/* Add some new drives, as many as will fit.
|
||||
|
@ -217,6 +217,9 @@ struct raid5_private_data {
|
||||
|
||||
/* used during an expand */
|
||||
sector_t expand_progress; /* MaxSector when no expand happening */
|
||||
sector_t expand_lo; /* from here up to expand_progress it out-of-bounds
|
||||
* as we haven't flushed the metadata yet
|
||||
*/
|
||||
int previous_raid_disks;
|
||||
|
||||
struct list_head handle_list; /* stripes needing handling */
|
||||
|
Loading…
Reference in New Issue
Block a user