dm raid: enhance super_sync() to support new superblock members
authorHeinz Mauelshagen <heinzm@redhat.com>
Thu, 19 May 2016 16:49:32 +0000 (18:49 +0200)
committerMike Snitzer <snitzer@redhat.com>
Tue, 14 Jun 2016 21:09:35 +0000 (17:09 -0400)
Add transferring the new takeover/reshape related superblock
members introduced to the super_sync() function:

 - add/move supporting functions

 - add failed devices bitfield transfer functions to retrieve the
   bitfield from superblock format or update it in the superblock

 - add code to transfer all new members

Signed-off-by: Heinz Mauelshagen <heinzm@redhat.com>
Signed-off-by: Mike Snitzer <snitzer@redhat.com>
drivers/md/dm-raid.c

index bc5a53d..2842033 100644 (file)
@@ -214,6 +214,12 @@ static bool _test_flags(uint32_t flags, uint32_t all_flags)
        return (flags & all_flags) ? true : false;
 }
 
+/* Clear (multiple) @flags in @all_flags */
+static void _clear_flags(uint32_t flags, uint32_t *all_flags)
+{
+       *all_flags &= ~flags;
+}
+
 /* Return true if single @flag is set in @*flags, else set it and return false */
 static bool _test_and_set_flag(uint32_t flag, uint32_t *flags)
 {
@@ -1289,31 +1295,54 @@ static void sb_retrieve_failed_devices(struct dm_raid_superblock *sb, uint64_t *
        }
 }
 
+static void sb_update_failed_devices(struct dm_raid_superblock *sb, uint64_t *failed_devices)
+{
+       int i = ARRAY_SIZE(sb->extended_failed_devices);
+
+       sb->failed_devices = cpu_to_le64(failed_devices[0]);
+       while (i--)
+               sb->extended_failed_devices[i] = cpu_to_le64(failed_devices[i+1]);
+}
+
+/*
+ * Synchronize the superblock members with the raid set properties
+ *
+ * All superblock data is little endian.
+ */
 static void super_sync(struct mddev *mddev, struct md_rdev *rdev)
 {
-       int i;
-       uint64_t failed_devices;
+       bool update_failed_devices = false;
+       unsigned int i;
+       uint64_t failed_devices[DISKS_ARRAY_ELEMS];
        struct dm_raid_superblock *sb;
        struct raid_set *rs = container_of(mddev, struct raid_set, md);
 
+       /* No metadata device, no superblock */
+       if (!rdev->meta_bdev)
+               return;
+
+       BUG_ON(!rdev->sb_page);
+
        sb = page_address(rdev->sb_page);
-       failed_devices = le64_to_cpu(sb->failed_devices);
 
-       for (i = 0; i < mddev->raid_disks; i++)
-               if (!rs->dev[i].data_dev ||
-                   test_bit(Faulty, &(rs->dev[i].rdev.flags)))
-                       failed_devices |= (1ULL << i);
+       sb_retrieve_failed_devices(sb, failed_devices);
 
-       memset(sb + 1, 0, rdev->sb_size - sizeof(*sb));
+       for (i = 0; i < rs->raid_disks; i++)
+               if (!rs->dev[i].data_dev || test_bit(Faulty, &rs->dev[i].rdev.flags)) {
+                       update_failed_devices = true;
+                       set_bit(i, (void *) failed_devices);
+               }
+
+       if (update_failed_devices)
+               sb_update_failed_devices(sb, failed_devices);
 
        sb->magic = cpu_to_le32(DM_RAID_MAGIC);
-       sb->compat_features = cpu_to_le32(0);   /* No features yet */
+       sb->compat_features = cpu_to_le32(0); /* Don't set reshape flag yet */
 
        sb->num_devices = cpu_to_le32(mddev->raid_disks);
        sb->array_position = cpu_to_le32(rdev->raid_disk);
 
        sb->events = cpu_to_le64(mddev->events);
-       sb->failed_devices = cpu_to_le64(failed_devices);
 
        sb->disk_recovery_offset = cpu_to_le64(rdev->recovery_offset);
        sb->array_resync_offset = cpu_to_le64(mddev->recovery_cp);
@@ -1321,6 +1350,32 @@ static void super_sync(struct mddev *mddev, struct md_rdev *rdev)
        sb->level = cpu_to_le32(mddev->level);
        sb->layout = cpu_to_le32(mddev->layout);
        sb->stripe_sectors = cpu_to_le32(mddev->chunk_sectors);
+
+       sb->new_level = cpu_to_le32(mddev->new_level);
+       sb->new_layout = cpu_to_le32(mddev->new_layout);
+       sb->new_stripe_sectors = cpu_to_le32(mddev->new_chunk_sectors);
+
+       sb->delta_disks = cpu_to_le32(mddev->delta_disks);
+
+       smp_rmb(); /* Make sure we access most recent reshape position */
+       sb->reshape_position = cpu_to_le64(mddev->reshape_position);
+       if (le64_to_cpu(sb->reshape_position) != MaxSector) {
+               /* Flag ongoing reshape */
+               sb->flags |= cpu_to_le32(SB_FLAG_RESHAPE_ACTIVE);
+
+               if (mddev->delta_disks < 0 || mddev->reshape_backwards)
+                       sb->flags |= cpu_to_le32(SB_FLAG_RESHAPE_BACKWARDS);
+       } else
+               /* Flag no reshape */
+               _clear_flags(cpu_to_le32(SB_FLAG_RESHAPE_ACTIVE|SB_FLAG_RESHAPE_BACKWARDS), &sb->flags);
+
+       sb->array_sectors = cpu_to_le64(mddev->array_sectors);
+       sb->data_offset = cpu_to_le64(rdev->data_offset);
+       sb->new_data_offset = cpu_to_le64(rdev->new_data_offset);
+       sb->sectors = cpu_to_le64(rdev->sectors);
+
+       /* Zero out the rest of the payload after the size of the superblock */
+       memset(sb + 1, 0, rdev->sb_size - sizeof(*sb));
 }
 
 /*