Linuxにおける、lvm2 on RAID1の構築方法。
忘れてしまうので、メモしておきます。
●ディスク追加(20120723) /dev/sda 既存の1台目 2000.3 GB /dev/sdb 既存の2台目 2000.3 GB /dev/sdc 今回追加1台目 2000.3 GB /dev/sdd 今回追加2台目 2000.3 GB 1)ディスクの現状確認 # fdisk -l /dev/sda ----- Disk /dev/sda: 2000.3 GB, 2000398934016 bytes 255 heads, 63 sectors/track, 243201 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sda1 * 64 79 128520 fd Linux raid autodetect /dev/sda2 80 6159 48837600 fd Linux raid autodetect /dev/sda3 6160 243201 1904039865 fd Linux raid autodetect ----- # fdisk -l /dev/sdb ----- Disk /dev/sdb: 2000.3 GB, 2000398934016 bytes 255 heads, 63 sectors/track, 243201 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sdb1 * 64 79 128520 fd Linux raid autodetect /dev/sdb2 80 6159 48837600 fd Linux raid autodetect /dev/sdb3 6160 243201 1904039865 fd Linux raid autodetect ----- ※参考まで。 2)fdiskでスライス。 下記のように、fdisk でスライス。 4096バイトセクタを考慮にいれ、スタートを64とする。 id は fd 。 # fdisk -l /dev/sdc ----- Disk /dev/sdc: 2000.3 GB, 2000398934016 bytes 255 heads, 63 sectors/track, 243201 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sdc1 64 243201 1953005985 fd Linux raid autodetect ----- # fdisk -l /dev/sdd ----- Disk /dev/sdd: 2000.3 GB, 2000398934016 bytes 255 heads, 63 sectors/track, 243201 cylinders Units = cylinders of 16065 * 512 = 8225280 bytes Device Boot Start End Blocks Id System /dev/sdd1 64 243201 1953005985 fd Linux raid autodetect ----- 3)RAID1を追加 # mdadm --create /dev/md3 --level=raid1 --raid-devices=2 /dev/sdc1 /dev/sdd1 ----- mdadm: array /dev/md3 started. ----- # mdadm --detail /dev/md3 ----- /dev/md3: Version : 0.90 Creation Time : Mon Jul 23 13:45:36 2012 Raid Level : raid1 Array Size : 1953005888 (1862.53 GiB 1999.88 GB) Used Dev Size : 1953005888 (1862.53 GiB 1999.88 GB) Raid Devices : 2 Total Devices : 2 Preferred Minor : 3 Persistence : Superblock is persistent Update Time : Mon Jul 23 13:47:55 2012 State : active, resyncing Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Rebuild Status : 1% complete UUID : XXXX Events : 0.2 Number Major Minor RaidDevice State 0 8 33 0 active sync /dev/sdc1 1 8 49 1 active sync /dev/sdd1 ----- # cat /proc/mdstat ----- Personalities : [raid1] md3 : active raid1 sdd1[1] sdc1[0] 1953005888 blocks [2/2] [UU] [>....................] resync = 0.1% (2493440/1953005888) finish=208.6min speed=155840K/sec md0 : active raid1 sda1[1] sdb1[0] 128448 blocks [2/2] [UU] md1 : active raid1 sda2[1] sdb2[0] 48837504 blocks [2/2] [UU] md2 : active raid1 sda3[1] sdb3[0] 1904039744 blocks [2/2] [UU] unused devices:----- 4)mdadm.confに追記 # cd /etc # cat mdadm.conf ----- # mdadm.conf written out by anaconda DEVICE partitions MAILADDR root ARRAY /dev/md0 level=raid1 num-devices=2 uuid=XXXX ARRAY /dev/md1 level=raid1 num-devices=2 uuid=XXXX ARRAY /dev/md2 level=raid1 num-devices=2 uuid=XXXX ----- # cp mdadm.conf mdadm.conf-12072301 # mdadm --detail --scan ----- ARRAY /dev/md2 level=raid1 num-devices=2 metadata=0.90 UUID=XXXX ARRAY /dev/md1 level=raid1 num-devices=2 metadata=0.90 UUID=XXXX ARRAY /dev/md0 level=raid1 num-devices=2 metadata=0.90 UUID=XXXX ARRAY /dev/md3 level=raid1 num-devices=2 metadata=0.90 UUID=XXXX ----- # mdadm --detail --scan >> mdadm.conf # vi mdadm.conf ----- # mdadm.conf written out by anaconda DEVICE partitions MAILADDR root ARRAY /dev/md0 level=raid1 num-devices=2 uuid=XXXX ARRAY /dev/md1 level=raid1 num-devices=2 uuid=XXXX ARRAY /dev/md2 level=raid1 num-devices=2 uuid=XXXX ARRAY /dev/md3 level=raid1 num-devices=2 uuid=XXXX ----- 5)lvm2を使うのでこちらも追加 ・kvm2 ボリューム作成 # pvcreate /dev/md3 Writing physical volume data to disk "/dev/md3" Physical volume "/dev/md3" successfully created # vgcreate -s 32m VolGroup01-kvm2 /dev/md3 Volume group "VolGroup01-kvm2" successfully created # lvcreate -n LogVol00 -l 100%FREE VolGroup01-kvm2 Logical volume "LogVol00" created ・できたか確認 # lvdisplay --- Logical volume --- LV Name /dev/VolGroup01-kvm2/LogVol00 VG Name VolGroup01-kvm2 LV UUID XXXX LV Write Access read/write LV Status available # open 0 LV Size 1.82 TB Current LE 59601 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 253:1 --- Logical volume --- LV Name /dev/VolGroup00/LogVol00 VG Name VolGroup00 LV UUID XXXX LV Write Access read/write LV Status available # open 1 LV Size 1.77 TB Current LE 58106 Segments 1 Allocation inherit Read ahead sectors auto - currently set to 256 Block device 253:0 6)ext3でfsを構築 # mkfs.ext3 /dev/mapper/VolGroup01--kvm2-LogVol00 ----- mke2fs 1.39 (29-May-2006) Filesystem label= OS type: Linux Block size=4096 (log=2) Fragment size=4096 (log=2) 244137984 inodes, 488251392 blocks 24412569 blocks (5.00%) reserved for the super user First data block=0 Maximum filesystem blocks=4294967296 14901 block groups 32768 blocks per group, 32768 fragments per group 16384 inodes per group Superblock backups stored on blocks: 32768, 98304, 163840, 229376, 294912, 819200, 884736, 1605632, 2654208, 4096000, 7962624, 11239424, 20480000, 23887872, 71663616, 78675968, 102400000, 214990848 Writing inode tables: done Creating journal (32768 blocks): done Writing superblocks and filesystem accounting information: done This filesystem will be automatically checked every 28 mounts or 180 days, whichever comes first. Use tune2fs -c or -i to override. ----- 7)仮マウント # mount -t ext3 /dev/mapper/VolGroup01--kvm2-LogVol00 /kvm2 # df -k -T Filesystem Type 1K-blocks Used Available Use% Mounted on /dev/mapper/VolGroup00-LogVol00 ext3 1844389860 190922896 1558266096 11% / /dev/md0 ext3 124387 47709 70256 41% /boot tmpfs tmpfs 16465284 0 16465284 0% /dev/shm /dev/mapper/VolGroup01--kvm2-LogVol00 ext3 1922360144 200164 1824509704 1% /kvm2 8)起動時に自動マウント # cat /etc/fstab ----- /dev/VolGroup00/LogVol00 / ext3 defaults 1 1 /dev/md0 /boot ext3 defaults 1 2 tmpfs /dev/shm tmpfs defaults 0 0 devpts /dev/pts devpts gid=5,mode=620 0 0 sysfs /sys sysfs defaults 0 0 proc /proc proc defaults 0 0 /dev/md1 swap swap defaults 0 0 /dev/VolGroup01-kvm2/LogVol00 /kvm2 ext3 defaults 0 0 ----- ※/dev/VolGroup01-kvm2/LogVol00 を追加 ・マウント確認 # df -k -T Filesystem Type 1K-blocks Used Available Use% Mounted on /dev/mapper/VolGroup00-LogVol00 ext3 1844389860 190922896 1558266096 11% / /dev/md0 ext3 124387 47709 70256 41% /boot tmpfs tmpfs 16465284 0 16465284 0% /dev/shm # mount -a # df -k -T Filesystem Type 1K-blocks Used Available Use% Mounted on /dev/mapper/VolGroup00-LogVol00 ext3 1844389860 190922896 1558266096 11% / /dev/md0 ext3 124387 47709 70256 41% /boot tmpfs tmpfs 16465284 0 16465284 0% /dev/shm /dev/mapper/VolGroup01--kvm2-LogVol00 ext3 1922360144 200164 1824509704 1% /kvm2 あとは、リブートしてみて、マウントされていればOK。 RAID1の再構築をひたすら待つ。 完了。