[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: Preparing for zfs filesystem



On Mon, 2013-06-17 at 22:37 -0500, Robert G. (Doc) Savage wrote:
> Well, it took about 18-1/2 hours altogether. That's with a hardware RAID
> card with four 15K SCSI3 drives on one channel and five on the other.
> 
> Now they all look like:
> 
>         Disk /dev/sdb: 300.0 GB, 299965284352 bytes
>         255 heads, 63 sectors/track, 36468 cylinders
>         Units = cylinders of 16065 * 512 = 8225280 bytes
>         Sector size (logical/physical): 512 bytes / 512 bytes
>         I/O size (minimum/optimal): 512 bytes / 512 bytes
>         Disk identifier: 0x00000000
> 
> Before doing anything, I needed to change SELinux's setting from
> enforcing to permissive in /dev/selinux/config. This is necessary until
> Red Hat makes ZFS a supported filesystem. To make this change effective
> I had to reboot.
> 
> Because I have two SCSI3 channels, and one of those 300GB hard drives
> (/dev/sda) is boot & root, I have to explicitly declare the zfs pool by
> extracting the info from /dev/disk/by-id/. Dynamically created at boot
> time, these are soft links to /dev/sd[b-j]. Note that when I have a SATA
> drive in the eSATA external "toaster" adapter, it's detected as /dev/sda
> and all the SCSI drives get bumped down one drive letter (/dev/sd[c-k]).
> That's not a problem because the /dev drive mapped to each of these
> links gets bumped too.
> 
>         # zpool create -f pub raidz2 \
>         /dev/disk/by-id/scsi-SAdaptec_0-1_4CABCE14 \
>         /dev/disk/by-id/scsi-SAdaptec_0-2_660FCE14 \
>         /dev/disk/by-id/scsi-SAdaptec_0-3_5A13DE14 \
>         /dev/disk/by-id/scsi-SAdaptec_0-4_DEA3EE14 \
>         /dev/disk/by-id/scsi-SAdaptec_1-0_92C3FE14 \
>         /dev/disk/by-id/scsi-SAdaptec_1-1_551BFE14 \
>         /dev/disk/by-id/scsi-SAdaptec_1-2_62E00E14 \
>         /dev/disk/by-id/scsi-SAdaptec_1-3_49941E14 \
>         /dev/disk/by-id/scsi-SAdaptec_1-4_5B0C1E14
> 
> I'd no sooner hit "Enter" than I had a 1.9T array mounted at /pub. Or
> did I? I pulled up 'man zfs' again and looked at 'zfs create' again.
> Explicitly turning dedup=on sounded like a good idea. So did making it
> shareable via NFS.
> 
>         # zfs create -p -o dedup=on -o sharenfs=on pub
> 
> That's it. I hope. I'm using rsync to migrate my repos from their temp
> home on the 4T drive in /dev/sda1 mounted at /mnt. The drive lights on
> the RAID array are blinking like mad. :-)

rsync jobs have been running all day long restoring from the 4T drive to
the new array. Only about 100GB to go. The new memory is apparently
doing its ZFS caching job because top shows:

Mem:  32883168k total, 32498740k used,   384428k free,   564792k buffers
Swap: 16515064k total,        0k used, 16515064k free, 18365320k cached

--Doc


-
To unsubscribe, send email to majordomo@silug.org with
"unsubscribe silug-discuss" in the body.