ZFS for Cyrus IMAP storage
Rudy Gevaert
Rudy.Gevaert at UGent.be
Sat May 5 17:20:32 EDT 2007
Vincent Fox wrote:
>
> I originally brought up the ZFS question.
>
> We seem to have arrived at a similar solution after much
> experimentation. Meaning using ZFS for the things it does well already,
> and leveraging proven hardware to fill in the weak spots.
>
> I have a pair of Sun 3510FC arrays we have exported 2 RAID5 LUNs (5
> disks each) with one on primary and the other on second controller. This
> is to exploit the active-active controller feature of the 3510FC. We
> are also doing multipathing through a pair of SAN fabric switches.
>
> On top of that we then use ZFS to join a LUN from each array into a
> mirror pair, and then add the other pair as well. I guess you could call
> it RAID 5+1+0. This architecture allows us to add more storage to the
> pool while online, by adding more 3510FC array pairs to the pool.
>
> Performance in benchmarking (Bonnie++ etc.) has shown to be little
> different from turning them into JBOD and doing everything with ZFS.
> Behavior is more predictable to me since I know that the 3510 firmware
> knows how to rebuild a RAID5 set using the assigned spare drive in that
> array. With ZFS I see no way to specify which disk is assigned as spare
> to a particular set of disks, which could mean a spare is pulled from
> another array.
>
> It's pretty nifty to be able to walk into the machine room and flip off
> the power to an entire array and things keep working without a blip.
> It's not the most efficient usage of disk space but with performance &
> safety this promising for an EMAIL SERVER it will definitely be
> welcome. I dread the idea of silent data corruption or long fsck time
> on a 1+ TB mail spool which ZFS should save us from. I have atime=off
> and compression=on. Our setup is slightly different from yours in that
> we are clustering 2 T2000 with 8GB RAM each, and we are currently
> setting up Solaris Cluster 3.2 software in failover configuration so we
> can patch without downtime.
>
> Thanks for the idea about daily snapshots for recovering recent data, I
> like that idea a lot. I'll tinker around with it I wonder if there'd be
> much penalty to upping the snapshots to every 8 hours. Depends on how
> much churn there is in your mail spool I suppose.
>
> This system should move into production later this month. We have
> 70,000 accounts that we'll begin a long and slow migration from our
> UW-IMAP pool of servers. We have an existing Perdition proxy server
> setup, which will allow us to migrate users transparently. Hopefully
> I'll have more good things to say about it sometime thereafter.
>
Are you going to do this with "1" perdition server? Make sure you have
compiled perdition with /dev/urandom, or an other sort of non blocking
entropy providing device :)
Rudy
More information about the Info-cyrus
mailing list