ZFS doing insane I/O reads

Eric Luyten Eric.Luyten at vub.ac.be
Mon Feb 27 05:46:30 EST 2012


On Mon, February 27, 2012 11:10 am, Ram wrote:
> I just deployed zfs on my newer cyrus servers.
> These servers get less than 2000 mails per hour and around 400
> concurrent pop/imap connections
>
>
> I have seen that even if there is no incoming pop or imap connection
> still there is large amount of READ happenning on the zfs partitions. Is this
> normal behaviour for an imap server. Iostat shows sometimes upto 2000 TPS
>
>
> The reads are infact more than 10x of what writes are. I am afraid I
> will be trashing the  harddisk. Do I need to tune ZFS specially for cyrus  ?
>
>
>
> This is the typical zpool iostat output
>
>
> zpool iostat 1
> pool        alloc   free   read  write   read  write
> ----------  -----  -----  -----  -----  -----  -----
> imap         145G   655G    418     58  18.0M  1.78M
> imap         146G   654G    258    118  8.28M   960K
> imap         145G   655G    447    146  19.4M  4.37M
> imap         145G   655G    413     32  19.4M  1.46M
> imap         145G   655G    339      4  14.8M  20.0K
> imap         145G   655G    341     40  15.7M   755K
> imap         145G   655G    305     10  15.0M  55.9K
> imap         145G   655G    328     12  14.8M   136K


Ram,

We have a single Cyrus server about ten times as busy as yours with four ZFS
pools (EMC Celerra iSCSI SAN) for message stores ; all the databases, quota
and seen information are on an internal server SSD based (mirror) pool.
We also have a few GB of SSD based ZIL (synchronous write cache) per pool.


Here is our 'zpool iostat 1' output :

               capacity     operations    bandwidth
pool        alloc   free   read  write   read  write
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T     22     32   422K   286K
cpool2      1.18T  2.66T     29     45   578K   459K
cpool3      1.00T  2.84T     24     34   456K   314K
cpool4       993G  2.87T     25     35   455K   328K
ssd         7.49G  22.3G      4     35  17.2K   708K
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T     45     16   670K   759K
cpool2      1.18T  2.66T     47     25   565K   603K
cpool3      1.00T  2.84T     33     13   410K   483K
cpool4       993G  2.87T     12      8   525K   244K
ssd         7.49G  22.3G     13    210  49.4K  10.8M
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T     20     22  77.9K  2.15M
cpool2      1.18T  2.66T     25      4   937K   128K
cpool3      1.00T  2.84T     20     91   324K  11.0M
cpool4       993G  2.87T     17     13   844K  83.9K
ssd         7.49G  22.3G      6    237  20.0K  20.9M
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T      0      0   1023      0
cpool2      1.18T  2.66T     12     21   146K  1.26M
cpool3      1.00T  2.84T      8     26  46.5K  2.28M
cpool4       993G  2.87T     11      4   353K  24.0K
ssd         7.49G  22.3G     17    135  99.4K  8.12M
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T      4      0  80.9K  4.00K
cpool2      1.18T  2.66T      7      6   133K  28.0K
cpool3      1.00T  2.84T      6      0  16.5K  4.00K
cpool4       993G  2.87T      4      4   149K  20.0K
ssd         7.49G  22.3G      9     76  51.0K  4.24M
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T     12      0   269K  4.00K
cpool2      1.18T  2.66T     19      0   327K  4.00K
cpool3      1.00T  2.84T      7      3  11.0K  16.0K
cpool4       993G  2.87T      5     95   167K  11.4M
ssd         7.49G  22.3G      4    226  17.5K  25.2M
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T     14     20   311K  1.22M
cpool2      1.18T  2.66T     19     15  85.4K  1.39M
cpool3      1.00T  2.84T      6      6  5.49K  40.0K
cpool4       993G  2.87T      4     15  17.0K  1.70M
ssd         7.49G  22.3G      6    151  21.5K  13.1M
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T     56     15  2.11M   559K
cpool2      1.18T  2.66T     13      7  18.5K  32.0K
cpool3      1.00T  2.84T      5      4  54.4K   392K
cpool4       993G  2.87T     17      2  66.4K   136K
ssd         7.49G  22.3G      6    109  45.9K  8.29M
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T     38     19   228K  1.89M
cpool2      1.18T  2.66T     29     11   160K   300K
cpool3      1.00T  2.84T      4      4  11.5K  24.0K
cpool4       993G  2.87T      9      8  31.5K  56.0K
ssd         7.49G  22.3G     12    150  46.0K  12.1M
----------  -----  -----  -----  -----  -----  -----
cpool1       901G  2.96T     32      1   106K   256K
cpool2      1.18T  2.66T     46      5   692K  95.9K
cpool3      1.00T  2.84T      7     13   189K   324K
cpool4       993G  2.87T      4      0  29.0K  4.00K
ssd         7.49G  22.3G     25     96   149K  8.08M
----------  -----  -----  -----  -----  -----  -----


Q1 : How much RAM does your server have ?
     Solaris 10 uses all remaining free RAM as ZFS read cache.
     We have 72 GB of RAM in our server.
     Or are you using ZFS on e.g. BSD ?


Q2 : What is your 'fsstat zfs 1' output ?



Regards,
Eric.




More information about the Info-cyrus mailing list