LARGE single-system Cyrus installs?
Wesley Craig
wes at umich.edu
Sat Oct 6 13:58:36 EDT 2007
Personally, I've seen Solaris bottlenecking on file opens in large
directories. This was a while ago, but it was one of the major
reason we switched to Linux -- the order of magnitude improvement in
directory scale was sure handy for 80-90K users with no quota. The
kind of blocking I'm talking about didn't show up in sar or iostat,
despite being "IO" in nature. Of course it was some sort of
algorithmic inefficiency in the directory data structures, not the
speed of moving data on & off the disk. As a general statement, tho,
finding bottlenecks like the one UCDavis is complaining about is done
by process sampling. No guessing required.
:wes
On 06 Oct 2007, at 05:50, Rob Mueller wrote:
>> The iostat and sar data disagrees with it being an I/O issue.
>>
>> 16 gigs of RAM with about 4-6 of it being used for Cyrus
>> leaves plenty for ZFS caching. Our hardware seemed more than
>> adequate to anyone we described it to.
>>
>> Yes beyond that it's anyone guess.
>
> If it wasn't IO limit related, and it wasn't CPU limit related,
> then there must be some other single resource that things were
> contending for.
>
> My only guess then is it's some global kernel lock or the like.
>
> When the load skyrocketed, it must have shown that lots of
> processes were not in S (sleep) state. Were they in R (run) state
> or D (io wait) state? Since you're on Solaris, you could use DTrace
> to find out what they were actually doing/waiting for...
More information about the Info-cyrus
mailing list