Re: [squid-users] Massive Squid Deployment

From: Chris Woodfield <rekoil_at_semihuman.com>
Date: Thu, 2 Apr 2009 10:05:29 -0400

One performance-specific 2.7 question - I recall hearing mention of an
issue with 2.6 where larger objects held in mem_cache required
exponentially more CPU cycles to serve up (i.e. n cycles for a 4KB
object, n*2 for an 8KB object, n*4 for a 12KB object, etc). Does
anyone know if this issue is still present in 2.7 code?

Thanks,

-C

On Apr 1, 2009, at 12:42 AM, Amos Jeffries wrote:

>>
>> Hello people, I'm having some bottlenecks on our squid deployement
>> and I
>> was
>> wondering if anyone had any recommendations because I'm near out of
>> ideas.
>> Would someone change my architecture or does anyone have any
>> experience
>> with
>> a squid deployment of this size? Basically we are pushing 600mb at
>> 240,000k
>> connections. When we reach speeds on our around that that number,
>> we start
>> seeing slow performance and getting alot of page timeouts. We are
>> running
>> 20
>> squid 2.6 boxes running dsr behind a single foundry gte load
>> balancer. I
>> recently had 18 squid boxes and thought we had a squid bottleneck
>> but no
>> change. I was kinda leaning in the direction of splitting it in
>> half and
>> doing multiple load balancers. Does anyone have any experience
>> pushing
>> this
>> much traffic?
>
> I'd go to 2.7 if it was squid being slow. It has a number of small
> performance boosters missing in 2.6.
>
> But, it does sound like a load balancer bottleneck if you had zero
> change
> from adding two more Squid. The Yahoo and Wikimedia guys are the
> poster
> installs for these types of Squid deployment and they use CARP
> meshing.
>
> PS: We are very interested in getting some real-world benchmarks from
> high-load systems like yours. Can you grab the data needed for
> http://wiki.squid-cache.org/KnowledgeBase/Benchmarks ?
>
> Amos
>
>
>
Received on Thu Apr 02 2009 - 14:05:40 MDT

This archive was generated by hypermail 2.2.0 : Fri Apr 03 2009 - 12:00:01 MDT