Re: [squid-users] throughput limitation from cache

From: Peter Smith <peter.smith@dont-contact.us>
Date: Thu, 12 Jan 2006 18:45:49 -0600

Peter Smith wrote:

> Richard Mittendorfer wrote:
>
>> Hello *,
>>
>> When downloading a cached file from the local squid, I just get about
>> 250 - 280kB/s. Even on localhost. Is this a limitation with diskd
>> serving files from cache or some intern limit? I also tried aufs, but
>> didn't get a better rate. I found a thread here about this, but it got
>> more into a diskd/aufs discussion :) and didn't provide a solution or
>> explains it. It doesn't look like it's disk/system related,
>> vmstat/top doesn't show
>> high load.
>>
>> linux 2.6 celeron 600MHz
>> squid-2.5.STABLE12 Debian GNU/Linux
>> 2 x cache_dir diskd 1G 8 128 on 2 x SCSI reiserfs
>> 32M cache_mem, acls, .. need more info?
>>
>> TIA, ritch
>>
>>
>
> I too have experience this. I believe this is because Squid has
> built-in safeguards to keep users from flooding the disk and to even
> the load across all disks for all users. Another reason is Squid
> handles all requests in sequential order--client requests, server
> requests, and memory/disk cache reads/writes.. It is like a
> virtualized network handler or VM. It really doesn't want one client
> to kill all others..
>
> HTH,
> Peter

Btw, it still does have a lot to do with diskd/aufs... Read here
http://squid-docs.sourceforge.net/latest/html/x220.html#AEN318

Peter
Received on Thu Jan 12 2006 - 17:45:55 MST

This archive was generated by hypermail pre-2.1.9 : Wed Feb 01 2006 - 12:00:01 MST