Re: [squid-users] throughput limitation from cache

From: Peter Smith <peter.smith@dont-contact.us>
Date: Thu, 12 Jan 2006 18:37:40 -0600

Richard Mittendorfer wrote:

>Hello *,
>
>When downloading a cached file from the local squid, I just get about
>250 - 280kB/s. Even on localhost. Is this a limitation with diskd
>serving files from cache or some intern limit? I also tried aufs, but
>didn't get a better rate. I found a thread here about this, but it got
>more into a diskd/aufs discussion :) and didn't provide a solution or
>explains it.
>It doesn't look like it's disk/system related, vmstat/top doesn't show
>high load.
>
>linux 2.6 celeron 600MHz
>squid-2.5.STABLE12 Debian GNU/Linux
>2 x cache_dir diskd 1G 8 128 on 2 x SCSI reiserfs
>32M cache_mem, acls, .. need more info?
>
>TIA, ritch
>
>

I too have experience this. I believe this is because Squid has built-in
safeguards to keep users from flooding the disk and to even the load
across all disks for all users. Another reason is Squid handles all
requests in sequential order--client requests, server requests, and
memory/disk cache reads/writes.. It is like a virtualized network
handler or VM. It really doesn't want one client to kill all others..

HTH,
Peter
Received on Thu Jan 12 2006 - 17:37:46 MST

This archive was generated by hypermail pre-2.1.9 : Wed Feb 01 2006 - 12:00:01 MST