Re[2]: [squid-users] Frequent cache rebuilding

From: Andreev Nikita <nik_at_kemsu.ru>
Date: Thu, 22 Jan 2009 12:17:59 +0700

>>> Why does squid eat 100% of processor if the problem is in FS?

> How is your cache_dir defined? aufs (in general) is a better choice
> than ufs, diskd might still have some stability issues under load, and
> coss is a good supplement as a small object cache. Conceivably if Squid
> is set up with a ufs cache_dir mounted as NFS, it's spending a lot of
> time in a wait state, blocked while the I/O completes.

For 6 days uptime:
# vmstat
procs -----------memory---------- ---swap-- -----io---- --system-- -----cpu------
 r b swpd free buff cache si so bi bo in cs us sy id wa st
 2 0 92 104052 235704 2309956 0 0 3 43 24 33 10 16 73 1 0

As you can see system has spent only 1% of CPU time in I/O wait.
(cpu-wa column).

My cache dir directive looks like:
cache_dir ufs /var/spool/squid 16384 64 1024

# vmstat -d
vmstat -d
disk- ------------reads------------ ------------writes----------- -----IO------
       total merged sectors ms total merged sectors ms cur sec
ram0 0 0 0 0 0 0 0 0 0 0
ram1 0 0 0 0 0 0 0 0 0 0
ram2 0 0 0 0 0 0 0 0 0 0
ram3 0 0 0 0 0 0 0 0 0 0
ram4 0 0 0 0 0 0 0 0 0 0
ram5 0 0 0 0 0 0 0 0 0 0
ram6 0 0 0 0 0 0 0 0 0 0
ram7 0 0 0 0 0 0 0 0 0 0
ram8 0 0 0 0 0 0 0 0 0 0
ram9 0 0 0 0 0 0 0 0 0 0
ram10 0 0 0 0 0 0 0 0 0 0
ram11 0 0 0 0 0 0 0 0 0 0
ram12 0 0 0 0 0 0 0 0 0 0
ram13 0 0 0 0 0 0 0 0 0 0
ram14 0 0 0 0 0 0 0 0 0 0
ram15 0 0 0 0 0 0 0 0 0 0
sda 50114 8198 1197972 239114 771044 986524 13061742 1616345 0 1239
sdb 125 1430 2383 100 3 20 184 43 0 0
sdc 547181 13909 6116481 6209599 2893943 6771249 77505040 42580590 0 8027
dm-0 6659 0 143594 45401 528574 0 4228592 1248409 0 269
dm-1 13604 0 408122 82828 883993 0 7071944 3118925 0 677
dm-2 150 0 1132 387 2 0 10 2 0 0
dm-3 36240 0 639146 173982 178529 0 1428232 540632 0 229
dm-4 164 0 1136 610 35 0 76 155 0 0
dm-5 216 0 1240 817 166439 0 332884 262910 0 185
hda 0 0 0 0 0 0 0 0 0 0
fd0 0 0 0 0 0 0 0 0 0 0
md0 0 0 0 0 0 0 0 0 0 0

If it's not an I/O wait problem then what can cause squid to use 100%
of CPU core? I tried to clear cache but after an hour or so squid
began to use as much CPU as usual (~100%).

I'm not sure but maybe it started after we enlarged our outer link
from 2Mbps to 4Mbps.

I will try to move squid cache to local disk but squid works in VMware
Virtual Infrastructure. So if I move any of virtual machine partitions
from shared to local storage I wouldn't have an ability to move squid
VM from one HA cluster node to the other ('cause local partitions on
cluster nodes are different from each other).

Regards,
Nikita.
Received on Thu Jan 22 2009 - 05:18:17 MST

This archive was generated by hypermail 2.2.0 : Fri Jan 23 2009 - 12:00:02 MST