Re: [squid-users] error with file system

From: Pavel Kazlenka <pavel.kazlenka_at_measurement-factory.com>
Date: Wed, 21 Aug 2013 10:36:30 +0300

Hi Carlos,

Please note, that client's requests also spend file descriptors. Use
netstat to find the exact number.
If you use ubuntu you could be interested in this thread too:
http://www.squid-cache.org/mail-archive/squid-users/201212/0276.html

Best wishes,
Pavel

On 08/20/2013 09:57 PM, Carlos Defoe wrote:
> Hello,
>
> Look at this:
>
> 2013/08/20 07:55:26 kid1| ctx: exit level 0
> 2013/08/20 07:55:26 kid1| Attempt to open socket for EUI retrieval
> failed: (24) Too many open files
> 2013/08/20 07:55:26 kid1| comm_open: socket failure: (24) Too many open files
> 2013/08/20 07:55:26 kid1| Reserved FD adjusted from 100 to 64542 due to failures
> 2013/08/20 07:55:26 kid1| WARNING! Your cache is running out of filedescriptors
> 2013/08/20 07:55:26 kid1| comm_open: socket failure: (24) Too many open files
>
> ulimit -n = 65535 (i have configured it in limits.conf myself)
>
> When squid starts, it shows no errors:
>
> 2013/08/20 13:38:11 kid1| Starting Squid Cache version 3.3.8 for
> x86_64-unknown-linux-gnu...
> 2013/08/20 13:38:11 kid1| Process ID 8087
> 2013/08/20 13:38:11 kid1| Process Roles: worker
> 2013/08/20 13:38:11 kid1| With 65535 file descriptors available
>
> running lsof gives no more than 8000 files opened when the problem occurs.
>
> Why should it say "Too many open files"? Do you think SELinux can be
> the cause of this issue?
>
> thanks
Received on Wed Aug 21 2013 - 07:41:01 MDT

This archive was generated by hypermail 2.2.0 : Wed Aug 21 2013 - 12:00:43 MDT