RE: [squid-users] Same Domain Caching

From: Dave Raven <dave@dont-contact.us>
Date: Mon, 22 Oct 2007 10:04:13 +0200

I had considered doing that as well - using a redirector to match on
youtube.com/get_video but then I'll need to save those to the disk and
manage them myself as opposed to using squids method. Is there a successful
use of ETags for something like this / is it worth looking at? As I
understand it that's essentially there point (if used correctly) ?

Something else that might be worth looking into (for me) is whats after the
? - I suppose that if it in some way identifies the video I could rewrite
the url to be the one I know is cache (e.g. the first ever request for that)

Thanks again for the help
Dave

-----Original Message-----
From: Adrian Chadd [mailto:adrian@creative.net.au]
Sent: Monday, October 22, 2007 9:38 AM
To: Dave Raven
Cc: squid-users@squid-cache.org
Subject: Re: [squid-users] Same Domain Caching

On Mon, Oct 22, 2007, Dave Raven wrote:
> Hi all,
> Is there a way to assume that anything under a certain domain is
> similar across servers? For example, www.youtube.com videos come from
> various servers --
>
> 1191839044.533 53841 10.10.108.250 TCP_MISS/200 1770189 GET
> http://sjc-v180.sjc.youtube.com/get_video? - DIRECT/64.15.120.171
video/flv
> 1140
> 1191917902.678 610481 10.10.100.198 TCP_MISS/200 9465378 GET
> http://v194.youtube.com/get_video? - DIRECT/208.65.154.167 video/flv 1068
>
> And so on.. if I force caching on video/flv files it should make for good
> caching of the content, but 100 users viewing a video could all go to
> different servers to get it - meaning instead of getting 100 hits I get
100x
> the content size in cache?
>
> Is there a way around this?

There's been a few attempts at it but noone yet seems to have implemented
what I've suggested. Someone recently posted how he does it via log
post-processing
and rewriter rules.

Adrian
Received on Mon Oct 22 2007 - 02:04:34 MDT

This archive was generated by hypermail pre-2.1.9 : Thu Nov 01 2007 - 13:00:01 MDT