Usually not in a reliable manner. Problem is that many major web sites
makes HTML content uncacheable. Reasons vary from customized content
and/or marketing needing to know the exact hit count, to plain
stupidity.
But yes, in most cases it is doable. You have two options:
a) Offline mirroring using wget and some web server. Drawback is that
the URL the user has to browse to is the URL to your local mirrors, not
the original URLs.
b) Offline caching using Squid or another cache. Probably requires
aggressive tweaking of refresh_pattern rules to defeat the published
uncacheable state of the pages, and might also require offline_mode in
Squid to further enforce this.
-- Henrik Nordstrom Squid hacker Davida Schiff wrote: > > Hi Henrik and thanks for replying, > > What I want to do is just cache the initial web page from various sites > like yahoo, apple, Netscape, etc on my server, then go off line and connect > 2 other PCs to this server and have them be able to view these 1 page sites > (no internet access) that were cached. Is this possible? > > Hope your weekend was good, > > Davida > > -----Original Message----- > From: Henrik Nordstrom [mailto:hno@hem.passagen.se] > Sent: Sunday, July 09, 2000 2:46 PM > To: Davida Schiff > Cc: squid-users@squid-cache.org > Subject: Re: Squid and wget > > Before I even attempt to answer this question, I must ask you to define > further what you mean? > > Access to pages cached from earlier sessions, or access to preselected > material? > > Access to pages cached from earlier sessions is Squid in offline_mode, > but not unfortunately not many sites can be cached in a way making that > useable.. (to much "dynamic" content today). > > Access to preselected web pages is mirroring, for which as I said wget > is a good tool, in combination with Apache or another web server for > giving the clients access to the mirrored content. Beware however that > many web site owners dislike this way of using their content. > > -- > Henrik Nordstrom > Squid hacker > > Davida Schiff wrote: > > > > Hi Henrik, > > > > Do you have any suggestions as to what I can use (Linux Platform) in order > > to access stored (cached) web pages on my server by the clients when they > > are not connected to the internet? Any help would be appreciated. > > > > Thanks, > > > > Davida > > > > -----Original Message----- > > From: Henrik Nordstrom [mailto:hno@hem.passagen.se] > > Sent: Thursday, July 06, 2000 2:36 PM > > To: Davida Schiff > > Cc: squid-users@ircache.net > > Subject: Re: Squid and wget > > > > What you describe sounds like the job of wget mirroring into a catalog > > which is then made available using Apache or another HTTP server. > > > > Squid is more useful if you want the clients to be able to interactively > > retreive contents from the internet. > > > > -- > > Henrik Nordstrom > > Squid Hacker > > > > Davida Schiff wrote: > > > > > > Hi, > > > > > > I have 3 notebooks: 1 server (running Squid and Apache) and 2 clients. I > > > would like to use Squid to store several web pages (from the internet) > and > > > have them available to the client notebooks when not attached to an > > outside > > > network. I have wget installed and can get pages from the web. My > > questions > > > are: where do these pages go and should I put them in another location > so > > > that my clients can access them? > > > > > > Thanks > > > > > > DavidaReceived on Mon Jul 10 2000 - 15:18:18 MDT
This archive was generated by hypermail pre-2.1.9 : Tue Dec 09 2003 - 16:54:24 MST