Re: Caching pages...again
This WebDNA talk-list message is from 2001
It keeps the original formatting.
numero = 36925
interpreted = N
texte = Glenn Busbin wrote:> > >> >> >This has nothing to do with caching.> > Google says it sure has to do with caching on proxy servers.> ROBOTS != caching serversMost HTTP caching servers operate in transparent mode; by this I mean that the client requests a page through the server and the server seeswhether it has that page cached. If you are unlucky (or gullible) enough to use AOL, all of your accesses are through a caching server.Robots or spiders, on the other hand, do not (as a rule) keep a cacheof the pages they visit (think of the storage requirements). They only apply whatever indexing algorythm they use to the page and storethe indexed information for later searching.Google, unlike most search engines, does keep a copy of the originalpage available for viewing _when the original page is no longeravailable_. Here is a representative search from Google:University Press of America, Inc.: Catalog/Advanced SearchClick here for details on Web Discount. University Press of America,Inc. Catalog/ Advanced Search. Click Here for Search Instructions ... www.univpress.com/Catalog/ - 18k - Cached - Similar pages ------If you click on the first line, you get the actual site; if you click on the hyperlink (unlined) labled Cached you get a copy of the page as it was when the spider walked the site.This is not a feature pf caching proxy servers at all; this is a feature of Google specifically. QED> >> >> >Use MIME headers correctly if you don't want any caching done.> >> > Will all servers obey MIME headers for cache rules?> The combination of the two headers I described in the other thread isthe only way _I_ know of to consistently defeat (or correctly apply)caching proxy services. If you do not use both, you are letting yourself in for whatever heuristic the caching server uses to determinewhether the page should be fresh or cached.John--John PeacockDirector of Information Research and TechnologyRowman & Littlefield Publishing Group4720 Boston WayLanham, MD 20706301-459-3366 x.5010fax 301-429-5747-------------------------------------------------------------This message is sent to you because you are subscribed to the mailing list
.To unsubscribe, E-mail to: To switch to the DIGEST mode, E-mail to Web Archive of this list is at: http://search.smithmicro.com/
Associated Messages, from the most recent to the oldest:
Glenn Busbin wrote:> > >> >> >This has nothing to do with caching.> > Google says it sure has to do with caching on proxy servers.> ROBOTS != caching serversMost HTTP caching servers operate in transparent mode; by this I mean that the client requests a page through the server and the server seeswhether it has that page cached. If you are unlucky (or gullible) enough to use AOL, all of your accesses are through a caching server.Robots or spiders, on the other hand, do not (as a rule) keep a cacheof the pages they visit (think of the storage requirements). They only apply whatever indexing algorythm they use to the page and storethe indexed information for later searching.Google, unlike most search engines, does keep a copy of the originalpage available for viewing _when the original page is no longeravailable_. Here is a representative search from Google:University Press of America, Inc.: Catalog/Advanced SearchClick here for details on Web Discount. University Press of America,Inc. Catalog/ Advanced Search. Click Here for Search Instructions ... www.univpress.com/Catalog/ - 18k - Cached - Similar pages ------If you click on the first line, you get the actual site; if you click on the hyperlink (unlined) labled Cached you get a copy of the page as it was when the spider walked the site.This is not a feature pf caching proxy servers at all; this is a feature of Google specifically. QED> >> >> >Use MIME headers correctly if you don't want any caching done.> >> > Will all servers obey MIME headers for cache rules?> The combination of the two headers I described in the other thread isthe only way _I_ know of to consistently defeat (or correctly apply)caching proxy services. If you do not use both, you are letting yourself in for whatever heuristic the caching server uses to determinewhether the page should be fresh or cached.John--John PeacockDirector of Information Research and TechnologyRowman & Littlefield Publishing Group4720 Boston WayLanham, MD 20706301-459-3366 x.5010fax 301-429-5747-------------------------------------------------------------This message is sent to you because you are subscribed to the mailing list .To unsubscribe, E-mail to: To switch to the DIGEST mode, E-mail to Web Archive of this list is at: http://search.smithmicro.com/
John Peacock
DOWNLOAD WEBDNA NOW!
Top Articles:
Talk List
The WebDNA community talk-list is the best place to get some help: several hundred extremely proficient programmers with an excellent knowledge of WebDNA and an excellent spirit will deliver all the tips and tricks you can imagine...
Related Readings:
Press Release hit the NewsWire!!! (1997)
WebCat2b13MacPlugin - nested [xxx] contexts (1997)
WebDelivery downloads alias, not original ? (1997)
RE: Automatic Forwarding using WebCat (1997)
vs (1997)
Time/date formatting %X doesn't work ... (1997)
Securing/hiding database file (2000)
Active Server Code... (1998)
WebCat2b13 Mac plugin - [sendmail] and checkboxes (1997)
Generating unique SKU from [cart] - Still Stumped... (1997)
WC2.0 Memory Requirements (1997)
Resume Catalog ? (1997)
Bug or syntax error on my part? (1997)
Can I invoke an ssi plugin from within a webcat page (1997)
WebCatalog for Postcards ? (1997)
Search Questions (2000)
WC2f3 (1997)
WC1.6 to WC2 date formatting (1997)
Question re: FlushDatabases (1997)
random in arizona (2003)