Re: how do we turn off staticpages page creation?

This WebDNA talk-list message is from

2003


It keeps the original formatting.
numero = 50118
interpreted = N
texte = >Set a robot.txt file in that directory and set whatever pages you do not >want crawled. That will work for good bots, but there are those that look for the names of pages and ignore the DISALLOW tag. That's why it's best to have no reference at all, either on a template or the robots.txt file, to anything you don't want visited.The ^*#&@^$! spammer's address harvesters will crawl anything they can get to while searching for email addresses. Make a honeypot page that does nothing but record hits and mention it only in the robots.txt file with a DISALLOW tag. I'll bet a nickle it gets hit.Glenn------------------------------------------------------------- This message is sent to you because you are subscribed to the mailing list . To unsubscribe, E-mail to: To switch to the DIGEST mode, E-mail to Web Archive of this list is at: http://webdna.smithmicro.com/ Associated Messages, from the most recent to the oldest:

    
  1. Re: how do we turn off staticpages page creation? (Kenneth Grome 2003)
  2. Re: how do we turn off staticpages page creation? (Lucas Rockwell 2003)
  3. Re: how do we turn off staticpages page creation? (WebCat @ Inkblot Media 2003)
  4. Re: how do we turn off staticpages page creation? (Glenn Busbin 2003)
  5. Re: how do we turn off staticpages page creation? (WebCat @ Inkblot Media 2003)
  6. Re: how do we turn off staticpages page creation? (Chris List Recipient 2003)
  7. Re: how do we turn off staticpages page creation? (dbrooke 2003)
  8. Re: how do we turn off staticpages page creation? (Charles Kline 2003)
  9. Re: how do we turn off staticpages page creation? (Charles Kline 2003)
  10. how do we turn off staticpages page creation? (Lucas Rockwell 2003)
>Set a robot.txt file in that directory and set whatever pages you do not >want crawled. That will work for good bots, but there are those that look for the names of pages and ignore the DISALLOW tag. That's why it's best to have no reference at all, either on a template or the robots.txt file, to anything you don't want visited.The ^*#&@^$! spammer's address harvesters will crawl anything they can get to while searching for email addresses. Make a honeypot page that does nothing but record hits and mention it only in the robots.txt file with a DISALLOW tag. I'll bet a nickle it gets hit.Glenn------------------------------------------------------------- This message is sent to you because you are subscribed to the mailing list . To unsubscribe, E-mail to: To switch to the DIGEST mode, E-mail to Web Archive of this list is at: http://webdna.smithmicro.com/ Glenn Busbin

DOWNLOAD WEBDNA NOW!

Top Articles:

Talk List

The WebDNA community talk-list is the best place to get some help: several hundred extremely proficient programmers with an excellent knowledge of WebDNA and an excellent spirit will deliver all the tips and tricks you can imagine...

Related Readings:

webcat, osx, and includes (2001) WebCat2b13MacPlugIn - [include] doesn't allow creator (1997) Showif, Hideif reverse logic ? (1997) Classified (1999) 2.0Beta Command Ref (can't find this instruction) (1997) Upload and WebCat (1998) Error: Too many nested [xxx] contexts (1997) Quickbooks import (2006) Caching [include] files ... (1997) Showing Results Alphabetically in Columns (top to bottom) (2002) Re:Help name our technology! (1997) [Fwd: Rotating Banners ... (was LinkExchange)] (1997) Signal Raised (1997) ACGI processing for .html (1997) Emailer setup (1997) Country & Ship-to address & other fields ? (1997) HELP WITH DATES (1997) Security Issues and WebCommerce Solution (1997) [isfile] ? (1997) Showif for mulitple variations (1997)