Should I Use Software To Create Hundreds Of Virtually Identical Doorway Pages?
There are software programs that generate hundreds of virtually identical pages, which use different keywords, without any human intervention. I have seen such pages. They are ugly and don't make any sense.
I recommend you avoid using such programs. Search engines consider these types of pages as spam and will catch up with you sooner or later.
According to the search engines, 30-40% of all webpages on the internet fall into the category of spam. I bet a large percentage is doorway pages created by software programs. So you see why search engines don't like them?
Search engines can, and do, detect spam doorway pages and are actively removing them from their index.
Beware of Scam Artists That Use Software Generated Doorway Pages
So what do the search engine optimizers who use software generated doorway pages do, when they're removed from the search engines? They go back to their clients and tell them that the search engines have dropped all their pages and they would need to spend more money to create and optimize new pages. Talk about getting ripped off!
How To Hide Doorway Pages From A Search Engine
On a final note, if you create a doorway page for a specific search engine, you do not want other search engines to find and index the page. Because if they do index a page not intended for it and also indexes the page created specifically for it, the two virtually identical pages could set off the spam alarm bell. So how do you avoid this predicament?
The answer is to use a robots.txt file to instruct each search engine spider which directories and pages they are allowed to crawl and index. Read the What Is A Robots.txt File? section for more information.
What's The Difference Between Hallway & Doorway Pages?
Hallway pages are similar to site maps, except that they only contain links to doorway pages. Hallway pages are so-called because they are like hallways leading to many doors. Do not include links to your doorway pages on your site map. Use a hallway page instead.
How To Get Search Engines To Index Dynamically Generated Webpages
Dynamic pages are database driven pages created "on the fly," (in real time) depending on the user's interest or request. Dynamically generated pages are commonly created using programming languages such like ASP (Active Server Pages), Cold Fusion, or Perl (CGI).
Anyone who surfs the world wide web would have come across a dynamically generated website. For example, all search engines dynamically generate a particular page of search results based on the search terms you enter.
There are 3 reasons why search engines do not index dynamically generated pages:
- Dynamically generated pages do not actually exist on the internet until someone submits a request, and since search engine spiders do not submit requests, they may not find all the pages in a site.
- Dynamically generated URLs usually contain question marks (?) or other symbols (&, %, + and $). When search engine spiders do follow dynamically generated URLs, they may only crawl part of the URL excluding everything after the question mark, which usually results in the spiders indexing nothing.
- Many search engine spiders avoid indexing URLs that have a reference to the CGI bin directory to ensure that they do not fall into an endless loop.
How To Turn Dynamically Generated Pages Into Search Engine Friendly Pages
The best way to turn dynamically generated webpages into search engine friendly pages is to use a system that converts your database into static pages whenever it is updated. Alternatively, you may use a software program to convert an entire database driven site into a static one.
3 Tools That Convert Dynamically Generated URLs Into Ones That Search Engines Can Crawl & Index
Use software fixes to convert ASP, Cold Fusion, or Perl generated URLs into ones that search engines can crawl and index. Note that these software fixes do not convert ASP, Cold Fusion, or Perl generated pages into HTML static pages. They simply convert the URLs into HTML versions, so that search engines will index the contents properly.
- A Users Guide to URL Rewriting with the Apache web server - A document that describes how you can use Apache's "mod_rewrite" module to enable you to turn URLs containing query strings into URLs that search engines can index. This guide is very advanced and is not recommended for beginners.
- ASPSpiderBait - An ISAPI Filter program that rewrites URLs on the fly to help search engines find and index every page on a dynamic ASP site. Cost: $100.
http://www.webanalyst.com.au/Products/ASPSpiderBait.htm (no longer available)
- XQASP - Offers a NT IIS 3.0+ C++ or Unix Java filter called XQASP that converts question marks (?) in ASP URLs into forward slashes (/). Cost: $250-$2,100.
http://www.xde.net/xq/tool.xqasp/qx/index.htm (no longer available)
Here is a tool that converts databases into static HTML files:
- DBtoHTML Express - Offers a template-based tool that converts database files into static HTML files. Supports ODBC, Oracle, and SQL databases. Operating Systems: Windows 95, 98, Me, NT and 2000. Cost: $129.