Creating a search engine friendly website structure

Feb 10
08:17

2006

Andrew Allfrey

Andrew Allfrey

  • Share this article on Facebook
  • Share this article on Twitter
  • Share this article on Linkedin

Creating a good web site structure is crucial for both your visitors and achieving a good rank with the major search engines. Most visitors who visit your site will be visiting for a particular reason. Either they want to find out more information about the products or services that you offer, or alternatively, they may wish to buy a product having carried out some initial research.

mediaimage

You may have only one opportunity to inform them about your services and have them contact you; therefore your site should be designed in such a way that encourages and makes contacting you easy to do.

In this regard,Creating a search engine friendly website structure Articles search engine optimizationshould not get in the way of having a clean and friendly site. The following design rules can help you build a site that will be both visitor friendly and ranked well by the major search engines. 

Home Pages 

The home page is the most important page of a Web site from an SEOstandpoint. If you are only going to optimize one page of your site, both for search engines and potential clients, it should be the home page. Even if your goal is to optimize your entire site, you should devote a good deal of your time and effort to optimizing your home page. 

In the search engine indexing process, the home page of a domain receives more weight than its internal pages. The home page is also the one page that will be seen by most of your visitors. Because of this, it is important to list and include your practice areas and other keywords on your home page, as well as provide links to information in your internal pages. The home page must be text rich and not contain extensive graphics, flash or video components. The page must also load quickly, and any JavaScript and style sheets should be placed in external files.  

You should never have a splash page if your objective is effective search engine optimization. Most visitors will skip by the splash page, and it may even keep a search engine crawler or spider from indexing your the other Web pages of your site.  

A Search Engine and Visitor Friendly Site Map 

You should always have a site map that contains links to your web site's various sections or pages. From an SEO perspective, having a link to your site map on your home page -- if not all pages of your Web site -- is important. Listing the links to all the major sections and/or pages of your site in one, easy-to-find place makes it much easier for a search engine crawler to grab and index your entire site. From a user-perspective, it is worthwhile having the site map linked on every page so that users can easily go there if they are having problems finding specific information. 

Multiple Sub-domains 

If you own a particular domain name then you can create an unlimited number of domain names that use the original name as a suffix. Once you register a domain, such as myfirm.com with a registrar, you are free to create as many sub-domains as you wish,  e.g., widgets.myfirm.com 

You will still need your system administrator or ISP to set these sub-domains up, and they may charge a setup cost, but there are no domain name registration costs. Each of the default/home pages of these domains will get the same preferential value as your default domain name or home page from search engines. If your Web developer can provide sub-domains for you, it may help increase the rankings of some of your Web pages in search engine indices, without having to purchase additional domains from one of the registries. 

Directory Depth and Search Engine Indexing and Rankings  

Directory depth refers to the number of directories in a URL. When your web site content and files are grouped into a directory structure, your URL will reflect this structure (similar to how files are grouped and stored into broad categories in folders in your computer). The deeper you go "down" the directory structure, the more directories your URL will show for a particular file.

For example: 

www.mydomain.com/index.html has a directory depth of 0

www.mydomain.com/products/index.html has a directory depth of 1 (products is a directory)

www.mydomain.com/products/widgets/index.html has a directory depth of 2 (products is a directory, and widgets is a subdirectory) 

Most of the major search engines will index a site regardless of its directory depth. Yet, some search engine algorithms will discount the weight given to pages that are multiple levels down in a web site's directory structure. As a general rule, it is advisable not to have extraneous directories that are not used in your web site's structure. For example, if you have an extraneous directory called "html" where you place all of your HTML files, you should remove it if possible. This can normally be accomplished by asking your webmaster to change the web server configuration to point to the "html" directory/folder as the default folder for the Web site. 

Domain, Directory and File Names  

You can place keywords in the name of a domain, secondary domain, directory or file. Some search engines give extra weight to these file name words. Here are a few tips to help you do this:  

Separate keywords with dashes (e.g., "blue-widgets"). Do not run keywords together, as they will be seen as one word by the search engine, and most people use spaces when searching for terms (e.g., people search for "blue widgets" not "bluewidgets".  

Do not use underscores to separate words (e.g., "blue_widgets"). If you separate your keywords with underscores, Google (and possibly other search engines) will not separate your keywords. The search engines will index the word for your file name as one long text string and not as separate words.  

Dynamically Generated URLs 

Dynamically generated URLs come from Web pages that are generated from "scripts," which put together page content from a database or multiple files "on the fly" by a server, when a user asks to see a Web page. These pages are different from those that statically sit on a server. If you see a URL containing a question mark ("?"), it likely means that page has been pulled from a database to be displayed in your browser. Other symbols you might see that indicate this are: &, $, =, +, and %. 

Some search engine crawlers, such as Google's, will index dynamically-generated pages, but others will not index pages when they see those symbols, especially the question mark. Some search engines will only index a dynamically-generated page that is linked from a static page, i.e., the search engine will not follow links from dynamically-generated pages. 

If you are going to have dynamic pages on your web site, it is important that, at a minimum, they not have session variables presented to the search engines. Session variables are unique strings of text in a URL that are generated each time a user visits a site. For example, Amazon has session variables in the URL when a user visits its site. Session variables cause the same page to appear with constantly changing URLs, and search engines will not index dynamic pages that deliver the same content over and over again with a changing URL. 

A Web server may be configured to have the URL of a dynamic page appear as a static HTML page, in which case spiders will index the Web page as though it is static and will follow links on the page. 

An additional issue that dynamic pages create, even if they appear in the URLs as static pages, is that they are often slower-loading than plain HTML pages, because of the data processing required by the server to put them together on the fly. The slow processing of these pages can lead to fewer pages being indexed. If a page takes too long to load, a spider will move on without indexing that page. As Google notes "Because our web crawler can easily overwhelm and crash sites serving dynamic content, we limit the amount of dynamic pages we index". 

Hypertext Links 

Hypertext links are the text and links placed between the link tags (<a href="http://www.mydomain.com/"> link text </a>), which take internet users to different sections and pages of your web site, as well as to other Web sites.  

Search engines like hypertext links as navigation tools for two reasons:  

(1) Spiders and crawlers can easily read the text placed between link tags; and  

(2) Spiders or crawlers can easily follow links that are listed. Remember that if a spider or crawler has a hard time finding or following the links you've placed on your pages; it may give up indexing your site. 

Drop-Down or Pull-Down Menus 

Using a drop-down or pull-down menu helps you save on screen space when you have a lot of material that needs to be displayed on your pages, or a complicated multi-level navigational structure. In general one should use XHTML and CSS style sheets to present drop-down or pull-down menu navigation so that the search engines will still see the Href links when indexing your Web page.  

If this type of navigational scheme uses a CGI script or JavaScript it is not good for search engine crawlers, and the links may not be indexed. If you decide to use drop-down or pull-down menus using JavaScript or a CGI script as your main navigational format, remember to include a secondary form of navigation to allow search engine spiders to follow links to the sections of your Web site easily. You can normally do this by linking to a page that contains links to the choices presented in the drop-down or pull-down menu. 

Image Maps 

Most search engines will not follow links placed in image maps. Thus, if you are going to use an image map it is best to also provide text links as well. Ideally, all of the text links should be included on the same page, but if needed, from a design perspective, put one link in text at the bottom of the page that takes your users to a list of links that were on the image map. This will also help those not loading graphics find the content they are looking for. 

Fast Downloads of HTML, XHTML, Site Layout and Navigation  

It is important that the HTML or XHTML of a web page load into a user's browser window quickly. The file size of a web page itself, excluding images, plug-ins or other files that are a part of the page, should to be under 40k when at all possible. One way to keep the file size of a page small and remove extra text from the top of the page is to keep JavaScript and style sheets in a separate file that is called by and integrated into the Web page at the time a user wants the page.  

These JavaScript and style sheet files can be called by multiple or even all of the pages on the site and then are cached by the browser. Having separate files for style sheets and JavaScript also helps when making site modifications -- by making changes to one file you can implement changes across the entire site.