Search our Shops        
 

Sunday, February 04, 2007

Google’s New Sitemap Protocol

The Sitemap Protocol allows you to inform search engine crawlers about URLs on your Web sites that are available for crawling. A Sitemap consists of a list of URLs and may also contain additional information about those URLs, such as when they were last modified, how frequently they change, etc.

Sitemaps are particularly beneficial when users can not reach all areas of a Web site through a browseable interface i.e. users are unable to reach certain pages or regions of a site by following links. For example, any site where certain pages are only accessible via a search form would benefit from creating a Sitemap and submitting it to search engines.This document describes the formats for Sitemap files and also explains where you should post your Sitemap files so that search engines can retrieve them.

Please note that the Sitemap Protocol supplements, but does not replace, the crawl-based mechanisms that search engines already use to discover URLs. By submitting a Sitemap (or Sitemaps) to a search engine, you will help that engine's crawlers to do a better job of crawling your site.Using this protocol does not guarantee that your Web pages will be included in search indexes. In addition, using this protocol may not influence the way your pages are ranked by a search engine.

XML Sitemap Format
The XML Sitemap Format allows you to provide a list of URLs and include additional information about those URLs in your Sitemap. This additional information includes the date the content at that URL last changed, how often that content can be expected to change and how important that URL is relative to other URLs on your site.

The XML Sitemap Format uses the following XML tags:

changefreq
: how frequently the content at the URL is likely to change

lastmod : the time the content at the URL was last modified

loc : the URL location

priority : the priority of the page relative to other pages on the same site

url : this tag encapsulates the first four tags in this list

urlset : this tag encapsulates the first five tags in this list

New "Google Sitemaps" Web Page Feed Program
Today, Google has unveiled a new Google Sitemaps program allowing webmasters and site owners to feed it pages they'd like to have included in Google's web index. Participation is free. Inclusion isn't guaranteed, but Google's hoping the new system will help it better gather pages than traditional crawling alone allows. Feeds also let site owners indicate how often pages change or should be revisited.

How the new feed program will work?
Webmasters create XML files containing the URLs they want crawled, along with optional hints about the URLs such as things like when the page last changed, and the rate of change. They host the Sitemap on their server and tell us where it is. We provide an open-source tool called Sitemap Generator to assist in this process. Eventually, we are hoping webservers will natively support the protocol so there are no extra steps for webmasters. When a Sitemap changes, we support auto-notifying us so we can pick up the newest version.

For more details on Google’s sitemap Protocol visit at www.halfvalue.com and www.halfvalue.co.uk

For more information on Books visit at www.lookbookstores.com

No comments: