Search Engines' fundamental protocols

Below is the list of basic elements for Search Engines their usefulness:

Sitemap:

A sitemap is designed to guide Search Engines through all the content of your website, lead them to each piece of information on your site and help them to classify your content. There are a few different forms of sitemap with highlights on different types of content such as news, mobile, images and videos, etc.


XML:

Extensible Markup Language (Recommended Format) is the commonly accepted sitemap format as is extremely convenient for Search Engines to crawl and it can be made by various sitemap generators. It also provides a consistent control on page parameters. However, file sizes required for XML format are huge.

RSS:

Really Simple Syndication or Rich Site Summary is easy to manage as you can code it to be updated automatically when you add new content. However, it is hard to under understand and use because of its updating properties even when it is actually a dialect of XML.

Txt:

Text File is the easiest format to use with one URL per line up to 50000 lines. But this format does not offer the feature to add meta data to pages.

Robots.txt

The robots.txt file is made by the Robots Exclusion Protocol and stored on a website's root directory (e.g., www.google.com/robots.txt). The robots.txt file guide automated web crawlers (search spiders) when they visit your site. Robots.txt give out th location of sitemaps file and crawl-delay parameters. It can help you to give hints on which part of the site you want bots to crawl and which you disallow.

Meta Robots

The meta robots tag creates page-level instructions for Search Engine bots. Thus it should be included in the head section of the HTML document.

Rel="Nofollow"


The rel=nofollow attribute allows you to tell Search Engines not to follow a particular link, but some engines still follow them for discovering new pages. These links certainly offer much less value than their followed counterparts, but are useful in situations where you link to untrusted sources.

Rel="canonical"

Two or more copies of identical content are often shown on your website under different URLs. For instant, the following URLs can all lead to a single homepage:
http://www.SEO-experts.com/
http://www.SEO-experts.com/default.asp
http://SEO-experts.com/
http://seo-experts.com/default.asp
http://SEO-experts /Default.asp

We can identify this by ourselves. However, Search Engines can consider these five URLs as separate pages. When they find out that the content on each of these page are the same, they would eventually devalue your page due to duplicate content. This issue can be solved by the canonical tag as it indicates the Search Engines which page is the singular “authoritative” version.

Sources: References from several reliable SEO forums and websites.

Tags: SEO Experts Services in Singapore

Post a Comment

Note: Only a member of this blog may post a comment.