SEO and duplicate content, the most important search engines attempt to filter their results by removing all the results whose content is already in their databases. This type of filter is called filter duplicate content.

It is important to understand and identify what the duplicate content. The duplicate content is generally defined as substantial blocks of text that are copied from one site to another. Many webmasters try to use duplicate content to manipulate and influence the results of search engines. The community of SEO continues to debate the legitimacy and existence of duplicate content filters. Most webmasters, they have simply accepted the fact that duplicates content could be a penalty by most search engines.



How a search engine can determine which version is the original content? It is difficult for a search engine called which site is causing a version of content, and websites "innocent" may find themselves penalized or banned for including duplicated content. After analyzing the behavior of search engines, we must assume that they will keep most of the time listed the contents from a source called "trusted". For this, they can refer to the number of links incoming from the source, at the age of the domain, or any factor optimization to determine the reputation of the area that contains the duplicate content. If one of the copies is considered by the search engine as originating from a source "known", then it will rank the page properly, while the other source that might yet be the original content will be penalized .

All representatives of the search engines have made it clear they preferred the unique content. Webmasters who wish to avoid a penalty would do well to follow these tips.

Redirects

If you recast your site, use 301 permanent redirects. It is a great way to route your web traffic from your old pages to new.

Uniqueness

Each page of your site must be unique. The theme of each page of a website, even if the theme is similar to another page, is to be composed of a unique and original.

Multiple Languages

If there are multiple versions of the site into different languages, use different domains for each of them. In this way the search engines will not be satisfied to see that an item is merely translated into different languages ​​and will not be considered duplicate content. Each language version is unique content in the eyes of a search engine.

META tags unique

Each page should have unique Meta tags (TITLE, DESCRIPTION and KEYWORDS).Two pages with the same TITLE can be considered as duplicate content and only one of them will be indexed.

Robots.txt

If you intentionally duplicate content on your site, make sure the presence of a robots.txt file to the root of your site to prevent search engines from indexing does not come any part of your site (or a game you want to keep private).

Membership

If you just use the same product descriptions from your affiliate, exactly as if you had copied its product sheets and your pages will be considered duplicate content (because it's likely that your affiliate be considered a source of confidence).

Copyright

Include a copyright notice at the foot of each page of your website.

Copyscape

If you find that another website has reproduced your content without your consent, please enforce your copyright. Use Copyscape (http://www.copyscape.com) to see which pages have been duplicated.

Then, if you see an offense, enter politely contact the offending site and ask them to make appropriate changes.

If satisfactory changes are not made within a reasonable time, you can get in touch with the web host offender with him and see what he can do. You can also submit a DMCA (Digital Millennium Copyright Act) to Google (http://www.google.com/dmca.html), if the pages are forged in the results of its indexes. The procedure is strict and lengthy, so be sure of the legitimacy of your action.

Avoid duplicate content

Avoid, whenever possible, to create a web page that contains content identical or closely linked to another web page of your site. If for any reason you have two pages with identical content, use robots.txt to block access to one of them to search engines. This will prevent any duplicate content indexed in search results.

In conclusion

Although one can always wonder if the search engines give a penalty or not to duplicate content, however, have all expressed that they had absolutely no desire to offer the same content in their results pages. If, on your website, you do not offer, do still active research to find out if other websites do not violate who reproduces in your content. This proactive approach will help avoid possible sanctions.