Skip to main content

Avoid Duplicate Content Penalties

Avoid Duplicate Content Penalties

Large search engines attempt to filter their search results by removing any
results that duplicate the content of other search results. Such filtering
is referred to as "duplicate content penalty".

It is important to understand and identify what "duplicate content" actually
is. Duplicate content is generally defined as substantive blocks of text
that are copied from one site to another. Some webmasters try to use
duplicated content in an attempt to manipulate and influence search engine
rankings. The search community still occasionally debates the legitimacy and
existence of duplicate content filters, but whether they exist today, or
will exist tomorrow, is really irrelevant. Most webmasters have simply
accepted the fact that the duplicate content penalty is currently enforced
by at least some of the major search engines.

With that in mind, how does the search engine determine which version of the
content is the original, and which is duplicated? It is difficult for the
search engine to tell which website is responsible for the original version
of any content, and some innocent websites might find themselves penalized
or banned for including duplicated content. After analyzing the behavior of
search engines, it is safe to assume that the search engines will often
retain the content listing from what it considers to be the most 'trusted'
source. They may look at the number of incoming related links, the age of
the domain, or any other SEO factors that reinforce the reputation of the
domain that contains the duplicated content. If one of the 'copies' is
considered by the search engine to be from a reputable source, they my find
themselves ranking well, while the actual source of the 'original' version
may find themselves unjustly banned or penalized.

Representatives from the major search engines have all made it clear that
they prefer search engines that contain unique content. Webmasters who want
to avoid any current or future bans will do well to follow these simple
guidelines in order to avoid duplicate content penalties:

1. Redirects

If you redesign your website, use permanent 301 redirects. Redirects are a
legitimate way of routing web traffic.

2. Unique

Each page within a website should be unique. The focus of each page on a
website, even if it's similar to the theme of another page, must contain
unique and original content.

3. Multi-Language

If there are multiple language versions of a website, consider using a
different domain for different versions; search engines do not view an
article translated into a variety of foreign languages as being duplicated
content - each language version is unique content in the eyes of the search
engine.

4. Unique Meta Tags

Each web page should contain unique meta tags.

5. Robots.txt

If you do have intentional duplicate content on your website, be sure to
have a "robots.txt" file for your site to prevent the search engines from
indexing the areas with duplicated content (or any areas of the website that
you wish to remain private, for that matter).

6. Affiliate Twist

If you are promoting products or services using an affiliate program, use
unique and distinctive product descriptions and web copy. If you simply use
the same descriptions provided by the product owner or service provider, it's
very likely that your copy could be viewed as duplicated content.

7. Copyright

Include a copyright notice on your website.

8. Enforce

If you discover that another website is scraping your unique web content and
replicating it, enforce your copyright! Use CopyScape at
http://www.copyscape.com/ , or use their "copy sentry" service to receive
notification of any infractions. If you discover a copyright violation,
contact the website and politely request appropriate changes.

If the changes are not made in a reasonable and satisfactory amount of time,
contact the ISP (web host) of the infringing site, and file a DMCA complaint
with Google http://www.google.com/dmca.html .

9. Avoid Identical Content

Do everything you can to avoid serving a web page that contains content
identical or closely related to another page. If for some reason you have
two pages that contain identical content, use a robots.txt to block the
search engines from spidering one version of the page.

Other Tools:

Duplicate Page Checker - http://www.webconfs.com/similar-page-checker.php

While it may still be debatable whether all the major search engines
currently employ a duplicate content penalty, all have made it abundantly
clear that they do not have any desire to provide search results that rehash
the same content over and over. Actively avoid any potential penalties by
taking a proactive approach to building unique content.


I'm protected by SpamBrave
http://www.spambrave.com/

Comments

Popular posts from this blog

What Is Internet Website Content?

What Is Internet Website Content? Content is made up of multiple elements, and is primarily the; * On-page visible text * Images and image Alt text * Anchor text in hyperlinks to internal or external pages * Hyperlink titles in links and menus * The descriptive Title and Description meta-data In the context of Google, a picture is NOT worth a thousand words! Moreover, words must be accessible, not embedded in images or Flash movies, JavaScript, slide shows etc. In 15 years as an SEO consultant, if there's one common denominator evident on websites, it's that there is a profound reluctance to expend time, money, and creative energy on unique text content. Brevity is the watchword - economical use of words is encouraged by design, branding and marketing advisers! * The branding gurus want you to use the textual equivalent of sound bites - bullet points and short sentences! * The website designers want the entire content of the page to be above ...

How to Write Web Copy

Actionable tips for software developers writing web copy. Scan Web site visitors read websites very differently than they might read a book or a newspaper. Web visitors scan the text, rather than reading each and every word. As a result, the web copy should be designed to be easily scannable. That is not to say the copy should not be well written, but it should be broken into small "chunks" so that the visitor can easily scan it and take away the main idea. White Space Avoid dense copy. Copy should be broken into readable, digestible "chunks" and surrounded by a good amount of white space. Font Type Font size matters. Avoid using micro fonts. Studies have shown that the easiest type faces to read on the Internet are san serif fonts. Popular sans fonts include Helvetica, Avant Garde, and Arial. Popular serif fonts include Times Roman, Courier, and Palatino. Sans-serif fonts have become the de facto standard for "body" text on-screen, because monitors pr...

The REAL Value of Keywords

An important question in SEO is how much intrinsic value resides in a specific keyword and, whether SEO has the potential to take everybody on a fool's errand? When it comes to bigger companies, for instance, can a massive SEO investment in trying to achieve top ranking for almost-generic, ultra-competitive keywords be worth all the disappointment and soul-searching? Surely, in so many cases, there has to be a better way? At the other end of the scale are smaller companies with a limited marketing budget, particularly in the business-to-business sphere. There is often a fine balance to achieve when it comes to investing in SEO for what can only be low-traffic keywords in niche sectors, even where higher gross margins per sale indicate otherwise. Realizing this, many companies will skip the on-line sales dance, or resign themselves to having a website that is little more than an 'on-line brochure' presence or a support mechanism for Pay-Per-Click or social media activities. ...