What is duplicate content
Content which is exactly copied from one already existing article is known as duplicate content. Here content can be of any type, such as literature, medias and pictures inside a web page.
Formerly: It is a term used in the field of search engine optimization to describe content that appears on more than one web page. – according to Wikipedia.
According to Google: Duplicate content generally refers to substantive blocks of content within or across domains that either completely matches other content or are appreciably similar.
Examples of duplicate content:
- Your blog or website content (thereby its literature) syndicated (copied) from another website.
- SEO non-friendly websites with different URLs pointing to the same content.
- Pages that have sorting options on the basis of time, date, color or other sorting criteria can produce duplicate pages.
- Pages with tracking codes and affiliate codes.
- Canonical Issue
- Thousands of e-commerce products and similar product variants.
What is Not Duplicate Content in the eyes of a search engine?
- Quotes from other sites when used in moderation on your page inside quotation marks. They must preferably be associated with a source link.
- Images and videos from other sites with different names. Or images repeated on your own website with similar names. (This is not considered duplicate content as search engines cannot crawl the content of images and videos).
- HTML Iframes
Impact of Duplicate content
Search engines read (thereby indexes) your web page’s content and identify your keywords. They then match your content with already existing data in their database. Once Search engines find that, your article contains duplicate contents which are already existing, it omits your web page from their results page. For example, check the below image.
Duplicate content happens all the time, on the internet. Not all duplicate contents get penalized. Why not? It is so because they are represented in a different way. Thus, we have a two categories of duplicate content on the internet.
- Non-malicious: This duplicate content issue occurs when a web page contains partial duplicate content from some other web page. It harms your SEO and lower your rankings in SERPs.
- Malicious: This duplicate content issue occurs when the author of a website intentionally duplicated the article, in an effort to manipulate search results and gain more traffic. In this case, your web page might get omitted from the search results.
In case, we have the web page about the history or some news which cannot be changed. Then, Search engines select the most user-friendly, well written, error free and informative version of web pages among the other duplicate web pages. And show the better one on the top. Plus, there are a lot of other factors that defines the rank of your web page in SERP.
How does google handles duplicate contents
How to resolve duplicate issues in your website.
Write unique content
Prevention is better than cure. It’s better not to copy anything from the internet while writing your blog post or your website’s content from the starting itself. If you are not a good writer you can use several online tools or you can take help from your friends. Either way, you will have a good and unique content for your blog or a website.
If you already have a content, then you can check it is duplicate or not through this tool.
Change your content and make it unique
If you have a duplicate content in your website, then you can change it using a spin bot. Spin bots are online software which changes your content via changing its words into its synonyms.
Use Rel=“canonical” Link tag
When you have multiple URLs serving the same content, choose the URL that you would prefer to be displayed in SERP. This will be your canonical URL. Then you should add a rel=“canonical” link tag in the <head> section of other web pages with the same duplicate content.
Make your website SEO friendly
- Use meta robots tag to no index and no follow your web pages that contain same information.
- Resolve Canonical Issues in your website
- Never include pages having duplicate content in sitemap
- alternatively, you can use robots.txt to block pages having duplicate content