by Stoney deGeyter
Search engine spiders can be very forgiving with a lot of types duplicate content. I've found that, given enough time, the engines can learn when two websites or web pages are complete duplicates of the other. Once they figure this out then they basically understand that a link to one is a link to the other, etc. One version will ultimately be dropped from the index in favor of the other.
There are two basic problems with this. First, it takes time. Until the search engines figure out which dupes should be "merged" you're essentially splitting link flow between pages. Two inbound links split between each of the duplicate pages produces only half the power than two links both pointing to a single page. So until that merging occurs in the the search index, you're losing valuable link power.
The second problem is that you leave it to the search engines to decide which of the duplicate pages (or sites) should be dropped from the index. When you let the search engines decide, you lose essential control over your marketing plan.
The less you make the search engines think the better. It's not only OK, but it's essential to tell the search engines what to think, especially when it comes to which pages of your site should or should not be included in the index. If you have duplicate pages and don't keep the search engines from finding and indexing all of them, then you're forcing them to make the decision for you. Instead, you need to be proactive, let them know which version is the "correct" version.
One issue we've come across, especially with e-commerce sites is when products can be accessed via both secure and non-secure URLs.

This issue is typically caused by poorly implemented site navigation and linking. What happens is that the shopper adds a product to the shopping cart. At that point they enter into the secure pages. But when the shopper continues shopping, instead of proceeding to checkout, they navigate back into the site keeping the https: in the browser URL. This opens up the entire site to be indexed using secure URLs, creating duplicates of the non-secure URLs
There are a couple fixes to this. The first is to not allow your visitors to enter the secure areas of the site until they are ready to check out.

There is no reason to pass your visitors into the secure part of your site when they add products to their cart. The place to go secure is when they hit the checkout button. But--and this is important--if they enter the secure check out process but want to leave it to continue shopping, they need to be placed back into non-secure pages.
This leads us to our second fix: Use absolute URLs in all site navigation and shopping cart pages.
Quick refresher: an absolute link uses the full domain name in the link:
http://www.site.com/category/product.html
A relative link only uses the path from current location to the destination:
../product.html
When using relative links, if the shopper is already on an secure (https:) URL then they'll stay on secure URLs. When you use absolute links then you are forcing the visitor to go http: instead of https:.
When shoppers can access secure and non secure versions of the same page, then likely the search engines can as well. This creates almost a complete duplicate of your site, one secure and one non-secure version. Using absolute links will ensure that at no point can a regular page be accessed in secure mode, thus preventing the duplication.
This article is a continuation in my series on duplicate content. Follow the links below to read more:
- Theories in Duplicate Content Penalties
- How Poor Product Categorization Creates Duplicate Content and Frustrates Your Shoppers
- Redirecting Alternate Domains to Prevent Duplicate Content
Want more from your web site?
Search Influence can help! Targeted Traffic. Increased Revenue. Results Guaranteed. Customized Internet Marketing you can afford.
No comments:
Post a Comment