Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Duplicate Titles caused by multiple variations of same URL
-
Hi. Can you please advise how I can overcome this issue. outdoorsrank.com crawle is indicating I have 100's of Duplicate Title tag errors. However this is caused because many URL's have been indexed multiple times in Google. For example.
www.abc.com
www.abc.com/?b=123What can I do to stop this issue being reported as duplictae Titles, as well as content?
I was thinking maybe I can use Robots.txt to block various query string parameters. I'm Open to ideas and examples.
-
Depending on how you implement the canonicals, you should see a decrease in your duplicate errors, which will be replaced by canonical notices. Ideally, there won't be anything to ignore.
-
Thank you for your response.
Does this mean for each main page I have i.e.
etc,
if I put a Rel="canonical", I can then ignore messages of duplicate content for URL's reported such as
abc.com/page1 (put a rel="canonical" on this page)
abc.com/page2 (put a rel="canonical" on this page)
abc.com/index.html (put a rel="canonical" on this page)
etc,
?
-
*Edit: Miki beat me to it, but here's a little more explanation.
The first thing to note here is that Google's indexing doesn't actually have any effect on your Moz crawl report. All of the data you see there comes from our very own rogerbot, which crawls similarly to googlebot.
Though Google's crawler has a wide variety of ways to locate and index content, rogerbot can only crawl links on your site. If your crawl report is picking up each of these URLs, then there must be links pointing to those URLs somewhere on your site. The danger here is that Google and the other search engines will pick up those variants and not be able to determine which of them is the "real" one. That could lead to a) Google listing a URL you'd rather it didn't, or b) Google not understanding how to list your site at all.
A few of these have pretty simple fixes—index.html should be 301 redirected to your root domain, for example. Rel="canonical" is very applicable here, too. Here are a couple resources you may want to check out:
http://outdoorsrank.com/learn/seo/canonicalization - Best practices article on canonicalization
http://outdoorsrank.com/learn/seo/redirection - Best practices article on redirectsI hope that helps!
Matt Roney
Moz Customer Mentor -
I would redirect all variations to www.abc.com as well as REL=Canonical back to www.abc.com. This should solve you issues.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Near Duplicate Title Tag Checker
Hi Everyone, I know there are a lot of tools like Siteliner, which can check the uniqueness of body copy, but are there any that can restrict the check to the title tags alone? Alternatively, is there an Excel or Google Sheets function that would allow me to do the same thing? Thanks, Andy
Intermediate & Advanced SEO | | AndyRSB0 -
Duplicate Title tags even with rel=canonical
Hello, We were having duplicate content in our blog (a replica of each post automatically was done by the CMS), until we recently implemented a rel=canonical tag to all the duplicate posts (some 5 weeks ago). So far, no duplicate content were been found, but we are still getting duplicate title tags, though the rel=canonical is present. Any idea why is this the case and what can we do to solve it? Thanks in advance for your help. Tej Luchmun
Intermediate & Advanced SEO | | luxresorts0 -
What are partial urls and why this is causing a sitemap error?
Hi mozzers, I have a client that recorded 7 errors when generating Xml sitemap. One of the errors appear to be coming from partial urls and apparently I would need to exclude them from sitemap. What are they exactly and why would they cause an error in the sitemap. Thanks!
Intermediate & Advanced SEO | | Ideas-Money-Art0 -
Problems with ecommerce filters causing duplicate content.
We have an ecommerce website with 700 pages. Due to the implementation of filters, we are seeing upto 11,000 pages being indexed where the filter tag is apphended to the URL. This is causing duplicate content issues across the site. We tried adding "nofollow" to all the filters, we have also tried adding canonical tags, which it seems are being ignored. So how can we fix this? We are now toying with 2 other ideas to fix this issue; adding "no index" to all filtered pages making the filters uncrawble using javascript Has anyone else encountered this issue? If so what did you do to combat this and was it successful?
Intermediate & Advanced SEO | | Silkstream0 -
Artist Bios on Multiple Pages: Duplicate Content or not?
I am currently working on an eComm site for a company that sells art prints. On each print's page, there is a bio about the artist followed by a couple of paragraphs about the print. My concern is that some artists have hundreds of prints on this site, and the bio is reprinted on every page,which makes sense from a usability standpoint, but I am concerned that it will trigger a duplicate content penalty from Google. Some people are trying to convince me that Google won't penalize for this content, since the intent is not to game the SERPs. However, I'm not confident that this isn't being penalized already, or that it won't be in the near future. Because it is just a section of text that is duplicated, but the rest of the text on each page is original, I can't use the rel=canonical tag. I've thought about putting each artist bio into a graphic, but that is a huge undertaking, and not the most elegant solution. Could I put the bio on a separate page with only the artist's info and then place that data on each print page using an <iframe>and then put a noindex,nofollow in the robots.txt file?</p> <p>Is there a better solution? Is this effort even necessary?</p> <p>Thoughts?</p></iframe>
Intermediate & Advanced SEO | | sbaylor0 -
Easy way to change wordpress category titles. Currently categories are appearing with the same title?!
I'm working on a wordpress adult dating review site and have started to set up categories for each of my main keywords. I have also started to add sub categories by county and town and so far have done so for the counties of 'Lincolnshire' and 'Derbyshire'. The problem is though that for each of my subcategories the page titles are appearing the same. For example: www.mysite.com/category/online-dating/lincolnshire/spalding (root category online dating) shows the title as 'Spalding'. www.mysite.com/category/adult-dating/lincolnshire/spalding also has the title 'Spalding' even though it's root category is different (adult dating). It's probably easier to go to http://www.top-10-dating-reviews.com to see how it's set up. If you click in the category text in the top menu and navigate to dating/derbyshire/alfreton for example and then adult dating/derbyshire/alfreton you'll notice the page titles are the same. I use all in one SEO pack and have rewrite titles checked with category titles set to %category_title% | %blog_title%. I also use category SEO updater. In order to prevent duplicate content issues how can I simply make the title of each category category root title/category subtitle(county)/category subtitle 2(town). The title of each category page would then read for example Online Dating Lincolnshire Spalding.
Intermediate & Advanced SEO | | SamCUK0 -
Overly-Dynamic URL
Hi, We have over 5000 pages showing under Overly-Dynamic URL error Our ecommerce site uses Ajax and we have several different filters like, Size, Color, Brand and we therefor have many different urls like, http://www.dellamoda.com/Designer-Pumps.html?sort=price&sort_direction=1&use_selected_filter=Y http://www.dellamoda.com/Designer-Accessories.html?sort=title&use_selected_filter=Y&view=all http://www.dellamoda.com/designer-handbags.html?use_selected_filter=Y&option=manufacturer%3A&page3 Could we use the robots.txt file to disallow these from showing as duplicate content? and do we need to put the whole url in there? like: Disallow: /*?sort=price&sort_direction=1&use_selected_filter=Y if not how far into the url should be disallowed? So far we have added the following to our robots,txt Disallow: /?sort=title Disallow: /?use_selected_filter=Y Disallow: /?sort=price Disallow: /?clearall=Y Just not sure if they are correct. Any help would be greatly appreciated. Thank you,Kami
Intermediate & Advanced SEO | | dellamoda2 -
Capitals in url creates duplicate content?
Hey Guys, I had a quick look around however I couldn't find a specific answer to this. Currently, the SEOmoz tools come back and show a heap of duplicate content on my site. And there's a fair bit of it. However, a heap of those errors are relating to random capitals in the urls. for example. "www.website.com.au/Home/information/Stuff" is being treated as duplicate content of "www.website.com.au/home/information/stuff" (Note the difference in capitals). Anyone have any recommendations as to how to fix this server side(keeping in mind it's not practical or possible to fix all of these links) or to tell Google to ignore the capitalisation? Any help is greatly appreciated. LM.
Intermediate & Advanced SEO | | CarlS0