Deepcrawl is now Lumar. Read more.
DeepcrawlはLumarになりました。 詳細はこちら

URL Parameters

URL parameters are additional query strings that follow the main URL for a page, typically contained after a question mark. Multiple query string parameters can be used within a single URL, which can impact a website’s SEO and organic search performance in several ways, including causing duplicate content and crawling issues.  
 
This collection of SEO Office Hours notes addresses common problems seen with URL parameter handling, with recommendations on how to fix these issues with best practices suggested by Google Search Central.

Geotargeting Can’t be Applied to Language Versions Which Are Split Out Using Parameters

Google is unable to automatically detect and apply geotargeting if different language or country versions of a site are separated out using parameters.

30 Oct 2018

Google Only Crawls Hashed URLs When Unique Content Detected

Google doesn’t usually crawl hash URLs unless it has detected over time that there is something unique it needs to pick up from the hashed version.

19 Oct 2018

Use Log Files to Identify Crawl Budget Wastage & Issues With URL Structure

When auditing eCommerce sites, John recommends first looking at what URLs are crawled by Googlebot. Then identify crawl budget wastage and perhaps change the site’s URL structure to stop Googlebot crawling unwanted URLs with parameters, filters etc.

13 Jul 2018

Google Can Ignore UTM Tracking Parameters in URLs

For URLs with UTM tracking parameters, Google will ignore the parameter and focus on the primary URL instead. Google will try to crawl UTM parameter URLs if they are linked to externally though, so use the parameter handling tool in GSC or the canonical tag in this instance.

10 Jul 2018

Parameter Handling Signals Are Stronger Than the Canonical Signal

Google won’t blindly follow URL parameter handling set in GSC, but John says that this is a stronger signal than using canonicals.

15 May 2018

Content Loaded On Hash URLs Won’t Be Indexed

Google doesn’t index URLs with a hash separately, so if content is only loaded when the hash is used on a URL rather than being loaded on the main URL, this content won’t be indexed.

1 May 2018

GSC URL Parameters Are Signals, Not Rules, For Crawling

Rules set in URL Parameters in Search Console are used as a signal by Google for what they shouldn’t crawl, but they aren’t obeyed in the same way as robots directives.

27 Mar 2018

Google Can Proactively Assume Duplicate Pages Before Crawling Them

Google will sometimes assumes that pages are duplicates before crawling them. This can happen when you have multiple parameters for your URLs that don’t actually change the content being served.

6 Mar 2018

Google Assumes Domains Are the Same If Duplicate Parameters Used on Shared Hosting

If you have shared hosting for different websites and specify the same URL parameters for each one then Google may assume that the domains are the same or equivalent.

9 Jan 2018

Parameters Applied at Site Level Recommended as They Are Easier to Manage

John recommends applying parameters in Search Console at a site level as this is easier to manage than at a page level. However, if the same parameter changes the content on one page but is irrelevant on another then you will need to separate these in Search Console.

3 Oct 2017

Back 2/4 Next