An SEO Guide to URL Parameter Handling

An SEO Guide to URL Parameter Handling
‘ );

h3_html = ‘



cta = ‘‘+cat_head_params.cta_text.textual content+’
atext = ‘


scdetails = scheader.getElementsByClassName( ‘scdetails’ );
sappendHtml( scdetails[0], h3_html );
sappendHtml( scdetails[0], atext );
sappendHtml( scdetails[0], cta );
// brand
sappendHtml( scheader, “” );
sc_logo = scheader.getElementsByClassName( ‘sc-logo’ );
logo_html = ‘‘;
sappendHtml( sc_logo[0], logo_html );

sappendHtml( scheader, ‘


‘ );

if(“undefined”!=typeof __gaTracker)
} // endif cat_head_params.sponsor_logo

While parameters are liked by builders and analytics aficionados, they’re typically an SEO nightmare. Endless mixtures of parameters can create hundreds of URL variations out of the identical content material.

The drawback is we will’t merely want parameters away. They play an vital position in a web site’s person expertise. So we’d like to perceive how to deal with them in an SEO-friendly means.

To achieve this we discover:

What Are URL Parameters?

url parameter elements

Also recognized by the aliases of question strings or URL variables, parameters are the portion of a URL that follows a query mark. They are comprised of a key and a worth pair, separated by an equal signal. Multiple parameters will be added to a single web page by utilizing an ampersand.

The most typical use circumstances for parameters are:

  • Tracking – For instance ?utm_medium=social, ?sessionid=123 or ?affiliateid=abc
  • Reordering – For instance ?type=lowest-price, ?order=highest-rated or ?so=latest
  • Filtering – For instance ?kind=widget, color=blue or ?price-range=20-50
  • Identifying – For instance ?product=small-blue-widget, categoryid=124 or itemid=24AU
  • Paginating – For instance, ?web page=2, ?p=2 or viewItems=10-30
  • Searching – For instance, ?question=users-query, ?q=users-query or ?search=drop-down-option
  • Translating – For instance, ?lang=fr, ?language=de or

SEO Issues with URL Parameters

1. Parameters Create Duplicate Content

Often, URL parameters make no vital change to the content material of a web page. A re-ordered model of the web page is commonly not so completely different from the unique. A web page URL with monitoring tags or a session ID is similar to the unique.

For instance, the next URLs would all return assortment of widgets.

  • Static URL:
  • Tracking parameter:
  • Reordering parameter:
  • Identifying parameter:
  • Searching parameter:

That’s fairly a number of URLs for what’s successfully the identical content material – now think about this over each class in your web site. It can actually add up.

The problem is that engines like google deal with each parameter primarily based URL is a brand new web page. So they see a number of variations of the identical web page. All serving duplicate content material and all concentrating on the identical key phrase phrase or semantic matter.

While such duplication is unlikely to trigger you to be fully filtered out of the search outcomes, it does lead to key phrase cannibalization and will downgrade Google’s view of your general web site high quality as these extra URLs add no actual worth.

2. Parameters Waste Crawl Budget

Crawling redundant parameter pages drains crawl price range, lowering your web site’s potential to index SEO related pages and rising server load.

Google sums up this level completely.

“Overly complex URLs, especially those containing multiple parameters, can cause a problems for crawlers by creating unnecessarily high numbers of URLs that point to identical or similar content on your site. As a result, Googlebot may consume much more bandwidth than necessary, or may be unable to completely index all the content on your site.”

three. Parameters Split Page Ranking Signals

If you will have a number of permutations of the identical web page content material, hyperlinks and social shares could also be coming in on numerous variations.

This dilutes your rating alerts. When you confuse a crawler, it turns into not sure which of the competing pages to index for the search question.

four. Parameters Make URLs Less Clickable

parameter based url clickability

Let’s face it. Parameter URLs are ugly. They’re onerous to learn. They don’t appear as reliable. As such, they’re much less doubtless to be clicked.

This will affect web page efficiency. Not solely as a result of CTR can affect rankings, but additionally as a result of it’s much less clickable on social media, in emails, when copy pasted into boards or anyplace else the complete URL could also be displayed.

While this will likely solely have a fractional affect on a single web page’s amplification, each tweet, like, share, e-mail, hyperlink, and point out issues for the area.

Poor URL readability might contribute to a lower in model engagement.

Assess the Extent of Your Parameter Problem

It’s vital to know each parameter used in your web site. But chances are high your builders don’t preserve an up to date checklist.

So how do you discover all of the parameter that wants dealing with? Or perceive how engines like google crawl and index such pages? Know the worth they bring about to customers?

Follow these 5 steps:

  • Run a crawler: With a instrument like Screaming Frog you may seek for “?” within the URL.
  • Look in Google Search Console URL Parameters Tool: Google auto-adds the question strings it finds.
  • Review your log information: See if Googlebot is crawling parameter-based URLs.
  • Search with web site: inurl: superior operators: Know how Google is indexing the parameters you discovered by placing the important thing in a web inurl:key mixture question.
  • Look in Google Analytics All Pages report: Search for “?” to see how every of the parameters you discovered are utilized by customers. Be certain to verify that URL question parameters haven’t been excluded within the view setting.

Armed with this knowledge, now you can determine how to finest deal with every of your web site’s parameters.

SEO Solutions to Tame URL Parameters

You have six instruments in your SEO arsenal to take care of URL parameters on a strategic stage.

Limit Parameter-Based URLs

A easy evaluation of how and why parameters are generated can present an SEO fast win. You will typically discover methods to scale back the variety of parameter URLs and so reduce the unfavorable SEO affect. There are 4 frequent points to start your evaluation.

1. Eliminate Unnecessary Parameters

remove unnecessary parameters

Ask your developer for an inventory of each web site parameters and its operate. Chances are, you’ll uncover parameters that not carry out a useful operate.

For instance, customers will be higher recognized by cookies than sessionIDs. Yet the sessionID parameter should exist in your web site because it was used traditionally.

Or you might uncover filter in your faceted navigation is never utilized by your customers.

Any parameters attributable to technical debt needs to be instantly eradicated.

2. Prevent Empty Values

no empty parameter values

URL parameters needs to be added to a URL solely after they have a operate. Don’t allow parameter keys to be added if the worth is clean.

In the above instance, key2 and key3 add no worth each actually and figuratively.

three. Use Keys Only Once

single key usage

Avoid making use of a number of parameters with the identical parameter title and a special worth.

For multi-select choices, it’s higher to mix the values collectively after a single key.

four. Order URL Parameters

order url parameters

If the identical URL parameter is rearranged, the pages are interpreted by engines like google as equal.

As such, parameter order doesn’t matter from a replica content material perspective. But every of these mixtures burns crawl price range and break up rating alerts.

Avoid these points by asking your developer to write a script to all the time place parameters in a constant order, no matter how the person chosen them.

In my opinion, it is best to begin with any translating parameters, adopted by figuring out, then pagination, then layering on filtering and reordering or search parameters and eventually monitoring.


  • Allows extra environment friendly use of crawl price range.
  • Reduces duplicate content material points.
  • Consolidates rating alerts to fewer pages.
  • Suitable for all parameter varieties.


  • Moderate technical implementation time

Rel=”Canonical” Link Attribute

rel=canonical for parameter handling

The rel=”canonical” hyperlink attribute calls out web page has similar or comparable content material to one other. This encourages engines like google to consolidate the rating alerts to the URL specified as canonical.

You can rel=canonical your parameter-based URLs to your SEO-friendly URL for monitoring, figuring out or reordering parameters. But this tactic isn’t appropriate when the parameter web page content material isn’t shut sufficient to the canonical, corresponding to pagination, looking, translating or some filtering parameters.


  • Relatively simple technical implementation.
  • Very doubtless to safeguard in opposition to duplicate content material points.
  • Consolidates rating alerts to the canonical URL.


  • Wastes crawl price range on parameter pages.
  • Not appropriate for all parameter varieties.
  • Interpreted by engines like google as a robust trace, not a directive.

Meta Robots Noindex Tag

meta robots noidex tag for parameter handling

Set a noindex directive for any parameter primarily based web page that doesn’t add SEO worth. This tag will stop engines like google from indexing the web page.

URLs with a “noindex” tag are additionally doubtless to be crawled much less incessantly and if it’s current for a very long time will finally lead Google to nofollow the web page’s hyperlinks.


  • Relatively simple technical implementation.
  • Very doubtless to safeguard in opposition to duplicate content material points.
  • Suitable for all parameter varieties you don’t want to be listed.
  • Removes present parameter-based URLs from the index.


  • Won’t stop engines like google from crawling URLs, however will encourage them to achieve this much less incessantly.
  • Doesn’t consolidate rating alerts.
  • Interpreted by engines like google as a robust trace, not a directive.

Robots.txt Disallow

robots txt disallow for parameter handling

The robots.txt file is what engines like google have a look at first earlier than crawling your web site. If they see one thing is disallowed, they received’t even go there.

You can use this file to block crawler entry to each parameter primarily based URL (with Disallow: /*?*) or solely to particular question strings you don’t need to be listed.


  • Simple technical implementation.
  • Allows extra environment friendly use of crawl price range.
  • Avoids duplicate content material points.
  • Suitable for all parameter varieties you don’t want to be crawled.


  • Doesn’t consolidate rating alerts.
  • Doesn’t take away present URLs from the index.

URL Parameter Tool in Google Search Console

GSC URL parameter handling

Configure Google’s URL parameter instrument to inform crawlers the aim of your parameters and the way you desire to them to be dealt with.

Google Search Console has a warning message that utilizing the instrument “could result in many pages disappearing from a search.”

This could sound ominous. But what’s extra menacing is hundreds of duplicate pages hurting your web site’s potential to rank.

So it’s finest to learn the way to configure URL parameters in Google Search Console, reasonably than letting Googlebot determine.

The secret is to ask your self how the parameter impacts the web page content material.

  • Tracking parameters don’t change web page content material. Configure them as “representative URLs”.
  • Configure parameters that reorder web page content material as “sorts”. If that is optionally added by the person, set crawl to “No URLs”. If a kind parameter it’s utilized by default, use “Only URLs with value”, getting into the default worth.
  • Configure parameters that filter web page down to a subset of content material as “narrows”. If these filters will not be SEO related, set crawl to “No URLs”. If they’re SEO related set to “Every URL”.
  • Configure parameters that present a sure piece or group of content material as “specifies”. Ideally, this needs to be static URL. If not potential, you’ll doubtless need to set this to “Every URL”.
  • Configure parameters that show a translated model of the content material as “translates”. Ideally, translation needs to be achieved through subfolders. If not potential, you’ll doubtless need to set this to “Every URL”.
  • Configuration parameters that show a part web page of an extended sequence as “paginates”. If you will have achieved environment friendly indexation with XML sitemaps, it can save you crawl price range and set crawl to “No URL”. If not, set to “Every URL” to assist crawlers to attain all the gadgets.

Google will routinely add parameters to the checklist beneath the default “Let Googlebot decide”. The problem is, these can by no means be eliminated, even when the parameter not exists.

So at any time when potential, it’s finest to proactively add parameters your self. So that if at any level that parameter not exists, you might delete it from GSC.

For any parameter you set in Google Search Console to “No URL”, you must also think about including it in Bing’s ignore URL parameters instrument.


  • No developer time wanted.
  • Allows extra environment friendly use of crawl price range.
  • Likely to safeguard in opposition to duplicate content material points.
  • Suitable for all parameter varieties.


  • Doesn’t consolidate rating alerts.
  • Interpreted by Google as a useful trace, not a directive.
  • Only works for Google and with lesser management for Bing.

Move From Dynamic to Static URLs

Many folks assume the optimum means to deal with URL parameters is to merely keep away from them within the first place. After all, subfolders surpass parameters to assist Google perceive web site construction and static, keyword-based URLs have all the time been a cornerstone of on-page SEO.

To obtain this, you need to use server-side URL rewrites to convert parameters into subfolder URLs.

For instance, the URL:

Would change into:

This method works nicely for descriptive keyword-based parameters, corresponding to these which establish classes, merchandise, or filter for search engine related attributes. It can also be efficient for translated content material.

But it turns into problematic for non-keyword related parts of faceted navigation, corresponding to value. Having such a filter as a static, indexable URL provides no SEO worth.

It’s additionally a problem for looking parameters, as each user-generated question would create a static web page that vies for rating in opposition to the canonical – or worse presents to crawlers low-quality content material pages at any time when a person has looked for an merchandise you don’t provide.

It’s considerably odd when utilized to pagination (though not unusual due to WordPress), which might give a URL corresponding to

Very odd for reordering, which might give a URL corresponding to

And is commonly not a viable choice for monitoring. Google Analytics won’t acknowledge a static model of UTM parameter.

More to the purpose, by changing dynamic parameters with static URLs for issues like pagination, onsite search field outcomes or sorting doesn’t deal with duplicate content material, crawl price range or inner hyperlink fairness dilution.

And having all of the mixtures of filters out of your faceted navigation as indexable URLs typically leads to skinny content material points. Especially when you provide multi-select filters.

Many SEO execs argue it’s potential to present the identical person expertise with out impacting the URL. For instance, by utilizing POST reasonably than GET requests to modify the web page content material. Thus, preserving the person expertise and avoiding the SEO issues.

But stripping out parameters on this method would take away the likelihood to your viewers to bookmark or share a hyperlink to that particular web page. And if clearly not possible for monitoring parameters and never optimum for pagination.

The crux of the matter is that for a lot of web sites, finishing avoiding parameters is just not potential in order for you to present the perfect person expertise. Nor wouldn’t it be finest apply SEO.

So we’re left with this. For parameters that you simply don’t need to be listed in search outcomes (paginating, reordering, monitoring, and many others) implement as question strings. For parameters that you simply do need to be listed, use static URL paths.


  • Shifts crawler focus from parameter-based to static URLs which have the next chance to rank.


  • Significant funding of improvement time for URL rewrites and 301 redirects.
  • Doesn’t stop duplicate content material points.
  • Doesn’t consolidate rating alerts.
  • Not appropriate for all parameter varieties.
  • May lead to skinny content material points.
  • Doesn’t all the time present a linkable or bookmarkable URL.

Best Practice URL Parameter Handling for SEO

So which of those six SEO techniques do you have to implement?

The reply can’t be all of them.

Not solely would that create pointless complexity. But typically the SEO options actively battle with each other.

For instance, when you implement robots.txt disallow, Google wouldn’t give you the option to see any meta noindex tag. You additionally shouldn’t mix a meta noindex tag with a rel=canonical hyperlink attribute.

What turns into clear is there isn’t any one good resolution.

Even Google’s John Mueller can’t determine on an method. In a Google Webmaster hangout, he initially really useful in opposition to disallowing parameters, however when questioned on this from a faceted navigation perspective, answered “it depends.”

There are events when crawling effectivity is extra vital than consolidating authority alerts.

Ultimately, what’s proper to your web site will rely in your priorities.

url parameter handling option pros and cons

Personally, I don’t use noindex or block entry to parameter pages. If Google can’t crawl and perceive all of the URL variables, it might probably’t consolidate the rating alerts to the canonical web page.

I take the next plan of assault for SEO-friendly parameter dealing with:

  • Do key phrase analysis to perceive what parameters needs to be search engine pleasant, static URLs.
  • Implement right pagination dealing with with rel=”subsequent & rel=”prev”.
  • For all remaining parameter-based URLs, implement constant ordering guidelines, which use keys solely as soon as and forestall empty values to restrict the variety of URLs.
  • Add a rel=canonical hyperlink attribute to appropriate parameter pages to mix rating potential.
  • Configure URL parameter dealing with in each Google and Bing as a failsafe to assist engines like google perceive every parameter’s operate.
  • Double verify that no parameter-based URLs are being submitted within the XML sitemap.

No matter what parameter dealing with technique you select to implement, be certain to doc the affect of your efforts on KPIs.

Image Credits

Featured Image: Paulo Bobita
In-Post Images/Screenshots: Created/Taken by creator

Source hyperlink SEO

Be the first to comment

Leave a Reply

Your email address will not be published.