Without a technically optimized foundation, any time and money you spend on other SEO activities, such as what’s commonly referred to as “on-page” and content marketing, is an inefficient use of time and budget. This article explains what technical SEO is, how good-practice can impact site performance, and what your team or agency should routinely check.
“Do you know what my favorite part of the game is? The opportunity to play.” – Mike Singletary
SEO includes a range of activities designed to result in more sales via more, relevant, organic search traffic to your website. Activities like:
- Making sure the keywords that accurately reflect what your page offers, used in a way that matches how people use a search engine, are inserted in the right place.
- Creating useful and original content to support your product to help increase the ways that search users can come across your brand.
These kinds of activities are critical once your site is competing in the search engine results. Here’s the thing… Have you entered the game? Do you know if all the right pages on your site are in the competition? Do you have any performance inhibitors that are stopping your pages from getting into the race?
Accessible and Unique
The web is quite a challenging place as an information repository. Due to the potential of hyperlinking connecting two documents in separate locations, this creates an infinite web-like structure and complexity that we don’t get with more traditional information repositories – such as a hierarchical filing structure. Think about a time you might have landed on an interesting Wikipedia page then found an interesting source (hyperlinked) which links to another Wikipedia page. It’s a never-ending click hole! This isn’t just a time-sink for humans. Robots, like spiders and crawlers, travel by links too; therefore we need to allow them a clear and easy path through our site to ensure that all the URLs we want search engines to index are accessible and unique. By “accessible” our concern is to ensure that search spiders can access the URLs that you (the brand, business, or site owner) need to get indexed and ranked. On the flip side, we need to ensure that URLs that aren’t supposed to be indexed and ranked are adequately blocked from such. There are a handful of ways of achieving that end result. The chosen solution that a technical SEO may recommend will consider the totality of your site, plus your high-level digital marketing goals, combined with the data they have available before making a recommendation as to how prevent any site URLs from being accessed or indexed. Here are the main considerations when it comes to an efficient crawl.
1. Website Architecture
Most popular search engines, such as Google and Bing, place a strong emphasis on link data in their core algorithms. Search engines try to surface the best answer to a human question. While numerous signals contribute towards this qualification, freely given links from other writers on the web can act as a signal that a page is perceived as useful and authoritative. This link data accumulates equity on the target URL and the stronger the equity the greater the higher the algorithmic value of this component. For this reason having a site structure that affords the efficient flow of equity through your site should be a high priority. An ideal structure should be a pyramid and not a waterfall and in as far as possible try to keep core pages no further than three clicks from home. The second reason it’s important to architect a site coherently is that spiders travel via links; that is how our site pages are discovered before the machine decides to index. A well architected site will naturally result in a number of internal links pointing to deeper pages, thus increasing the chances of such a page being spidered. Layers of complexity are added when it comes to larger ecommerce sites that may have numerous categorizations, products, and variation within the same product (e.g., an online clothing store that caters to men and women; with numerous collections, types of item – within a collection – and variables such as size, color, and use-case – party dress, work dress, evening dress, etc.). Remembering that, whenever possible, we want the core product page to be no more than three levels from home, this is where a technical SEO will be looking for your use of attributes and directives.
2. Attributes & Directives
Attributes and directives can help solve similar common problems that occur on larger sites; particularly ecommerce sites. One of the most common issues preventing full crawl and indexing of larger sites is that caused by substantially similar pages. A search engine crawler is looking for new content to add to the index. A page has to be different enough (from other pages on the site) to warrant space in the index. With ecommerce sites, pages often aren’t that different to each other (e.g., category landing pages with multiple pages of results). If I navigate to “black shoes” on my favorite online clothing store I will have hundreds if not thousands of results returned. Using attributes correctly we can describe the relationship between one page and another, when to all intents Page 2 of my results is similar to Page 1 or 3, etc. Considering rel attributes in mark-up, a technical SEO will assess what kind of attributes you’re using and for what type of pages. Here are some that would be commonly considered:
- Rel=prev or rel=next: pagination attributes can be used for category landing pages, and help describe the relationship between multiple pages of results when there’s little other point of difference.
- Rel=canonical: can be used to explain a preferred URL when a set of URLs are substantially similar.
Another way to consider treating substantially similar pages is with strategic decisions about whether all versions should be indexed. In such situations we have to determine if our substantially similar pages have any point of difference that makes all versions worthy of indexing.
3. Index Or Not?
Setting aside the obvious areas you can’t allow to be indexed (site functions, secure processes, login areas and such), there are considerations as to what other pages may be prevented from crawl or index, for the greater good. Search engine crawlers are looking for new unique content. So it is a waste of crawl budget and index space to occasionally allow large areas of a site that may add no point of difference to be crawled and indexed. At worst, such issues could result in algorithmic demotions triggered by the Panda algorithm filter. For example if at the end of season, I move a lot of unsold stock to a “sale” folder, it may be best to keep this content out of the index so as not to potentially compete with my core, full-priced inventory. There are a number of ways to prevent crawl and index. A technical SEO will often need to understand and quantify all a website variables before suggesting a way, (or even combination of ways) of preventing URLs from being crawled and indexed though options may include:
- Securing pages behind a login.
- Using instructions in a Robots.txt file to prevent URLs and folders being crawled.
- Using a “noindex” directive on an individual page or set of pages.
- Using parameter handling in Search Console for URLs that have functional parameters applied though the content hardly alters.
Keeping indexed pages at an optimal level means that we’re keeping any equity moving efficiently through the site.
4. Crawl Efficiency
To get an efficient crawl we want to make sure that spiders can get around our site with as few barriers as possible so that there’s a greater chance of discovering our important URLs that we want in the index. Our concerns here as technical SEOs are:
- Do all URLs return a 200 (success) header status?
- Do 200 pages all render the correct content?
- Do we have any 404 (not found/error) pages?
- Are we linking to any URLs that return a 404?
- Does our intended 404 page return a customized page that helps get people (and robots) to another page?
- Can we find any redirected URLs on crawling the site?
It’s important to keep a site as tidy as possible, with no links to URLs that return a 404 and no links to URLs that redirect to another URL. Instead all links on our site should point directly to their destination. While in many situations there may be reasons for having to redirect URLs, using a 301 redirection header response code (not 302 for example) will also redirect some of the link equity we talked about earlier, as well as human visitors. However there’s no need to link to a URL that 301 redirects to another URL on our own site.
5. Index Efficiency
Understanding how our site is perceived in search indices is an important part of the benchmarking process. It can help us understand how much of our data is indexed, might we have a problem with duplicate content, or are we wasting space with old URLs still there? Using combinations of advanced queries a technical SEO will try to understand how many results from your site are indexed; are these the right results, is there an issue with lack of URL, domain or subdomain canonicalization. A query like [site:thisisanexample.com] may reveal results that show both HTTP and HTTPs results for my website. If that is the case, is this as intended? Was this intended? Is this impacting performance? These are all the questions that would need to be assessed as part of a technical audit. When problems are found and treatments agreed (such as a disallow rule in robots.txt file), it’s then good-practice to clean up the index by submitting a removal request in Search Console to clean out a specific URL or even folder. Optimizing crawl and index work together as a cornerstone of technical SEO. We’re making efficient use of search robots moving through our site without dead ends and redirections, as well as preventing a build-up of substantially similar or even functional duplicate pages from getting indexed. Time and again we see some of our most significant gains in visibility, traffic and ranking terms resulting from cleaning up crawl and index.
6. User-Friendly
In recent years, the most significant algorithm updates from Google have put the user experience front and center. In April this year we had the mobile update and from as early as 2010 Google have been confirming usability aspects such as page speed are factors included in their ranking algorithm.
Device Strategy
A technical SEO review should take account of how your website performs on different kinds of devices. Does your analytics package show any large disparity in performance? Is there a significant difference in your rankings when comparing mobile to desktop search queries?
How to Cater for Different Devices
Routine checks should include sample device comparisons from a front-end perspective, as well as looking at how you’re delivering content to non-desktop users. Choices exist around using one website that may respond and scale according to the device information sent in the server request; or to use more than one website version, each optimized for a particular kind of screen (e.g. a desktop version and a smartphone version) and use automatic redirection according to user agent. There’s no single perfect solution for all websites. The best strategy must take account of your business direction and audience needs. That said – one of the most common technical SEO issues is caused by duplicated websites intended for different audiences. Calling a site, or subdomain “the mobile website” doesn’t make it a mobile website. A website is available on any device that can connect and receive data through HTTP unless we specifically set to be treated differently. If we don’t, the result will be the kind of duplicate content issues mentioned above, that can cause serious problems.
Speed Matters
Pages on your site should render quickly irrespective of the end-user device. There are several ways to improve performance. Checks could include:
- Host performance, security, and server specification.
- How page content is delivered and use of any CDN (content delivery network).
- Reducing any unnecessary scripts and tightening code.
- Parallelize simultaneous calls.
We’ve experienced tangible incremental traffic gains that appear to directly correlate with the completion of speed optimization work on a handful of client sites, so it’s definitely an area to make sure if on the checklist as part of your technical SEO plan. In case you’re interested in monitoring or implementing a technical SEO routine with your team there’s a handy checklist on our site that you might want to check out that includes all of the above aspects to work through and more.