Search engine optimization is an ongoing process. It takes a lot of work to achieve those elusive Page One rankings—and a lot of work to retain them. Part of the required onsite work is a systematic review of sources that might indicate there's trouble brewing or already in full storm mode.
This article outlines 10 health checks of indicators we include at our firm as part of clients' long-term SEO strategy, along with suggested check frequencies (there are no hard-and-fast rules on frequency; every company will have a different requirement). All examples are based on issues we've actually encountered.
1. Search Console health check
Search Console information is as close as you're going to get to understanding what Google does and doesn't want. It's advice straight from the horse's mouth. Though it contains lots of useful info (checked your average page CTR against site outliers recently, for example? Need to swap up your meta descriptions to make them more compelling?), reviewing the Coverage and Performance sections and any messages related to slow loading pages is essential.
2. Google Analytics traffic analysis
Review your organic traffic: Any peaks/troughs? pages performing as expected? Correlation between offline marketing activities and branded searches?
3. Meta tag and HTTP header analysis
To be honest, most SEO software will quickly pick up meta tag and HTTP header issues and alert you, but if you're not signed up with one of the SEO software houses, then download the free version of Screaming Frog and crawl your site using that instead. You're looking for any noindex meta tags in the section of the source code or X-Robots-Tag nofollow/noindex directives in the http header.
4. Technical health check
The most common technical errors we see are canonical errors and mislabeled hreflang tags ("uk" is not country code for Great Britain...). Common canonical errors:
- They disappear; suddenly you've got massive duplicate content problems.
- They all start referencing the unsecure http version of the domain.
- They reference 404s. Yep. Here's the original version of my page...
- They clash with hreflang tags. If you have a regional page indicated appropriately with hreflang tags but the canonical suggests the original version is actually the page dedicated to another region, then this is problematic for a search engine trying to figure out what's going on.
I won't go in to too much detail about hreflang tags, but suffice it to say, use the right country codes.
It's also well worth using the aforementioned Screaming Frog to run a custom search for Google Analytics/Tag Manager code on every page you're tracking. Massive traffic drop? Maybe the homepage has mysteriously shed its tracking code...
5. Robots.txt review
A simple but obvious one. Anyone been tampering with the robots.txt file? Accidentally copied the file over from a staging site and blocked all search engines from crawling any page on the site with a cheeky "Disallow: /"?
6. XML sitemap review
Many of the SEO CMS plugins will automatically update xml sitemaps as new pages are added to the site. Check to make sure that's definitely happening, and be sure to submit your sitemaps in Search Console.
7. Meta data review
You're ranking for your target keywords, but your organic CTR is awful. Have you accidentally written your shopping list as the meta description? Or, even "better," has a disgruntled ex-employee/agency changed all your meta titles and descriptions as a final "up yours" before leaving? Does that airline really offer complementary sexually charged massages with a surprising ending? And does the CEO really have male genitalia on his head? Believe me, this sort of thing happens.
8. Internal linking review
Frequency: every six months
Google spends a lot of time telling webmasters to use descriptive internal linking and not simply the words "click here." Want search engines to truly understand your page content? Then use the right keywords and their variations as your internal anchor text. Also, make sure your most important pages are linked to from multiple locations on your site; doing so will help the flow of PageRank while simultaneously telling search engines you really want them crawled.
9. Fetch and render priority product pages
Frequency: every six months
Unfortunately, blank pages struggle to rank, so if you do have dynamically loaded webpage content, it's well worth establishing what Google and Bing can and cannot see.
10. Schema review
Frequency: every six months
As we see more and more featured snippets in Google's results pages, it's important to make sure you're on point with schema mark-up that could snare you that elusive Position 0. Use Google's Structure Data Testing Tool to ensure site revisions haven't accidentally broken the code or removed it from the section of the relevant page. Also make sure your highest priority mark-up isn't nested under less important mark-up and your social mark-up is all in order so you smash those social CTRs.
BONUS: Full onsite/offsite audit
Simply a well worthwhile exercise!
You may like these other MarketingProfs articles related to Search:
- A Complete Guide to Anchor Text Optimization in Four Steps
- An 11-Step Plan for Improving Your SEO Strategy [Infographic]
- Five Ways to Get Keyword Ideas for Your Website: A Beginner's Guide
- A Marketer's Guide to SEO in 2022: Franco Valentino on Marketing Smarts [Podcast]
- How Many Words Do People Use When Searching Online?
- Three SEO Trends Marketers Need to Know in 2022