Back in 2011, Google released an algorithmic update called Panda, and in the five years since, it has become one of the backbones of Google’s web search. Panda, and many of the tertiary algorithms surrounding it – the zoo of Penguin, Hummingbird, Pigeon, and all the rest – tend to focus on analyzing the quality of content on a site. Prior to Panda, links and keywords were the most important metrics. Yet, if you remember those old days, you would recognize how easy it was to game those metrics and get a mediocre or spammy page ranked in the top spots on Google search results.
Why do I bring this up? Well, the reason is that even if Panda has not be publicly updating as a stand-alone thing for a while, it’s still there and it still matters. What it means is that the quality of content on your site matters. If you have old, low quality content floating around, don’t ignore it. Chances are pretty good that it’s hurting your search ranking, and dealing with it could give you quite a significant boost to your search ranking.
On a large old site, though, you could have thousands, tens of thousands, or more pages all over the place. It would be a massive pain to try to manually go through them and analyze them for quality. Thankfully, you can use a few tools to make it all significantly easier.
The first thing you need to do is make a comprehensive list of all of the content on your site. Again, this would be a huge pain to do manually, so I’m going to recommend a tool to you. The tool is called Screaming Frog, and you can find it here.
Screaming Frog is a very comprehensive web crawling spider. It has a ton of features, from auditing content to analyzing meta data, checking links, creating a sitemap, integrating with Google Analytics, and a whole lot more. The free version is sufficient for what we’re doing today, with one caveat; it only works on sites with under 500 URLs to crawl. If you have more than that, and you probably will, you want to buy a license to the paid version. This costs about $200 (150 British Pounds) for a one year license.
I’m aware that it’s sometimes a tough sell to spend that much money on a tool without knowing how to use it, so feel free to get the free version and experiment. The wide array of features means the cost is well worth it if you learn to make use of them, but I understand if you don’t want to dive into it all right away. Still, I’m going to assume that bought the paid version and go from there.
With Screaming Frog, you will need to put in a few options to crawl your site. With the program open, click Configuration and click Spider. Under basic options, make sure every box is unchecked, then check “crawl all subdomains” to make sure you get every page on your site. If you know all of content pages are in a specific subdomain, you can crawl that subdomain and the subdomains of that subdomain, without crawling other top level subdomains. If that makes sense. With those options set, input your website URL in the “URL to Spider” box up at the top, then hit start.
Depending on how large your site is, how good your internet connection is, and how much memory your computer has, the crawl could take a while. Just take a break and let it do its work. When it finishes, you’ll be presented with a ton of data about your site, all organized based on the URL of each individual page. You can see the URL, the page titles and meta data, link information, HTTP response codes, and a lot more.
Once you have all of that data on hand, you can start to run an audit. Before you do, though, you should understand the markers of quality on the modern web.
What you’re looking for on your site are pages that could be holding you back as a whole, because of some lackluster element. I’m not going to cover site-wide issues here, like having poorly optimized meta data throughout your site or not using human-readable URLs. Those are good to fix, but they’re outside the scope of this article.
So, what constitutes poor quality content? You can look at it like this:
Essentially, what I’m going to do is consider word count to be the number one indicator of content quality, and let you use the other metrics to help you judge what to do with that content once you have identified it.
Going back to your data in Screaming Frog, go to the “internal” tab and sort your pages by word count. Export this data as a CSV file so you can work on it. Anything under, say, 1,500 words should be pulled aside. Everything else can be ignored; it’s probably fine.
Any piece of content under 1,500 words has to have a good reason to be that short. Here are some valid reasons:
If, however, the page is short because it’s a simple blog post and you didn’t think you needed to write much about it, that’s probably thin content. Covering an issue on a superficial level is fine, but you’re always going to be outranked by anyone covering it in more depth. It’s better if you find ways to expand the content, or remove it so it isn’t holding you back.
Go ahead and filter through your potentially thin URLs to find pages that meet the criteria for potentially being thin content. Once you have that list, you can move on to the next step.
At this point, you have a list of the thin content on your site. Now you have to decide what to do with it. This is the point where you want to have some idea of those other metrics, like traffic, social shares, and incoming links. Use your analytics suite of choice to gather that data.
The way I see it, there are generally four possible options.
What you do specifically will depend on your own analysis and your own personal thresholds. If you’re on the edge, go ahead and err on the side of saving the content and improving it in some way.
Before you go and start making changes, you want to do a few things. First, make a back up of your site. You don’t want to make changes that tank your site and have no option for recovering the lost value. Sometimes a page you think is thin is a page Google doesn’t mind, and removing it can hurt you.
Second, take snapshots of your important traffic and search ranking metrics. You want to know where your site starts so you know what effect your changes have.
Third, identify any other issues that could be having a negative effect on your SEO, particularly on those pages. Image captions, page meta data, URL structure, scripts, site speed; these are changes you can make while you’re making large changes to your site, in hopes of getting all of the improvement at once.
Once you’ve set up that groundwork, go ahead and make the content changes you want to make. You have two choices here; you can either change them gradually or you can change them all at once. Doing it all at once is likely to make your search ranking fluctuate a lot for a week or two, while Google figures out what you did. You might go up and you might go down before you settle in your new position, so don’t freak out until your metrics have stabilized.
Next, just to make sure Google understands what you’ve done, create a comprehensive site map and upload it to Google Webmaster Tools. This ensures that any lingering pages, new pages, or potentially missing old pages are indexed by the search engine.
That’s it! Once you’ve made your chances, all that’s left is to monitor your metrics and see what they did. Ideally, your site will have improved.