In this series of articles, I'd love to shed some light on how The Next Web approaches marketing through Web analytics, Search Engine Optimization (SEO), Conversion Rate Optimization (CRO), social media and more. This time around, we're focusing on SEO and how we analyze and audit a website the size of The Next Web.
Why do we run SEO audits?We publish 30-40 articles a day and try to improve the site the best we can. It's becoming an increasing challenge to analyze all the pages we have online, but sometimes we can't always abide by the guidelines and best practices of our big friend, Google.
Before we dive into the data, analyze problems and get the website into the best shape we can, we need to know what is happening to the site in real time. Therefore we need to crawl everything.
The main purpose of these audits are:
Accessibility and Indexability
Is our website accessible for search engines, or are there pages we don't want to be found? If they can't find them, nobody will.
Technical errors
There are many factors search engines use to rank result pages. By crawling all pages, from domain to page level, you can spot possibilities to optimize for SEO more quickly.
After the crawl you still need to do the most important work, analyzing the data and turn it into an actionable plan. But let's start at the beginning, the setup.
Why Screaming Frog?Site audits can be very time-consuming. The biggest challenge we came across was that we wanted to have all URLs visible, but that was a much bigger task than we expected. However, we found Screaming Frog SEO Spider was more than up for the task.
An all-round tool which allows you to find broken links, check for Google Analytics (or any other) code on all pages, monitor all the redirects and find out the redirect paths in a website. But it also has its limitations.
The amount of URLs you are able to crawl is directly linked with the RAM capacity on you computer. My laptop only has 4GB of RAM, which meant I wasn't able to crawl thenextweb.com in its entirety. Things ground to a halt around the 40,000 URL mark, but the site actually has nearly 200,000. So we had a problem.
We looked into the alternatives that enabled us to keep down costs while maintaining the same level of flexibility we had before. We looked into a few cloud tools and since we'd like to use the data in an easy way and still be able to use the raw data at some point we decided that Screaming Frog was the way to go.
Still, the problem of not being able to monitor more URLs was still there.
Yes, it's possible to get yourself a monster of a computer, but we decided to go to the Google Cloud Platform and run it virtually. The biggest source of information we used for it was this article (thanks for that @FiliWiese).
What is the next step that we'd like to take to get more data on our SEO performance?
We're going to make sure that we can leverage the data from Screaming Frog even better so we can monitor our changes and performance over time. Hopefully in six months we'll be able to blog more about this. We're also looking into making this process more automated to save time and costs for running the server.
How you can run this?You don't need to be a complete geek to get how this works. So here's the complete set up on how to run ScreamingFrog in the cloud:
Hope you can benefit from this steps and tips.How are you trying to make your SEO process more innovative and efficient?
If you missed the previous posts in this series, don't forget to check them out: #1: Heat maps , #2: Deep dive on A/B testing, #3: Learnings from our A/B tests and #4: From Manager to Recruiter.
Source: How we manage SEO for a site the size of The Next Web
No comments:
Post a Comment