8 Alternative Ways To Use Screaming Frog for SEO

If you don’t know what Screaming Frog is, (then where have you been!?) It’s a tool that lets you crawl an entire site on demand. But that’s not all it’s good for...I was recently working on a clients site and I found myself turning to the tool more and more for tasks other than the standard crawl feature. I thought it might be useful to put together a list of other ways to use Screaming Frog.

1 - Verifying Google Webmaster Tools Crawl Errors

I recently wrote a post on how to do a site audit using Webmaster Tools. I’m a big fan of its features, but I find that the GWT doesn’t update the crawl errors frequently enough. This means there are often things like 404s reported when they’ve actually already been fixed. I’ve been using Screaming Frog to solve this problem. Below is my new process for 404s and other common errors.

  1. Crawl the entire site and put all 404s into a spreadsheet.
  2. Download all 404s from Google Webmaster tools and put them into the same spreadsheet.
  3. Remove duplicates.
  4. Copy all URLs into a text document and save it as 404.txt
  5. Using the Screaming Frog list mode, upload the 404.txt document and start the crawl

Screaming Frog List Mode

You can then export all the remaining 404s and fix them.

2 - Checking If a Site Migration was Successful

A site I was working on recently changed their URL structure. For a couple of reasons, some of the URLs were not redirecting correctly. The list in Screaming Frog mode came in useful for checking which URLs were not redirecting correctly.

I got the client to send me a list of all the old URLs and followed the same process as above to find out which of the URLs were returning a 404. I then downloaded all the URLs with the problem and passed them to the developer to fix. This made identifying the problem really easy.

3 - Finding Infinite Links

Sometimes websites that use relative URLs can create never ending chains of links. Again, this recently happened on a client’s site. They were using relative URLs everywhere except some pages on the blog.

This meant that sometimes when they linked to a page on the blog, it was being appended to the existing URL. For example:

http://www.example.com linking to www.example.com/page1

Was creating http://www.example.com/www.example.com/page1

This was causing infinite lists of URLs. This means that search engines could be wasting their time crawling pages that technically don’t exist. Because this wasn’t the case on every page, I had to identify where on the site the issue was. When looking at some of the examples, the cause was using links that didn’t include the “http”. To find where this was happening I used the “custom” feature.

Screaming Frog Custom Feature

This is under the configuration tab. I asked it to include only pages that included the html :

“<href=”www.“

This then returned all of the pages that were linking to other pages in this way.

4 - Checking a List of Links

During processes of outreach, you often end up creating a large list of pages you are expecting links from. Going through each one to check that the link actually exists on the page can be a tiring job. To speed up the process, the Screaming Frog list mode lets you check a stream of URLs very quickly. There is already a post on the Screaming Frog blog on how to do this: Auditing Backlinks Using Screaming Frog.

5 - Creating a Sitemap

Screaming Frog makes creating an XML sitemap really easy, but it’s important that you set up the crawl correctly before you start. If you don’t limit the spider to crawl only those pages that you want in your sitemap, you can end up with a bunch or URLs that shouldn’t be in there. An example of this is with Wordpress, which I discovered when I crawled my site.

A common problem with Wordpress is that it creates pages like http://www.craigbradford.co.uk/about-craig/?replytocom=12 when people leave comments.

I don’t want these pages indexed and definitely not in my sitemap, so I can use the exclude tool (which is under the configuration menu), to ensure anything with this style of URL tail is excluded.

Screaming Frog Exclude Feature

Once you have set up the configuration, let Screaming Frog compete a full crawl of the site. Once complete you have the option to export the sitemap. Under the main navigation go to “Export” then select sitemap. You can then upload it to your site and submit it through webmaster tools.

6 - Check Sitemap for Errors

Duane Forrester from Bing recently said that Bing allow a 1% level of ‘dirt’ in a sitemap. “Dirt” could be anything from 404 errors to redirects.

Screaming Frog can be used to keep your sitemap clean and healthy. If you have the XML file like shown above, you can simply change to list mode and upload the XML file. Screaming frog can then crawl all of the URLs and tell you if there are any errors such as 404s or pages that are redirecting.

7 - Using Screaming Frog for Linkbuilding

When doing outreach, I often find it easier if I first contact a link target through something like Twitter.

Taking this one step further, an easy and innocent way to get on someones radar would be to crawl their site for them, find a blog post that returns a 404 and tell them about it.

Outreach Tweet

Assuming you don’t use an SEO profile, this is a good way to be nice and draw attention to the fact that you read their blog. Now, when you actually do contact them for outreach, it’s not out of nowhere and they’ll at least recognise your name and face.

8 - IP and User Agent Redirection

Two features that don’t get as much attention are the proxy option and the ability to change user agent. Taking them in order, the proxy feature can be useful for clients that are using IP based redirects. To see what’s going on you can use buy a set of international proxies or else try some of the free ones at Hide My Ass. You can them compare the results. To use this, select “Configure” then “Proxy”.

Screaming Frog Proxy Setting

Tick the box for “User Proxy Server” and enter your proxy details. When you crawl the site now, it will be using the international address instead of yours. If you are going to do this, I would recommend paying for private proxies as the free ones can be quite temperamental.

Changing user agent can be useful for checking if websites are treating search engine crawlers such as Googlebot differently to users. It can also pick up if robots.txt is explicitly blocking certain content from individual search engines. To use this feature just select “Configuration” and then “User agent”, it doesn’t get much easier than that.

That’s all folks! I’m sure there are Lots of other ways to use Screaming Frog or other crawlers, if you have any tips, please leave a comment and I’ll update the post with any other tips. If you have any questions, you can get me on Twitter @CraigBradford.

Craig Bradford

Craig Bradford

Craig joined Distilled in March 2011. Originally from Scotland, Craig moved to London in search of becoming an SEO ninja. After spending 4 years at Strathclyde University studying Sports Engineering, he decided that he didn’t fancy designing sports...   read more

Get blog posts via email

20 Comments

  1. Hi Craig,
    Great article!
    One small observation, though: rather than searching for
    a href=”www.,
    i think it's best just to search for
    href=”www.,
    because between "a" and "href" can be other elements, like target or class.

    reply >
  2. Hi Craig, Nice article. Distilled seem to be great advocates of Screaming Frog over Xenu and so am I! Thanks again

    reply >
  3. Thanks Craig for those great tips! And I agree with Olivian, it would be better to target just the href because there is a high change you could find something in the middle of the a and the href element.

    reply >
  4. Nice article Craig. I've been doing much of the same but with Xenu.

    One thing I have found when using broken links in outreach is that the more information you can give to the prospective webmaster the better the chance of success.

    I think its a great way to build links and help someone out at the same time. Everyone's #winning.

    reply >
  5. really nice article and informative, especially like the use of the back link for link building :-)

    reply >
  6. Craig, thanks for the features roundup. A feature I keep looking for, though, and haven't been able to figure out, is how to spider a list of sites in search for urls that contains specific source code. Any hints on that?

    reply >
  7. Code Buddy

    @Metapilot you can search the source of all pages during a crawl with Screaming Frog using the custom filter.

    reply >
  8. 7 is a really good idea. That said, although I'm sure many would appreciate the sentiment - and the effort you're making to help them out - I imagine one or two people might take offence to pointing out mistakes to them. I guess you've got to play it by ear, depending on the person's personality...

    reply >
  9. Beth

    I do like this program however, #6 above doesn't seem to work with any xml sitemap I upload. Is this a bug in the app?

    reply >
  10. Is there a guide like this for XENU?

    reply >
  11. One thing one need to keep in mind - run Screaming Frog (or Xenu) at off pick hours. Last thing you want is to compete with your site visitors for the server resourses.

    Great article, Craig!

    reply >
  12. Great, I just started using screaming frog and It's really simple to use. I am a bit confused as when I search for 404 errors on my site it provides me with a list of url's all ending with /Put Your G+ Page Url Here. Anyone know what the meaning of this is?

    reply >
  13. how many url's can screaming frog gather? it seems it stopped near 100,000

    reply >
  14. Hey Craig,

    I love your ideas about how to use screaming frog in a slightly different way. Especially, the tip with Google Webmaster Tools helped me a lot because that was the question I always had. How do I know what my remaining 404 are?

    I will certainly do that for all of my sites.

    reply >
  15. Indeed an essential tool today. And Dan at Screaming Frog is excellent in responding to questions etc. Great post Craig,

    reply >
  16. I was looking for a full detail about screaming frog until I dropped here... Thank you for your post, very interesting and informative tools for SEO! :D

    reply >
  17. Screaming Frog is really an amazing tool for website audit and as you mention for getting links, building relationship with other webmaster or blogger. I really like the broken link building efforts you mentioned. This is tool popped-up first when I think of response code check, broken links and other website errors. I highly recommend this tool because of it easy usability and customization according as per your need.....

    Thanks for sharing this!

    reply >
  18. Stefan

    Hey, Craig,
    Great tip #7, that's insightful, I'm going to try that!
    I've been using a combination of Screaming Frog and WebMeUp on my site. I've found that these two are enough for my SEO and site analysis needs.
    Cheers, Stefan

    reply >
  19. Helen Lunde

    There are some fantastic tips here! I get so excited when happen to discover nifty features which not lay on the surface, as for instance internal linking optimization in WebSite Auditor or finding infinite links in Screaming Frog (thank you Craig for this one). And tip #7 simply rocks!

    reply >
  20. Thanks Craig. I absolutely love Screaming Frog and couldn't do my job without it. It's an essential tool for website audits and getting the on page stuff sorted.

    reply >

Leave a Reply

Your email address will not be published. Required fields are marked *

*
*

You may use these HTML tags and attributes: <a href="" title=""> <abbr title=""> <acronym title=""> <b> <blockquote cite=""> <cite> <code> <del datetime=""> <em> <i> <q cite=""> <strike> <strong>