Select Page

CCTLD vs DIR Pros and Cons

There is a lot of different discussions around the web about whether you should use a Country Code Top Level Domain (ccTLD) or a Directory when scaling a website up into multiple languages or countries, so I am going to give an overview of the pros and cons for each method.

The main question many people ask is that when moving into a new market is which should be used to get the maximum benefit? A country code top level domain or a directory?

 

Country Code Top Level Domain (ccTLD)

A country code top level domain is a domain that is bound to a specific country, such as .uk for United Kingdom, .au for Australia, .cx for Christmas Islands, .fr for France, .th for Thailand etc.

 

Pros of Country Code Top Level Domains

  • Clear signal to Google who the website is targeting. There is no ambiguity, when Google looks at your website www.example.com.au it is clear that the website is targeting Australian users. Usually in Google Webmaster Tools you can set which country you want your website to target, when your website is on a generic top level domain such as .com. Although when your website is on a ccTLD this information is not available since Google assumes (correctly or not) that you are targeting customers within that country.
Screenshot below showing generic top level domain settings in Google Webmaster Tools where you can set the country you would like to target.
Where as when you look at the same information in Google Webmaster Tools when you are on a ccTLD such as .uk this is what you see
  • Another benefit to having a ccTLD is that some users in certain countriesprefer to buy from those domains. For example people in Australia are extremely patriotic and prefer to buy from .com.au domains (company in Australia domain). Where as in somewhere like Thailand, people prefer to buy from a .com domain opposed to a .co.th (company in Thailand) domain since there can be a lot of poor quality websites on .co.th domains. So Thai people believe .com businesses will provide a better service or product. This isn’t always the case and there will always be people with different preferences, so when moving into a new market it is essential to do your own research into what would work best for your business.
  • It is a lot easier for a user to link to www.website.co.th than it is for a user to link to www.website.com/th/. By making this process simpler then it could lead to an increase in natural links for your website.
  • By having several country code top level domains it is possible to occupy more of the search results pages for branded search queries. For example if someone searched for “your brand” then it is possible that your .co.uk website, .com website and .co.th websites could all rank. Google is no where near perfect when it comes to language differentiation although it is getting better.

 

Cons of Country Code Top Level Domains

  • The obvious one here is that it kind of limits where you can operateyour business. If you are on a .co.uk domain, then it is unlikely that you are going to rank well in other markets due to how Google perceives ccTLDs as shown above. That said, if you have plenty of local links say from .fr domains, then Google isn’t going to totally ignore them. It is likely going to ignore the ccTLD and change its mind and say ‘even though the domain is a .co.uk domain, this website has lots of links from .fr domains, so it must be relevant for French users too’ as shown in the image below.  I have seen a lot of random and non relevant websites on other ccTLDs rank in strange places in the past, but I would put that down to Google’s poor algorithm.

 

  • One potential pitfall of using a country code top level domain is that all of the links from your main aren’t fully being utilised and there is no history to your new domain. That said, what most larger websites tend to do is link all of their ccTLD domains together so that each page in a different language links to all of the other languages, so www.website.com/en/pageone.html would also link through to www.website.com/fr/pageone.html, or www.website.co.th/pageone.html and www.website.com.au/pageone.html would like together and vice versa.
  • To successfully launch a new language / localised website then it is important to consider what marketing budget is available to really push the new website. It does require significant effort to launch a new website into another market so think carefully about the options. If the website is simply being translated in the hope that some traffic will magically appear then a ccTLD is probably not the best option, instead just use the main domain with the new languages in a directory.

 

Language Directories

The second option when scaling a site up into multiple languages is to use a directory for each language such as www.website.com/fr/ for France (or French) and www.website.com/th/ for Thailand (or Thai). One important point to think about here is if you are targeting a certain country or a certain language, or both?

Taking French as an example, if you are only targeting France as a country then you will be missing out on a massive market from other French speaking countries and this can be quite big. There are 28 countries around the world where French is a national language which include: Belgium, Benin, Burkina-Faso, Burundi, Cameroon, Canada, Central African Republic, Chad, Comoros, Democratic Republic of Congo, Djibouti, France, Gabon, Guinea, Haiti, Ivory Coast, Luxembourg, Madagascar, Mali, Monaco, Niger, Republic of Congo, Rwanda, Senegal, Seychelles, Switzerland, Togo and Vanuatu.

Another option here is to target both language and country by providing a full localised directory such as www.website.com/ca-fr/ for Canadian people who speak French and www.website.com/be-fr/ for Belgium people who speak French.

 

Pros of Language Directories

  • One of the main pros to having all new languages in a directory is that all of the inbound links to your website will flow through to the new language pages. That said, this would also happen when linking your different ccTLDs together correctly although is it not fully known if both of these methods would pass the same amount of Google Juice through.

 

Cons of Language Directories

  • One of the potential pitfalls of having languages in a directory opposed to on a ccTLD is that Google could confuse the site as being a total mess in multiple languages. Although this isn’t likely to happen and you can set each directory to be targeting a specific country within Google Webmaster Tools as shown above.
  • It is possible that for branded search queries that you will find your main pages ranking in the incorrect language which isn’t a great user experience. For example if someone is searching for “your brand” from France then it is very likely that the main English website will rank since the English page has far more Google Juice strength and it can be difficult to get this correct.
  • The issue of what content you should display on the ROOT of the domain, www.website.com? Should it be English or one of the other languages? Then this takes you into the realm of potentially redirecting users either based on the IP address or on their language settings from the browser. Bearing in mind that the incorrect language could also show as a snippet in the search results which would dramatically reduce the click through rate.
  • If you are thinking of automatic redirects for users then you may havepotential Google Bot issues with cloaking. There is no clear info from Google as to what is best here and they have even totally contradicted their self several times on this issue. There is also the possibility that if the automatic redirects haven’t been set up correctly that you could accidentally block Google from accessing most of your website.  For example since Google Bot always crawls from America then you wouldn’t want to be redirecting Google Bot to the English page all of the time since it would never discover any of the other content!
  • People within the market you are entering are likely to link to the main website and not the localised directory, for example you will find a lot of people would link to www.website.com opposed to www.website.com/th-th/ which could hinder how well the language directories rank within the local search engines.

 

Summary

There is no quick win to breaking into a new market and there are a lot of pros and cons to each of the different methods, country code top level domains and directories. Personally I would prefer to use ccTLDs where possible if there is going to be a real push into that market.

If you are just thinking about translating and not promoting then it is a waste of money, similar to buying a phone and expecting it to ring on its own! Nothing worth while is easy, so it is going to take considerable effort to break into the market.

You may have noticed that I have not mentioned the option for sub-domains within this post and that is because I don’t like sub-domains :-)But if you are interested then I am sure you can pull some of the above points for/against sub-domains too.

Hopefully this has provided you a wide range of information on the topic and will help you decide what is best for your website.

Here is a summary table outlining all of the different points listed above

 

The 200 Signals Used In Google’s Ranking Algorithm

Whilst I would love to think I knew the full 200+ signals being used in Google’s ranking algorithm, I don’t think I would quite get to 200 on my own. :-) I am asking for help from the SEO community in this post to share their knowledge on how Google rank websites.

This post is designed to list all of the possible data sources / signals Google could be using as part of their ranking algorithm. My feeling is that if they have access to data, then they will be using it in someway or another. Weather this is to help boost websites or to flag websites as spam and demote them. We aren’t going to be discussing the actual weight Google is giving to each of the different signals (that can be for another post :-) ), just to see if we can get all 200 (or more) signals theycould be using.

I have broken down the main areas to help categorise the different signals. Anyone who can add to this list will get a link back to their blog (as long as you don’t have a spammy website!) since this is my way of saying thanks for contributing.

Anyone wishing to contribute can add a comment to this blog post and I will update the blog post. When commenting please state;

  • What piece of information Google could be using
  • How they would judge / rank this information (I.e. how does this show quality or not?)
  • Where they would get this data from
  • Any relevant official sources of information to support your theory

A few of the easy ones have already been put in :-) I want this post to be generated by the SEO community to see how many we can get between us all.

So lets get started then!

On Site Signals

  1.  Meta title
  2.  H1 & heading tags
  3.  Keywords within the text
  4.  Amount of useful text when subtracting templated information and adverts
  5.  Phone number, which signifies that the website is a genuine business
  6.  Amount of unique images which cannot be found anywhere else
  7.  Amount of unique high quality content
  8.  Is there regular fresh content appearing on the website?
  9.  Internal linking
  10.  Site speed
  11.  Image ALT attribute
  12.  Amount of adverts on the page & how much useful content is left once the adverts are removed
  13.  Video markup
  14.  Microformats and rich snippets such as hreview or hrecipe
  15.  IP C class and location of server
  16.  Content above the fold VS content below the fold
  17.  Top level domain
  18.  How long the domain has been owned by the current owner, not just how long the domain has been registered to ‘someone’
  19.  Keyword over usage or keyword stuffing (negative factor)
  20.  Cloaking (negative factor)
  21.  Hidden text (negative factor)
  22.  Duplicated content from external websites
  23.  301 redirect flags such as redirect chains (301 –> 301 –> 301 –> Final destination), redirect loops or redirects ending on a 404 page.
  24.  Keyword rich URLs opposed to page numbers or product ID’s
  25.  Number of crawl errors found when crawling the site
  26.  Does the HTML conform to W3C standards

Off Site Signals

  1.  Anchor text for inbound links
  2.  Number of linking root domains
  3.  PageRank from inbound links
  4.  Followed / Nofollowed attributes on inbound links
  5.  Number of mentions (ie, ‘website.com’ without an actual link)
  6.  Total number of external links in addition to linking root domains
  7.  Surrounding text on external links
  8.  IP C class of linking websites
  9. Existence of a Google places profile
  10.  Link profile of competitor websites
  11.  Age of the back link
  12.  Link growth rate and how fast links to certain pages have gained links
  13.  Evidence of paid links
  14.  Selling ‘followed’ links to other websites (negative factor)
  15.  History of comment spamming on forums, blogs or other link spam (negative factor)

Brand Signals

  1.  Does the website have a brand logo which appears throughout the website?
  2.  Does the brand logo appear anywhere else on the web?
  3.  Does the website send out branded emails to its customers? This data could be gathered by analysing gMail accounts.
  4.  How many gMail accounts are receiving, opening and engaging with emails sent from this brand website?
  5.  Does the website have a LinkedIn company page with real people working at the company?
  6.  Search volume for branded keywords and brand name
  7. Presence of a physical address on the website
  8.  Age of the domain
  9.  Number of branded external links

Social Signals

  1. Does the website have a Twitter account?
  2. Does the website have a Google+ brand page?
  3. How many Plus 1’s does the website/web page have?
  4. How many Facebook fans does the brand page have?
  5. How many people are Tweeting about the website?
  6. How many Twitter followers the website has
  7. Surrounding text of links in tweets
  8. Number and rate of growth for page views and replies on social media pages
  9. Presence of user generated content (UGC) on the website
  10. Growth rate of social media mentions, such as a slow growing amount of mentions or a large outbreak

Other Signals

  1.  Is the website advertising on AdWords?
  2.  Is the website advertising on Google Display Network?
  3.  Is the website listed on any official business websites such as Business Link or Yell.com?
  4.  Does the website have a physical address?
  5.  Is there an email address where customers / website visitors can contact the owner?
  6.  Click through rate (CTR) from the search results
  7.  Bounce rate from the search results
  8.  Review rating of the website / business on third party websites
  9.  Google seller reviews for eCommerce websites
  10.  Quantity of pages in Google’s index
  11.  History of past penalties for this domain (negative factor)
  12.  History of past penalties from this owner of the domain (negative factor)
  13.  History of past penalties for other domains from the same owner (negative factor)
  14.  Amount of time spent on the page / site
  15.  Number of page views per visit

 

Please add your comments so we can get this list full. I would be surprised if we couldn’t fill this list with all the knowledge out in the SEO industry.

Contributors

Michael Cropper – SEO and Internet Geek (1, 2, 3, 4, 5, 6, 7, 8, 41, 42, 43, 44, 45, 81, 82, 83, 84, 85, 121, 122, 123, 124, 125, 161, 162, 163, 164, 165)

Felix Lueneberger (9, 10, 11, 12, 13, 14, 15, 16, 17, 46, 47, 48, 49, 50, 86, 87, 88, 89, 126, 127, 128, 129, 130, 166, 167, 168, 169, 170)

John – Forest Software (18, 51, 52, 53)

Sean – http://01100111011001010110010101101011.co.uk (19, 54, 55, 20, 21, 22, 171, 172, 173, 23)

Jeremy Quinn (24)

Brahmadas from SEOZooms (25, 26, 174, 175)

 

12 thoughts on “The 200 Signals Used In Google’s Ranking Algorithm”

  1. Interesting approach ;-).
    I’d like to add:

    “on site”:
    -internal linking (!)
    -site speed
    – TLD
    -image alt tags
    – amount of ads on the page
    -video markup
    -microformats such as hreview or hrecipe
    -IP C class, location of server
    -content above Vs. below the fold

    “off site”:
    – total number of external links (in addition to LRDs already mentioned)
    – surrounding text of external links
    – IP C class of linking sites
    – Google places profile exists
    – Link profile of competitor sites

    “brand signals”:
    -search volume of brand name
    – physical address on page
    – age of domain / site
    – number of branded external links

    “social signals”:
    – how many twitter followers does the site have
    – Surrounding text of links in tweets
    – what is the rate of pageviews / replies on social media
    – is there UGC on the site?
    – how fast is the growth of social mentions (slwoly growing or outbreak?)

    “other signals”:
    – CTR from SERPs
    – bounce rate from SERPs
    – Reviews rating on third party sites
    – Google seller reviews (for e-commerce sites)
    – number of pages in Google index

  2. I would add :

    Onpage :
    – Age of domain (although I see that that is in Brand signals) more specifically, how long it has been owned by “you” not just registered

    External links :
    – How old the link is
    – how quickly these links have been obtained
    – any evidence that links have been paid for

    1. Of course it is, what else would this be. :-) It’s also good to see what people within the SEO community think could be being used as a ranking signal.

      But the real question is…are you the real Matt Cutts? (It would be awesome if you are and knowing that you are reading my blog :-) But it could also be a fake)

      Update: The above comment is not from Matt Cutts, the real Matt Cutts has clarified.

  3. What about Crawl rate, crawl errors, validation, page overal visibility, user side reputations, bounce rate, time on site and pages, number of page views . there are many things I think…. so it may not be 200+ may be more than 300 or 400. for the time being and Google may identify and add many and many new factors in the future so that the SEO game will in more tough plat form….. he he

    1. Technically correct. But it is a common understanding that when one says “Meta Title” that they are talking about the <title> tag and not the old skool tag which is never used: <meta name=”title” content”something”/>

  4. Think that is jumped ! one or serious forget to attend skool before attending college

    Any way nice post Mick. Keep it up. Think more time will take to fill the blank space of signals. Or the Google God is not getting right signals now? Regards

    1. Thanks for sharing Robert, the website provides a nice scientific approach whereas this list is a nice quick reference point based purely on peoples opinions and experience within the industry.

3 Essential Excel Plugins for SEO

Here I am going got talk through some of the Excel plugins I have found very useful for SEO tasks. Best thing I can advise is to get them installed and give them a go for yourself. The three plugins to look through today are;

  1. Regular expressions plugin for Excel
  2. SEO Tools plugin for Excel
  3. Excellent Analytics plugin for Excel

 

Regular Expressions Plugin for Excel

Link: Regex Plugin Excel

This can be a great plugin when working with large data sets within Excel as it allows you to use all of your normal regular expressions to find the data you need. See this site for a regular expressions cheat sheetfor reference, alternatively if you aren’t too familiar with regular expressions then I suggest reading the following tutorial.

 

There is not much more to say about the regular expressions plugin for Excel apart from that you can filter out a lot more than you can using the normal string matches which Excel has be default. It can be a lot easier to get the information you require.

 

SEO Tools Plugin for Excel

I briefly mentioned this in an earlier post about why programming is an essential SEO skill but I wanted to expand on the amazing excel plugin which is SEO Tools by Neils Bosma and why you need to be using this if you aren’t already.

Below is a screenshot of the additional tab you get in Excel once you have installed the plugin (60 second install, follow the instructions in the ‘read me’ file!)

 

 

So what are some of the cool things you can do with SEO Tools once it is installed? Well an awful lot actually.

Check the HTTP Status of pages

Simply add a lot of different URLs in column A then use the function=HttpStatus(A2) to find out if it is returning a 200, 404, 503 etc. This can be a really useful tool for both SEO since Google Webmaster Tools reports on errors for your site for a long time, often when these pages are actually no longer showing any errors. So it can be a good method to list all of the URLs which Google Webmaster Tools is showing as errors and double checking these to spot the real pages that are showing as errors.

To find this function you can simple click on the ‘onpage’ button highlighted and select the HttpStatus() option from the list

 

Filter through back link targets

Lets say you have a list of websites / webpages you want to target for back links but it isn’t easy to figure out which ones to target first. From a large list, it is important to be able to filter through these to identify the best websites to contact first. There is no point in contacting poor quality websites first!

Maybe you want to prioritise websites to contact who have “SEO” in their meta title and have a page rank of at least 1

 

Identify popular content on competitor websites

How about trying to identify which content is the best on competitor websites? Maybe a competitor has a blog or a content section which you can easily scrape all of their URLs from. Simply go to Google, view 100 results and search for “site:blog.website.com” then use Scrape Similar to get all of the URLs where you can paste these into Excel.

Once you have all of these URLs then you can check how popular they have been on the social media channels. In the screenshot below I have pulled a few of my blog posts and identified how many tweets they have had recently using the =TwitterCount(A2) function from SEO Tools.

 

Whilst I am not a massive fan of copying off competitors (I believe you should be doing your own thing far better so they are copying off you!) this can be a useful tool to get a rough idea of different content that people are interested in. It could even be a good guide to identifying your own ‘linkable content’ since Google Analytics can only show the amount of traffic that each of the pages gets.

So looking at the above screenshot, if those websites were for competitors then I would think about doing a post about how to see competitors social media link statistics since this has gained the most traction.

 

Website Scraping with XPathOnURL

This is one of the most powerful bits of SEO Tools and one of the best bits with this Excel plugin. I am not going to go into too much detail here about XPathOnURL since this is a full blog post on its own. To give you an idea of what it can do though it basically scrapes certain bits of the HTML that you need such as “get me all the links on this page” which would translate to “get me all the HREF attributes within all of the <a> tags on the page”.

This tool can be extremely useful when scraping websites for certain pieces of content / information etc. Have a play, see what you can do.

 

Excellent Analytics

Link: Excellent Analytics

This can be a really useful Excel plugin for SEO since it allows you to export all the data you need from Google Analytics straight into Excel. So no more exporting large data sets then matching up the relevant data with a =VLOOKUP() function instead, just get the precise data you need from GA straight into the adjacent column in Excel where your other data is.

I am not going to cover this in a large amount since it is quite self explanatory, it pulls in data from Google Analytics based on all the normal segments, filters and visitor data you are used to using via the normal Google Analytics dashboard. Simply link up your GA account with the plugin and away you go. Below is a screenshot showing one of the user interfaces.

 

On larger websites which 10’s of millions of page views per month you may have noticed how slow Google Analytics is to do anything, so this can help speed this process up for extracting the data you need.

 

Summary

These are the three essential Excel plugins SEO I use on a regular basis. I have found them to be real time savers for various tasks as outlined above so go and give them a go.

If you have any other excel plugins that you use for SEO then please leave a comment.

 

2 thoughts on “3 Essential Excel Plugins for SEO”

How To Create An Embed Code For Your Infographic

Here I am going to show you how easy it is to create your own embed code for either your widgets or infographics you have created. They are an essential tool if you want to help build natural links.

This is what the final piece of code looks like to a user, with the images / info graphic / widget above and the text to embed this below.

 

Click below to embed the above image into your website

 

So how do you do that? Well it is quite simple really, just a few bits of HTML and JavaScript will enable you to create a box with your embed code in so people can place your images / infographics / widgets on their site with that all important link back to your website as the original source.

The first step is to include the image on the page you are wanting the embed code to be one. Once you have done that you simply add the code below to your HTML and tweak the relevant bits which will be explained in a moment.

 

Complete Embed code:

 <textarea style=”height: 44px; width: 300px; margin: 2px;” onclick=”this.focus(); this.select(); _gaq.push([‘_trackEvent’, ‘Infographic’, ‘Click’, ‘guerilla-marketing-infographic’]);” rows=”2″>

&lt;img src=http://www.michaelcropper.co.uk/wp-content/uploads/2011/09/guerilla-marketing-at-its-best-in-manchester.jpg width=300 height=300 /&gt;

&lt;br&gt;

View full image

&lt;a href=http://www.michaelcropper.co.uk/wp-content/uploads/2011/09/guerilla-marketing-at-its-best-in-manchester.jpg title=”Guerilla Marketing Manchester” style=”color:#0000FF;text-align:left”&gt; Guerilla Marketing Manchester&lt;/a&gt;

</textarea>

 

So if we take a look through each of the different sections to explain what each part does.

 

Step 1 – Create a text box

Create a text box on the screen which is 44 pixels high, 300 pixels wide and has a bit of a margin. This section also has some JavaScript so when a user clicks (onclick section) the text box then the text is highlighted (this.focus(); this.select() section) which makes it easier for the user to then right click and copy the text. In addition, the code below also has an event tracking Google Analytics tag which means that when someone clicks the text box an event is fired off to the Google Analytics account you have installed on the page which says “someone has clicked an ‘infographic’ which is called ‘guerilla-marketing-infographic’ “

Below the code is what the code actually creates, the text box (although yours will be empty to begin with until the next step.

 

<textarea style=”height: 44px; width: 300px; margin: 2px;” onclick=”this.focus(); this.select(); _gaq.push([‘_trackEvent’, ‘Infographic’, ‘Click’, ‘guerilla-marketing-infographic’]);” rows=”2″>

 

If interested, read full details about Google Analytics event tracking, but in summary is _gaq.push([‘_trackEvent’, ‘category‘, ‘action‘, ‘label‘]);where you can choose the category, action and label you want to use.

 

Step 2 – Add the Image

You’ve got something to embed right? Well this is the next step, place the HTML code of your original image / infographic / widget in between the <textarea> and </textarea> tags. One key thing to note though is that you need to replace HTML characters such as < and > with their HTML code equivalent. Take a look through a full list of HTML codes for reference. It can be a bit of a pain to do this, but hopefully you don’t have much to include here!

 

&lt;img src=http://www.michaelcropper.co.uk/wp-content/uploads/2011/09/guerilla-marketing-at-its-best-in-manchester.jpg width=300 height=300 /&gt;

 

Step 3 – Add the Link Back to You

Here is the most important part, adding a link back to your website as the original source of the content. Same rules apply as in step 2 where by you need to replace the HTML characters with their HTML codes. The great thing here is that you can choose the anchor text which is used when people are linking back to you.

 

View full image

&lt;a href=http://www.michaelcropper.co.uk/wp-content/uploads/2011/09/guerilla-marketing-at-its-best-in-manchester.jpg title=”Guerilla Marketing Manchester” style=”color:#0000FF;text-align:left”&gt; Guerilla Marketing Manchester&lt;/a&gt;

 

I have seen some people include cheeky links within this section previously which link back to one of their key pages they want to rank with the best anchor text. Whilst this isn’t a bad idea, personally I would only add it where relevant. A lot of people will likely remove the link anyways, but it is certainly worth a try :-)

 

Step 4 – Drive Traffic to Share Your Content

Now comes the easy bit! Go out there and tell people about the great content you have just created. If the content truly is great then it is going to make people want to embed this item in their own websites and tell their friends about it naturally.

Here is what the final piece will look like when a user embeds the code in their own website:

 

So there we go, simple as that! Now time to go and create some great content and promote it well. For anyone who is interested in what the massive gorilla is, then have a read of my guerilla marketing Manchesterpost.

Here is one we use in our info graphic series for A day out on the District line in London. Great option for promoting some great content.

Note: You may also need to wrap ” ” quotes around some of the HTML tags within the embed code itself. Strangely when I was creating this post additional quotation marks were being added around the IMG SRC and A HREF attributes causing invalid URLs, so if you come across this then just tweak accordingly.

 

10 thoughts on “How To Create An Embed Code For Your Infographic”

  1. Great code, thanks.

    The speech marks are formatted so it didn’t work for me until I changed them:

    ” – didn’t work (ASCII)

    ” – did work (type into notepad)

    Hope that helps someone. :)

  2. I have been trying to edit the code so it displays properly but I am having issues. I don’t know what I am not replacing correctly

    this code doesnt work but i gave it a valiant effort. if you can provide any help i would appreciated it! thanks for the great code anyways

    1. Hi Joffrey,

      Your code didn’t come through properly, likely because WordPress filters certain things out in comments by default. If you want to explain your issue in the comments or email me with more details then I will see if I can point you in the right direction.

      Thanks
      Michael

  3. Hi I am trying to help my friend do this on her friends website here.
    I have tried following the instructions, tried changing the ” and at this point in time, I can’t figure it out.

    It’s a wordpress site and I owe her a favour so it would be really good, if I could get this working for her.

    Thanks in advance

    Peter

37 Essential SEO Tools And Resources

There have been hundreds if not thousands of blog posts on SEO tools before, so I am not going to waffle on about why you should use each one. The main reason for this blog post is more of a reference manual for me  ????

These are the tools that I have found invaluable being an SEO so I hope they are of some use for anyone reading too.

 

Google Webmaster Tools

Link: Google Webmaster Tools

Great SEO tool to easily spot duplicate content, duplicate title tags, malware on the site, HTML issues and URLs Google has found that are either 404’ing or 500/503’ing.

This is a great starting point to diagnose websites to see what is potentially going wrong with them and it can also be a great way to spot malformed back links that you can 301 to the correct place.

 

Google Analytics

Link: Google Analytics

What a powerful SEO tool. I am not going to go into details about why this is so amazing, but if you are not using this (or an equivalent analytics package) then you need to get this installed and find out for yourself why it is so amazing.

 

Microsoft Excel

You need to make sure you have the full version of Excel and not just the free version that is available since there are a lot of additional features you will need to use such as advanced filtering and customising your Excel with some nice plugins.

When using Excel I also have additional plugins installed which help with more advanced functions and save a lot of time. I did another blog post about essential Excel plugins for SEO recently which goes into more detail about these three plugins.

  1. Regexp Plugin
  2. SEO Tools
  3. Excellent Analytics

 

Google Docs

Link: Google Docs

The main reason I use Google Docs is for their spreadsheets. This can be a great way to share certain bits of information between a team of people, may that be list of websites that have been contacted for links if there are teams of link builders or it can also be used for more creative reasons such as the ImportXML function that is available.

I tend to use this in addition to Excel since there are some things that Google Docs can do better than Excel, the main one being that Excel is limited to importing only one XML feed URL per document (without a lot of effort tweaking things and writing Macros at least) whereas Google Docs has a nice limit of 50 XML feed URLs that can be pulled in within one spreadsheet, which can be great when scraping content from other websites.

 

Open Site Explorer

Link: Open Site Explorer

I briefly mentioned this earlier, but this tool was created by SEOmoz and is based on their Link Scape index. Basically SEOmoz have Link Scape which crawls the web just like any other search engine would do and then indexes the information. From this information they then provide all of the juicy statistics that SEOs like to get their Excel spreadsheets on.

Go and have a play with what is available, but to get the full data set then you need to sign up to become a Pro Member which is well worth it for anyone doing a lot of SEO.

One thing to bear in mind is that it does provide a lot of data and one could argue that it is just data and only useful if there can be some actionable insights that can be gained. I have seen a lot of SEOs in the past creating reports, pie charts, graphs of all this wonderful data and comparing it with competitors without anything coming from this. So just make sure you know why you are gathering and analysing all of the data before you get started or you will just end up with a buffet of pie charts in Excel!

 

Fiddler

Link: Fiddler

Fantastic piece of software whereby Internet Explorer and Chrome internet traffic is piped through the program (automatically) and you can clearly see every request/response that is happening along with all the HTTP header information and the time taken to download each of the different items on the page. Really useful program which I use on a regular basis.

If you are using this software and Firefox as your browser, then you need to configure Firefox to go through the program. It is a bit more hassle than it is worth to be honest, best to just use IE/Chrome which automatically pipes through Fiddler if it is installed.

 

Xenu

Link: Xenu

Nice scraping tool where you can simply enter in a domain, press go and receive a list of all of the URLs on the website. This works by spidering the website to identify all of the pages.

 

Screaming Frog

Link: Screaming Frog

Similar to Xenu where you can enter a domain and scrape all of the URLs on the site. The free version is limited to 500 URLs that you can scrape though, so good for small projects but probably not the best choice for larger projects.

 

Simple Sitemap Generator

Link: Simple Sitemap Generator

This is a tool which I built to help speed up the process of quickly generating an XML sitemap. The tool isn’t like other XML sitemap tools as it doesn’t crawl the website to identify all the URLs as there are other tools which do this already such as Xenu. Simply paste in a list of up to 50,000 URLs into the XML sitemap tool then click on generate to get your fully search engine friendly XML sitemap in no time at all.

 

 

Useful SEO  Blogs

 

Matt Cutts Blog

Link: http://www.mattcutts.com/blog/

Matt Cutts is one of the original engineers at Google and is the regarded as ‘the face of Google’. Want some information straight from the horses mouth about new developments? Then look no further. He doesn’t update his blog very often, but it is generally some good information when he does.

 

SEOmoz’s Blog

Link: http://www.seomoz.org/blog

SEOmoz’s blog is a great place to keep a close eye on for lots of useful SEO industry news. They also have a lot of guest blog posters who can provide some good content. In addition, they have a ‘Whiteboard Friday’ video every (you guessed it) Friday which is a must watch every week.

 

Google’s Official Blogs

There are quite a few official Google blogs for different areas so I will list some of the ones I find most useful. If the one you are looking for isn’t listed below try searching Google’s full blog directory. There are a lot of industry related blogs within the full directory so see if there is one for your industry.

Google Analytics Blog

Google Mobile Blog

Google Webmaster Central Blog – News on crawling and indexing websites

Google Inside Search Blog – The official Google search blog

Google Inside AdWords Blog

Google General Blog – Insights from Googlers into products, technology and the Google culture.

 

SEO By The Sea

Link: SEO by the sea

SEO by the sea is a great blog which tends to publish less mainstream topics which can be very useful. For example a recent post was about the 10 most important SEO patents and phrase based indexing which is well worth a read.

 

Stone Temple

Link: Stone Temple

Stone Temple’s blog is run by Eric Enge who tends to focus on interviews with high profile people within the search industry, such as Matt Cutts and Vanessa Fox. The interviews he does can be really insightful hearing all of the great information from the source instead of being paraphrased in blog posts as sometimes happens. Well worth keeping an eye out on.

 

Google Webmaster Central’s YouTube Channel

Link: Google Webmaster Help YouTube

Keep a close eye out on what content is uploaded to this channel, it is one of Google’s main way of  communicating new information and answering webmaster questions related to SEO.

It is good to subscribe to this channel so you are instantly updated when new videos are uploaded.

 

SEO Plugins for Google Chrome

 

Awesome Screenshot

Link: Awesome Screenshot

Great tool for capturing the whole of the page and not just the visual section. In addition, there is a nice simple editor to adding highlights and shapes to the screenshot you have just taken which saves a little bit of time editing it in another piece of software.

 

Nofollow Spotter

Link: Nofollow Spotter

Handy little tool to quickly identify any possible errors on the page, such as if everything is nofollowed by mistake. I usually always have this turned on whilst working as it is speeds up the checking process instead of having to run through the HTML.

 

Firebug

Link: Firebug

What a fantastic tool for quickly viewing certain parts of the HTML. In addition it can tell you the different header codes which are happening for each requested item on the page. Firebug also allows you to edit the HTML inline so you can add/delete certain parts of the page which can be useful when debugging something.

 

Google Analytics Debugger

Link: Google Analytics debugger

Extremely handy tool when wanting to test your new implementation of Google Analytics and to make sure everything is sent across to Google correctly. When using this tool you will see a ‘tracking beacon sent!’ message when everything has gone correctly.

Note, press ‘CTRL + Shift + J’ to activate.

 

Scrape Similar

Link: Scrape Similar

Great tool for easily and quickly scraping content from websites you are viewing. This can be useful to scrape a lot of URLs from Google or to scrape a lot of information from websites with a lot of listings on one page such as products or prices etc. Quickly copy this information into Excel then you can use it to compare with other data.

 

XPath Helper

Link: XPath Helper

Extremely useful tool for quickly getting the XPath of some content you want on the page. Press CTRL + Shift + X to activate while hovering over the content on the page you want to scrape. To change the content you can either edit the XPath directly within the tool or you can hold Shift down and hover over the content you want.

 

User-Agent Switcher

Link: User-Agent Switcher

This tool can be really useful to see how a website behaves when being accessed from different devices or browsers. Some websites, do some really crazy stuff in the background and this can help understand what is happening so you can recommend the correct solutions.

As the simplest example, you can select to view the website where using the default user-agent for your Chrome browser, and you can compare this with how it looks when you change the user-agent to an iPhone which you can see below. Since the recommendations will be slightly different depending on how things are technically set up, it is important to understand these areas.

 

 

Above: Amazon website while accessing via normal User-Agent

Below: Amazon website using iPhone User-Agent

 

Redirect Path

Link: Redirect Path

This is a great little tool that shows you the full redirect path in a really user friendly way. There are other tools that give you much more details, such as the Live HTTP Headers plugin for Firefox (mentioned below) if you want to dig deeper – but this tool is great for quick snapshots of information and it clearly displays within your toolbar when a redirect has happened (in case you blinked!).

Below is an example of how this displays the redirect path when a redirect(s) has happened;

 

 

SEO Plugins for Firefox

 

Firebug

Link: Firebug

As already talked about above, this is Firebug for Firefox.

 

Netcraft Toolbar

Link: Netcraft toolbar

Handy little toolbar showing different stats about the website including its age and where it is hosted. It is not something that I use every day but it is a handy tool when assessing lots of different websites.

 

Live HTTP Headers

Link: Live HTTP headers

Extremely useful tool for seeing what header code is being returned on a certain page, it can also help identify any 301 redirect chains.

 

Fire FTP

Link: Fire FTP

You may come across a time where your FTP access is limited on an internal network due to ‘security reasons’ so this is a great way to get around that. Simply install the Firefox plugin and away you go, your own basic FTP client bypassing any restrictions.

 

Server Spy

Link: Server Spy

Handy little tool to show what brand of HTTP server the website is running, such as IIS or Apache

 

People to follow on Twitter

Check out who I am following and follow them too :-) Be wary of people talking bollocks though, always best to use your own judgement and not to take everything as fact. Test for yourself, but Twitter is a great source of information in the SEO area.

@MickCropper

5 thoughts on “37 Essential SEO Tools And Resources”

  1. Great list of tools here, I can see that this list is being put together by someone who’s got on-the-job quality experience. the only other thing I would have added, would have been Majestic SEO, Screaming Frog and SEMRush.com.

    1. Hey Steve, thanks for the additions. I’ve personally not used the tools mentioned (although I’m aware of them), as use Open Site Explorer, Xenu and Searchmetrics instead which do similar jobs respectively. Lots of great tools within SEO to play around with that’s for sure!