Select Page

CCTLD vs DIR Pros and Cons

There is a lot of different discussions around the web about whether you should use a Country Code Top Level Domain (ccTLD) or a Directory when scaling a website up into multiple languages or countries, so I am going to give an overview of the pros and cons for each method.

The main question many people ask is that when moving into a new market is which should be used to get the maximum benefit? A country code top level domain or a directory?

 

Country Code Top Level Domain (ccTLD)

A country code top level domain is a domain that is bound to a specific country, such as .uk for United Kingdom, .au for Australia, .cx for Christmas Islands, .fr for France, .th for Thailand etc.

 

Pros of Country Code Top Level Domains

  • Clear signal to Google who the website is targeting. There is no ambiguity, when Google looks at your website www.example.com.au it is clear that the website is targeting Australian users. Usually in Google Webmaster Tools you can set which country you want your website to target, when your website is on a generic top level domain such as .com. Although when your website is on a ccTLD this information is not available since Google assumes (correctly or not) that you are targeting customers within that country.
Screenshot below showing generic top level domain settings in Google Webmaster Tools where you can set the country you would like to target.
Where as when you look at the same information in Google Webmaster Tools when you are on a ccTLD such as .uk this is what you see
  • Another benefit to having a ccTLD is that some users in certain countriesprefer to buy from those domains. For example people in Australia are extremely patriotic and prefer to buy from .com.au domains (company in Australia domain). Where as in somewhere like Thailand, people prefer to buy from a .com domain opposed to a .co.th (company in Thailand) domain since there can be a lot of poor quality websites on .co.th domains. So Thai people believe .com businesses will provide a better service or product. This isn’t always the case and there will always be people with different preferences, so when moving into a new market it is essential to do your own research into what would work best for your business.
  • It is a lot easier for a user to link to www.website.co.th than it is for a user to link to www.website.com/th/. By making this process simpler then it could lead to an increase in natural links for your website.
  • By having several country code top level domains it is possible to occupy more of the search results pages for branded search queries. For example if someone searched for “your brand” then it is possible that your .co.uk website, .com website and .co.th websites could all rank. Google is no where near perfect when it comes to language differentiation although it is getting better.

 

Cons of Country Code Top Level Domains

  • The obvious one here is that it kind of limits where you can operateyour business. If you are on a .co.uk domain, then it is unlikely that you are going to rank well in other markets due to how Google perceives ccTLDs as shown above. That said, if you have plenty of local links say from .fr domains, then Google isn’t going to totally ignore them. It is likely going to ignore the ccTLD and change its mind and say ‘even though the domain is a .co.uk domain, this website has lots of links from .fr domains, so it must be relevant for French users too’ as shown in the image below.  I have seen a lot of random and non relevant websites on other ccTLDs rank in strange places in the past, but I would put that down to Google’s poor algorithm.

 

  • One potential pitfall of using a country code top level domain is that all of the links from your main aren’t fully being utilised and there is no history to your new domain. That said, what most larger websites tend to do is link all of their ccTLD domains together so that each page in a different language links to all of the other languages, so www.website.com/en/pageone.html would also link through to www.website.com/fr/pageone.html, or www.website.co.th/pageone.html and www.website.com.au/pageone.html would like together and vice versa.
  • To successfully launch a new language / localised website then it is important to consider what marketing budget is available to really push the new website. It does require significant effort to launch a new website into another market so think carefully about the options. If the website is simply being translated in the hope that some traffic will magically appear then a ccTLD is probably not the best option, instead just use the main domain with the new languages in a directory.

 

Language Directories

The second option when scaling a site up into multiple languages is to use a directory for each language such as www.website.com/fr/ for France (or French) and www.website.com/th/ for Thailand (or Thai). One important point to think about here is if you are targeting a certain country or a certain language, or both?

Taking French as an example, if you are only targeting France as a country then you will be missing out on a massive market from other French speaking countries and this can be quite big. There are 28 countries around the world where French is a national language which include: Belgium, Benin, Burkina-Faso, Burundi, Cameroon, Canada, Central African Republic, Chad, Comoros, Democratic Republic of Congo, Djibouti, France, Gabon, Guinea, Haiti, Ivory Coast, Luxembourg, Madagascar, Mali, Monaco, Niger, Republic of Congo, Rwanda, Senegal, Seychelles, Switzerland, Togo and Vanuatu.

Another option here is to target both language and country by providing a full localised directory such as www.website.com/ca-fr/ for Canadian people who speak French and www.website.com/be-fr/ for Belgium people who speak French.

 

Pros of Language Directories

  • One of the main pros to having all new languages in a directory is that all of the inbound links to your website will flow through to the new language pages. That said, this would also happen when linking your different ccTLDs together correctly although is it not fully known if both of these methods would pass the same amount of Google Juice through.

 

Cons of Language Directories

  • One of the potential pitfalls of having languages in a directory opposed to on a ccTLD is that Google could confuse the site as being a total mess in multiple languages. Although this isn’t likely to happen and you can set each directory to be targeting a specific country within Google Webmaster Tools as shown above.
  • It is possible that for branded search queries that you will find your main pages ranking in the incorrect language which isn’t a great user experience. For example if someone is searching for “your brand” from France then it is very likely that the main English website will rank since the English page has far more Google Juice strength and it can be difficult to get this correct.
  • The issue of what content you should display on the ROOT of the domain, www.website.com? Should it be English or one of the other languages? Then this takes you into the realm of potentially redirecting users either based on the IP address or on their language settings from the browser. Bearing in mind that the incorrect language could also show as a snippet in the search results which would dramatically reduce the click through rate.
  • If you are thinking of automatic redirects for users then you may havepotential Google Bot issues with cloaking. There is no clear info from Google as to what is best here and they have even totally contradicted their self several times on this issue. There is also the possibility that if the automatic redirects haven’t been set up correctly that you could accidentally block Google from accessing most of your website.  For example since Google Bot always crawls from America then you wouldn’t want to be redirecting Google Bot to the English page all of the time since it would never discover any of the other content!
  • People within the market you are entering are likely to link to the main website and not the localised directory, for example you will find a lot of people would link to www.website.com opposed to www.website.com/th-th/ which could hinder how well the language directories rank within the local search engines.

 

Summary

There is no quick win to breaking into a new market and there are a lot of pros and cons to each of the different methods, country code top level domains and directories. Personally I would prefer to use ccTLDs where possible if there is going to be a real push into that market.

If you are just thinking about translating and not promoting then it is a waste of money, similar to buying a phone and expecting it to ring on its own! Nothing worth while is easy, so it is going to take considerable effort to break into the market.

You may have noticed that I have not mentioned the option for sub-domains within this post and that is because I don’t like sub-domains :-)But if you are interested then I am sure you can pull some of the above points for/against sub-domains too.

Hopefully this has provided you a wide range of information on the topic and will help you decide what is best for your website.

Here is a summary table outlining all of the different points listed above

 

The 200 Signals Used In Google’s Ranking Algorithm

Whilst I would love to think I knew the full 200+ signals being used in Google’s ranking algorithm, I don’t think I would quite get to 200 on my own. :-) I am asking for help from the SEO community in this post to share their knowledge on how Google rank websites.

This post is designed to list all of the possible data sources / signals Google could be using as part of their ranking algorithm. My feeling is that if they have access to data, then they will be using it in someway or another. Weather this is to help boost websites or to flag websites as spam and demote them. We aren’t going to be discussing the actual weight Google is giving to each of the different signals (that can be for another post :-) ), just to see if we can get all 200 (or more) signals theycould be using.

I have broken down the main areas to help categorise the different signals. Anyone who can add to this list will get a link back to their blog (as long as you don’t have a spammy website!) since this is my way of saying thanks for contributing.

Anyone wishing to contribute can add a comment to this blog post and I will update the blog post. When commenting please state;

  • What piece of information Google could be using
  • How they would judge / rank this information (I.e. how does this show quality or not?)
  • Where they would get this data from
  • Any relevant official sources of information to support your theory

A few of the easy ones have already been put in :-) I want this post to be generated by the SEO community to see how many we can get between us all.

So lets get started then!

On Site Signals

  1.  Meta title
  2.  H1 & heading tags
  3.  Keywords within the text
  4.  Amount of useful text when subtracting templated information and adverts
  5.  Phone number, which signifies that the website is a genuine business
  6.  Amount of unique images which cannot be found anywhere else
  7.  Amount of unique high quality content
  8.  Is there regular fresh content appearing on the website?
  9.  Internal linking
  10.  Site speed
  11.  Image ALT attribute
  12.  Amount of adverts on the page & how much useful content is left once the adverts are removed
  13.  Video markup
  14.  Microformats and rich snippets such as hreview or hrecipe
  15.  IP C class and location of server
  16.  Content above the fold VS content below the fold
  17.  Top level domain
  18.  How long the domain has been owned by the current owner, not just how long the domain has been registered to ‘someone’
  19.  Keyword over usage or keyword stuffing (negative factor)
  20.  Cloaking (negative factor)
  21.  Hidden text (negative factor)
  22.  Duplicated content from external websites
  23.  301 redirect flags such as redirect chains (301 –> 301 –> 301 –> Final destination), redirect loops or redirects ending on a 404 page.
  24.  Keyword rich URLs opposed to page numbers or product ID’s
  25.  Number of crawl errors found when crawling the site
  26.  Does the HTML conform to W3C standards

Off Site Signals

  1.  Anchor text for inbound links
  2.  Number of linking root domains
  3.  PageRank from inbound links
  4.  Followed / Nofollowed attributes on inbound links
  5.  Number of mentions (ie, ‘website.com’ without an actual link)
  6.  Total number of external links in addition to linking root domains
  7.  Surrounding text on external links
  8.  IP C class of linking websites
  9. Existence of a Google places profile
  10.  Link profile of competitor websites
  11.  Age of the back link
  12.  Link growth rate and how fast links to certain pages have gained links
  13.  Evidence of paid links
  14.  Selling ‘followed’ links to other websites (negative factor)
  15.  History of comment spamming on forums, blogs or other link spam (negative factor)

Brand Signals

  1.  Does the website have a brand logo which appears throughout the website?
  2.  Does the brand logo appear anywhere else on the web?
  3.  Does the website send out branded emails to its customers? This data could be gathered by analysing gMail accounts.
  4.  How many gMail accounts are receiving, opening and engaging with emails sent from this brand website?
  5.  Does the website have a LinkedIn company page with real people working at the company?
  6.  Search volume for branded keywords and brand name
  7. Presence of a physical address on the website
  8.  Age of the domain
  9.  Number of branded external links

Social Signals

  1. Does the website have a Twitter account?
  2. Does the website have a Google+ brand page?
  3. How many Plus 1’s does the website/web page have?
  4. How many Facebook fans does the brand page have?
  5. How many people are Tweeting about the website?
  6. How many Twitter followers the website has
  7. Surrounding text of links in tweets
  8. Number and rate of growth for page views and replies on social media pages
  9. Presence of user generated content (UGC) on the website
  10. Growth rate of social media mentions, such as a slow growing amount of mentions or a large outbreak

Other Signals

  1.  Is the website advertising on AdWords?
  2.  Is the website advertising on Google Display Network?
  3.  Is the website listed on any official business websites such as Business Link or Yell.com?
  4.  Does the website have a physical address?
  5.  Is there an email address where customers / website visitors can contact the owner?
  6.  Click through rate (CTR) from the search results
  7.  Bounce rate from the search results
  8.  Review rating of the website / business on third party websites
  9.  Google seller reviews for eCommerce websites
  10.  Quantity of pages in Google’s index
  11.  History of past penalties for this domain (negative factor)
  12.  History of past penalties from this owner of the domain (negative factor)
  13.  History of past penalties for other domains from the same owner (negative factor)
  14.  Amount of time spent on the page / site
  15.  Number of page views per visit

 

Please add your comments so we can get this list full. I would be surprised if we couldn’t fill this list with all the knowledge out in the SEO industry.

Contributors

Michael Cropper – SEO and Internet Geek (1, 2, 3, 4, 5, 6, 7, 8, 41, 42, 43, 44, 45, 81, 82, 83, 84, 85, 121, 122, 123, 124, 125, 161, 162, 163, 164, 165)

Felix Lueneberger (9, 10, 11, 12, 13, 14, 15, 16, 17, 46, 47, 48, 49, 50, 86, 87, 88, 89, 126, 127, 128, 129, 130, 166, 167, 168, 169, 170)

John – Forest Software (18, 51, 52, 53)

Sean – http://01100111011001010110010101101011.co.uk (19, 54, 55, 20, 21, 22, 171, 172, 173, 23)

Jeremy Quinn (24)

Brahmadas from SEOZooms (25, 26, 174, 175)

 

12 thoughts on “The 200 Signals Used In Google’s Ranking Algorithm”

  1. Interesting approach ;-).
    I’d like to add:

    “on site”:
    -internal linking (!)
    -site speed
    – TLD
    -image alt tags
    – amount of ads on the page
    -video markup
    -microformats such as hreview or hrecipe
    -IP C class, location of server
    -content above Vs. below the fold

    “off site”:
    – total number of external links (in addition to LRDs already mentioned)
    – surrounding text of external links
    – IP C class of linking sites
    – Google places profile exists
    – Link profile of competitor sites

    “brand signals”:
    -search volume of brand name
    – physical address on page
    – age of domain / site
    – number of branded external links

    “social signals”:
    – how many twitter followers does the site have
    – Surrounding text of links in tweets
    – what is the rate of pageviews / replies on social media
    – is there UGC on the site?
    – how fast is the growth of social mentions (slwoly growing or outbreak?)

    “other signals”:
    – CTR from SERPs
    – bounce rate from SERPs
    – Reviews rating on third party sites
    – Google seller reviews (for e-commerce sites)
    – number of pages in Google index

  2. I would add :

    Onpage :
    – Age of domain (although I see that that is in Brand signals) more specifically, how long it has been owned by “you” not just registered

    External links :
    – How old the link is
    – how quickly these links have been obtained
    – any evidence that links have been paid for

    1. Of course it is, what else would this be. :-) It’s also good to see what people within the SEO community think could be being used as a ranking signal.

      But the real question is…are you the real Matt Cutts? (It would be awesome if you are and knowing that you are reading my blog :-) But it could also be a fake)

      Update: The above comment is not from Matt Cutts, the real Matt Cutts has clarified.

  3. What about Crawl rate, crawl errors, validation, page overal visibility, user side reputations, bounce rate, time on site and pages, number of page views . there are many things I think…. so it may not be 200+ may be more than 300 or 400. for the time being and Google may identify and add many and many new factors in the future so that the SEO game will in more tough plat form….. he he

    1. Technically correct. But it is a common understanding that when one says “Meta Title” that they are talking about the <title> tag and not the old skool tag which is never used: <meta name=”title” content”something”/>

  4. Think that is jumped ! one or serious forget to attend skool before attending college

    Any way nice post Mick. Keep it up. Think more time will take to fill the blank space of signals. Or the Google God is not getting right signals now? Regards

    1. Thanks for sharing Robert, the website provides a nice scientific approach whereas this list is a nice quick reference point based purely on peoples opinions and experience within the industry.

3 Essential Excel Plugins for SEO

Here I am going got talk through some of the Excel plugins I have found very useful for SEO tasks. Best thing I can advise is to get them installed and give them a go for yourself. The three plugins to look through today are;

  1. Regular expressions plugin for Excel
  2. SEO Tools plugin for Excel
  3. Excellent Analytics plugin for Excel

 

Regular Expressions Plugin for Excel

Link: Regex Plugin Excel

This can be a great plugin when working with large data sets within Excel as it allows you to use all of your normal regular expressions to find the data you need. See this site for a regular expressions cheat sheetfor reference, alternatively if you aren’t too familiar with regular expressions then I suggest reading the following tutorial.

 

There is not much more to say about the regular expressions plugin for Excel apart from that you can filter out a lot more than you can using the normal string matches which Excel has be default. It can be a lot easier to get the information you require.

 

SEO Tools Plugin for Excel

I briefly mentioned this in an earlier post about why programming is an essential SEO skill but I wanted to expand on the amazing excel plugin which is SEO Tools by Neils Bosma and why you need to be using this if you aren’t already.

Below is a screenshot of the additional tab you get in Excel once you have installed the plugin (60 second install, follow the instructions in the ‘read me’ file!)

 

 

So what are some of the cool things you can do with SEO Tools once it is installed? Well an awful lot actually.

Check the HTTP Status of pages

Simply add a lot of different URLs in column A then use the function=HttpStatus(A2) to find out if it is returning a 200, 404, 503 etc. This can be a really useful tool for both SEO since Google Webmaster Tools reports on errors for your site for a long time, often when these pages are actually no longer showing any errors. So it can be a good method to list all of the URLs which Google Webmaster Tools is showing as errors and double checking these to spot the real pages that are showing as errors.

To find this function you can simple click on the ‘onpage’ button highlighted and select the HttpStatus() option from the list

 

Filter through back link targets

Lets say you have a list of websites / webpages you want to target for back links but it isn’t easy to figure out which ones to target first. From a large list, it is important to be able to filter through these to identify the best websites to contact first. There is no point in contacting poor quality websites first!

Maybe you want to prioritise websites to contact who have “SEO” in their meta title and have a page rank of at least 1

 

Identify popular content on competitor websites

How about trying to identify which content is the best on competitor websites? Maybe a competitor has a blog or a content section which you can easily scrape all of their URLs from. Simply go to Google, view 100 results and search for “site:blog.website.com” then use Scrape Similar to get all of the URLs where you can paste these into Excel.

Once you have all of these URLs then you can check how popular they have been on the social media channels. In the screenshot below I have pulled a few of my blog posts and identified how many tweets they have had recently using the =TwitterCount(A2) function from SEO Tools.

 

Whilst I am not a massive fan of copying off competitors (I believe you should be doing your own thing far better so they are copying off you!) this can be a useful tool to get a rough idea of different content that people are interested in. It could even be a good guide to identifying your own ‘linkable content’ since Google Analytics can only show the amount of traffic that each of the pages gets.

So looking at the above screenshot, if those websites were for competitors then I would think about doing a post about how to see competitors social media link statistics since this has gained the most traction.

 

Website Scraping with XPathOnURL

This is one of the most powerful bits of SEO Tools and one of the best bits with this Excel plugin. I am not going to go into too much detail here about XPathOnURL since this is a full blog post on its own. To give you an idea of what it can do though it basically scrapes certain bits of the HTML that you need such as “get me all the links on this page” which would translate to “get me all the HREF attributes within all of the <a> tags on the page”.

This tool can be extremely useful when scraping websites for certain pieces of content / information etc. Have a play, see what you can do.

 

Excellent Analytics

Link: Excellent Analytics

This can be a really useful Excel plugin for SEO since it allows you to export all the data you need from Google Analytics straight into Excel. So no more exporting large data sets then matching up the relevant data with a =VLOOKUP() function instead, just get the precise data you need from GA straight into the adjacent column in Excel where your other data is.

I am not going to cover this in a large amount since it is quite self explanatory, it pulls in data from Google Analytics based on all the normal segments, filters and visitor data you are used to using via the normal Google Analytics dashboard. Simply link up your GA account with the plugin and away you go. Below is a screenshot showing one of the user interfaces.

 

On larger websites which 10’s of millions of page views per month you may have noticed how slow Google Analytics is to do anything, so this can help speed this process up for extracting the data you need.

 

Summary

These are the three essential Excel plugins SEO I use on a regular basis. I have found them to be real time savers for various tasks as outlined above so go and give them a go.

If you have any other excel plugins that you use for SEO then please leave a comment.

 

2 thoughts on “3 Essential Excel Plugins for SEO”

How To Create An Embed Code For Your Infographic

Here I am going to show you how easy it is to create your own embed code for either your widgets or infographics you have created. They are an essential tool if you want to help build natural links.

This is what the final piece of code looks like to a user, with the images / info graphic / widget above and the text to embed this below.

 

Click below to embed the above image into your website

 

So how do you do that? Well it is quite simple really, just a few bits of HTML and JavaScript will enable you to create a box with your embed code in so people can place your images / infographics / widgets on their site with that all important link back to your website as the original source.

The first step is to include the image on the page you are wanting the embed code to be one. Once you have done that you simply add the code below to your HTML and tweak the relevant bits which will be explained in a moment.

 

Complete Embed code:

 <textarea style=”height: 44px; width: 300px; margin: 2px;” onclick=”this.focus(); this.select(); _gaq.push([‘_trackEvent’, ‘Infographic’, ‘Click’, ‘guerilla-marketing-infographic’]);” rows=”2″>

&lt;img src=http://www.michaelcropper.co.uk/wp-content/uploads/2011/09/guerilla-marketing-at-its-best-in-manchester.jpg width=300 height=300 /&gt;

&lt;br&gt;

View full image

&lt;a href=http://www.michaelcropper.co.uk/wp-content/uploads/2011/09/guerilla-marketing-at-its-best-in-manchester.jpg title=”Guerilla Marketing Manchester” style=”color:#0000FF;text-align:left”&gt; Guerilla Marketing Manchester&lt;/a&gt;

</textarea>

 

So if we take a look through each of the different sections to explain what each part does.

 

Step 1 – Create a text box

Create a text box on the screen which is 44 pixels high, 300 pixels wide and has a bit of a margin. This section also has some JavaScript so when a user clicks (onclick section) the text box then the text is highlighted (this.focus(); this.select() section) which makes it easier for the user to then right click and copy the text. In addition, the code below also has an event tracking Google Analytics tag which means that when someone clicks the text box an event is fired off to the Google Analytics account you have installed on the page which says “someone has clicked an ‘infographic’ which is called ‘guerilla-marketing-infographic’ “

Below the code is what the code actually creates, the text box (although yours will be empty to begin with until the next step.

 

<textarea style=”height: 44px; width: 300px; margin: 2px;” onclick=”this.focus(); this.select(); _gaq.push([‘_trackEvent’, ‘Infographic’, ‘Click’, ‘guerilla-marketing-infographic’]);” rows=”2″>

 

If interested, read full details about Google Analytics event tracking, but in summary is _gaq.push([‘_trackEvent’, ‘category‘, ‘action‘, ‘label‘]);where you can choose the category, action and label you want to use.

 

Step 2 – Add the Image

You’ve got something to embed right? Well this is the next step, place the HTML code of your original image / infographic / widget in between the <textarea> and </textarea> tags. One key thing to note though is that you need to replace HTML characters such as < and > with their HTML code equivalent. Take a look through a full list of HTML codes for reference. It can be a bit of a pain to do this, but hopefully you don’t have much to include here!

 

&lt;img src=http://www.michaelcropper.co.uk/wp-content/uploads/2011/09/guerilla-marketing-at-its-best-in-manchester.jpg width=300 height=300 /&gt;

 

Step 3 – Add the Link Back to You

Here is the most important part, adding a link back to your website as the original source of the content. Same rules apply as in step 2 where by you need to replace the HTML characters with their HTML codes. The great thing here is that you can choose the anchor text which is used when people are linking back to you.

 

View full image

&lt;a href=http://www.michaelcropper.co.uk/wp-content/uploads/2011/09/guerilla-marketing-at-its-best-in-manchester.jpg title=”Guerilla Marketing Manchester” style=”color:#0000FF;text-align:left”&gt; Guerilla Marketing Manchester&lt;/a&gt;

 

I have seen some people include cheeky links within this section previously which link back to one of their key pages they want to rank with the best anchor text. Whilst this isn’t a bad idea, personally I would only add it where relevant. A lot of people will likely remove the link anyways, but it is certainly worth a try :-)

 

Step 4 – Drive Traffic to Share Your Content

Now comes the easy bit! Go out there and tell people about the great content you have just created. If the content truly is great then it is going to make people want to embed this item in their own websites and tell their friends about it naturally.

Here is what the final piece will look like when a user embeds the code in their own website:

 

So there we go, simple as that! Now time to go and create some great content and promote it well. For anyone who is interested in what the massive gorilla is, then have a read of my guerilla marketing Manchesterpost.

Here is one we use in our info graphic series for A day out on the District line in London. Great option for promoting some great content.

Note: You may also need to wrap ” ” quotes around some of the HTML tags within the embed code itself. Strangely when I was creating this post additional quotation marks were being added around the IMG SRC and A HREF attributes causing invalid URLs, so if you come across this then just tweak accordingly.

 

10 thoughts on “How To Create An Embed Code For Your Infographic”

  1. Great code, thanks.

    The speech marks are formatted so it didn’t work for me until I changed them:

    ” – didn’t work (ASCII)

    ” – did work (type into notepad)

    Hope that helps someone. :)

  2. I have been trying to edit the code so it displays properly but I am having issues. I don’t know what I am not replacing correctly

    this code doesnt work but i gave it a valiant effort. if you can provide any help i would appreciated it! thanks for the great code anyways

    1. Hi Joffrey,

      Your code didn’t come through properly, likely because WordPress filters certain things out in comments by default. If you want to explain your issue in the comments or email me with more details then I will see if I can point you in the right direction.

      Thanks
      Michael

  3. Hi I am trying to help my friend do this on her friends website here.
    I have tried following the instructions, tried changing the ” and at this point in time, I can’t figure it out.

    It’s a wordpress site and I owe her a favour so it would be really good, if I could get this working for her.

    Thanks in advance

    Peter

37 Essential SEO Tools And Resources

There have been hundreds if not thousands of blog posts on SEO tools before, so I am not going to waffle on about why you should use each one. The main reason for this blog post is more of a reference manual for me  ????

These are the tools that I have found invaluable being an SEO so I hope they are of some use for anyone reading too.

 

Google Webmaster Tools

Link: Google Webmaster Tools

Great SEO tool to easily spot duplicate content, duplicate title tags, malware on the site, HTML issues and URLs Google has found that are either 404’ing or 500/503’ing.

This is a great starting point to diagnose websites to see what is potentially going wrong with them and it can also be a great way to spot malformed back links that you can 301 to the correct place.

 

Google Analytics

Link: Google Analytics

What a powerful SEO tool. I am not going to go into details about why this is so amazing, but if you are not using this (or an equivalent analytics package) then you need to get this installed and find out for yourself why it is so amazing.

 

Microsoft Excel

You need to make sure you have the full version of Excel and not just the free version that is available since there are a lot of additional features you will need to use such as advanced filtering and customising your Excel with some nice plugins.

When using Excel I also have additional plugins installed which help with more advanced functions and save a lot of time. I did another blog post about essential Excel plugins for SEO recently which goes into more detail about these three plugins.

  1. Regexp Plugin
  2. SEO Tools
  3. Excellent Analytics

 

Google Docs

Link: Google Docs

The main reason I use Google Docs is for their spreadsheets. This can be a great way to share certain bits of information between a team of people, may that be list of websites that have been contacted for links if there are teams of link builders or it can also be used for more creative reasons such as the ImportXML function that is available.

I tend to use this in addition to Excel since there are some things that Google Docs can do better than Excel, the main one being that Excel is limited to importing only one XML feed URL per document (without a lot of effort tweaking things and writing Macros at least) whereas Google Docs has a nice limit of 50 XML feed URLs that can be pulled in within one spreadsheet, which can be great when scraping content from other websites.

 

Open Site Explorer

Link: Open Site Explorer

I briefly mentioned this earlier, but this tool was created by SEOmoz and is based on their Link Scape index. Basically SEOmoz have Link Scape which crawls the web just like any other search engine would do and then indexes the information. From this information they then provide all of the juicy statistics that SEOs like to get their Excel spreadsheets on.

Go and have a play with what is available, but to get the full data set then you need to sign up to become a Pro Member which is well worth it for anyone doing a lot of SEO.

One thing to bear in mind is that it does provide a lot of data and one could argue that it is just data and only useful if there can be some actionable insights that can be gained. I have seen a lot of SEOs in the past creating reports, pie charts, graphs of all this wonderful data and comparing it with competitors without anything coming from this. So just make sure you know why you are gathering and analysing all of the data before you get started or you will just end up with a buffet of pie charts in Excel!

 

Fiddler

Link: Fiddler

Fantastic piece of software whereby Internet Explorer and Chrome internet traffic is piped through the program (automatically) and you can clearly see every request/response that is happening along with all the HTTP header information and the time taken to download each of the different items on the page. Really useful program which I use on a regular basis.

If you are using this software and Firefox as your browser, then you need to configure Firefox to go through the program. It is a bit more hassle than it is worth to be honest, best to just use IE/Chrome which automatically pipes through Fiddler if it is installed.

 

Xenu

Link: Xenu

Nice scraping tool where you can simply enter in a domain, press go and receive a list of all of the URLs on the website. This works by spidering the website to identify all of the pages.

 

Screaming Frog

Link: Screaming Frog

Similar to Xenu where you can enter a domain and scrape all of the URLs on the site. The free version is limited to 500 URLs that you can scrape though, so good for small projects but probably not the best choice for larger projects.

 

Simple Sitemap Generator

Link: Simple Sitemap Generator

This is a tool which I built to help speed up the process of quickly generating an XML sitemap. The tool isn’t like other XML sitemap tools as it doesn’t crawl the website to identify all the URLs as there are other tools which do this already such as Xenu. Simply paste in a list of up to 50,000 URLs into the XML sitemap tool then click on generate to get your fully search engine friendly XML sitemap in no time at all.

 

 

Useful SEO  Blogs

 

Matt Cutts Blog

Link: http://www.mattcutts.com/blog/

Matt Cutts is one of the original engineers at Google and is the regarded as ‘the face of Google’. Want some information straight from the horses mouth about new developments? Then look no further. He doesn’t update his blog very often, but it is generally some good information when he does.

 

SEOmoz’s Blog

Link: http://www.seomoz.org/blog

SEOmoz’s blog is a great place to keep a close eye on for lots of useful SEO industry news. They also have a lot of guest blog posters who can provide some good content. In addition, they have a ‘Whiteboard Friday’ video every (you guessed it) Friday which is a must watch every week.

 

Google’s Official Blogs

There are quite a few official Google blogs for different areas so I will list some of the ones I find most useful. If the one you are looking for isn’t listed below try searching Google’s full blog directory. There are a lot of industry related blogs within the full directory so see if there is one for your industry.

Google Analytics Blog

Google Mobile Blog

Google Webmaster Central Blog – News on crawling and indexing websites

Google Inside Search Blog – The official Google search blog

Google Inside AdWords Blog

Google General Blog – Insights from Googlers into products, technology and the Google culture.

 

SEO By The Sea

Link: SEO by the sea

SEO by the sea is a great blog which tends to publish less mainstream topics which can be very useful. For example a recent post was about the 10 most important SEO patents and phrase based indexing which is well worth a read.

 

Stone Temple

Link: Stone Temple

Stone Temple’s blog is run by Eric Enge who tends to focus on interviews with high profile people within the search industry, such as Matt Cutts and Vanessa Fox. The interviews he does can be really insightful hearing all of the great information from the source instead of being paraphrased in blog posts as sometimes happens. Well worth keeping an eye out on.

 

Google Webmaster Central’s YouTube Channel

Link: Google Webmaster Help YouTube

Keep a close eye out on what content is uploaded to this channel, it is one of Google’s main way of  communicating new information and answering webmaster questions related to SEO.

It is good to subscribe to this channel so you are instantly updated when new videos are uploaded.

 

SEO Plugins for Google Chrome

 

Awesome Screenshot

Link: Awesome Screenshot

Great tool for capturing the whole of the page and not just the visual section. In addition, there is a nice simple editor to adding highlights and shapes to the screenshot you have just taken which saves a little bit of time editing it in another piece of software.

 

Nofollow Spotter

Link: Nofollow Spotter

Handy little tool to quickly identify any possible errors on the page, such as if everything is nofollowed by mistake. I usually always have this turned on whilst working as it is speeds up the checking process instead of having to run through the HTML.

 

Firebug

Link: Firebug

What a fantastic tool for quickly viewing certain parts of the HTML. In addition it can tell you the different header codes which are happening for each requested item on the page. Firebug also allows you to edit the HTML inline so you can add/delete certain parts of the page which can be useful when debugging something.

 

Google Analytics Debugger

Link: Google Analytics debugger

Extremely handy tool when wanting to test your new implementation of Google Analytics and to make sure everything is sent across to Google correctly. When using this tool you will see a ‘tracking beacon sent!’ message when everything has gone correctly.

Note, press ‘CTRL + Shift + J’ to activate.

 

Scrape Similar

Link: Scrape Similar

Great tool for easily and quickly scraping content from websites you are viewing. This can be useful to scrape a lot of URLs from Google or to scrape a lot of information from websites with a lot of listings on one page such as products or prices etc. Quickly copy this information into Excel then you can use it to compare with other data.

 

XPath Helper

Link: XPath Helper

Extremely useful tool for quickly getting the XPath of some content you want on the page. Press CTRL + Shift + X to activate while hovering over the content on the page you want to scrape. To change the content you can either edit the XPath directly within the tool or you can hold Shift down and hover over the content you want.

 

User-Agent Switcher

Link: User-Agent Switcher

This tool can be really useful to see how a website behaves when being accessed from different devices or browsers. Some websites, do some really crazy stuff in the background and this can help understand what is happening so you can recommend the correct solutions.

As the simplest example, you can select to view the website where using the default user-agent for your Chrome browser, and you can compare this with how it looks when you change the user-agent to an iPhone which you can see below. Since the recommendations will be slightly different depending on how things are technically set up, it is important to understand these areas.

 

 

Above: Amazon website while accessing via normal User-Agent

Below: Amazon website using iPhone User-Agent

 

Redirect Path

Link: Redirect Path

This is a great little tool that shows you the full redirect path in a really user friendly way. There are other tools that give you much more details, such as the Live HTTP Headers plugin for Firefox (mentioned below) if you want to dig deeper – but this tool is great for quick snapshots of information and it clearly displays within your toolbar when a redirect has happened (in case you blinked!).

Below is an example of how this displays the redirect path when a redirect(s) has happened;

 

 

SEO Plugins for Firefox

 

Firebug

Link: Firebug

As already talked about above, this is Firebug for Firefox.

 

Netcraft Toolbar

Link: Netcraft toolbar

Handy little toolbar showing different stats about the website including its age and where it is hosted. It is not something that I use every day but it is a handy tool when assessing lots of different websites.

 

Live HTTP Headers

Link: Live HTTP headers

Extremely useful tool for seeing what header code is being returned on a certain page, it can also help identify any 301 redirect chains.

 

Fire FTP

Link: Fire FTP

You may come across a time where your FTP access is limited on an internal network due to ‘security reasons’ so this is a great way to get around that. Simply install the Firefox plugin and away you go, your own basic FTP client bypassing any restrictions.

 

Server Spy

Link: Server Spy

Handy little tool to show what brand of HTTP server the website is running, such as IIS or Apache

 

People to follow on Twitter

Check out who I am following and follow them too :-) Be wary of people talking bollocks though, always best to use your own judgement and not to take everything as fact. Test for yourself, but Twitter is a great source of information in the SEO area.

@MickCropper

5 thoughts on “37 Essential SEO Tools And Resources”

  1. Great list of tools here, I can see that this list is being put together by someone who’s got on-the-job quality experience. the only other thing I would have added, would have been Majestic SEO, Screaming Frog and SEMRush.com.

    1. Hey Steve, thanks for the additions. I’ve personally not used the tools mentioned (although I’m aware of them), as use Open Site Explorer, Xenu and Searchmetrics instead which do similar jobs respectively. Lots of great tools within SEO to play around with that’s for sure!

How To Get Personalised Images On Blog Comments

Recently I have had a few people ask me about how to get your own personalised image when commenting on blog posts so I thought I would write a bit of information on how to do this, as it is extremely simple. Below is a screenshot of what your blog comments will look like once you have completed the simple steps outlined in this blow post.

 

 

All you need to do so that your image will show next to all blog comments is to sign up to Gravatar (aka Globally Recognised Avatar – see where the name came from?). Within Gravatar you can personalise the image you would like to display and it is as simple as that!

Then next time you leave a comment on a blog post, the email address you have to enter when submitting a comment will be referenced with Gravatar and your image pulled in from there. Simple.

 

This technique works when commenting on all WordPress blogs and many other large blogs around the web. Great for branding yourself when commenting on blog posts.

So there you have it, a quick process which shows you how to get personalised images on blog comments.

Google Introduces No Panda Penalty Meta Tag

With all the recent Panda issues that Google has caused throughout 2011, they have recently announced a new tag that will help websites recover from a Panda penalty when they believe they have been incorrectly punished (aka Pandalised)

All you need to do to implement this ‘quick fix’ is to add the following meta tag to your website across all pages that have been impacted, between the <head> and </head> tags within your HTML. The next time Google crawls your website and can see the ‘no Panda penalty’ meta tag it will update its index accordingly and you should begin to see your rankings improve again.

 

<meta name=”panda” content=”no-penalty”/>

 

Here is what happened when I introduced this onto a site I have been working on. Sorry I cannot provide the website URL, so you will have to take my word for it. :-)

 

 

Wow! A full recovery from a simply one line change in the code.

Well if you have got this far reading the post, let me just add a small disclaimer. This post is a JOKE, so don’t go running to your developers asking them to implement this!

There have been a lot of stupid posts around the web recently about how websites have recovered from Panda by making silly little changes.  Upon further looking at these they turn out to either be nothing valid since the site returned at another Panda update (ie, it was a tweak in the algorithm that caused the positive change opposed to any work they have done), or they won’t provide details about the website in question and the specific details/data behind their claim.

Sorry if I got your hopes up here about the ‘quick fix’ for Panda, but the reality is that to get your rankings back then you need significantly better content throughout your website. Check out the recent epic blog post I did about how to recover from Google Panda for some real advice.

 

4 thoughts on “Google Introduces No Panda Penalty Meta Tag”

    1. …… ????

      Panda is an iterative update, so maybe you just need to wait until the next iteration ????

      < / joke >

      (sorry, I just have to keep saying that this meta tag is a joke as I know there will be someone, somewhere, who will actually try to implement this thinking it will fix issues – hell, I have seen enough other crazy ideas around the subject, people will try anything!)

Let Me Share Your Content!

One of my pet hates about websites is when they make it difficult or impossible to share their content on social media. Please webmasters & companies, come into the 21st century and understand how important sharing content is. Lets take a look through a few different websites and what sharing options they have on their site.

Websites With Poor Social Sharing Options

PC World

Looking at the product pages on PC World, the only options for a user are either to print or email the page. They are missing a massive opportunity by not including some of the more social buttons such as AddThis, Facebook Like button or the Tweet button.

 

Expedia

Expedia do a little better than PC World since they offer the ability to Like / Send buttons from Facebook, although what about if I want to share the content on Twitter or StumbleUpon?

JJB Sports

JJB Sports have a better implementation of social sharing buttons than the previous websites listed as you can see below. They have included the AddThis button, Tweet button, Google Plus One button and the Facebook Like button. It doesn’t appear to have gained much traction so far since most of the products I viewed had no social shares, although I did only look through a handful of random product pages.

Google

With Google having recently launched their own social media platform, Google Plus, their Plus One button has been showing up everywhere and almost forcing people to begin to Plus One different websites. Here is an example of one of their Google Webmaster Guidelines posts and you can see that they have actually been nice to the other platforms (for a change!) and given users the ability to share the content on a range of platforms.

They have also integrated the Plus One button into the search results to try and get more people to use it, although more recently this has disappeared and will only appear when you mouse over a listing.

Whether you agree with how Google is pushing their own services or not, this is a good implementation of social sharing.

Website With Great Social Sharing Options

 

Now we can take a look at some websites who have nailed social sharing on the site, so much so that the number of shares their content gets of phenomenal.

SEO Book

SEO Book do a fantastic job of making social sharing buttons very visible, which is one of the reasons why their content is shared a lot. Below is an image of their recent blog posts and you can see it has had over 700 social shares, and likely many more which aren’t directly listed.

 

Search Engine Land

Search Engine Land do a good job of social sharing buttons as well, with the main sharing buttons placed above the story then additional sharing buttons offered by Share This at the end of the article. This way, the most frequently used buttons are easily identifiable without cluttering the page and the additional less used ones are still available.

Main social sharing buttons;

Other social sharing buttons at bottom of page;

Search Engine Round Table

Search Engine Round Table also do a great job of social sharing in a very similar way to how SEO Book implement it. They also list the main social sharing buttons with the additional option to share more content via theAdd This button. The screenshot below was taken approximately 6 hours after the content was put online, so not bad to generate 60+ social shares within that time. I certainly couldn’t build that many links manually in such a short space of time!

 

Hotels-London

Over at Hotels-London.co.uk we have recently introduced some fantastic (if I do say so myself!) social sharing options so a user can easily see how to share the content throughout the site. The image below is a toolbar which is always visible at the bottom of every page and allows customers to create nice friendly shortlists and view recently viewed hotels. Go check it out, its ace :-)

We also have the Add This button installed throughout the site in prominent places which allows the user to share via other popular social services.

Why Add Social Buttons?

If you website doesn’t excel at social sharing or you are still in doubt about why you need to be using social sharing buttons then here are some obvious tips about why you need to be promoting them;

  • Social sharing helps promote your brand to more people FOR FREE. When a user shares the content on Facebook or Twitter, this content (and your brand) is then seen by all of their friends, family, followers and fans. Imagine the cost of trying to advertise to all of these people…
  • By allowing people to share your content, they are likely going to come back and visit your website again. I personally use Twitter as a ‘read later’ list when I re-tweet random things that I spot. Other people will email things to their self for reading later at, generally when they spot something interesting (such as a nice holiday!) in works time and don’t want to get caught by their manager browsing this.
  • Social signals help your SEO efforts!
  • Social involves real people in the real world who are going to share and talk about your brand with other real people. The more real people you can get involved with your brand the better. This will ultimately lead to more sales, as long as people are talking about you for the right reasons!
  • When someone has just purchased a product, left a review or commented on some content then they like to brag about it. Let them tell their friends about it, it will ultimately result in more people coming to your website.

Here is a great example illustrating the last point. The example is when you have purchased an item on Amazon.

 

When it comes to link building, there are a couple of options

Option 1: Manually build low quantity and quality of links

Option 2: Build and promote great content which will attract natural links.

I certainly know which I would prefer to do. Get the great content and promote well, since this starts the ball rolling for content to be seen by many more people when it is shared over social media.

I hope this gives you something to go off and some good examples about the different options available. Most importantly is just getting something there at least, you can always improve on it later. For example, the Add This button takes literally less than 5 minutes to set up and install site wide. Just get it done. :-)

 

4 thoughts on “Let Me Share Your Content!”

  1. Nice one!
    I must admit I frequently share my Amazon purchases, just to show off what awesome things I buy ;-).

    The only point I’d like to add is that you should check where your audience is, before blindly implementing all kinds of social services. Should JJB Sports offer LinkedIn Share? Possibly not. Do put social buttons on your site but don’t necessarily overwhelm your users with too many (unnecessary) choices :).

    1. Very good point Felix! Knowing who your customers and how they use your services is imperative. Just as adding a LinkedIn Share button onto a children’s TV show website wouldn’t really provide much use.

  2. Great article Michael! I’m all for introducing share buttons as long as the WHAT (type of button) & WHY (the objective) & WHERE (on the page) have been considered for each page of your site. It’s really important to bear in mind the customer journey particularly on e-commerce sites and ensure you are not distracting them from completing the purchase. Would be great to hear the performance those share button had for Hotels-London. ????

Why Programming Is An Essential SEO Skill

As a search engine optimiser you will come across programming every day in one way or another and while it isn’t a must have skill for a lot of roles, I believe it is essential to be able to program and have a good understanding of programming in general.

Some SEO roles require you to get hands on with the code and make the changes straight away, where as within other roles the actual hands on work is completed by the technical team/department.

Either way here are a few reasons why I believe programming is an essential skill to have.

Improve Working Relationship with Developers

If you can speak the language of the developers then it goes a long way to building trust with them. It is important for developers to feel that the work they are doing is contributing to the wider business objectives, so by being able to speak in technical terms means you can bridge that gap between business requirements and technical implementations. I am not saying everyone needs to be an expert programmer or website developer but it is important to understand how things work.

If you don’t understand how things work then how can you go about making things better?

Working on several large scale multi-national travel websites has taught me that changes to the site are much easier to make once you understand how the site has been built. If you want to change the H1 or Meta Title on a selection of pages, then if you understand that the pages are built based on templates then this can help you devise a solution with the development team which allows you to make the required changes. Possibly some kind of mini content management system would be useful?

Understanding how to program also means that you don’t go asking for changes that are either impossible to achieve, that take a long time to implement and provide little or no return on investment.

For technical people, getting things right first time and developing code properly is what drives them. For non-technical people, they are at the opposite end of the spectrum and are only interested in the end product. To achieve a successful product it is important for all parties to work together and ensure that code is developed in a robust way and in a way that meets the original specification.

Whether you lean a bit of Java, PHP, C# or ASP.Net, it doesn’t matter since it gives you an understanding of how things work. It will help you have a real conversation about a technical problem and enable you to understand why seemingly simple changes can actually take a few days work. On the plus side, learning to program is fun too! :-)

Ability to use Macros in Excel

If you can program, then you will soon be able to pick up the programming language in Excel, VBA (Visual Basic for Applications). With this you can quickly develop tools which will make your job and your life much easier by automating the regular tasks or pulling in the required information quickly.

There are alternatives to using Macros in Excel, such as using Microsoft Access, but my preferred method is still traditional programming as it offers much more flexibility since you can integrate this information within a database.

 

Website Scraping

As a search engine optimiser I have no doubt that at some point in your career you have needed to scrape a lot of information from a certain website. Weather that be for price comparison, scraping data from the HTML or just playing around for fun. Website scraping can be done at a basic level using tools such as SEO Tools for Excel, but if you want the more juicy data then you really want to be able to put this into a database and either use on your own website or use it for further research.

To Spot The Rodney McKay

If you don’t anything about programming then when you get told by the technical team that something can’t be done because “the CSS file is located on an external server” you will have no knowledge to be able to question what has been said. You will find a Rodney McKay in every organisation who will say everything is impossible to fix, so they seem like a savior when they solve the problem that didn’t really exist in the first place!

Knowing your onions and being able to question things is extremely important.

 

Using APIs For Extra Data

There is a lot of data analysis within the SEO industry and some of this can include pulling in data from external sources such as Open Site Explorer via their API, Application Programming Interface. Some other APIs that you will no doubt use at some point will be the free Google APIs.

Whether that is using the KML API which is used for creating cool Google Maps visualisations or using the YouTube API to automatically post new content to YouTube when you add it to your website.

APIs can be an extremely powerful tool to use when joining bits of information together and connecting online applications and will certainly help your normal SEO efforts by taking the manual work out of things.

 

Summary

Take what you will from this long winded rant about programming, but I can strongly suggest learning to program and learning more about how ecommerce websites work in general. All of this knowledge is never lost and it will help you bridge the gap between marketing, external demand (ie search volumes on Google) and the technical team.

 

3 thoughts on “Why Programming Is An Essential SEO Skill”

  1. Like the post and examples; I think it’s a fair point and I agree in the main. Essential is a strong word; there are a lot of areas of SEO where this isn’t really the case, but at the same time I think a lot of SEO’s who shy away from the technical aspects could really improve what they can offer their company or clients by learning to code.

    One of my pet hates at the moment is the whole mobile same url versus subdomain versus sub folder argument. I see so many SEO’s who think for a large e-commerce website, it’s as simple as changing a CSS style sheet to make it mobile friendly (hey, if it’s this easy for WordPress, it must be this easy for everything else!). This leads to mindless comments like “if you don’t serve your mobile site on the same URL as your main site, you should sack your devs”.

    As I guy with technical knowledge, I understand this is of course bollocks; I see e-commerce sites with masses of extra conditional statements, huge code bloat, div hiding and so on just so the mobile site can be run from the same URL. A lot of the time it’s better on so many development and technical levels to run the mobile site from a subdomain/folder (just ask eBay, Next, Argos etc) but a non-tech savvy SEO just wouldn’t see this – it’s so different from the simple task of setting up a mobile version of a blog or content site.

    That’s just one example that’s current for me at the moment, but there are a lot of other cases too where tech knowledge is useful.

    Just to finish, I thought you might like to know that you do have quite a well known ally in this line of thinking! I was at SES London 2012, and one of my favourite quotes was Dave Naylor saying in a session “If you want to be an SEO, learn to code”. So you’re not alone :)

  2. I disagree. I can’t code worth a damn, but I do think I’m rather competent at this SEO malarkey.

    I think what you really mean is that a certain level of technical understanding is essential. I would agree with that. I can’t code, but I have a solid understanding of how the internet works, how websites work, and how HTML works. Those are required aspects of being an effective SEO.

    The ability to code is a ‘nice-to-have’ extra in my book. It helps, but it’s not necessary. Then again, as a non-coder, I would say that. :)

    1. As a coder, I would say it is essential :-p

      But as you say, actually being able to do the code yourself is more of a nice-to-have – as long as there is the deeper understanding of how websites work, how the internet works, understanding HTML. These types of skills are not always as common as one would hope.

      What I have found is that having this deeper understanding of ‘how stuff works’ actually helps things get done easier, as it helps to understand technical difficulties with implementing SEO recommendations then work through the solutions together with people to come to an amicable solution.

      People may not be able to do it their self (hey, I can read PHP & ASP and understand what is going on if forced but I wouldn’t have a clue where to begin writing it – my background is Java – and I also have no desire to learn those either!) but at least being able to understand things at a deeper level helps an awful lot.

      Also having the ability to ‘muck in’ can help get some of the smaller SEO tasks through which people often don’t value. E.g. messed up meta titles, missing H1s, fixing broken canonicals and removing the line from robots.txt which is telling Google to go away. In larger organisations often the technically smaller items are often pushed back even with huge amounts of evidence to back up how important they actually are, as there is always something more important. I have yet to come across a company (which isn’t an SEO company!) where SEO is prioritised above everything else.

Is Google Turning Into A Hotel Aggregator?

There has been a lot of developments over the past few years with Google in the travel industry and their actions are ones I have been closely watching. With Google Hotel Finder launching earlier this year it is interesting to see how Google is changing.

When searching for hotels online there are two main categories of searches

  • Destination searches such as London Hotels
  • Hotel name searches such as The Lowry Hotel

So if we have a look at each of these to see how Google is handling these now we may get an idea of what lies in store in the next year or so.

 

Destination Searches

Looking at the search term Manchester Hotels we can see that there are several important parts to this.

Here we can see several parts to this section which are

  • The link for “Places for hotels near Manchester” and “More results near Manchester”.
  • The travel dates are listed
  • A selection of individual hotel websites are listed
  • Different recommended search terms related to “Manchester Hotels”
  • Nearby locations

Lets take a look at each one of these and see how Google are handling this.

 

Places for hotels near Manchester

This is where the link goes to.

As we can see above, there are what you would expect to see on any hotel aggregator website such as

  • Hotel name
  • Brief snippet of information (albeit a very poor snippet)
  • Link to show where the hotel is located on a map
  • Average rating of the hotel
  • Number of reviews of this hotel
  • Price of the hotel for today

When comparing this to some of the hotel aggregators, it seems Google has got a lot of the areas covered already. Hotel aggregators really need to step up in terms of innovation to ensure they are staying ahead of Google which is going to be quite a challenge.

 

Travel Dates Listed in SERPs

This is a great option for customers since this will automatically update the prices from various different hotel aggregators when the date is chosen.

 

Individual Hotels Listed

Again great for the user since this allows them to go direct to the hotel who may be able to deal with their queries better.

 

Related Search Terms and Nearby Locations

These ultimately send the user through the same process as above but with a more refined search.

 

Hotel Name Searches

Now if we have a look through the same process for hotel name searches to see how Google is handling these.

These listings within the SERPs are a first step towards integrating all of the universal search items into one nice area, which nicely takes up a large part of the results and is positioned prominently above everything else. Now for a user, this again does make sense to have the hotels listing first with other good content pulled together from various sources. My main issue with this though is that Google is notoriously bad at grouping things together like this and there is often a lot of errors with either images or reviews that are listed.

This listing has got all of the items that have been discussed but is just formatted in a nicer way. The interesting point will come when this switch is made from a listing like above to a similar listing like found on Google Hotel Finder as seen below

I have no doubt that this is on the cards for Google, it is just a matter of when they will do this or begin directing users towards their own Hotel Finder product.

 

How Much Can Google Reviews Be Trusted?

One item that Google has yet to work on though is the quality of the reviews left for hotels. Since there is no way of Google knowing if the person has actually stayed in the hotel and this is there hotel aggregators can stand out as they have the full booking details of the customer and know if they have stayed there. In addition hotel aggregators also have manual checking process in place to keep hotel reviews high quality.

As for how Google may overcome this issue, I am not sure yet. Although I imagine there will be some kind of information linked in with Google Plus in a very similar way of how Page Rank works. I.e. If the person leaving a review has an active Google Plus profile, lots of friends & followers and is also active on other Google products. This would be one way to class a reviewer as trustworthy or not. The same is on Twitter when it is obvious someone who has just began following you is a bot due to their statistics.

 

Is Google Going To Become a Competitor?

In a recent article from Marketing Week, apparently not.

“We have no intention of entering the booking space” – Google

Although I am not convinced. There are a lot of options open to Google in this industry and it is an extremely lucrative area for them, so only time will tell exactly what they will do. One thing for certain is that the travel industry needs to step up their game to compete against Google.