Category Archives: seo

Chrome 7 – Auto Preview In Window

I don’t recall seeing this before, but I stumbled across this quite by chance this morning after upgrading to Chrome 7 yesterday. It seems Chrome is now showing auto previews of sites and searches in the main window, as you type them into the address/search bar.

Starting at, I began typing in the address of my blog into the search bar. Note, this is without hitting return/enter:

Chrome auto-previews site

The same happens when searching for a term, and note, the search results autoupdate as you type:

Chrome autopreviews searches

Google Instant – UnBlacklist

Google Instant was rolled out earlier this month, allowing users to see the search results for words as they are typing them. I won’t bore with the details, as if you haven’t seen it already, there are plenty of blog posts out there showing it off in great detail! There has also been a rush of talk about the dreaded “blacklist” that Google has applied to Instant, and how seemingly reasonable searches will not show up unless you complete the search word and click “Search” or hit Enter.

In light of the “naughty” filter (as it’s being referred to), 2600 (the hacker group) have begun compiling a list of words that have been blacklisted by Google Instant, fairly or not. At the end of said blacklist, is a list of words which are not blacklisted, even though they would do well to be, given the guidelines Google seems to be following at the moment. You can, for example, search unhindered for the phrase “how to commit suicide” without the big G sticking its beak in.

Hungry to know more, I quickly thought of a word, to see which list it would fit in… out of curiosity. Imagine my surprise, when searching for “clung” – as in: “the climber clung to the cliff-face for dear life after his safety line snapped leaving him dangling above a great big gorge” – I was given the following suggestion from Google, as well as a full Instant result:
Google Instant Search - Clung

Note the lightly shaded letter “e” urging me to complete the search.

I also found the third result for the full, dirty, uncensored search reasonably amusing!
Google Instant Search - Clunge

But you can’t search for “teen”!

Google Maps Fail

A search for a place to rest my head near to Exeter Airport (I had an early flight to catch yesterday) led me to perhaps the closest airport hotel, ever! Situated between the two runways at Exeter, this was going to be ultra, or even ├╝ber-convenient!

Alas, Google Maps had failed me, as a quick look at the postcode of said hotel, and indeed the image which showed no nearby buildings, quickly alerted my suspicions. The Trelawney Hotel is actually in Torquay:

Google Maps, well and truly, failed!

Google Analytics and sub-domain tracking

I am always championing Google Analytics as my analytical package of choice, not least because it’s free, but because I feel the functionality you get out of it, as a free package, is second to none when compared to the rest of the free/nearly free products out there. Out of the box, GA works with minimal setup or configuration, and works really bloody well.

Tracking sub-domains with Google Analytics also works, with the minimal of configuration, by simply amending your GA code (this is the most up to date version, just by the way):

_gaq.push([‘_setAccount’, ‘UA-XXXXXXX-1’]);

To read:

_gaq.push([‘_setAccount’, ‘UA-XXXXXXX-1’]);
_gaq.push([‘_setDomainName’, ‘’]);

Simple enough? Well, here is a a nice tidbit of information, if you have two pages on your primary and sub-domains, that share the same page name (i.e. index.html), then Google Analytics will combine the statistics for these two pages artificially inflating the statistics for index.html on the primary domain. As my good friend Rob put it: “I wonder who decided that this was a sensible default?” Sensible default indeed!

More endless searching led me to information on setting up a filter, which will supposedly split your data and track traffic on both sub-domains and the primary domain:
Google Analytics filter for sub-domains

Unfortunately, this didn’t work out too well either in my tests:
sub-domain tracking really doesn't work

So, calling on the Twitterverse, David Whitehouse from Bronco volunteered to help out and see if we could get this fixed, and came across the following filter configuration from Brian Clifton’s book:
Google Analytics subdomain tracking filter

This has now been set up, and I’m running a few more tests on some new pages and will report on results as soon as they’re in…


So it’s the morning after the afternoon before, and I can’t say the results are overly brilliant, especially when I spot the following duplication in my analytics overview:
Statistics duplicated in Google Analytics
Furthermore, the subdomain I have been testing is being treated like a subdirectory in the reporting:
Subdirectories and subdomains crossed over in Google Analytics
So when you hover over the “Visit this page” link in GA, you get the following in your status bar:

Not what I would call perfect, by any means.

Robots.txt is case sensitive!

What are robots?

Robots, crawlers, spiders or agents are programs which are used to traverse the wobbly world wide web automatically, taking note on which content is where. Search engines use these programs to index content for their indices, spammers use them to scrape content for their own sites, or even to crawl the web for email addresses to spam you even more. Understanding of how search robots work is an intrinsic part of SEO!

What is robots.txt?

Robots.txt is a file which, when placed in the root of a publicly available webserver, tells search engine robots and agents which content they can and cannot/should and shouldn’t access. It can be used to block access to members’ only directories for example, or pages which nobody should really be finding through search engines, or, as I was trying to do earlier today, block search engines from indexing pages which are a part of your affiliate system or associated with an affiliate ID.

What is an affiliate system?

Many e-businesses run affiliate systems, which can be described as an semi-automated process by which the e-business takes referrals from partner websites, and remunerates them for any transactions that result from the referral. Having worked in the online gaming industry for over 10 years now, I’m quite familiar with many of them, as they an intricate part of just about every online gambling business model.

Most affiliate systems use standard html links which contain affiliate ID parameters in order to track referrals from their partner or affiliated sites. Anyone who clicks the link which contains the affiliate parameters will be associated with that affiliate’s account. If they go on to purchase something from the site, the affiliate will take their share of that revenue.

Search engine optimisation and brand protection

This is leading somewhere, I promise!

Many companies, when taking on an SEO consultant, agency or in-house employee, go straight for the proverbial jugular. They want to target the big, juicy keywords which will drive mountains of good, converting, valuable traffic. Because of this, they usually overlook the basics, ensuring you’re dominating the search results for your brand names. Imagine, if you will, the panic in the office this morning, when after no more than a few hours, I spot a discrepancy in our brand term search results: an affiliate tracked URL is ranking in second spot, taking a nice bounty per referral as well as a share of any future revenue from any clients that came through that link!

SEO, brand protection and affiliate URLs

Whoever was in here before me, had not taken the time to ensure that search engines, and especially Google (ye olde search dominator) were not allowed to index URLs tagged with affiliate IDs. Due to the high volume of traffic this affiliate was sending through his affiliate ID, that URL got indexed for our brand name, and is still ranking in second place. I immediately submitted a removal request, and asked for a change to the robots.txt to ensure that affiliate parameters were blocked from this point forward… the reply: “what robots.txt?”

Robots.txt to block affiliate IDs

Given the fact the content/web developer people here hadn’t implemented one, and the urgency required to get this resolved, I quickly typed up a robots.txt file for them to upload… ran it through the nicely-provided-by-Google testing utility:

Allowed by line 5: Disallow:

But what about line 6 you stupid test tool? The one that says:

Disallow: /?affid=*

No amount of fiddling would get it to work! I tried and tried and tried. And then I tested a second URL, one I typed up myself, and not copied and pasted:

Blocked by line 6: Disallow: /?affid=*

And that is when it struck me… affid and affId are two completely, and utterly different things according to robots.txt… why? Because ROBOTS.TXT IS BLOODY CASE SENSITIVE!