You Are Browsing The Technology Category

Facebook Comments and SEO

March 16 2011 // SEO + Social Media + Technology // 24 Comments

Facebook Comments could be the most disruptive feature released by Facebook. Why? Comments are one of the largest sources of meta content on the web. Our conversations provide a valuable feedback mechanism, giving greater context to both users and to search engines.

The Walled Garden

Using Firebug you can quickly locate Facebook Comments and determine how they're being rendered. Facebook Comments are served in an iframe.

Facebook Comments Delivered in iFrame

This means that the comments are not going to be attributed to that page or site nor seen by search engines. In short, Facebook Comments reside in the walled garden. All your comments are belong to Facebook.

This differs from implementations like Disqus or IntenseDebate where the comments are 'on the page' or 'in-line'. One of the easier ways to understand this is to grab comment text from each platform and search for it on Google. Remember to put the entire text in quotes so you're searching for that exact comment phrase.

Disqus Comments

Here's a comment I made at Search Engine Roundtable via Disqus.

Comment on Disqus

Here's a search for that comment on Google.

Disqus Comment SERP

Sure enough you can find my comment directly at Search Engine Roundtable or at FriendFeed, where I import my Disqus comments.

Facebook Comments

Here's a comment made via Facebook Comments on TechCrunch.

Comment made via Facebook Comments

Here's a search for this comment on Google.

Facebook Comments SERP

In this instance you can't find this comment via search (even on Bing). The comment doesn't exist outside of Facebook's walled garden. It doesn't resolve back to TechCrunch.

I thought of an edge case where Facebook Comments might show up on FriendFeed (via Facebook), but my test indicates they do not.

Comments and SEO

Search engines won't see Facebook Comments. That is a big deal. Comments reflect the user syntax. They capture how people are really talking about a topic or product. Comments help search engines to create keyword clusters and deliver long-tail searches. Comments may signal that the content is still fresh, important and popular. All that goes by the wayside.

It's no secret that search engines crave text. Depriving Google of this valuable source of text is an aggressive move by Facebook.

Is this on purpose? I have to believe it is. I can't know for sure but it's curious that my Quora question has gone unanswered by Facebook, even when I've asked a specific Facebook Engineer to answer.

[Update] Ray C. He did wind up answering my question and provided some examples of how Facebook comments could be made visible to search engines. (Thank you.) Essentially you grab the comments via the API and display them inline behind the comment box, similar to using a noscript tag. It's nice that they have this capability but most will simply use the default version without question or not apply this hack due to lack of technical expertise or time.

In addition, many have since noted that Google has started indexing Facebook comments. Problem solved right? Wrong! Google has always reserved the right to associate iframe content with a URL when it felt it was important. It just rarely did so. The truth of the matter is Google is still only indexing a small fraction of Facebook comments overall. So don't count on Google indexing your Facebook comments.

Comment Spam

Comment Spam

Comment spam is a huge problem. You know this if you've managed a blog for any amount of time. Google's implementation of nofollow didn't do much to stop this practice. So Facebook Comments is appealing to many since the forced identity will curtail most, if not all, of the comment spam.

This also means that the meta content for sites using Facebook Comments may be more pristine. This should be an advantage when Facebook does any type of Natural Language Processing on this data. A cleaner data set can't hurt.

Article Sentiment

Extending this idea, you begin to realize that Facebook could have a real leg up on determining the sentiment of an article or blog post. Others might be able to parse Tweets or other indicators, but Facebook would have access to a large amount of proprietary content to mine page level and domain level sentiment.

Comment Reputation

Facebook can improve on sentiment by looking at comment reputation. Here's where it gets exciting and scary all at the same time. Facebook can map people and their comments to Open Graph objects. It sounds a bit mundane but I think it's a huge playground.

Suddenly, Facebook could know who carries a high reputation on certain types of content. Where did you comment? How many replies did you receive? What was the sentiment of those replies? What was the reputation for those who replied to you? How many Likes did you receive? How many times have you commented on the same Open Graph object as someone else?

You might be highly influential when commenting on technology but not at all when commenting on sports.

The amount of analysis that could be performed at the intersection of people, comments and objects is ... amazing. Facebook knows who is saying what as well as when and where they're saying it.

PeopleRank

PeopleRank

Facebook Comments could go a long way in helping Facebook create a PeopleRank algorithm that would help them better rank pages for their users. If I haven't said it recently, Facebook's Open Graph is just another version of Google's Search Index.

In this instance, Facebook seems to be doing everything it can to develop an alternate way of ranking the web's content while preventing Google from doing so. (Or am I projecting my own paranoia on the situation?)

PeopleRank could replace PageRank as the dominant way to organize content.

Traffic Channel Disruption

The traffic implications of Facebook Comments are substantial. By removing this content from the web, Facebook could reduce the ability of Google and Bing to send traffic to these sites. The long tail would get a lot shorter if Facebook Comments were widely adopted as is.

We've seen some anecdotal evidence that referring traffic from Facebook has increased after implementing Facebook Comments. That makes sense, particularly in the short-term.

The question is whether this is additive or a zero-sum game. In the long-run, would implementing Facebook Comments provide more traffic despite the potential loss in search engine traffic via fewer long-tail visits?

For publishers, the answer might be yes. For retailers, the answer might be no. That has a lot to do with the difference between informational and transactional search.

Even posing the question shows how disruptive Facebook Comments could be if it is widely adopted. It could be the true start of a major shift in website traffic channel mix.

Google Search Quality Decline or Elitism?

January 27 2011 // Marketing + SEO + Technology // 8 Comments

Are content farms really the problem or are you just a snob?

The recent complaints about Google's search quality (here, here, here and here) range from real spam to indictments of content farms. I think we can all agree that spam (cloaking, scrapers, splogs, status code manipulation etc.) should be weeded out. But that leaves us with the larger issue: the quality of results.

Quality

The definition of quality usually refers to a 'degree of excellence' or 'superiority of kind'. It's often associated with grade. Think back to your time in school. Did you ever get back a paper you thought deserved a higher grade? You were certain it was an A paper and you got a B+ instead!

B+ Grade

Quality is a matter of taste.

Taste

Ruination IPA or Coors Light

What about beer? I adore Stone's Ruination IPA. But I'm certain a lot more Coors Light is sold in a day than Ruination IPA in a month, maybe even a year. Even if I were to try to determine the best IPA, there would be many conflicting and passionate opinions on the topic.

Value

Perhaps it's about value instead? Ruination IPA costs a pretty penny while Coors Light is cheap. Maybe Coors Light is the best value because of the ratio of price to quality. But people value things in very different ways. This is clear when looking at restaurant reviews.

Applebees vs The French Laundry

When I read restaurant reviews I can tell whether the reviewer has the same food bias as I do. I treat reviews which laud huge portions, or rock bottom prices, or extol the virtues of never-ending refills differently. Their view of what a good meal is differs from mine. They're looking for quantity, no matter how mediocre the food. I'm looking for quality and generally don't want a pound and a half of garlic mashed potatoes.

There's nothing wrong with either perspective. But they are different.

Popularity

Google Serves Lots of People

Look around folks. What do you see more of? Fast food or fine dining? It's fast food hands down.

And you can see this in nearly every area of life. Justin Bieber and Miley Cyrus are wildly popular musicians but I'm listening to Kasabian and Kaiser Chiefs. I haven't touched Internet Explorer in years but it's (sadly) still the most popular browser.

Mahalo, Squidoo and eHow get millions of visitors a month. These site are popular, and while you might find them distasteful, lacking quality or providing little value, many others (clearly) disagree.

Do I like these sites? No. Perhaps I'm a snob. Maybe you are too.

Numbers

The number of searches has skyrocketed in the last five years. Using comScore's monthly numbers, core searches has gone from 6.9 billion at the beginning of 2007 to 16.4 billion at the beginning of 2011.

US Search Volume 2007 to 2011

At the same time Pew reports a growing percentage of adults are now online and using search engines on a daily basis.

Audience

The search audience has changed. One way to measure this is to plot daily search engine usage by adults against the innovation curve.

Diffusion of Innovation

The U.S. Census Bureau puts the population of the US at around 300 million. Using the CIA World Factbook we can estimate that 80% of those are over the age of 14. I'm going to use the resulting number (240 million) as my adult population number.

In 2007 Pew reported that 70% of adults were online and that 40% of them used search on a daily basis.

  • 240,000,000 X 70% X 40% = 67,200,000

In 2010 Pew reported that 79% of adults were online and that 49% of them used search on a daily basis.

  • 240,000,000 X 79% X 49% = 92,904,000

innovation adoption of search

In both 2007 and 2010 daily search usage penetrated the Early Majority. The difference is that the Early Majority now outnumber the Innovator and Early Adopter groups combined.

Early Majority Rule Search Volume

That's just in three years, imagine the difference between 2005 and 2010. The picture of a daily search user is very different today.

Mental Models

The nature of our searches (as a whole) is likely changing because of who is now searching. The mental model of an Innovator or Early Adopter is going to be different than that of someone in the Early Majority.

Each group is going to approach search with different ideas and baggage. The Innovator and Early Adopter are more likely to be open to new experiences and to explore. They are more risk tolerant.

The Early Majority and Late Majority are more likely to apply their information seeking behaviors from other mediums to search. They're looking for the familiar.

Brands

Many seemed surprised when Google Instant revealed a 'bias' toward brands. It has since been confirmed that Google is not engaging in any internal bias. That bias is a user bias. It's a predication based, in large part, on the volume of searches.

Should we really be surprised? Many of these companies are spending a fortune to advertise and market their brand. Their goal is to capture mindshare and they are succeeding. So much so that people, particularly the Early and Late Majority, go online to search for those brands.

Brand Search Acceleration

In 2005, a DoubleClick report (Search Before The Purchase) showed relatively low levels of brand search. While it accelerated closer to the actual purchase, in some instances only 27% of searches were on brand. Do you honestly think that's still true today?

eCommerce has certainly grown in that time. The number of navigation searches has climbed, which is closely related to brand. People continue to search (a lot) for Facebook or Craigslist as a way to get to those destination. But last year Bing also reported that Walmart was the 8th most searched term.

Users

Matt Cutts tells us not to chase the algorithm but to chase the user. But who is the user really? The audience has changed! And if the algorithm is trying to use human feedback as a signal, wouldn't the results reflect that new composition?

Might that be why in October of 2010 many people noticed an algorithm change that seemed to skew toward bigger brands. It's what Jonathan Mendez called 'gentrification of the SERPs'. (I wish I'd come up with that term!)

I may not think the results got better, but perhaps someone from the Early Majority or Late Majority did. They look at those results and see a lot of familiar brands and that instills confidence.

Content Farms

So when you see eHow at the top of a result and cringe, others might be thinking Google has led them to the easiest and best result. When you find a Mahalo page you might grind your teeth, but others could walk away thinking they got exactly what they needed.

I may enjoy reading the works of Shakespeare but plenty of others will be super happy to have the CliffsNotes version instead.

Which User is Google Optimizing For?

McGoogle

I believe Google when they say they want to provide the most relevant results. But there is a fair bit of subjectivity involved because the user is not some monolithic, homogeneous blob. Quality, taste, value and popularity are all going to inform what people think is relevant.

If Google is optimizing for the majority, that may mean a very different interpretation of relevancy. There's nothing really wrong with that, but if you're an Innovator or Early Adopter, you might think things are getting worse and not better.

There's usually a better place to eat right down the street from a McDonald's, but it's McDonald's that still gets most of the business. There are some places (North Beach in San Francisco for instance) that have a 'no-chains' policy.

Google could certainly do that. They could stand up and say that fast food content from Demand Media wouldn't gain prime SERP real estate. Google could optimize for better instead of good enough. They could pick fine dining over fast food.

But is that what the 'user' wants?

2011 Predictions

December 31 2010 // Analytics + Marketing + SEO + Social Media + Technology + Web Design // 3 Comments

Okay, I actually don't have any precognitive ability but I might as well have some fun while predicting events in 2011. Lets look into the crystal ball.

2011 Search Internet Technology Predictions

Facebook becomes a search engine

The Open Graph is just another type of index. Instead of crawling the web like Google, Facebook lets users do it for them. Facebook is creating a massive graph of data and at some point they'll go all Klingon on Google and uncloak with several bird of prey surrounding search. Game on.

Google buys Foursquare

Unless you've been under a rock for the last 6 months it's clear that Google wants to own local. They're dedicating a ton of resources to Places and decided that getting citations from others was nice but generating your own reviews would be better. With location based services just catching on with the mainstream, Google will overpay for Foursquare and bring check-ins to the masses.

UX becomes more experiential

Technology (CSS3, Compass, HTML5, jQuery, Flash, AJAX and various noSQL databases to name a few) transforms how users experience the web. Sites that allow users to seamlessly understand applications through interactions will be enormously successful.

Google introduces more SEO tools

Google Webmaster Tools continues to launch tools that will help people understand their search engine optimization efforts. Just like they did with Analytics, Google will work hard in 2011 to commoditize SEO tools.

Identity becomes important

As the traditional link graph becomes increasingly obsolete, Google seeks to leverage social mentions and links. But to do so (in any major way) without opening a whole new front of spam, they'll work on defining reputation. This will inevitably lead them to identity and the possible acquisition of Rapleaf.

Internet congestion increases

Internet congestion will increase as more and more data is pushed through the pipe. Apps and browser add-ons that attempt to determine the current congestion will become popular and the Internati will embrace this as their version of Greening the web. (Look for a Robert Scoble PSA soon.)

Micropayments battle paywalls

As the appetite for news and digital content continues to swell, a start-up will pitch publications on a micropayment solution (pay per pageview perhaps) as an alternative to subscription paywalls. The start-up may be new or may be one with a large installed user base that hasn't solved revenue. Or maybe someone like Tynt? I'm crossing my fingers that it's whoever winds up with Delicious.

Gaming jumps the shark

This is probably more of a hope than a real prediction. I'd love to see people dedicate more time to something (anything!) other than the 'push-button-receive-pellet' games. I'm hopeful that people do finally burn out, that the part of the cortex that responds to this type of gratification finally becomes inured to this activity.

Curation is king

The old saw is content is king. But in 2011 curation will be king. Whether it's something like Fever, my6sense or Blekko, the idea of transforming noise into signal (via algorithm and/or human editing) will be in high demand, as will different ways to present that signal such as Flipboard and Paper.li.

Retargeting wins

What people do will outweigh what people say as retargeting is both more effective for advertisers and more relevant for consumers. Privacy advocates will howl and ally themselves with the government. This action will backfire as the idea of government oversight is more distasteful than that of corporations.

Github becomes self aware

Seriously, have you looked at what is going on at Github? There's a lot of amazing work being done. So much so that Github will assemble itself Voltron style and become a benevolently self-aware organism that will be our digital sentry protecting us from Skynet.

Facebook Like Numbers Are Inflated

November 05 2010 // Social Media + Technology // 15 Comments

As you surf the web today you'll inevitably run into Facebook's Like button.

number of likes

There are a number of implementations but they all tell you how many Likes that item (or object in Open Graph speak) has received.

When a Like is not a Like

But did 938 people really Like this rather interesting Slate article about Netflix? No.

actual like count

Only 130 actually liked this article. The rest of that 938 is composed of shares and comments.

What you're looking at above is XML output from a links.getStats call from Facebook's old REST API. The data definitions for the link_stat table detail what share, like, comment and total represent.

Link_Stat Data Definitions

The Like number shown to users is actually the total_count - "the total number of times the URL has been shared, liked, or commented on."

I'm not particularly perturbed by lumping share and like together - those two actions are similar. In both cases I'm explicitly choosing to interact and promote that item. And I suspect that they're doing this for some amount of backwards compatibility.

But comments seems like a stretch to me. I'm choosing to interact with a combination of item and person. My comment might have little to do with the item and more to do with the person sharing it. In this instance I could have commented on the movie viewing habits of the person sharing the item. Does that mean I 'Like' that item?

Like Number Inflation

At a minimum, I think this is a manipulation of perception. The numbers are part of a Like marketing campaign. Large Like numbers throughout the Internet make it seem like the functionality is being used frequently. Yet, here we see that the specific Like feature isn't as popular as we might have suspected.

I'm still a fan (pun intended) of the Like button and the Open Graph, but showing this inflated number (even if it can be rationalized) seems disingenuous. What do you think?

The Best SEO Tools Not About SEO

June 22 2010 // SEO + Technology // 5 Comments

There are plenty of great blog posts about SEO tools, though you should be careful to look at a curated and updated list. Actually an SEO tool wiki would be an interesting idea. But I digress.

Instead of discussing the SEO tools I use I thought I'd share the other tools I use each and every day. Tools that have become indispensable, saving me time, energy and headaches.

Dropbox

Dropbox Logo

Sharing files can be hassle unless you have Dropbox.

Dropbox is essentially a cloud based storage system. I started using it to sync files between my laptop and desktop computers. But what Dropbox is really good for is sharing files with clients.

Email is unreliable and you often wind up spending time waiting for folks to find and download files. Dropbox lets you create a shared folder for each client where you can keep all related materials. Not only can your client find the materials, they can point internal resources to it with ease. This is particularly useful if a client uses contracted or offshore developers.

You may have to convince clients to install Dropbox. Don't worry, the 2GB plan is free, installation is easy and the instant value it delivers will earn you quick kudos.

Adium

Adium Logo

I love Instant Messaging. Short of being on-site, this is often the best way to communicate, clarify and remove roadblocks. Email is slow and asynchronous. The phone doesn't provide the added context of links or screen shots. IM is fast and effective. It can also be a hassle if you have clients on multiple IM platforms. Yahoo! Messenger, Google Talk, Jabber, AIM and more.

Setting up accounts with each is easy, but having every IM client up and running at once creates problems. Context switching between each platform's UI is not trivial. The messages arrive in different ways (with different sounds) and you wind up having multiple windows begging for your attention.

That's where Adium comes in.

Adium is a free instant messaging application for Mac OS X that can connect to AIM, MSN, Jabber, Yahoo, and more.

Adium unifies all your IM programs into one slick interface. The tabbed chat feature is particularly nice so that you don't have a new window for each IM conversation cluttering up your monitor.

You can even combine contacts (the same person on multiple IM platforms) "so that each one represents a person, not an account." This is nice when you don't care how you reach them, just that you reach them. Like the phone, you don't care who the carrier is, you just want to connect.

The good news is you can use Adium even if your friends or clients don't. The bad news, it's Mac only. Windows users might want to check out Trillian instead.

TinyGrab

TinyGrab Logo

If a picture is worth a thousand words, perhaps a screen shot is worth a few hundred.

Screen grabs are a vital part of the SEO process. You want to show clients what you're seeing and how to fix it. If you're building a presentation deck this isn't a huge problem.

If you're having an IM conversation about an issue (with Adium I hope), the traditional screen grab can be slow and clunky. Enter TinyGrab.

Download this tool and each time you take a screen grab it saves it to the cloud and copies a tiny URL of it to your clipboard. Then simply paste it into your conversation and you'll be looking at the same thing in no time.

The free version of TinyGrab gives you 10 grabs a day. For a one-time fee of £10 you can upgrade to the premium version for unlimited grabs.

These tools make me more productive every day. Do you have other tools that make a difference in your daily life? Share them here.

Have Facebook and Google Killed Permission Marketing?

May 06 2010 // Advertising + Marketing + Technology // 3 Comments

Have Facebook and Google Killed Permission Marketing

Back in 1999 I sat in the San Diego County Courthouse reading Seth Godin's Permission Marketing, hoping that I didn't get selected to serve on the class-action lawsuit against grocery chains who had allegedly conspired to fix prices on eggs.

I run hot and cold on Godin these days but Permission Marketing made a lot of sense and still does to a large extent. The core principle was that you needed permission to market to your customer.

Make the Permission Overt and Clear - Chapter 9, p 163.

As an early email marketer I recall the days when double opt-in lists were all the rage. Opt-in just wasn't enough because the methods of collection could have been less than overt and clear. A double opt-in list ensured that you were getting the best list, the Glengarry list.

Opt-In versus Opt-Out

The difference between opt-in and opt-out can be substantial. Opt-in is the active choice to accept something, while opt-out is the passive acceptance of something. The problem here is that inertia can be quite powerful. The default presentation is often used by users as they seek to efficiently complete a task.

That's not to say all opt-ins are created equal. The acceptance of terms of use (and privacy) before completing a download or registration is a weak opt-in since the majority of people don't read it and those that do often don't understand it. This type of coerced opt-in may be better than an opt-out but not by much.

Is Opt-Out Bad?

As a marketer, opt-in can be frustrating. A product or service that you just know would be valuable to a user is gated by their natural inertia. You run the numbers and it's clear that an opt-out would be better for both the business and the user. Quite simply, you'd be able to deliver a valuable product to more of the right users. Those who don't see that value can opt-out. No fuss, no muss right?

Well, permission marketing would tell you that you need overt and clear permission from a user to start that relationship. A user must raise their hand. Is opt-out overt enough? That's debatable but it brings us to another permission marketing principle. Once given permission, you can't abuse that permission. That's where things have gone awry.

Opt-out got a bad name because (way) too many businesses abused that weak permission by not being relevant. It's a shame since a good marketer could probably pull off an opt-out program. And that's just what Facebook and Google are doing.

Value and Relevance

The value of your product or service and the relevance you deliver to the user are going to be paramount to maintaining that permission, no matter how it was attained. Think about that for a minute.

What I'm saying is that if your product or service is that good, you can acquire those customers in nearly any way. Opt-in, Opt-out, Optimus Prime, it won't matter. Sure, some people will claim it does, but there's evidence to the contrary.

Google is Good ... Enough

Google tracks and uses your search and site history to personalize your search results. They actually do this when you're signed-in and signed-out. Here's a look at how you sign up for Web History.

Google Web History

It's opt-out and it's relatively overt, but is it clear? It communicates the benefits quite nicely but what the feature actually does ... not so much. But hey, that's why there's a Learn More link, right?

Web History actually can make your Google experience better. For most users I'd guess the Web History feature is completely transparent and they have no idea that their actions are being recorded. They simply think Google works great.

But what happens when someone figures out what's going on?

What People Say and What People Do

People may say they would turn Web History off but how many really do? Sure, sometimes there's a meme that takes hold and a few folks will very publicly call it quits. But the majority don't ... even when they say they will. The bark is much worse than the bite. And both Google and Facebook know it.

Lets take behavioral targeting (BT) as an example.

Behavioral targeting uses information collected on an individual's web-browsing behavior, such as the pages they have visited or the searches they have made, to select which advertisements to display to that individual.

When people are asked whether they want this type of advertising, the response is generally negative.

Users Say They Don't Want Behavior Targeting

Yet, behavioral targeting has proven to be very successful with click through rates substantially higher, often cited at three times the normal click through rate and recently noted in one study (pdf) as having the ability to achieve a 1000% lift. The ads are more relevant and people are voting with their clicks.

Google's DoubleClick has a BT program. They call it interest-based advertising. The program is opt-out and Matt Cutts recently commented on the opt-out behavior.

Only a relatively small number of people visit that opt-out page each week, and the majority of them change their interests rather than opting out.

Once again, we see a product delivering enough value and certainly enough relevance to overcome any ire users might have about the 'auto' opt-in. In fact, the product produces such relevance (as seen by the high CTR) that most users simply think the ads are getting better. They're not giving much thought to the how, just that it's a better experience.

What about Privacy?

I still believe in privacy. I actually have Web History turned off and I don't share much on Facebook. I consciously made those choices. Just like I make the choice not to give my name and address away at the drop of a hat to enter to win the new car parked in the middle of the mall. There's a certain level of personal responsibility and common sense that must be levied on the user.

I believe that you would see users opt-out of these services if they didn't provide the requisite relevance and value. Right now, Google and Facebook do for the majority of users.

Marketing Privacy

Google has been careful, outside of Buzz, to not provoke negative user interest. Instead, they've worked and publicized their attempts to make opt-out and privacy settings more available. Why? They've seen that users are willing to give up a certain amount of privacy to engage in their products. So they're happy to have 100,000 people a day visit their dashboard.

Facebook, on the other hand, has provoked negative user interest. They make broad sweeping changes that highlight the exchange of privacy for value. Coupled with a poor user interface for the various opt-out settings and Facebook has caught substantially more flak.

Google has been marketing privacy while Facebook has been marketing value.

Intravenous Permission

Have Google and Facebook killed Permission Marketing? Not really. Google, and Facebook to a lesser degree, has short-circuited the natural progression of permission and achieved a type of intravenous permission (the highest level) through the release of great and free products. (Free is important. It creates a subtle user obligation.)

Users can always revoke this level of permission. It will take a break in trust, an abuse of permission, to force users to evaluate their exchange of privacy for value. Even then, that balance will have to be substantially different for users to make a change.

Google Text-Only Cache Bookmarklet

March 22 2010 // SEO + Technology // 1 Comment

text only no html

Last year I wrote about Google's text-only cache which lets you see what Googlebot sees.  This fits in well with Blind Five Year Old philosophy since Google doesn't care if your site is pretty. I know, it's not a perfect analogy because Google would need to read the text but ... think of it as braille for Googlebot.

I still recommend the technique but wanted to pass along a better way to access Google's text-only cache.

Google Text-Only Cache Bookmarklet

A bookmarklet is a bookmark that delivers one-click functionality to a browser or web page. You're probably using a few already (bit.ly anyone?). Following is a bookmarklet that shows the Google text-only cache of any page.

Text Cache

Simply drag the link above to your bookmark bar to have one-click access to Google's text-only cache of the page you're on. This bookmarklet comes from SEOmoz, where they've compiled a list of 30 SEO bookmarklets along with instructions on how to create your own.

Create Your Own Bookmarklet

Creating your own bookmarklet really is easy. Here's one I created that gives you one-click access to SEM Rush.

SEM Rush

The instructions SEOmoz provides are solid, but limited to simple queries. Anything more and you'll need to learn additional javascript commands and syntax. If you're technically inclined that's not a huge task, but start out by futzing around with the simple stuff. As always, doing it is the best way to learn.

So grab or create SEO bookmarklets so you're spending less time navigating and more time analyzing.

2010 Internet, SEO and Technology Predictions

January 03 2010 // Advertising + Marketing + SEO + Social Media + Technology // 5 Comments

As we begin 2010, it's time for me to go on the record with some predictions. A review of my 2009 predictions shows a few hits, a couple of half-credits and a few more misses. Then again, many of my predictions were pretty bold.

2010 Technology Predictions

This year is no different.

The Link Bubble Pops

At some point in 2010, the link bubble will pop. Google will be forced to address rising link abuse and neutralize billions of links. This will be the largest change in the Google algorithm in many years, disrupting individual SEO strategies as well as larger link based models such as Demand Media.

Twitter Finds a Revenue Model

As 2010 wears on Twitter will find and announce a revenue model. I don't know what it will be and I'm unsure it will work, but I can't see Twitter waving their hands for yet another year. Time to walk the walk Twitter.

Google Search Interface Changes

We've already seen the search mode test that should help users navigate and refine search results. However, I suspect this is just the beginning and not the end. The rapid rate of iteration by the Google team makes me believe we could see something as radical as LazyFeed's new UI or the New York Times Skimmer.

Behavioral Targeting Accelerates

Government and privacy groups continue to rage against behavioral targeting (BT), seeing it as some Orwellian advertising machine hell bent on destroying the world. Yet, behavioral targeting works and savvy marketers will win against these largely ineffectual groups and general consumer apathy. Ask people if they want targeted ads and they say no, show them targeted ads and they click.

Google Launches gBooks

The settlement between Google, the Authors Guild and the Association of American Publishers will (finally) be granted final approval and then the fireworks will really start. That's right, the settlement brouhaha was the warm up act. Look for Google to launch an iTunes like store (aka gBooks) that will be the latest in the least talked about war on the Internet: Google vs. Amazon.

RSS Reader Usage Surges

What, isn't RSS dead? Well, Marshall Kirkpatrick doesn't seem to think so and Louis Gray doesn't either. I'll side with Marshall and Louis on this one. While I still believe marketing is the biggest problem surrounding RSS readers, advancements like LazyFeed and Fever make me think the product could also advance. I'm still waiting for Google to provide their reader as a while label solution for eTailers fed up with email overhead.

Transparent Traffic Measurement Arrives

Publishers and advertisers are tired of ballpark figures or trends which are directionally accurate. Between Google Analytics and Quantcast people now expect a certain level of specificity. Even comScore is transitioning to beacon based measurement. Panel based traffic measurement will recede, replaced by transparent beacon based measurement ... and there was much rejoicing.

Video Turns a Profit

Online video adoption rates have soared and more and more premium content is readily available. Early adopters bemoan the influx of advertising units, trying to convince themselves and others that people won't put up with it. But they will. Like it or not, the vast majority of people are used to this form of advertising and this is the year it pays off.

Chrome Grabs 15% of Browser Market

Depending on who you believe, Chrome has already surpassed Safari. And this was before Chrome was available for Mac. That alone isn't going to get Chrome to 15%. But you recall the Google 'What's a Browser?' video, right? Google will disrupt browser inertia through a combination of user disorientation and brand equity. Look for increased advertising and bundling of Chrome in 2010.

Real Time Search Jumps the Shark

2009 was, in many ways, the year of real time search. It was the brand new shiny toy for the Internati. Nearly everyone I meet thinks real time search is transformational. But is it really?

A Jonathan Mendez post titled Misguided Notions: A Study of Value Creation in Real-Time Search challenges this assumption. A recent QuadsZilla post also exposes a real time search vulnerability. The limited query set and influx of spam will reduce real time search to an interesting, though still valuable, add-on. The Internati? They'll find something else shiny.

Yahoo Strong-Arms comScore

December 22 2009 // Advertising + Technology // Comment

The other day I received an interesting email from Yahoo!

Yahoo and comScore beacon

It's pretty easy to read between the lines here. In fact, little line reading is necessary. The new comScore beacon is providing more accurate results. Yahoo is not currently participating in the beacon program. Yahoo wasn't keen on the "apples-to-oranges" comparison that "could create confusion for advertisers" because it would likely negatively impact their display business.

Don't Forget Yahoo!

Bashing Yahoo! seems to be the cool thing to do these days, and they've certainly driven themselves into a ditch. But Yahoo! still holds a powerful position as a portal, content and email provider. This email seems like a not-so-gentle reminder that Yahoo! is still a 900 pound gorilla in some circles.

Beacons and Panel Data

The other takeaway here is the fact that beacons are fast becoming the best way to measure traffic. I see comScore's introduction of beacon technology as a direct reaction to Quantcast.

The rise of Google Analytics allows more and more companies to know exactly how much traffic they receive. The result of this knowledge is a growing dissatisfaction with panel based measurements that aren't just inaccurate but are sometimes flat out wrong.

No More Hand Waving

Whether it was the Web 1.0 darling Alexa or recent upstart Compete, panel based services continue to fail. The difference this time around is that we have beacons (like Google Analytics and Quantcast) that let us know when they fail and by how much.

So while Yahoo! has secured a 6-month repreive, the future will be in accurate and transparent traffic measurement.

Comcast Upgrade Disrespects Customers

October 31 2009 // Rant + Technology + Web Design // 1 Comment

Saturday is bill paying day. One of those bills was Comcast. I'm signed up for automatic payments but I generally check to make sure everything is okay. I'm a bit paranoid that way and it usually only takes a few minutes with a cup of coffee steaming next to me to confirm that all is well.

Comcast was last on the list since I review my bills in reverse chronological order and the Comcast bill notification arrived in my inbox on Friday.

Comcast Fail

I clicked through on the bill and entered my user name.

comcast fail

I tried three times, paying special attention to ensure I didn't fat finger something. Each time, same thing.

So I contacted customer support using their Live Chat feature. I was quickly connected with Vanessa who after a brief back and forth provided this explanation.

Vanessa > I wish to inform you that we did an upgrade with our system and we merged the 2 accounts which is the comcast.net and .com

Vanessa > And due to this upgrade since you do not have internet service I am afraid that you need to register it again online, AJ.

That makes as much sense as a fish riding a bicycle! To Vanessa's credit she was apologetic (even though it wasn't her fault) and very helpful. Thank you Vanessa.

Comcast System Upgrade?

I'm not an engineer or a coder, but I know enough to know that a database merge can be done far more elegantly. Boiled down, isn't this a simple left outer join?

Even if there is more complexity (and there usually is) wouldn't it be wise to deal with those issues instead of inconveniencing your customers?

Comcast Error Messaging

Even if Comcast chose to go ahead as planned, they could have avoided in-bound customer service issues by applying proper error messaging.

A simple statement about a system upgrade requiring users to re-register would have made the situation clear. Inconvenient but at least Comcast would have provided an immediate answer to the problem they created.

But here's the rub. You need your account number to register!

That account number is not on the billing email and since I use paperless billing I have no material with my Comcast account number. No matter what, I'd still have to contact customer support to retrieve my account number.

Comcast Customer Service

Despite the buzz Comcast has generated around their use of Twitter they still don't seem to understand customer service.

If they did, they'd have created use cases from a customer perspective and realized that this upgrade would be detrimental for users and would cause added customer support costs.

AT&T U-verse is looking better all the time.