Peak Google? Not Even Close
Search vs Native Ads
Google owns search, but are they a one trick pony?
A couple weeks ago Ben Thompson published an interesting article suggesting Google may follow IBM and Microsoft in peaking, perhaps with native ads becoming more dominant than online search ads.
According to Forrester, in a couple years digital ad spend will overtake TV ad spend. In spite of the rise of sponsored content, native isn’t even broken out as a category.
Part of the issue with native advertising is it can be blurry to break out some of it. Some of it is obvious, but falls into multiple categories, like video ads on YouTube. Some of it is obvious, but relatively new & thus lacking in scale. Amazon is extending their payment services & Prime shipping deals to third party sites of brands like AllSaints & listing inventory from those sites on Amazon.com, selling them traffic on a CPC basis. Does that count as native advertising? What about a ticket broker or hotel booking site syndicating their inventory to a meta search site?
And while native is not broken out, Google already offers native ad management features in DoubleClick and has partnered with some of the more well known players like BuzzFeed.
The Penny Gap’s Impact on Search
Time intends to test paywalls on all of its major titles next year & they are working with third parties to integrate affiliate ads on sites like People.com.
The second link in the above sentence goes to an article which is behind a paywall. On Twitter I often link to WSJ articles which are behind a paywall. Any important information behind a paywall may quickly spread beyond it, but typically a competing free site which (re)reports on whatever is behind the paywall is shared more, spreads further on social, generates more additional coverage on forums and discussion sites like Hacker News, gets highlighted on aggregators like TechMeme, gets more links, ranks higher, and becomes the default/canonical source of the story.
Part of the rub of the penny gap is the cost of the friction vastly exceeds the financial cost. Those who can flow attention around the payment can typically make more by tracking and monetizing user behavior than they could by charging users incrementally a cent here and a nickel there.
Well known franchises are forced to offer a free version or they eventually cede their market position.
There are sites which do roll up subscriptions to a variety of sites at once, but some of them which had stub articles requiring payment to access like Highbeam Research got torched by Panda. If the barrier to entry to get to the content is too high the engagement metrics are likely to be terrible & a penalty ensues. Even a general registration wall is too high of a barrier to entry for some sites. Google demands whatever content is shown to them be visible to end users & if there is a miss match that is considered cloaking – unless the miss match is due to monetizing by using Google’s content locking consumer surveys.
Who gets to the scale needed to have enough consumer demand to be able to charge an ongoing subscription for access to a variety of third party content? There are a handful of players in music (Apple, Spotify, Pandora, etc) & a handful of players in video (Netflix, Hulu, Amazon Prime), but outside of those most paid subscription sites are about finance or niche topics with small scale. And whatever goes behind the paywalls gets seen by almost nobody when compared against to the broader public market at the free pricepoint.
Even if you are in a broad niche industry where a subscription-based model works, it still may be brutally tough to compete against Google. Google’s chief business officer joined the board of Spotify, which means Spotify should be safe from Google risk, except…
- In spite of billions of dollars of aggregate royalty payouts by Spotify, Taylor Swift pulled her catalog from Spotify
- shortly after Taylor Swift pulled her catalog from Spotify, YouTube announced their subscription service, which will include Taylor Swift’s catalog & will offer a free 6-month trial
Google’s Impact on Premium Content
I’ve long argued Google has leveraged piracy to secure favorable media deals (see the second bullet point at the bottom of this infographic). Some might have perceived my take as being cynical, but when Google mentioned their “continued progress on fighting piracy” the first thing they mentioned was more ad units.
There are free options, paid options & the blurry lines in between which Google & YouTube ride while they experiment with business models and monetize the flow of traffic to the paid options.
“Tech companies don’t believe in the unique value of premium content over the long term.” – Jessica Lessin
There is a massive misalignment of values which causes many players to have to refine their strategy over and over again. The gray area is never static.
Many businesses only have a 10% or 15% profit margin. An online publishing company which sees 20% of its traffic disappear might thus go from sustainable to bleeding cash overnight. A company which can arbitrarily shift / redirect a large amount of traffic online might describe itself as a “kingmaker.”
In Germany some publishers wanted to be paid to be in the Google index. As a result Google stopped showing snippets near their listings. Google also refined their news search integration into the regular search results to include a broader selection of sources including social sites like Reddit. As a result Axel Springer quickly found itself begging for things to go back to the way they were before as their Google search traffic declined 40% and their Google News traffic declined 80%. Axel Springer got their descriptions back, but the “in the news” change remains.
Google’s Impact on Weaker Players
If Google could have that dramatic of an impact on Axel Springer, imagine what sort of influence they have on millions of other smaller and weaker online businesses.
One of the craziest examples is Demand Media.
Demand Media’s market cap peaked above $1.9 billion. They spun out the domain name portion of the business into a company named Rightside Group, but the content portion of the business is valued at essentially nothing. They have about $40 million in cash & equivalents. Earlier this year they acquired Saatchi Art for $17 million & last year they acquired ecommerce marketplace Society6 for $94 million. After their last quarterly result their stock closed down 16.83% & Thursday they were down another 6.32%, given them a market capitalization of $102 million.
On their most recent conference call, here are some of the things Demand Media executives stated:
- By the end of 2014, we anticipate more than 50.000 articles will be substantially improved by rewrites made rich with great visuals.
- We are well underway with this push for quality and will remove $1.8 million underperforming articles in Q4
- as we strive to create the best experience we can we have removed two ad units with the third unit to be removed completely by January 1st
- (on the above 2 changes) These changes are expected to have a negative impact on revenues and adjusted EBITDA of approximately $15 million on an annualized basis.
- Through Q3 we have invested $1.1 million in content renovation costs and expect approximately another $1 million in Q4 and $2 million to $4 million in the first half of next year.
- if you look at visits or you know the mobile mix is growing which has lower CPM associated with it and then also on desktop we’re seeing compression on the pricing line as well.
- we know that sites that have ad density that’s too high, not only are they offending audiences in near term, you are penalized with respect to where you show up in search indexes as well.
Google torched eHow in April of 2011. In spite of over 3 years of pain, Demand Media is still letting Google drive their strategy, in some cases spending millions of dollars to undo past investments.
Yet when you look at Google’s search results page, they are doing the opposite of the above strategy: more scraping of third party content coupled with more & larger ad units.
“@CyrusShepard: Google’s got you covered: pic.twitter.com/ZU2AOl5EGn” Google copies entire web page?— john andrews (@searchsleuth999) October 30, 2014
Originally the source links in the scrape-n-displace program were light gray. They only turned blue after a journalist started working on a story about 10 blue links.
The Blend
The search results can be designed to have some aspects blend in while others stand out. Colors can change periodically to alter response rates in desirable ways.
The banner ad got a bad rap as publishers have fought declining CPMs by adding more advertisements to their pages. When it works, Google’s infrastructure still delivers (and tracks) billions of daily banner ads.
Search ads have never had the performance decline banner ads have.
The closest thing Google ever faced on that front was when AdBlock Plus became popular. Since it was blocking search ads, Google banned them & then restored them as they eventually negotiated a deal to pay them to display ads on Google while they continued to block ads on other third party sites.
Search itself *is* the ultimate native advertising platform.
Google is doing away with the local carousel in favor of a 3 pack local listing in categories like hotels. Once a person clicks on one of the hotel listings, Google then inserts an inline knowledge graph listing for that hotel with booking affiliate links inline in the search results, displacing the organic result set below the fold.
Notice in the above graphic how the “website” link uses generic text, is aligned toward the right, and is right next to an image so that it looks like an ad. It is engineered to feel like an ad and be ignored. The actual ads are left aligned and look like regular text links. They have an ad label, but that label is a couple lines up from them & there are multiple horizontal lines between the label and the actual ad units.
Not only does Google have the ability to shift the layout in such a drastic format, but then with whatever remains they also get to determine who they act against & who they don’t. While the SEO industry debates the “ethics” of various marketing techniques Google has their eye on the prize & is displacing the entire ecosystem wholesale.
Users were generally unable to distinguish between ads and organic listings *before* Google started mixing the two in their knowledge graph. That is a big part of the reason search ads have never seen the engagement declines banner ads have seen.
Mobile has been redesigned with the same thinking in mind. Google action items (which can eventually be monetized) up top & everything else pushed down the page.
The blurring of “knowledge” and ads allows Google to test entering category after category (like doctor calls from the search results) & forcing advertisers to pay for the various tests while Google collects data.
And as Google is scraping information from third party sites, they can choose to show less information on their own site if doing so creates additional monetization opportunities. As far back as 2009 Google stripped phone numbers off of non-sponsored map listings. And what happened with the recent 3 pack? While 100% of the above the fold results are monetized, …
“Go back to an original search that turns up the 3 PAC. Its completely devoid of logical information that a searcher would want:
- No phone number
- No address
- No map
- NO LINK to the restaurant website.
Anything that most users would want is deliberately hidden and/or requires more clicks.” – Dave Oremland
Google justifies their scrape-n-displace programs by claiming they are put users first. Then they hide some of the information to drive incremental monetization opportunities. Google may eventually re-add some of those basic features which are now hidden, but as part of sponsored local listings.
After all – ads are the solution to everything.
Do branded banner ads in the search results have a low response rate? Or are advertisers unwilling to pay a significant sum for them? If so, Google can end the test and instead shift to include a product carousel in the search results, driving traffic to Google Shopping.
“I see this as yet another money grab by Google. Our clients typically pay 400-500% more for PLA clicks than for clicks on their PPC Brand ads. We will implement exact match brand negatives in Shopping campaigns.” – Barb Young
That money grab stuff has virtually no limit.
The Click Casino
Off the start keywords defaulted to broad match. Then campaigns went “enhanced” so advertisers were forced to eat lower quality clicks on mobile devices.
Then there was the blurring exact match targeting into something else, forcing advertisers to buy lower quality variations of searches & making them add tons of negative keywords (and keep eating new garbage re-interpretations of words) in order to run a fine tuned campaign specifically targeted against a term.
In the past some PPC folks cheered obfuscation of organic search, thinking “this will never happen to me.”
Oops.
And of course Google not only wants to be the ad auction, but they want to be your SEM platform managing your spend & they are suggesting you can leverage the “power” of automated auction time biding.
Advertisers RAVE about the success of Google’s automatic bidding features: “It received one click. That click cost $362.63.”
The only thing better than that is banner ads in free mobile tap games targeted at children.
Adding Friction
Above I mentioned how Google arbitrarily banned the AdBlock Plus extension from the Play store. They also repeatedly banned Disconnect Mobile. If you depend on mobile phones for distribution it is hard to get around Google. What’s more they also collect performance data, internally launch competing apps, and invest in third party apps. And they control the prices various apps pay for advertising across their broad network.
So maybe you say “ok, I’ll skip search & mobile, I’ll try to leverage email” but this gets back to the same issue again. In Gmail social or promotional emails get quarantined into a ghetto where they are rarely seen:
“online stores, if they get big enough, can act as chokepoints. And so can Google. … Google unilaterally misfiled my daily blog into the promotions folder they created, and I have no recourse and no way (other than this post) to explain the error to them” – Seth Godin
Those friction adders have real world consequences. A year ago Groupon blamed Gmail’s tabs for causing them to have poor quarterly results. The filtering impact on a start up can be even more extreme. A small shift in exposure can lower the K factor to something below 1 & require the startups to buy exposure rather than generating it virally.
In addition to those other tabs, there are a host of other risks like being labeled as spam or having a security warning. Few sites promote Google’s view as often as people like Greg Sterling do on Search Engine Land, yet even their newsletter was showing a warning in Gmail.
Google can also add friction to
- websites using search rankings, vertical search result displacement, hiding local business information (as referenced above), search query suggestions, and/or leveraging their web browser to redirect consumer intent
- video on YouTube by counting ad views as organic views, changing the relevancy metrics, and investing in competing channels & giving them preferential exposure as part of the deal. YouTube gets over half their views on mobile devices with small screens, so any shift on Google’s rank preference is going to have a major shift in click distributions.
- mobile apps using default bundling agreements which require manufactures to set Google’s apps as defaults
- other business models by banning bundling-based business models too similar to their own (bundling is wrong UNLESS it is Google doing it)
- etc.
The launch of Keyword (not provided) which hid organic search keyword data was friction for the sake of it in organic search. When Google announced HTTPS as a ranking signal, Netmeg wrote: “It’s about ad targeting, and who gets to profile you, and who doesn’t. Mark my words.”
Facebook announced their relaunch of Atlas and Google immediately started cracking down on data access:
In the conversations this week, with companies like Krux, BlueKai and Lotame, Google told data management platform players that they could not use pixels in certain ads. The pixels—embedded within digital ads—help marketers target and understand how many times a given user has seen their messages online.
“Google is only allowing data management platforms to fire pixels on creative assets that they’re serving, on impressions they bought, through the Google Display Network,” said Mike Moreau, chief solutions officer at Krux. “So they’re starting with a very narrow scope.”
Around the same time Google was cracking down on data sharing, they began offering features targeting consumers across devices & announced custom affinity audiences which allow advertisers to target audiences who visit any particular website.
Google’s special role is not only as an organizer (and obfuscate) of information, but then they get to be the source measuromg how well marketing works via their analytics, which can regularly launch new reports which may causually over-represent their own contribution while under-representing some other channels, profiting from activity bias. The industry default of last click attribution driving search ad spending is one of the key issues which has driven down display ad values over the years.
Investing in Competition
Google not only ranks the ecosystem, but they actively invest in it.
Google tried to buy Yelp. When Facebook took off Google invested in Zynga to get access to data, in spite of a sketchy background. When Google’s $6 billion offer for Groupon didn’t close the deal, Google quickly partnered with over a dozen Groupon competitors & created new offer ad units in the search results.
Inside of the YouTube ecosystem Google also holds equity stakes in leading publishers like Machinima and Vevo.
There have been a few examples of investments getting special treatment, getting benefit of the doubt, or access to non-public information.
The scary scenario for publishers might sound something like this: “in Baidu Maps you can find a hotel, check room availability, and make a booking, all inside the app.” There’s no need to leave the search engine.
Take a closer look & that scary version might already be here. Google’s same day delivery boss moved to Uber and Google added Uber pickups and price estimates to their mobile Maps app.
Google, of course, also invested in Uber. It would be hard to argue that Uber is anything but successful. Though it is also worth mentioning winning at any cost often creates losses elsewhere:
- their insurance situation sounds fuzzy
- the founder doesn’t sound like a nice guy
- they’ve repeatedly engaged in shady endeavors like flooding a competing network with bogus requests & trying to screw with a competitor’s ability to raise funds
- as subprime auto loans have become widespread, Uber has pushed an extreme version of them onto their drivers while cutting their rates – an echo of the utopia of years gone by.
Google invests in disruption as though disruption is its own virtue & they leverage their internal data to drive the investments:
“If you can’t measure and quantify it, how can you hope to start working on a solution?” said Bill Maris, managing partner of Google Ventures. “We have access to the world’s largest data sets you can imagine, our cloud computer infrastructure is the biggest ever. It would be foolish to just go out and make gut investments.”
Combining usage data from their search engine, web browser, app store & mobile OS gives them unparalleled insights into almost any business.
Google is one of the few companies which can make multi-billion dollar investments in low margin areas, just for the data:
Google executives are prodding their engineers to make its public cloud as fast and powerful as the data centers that run its own apps. That push, along with other sales and technology initiatives, aren’t just about grabbing a share of growing cloud revenue. Executives increasingly believe such services could give Google insights about what to build next, what companies to buy and other consumer preferences
Google committed to spending as much as a half billion dollars promoting their shopping express delivery service.
Google’s fiber push now includes offering business internet services. Elon Musk is looking into offering satellite internet services – with an ex-Googler.
The End Game
Google now spends more than any other company on federal lobbying in the US. A steady stream of Google executives have filled US government rolls like deputy chief technology officer, chief technology officer, and head of the patent and trademark office. A Google software engineer went so far as suggesting President Obama
- Retire all government employees with full pensions.
- Transfer administrative authority to the tech industry.
- Appoint Eric Schmidt CEO of America.
That Googler may be crazy or a troll, but even if we don’t get their nightmare scenario, if the regulators come from a particular company, that company is unlikely to end up hurt by regulations.
President Obama has stated the importance of an open internet: “We cannot allow Internet service providers to restrict the best access or to pick winners and losers in the online marketplace for services and ideas.”
If there are relevant complaints about Google, who will hear them when Googlers head key government roles?
Larry Page was recently labeled businessperson of the year by Fortune:
It’s a powerful example of how Page pushes the world around him into his vision of the future. “The breadth of things that he is taking on is staggering,” says Ben Horowitz, of Andreessen Horowitz. “We have not seen that kind of business leader since Thomas Edison at GE or David Packard at HP.”
A recent interview of Larry Page in the Financial Times echos the theme of limitless ambition:
- “the world’s most powerful internet company is ready to trade the cash from its search engine monopoly for a slice of the next century’s technological bonanza.” … “As Page sees it, it all comes down to ambition – a commodity of which the world simply doesn’t have a large enough supply.”
- “I think people see the disruption but they don’t really see the positive,” says Page. “They don’t see it as a life-changing kind of thing . . . I think the problem has been people don’t feel they are participating in it.”
- “Even if there’s going to be a disruption on people’s jobs, in the short term that’s likely to be made up by the decreasing cost of things we need, which I think is really important and not being talked about.”
- “in a capitalist system, he suggests, the elimination of inefficiency through technology has to be pursued to its logical conclusion.”
There are some dark layers which are apparently “incidental side effects” of the techno-utopian desires.
Mental flaws could be reinforced & monetized by hooking people on prescription pharmaceuticals:
It takes very little imagination to foresee how the kitchen mood wall could lead to advertisements for antidepressants that follow you around the Web, or trigger an alert to your employer, or show up on your Facebook page because, according to Robert Scoble and Shel Israel in Age of Context: Mobile, Sensors, Data and the Future of Privacy, Facebook “wants to build a system that anticipates your needs.”
Or perhaps…
Those business savings are crucial to Rifkin’s vision of the Third Industrial Revolution, not simply because they have the potential to bring down the price of consumer goods, but because, for the first time, a central tenet of capitalism—that increased productivity requires increased human labor—will no longer hold. And once productivity is unmoored from labor, he argues, capitalism will not be able to support itself, either ideologically or practically.
That is not to say “all will fail” due to technology. Some will succeed wildly.
Michelle Phan has been able to leverage her popularity on YouTube to launch a makeup subscription service which is at an $84 million per year revenue run rate.
Those at the top of the hierarchy will get an additional boost. Such edge case success stories will be marketed offline to pull more people onto the platform.
Google is promoting Japanese Youtube creators’ personal brands in heavy TV, online, and print ads. Trying to recruit more maybe?— Patrick McKenzie (@patio11) November 1, 2014
While a “star based” compensation system makes a few people well off, most people publishing on those platforms won’t see any financial benefit from their efforts. Worse yet, a lot of the “viral” success stories are driven by a large ad budget.
Category after category gets commoditized, platform after platform gets funded by Google, and ultimately employees working on them will long for the days where their wages were held down by illegal collusion rather than the crowdsourcing fate they face:
Workers, in turn, have more mobility and a semblance of greater control over their working lives. But is any of it worth it when we can’t afford health insurance or don’t know how much the next gig might pay, or when it might come? When an app’s terms of service agreement is the closest thing we have to an employment contract? When work orders come through a smartphone and we know that if we don’t respond immediately, we might not get such an opportunity again? When we can’t even talk to another human being about the task at hand and we must work nonstop just to make minimum wage?
Just as people get commoditized, so do other layers of value:
- those who failed to see the value of & invest in domain names cheered as names lost value
- Google philosophically does not believe in customer service, but they feel they should grade others on their customer service and remove businesses which don’t respond well to culturally unaware outsourced third world labor.
- Google gets people to upload pirate content to YouTube & then uses the existence of the piracy to sell ads. They can also use YouTube to report on the quality of various ISPs, while also blocking competing search engines and apps from being able to access YouTube.
- PageRank extracts the value of editorial votes, allowing Google to leverage the work of editors to refine their search rankings without requiring people to visit the sites. Mix in the couple years Google spent fearmongering about links and it is not surprising to see the Yahoo Directory shut down.
- Remember the time a Google “contractor” did a manual scrape of Mocality in Kenya, then called up the businesses and lied to them to try to get them onto Google? I’d link to the original Mocality blog post, but Mocality Kenya has shut down.
- Google also scraped reviews (sometimes without attribution) & only failed at it when players large enough to be heard by Government regulators complained. This is why government scrutiny of Google is so important. It is one of the few things they actually respond to.
In SEO for a number of years many people have painted brand as the solution to everything. But consider the hotel search results which are 100% monetized above the fold – even if you have a brand, you still must pay to play. Or consider the Google Shopping ads which are now being tested on branded navigational searches.
Google even obtained a patent for targeting ads aimed at monetizing named entities.
You paid to build the brand. Then you pay Google again – “or else.”
One could choose to opt out of Google ad products so as not to pay to arbitrage themselves, but Google is willing to harm their own relevancy to extract revenues.
A search in the UK for the trademark term [cheapflights] is converted into the generic search [cheap flights]. The official site is ranking #2 organically and is the 20th clickable link in the left rail of the search results.
As much as brand is an asset, it also becomes a liability if you have to pay again for every time someone looks for your brand.
Mobile apps may be a way around Google, but again it is worth noting Google owns the operating system and guarantees themselves default placement across a wide array of verticals through bundling contracts with manufacturers. Another thing worth considering with mobile is new notification features tied to the operating systems are unbundling apps & Google has apps like Google Now which tie into many verticals.
As SEOs for a long time we had value in promoting the adoption of Google’s ecosystem. As Google attempts to capture more value than they create we may no longer gain by promoting the adoption of their ecosystem, but given their…
- cash hoard
- lobbyists
- ex-employees in key government rolls
- control over video, mobile, apps, maps, email, analytics (along with search)
- broad portfolio of investments
… it is hard to think they’ve come anywhere close to peaking.
How Does Yahoo! Increase Search Ad Clicks?
One wonders how Yahoo Search revenues keep growing even as Yahoo’s search marketshare is in perpetual decline.
Then one looks at a Yahoo SERP and quickly understands what is going on.
Here’s a Yahoo SERP test I saw this morning
I sometimes play a “spot the difference” game with my wife. She’s far better at it than I am, but even to a blind man like me there are about a half-dozen enhancements to the above search results to juice ad clicks. Some of them are hard to notice unless you interact with the page, but here’s a few of them I noticed…
Yahoo Ads | Yahoo Organic Results | |
Placement | top of the page | below the ads |
Background color | none / totally blended | none |
Ad label | small gray text to right of advertiser URL | n/a |
Sitelinks | often 5 or 6 | usually none, unless branded query |
Extensions | star ratings, etc. | typically none |
Keyword bolding | on for title, description, URL & sitelinks | off |
Underlines | ad title & sitelinks, URL on scroll over | off |
Click target | entire background of ad area is clickable | only the listing title is clickable |
What is even more telling about how Yahoo disadvantages the organic result set is when one of their verticals is included in the result set they include the bolding which is missing from other listings. Some of their organic result sets are crazy with the amount of vertical inclusions. On a single result set I’ve seen separate “organic” inclusions for
- Yahoo News
- stories on Yahoo
- Yahoo Answers
They also have other inclusions like shopping search, local search, image search, Yahoo screen, video search, Tumblr and more.
Here are a couple examples.
This one includes an extended paid affiliate listing with SeatGeek & Tumblr.
This one includes rich formatting on Instructibles and Yahoo Answers.
This one includes product search blended into the middle of the organic result set.
Google SEO Services (BETA)
When Google acquired DoubleClick Larry Page wanted to keep the Performics division offering SEM & SEO services just to see what would happen. Other Google executives realized the absurd conflict of interest and potential anti trust issues, so they overrode ambitious Larry: “He wanted to see how those things work. He wanted to experiment.”
Webmasters have grown tired of Google’s duplicity as the search ecosystem shifts to pay to play, or go away.
@davidiwanow I understand the problem, just not the complaints. Google won. Find another oppty, or pay Google. Simple.— john andrews (@searchsleuth999) November 5, 2014
Google’s webmaster guidelines can be viewed as reasonable and consistent or as an anti-competitive tool. As Google eats the ecosystem, those thrown under the bus shift their perspective.
Scraping? AOG (unless we do it) Affiliate? Fucking scumbags mainly AOG (unless we get into the space) Thin content? AOG (unless we do it)— Rae Hoffman (@sugarrae) November 5, 2014
Within some sectors larger players can repeatedly get scrutiny for the same offense with essentially no response, whereas smaller players operating in that same market are slaughtered because they are small.
At this point, Google should just come out and be blunt, “any form of promotion that does not involve paying us is against our guidelines.”— Rae Hoffman (@sugarrae) November 5, 2014
Access to lawyers, politicians & media outlets = access to benefit of the doubt.
Lack those & BEST OF LUCK TO YOU ;)
And most of all, I’m tired of having to tell SMBs that Google gives zero fucks when it comes to them— Rae Hoffman (@sugarrae) November 5, 2014
Google’s page asking “Do you need an SEO?” uses terms like: scam, illicit and deceptive to help frame the broader market perception of SEO.
If ranking movements appear random & non-linear then it is hard to make sense of continued ongoing investment. The less stable Google makes the search ecosystem, the worse they make SEOs look, as…
- anytime a site ranks better, that anchors the baseline expectation of where rankings should be
- large rank swings create friction in managing client communications
- whenever search traffic falls drastically it creates real world impacts on margins, employment & inventory levels
Matt Cutts stated it is a waste of resources for him to be a personal lightning rod for criticism from black hat SEOs. When Matt mentioned he might not go back to his old role at Google some members of the SEO industry were glad. In response some other SEOs mentioned black hats have nobody to blame but themselves & it is their fault for automating things.
After all, it is not like Google arbitrarily shifts their guidelines overnight and drastically penalizes websites to a disproportionate degree ex-post-facto for the work of former employees, former contractors, mistaken/incorrect presumed intent, third party negative SEO efforts, etc.
Oh … wait … let me take that back.
Indeed Google DOES do that, which is where much of the negative sentiment Matt complained about comes from.
Recall when Google went after guest posts, a site which had a single useful guest post on it got a sitewide penalty.
Around that time it was noted Auction.com had thousands of search results for text which was in some of their guest posts.
Enjoying Aaron murdering http://t.co/UadnmwekM7 RT @aaronwall: “about 9,730 results” http://t.co/Sms5L2BFGY— Brian Provost (@brianprovost) April 9, 2014
About a month before the guest post crack down, Auction.com received a $50 million investment from Google Capital.
- Publish a single guest post on your site = Google engineers shoot & ask questions later.
- Publish a duplicated guest post on many websites, with Google investment = Google engineers see it as a safe, sound, measured, reasonable, effective, clean, whitehat strategy.
The point of highlighting that sort of disconnect was not to “out” someone, but rather to highlight the (il)legitimacy of the selective enforcement. After all, …
@mvandemar @brianprovost if anyone should have the capital needed to “do things the right way, as per G” it should be G & those G invests in— aaron wall (@aaronwall) April 9, 2014
But perhaps Google has decided to change their practices and have a more reasonable approach to the SEO industry.
An encouraging development on this front was when Auction.com was once again covered in Bloomberg. They not only benefited from leveraging Google’s data and money, but Google also offered them another assist:
Closely held Auction.com, which is valued at $1.2 billion, based on Google’s stake, also is working with the Internet company to develop mobile and Web applications and improve its search-engine optimization for marketing, Sharga said.
“In a capitalist system, [Larry Page] suggests, the elimination of inefficiency through technology has to be pursued to its logical conclusion.” ― Richard Waters
With that in mind, one can be certain Google didn’t “miss” the guest posts by Auction.com. Enforcement is selective, as always.
“The best way to control the opposition is to lead it ourselves.” ― Vladimir Lenin
Whether you turn left or right, the road leads to the same goal.
Use Verticals To Increase Reach
In the last post, we looked at how SEO has always been changing, but one thing remains constant – the quest for information.
Given people will always be on a quest for information, and given there is no shortage of information, but there is limited time, then there will always be a marketing imperative to get your information seen either ahead of the competition, or in places where the competition haven’t yet targeted.
Channels
My take on SEO is broad because I’m concerned with the marketing potential of the search process, rather than just the behaviour of the Google search engine. We know the term SEO stands for Search Engine Optimization. It’s never been particularly accurate, and less so now, because what most people are really talking about is not SEO, but GO.
Google Optimization.
Still, the term SEO has stuck. The search channel used to have many faces, including Alta Vista, Inktomi, Ask, Looksmart, MSN, Yahoo, Google and the rest, hence the label SEO. Now, it’s pretty much reduced down to one. Google. Okay, there’s BingHoo, but really, it’s Google, 24/7.
We used to optimize for multiple search engines because we had to be everywhere the visitor was, and the search engines had different demographics. There was a time when Google was the choice of the tech savvy web user. These days, “search” means “Google”. You and your grandmother use it.
But people don’t spend most of their time on Google.
Search Beyond Google
The techniques for SEO are widely discussed, dissected, debated, ridiculed, encouraged and we’ve heard all of them, many times over. And that’s just GO.
The audience we are trying to connect with, meanwhile, is on a quest for information. On their quest for information, they will use many channels.
So, who is Google’s biggest search competitor? Bing? Yahoo?
Eric Schmidt thinks it’s Amazon:
Many people think our main competition is Bing or Yahoo,” he said during a visit to a Native Instruments, software and hardware company in Berlin. “But, really, our biggest search competitor is Amazon. People don’t think of Amazon as search, but if you are looking for something to buy, you are more often than not looking for it on Amazon….Schmidt noted that people are looking for a different kind of answers on Amazon’s site through the slew of reviews and product pages, but it’s still about getting information
An important point. For the user, it’s all about “getting information”. In SEO, verticals are often overlooked.
Client Selection & Getting Seen In The Right Places
I’m going to digress a little….how do you select clients, or areas to target?
I like to start from the audience side of the equation. Who are the intended audience, what does that audience really need, and where, on the web, are they? I then determine if it’s possible/plausible to position well for this intended audience within a given budget.
There is much debate amongst SEOs about what happens inside the Google black box, but we all have access to Google’s actual output in the form of search results. To determine the level of competition, examine the search results. Go through the top ten or twenty results for a few relevant keywords and see which sites Google favors, and try to work out why.
Once you look through the results and analyze the competition, you’ll get a good feel for what Google likes to see in that specific sector. Are the search results heavy on long-form information? Mostly commercial entities? Are sites large and established? New and up and coming? Do the top sites promote visitor engagement? Who links to them and why? Is there a lot news mixed in? Does it favor recency? Are Google pulling results from industry verticals?
It’s important to do this analysis for each project, rather than rely on prescriptive methods. Why? Because Google treats sectors differently. What works for “travel” SEO may not work for “casino” SEO because Google may be running different algorithms.
Once you weed out the wild speculation about algorithms, SEO discussion can contain much truth. People convey their direct experience and will sometimes outline the steps they took to achieve a result. However, often specific techniques aren’t universally applicable due to Google treating topic areas differently. So spend a fair bit of time on competitive analysis. Look closely at the specific results set you’re targeting to discover what is really working for that sector, out in the wild.
It’s at this point where you’ll start to see cross-overs between search and content placement.
The Role Of Verticals
You could try and rank for term X, and you could feature on a site that is already ranked for X. Perhaps Google is showing a directory page or some industry publication. Can you appear on that directory page or write an article for this industry publication? What does it take to get linked to by any of these top ten or twenty sites?
Once search visitors find that industry vertical, what is their likely next step? Do they sign up for a regular email? Can you get placement on those emails? Can you get an article well placed in some evergreen section on their site? Can you advertise on their site? Figure out how visitors would engage with that site and try to insert yourself, with grace and dignity, into that conversation.
Users may by-pass Google altogether and go straight to verticals. If they like video then YouTube is the obvious answer. A few years ago when Google was pushing advertisers to run video ads they pitched YouTube as the #2 global search engine. What does it take to rank in YouTube in your chosen vertical? Create videos that will be found in YouTube search results, which may also appear on Google’s main search results.
With 200,000 videos uploaded per day, more than 600 years required to view all those videos, more than 100 million videos watched daily, and more than 300 million existing accounts, if you think YouTube might not be an effective distribution channel to reach prospective customers, think again.
There’s a branding parallel here too. If the field of SEO is too crowded, you can brand yourself as the expert in video SEO.
There’s also the ubiquitous Facebook.
Facebook, unlike the super-secret Google, has shared their algorithm for ranking content on Facebook and filtering what appears in the news feed. The algorithm consists of three components…..
If you’re selling stuff, then are you on Amazon? Many people go directly to Amazon to begin product searches, information gathering and comparisons. Are you well placed on Amazon? What does it take to be placed well on Amazon? What are people saying? What are their complaints? What do they like? What language do they use?
In 2009, nearly a quarter of shoppers started research for an online purchase on a search engine like Google and 18 percent started on Amazon, according to a Forrester Research study. By last year, almost a third started on Amazon and just 13 percent on a search engine. Product searches on Amazon have grown 73 percent over the last year while searches on Google Shopping have been flat, according to comScore
All fairly obvious, but may help you think about channels and verticals more, rather than just Google. The appropriate verticals and channels will be different for each market sector, of course. And they change over time as consumer tastes & behaviors change. At some point each of these were new: blogging, Friendster, MySpace, Digg, Facebook, YouTube, Twitter, LinkedIn, Instagram, Pinterest, Snapchat, etc.
This approach will also help us gain a deeper understanding of the audience and their needs – particularly the language people use, the questions they ask, and the types of things that interest them most – which can then be fed back into your search strategy. Emulate whatever works in these verticals. Look to create a unique, deep collection of insights about your chosen keyword area. This will in turn lead to strategic advantage, as your competition is unlikely to find such specific information pre-packaged.
This could also be characterised as “content marketing”, which it is, although I like to think of it all as “getting in front of the visitors quest for information”. Wherever the visitors are, that’s where you go, and then figure out how to position well in that space.
The Only Thing Certain In SEO Is Change
SEO is subject to frequent change, but in the last year or two, the changes feel both more frequent and significant than changes in the past. Florida hit in 2003. Since then, it’s like we get a Florida every six months.
Whenever Google updates the underlying landscape, the strategies need to change in order to deal with it. No fair warning. That’s not the game.
From Tweaks To Strategy
There used to be a time when SEOs followed a standard prescription. Many of us remember a piece of software called Web Position Gold.
Web Position Gold emerged when SEO could be reduced to a series of repeatable – largely technical – steps. Those steps involved adding keywords to a page, repeating those keywords in sufficient density, checking a few pieces of markup, then scoring against an “ideal” page. Upload to web. Add a few links. Wait a bit. Run a web ranking report. Viola! You’re an SEO. In all but the most competitive areas, this actually worked.
Seems rather quaint these days.
These days, you could do all of the above and get nowhere. Or you might get somewhere, but when so many more factors in play, they can’t be isolated to an individual page score. If the page is published on a site with sufficient authority, it will do well almost immediately. If it appears on a little known site, it may remain invisible for a long time.
Before Google floated in 2004, they released an investor statement signalling SEO – well, “index spammers” – as a business risk. If you ever want to know what Google really feels about people who “manipulate” their results, it’s right here:
We are susceptible to index spammers who could harm the integrity of our web search results.
There is an ongoing and increasing effort by “index spammers” to develop ways to manipulate our web search results. For example, because our web search technology ranks a web page’s relevance based in part on the importance of the web sites that link to it, people have attempted to link a group of web sites together to manipulate web search results. We take this problem very seriously because providing relevant information to users is critical to our success. If our efforts to combat these and other types of index spamming are unsuccessful, our reputation for delivering relevant information could be diminished. This could result in a decline in user traffic, which would damage our business.
SEO competes with the Adwords business model. So, Google “take very seriously” the activities of those who seek to figure out the algorithms, reverse engineer them, and create push-button tools like Web Position Gold. We’ve had Florida, and Panda, and Penguin, and Hummingbird, all aimed at making the search experience better for users, whilst having the pleasant side effect, as far as Google is concerned, of making life more difficult for SEOs.
I think the key part of Google’s statement was “delivering relevant information”.
From Technical Exercise To PR
SEO will always involve technical aspects. You get down into code level and mark it up. The SEO needs to be aware of development and design and how those activities can affect SEO. The SEO needs to know how web servers work, and how spiders can sometimes fail to deal with their quirks.
But in the years since Florida, marketing aspects have become more important. An SEO can perform the technical aspects of SEO and get nowhere. More recent algorithms, such as Panda and Penguin, gauge the behaviour of users, as Google tries to determine information quality of pages. Hummingbird attempts to discover the intent that lays behind keywords.
As a result, Keyword-based SEO is in the process of being killed off. Google withholds keyword referrer data and their various algorithms attempt to deliver pages based on a users intent and activity – both prior and present – in order to deliver relevant information. Understanding the user, having a unique and desirable offering, and a defensible market position is more important than any keyword markup. The keyword match, on which much SEO is based, is not an approach that is likely to endure.
The emphasis has also shifted away from the smaller operators and now appears to favour brands. This occurs not because brands are categorized as “brands”, but due to the side effects of significant PR activities. Bigger companies tend to run multiple advertising and PR campaigns, so produce signals Google finds favorable i.e. search volume on company name, semantic associations with products and services, frequent links from reputable media, and so on. This flows through into rank. And it also earns them leeway when operating in the gray area where manual penalties are handed out to smaller & weaker entities for the same activities.
Rankings
Apparently, Google killed off toolbar PageRank.
We will probably not going to be updating it [PageRank] going forward, at least in the Toolbar PageRank.
A few people noted it, but the news won’t raise many eyebrows as toolbar PR has long since become meaningless. Are there any SEOs altering what they do based on toolbar PR? It’s hard to imagine why. The reality is that an external PR value might indicate an approximate popularity level, but this isn’t an indicator of the subsequent ranking a link from such a page will deliver. There are too many other factors involved. If Google are still using an internal PR metric, it’s likely to be a significantly more complicated beast than was revealed in 1997.
A PageRank score is a proxy for authority. I’m quite sure Google kept it going as an inside joke.
A much more useful proxy for authority are the top ten pages in any niche. Google has determined all well-ranking pages have sufficient authority, and no matter what the toolbar, or any other third-party proxy, says, it’s Google’s output that counts. A link from any one of the top ten pages will likely confer a useful degree of authority, all else being equal. It’s good marketing practice to be linked from, and engage with, known leaders in your niche. That’s PR, as in public relations thinking, vs PR (Page rank), technical thinking.
The next to go will likely be keyword-driven SEO. Withholding keyword referral data was the beginning of the end. Hummingbird is hammering in the nails. Keywords are still great for research purposes – to determine if there’s an audience and what the size of that audience may be – but SEO is increasingly driven by semantic associations and site categorizations. It’s not enough to feature a keyword on a page. A page, and site, needs to be about that keyword, and keywords like it, and be externally recognized as such. In the majority of cases, a page needs to match user intent, rather than just a search term. There are many exceptions, of course, but given what we know about Hummingbird, this appears to be the trend.
People will still look at rank, and lust after prize keywords, but really, rankings have been a distraction all along. Reach and specificity is more important i.e. where’s the most value coming from? The more specific the keyword, typically the lower the bounce rate and the higher the conversion rate. The lower the bounce-rate, and higher the conversion rate, the more positive signals the site will generate, which will flow back into a ranking algorithm increasing being tuned for engagement. Ranking for any keyword that isn’t delivering business value makes no sense.
There are always exceptions. But that’s the trend. Google are looking for pages that match user intent, not just pages that match a keyword term. In terms of reach, you want to be everywhere your customers are.
Search Is The Same, But Different
To adapt to change, SEOs should think about search in the widest possible terms. A search is quest for information. It may be an active, self-directed search, in the form of a search engine query. Or a more passive search, delivered via social media subscriptions and the act of following. How will all these activities feed into your search strategy?
Sure, it’s not a traditional definition of SEO, as I’m not limiting it to search engines. Rather, my point is about the wider quest for information. People want to find things. Eric Schmidt recently claimed Amazon is Google’s biggest competitor in search. The mechanisms and channels may change, but the quest remains the same. Take, for example, the changing strategy of BuzzFeed:
Soon after Peretti had turned his attention to BuzzFeed full-time in 2011, after leaving the Huffington Post, BuzzFeed took a hit from Google. The site had been trying to focus on building traffic from both social media marketing and through SEO. But the SEO traffic — the free traffic driven from Google’s search results — dried up.
Reach is important. Topicality is important. Freshness, in most cases, is important. Engagement is important. Finding information is not just about a technical match of a keyword, it’s about an intellectual match of an idea. BuzzFeed didn’t take their eye off the ball. They know helping users find information is the point of the game they are in.
And the internet has only just begun.
In terms of the internet, nothing has happened yet. The internet is still at the beginning of its beginning. If we could climb into a time machine and journey 30 years into the future, and from that vantage look back to today, we’d realize that most of the greatest products running the lives of citizens in 2044 were not invented until after 2014. People in the future will look at their holodecks, and wearable virtual reality contact lenses, and downloadable avatars, and AI interfaces, and say, oh, you didn’t really have the internet (or whatever they’ll call it) back then.
In 30 years time, people will still be on the exact same quest for information. The point of SEO has always been to get your information in front of visitors, and that’s why SEO will endure. SEO was always a bit of a silly name, and it often distracts people from the point, which is to get your stuff seen ahead of the rest.
Some SEOs have given up in despair because it’s not like the old days. It’s becoming more expensive to do effective SEO, and the reward may not be there, especially for smaller sites. However, this might be to miss the point, somewhat.
The audience is still there. Their needs haven’t changed. They still want to find stuff. If SEO is all about helping users find stuff, then that’s the important thing. Remember the “why”. Adapt the “how”
In the next few articles, we’ll look at the specifics of how.
Measuring SEO Performance After “Not Provided”
In recent years, the biggest change to the search landscape happened when Google chose to withhold keyword data from webmasters. At SEOBook, Aaron noticed and wrote about the change, as evermore keyword data disappeared.
The motivation to withold this data, according to Google, was privacy concerns:
SSL encryption on the web has been growing by leaps and bounds. As part of our commitment to provide a more secure online experience, today we announced that SSL Search on https://www.google.com will become the default experience for signed in users on google.com.
At first, Google suggested it would only affect a single-digit percentage of search referral data:
Google software engineer Matt Cutts, who’s been involved with the privacy changes, wouldn’t give an exact figure but told me he estimated even at full roll-out, this would still be in the single-digit percentages of all Google searchers on Google.com
…which didn’t turn out to be the case. It now affects almost all keyword referral data from Google.
Was it all about privacy? Another rocket over the SEO bows? Bit of both? Probably. In any case, the search landscape was irrevocably changed. Instead of being shown the keyword term the searcher had used to find a page, webmasters were given the less than helpful “not provided”. This change rocked SEO. The SEO world, up until that point, had been built on keywords. SEOs choose a keyword. They rank for the keyword. They track click-thrus against this keyword. This is how many SEOs proved their worth to clients.
These days, very little keyword data is available from Google. There certainly isn’t enough to keyword data to use as a primary form of measurement.
Rethinking Measurement
This change forced a rethink about measurement, and SEO in general. Whilst there is still some keyword data available from the likes of Webmaster Tools & the AdWords paid versus organic report, keyword-based SEO tracking approaches are unlikely to align with Google’s future plans. As we saw with the Hummingbird algorithm, Google is moving towards searcher-intent based search, as opposed to keyword-matched results.
Hummingbird should better focus on the meaning behind the words. It may better understand the actual location of your home, if you’ve shared that with Google. It might understand that “place” means you want a brick-and-mortar store. It might get that “iPhone 5s” is a particular type of electronic device carried by certain stores. Knowing all these meanings may help Google go beyond just finding pages with matching words
The search bar is still keyword based, but Google is also trying to figure out what user intent lays behind the keyword. To do this, they’re relying on context data. For example, they look at what previous searches has the user made, their location, they are breaking down the query itself, and so on, all of which can change the search results the user sees.
When SEO started, it was in an environment where the keyword the user typed into a search bar was exact matching that with a keyword that appears on a page. This is what relevance meant. SEO continued with this model, but it’s fast becoming redundant, because Google is increasingly relying on context in order to determine searcher intent & while filtering many results which were too aligned with the old strategy. Much SEO has shifted from keywords to wider digital marketing considerations, such as what the visitor does next, as a result.
We’ve Still Got Great Data
Okay, if SEO’s don’t have keywords, what can they use?
If we step back a bit, what we’re really trying to do with measurement is demonstrate value. Value of search vs other channels, and value of specific search campaigns. Did our search campaigns meet our marketing goals and thus provide value?
Do we have enough data to demonstrate value? Yes, we do. Here are a few ideas SEOs have devised to look at the organic search data they are getting, and they use it to demonstrate value.
1. Organic Search VS Other Activity
If our organic search tracking well when compared with other digital marketing channels, such as social or email? About the same? Falling?
In many ways, the withholding of keyword data can be a blessing, especially to those SEOs who have a few ranking-obsessed clients. A ranking, in itself is worthless, especially if it’s generating no traffic.
Instead, if we look at the total amount of organic traffic, and see that it is rising, then we shouldn’t really care too much about what keywords it is coming from. We can also track organic searches across device, such as desktop vs mobile, and get some insight into how best to optimize those channels for search as a whole, rather than by keyword. It’s important that the traffic came from organic search, rather than from other campaigns. It’s important that the visitors saw your site. And it’s important what that traffic does next.
2. Bounce Rate
If a visitor comes in, doesn’t like what is on offer, and clicks back, then that won’t help rankings. Google have been a little oblique on this point, saying they aren’t measuring bounce rate, but I suspect it’s a little more nuanced, in practice. If people are failing to engage, then anecdotal evidence suggests this does affect rankings.
Look at the behavioral metrics in GA; if your content has 50% of people spending less than 10 seconds, that may be a problem or that may be normal. The key is to look below that top graph and see if you have a bell curve or if the next largest segment is the 11-30 second crowd.
Either way, we must encourage visitor engagement. Even small improvements in terms of engagement can mean big changes in the bottom line. Getting visitors to a site was only ever the first step in a long chain. It’s what they do next that really makes or breaks a web business, unless the entire goal was that the visitor should only view the landing page. Few sites, these days, would get much return on non-engagement.
PPCers are naturally obsessed with this metric, because each click is costing them money, but when you think about it, it’s costing SEOs money, too. Clicks are getting harder and harder to get, and each click does have a cost associated with it i.e. the total cost of the SEO campaign divided by the number of clicks, so each click needs to be treated as a cost.
3. Landing Pages
We can still do landing page analysis. We can see the pages where visitors are entering the website. We can also see which pages are most popular, and we can tell from the topic of the page what type of keywords people are using to find it.
We could add more related keyword to these pages and see how they do, or create more pages on similar themes, using different keyword terms, and then monitor the response. Similarly, we can look at poorly performing pages and make the assumption these are not ranking against intended keywords, and mark these for improvement or deletion.
We can see how old pages vs new pages are performing in organic search. How quickly do new pages get traffic?
We’re still getting a lot of actionable data, and still not one keyword in sight.
4. Visitor And Customer Acquisition Value
We can still calculate the value to the business of an organic visitor.
We can also look at what step in the process are organic visitors converting. Early? Late? Why? Is there some content on the site that is leading them to convert better than other content? We can still determine if organic search provided a last click-conversion, or a conversion as the result of a mix of channels, where organic played a part. We can do all of this from aggregated organic search data, with no need to look at keywords.
5. Contrast With PPC
We can contrast Adwords data back against organic search. Trends we see in PPC might also be working in organic search.
For AdWords our life is made infinitesimally easier because by linking your AdWords account to your Analytics account rich AdWords data shows up automagically allowing you to have an end-to-end view of campaign performance.
Even PPC-ers are having to change their game around keywords:
The silver lining in all this? With voice an mobile search, you’ll likely catch those conversions that you hadn’t before. While you may think that you have everything figured out and that your campaigns are optimal, this matching will force you into deeper dives that hopefully uncover profitable PPC pockets.
6. Benchmark Against Everything
In the above section I highlighted comparing organic search to AdWords performance, but you can benchmark against almost any form of data.
Is 90% of your keyword data (not provided)? Then you can look at the 10% which is provided to estimate performance on the other 90% of the traffic. If you get 1,000 monthly keyword visits for [widgets], then as a rough rule of thumb you might get roughly 9,000 monthly visits for that same keyword shown as (not provided).
Has your search traffic gone up or down over the past few years? Are there seasonal patterns that drive user behavior? How important is the mobile shift in your market? What landing pages have performed the best over time and which have fallen hardest?
How is your site’s aggregate keyword ranking profile compared to top competitors? Even if you don’t have all the individual keyword referral data from search engines, seeing the aggregate footprints, and how they change over time, indicates who is doing better and who gaining exposure vs losing it.
Numerous competitive research tools like SEM Rush, SpyFu & SearchMetrics provide access to that type of data.
You can also go further with other competitive research tools which look beyond the search channel. Is most of your traffic driven from organic search? Do your competitors do more with other channels? A number of sites like Compete.com and Alexa have provided estimates for this sort of data. Another newer entrant into this market is SimilarWeb.
And, finally, rank checking still has some value. While rank tracking may seem futile in the age of search personalization and Hummingbird, it can still help you isolate performance issues during algorithm updates. There are a wide variety of options from browser plugins to desktop software to hosted solutions.
By now, I hope I’ve convinced you that specific keyword data isn’t necessary and, in some case, may have only served to distract some SEOs from seeing other valuable marketing metrics, such as what happens after the click and where do they go next.
So long as the organic search traffic is doing what we want it to, we know which pages it is coming in on, and can track what it does next, there is plenty of data there to keep us busy. Lack of keyword data is a pain, but in response, many SEOs are optimizing for a lot more than keywords, and focusing more on broader marketing concerns.
Further Reading & Sources:
Loah Qwality Add Werds Clix Four U
Google recently announced they were doing away with exact match AdWords ad targeting this September. They will force all match types to have close variant keyword matching enabled. This means you get misspelled searches, plural versus singular overlap, and an undoing of your tight organization.
In some cases the user intent is different between singular and plural versions of a keyword. A singular version search might be looking to buy a single widget, whereas a plural search might be a user wanting to compare different options in the marketplace. In some cases people are looking for different product classes depending on word form:
For example, if you sell spectacles, the difference between users searching on ‘glass’ vs. ‘glasses’ might mean you are getting users seeing your ad interested in a building material, rather than an aid to reading.
Where segmenting improved the user experience, boosted conversion rates, made management easier, and improved margins – those benefits are now off the table.
CPC isn’t the primary issue. Profit margins are what matter. Once you lose the ability to segment you lose the ability to manage your margins. And this auctioneer is known to bid in their own auctions, have random large price spikes, and not give refunds when they are wrong.
An offline analogy for this loss of segmentation … you go to a gas station to get a bottle of water. After grabbing your water and handing the cashier a $20, they give you $3.27 back along with a six pack you didn’t want and didn’t ask for.
Why does a person misspell a keyword? Some common reasons include:
- they are new to the market & don’t know it well
- they are distracted
- they are using a mobile device or something which makes it hard to input their search query (and those same input issues make it harder to perform other conversion-oriented actions)
- their primary language is a different language
- they are looking for something else
In any of those cases, the typical average value of the expressed intent is usually going to be less than a person who correctly spelled the keyword.
Even if spelling errors were intentional and cultural, the ability to segment that and cater the landing page to match disappears. Or if the spelling error was a cue to send people to an introductory page earlier in the conversion funnel, that option is no more.
In many accounts the loss of the granular control won’t cause too big of a difference. But some advertiser accounts in competitive markets will become less profitable and more expensive to manage:
No one who’s in the know has more than about 5-10 total keywords in any one adgroup because they’re using broad match modified, which eliminated the need for “excessive keyword lists” a long time ago. Now you’re going to have to spend your time creating excessive negative keyword lists with possibly millions upon millions of variations so you can still show up for exactly what you want and nothing else.
You might not know which end of the spectrum your account is on until disaster strikes:
I added negatives to my list for 3 months before finally giving up opting out of close variants. What they viewed as a close variant was not even in the ballpark of what I sell. There have been petitions before that have gotten Google to reverse bad decisions in the past. We need to make that happen again.
Brad Geddes has held many AdWords seminars for Google. What does he think of this news?
In this particular account, close variations have much lower conversion rates and much higher CPAs than their actual match type.
…
Variation match isn’t always bad, there are times it can be good to use variation match. However, there was choice.
…
Loss of control is never good. Mobile control was lost with Enhanced Campaigns, and now you’re losing control over your match types. This will further erode your ability to control costs and conversions within AdWords.
A monopoly restricting choice to enhance their own bottom line. It isn’t the first time they’ve done that, and it won’t be the last.
Have an enhanced weekend!
Understanding The Google Penguin Algorithm
Whenever Google does a major algorithm update we all rush off to our data to see what changed in terms of rankings, search traffic, and then look for the trends to try to figure out what changed.
The two people I chat most with during periods of big a…
Guide To Optimizing Client Sites 2014
For those new to optimizing clients sites, or those seeking a refresher, we thought we’d put together a guide to step you through it, along with some selected deeper reading on each topic area.
Every SEO has different ways of doing things, but we’ll cover the aspects that you’ll find common to most client projects.
Few Rules
The best rule I know about SEO is there are few absolutes in SEO. Google is a black box, so complete data sets will never be available to you. Therefore, it can be difficult to pin down cause and effect, so there will always be a lot of experimentation and guesswork involved. If it works, keep doing it. If it doesn’t, try something else until it does.
Many opportunities tend to present themselves in ways not covered by “the rules”. Many opportunities will be unique and specific to the client and market sector you happen to be working with, so it’s a good idea to remain flexible and alert to new relationship and networking opportunities. SEO exists on the back of relationships between sites (links) and the ability to get your content remarked upon (networking).
When you work on a client site, you will most likely be dealing with a site that is already established, so it’s likely to have legacy issues. The other main challenge you’ll face is that you’re unlikely to have full control over the site, like you would if it were your own. You’ll need to convince other people of the merit of your ideas before you can implement them. Some of these people will be open to them, some will not, and some can be rather obstructive. So, the more solid data and sound business reasoning you provide, the better chance you have of convincing people.
The most important aspect of doing SEO for clients is not blinding them with technical alchemy, but helping them see how SEO provides genuine business value.
1. Strategy
The first step in optimizing a client site is to create a high-level strategy.
“Study the past if you would define the future.” – Confucious
You’re in discovery mode. Seek to understand everything you can about the clients business and their current position in the market. What is their history? Where are they now and where do they want to be? Interview your client. They know their business better than you do and they will likely be delighted when you take a deep interest in them.
- What are they good at?
- What are their top products or services?
- What is the full range of their products or services?
- Are they weak in any areas, especially against competitors?
- Who are their competitors?
- Who are their partners?
- Is their market sector changing? If so, how? Can they think of ways in which this presents opportunities for them?
- What keyword areas have worked well for them in the past? Performed poorly?
- What are their aims? More traffic? More conversions? More reach? What would success look like to them?
- Do they have other online advertising campaigns running? If so, what areas are these targeting? Can they be aligned with SEO?
- Do they have offline presence and advertising campaigns? Again, what areas are these targeting and can they be aligned with SEO?
Some SEO consultants see their task being to gain more rankings under an ever-growing list of keywords. Ranking for more keywords, or getting more traffic, may not result in measurable business returns as it depends on the business and the marketing goals. Some businesses will benefit from honing in on specific opportunities that are already being targeted, others will seek wider reach. This is why it’s important to understand the business goals and market sector, then design the SEO campaign to support the goals and the environment.
This type of analysis also provides you with leverage when it comes to discussing specific rankings and competitor rankings. The SEO can’t be expected to wave a magic wand and place a client top of a category in which they enjoy no competitive advantage. Even if the SEO did manage to achieve this feat, the client may not see much in the way of return as it’s easy for visitors to click other listings and compare offers.
Understand all you can about their market niche. Look for areas of opportunity, such as changing demand not being met by your client or competitors. Put yourself in their customers shoes. Try and find customers and interview them. Listen to the language of customers. Go to places where their customers hang out online. From the customers language and needs, combined with the knowledge gleaned from interviewing the client, you can determine effective keywords and themes.
Document. Get it down in writing. The strategy will change over time, but you’ll have a baseline point of agreement outlining where the site is at now, and where you intend to take it. Getting buy-in early smooths the way for later on. Ensure that whatever strategy you adopt, it adds real, measurable value by being aligned with, and serving, the business goals. It’s on this basis the client will judge you, and maintain or expand your services in future.
Further reading:
– 4 Principles Of Marketing Strategy In The Digital Age
– Product Positioning In Five Easy Steps [pdf]
– Technology Marketers Need To Document Their Marketing Strategy
2. Site Audit
Sites can be poorly organized, have various technical issues, and missed keyword opportunities.
We need to quantify what is already there, and what’s not there.
- Use a site crawler, such as Xenu Link Sleuth, Screaming Frog or other tools that will give you a list of URLs, title information, link information and other data.
- Make a list of all broken links.
- Make a list of all orphaned pages
- Make a list of all pages without titles
- Make a list of all pages with duplicate titles
- Make a list of pages with weak keyword alignment
- Crawl robots txt and hand-check. It’s amazing how easy it is to disrupt crawling with a robots.txt file
Broken links are a low-quality signal. It’s debatable if they are a low quality signal to Google, but certainly to users. If the client doesn’t have one already, implement a system whereby broken links are checked on a regular basis. Orphaned pages are pages that have no links pointing to them. Those pages may be redundant, in which case they should be removed, or you need to point inbound links at them, so they can be crawled and have more chance of gaining rank. Page titles should be unique, aligned with keyword terms, and made attractive in order to gain a click. A link is more attractive if it speaks to a customer need. Carefully check robots.txt to ensure it’s not blocking areas of the site that need to be crawled.
As part of the initial site audit, it might make sense to include the site in Google Webmaster Tools to see if it has any existing issues there and to look up its historical performance on competitive research tools to see if the site has seen sharp traffic declines. If they’ve had sharp ranking and traffic declines, pull up that time period in their web analytics to isolate the date at which it happened, then look up what penalties might be associated with that date.
Further Reading:
– Broken Links, Pages, Images Hurt SEO
– Three Easy Ways To Fix Broken Links And Stop Unnecessary Visitor Loss
– 55 Ways To Use Screaming Frog
– Robots.txt Tutorial
3. Competitive Analysis
Some people roll this into a site audit, but I’ll split it out as we’re not looking at technical issues on competitor sites, we’re looking at how they are positioned, and how they’re doing it. In common with a site audit, there’s some technical reverse engineering involved.
There are various tools that can help you do this. I use SpyFu. One reporting aspect that is especially useful is estimating the value of the SEO positions vs the Adwords positions. A client can then translate the ranks into dollar terms, and justify this back against your fee.
When you run these competitive reports, you can see what content of theirs is working well, and what content is gaining ground. Make a list of all competitor content that is doing well. Examine where their links are coming from, and make a list. Examine where they’re mentioned in the media, and make a list. You can then use a fast-follow strategy to emulate their success, then expand upon it.
Sometimes, “competitors”, meaning ranking competitors, can actually be potential partners. They may not be in the same industry as your client, just happen to rank in a cross-over area. They may be good for a link, become a supplier, welcome advertising on their site, or be willing to place your content on their site. Make a note of the sites that are ranking well within your niche, but aren’t direct competitors.
Using tools that estimate the value of ranks by comparing Adwords keywords prices, you can estimate the value of your competitors positions. If your client appears lower than the competition, you can demonstrate the estimated dollar value of putting time and effort into increasing rank. You can also evaluate their rate of improvement over time vs your client, and use this as a competitive benchmark. If your client is not putting in the same effort as your competitor, they’ll be left behind. If their competitors are spending on ongoing-SEO and seeing tangible results, there is some validation for your client to do likewise.
Further reading:
– Competitor Analysis [pdf]
– Illustrated SEO Competitive Workflow
– Competitive Analysis: How To Become A SEO Hero In 4 Steps
4. Site Architecture
A well organised site is both useful from a usability standpoint and an SEO standpoint. If it’s clear to a user where they need to go next, then this will flow through into better engagement scores. If your client has a usability consultant on staff, this person is a likely ally.
It’s a good idea to organise a site around themes. Anecdotal evidence suggests that Google likes pages grouped around similar topics, rather than disparate topics (see from 1.25 onwards).
- Create spreadsheet based on a crawl after any errors have been tidied up
- Identify best selling products and services. These deserve the most exposure and should be placed high up the site hierarchy. Items and categories that do not sell well, and our less strategically important, should be lower in the hierarchy
- Pages that are already getting a lot of traffic, as indicated by your analytics, might deserve more exposure by moving them up the hierarchy.
- Seasonal products might deserve more exposure just before that shopping season, and less exposure when the offer is less relevant.
- Group pages into similar topics, where possible. For example, acme.com/blue-widgets/ , acme.com/green-widgets/.
- Determine if internal anchor text is aligned with keyword titles and page content by looking at a backlink analysis
A spreadsheet of all pages helps you group pages thematically, preferably into directories with similar content. Your strategy document will guide you as to which pages you need to work on, and which pages you need to religate. Some people spend a lot of time sculpting internal pagerank i.e. flowing page rank to some pages, but using nofollow on other links to not pass link equity to others. Google may have depreciated that approach, but you can still link to important products or categories sitewide to flow them more link equity, while putting less important sites lower in the site’s architecture. Favour your money pages, and relegate your less important pages.
Think mobile. If your content doesn’t work on mobile, then getting to the top of search results won’t do you much good.
Further Reading:
– Site Architecture & Search Engine Success Factors
– Optimiing Your Websites Architecture For SEO (Slide Presentation)
– The SEO Guide To Information Archetecture
5. Enable Crawling & Redirects
Ensure your site is deep crawled. To check if all your URLs are included in Google’s index, sign up with Webmaster Tools and/or other index reporting tools.
- Include a site map
- Check the existing robots.txt. Kep robots out of non-essential areas, such as script repositories and other admin related directories.
- If you need to move pages, or you have links to pages that no longer exist, use page redirects to tidy them up
- Make a list of 404 errors. Make sure the 404 page has useful navigation into the site so visitors don’t click back.
The accepted method to redirect a page is to use a 301. The 301 indicates a page has permanently moved location. A redirect is also useful if you change domains, or if you have links pointing to different versions of the site. For example, Google sees http://www.acme.com and http://acme.com as different sites. Pick one and redirect to it.
Here’s a video explaining how:
If you don’t redirect pages, then you won’t be making full use of any link juice allocated to those pages.
Further Reading:
– What Are Google Site Maps?
– The Ultimate Guide To 301 Redirects
– Crawling And Indexing Metrics
6. Backlink Analysis
Backlinks remain a major ranking factor. Generally, the more high quality links you have pointing to your site, the better you’ll do in the results. Of late, links can also harm you. However, if your overall link profile is strong, then a subset of bad links is unlikely to cause you problems. A good rule of thumb is the Matt Cutts test. Would you be happy to show the majority of your links to Matt Cutts? :) If not, you’re likely taking a high risk strategy when it comes to penalties. These can be manageable when you own the site, but they can be difficult to deal with on client sites, especially if the client was not aware of the risks involved in aggressive SEO.
- Establish a list of existing backlinks. Consider trying to remove any that look low quality.
- Ensure all links resolve to appropriate pages
- Draw up a list of sites from which your main competitors have gained links
- Draw up a list of sites where you’d like to get links from
Getting links involves either direct placement or being linkworthy. On some sites, like industry directories, you can pay to appear. In other cases, it’s making your site into an attractive linking target.
Getting links to purely commercial sites can be a challenge. Consider sponsoring charities aligned with your line of business. Get links from local chambers of commerce. Connect with education establishments who are doing relevant research and consider sponsoring or become involved in some way.
Look at the sites that point to your competitors. How were these links obtained? Follow the same path. If they successfully used white papers, then copy that approach. If they successfully used news, do that, too. Do whatever seems to work for others. Evaluate the result. Do more/less of it, depending on the results.
You also need links from sites that your competitors don’t have. Make a list of desired links. Figure out a strategy to get them. It may involve supplying them with content. It might involve participating in their discussions. It may involve giving them industry news. It might involve interviewing them or profiling them in some way, so they link to you. Ask “what do they need”?. Then give it to them.
Of course, linking is an ongoing strategy. As a site grows, many links will come naturally, and that in itself, is a link acquisition strategy. To grow in importance and consumer interest relative to the competition. This involves your content strategy. Do you have content that your industry likes to link to? If not, create it. If your site is not something that your industry links to, like a brochure site, you may look at spinning-off a second site that is information focused, and less commercial focused. You sometimes see blogs on separate domains where employees talk about general industry topics, like Signal Vs Noise, Basecamps blog. These are much more likely to receive links than sites that are purely commercial in nature.
Before chasing links, you should be aware of what type of site typically receives links, and make sure you’re it.
Further Reading:
– Interview Of Debra Mastaler, the Link Guru
– Scaleable Link Building Techniques
– Creative Link Building Ideas
7 Content Assessment
Once you have a list of keywords, an idea of where competitors rank, and what the most valuable terms are from a business point of view, you can set about examining and building out content.
Do you have content to cover your keyword terms? If not, add it to the list of content that needs to be created. If you have content that matches terms, see if compares well with client content on the same topic. Can the pages be expanded or made more detailed? Can more/better links be added internally? Will the content benefit from amalgamating different content types i.e. videos, audio, images et al?
You’ll need to create content for any keyword areas you’re missing. Rather than copy what is already available in the niche, look at the best ranking/most valuable content for that term and ask how it could be made better. Is there new industry analysis or reports that you can incorporate and/or expand on? People love the new. They like learning things they don’t already know. Mee-too content can work, but it’s not making the most of the opportunity. Aim to produce considerably more valuable content than already exists as you’ll have more chance of getting links, and more chance of higher levels of engagement when people flip between sites. If visitors can get the same information elsewhere, they probably will.
Consider keyword co-occurrence. What terms are readily associated with the keywords you’re chasing? Various tools provide this analysis, but you can do it yourself using the Adwords research tool. See what keywords it associates with your keywords. The Google co-occurrence algorithm is likely the same for both Adwords and organic search.
Also, think about how people will engage with your page. Is it obvious what the page is about? Is it obvious what the user must do next? Dense text and distracting advertising can reduce engagement, so make sure the usability is up to scratch. Text should be a reasonable size so the average person isn’t squinting. It should be broken up with headings and paragraphs. People tend to scan when reading online,searching for immediate confirmation they’ve found the right information. This was written a long time ago, but it’s interesting how relevant it remains.
Further Reading:
– Content Marketing Vs SEO
– Content Analysis Using Google Analytics
– Content Based SEO Strategy Will Eventually Fail
8. Link Out
Sites that don’t link out appear unnatural. Matt Cutts noted:
Of course, folks never know when we’re going to adjust our scoring. It’s pretty easy to spot domains that are hoarding PageRank; that can be just another factor in scoring. If you work really hard to boost your authority-like score while trying to minimize your hub-like score, that sets your site apart from most domains. Just something to bear in mind.
- Make a list of all outbound links
- Determine if these links are complementary i.e. similar topic/theme, or related to the business in some way
- Make a list of pages with no links out
Links out are both a quality signal and good PR practise. Webmaster look at their inbound links, and will likely follow them back to see what is being said about them. That’s a great way to foster relationships, especially if your client’s site is relatively new. If you put other companies and people in a good light, you can expect many to reciprocate in kind.
Links, the good kind, are about human relationships.
It’s also good for your users. Your users are going to leave your site, one way or another, so you can pick up some kudos if you help them on their way by pointing them to some good authorities. If you’re wary about linking to direct competitors, then look for information resources, such as industry blogs or news sites, or anyone else you want to build a relationship with. Link to suppliers and related companies in close, but non-competing niches. Link to authoritative sites. Be very wary about pointing to low value sites, or sites that are part of link schemes. Low value sites are obvious. Sites that are part of link schemes are harder to spot, but typically feature link swapping schemes or obvious paid links unlikely to be read by visitors. Avoid link trading schemes. It’s too easy to be seen as a part of a link network, and it’s no longer 2002.
Further Resources:
– Five Reasons You Should Link Out
– The Domino Effects Of Links And Relationships
– Link Building 101: Utilizing Past Relationships
9. Ongoing
It’s not set and forget.
Clients can’t expect to do a one off optimisation campaign and expect it to keep working forever. It may be self-serving for SEOs to say it, but it’s also the truth. SEO is ongoing because search keeps changing and competitors and markets move. Few companies would dream of only having one marketing campaign. The challenge for the SEO, like any marketer, is to prove the on-going spend produces a return in value.
- Competition monitoring i.e. scan for changes in competitors rank, new competitors, and change of tactics. Determine what is working, and emulate it.
- Sector monitoring – monitor Google trends, keywords trends, discussion groups, and news releases. This will give you ideas for new campaign angles.
- Reporting – the client needs to be able to see the work you’ve done is paying off.
- Availability – clients will change things on their site, or bring in other marketers, so will want you advice going forward
Further Reading
Whole books can be written about SEO for clients. And they have. We’ve skimmed across the surface but, thankfully, there is a wealth of great information out there on the specifics of how to tackle each of these topic areas.
Perhaps you can weigh in? :) What would your advice be to those new to optimizing client sites? What do you wish someone had told you when you started?
Google Search Censorship for Fun and Profit
Growing Up vs Breaking Things
Facebook’s early motto was “move fast and break things,” but as they wanted to become more of a platform play they changed it to “move fast with stability.” Anything which is central to the web needs significant stability, or it destroys many other businesses as a side effect of its instability.
As Google has become more dominant, they’ve moved in the opposite direction. Disruption is promoted as a virtue unto itself, so long as it doesn’t adversely impact the home team’s business model.
There are a couple different ways to view big search algorithm updates. Large, drastic updates implicitly state one of the following:
- we were REALLY wrong yesterday
- we are REALLY wrong today
Any change or disruption is easy to justify so long as you are not the one facing the consequences:
“Smart people have a problem, especially (although not only) when you put them in large groups. That problem is an ability to convincingly rationalize nearly anything.” … “Impostor Syndrome is that voice inside you saying that not everything is as it seems, and it could all be lost in a moment. The people with the problem are the people who can’t hear that voice.” – Googler Avery Pennarun
Monopoly Marketshare in a Flash
Make no mistake, large changes come with false positives and false negatives. If a monopoly keeps buying marketshare, then any mistakes they make have more extreme outcomes.
Here’s the Flash update screen (which hits almost every web browser EXCEPT Google Chrome).
Notice the negative option installs for the Google Chrome web browser and the Google Toolbar in Internet Explorer.
Why doesn’t that same process hit Chrome? They not only pay Adobe to use security updates to steal marketshare from other browsers, but they also pay Adobe to embed Flash inside Chrome, so Chrome users never go through the bundleware update process.
Anytime anyone using a browser other than Chrome has a Flash security update they need to opt out of the bundleware, or they end up installing Google Chrome as their default web browser, which is the primary reason Firefox marketshare is in decline.
Google engineers “research” new forms of Flash security issues to drive critical security updates.
Obviously, users love it:
Has anyone noticed that the latest Flash update automatically installs Google Toolbar and Google Chrome? What a horrible business decision Adobe. Force installing software like you are Napster. I would fire the product manager that made that decision. As a CTO I will be informing my IT staff to set Flash to ignore updates from this point forward. QA staff cannot have additional items installed that are not part of the base browser installation. Ridiculous that Adobe snuck this crap in. All I can hope now is to find something that challenges Photoshop so I can move my design team away from Adobe software as well. Smart move trying to make pennies off of your high dollar customers.
In Chrome Google is the default search engine. As it is in Firefox and Opera and Safari and Android and iOS’s web search.
In other words, in most cases across most web interfaces you have to explicitly change the default to not get Google. And then even when you do that, you have to be vigilant in protecting against the various Google bundleware bolted onto core plugins for other web browsers, or else you still end up in an ecosystem owned, controlled & tracked by Google.
Those “default” settings are not primarily driven by user preferences, but by a flow of funds. A few hundred million dollars here, a billion there, and the market is sewn up.
Google’s user tracking is so widespread & so sophisticated that their ad cookies were a primary tool for government surveillance efforts.
Locking Down The Ecosystem
And Chrome is easily the most locked down browser out there.
- Chromium is turning into abandonware, with Google stripping features to try to push people over to Chrome.
- Extensions must be installed from the official store. If those extensions deliver malware, no worries. But if those extensions are not aligned with Google’s business model – they will be banned until a commercial relationship aligned with Google’s business model is established. #censorship
- If someone other than Google changes default search settings, it’s time to reset hijacked settings.
- Chrome is so locked down that Yahoo! is canceling their search toolbar for Chrome to comply with recent Google Chrome policy updates, even as Google distributes toolbars in other browsers. #censorship
Whenever Google wants to promote something they have the ability to bundle it into their web browser, operating system & search results to try to force participation. In a fluid system with finite attention, over-promoting one thing means under-promoting or censoring other options. Google likes to have their cake & eat it too, but the numbers don’t lie.
I am frustrated @JohnMu saying that it will not cost CTR. Either Google lied about the increase in CTR with photos, or they’re lying now.— Rand Fishkin (@randfish) June 25, 2014
The Right to Be Forgotten
This brings us back to the current snafu with the “right to be forgotten” in Europe.
Google notified publishers like the BBC & The Guardian of their links being removed due to the EU “right to be forgotten” law. Their goal was to cause a public relations uproar over “censorship” which seems to have been a bit too transparent, causing them to reverse some of the removals after they got caught with their hand in the cookie jar.
The breadth of removals is an ongoing topic of coverage. But if you are Goldman Sachs instead of a government Google finds filtering information for you far more reasonable.
Some have looked at the EU policy and compared it to state-run censorship in China.
Google already hires over 10,000 remote quality raters to rate search results. How exactly is receiving 70,000 requests a monumental task? As their public relations propagandists paint this as an unbelievable burden, they are also highlighting how their own internal policies destroy smaller businesses: “If a multi-billion dollar corporation is struggling to cope with 70,000 censor requests, imagine how the small business owner feels when he/she has to disavow thousands or tens of thousands of links.”
The World’s Richest Librarian
Google aims to promote themselves as a digital librarian: “It’s a bit like saying the book can stay in the library, it just cannot be included in the library’s card catalogue.”
That analogy is absurd on a number of levels. Which librarian…
- tracks people to target ads at them?
- blends ads into their recommendations so aggressively that most users are unable to distinguish the difference between ads and regular recommendations?
- republishes the works of others, offers ultimatums while taking third party content, and obscures or entirely strips the content source?
- invests in, funds & defunds entire lines of publishing?
- claims certain book publishers shall be banned from the library due to nothing other than their underlying business model?
Sorry About That Incidental Deletion From the Web…
David Drummond’s breathtaking propaganda makes it sound like Google has virtually no history in censoring access to information:
In the past we’ve restricted the removals we make from search to a very short list. It includes information deemed illegal by a court, such as defamation, pirated content (once we’re notified by the rights holder), malware, personal information such as bank details, child sexual abuse imagery and other things prohibited by local law (like material that glorifies Nazism in Germany).
Yet Google sends out hundreds of thousands of warning messages in webmaster tools every single month.
Google is free to force whatever (often both arbitrary and life altering) changes they desire onto the search ecosystem. But the moment anyone else wants any level of discourse or debate into the process, they feign outrage over the impacts on the purity of their results.
Despite Google’s great power they do make mistakes. And when they do, people lose their jobs.
Consider MetaFilter.
They were penalized November 17, 2012.
At a recent SMX conference Matt Cutts stated MetaFilter was a false positive.
People noticed the Google update when it happened. It is hard to miss an overnight 40% decline in your revenues. Yet when they asked about it, Google did not confirm its existence. That economic damage hit MetaFilter for nearly two years & they only got a potential reprieve from after they fired multiple employees and were able to generate publicity about what had happened.
As SugarRae mentioned, those false positives happen regularly, but most the people who are hit by them lack political and media influence, and are thus slaughtered with no chance of recovery.
MetaFilter is no different than tens of thousands of other good, worthy small businesses who are also laying off employees – some even closing their doors – as a result of Google’s Panda filter serving as judge, jury and executioner. They’ve been as blindly and unfairly cast away to an island and no one can hear their pleas for help.
The only difference between MetaFilter and tons of other small businesses on the web is that MetaFilter has friends in higher places.
If you read past the headlines & the token slaps of big brands, these false positive death sentences for small businesses are a daily occurrence.
And such stories are understated for fear of coverage creating a witch-hunt:
Conversations I’ve had with web publishers, none of whom would speak on the record for fear of retribution from Cutts’ webspam team, speak to a litany of frustration at a lack of transparency and potential bullying from Google. “The very fact I’m not able to be candid, that’s a testament to the grotesque power imbalance that’s developed,” the owner of one widely read, critically acclaimed popular website told me after their site ran afoul of Cutts’ last Panda update.
Not only does Google engage in anti-competitive censorship, but they also frequently publish misinformation. Here’s a story from a week ago of a restaurant which went under after someone changed their Google listing store hours to be closed on busy days. That misinformation was embedded directly in the search results. That business is no more.
Then there are areas like locksmiths:
I am one of the few Real Locksmiths here in Denver and I have been struggling with this for years now. I only get one or two calls a day now thanks to spammers, and that’s not calls I do, it’s calls for prices. For instance I just got a call from a lady locked out of her apt. It is 1130 pm so I told her 75 dollars, Nope she said someone told her 35 dollars….a fake locksmith no doubt. She didn’t understand that they meant 35 dollars to come out and look at it. These spammers charge hundreds to break your lock, they don’t know how to pick a lock, then they charge you 10 times the price of some cheap lock from a hardware store. I’m so lost, I need help from google to remove those listings. Locksmithing is all I have ever done and now I’m failing at it.
There are entire sectors of the offline economy being reshaped by Google policies.
When those sectors get coverage, the blame always goes to the individual business owner who was personal responsible for Google’s behaviors, or perhaps some coverage of the nefarious “spammers.”
Never does anybody ask if it is reasonable for Google to place their own inaccurate $0 editorial front and center. To even bring up that issue makes one an anti-capitalist nut or someone who wishes to impede on free speech rights. This even after the process behind the sausage comes to light.
And while Google arbitrarily polices others, their leaked internal documents contain juicy quotes about their ad policies like:
- “We are the only player in our industry still accepting these ads”
- “We do not make these decisions based on revenue, but as background, [redacted].”
- “As with all of our policies, we do not verify what these sites actually do, only what they claim to do.”
- “I understand that we should not let other companies, press, etc. influence our decision-making around policy”
Is This “Censorship” Problem New?
This problem of control to access of information is nothing new – it is only more extreme today. Read the (rarely read) preface to Animal Farm, or consider this:
John Milton in his fiery 1644 defense of free speech, Areopagitica, was writing not against the oppressive power of the state but of the printers guilds. Darnton said the same was true of John Locke’s writings about free speech. Locke’s boogeyman wasn’t an oppressive government, but a monopolistic commercial distribution system that was unfriendly to ways of organizing information that didn’t fit into its business model. Sound familiar?
When Google complains about censorship, they are not really complaining about what may be, but what already is. Their only problem is the idea that someone other than themselves should have any input in the process.
“Policy is largely set by economic elites and organized groups representing business interests with little concern for public attitudes or public safety, as long as the public remains passive and obedient.” ― Noam Chomsky
Many people have come to the same conclusion
Turn on, tune in, drop out
“I think as technologists we should have some safe places where we can try out some new things and figure out what is the effect on society, what’s the effect on people, without having to deploy kind of into the normal world. And people like those kind of things can go there and experience that and we don’t have mechanisms for that.” – Larry Page
I have no problem with an “opt-in” techno-utopia test in some remote corner of the world, but if that’s the sort of operation he wants to run, it would be appreciated if he stopped bundling his software into billions of electronic devices & assumed everyone else is fine with “opting out.”
{This | The Indicated} {Just | True} {In | Newfangled}
A couple years ago we published an article named Branding & the Cycle, which highlighted how brands would realign with the algorithmic boost they gained from Panda & leverage their increased level of trust to increase their profit margins by leveraging algorithmic journalism.
Narrative Science has been a big player in the algorithmic journalism game for years. But they are not the only player in the market. Recently the Associated Press (AP) announced they will use algorithms to write articles based on quarterly earnings reports, working with a company named Automated Insights:
We discovered that automation technology, from a company called Automated Insights, paired with data from Zacks Investment Research, would allow us to automate short stories – 150 to 300 words — about the earnings of companies in roughly the same time that it took our reporters.
And instead of providing 300 stories manually, we can provide up to 4,400 automatically for companies throughout the United States each quarter.
…
Zacks maintains the data when the earnings reports are issued. Automated Insights has algorithms that ping that data and then in seconds output a story.
In the past Matt Cutts has mentioned how thin rewrites are doorway page spam:
you can also have more subtle doorway pages. so we ran into a directv installer in denver, for example. and that installer would say I install for every city in Colorado. so I am going to make a page for every single city in Colorado. and Boulder or Aspen or whatever I do directv install in all of those. if you were just to land on that page it might look relatively reasonable. but if you were to look at 4 or 5 of those you would quickly see that the only difference between them is the city, and that is something that we would consider a doorway.
One suspects these views do not apply to large politically connected media bodies like the AP, which are important enough to have a direct long-term deal with Google.
In the above announcement the AP announced they include automated NFL player rankings. One interesting thing to note about the AP is they have syndication deals with 1,400 daily newspapers nationwide, as well as thousands of TV and radio stations..
A single automated AP article might appear on thousands of websites. When thousands of articles are automated, that means millions of copies. When millions of articles are automated, that means billions of copies. When billions … you get the idea.
To date Automated Insights has raised a total of $10.8 million. With that limited funding they are growing quickly. Last year their Wordsmith software produced 300 million stories & this year it will likely exceed a billion articles:
“We are the largest producer of content in the world. That’s more than all media companies combined,” [Automated Insights CEO Robbie Allen] said in a phone interview with USA TODAY.
The Automated Insights homepage lists both Yahoo! & Microsoft as clients.
The above might sound a bit dystopian (for those with careers in journalism and/or lacking equity in Automated Insights and/or publishers who must compete against algorithmically generated content), but the story also comes with a side of irony.
Last year Google dictated press releases shall use nofollow links. All the major press release sites quickly fell in line & adopted nofollow, thinking they would remain in Google’s good graces. Unfortunately for those sites, they were crushed by Panda. PR Newswire’s solution their penalty was greater emphasis on manual editorial review:
Under the new copy quality guidelines, PR Newswire editorial staff will review press releases for a number of message elements, including:
- Inclusion of insightful analysis and original content (e.g. research, reporting or other interesting and useful information);
- Use of varied release formats, guarding against repeated use of templated copy (except boilerplate);
- Assessing release length, guarding against issue of very short, unsubstantial messages that are mere vehicles for links;
- Overuse of keywords and/or links within the message.
So now we are in a situation where press release sites require manual human editorial oversight to try to get out of being penalized, and the news companies (which currently enjoy algorithmic ranking boosts) are leveraging those same “spammy” press releases using software to auto-generate articles based on them.
That makes sense & sounds totally reasonable, so long as you don’t actually think about it (or work at Google)…
Building a Legal Moat for Your SEO Castle
About a month ago a year old case of an SEO firm being sued by it’s client resurfaced via a tweet from Matt Cutts.
I’d like to add something to this conversation that will be helpful for you as a service provider seeking to avoid that really, really scary issue.
Some quick background information if I may? No specifics are allowed but I’ve been a party, on both sides, to actual litigation pertaining to SEO contracts (not services rendered, just contractual issues with a third-party).
I’ve been the plaintiff and the defendant in cases involving contractual disputes and legal obligations so I, much to my dismay, speak from experience.
Suffice to say I’m not a lawyer, don’t act on any of this advice without talking it over with your counsel so they can tailor it to your specific needs and state law.
Basic Protections
There are essentially 3 ways to legally protect yourself and/or your company objectively. I say objectively because anyone can sue you for anything and “service” is a subjective term as are “results” unless they are specifically spelled out in your contract.
Objectively speaking, the law gives you 3 broad arenas for protective measures:
- Contracts
- Entity Selection
- Insurance
Contracts
Get a real lawyer, do not use internet “templates” and do not modify any piece of the contract yourself. Make sure your attorney completely understands what you do. A good lawyer will listen to you. Heck, mine now knows who Matt Cutts is and where the Webmaster Guidelines are located and what “anchor text” is :)
Your contracts need to cover the following scenarios:
- Client services
- Vendor relationships
- Employee/Contractor relationships
For standard client agreements you’ll want to cover some basic areas:
- Names of the legal entities partaking in the agreement
- Duties and nature of services
- Term and termination (who can cancel and when, what are the ramifications, etc)
- Fees
- No exclusive duty (a clause that says you can work with other clients and such)
- Disclaimer, Limitation of Liability
- Confidentiality
- Notices (what is considered legal notice? a letter? certified mail? email?)
- Governing law
- Attorney’s fees (if you need to enforce the contract make sure you can also collect fees)
- Relationship of Parties (spell out the relationship; independent entities? partners? joint ventures? spell out exactly what you are and what you are not
- Scope of Work
- Signatures (you should sign as you are in your entity; member, president, CEO, etc)
Some important notes are needed to discussion a couple of core areas of the contract:
For Governing law go with your home state if possible. Ideally, I try to get an arbitration clause in there rather than state law so in case there is a dispute it goes to a much less expensive form of resolution.
However, you can make an argument that if your contract is signed with your home state as governing law and your language is strong you are better off doing that instead of arbitration where one person makes a decision and no appeal is available.
For Limit of Liability go broad, real broad. You want to spell out that organic search (or just about any service) is not guaranteed to produce results, no promises were made, Google does not fully publish the algorithim thus you can’t be held liable for XYZ that happens.
Also, if your client is asking you to do things against webmaster guidelines, and you decide to do them, you NEED to get that documented. Have them email it to you, record the call, something. Here is the liability clause in my contract:
Client agrees and acknowledges that the internet is an organic, constantly shifting entity, and that Client’s ranking and/or performance in a search engine may change for many reasons and be affected by many factors, including but not limited to any actual or alleged non-compliance by Provider to guidelines set forth by Google related to search engine optimization.
Client agrees that no representation, express or implied, and no warranty or guaranty is provided by Provider with respect to the services to be provided by Provider under this Agreement. Provider’s services may be in the form of rendering consultation which Client may or may not choose to act on. To the maximum extent permitted by law, Client agrees to limit the liability of Provider and its officers, owners, agents, and employees to the sum of Provider’s fees actually received from Client.
This limitation will apply regardless of the cause of action or legal theory pled or asserted. In no event shall Provider be liable for any special, incidental, indirect, or consequential damages arising from or related to this Agreement or the Project. Client agrees, as a material inducement for Provider to enter into this Agreement that the success and/or profitability of Client’s business depends on a variety of factors and conditions beyond the control of Provider and the scope of this Agreement. Provider makes no representations or warranties of any kind regarding the success and/or profitability of Client’s business, or lack thereof, and Provider will not be liable in any manner respecting the same.
Client agrees to indemnify and hold harmless Provider and its officers, owners, agents, and employees from and against any damages, claims, awards, and reasonable legal fees and costs arising from or related to any services provided by Provider, excepting only those directly arising from Provider’s gross negligence or willful misconduct.
For vendor and independent contractor agreements you’ll want most of the aforementioned clauses (especially the relationship of parties) in addition to a few more things (for employee stuff, get with your lawyer because states are quite different and a lot of us use remote workers in different states)
- Non-Competition and non-interference
- Non-Solicitation and non-contact
These clauses essentially prohibit the pursuit of your clientele and employees by a vendor/contractor for a specified period of time.
Legal Entity
Don’t be a sole proprietor, ever. If you’re a smaller shop you might consider being a single member LLC (just you), an LLC (you and employees), or an S Corp. If you’re a larger operation you might want to incorporate and go Inc.
The benefits of the LLC set up are:
- Your personal assets are generally untouchable (providing you are not co-mingling funds in a bank account)
- Very easy to administer compared to other options
- Your liability is limited to company assets (pro tip: clear out your business bank account each month minus some operating margin, move it to personal savings)
Benefits of an S Corp are:
- Same protections as LLC
- You save a fair amount on self employment taxes (more below)
With the S Corp there’s more paperwork and filings but if you are earning a fair bit of money it may be worth it to you. Here’s a good article breaking this all down, and a excerpt:
“If you operate your business as a sole proprietorship or partnership/LLC, you will pay roughly 15.3% in self-employment taxes on your $100,000 of profits. The calculations get a little tricky if you want to be really super-precise but you can think about self-employment tax as roughly a 15% tax. So 15% on $100,000 equals $15,000. Roughly.”
“With an S corporation, you split your business profits into two categories: “shareholder wages” and “distributive share.” Only the “shareholder wages” get subjected to the 15.3% tax. The leftover “distributive share” is not subject to 15.3% tax.”
Be careful here (and I’m not a CPA so don’t do anything without consulting with your accountant) not to be absurd with your wages. So, if your net income is 1 million don’t take 25k in wages and 975k as a distribution.
Some final thoughts on entities:
- Most of you will probably fall into the LLC/S S Corp category, get with your attorney and accountant
- Keep everything separate because if you don’t (credit cards, bank accounts, etc) your personal assets might be at risk due to the “piercing of the corporate veil”
Insurance
As you would imagine, insurance policies are few and far between for our industry. You can get general liability for your office, workers comp for your employees, disability for yourself, and so on. However, what you might want to look into is a professional liability policy.
You’ll probably end up looking at a miscellaneous one like the one here (marketing consultant?) offered by Travelers. You’ll probably have to educate your agent on your business practices to ensure proper coverage.
This might be worth it just due to the legal protection clause; meaning they will pay for a lawyer to defend you. Having the proper entity classification might protect your assets but paying lawyers is expensive to defend even frivolous lawsuits.
Record Keeping
This is a bit out of the “contract” topic but good record keeping is essential. If you use a project management and/or a CRM system you really should make sure you can export when you need it.
Many online CRM applications and project management applications have limited export capabilities especially when it comes to export comments and notes on things like tasks and records. Most have an API that you can have a developer custom code to export your stuff. I’d look into this as well.
Final Thoughts
Get with your attorney and CPA to get your specific situations up to legal snuff if you haven’t already. Don’t act on my advice as I’m not a lawyer nor a CPA. Contracts and agreements are not fun to negotiate and can be even harder when you work with people you generally trust.
However, when it comes to business dealings and contracts I would save my trust for my lawyer :)
Have We Reached Peak Advertising?
The internet runs on advertising. Google is funded almost entirely by advertising. Facebook , likewise. Digital marketing spends continue to rise:
Internet advertising revenues in the United States totaled $12.1 billion in the fourth quarter of 2013, an increase of 14% from the 2013 third-quarter total of $10.6 billion and an increase of 17% from the 2012 fourth-quarter total of $10.3 billion. 2013 full year internet advertising revenues totaled $42.78 billion, up 17% from the $36.57 billion reported in 2012.
Search advertising spend comes out on top, but that’s starting to change:
Search accounted for 41% of Q4 2013 revenues, down from 44% in Q4 2012, as mobile devices have shifted. Search-related revenues away from the desktop computer. Search revenues totaled $5.0 billion in Q4 2013, up 10% from Q4 2012, when Search totaled $4.6 billion
The growth area for digital advertising lays in mobile:
Mobile revenues totaled 19% of Q4 2013 revenues, or $2.3 billion, up 92% from the $1.2 billion (11% of total) reported in Q4 2012
Prominent venture capitalist, Mary Meeker, recently produced an analysis that also highlights this trend.
So, internet advertising is growing, but web internet adoption is slowing down. Meanwhile, mobile and tablet adoption is increasing fast, yet advertising spend on these mediums is comparatively low. Nice opportunity for mobile, however mobile advertising is proving hard to crack. Not many people are clicking on paid links on mobile. And many mobile ad clicks are accidental, driving down advertiser bids.
This is not just a problem for mobile. There may be a problem with advertising in general. It’s about trust, and lack thereof. This situation also presents a great opportunity for selling SEO.
But first, a little background….
People Know More
Advertising’s golden age was in the 50’s and 60’s.
Most consumers were information poor. At least, they were information poor when it came to getting timely information. This information asymmetry played into the hands of the advertising industry. The advertising agency provided the information that helped match the problems people had with a solution. Of course, they were framing the problem in a way that benefited the advertiser. If there wasn’t a problem, they made one up.
Today, the internet puts real time information about everything in the hands of the consumer. It is easy for people to compare offers, so the basis for advertising – which is essentially biased information provision – is being eroded. Most people see advertising as an intrusion. Just because an advertiser can get in front of a consumer at “the right time” does not necessarily mean people will buy what the advertiser has to offer with great frequency.
Your mobile phone pings. “You’re passing Gordon’s Steak House….come in and enjoy our Mega Feast!” You can compare that offer against a wide range of offers, and they can do so in real time. More than likely, you’ll just resent the intrusion. After all, you may be a happy regular at Susan’s Sushi.
“Knowing things” is not exclusive. Being able to “know things” is a click away. If information is freely available, then people are less likely to opt for whatever is pushed at them by advertisers at that moment. If it’s easy to research, people will do so.
This raises a problem when it comes to the economics of content creation. If advertising becomes less effective for the advertiser, then the advertisers is going to reduce spend, or shift spend elsewhere. If they do, then what becomes of the predominant web content model which is based on advertising?
Free Content Driven By Ads May Be An Unsustainable Model
We’re seeing it in broadcast television, and we’ll see it on the web.
Television is dying and being replaced by the Netflix model. There is a lot of content. There are not enough advertisers paying top dollar as the audience is now highly fragmented. As a result, a lot of broadcast television advertising can be ineffective. However, as we’ve seen with Netflix and Spotify, people are prepared to pay directly for the content they consume in the form of a monthly fee.
The long term trend for advertising engagement on the web is not favourable.
The very first banner advertisement appeared in 1994. The clickthru rate of that banner ad was a staggering 44% It had a novelty value, certainly. The first banner ad also existed in an environment where there wasn’t much information. The web was almost entirely about navigation.
Today, there is no shortage of content. The average Facebook advertisement clickthrough rate is around 0.04%. Advertisers get rather excited if they manage to squeeze 2% or 3% click-thrus rates out of Facebook ads.
Digital advertising is no longer novel, so the click-thru rate has plummeted. Not only do people feel that the advertising isn’t relevant to them, they have learned to ignore advertising even if the ad is talking directly to their needs. 97-98% of the time, people will not click on the ad.
And why should they? Information isn’t hard to come by. So what is the advertiser providing the prospective customer?
Even brand engagement is plummeting on Facebook as the novelty wears off, and Facebook changes policy:
According to a new report from Simply Measured, the total engagement for the top 10 most-followed brands on Facebook has declined 40 percent year-over-year—even as brands have increased the amount of content they’re posting by 20.1 percent.
Is Advertising Already Failing?
Our industry runs on advertising. Much of web publishing runs on advertising.
However, Eric Clemons makes the point that the traditional method of advertising was always bound to fail, mainly because after the novelty wears off, it’s all about interruption, and nobody likes to be interrupted.
But wait! Isn’t the advantage of search that it isn’t interruption advertising? In search, the user requests something. Clemons feels that search results can still be a form of misdirection:
Misdirection, or sending customers to web locations other than the ones for which they are searching. This is Google’s business model. Monetization of misdirection frequently takes the form of charging companies for keywords and threatening to divert their customers to a competitor if they fail to pay adequately for keywords that the customer is likely to use in searches for the companies’ products; that is, misdirection works best when it is threatened rather than actually imposed, and when companies actually do pay the fees demanded for their keywords. Misdirection most frequently takes the form of diverting customers to companies that they do not wish to find, simply because the customer’s preferred company underbid.
He who pays becomes “relevant”:
it is not scalable; it is not possible for every website to earn its revenue from sponsored search and ultimately at least some of them will need to find an alternative revenue model.
The companies that appear high on PPC are the companies who pay. Not every company can be on top, because not every company can pay top dollar. So, what the user sees is not necessarily what the user wants, but the company that has paid the most – along with their quality score – to be there.
But nowadays, the metrics of this channel have changed dramatically, making it impossible or nearly impossible for small and mid-sized business to turn a profit using AdWords. In fact, most small businesses can’t break even using AdWords.This goes for many large businesses as well, but they don’t care. And that is the key difference, and precisely why small brands using AdWords nowadays are being bludgeoned out of existence
Similarly, the organic search results are often dominated by large companies and entities. This is a direct or side-effect of the algorithms. Big entities create a favourable footprint of awareness, engagement and links as a result of PR, existing momentum, brand recognition, and advertising campaigns. It’s a lot harder for small companies to dominate lucrative competitive niches as they can’t create those same footprints.
Certainly when it comes to PPC, the search visitor may be presented with various big player links at the expense of smaller players. Google, like every other advertising driven medium, is beholden to it’s big advertisers. Jacob Nielsen noted in 1998:
Ultimately, those who pay for something control it. Currently, most websites that don’t sell things are funded by advertising. Thus, they will be controlled by advertisers and will become less and less useful to the users”
If Interruption Advertising Is Failing, Is Advertising Scalable?
Being informed has changed customer behaviour.
The problem is not the medium, the problem is the message, and the fact that it is not trusted, not wanted, and not needed.
People don’t trust ads. There is a vast literature to support this. Is it all wrong?
People don’t want ads. Again, there is a vast literature to support this. Think about your own behavior, you own channel surfing and fast forwarding and the timing of when you leave the TV to get a snack. Is it during the content or the commercials?
People don’t need ads. There is a vast amount of trusted content on the net. Again, there is literature on this. But think about how you form your opinion of a product, from online ads or online reviews?
There is no shortage of places to put ads. Competition among them will be brutal. Prices will be driven lower and lower, for everyone but Google.
If the advertising is not scaleable, then a lot of content based on advertising will die. Advertising may not be able to support the net:
Now reality is reasserting itself once more, with familiar results. The number of companies that can be sustained by revenues from internet advertising turns out to be much smaller than many people thought, and Silicon Valley seems to be entering another “nuclear winter”
A lot of Adsense publishers are being kicked from the program. Many are terminated, without reason. Google appear to be systematically culling the publisher herd. Why? Shouldn’t web publishing, supported by advertising, be growing?
The continuing plunge in AdSense is in sharp contrast to robust 20% revenue growth in 2012, which outpaced AdWords’ growth of 19%…..There are serious issues with online advertising affecting the entire industry. Google has reported declining value from clicks on its ads. And the shift to mobile ads is accelerating the decline, because it produces a fraction of the revenue of desktop ads.
Matt Sanchez, CEO of San Francisco based ad network Say Media, recently warnedthat, “Mobile Is Killing Media.”
Digital publishing is headed off a cliff … There’s a five fold gap between mobile revenue and desktop revenue… What makes that gap even starker is how quickly it’s happening… On the industry’s current course, that’s a recipe for disaster.
Prices tumble when consumers have near-perfect real time information. Travel. Consumer goods. Anything generic that can be readily compared is experiencing falling prices and shrinking margins. Sales growth in many consumer categories is coming from the premium offerings. For example, beer consumption is falling across the board except in one area: boutique, specialist brews. That market sector is growing as customers become a lot more aware of options that are not just good enough, but great. Boutique breweries offer a more personal relationship, and they offer something the customer perceives as being great, not just “good enough”.
Mass marketing is expensive. Most of the money spent on it is wasted. Products and services that are “just good enough” will be beaten by products and services that are a precise fit for consumers needs. Good enough is no longer good enough, products and services need to be great and precisely targeted unless you’ve got advertising money to burn.
How Do We Get To These Consumers If They No Longer Trust Paid Advertising?
Consumers will go to information suppliers they trust. There is always demand for a trusted source.
Trip Advisor is a great travel sales channel. It’s a high trust layer over a commodity product. People don’t trust Trip Advisor, per se, they trust the process. Customers talk to each other about the merits, or otherwise, of holiday destinations. It’s transparent. It’s not interruption, misleading or distracting. Consumers seek it out.
Trust models will be one way around the advertising problem. This suits SEOs. If you provide trusted information, especially in a transparent, high-trust form, like Trip Advisor, you will likely win out over those using more direct sales methods. Consumers are getting a lot better at tuning those out.
The trick is to remove the negative experience of advertising by not appearing to be advertising at all. Long term, it’s about developing relationships built on trust, not on interruption and misdirection. It’s a good idea to think about advertising as a relationship process, as opposed to the direct marketing model on which the web is built – which is all about capturing the customer just before point of sale.
Rand Fishkin explained the web purchase process well in this presentation. The process whereby someone becomes a customer, particularly on the web, isn’t all about the late stages of the transaction. We have to think of it in terms of a slow burning relationship developed over time. The consumer comes to us at the end of an information comparison process. Really, it’s an exercise in establishing consumer trust.
Amazon doesn’t rely on advertising. Amazon is a trusted destination. If someone wants to buy something, they often just go direct to Amazon. Amazon’s strategy involves what it calls “the flywheel”, whereby the more things people buy from Amazon, the more they’ll buy from Amazon in future. Amazon builds up a relationship rather than relying on a lot of advertising. Amazon cuts out the middle man and sells direct to customers.
Going viral with content, like Buzzfeed, may be one answer, but it’s likely temporary. It, too, suffers from a trust problem and the novelty will wear off:
Saying “I’m going to make this ad go viral” ignores the fact that the vast majority of viral content is ridiculously stupid. The second strategy, then, is the high-volume approach, same as it ever was. When communications systems wither, more and more of what’s left is the advertising dust. Junk mail at your house, in your email; crappy banner ads on MySpace. Platforms make advertising cheaper and cheaper in a scramble to make up revenue through volume.
It’s not just about supplying content. It could be said newspapers are suffering because bundled news is just another form of interruption and misdirection, mainly because it isn’t specifically targeted:
Following The New York Times on Twitter is just like paging through a print newspaper. Each tweet is about something completely unrelated to the tweets before it. And this is the opposite of why people usually follow people and brands online. It’s not surprising that The New York Times have a huge problem with engagement. They have nothing that people can connect and engage with
Eventually, the social networks will likely suffer from a trust problem, if they don’t already. Their reliance on advertising makes them spies. There is a growing awareness of data privacy and users are unlikely to tolerate invasions of privacy, especially if they are offered an alternative. Or perhaps the answer is to give users a cut themselves. Lady Gaga might be onto something.
Friends “selling” (recommending) to friends is a high trust environment.
A Good Approach To SEO Involves Building Consumer Trust
The serp is low trust. PPC is low trust. Search keyword plus a site that is littered with ads is low trust. So, one good long term future strategy is to move from low to high trust advertising.
A high trust environment doesn’t really look like advertising. It’s could be characterised as a transparent platform. Amazon and Trip Advisor are good examples. They are honest about what they are, and they provide the good along with the bad. It could be something like Wikipedia. Or an advisory site. There are many examples, but it’s fair to say we know it when we see it.
A search on a keyword that finds a specific, relevant site that isn’t an obvious advertisement is high trust. The first visit is the start of a relationship. This is not the time to bombard visitors with your needs. Instead, give the visitor something they can trust. Trip Advisor even spells it out: “Find hotels travelers trust”.
Telsla understands the trust relationship. Recently, they’ve made their patents open-source, which, apart from anything else, is a great form of reputation marketing. It’s clear Telsa is more interested in long term relationships and goodwill than pushing their latest model on you at a special price. Their transparency is endearing.
First, you earn trust. Then you sell them something later. If you don’t earn their trust, then you’re just like any other advertiser. People will compare you. People will seek out information. You’re one of many options, unless you have formed a prior relationship. SEO is a brilliant channel to develop a relationship based on trust. If you’re selling SEO to clients, think about discussing the trust building potential – and value proposition – of SEO with them.
It’s a nice side benefit of SEO. And it’s a hedge against the problems associated with other forms of advertising.
The Rigged Search Game
SEO was all about being clever. Still is, really. However, SEO used to reward the clever, too. The little guy could take on the big guys and munch their lunch by outsmarting them.
It was such an appealing idea.
The promise of the internet was that the old power structures would be swept aside, the playing field would be made level again, and those who played the smartest game would prosper.
Sadly, this promise didn’t last long.
Power
The names may have changed, but traditional power structures were soon reasserted. The old gatekeepers were replaced with the new gatekeepers. The new gatekeepers, like Google, grew fat, rich and powerful. They controlled the game and the game was, once again, rigged in favor of those with the most power. That’s not a Google-specific criticism, it’s just the way commerce works. You get big, you move markets simply by being big and present. In search, we see the power imbalance as a side-effect, namely the way big players are treated in the SERPs compared to small players.
SEO for big, established companies, in terms of strategy, is simple. Make sure the site is crawlable. Run PR campaigns that frequently mention the name of the big, established company – which PR campaigns do anyway – and ensure those mentions include a back link. Talk to a lof of friendly reporters. Publish content, do so often, and make sure the important content is somewhere near the top.
That’s it.
The market reputation of the entity does most of the grunt work when it comes to ranking. So long as their ship is pointed in the right direction, they’re golden.
The main aim of the SEO who works for a big, established entity is to stop the big, established entity doing something stupid. So long as the SEO can prevent the entity doing stupid things – often a difficult task, granted – the big, established entity will likely dominate their niche simply by virtue of established market power.
That didn’t used to be the case.
When SEO started, and for a number of years after, the little guy could dominate niches by being the most relevant. The little guy could become the most relevant by carefully deconstructing the algorithm and giving the search engine what the search engine wanted. If the search engines weren’t careful, they were in very real danger of getting exactly what they asked for!
That temporary inversion of the traditional power structure made SEO a lot of fun. You did some clever stuff. You rose to the top. You collected the rewards. I think it’s grown less fun now because being clever isn’t enough. SEO works, but not quite as well as it used to for small players as the cost/reward equation favors big players.
Do A Lot Of Clever SEO Stuff, Get Nowhere
These days, a glass ceiling exists. SEOBook members can read a detailed post by Aaron outlining the glass ceiling here.
Here’s how it often plays out…
About a 8 months ago we launched one of the most viral pieces of content that we have ever done (particulary for a small site that doesn’t have a huge following) … it was done so well that it was organically referenced/hardcoded into Wikipedia. In addition it was cited on news sites, dozens and dozens of blogs (likely north of 100), a number of colleges, etc. It got like a couple hundred unique linking domains….which effectively doubled the unique linking domains that linked into the parent site. What impact did that have on rankings? Nada.
For link building to work well, the right signals need to exist, too. There needs to have high levels of reach and engagement. Big companies tend to have high levels of reach and engagement due to their market position and wider PR and advertising campaigns. This creates search keyword volume, keyword associations, engagement, and frequent mentions in important places, and all this is difficult to compete with if you have a small budget. The exception appears to be in relatively new niches, and in the regions, where the underlying data concerning engagement, reach and interest is unlikely to be particularly deep and rich.
Yet.
So, the little guy is often fighting a losing battle when it comes to search. Even if they choose a new, fast growing niche, as soon as that niche becomes lucrative enough to attract big players, traditional power will reassert itself. The only long term option for the little guy is to become a big guy, or get bought up by one, or go work for one.
Slavery
Abraham Lincoln thought wage labour was a stage workers pass through, typically in their 20’s or early 30s. Eventually, they become self-employed and keep all the profits of their labour.
Adam Smith maintained markets only work as intended if everyone had enough to participate. They also must have sufficient control over their own means of production. Adam Smith, father of modern capitalism, was not a big fan of corporate capitalism:
Merchants and master manufacturers are . . . the two classes of people who commonly employ the largest capitals, and who by their wealth draw to themselves the greatest share of the public consideration.
A side-effect of big players is they can distort markets. They have more purchasing power and that purchasing power sends a signal about what’s important. To big companies. The result is less diversity.
It’s self evident that power changes the search game. The search results become more about whoever is the most powerful. It seems ironic that Google started as an upstart outsider. The search results are difficult to conquer if you’re an upstart outsider, but pretty easy to do if you’re already a major player. Adwords, quality score being equal, favours those with deep pockets.
What’s happening is the little guy is getting squeezed out of this landscape and many of them will become slaves.
Huh? “Slaves”? By Aristotle’s and Lincoln’s definition, quite possibly:
If we want to have markets, we have to give everybody an equal chance to get into them, or else they don’t work as a means of social liberation; they operate as a means of enslavement.
Enslavement in the sense that the people with enough power, who can get the market to work on their behalf…
Right — bribing politicians to set up the system so that they accumulate more, and other people end up spending all their time working for them. The difference between selling yourself into slavery and renting yourself into slavery in the ancient world was basically none at all, you know. If Aristotle were here, he’d think most people in a country like England or America were slaves.
What’s happening in search is a microcosm of what is happening elsewhere in society. Markets are dominated and distorted by corporations at the direct expense of the small players. Yes, it’s nothing new, but it hasn’t always been this way in search.
So what is my point?
My point is that if you’re not getting the same business benefits from search as you used to, and the game seems that much harder, then it’s not because you’re not clever. It’s because the game is rigged.
Of course, small companies can prosper. You’ll find many examples of them in the SERPs. But their path to getting there via the search channel is now much longer and doesn’t pay as well as it used to. This means fewer SEOs will be hired by small companies because the cost of effective SEO is rising fast whilst the rewards are shrinking. Meanwhile, the big companies are increasing their digital budgets.
Knowing all this, the small operator can change their approach. The small operator has one advantage. They can be nimble, flexible, and change direction quickly. So, looking forward over the not-too-distant horizon, we either need a plan to take advantage of fast emerging markets before the big guys enter them, or we need a plan to scale, or we need to fight differently, such as taking brand/USP centric approaches.
Or go work for one of the big guys.
As a “slave”. Dude :) (Just kidding)
What’s In A Name?
Many SEO love keyword-loaded domain names. The theory is that domains that feature a keyword will result in a boost in ranking. It’s still a contentious topic:
I’ve seen bloggers, webmasters and search aficionados argue the case around the death of EMDs time and time again, despite the evidence staring them in the face: EMDs are still all over the place. What’s more, do a simple bulk backlink analysis via Majestic, and you will find tons which rank in the top 10 while surrounded by far more authoritative domains.
No matter what the truth of the matter as to the ranking value of EMDs, most would agree that finding the right language for describing and profiling our business is important.
Terminology Changes
Consider the term “startup”.
This term, which describes a new small business, feels like it has been around forever. Not so. Conduct a search on the time period 1995-1998 and you won’t find results for start-up:
It’s a word that has grown up with the web and sounds sexier than just business. Just like the word “consultant” or “boutique” sounds better than “mom and pop” or “1 person business”. (You must remember of course when “sanitation engineer” replaced “trash man”.) oI just did a search to see the use of the word startup from the period 1995 to 1998 and came up with zilch in terms of relation to business
Start up does sound sexier than “mom n pop” or “one person business”, or “a few stoner mates avoiding getting a job”. A pitch to a VC that described the business as a “mom n pop” may not be taken seriously, whereas calling it a startup will.
If we want to be taken seriously by our audience, then finding the audience’s language is important.
SEO or Digital Marketing Or…..?
Has SEO become a dirty word? Has it always been a dirty word?
SEO’s don’t tend to see it that way, even if they are aware of the negative connotations. They see SEO as a description of what they do. It’s always been a bit of a misnomer, as we don’t optimize search engines, but for whatever reason, it stuck.
The term SEO is often associated with spam. The ever-amiable Matt Cutts video’s could be accompanied by a stern, animated wagging finger and a “tut tut tut” subtext. The search engines frown on a lot when it comes to SEO. SEO is permanent frown territory. Contrast this with PPC. PPC does not have that negative connotation. There is no reputation issue in saying you’re a PPC provider.
Over the years, this propaganda exercise that has resulted in the “SEO questionable/PPC credible” narrative has been pretty effective. The spammer label, borrowed from the world of email spam, has not been a term the SEO has managed to shrug off. The search engines have even managed to get SEOs to use the term “spammer” as a point of differentiation. “Spam is what the other SEOs do. Not me, of course.” This just goes to show how effective the propaganda has been. Once SEOs used spam to describe their own industry, the fate of the term SEO was sealed. After all, you seldom hear doctors, lawyers and retailers defining what they do against the bad actors in their sector.
As traffic acquisition gets broader, encompassing PR and social media, new titles like Digital Marketer have emerged. These terms have the advantage of not being weighed down by historical baggage. I’m not suggesting people should name themselves one thing or the other. Rather, consider these terms in a strategic sense. What terms best describe who you are and what you do, and cast you in the best possible light to those you wish to serve, at this point in time?
The language moves.
Generic Name Or Brandable?
Keyword loaded names, like business.com, are both valuable and costly. The downside of such names, besides being costly, is they severely limit branding opportunities. The better search engines get, and the more people use social media and other referral channels, the less these generic names will matter.
What matters most in crowded markets is being memorable.
A memorable, unique name is a valuable search commodity. If that name is always associated with you and no one else, then you’ll always be found in the search results. SEMRush, MajesticSEO, and Mo are unlikely to be confused with other companies. “Search Engine Tools”, not so much.
Will the generic name become less valuable because generic names are perhaps only useful at the start of an industry? How mature is your industry? How can you best get differentiation in a crowded market through language alone?
The Strategy Behind Naming
Here are a few points to consider.
1. Start Early
Names are often an afterthought. People construct business plans. They think about how their website looks. They think about their target market. They don’t yet have a name. Try starting with a name and designing everything else around it. The name can set the tone of every other decision you make.
2. Positioning
In mature markets, differentiation is strategically important. Is your proposed name similar to other competitors names? Is it unique enough? If you’re in at the start of a new industry, would a generic, keyword loaded name work best? Is it time for a name change because you’ve got lost in the crowd? Has your business focus changed?
Does your name go beyond mere description and create an emotional connection with your audience? Names that take on their own meaning, like Amazon, are more likely to grow with the business, rather than have the business outgrow the name. Imagine if Amazon.com had called itself Books.com.
3. What Are You All About?
Are you a high-touch consultative company? Or a product based, functional company? Are you on the cutting edge? Or are you catering to a market who like things just the way they are?
Writing down a short paragraph about how you see yourself, how the customers see you, and your position in the market, will help you come up with suitable names. Better yet, write a story.
4. Descriptive Vs Differentiation
Descriptive can be safe. “Internet Search Engine” or “Web Crawler”. There’s no confusing what those businesses do. Compare them with the name Google. Google gives you no idea what the company does, but it’s more iconic, quirky and memorable. There’s no doubt it has grown with the company and become a natural part of their identity in ways that “Internet Search Engine” never could.
Sometimes, mixing descriptions to create something quirky works well. Airbnb is a good example. The juxtaposition of those two words creates something new, whilst at the same time having a ring of the familiar. It’s also nice to know if the domain name is available, and if the name can be trademarked. The more generic the name, the harder it is to trademark, and the less likely the domain name is available.
5. Does Your Name Travel Well?
Hopefully, your name isn’t a swear word in another culture. Nor have negative connotations. Here are a few comical examples where it went wrong:
Nokia’s new smartphone translates in Spanish slang to prostitute, which is unfortunate, but at least the cell phone giant is in good company. The name of international car manufacturer Peugeot translates in southern China to Biao zhi, which means the same thing.
This is not such an issue if your market is local, but if you plan to expand into other markets in future, then it pays to consider this angle.
6. There’s No Right Answer
There is probably no universally good name. At least, when you first come up with a name, you can be assured some people will hate it, some will be indifferent, and some will like it – no matter what name you choose.
This is why it’s important to ground the subjective name-choosing process in something concrete, like your business strategy, or positioning in the market. You name could have come before the business plan. Or it could reflect it. You then test your name with people who will likely buy your product or service. It doesn’t matter what your Mom or your friends think of the name, it’s what you think of the name and what your potential customers think of the name that counts.
7. Diluting Your Name
Does each service line and product in your company need a distinctive name? Maybe, but the risk is that it could dilute the brand. Consider Virgin. They put the exact same name on completely different service lines. That same brand name carries the values and spirit of Virgin to whatever new enterprise they undertake. This also reduces the potential for customer confusion.
Creating a different name for some of your offerings might be a good idea, Say, if you’re predominantly a service-based company, yet you also have one product that you may spin off at some point in future. You may want to clearly differentiate the product from the service so as not to dilute the focus of the service side. Again, this is where strategy comes in. If you’re clear about what your company does, and your position in the market, then it becomes easier to decide how to name new aspects of your business. Or whether you should give them a name at all.
7. Is your name still relevant?
Brands evolve. They can appear outdated if the market moves on. On the other hand, they can built equity through longevity. It seems especially difficult to change internet company names as the inbound linking might be compromised as a result. Transferring the equity of a brand is typically expensive and difficult. All the more reason to place sufficient importance on naming to begin with.
8. More Than A Name
The branding process is more than just a name and identity. It’s the language of your company. It’s the language of your customers. It becomes a keyword on which people search. Your customers have got to remember it. You, and your employees, need to be proud of it. It sets you apart.
The language is important. And strategic.
Pandas, Penguins, and Popsicles
Are you still working through your newsfeed of SEO material on the 101 ways to get out of panda 4.0 written by people that have never actually practiced SEO on their own sites? Aaron and I had concluded that what was rolling through was panda before it was announced that it was panda, but I’m not going to walk here on my treadmill and knock out yet another post on the things you should be doing if you were gut punched by that negative a priori algorithm (hat tip to Terry, another fine SEObook member, for pointing out to me those public discussions that showed the philosophical evolutionary shift towards the default assumption that sites likely deserve to be punished). I’d say 90% of those posts are thinly veiled sales pitches; I should know since I sell infographics to support my nachos habit. Speaking of infographics, there’s already a great one that covers recovery strategies that still work right here.
Should I write about penguin? Analysis of that beast consumed the better part of 2 year years of my waking time. Nope. Again, I think it has already been adequately covered in a previous blog post. There’s nothing particularly new to report there either since the next update may be completely different, might be just another refresh that doesn’t take into account those slapped in the 1.0 incarnation of the update, or may actually be the penguin everyone hopes it is, taking into account the countless hours agencies have spent disavowing links and spamming me with fake legal threads should I not remove links they themselves placed. I wouldn’t hold your breathe on that last one. Outside of crowdsourcing pain for future manual penalties, I don’t expect much relief on that front.
Instead, I think I’m going to talk about popsicles. That seems like the kind of tripe that a SEO blog might discuss. I bet I can make it work though. I’m a fat dude in the Phoenix area and we already had our first 100F day, so I’m thinking of frozen treats. Strap in.
Search tactics and I’d even go so far as to say even certain strategies are like popsicles. When they are brand new they are cool and refreshing, but once exposed to the public heat they fade…fast. Really fast. Like a goop of sticky mess, which users of ALN and BMR can probably tell you.
Bear with me.
If you have a tactic that works, why would you expose it to the public? Nothing good can come of that. Sure, you have a tactic that works 100% but since I’m a loyal subscriber you’re willing to share it with me for $297. Seems legit. I’m not saying all services/products pitched this way are inherently ‘bad’, I’m just saying you aren’t going to get a magic bullet, yet alone one hand-wrapped and delivered by filling out a single wufoo form…sans report.
Would you share with a really close friend? I suppose, but even still the popsicle isn’t going to last as long since it is now being consumed at an accelerated rate. There’s the thought of germs, contamination, and other nasty thoughts that’d prevent me from going down that route. Cue the “Two SEOs, one popsicle” reaction videos. No. There are two ways to make the best use out of that popsicle.
- Practionioner: eat it quietly, savor it, make it last.
- Strategist w/ resources: figure out the recipe and mass produce it as quickly as possible, knowing that after enough public heat is on, the popsicles will start melting before they can be eaten, and no one likely that weird, warm orange sticky stuff that tastes like a glucose intolerance test.
There’s another caveat to the two above scenarios. Even if you’re a strategist with deep resources, unless you’re willing to test on your own sites, you’re just effectively selling smoke on an unproven tactic.
So there you have it, tactics are like popsicles. Disappointed? Good. I’ve been doing SEO since 1997, so here’s a secret: try to create engaging content, supported by authoritative off-page signals. There’s an ebb and flow to this of course, but it can be translated across the full black/white spectrum. Markov content in a free wordpress theme can be engaging when it is cloaked with actionable imagery, with certain % of back-buttons disabled, or when you make the advertising more compelling than the content (just ask eHow). Similarly, well-researched interactive infographics can engage the user on the other side of the spectrum…just more expensive. Comment spam and parasitic hosting on “authority” sites can tap into those authority signals on dark side, as can a thorough native campaign across a bunch of relevant sites backed by a PR campaign, TV commercials, and radio spots for the light side. Budget and objectives are the only difference.
Go enjoy a popsicle everyone. Summer is here; I expect a lot more heat from Google, so you might need one.
About the author: Joe Sinkwitz is the Chief Revenue Officer at CopyPress. He {Tweets / posts / comments / shares his thoughts} on navigating the evolving SEO landscape on Twitter here.
Please Remove My Link. Or Else.
Getting links removed is a tedious business.
It’s just as tedious for the site owner who must remove the links. Google’s annoying practice of “suggesting” webmasters jump through hoops in order to physically remove links that the webmaster suspects are bad, rather than Google simply ignoring the links that they’ve internally flagged, is causing frustration.
Is it a punitive punishment? If so, it’s doing nothing to endear Google to webmasters. Is it a smokescreen? i.e. they don’t know which links are bad, but by having webmasters declare them, this helps Google build up a more comprehensive database? Bit of both? It might also be adding costs to SEO in order to put SEO out of reach of small companies. Perhaps it’s a red herring to make people think links are more important than they actually are.
Hard to be sure.
Collateral Damage
SEOs are accustomed to search engines being coy, punitive and oblique. SEOs accept it as part of the game. However, it becomes rather interesting when webmasters who are not connected to SEO get caught up in the collateral damage:
I received an interesting email the other day from a company we linked to from one of our websites. In short, the email was a request to remove links from our site to their site. We linked to this company on our own accord, with no prior solicitation, because we felt it would be useful to our site visitors, which is generally why people link to things on the Internet.
And check out the subsequent discussion on Hacker News. Matt Cutts first post is somewhat disingenuous:
Situation #1 is by far the most common. If a site gets dinged for linkspam and works to clean up their links, a lot of them send out a bunch of link removal requests on their own prerogative
Webmasters who receive the notification are encouraged by Google to clean up their backlinks, because if they don’t, then their rankings suffer.
But, essentially from our point of view when it comes to unnatural links to your website we want to see that you’ve taken significant steps to actually remove it from the web but if there are some links that you can’t remove yourself or there are some that require payment to be removed then having those in the disavow file is fine as well.
(Emphasis mine)
So, of course webmasters who have received a notification from Google are going to contact websites to get links removed. Google have stated they want to see that the webmaster has gone to considerable effort to remove them, rather than simply use the disavow tool.
The inevitable result is that a webmaster who links to anyone who has received a bad links notification may receive the latest form of email spam known as the “please remove my link” email. For some webmasters, this email has become more common that the “someone has left you millions in a Nigerian bank account” gambit, and is just as persistent and annoying.
From The Webmasters Perspective
Webmasters could justifiably add the phrase “please remove my link” and the word “disavow” to their spam filters.
Let’s assume this webmaster isn’t a bad neighbourhood and is simply caught in the cross-fire. The SEO assumes, perhaps incorrectly, the link is bad and requests a take-down. From the webmasters perspective, they incur a time cost dealing with the link removal requests. A lone request might take a few minutes to physically remove – but hang on a minute – how does the webmaster know this request is coming from the site owner and not from some dishonest competitor? Ownership takes time to verify. And why would the webmaster want to take down this link, anyway? Presumably, they put it up because they deemed it useful to their audience. Or, perhaps some bot put the link there – perhaps as a forum or blog comment link – against the webmasters wishes – and now, to add insult to injury, the SEO wants the webmaster to spend his time taking it down!
Even so, this might be okay if it’s only one link. It doesn’t take long to remove. But, for webmasters who own large sites, it quickly becomes a chore. For large sites with thousands of outbound links built up over years, removal requests can pile up. That’s when the spam filter kicks in.
Then come the veiled threats. “Thanks for linking to us. This is no reflection on you, but if you don’t remove my link I’ll be forced to disavow you and your site will look bad in Google. I don’t want to do this, but I may have to.”
What a guy.
How does the webmaster know the SEO won’t do that anyway? Isn’t that exactly what some SEO conference speakers have been telling other SEOs to do regardless of whether the webmaster takes the link down or not?
So, for a webmaster caught in the cross-fire, there’s not much incentive to remove links, especially if s/he’s read Matt’s suggestion:
higherpurpose, nowhere in the original article did it say that Google said the link was bad. This was a request from a random site (we don’t know which one, since the post dropped that detail), and the op can certainly ignore the link removal request.
In some cases Google does specify links:
We’ve reviewed the links to your site and we still believe that some of them are outside our quality guidelines.
Sample URLs:
ask.metafilter.com/194610/get-me-and-my-stuff-from-point-a-to-point-b-possibly-via-point-cPlease correct or remove all inorganic links, not limited to the samples provided above. This may involve contacting webmasters of the sites with the inorganic links on them.
And they make errors when they specify those links. They’ve flagged DMOZ & other similar links: “Every time I investigate these “unnatural link” claims, I find a comment by a longtime member of MetaFilter in good standing trying to help someone out, usually trying to identify something on Ask MetaFilter.”
Changing Behaviour
Then the webmaster starts thinking.
“Hmmm…maybe linking out will hurt me! Google might penalize me or, even worse, I’ll get flooded with more and more “please remove my link” spam in future.”
So what happens?
The webmaster becomes very wary about linking out. David Naylor mentioned an increasing number of sites adopting a “no linking” policy. Perhaps the webmaster no-follows everything as a precaution. Far from being the life-giving veins of the web, links are seen as potentially malignant. If all outbound links are made no-follow, perhaps the chance of being banned and flooded with “please remove my link”spam is reduced. Then again, even nofollowed links are getting removal requests.
As more webmasters start to see links as problematic, fewer legitimate sites receive links. Meanwhile, the blackhat, who sees their sites occasionally getting burned as a cost of doing business, will likely see their site rise as they’ll be the sites getting all the links, served up from their curated link networks.
A commenter notes:
The Google webspam team seems to prefer psychology over technology to solve the problem, especially recently. Nearly everything that’s come out of Matt Cutt’s mouth in the last 18 months or so has been a scare tactic.
IMO all this does is further encourage the development of “churn and burn” websites from blackhats who have being penalized in their business plan. So why should I risk all the time and effort it takes to generate quality web content when it could all come crashing down because an imperfect and overzealous algorithm thinks it’s spam? Or worse, some intern or non-google employee doing a manual review wrongly decides the site violates webmaster guidelines?
And what’s the point of providing great content when some competitor can just take you out with a dedicated negative SEO campaign, or if Google hits you with a false positive? If most of your traffic comes from Google, then the risk of the web publishing model increases.
Is Google broken? Or is your site broken? That’s the question any webmaster asks when she sees her Google click-throughs drop dramatically. It’s a question that Matt Haughey, founder of legendary Internet forum MetaFilter, has been asking himself for the last year and a half, as declining ad revenues have forced the long-running site to lay off several of its staff.
Then again, Google may just not want what MetaFilter has to offer anymore.
(In)Unintended Consequences
Could this be uncompetitive practice from Google? Are the sites getting hit with penalties predominantly commercial sites? It would be interesting to see how many of them are non-commercial. If so, is it a way to encourage commercial sites to use Adwords as it becomes harder and harder to get a link by organic means? If all it did was raise the cost of doing SEO, it would still be doing its job.
I have no idea, but you could see why people might ask that question.
Let’s say it’s benevolent and Google is simply working towards better results. The unintended consequence is that webmasters will think twice about linking out. And if that happens, then their linking behaviour will start to become more exclusive. When links become harder to get and become more problematic, then PPC and social-media is going to look that much more attractive.
Google Helpouts Twitter Spam (Beta)
Google is desperate to promote Helpouts. I first realized this when I saw the following spam message in my email inbox.
Shortly after a friend sent me a screenshot of a onebox promoting Helpouts in the SERPs.
That’s Google monopoly and those are Google’s services. It is not like they are:
- being anti-competitive
- paying others to spam other websites
Let’s slow down though. Maybe I am getting ahead of myself:
Google has its own remote technology support service similar to Mr. Gupta’s called Google Helpouts. Mr. Gupta’s complaint alleges Google may have been blocking his advertisements so Google Helpouts could get more customers.
Oh, and that first message looked like it could have been an affiliate link. Was it?
Hmm
Let me see
What do we have here?
Google Helpouts connects you to a variety of experts–from doctors, parenting experts, tutors, personal trainers, and more–over live video call. The Google Helpouts Ambassador Program is a unique opportunity to spread the word about Helpouts, earn money, and influence a new Google product–all on your own schedule.
As an Ambassador, you will:
- Earn extra income–receive $25 for each friend you refer who takes their first paid Helpout, up to $1,000 per month for the first 4 months.
- Give direct feedback and help shape a new Google product
- Join a community of innovative Ambassadors around the country
- Receive a Helpouts gift and the chance to win prizes
We all know HELPFUL hotel affiliate websites are spam, but maybe Google HELPouts affiliate marketing isn’t spam.
After all, Google did promise to teach people how to do their affiliate marketing professionally: “We will provide you with an Ambassador Toolkit with tips and suggestions on creative ways you can spread the word. You are encouraged to get creative, be innovative, and utilize different networks (i.e. social media, word of mouth, groups & associations, blogs, etc.) to help you.”
Of course the best way to lead is by example.
And lead they do.
They are highly inclusive in their approach.
@homosexualwentz Help is here! Use the code IFOUNDHELP for $20 off a 1-on-1 session w/ a trusted tech expert today: http://t.co/OtiePzjOIS— Helpouts by Google (@Helpouts) May 12, 2014
Check out this awesome Twitter usage
They’ve more Tweets in the last few months than I’ve made in 7 years. There are 1,440 minutes in a day, so it is quite an achievement to make over 800 Tweets in a day.
@Helpouts did google helpouts really just reply to a tweet of mine?— Emma Klinger (@Emma_Klinger) May 15, 2014
You and many many many many thousands of others, Emma.
Some minutes they are making 2 or 3 Tweets.
And with that sort of engagement & the Google brand name, surely they have built a strong following.
Uh, nope.
They are following over 500 people and have about 4,000 followers. And the 4,000 number is generous, as some of them are people who sell on that platform or are affiliates pushing it.
Let’s take a look at the zero moment of truth:
@Helpouts I really don’t want to pay for support.— Krishna M. Sadasivam (@pcweenies) May 14, 2014
Thanks for your unsolicited commercial message, but I am not interested.
@baileyboo612 @Helpouts what the heck??— logan™ (@logangaspard) May 13, 2014
You’re confusing me. Some context would help.
@Helpouts Hi, who can help me? , do u have an email for talk there?— Pamela López (@karlismkt) May 15, 2014
No email support, but support “sessions”? What is this?
@Helpouts $1.00 a minute? No thanks…— (@Owennn_Marsh) May 14, 2014
Oh, I get it now. Is this a spam bot promoting phone sex?
RT “@Helpouts: Help is just a click away! Helpouts by #Google has Apple tech experts who can help out today.” … No, just no— Johnny Blanchard (@JonnBlanchard) May 15, 2014
Ah, so it isn’t phone sex, but you can help with iPhones. Um, did we forget that whole Steve Jobs thermonuclear war bit? And why is Google offering support for Apple products when Larry Page stated the whole idea of customer support was ridiculous?
@Helpouts MAN I’VE ALREADY TRIED GOOGLE— mya ☪ (@gleebegay) May 15, 2014
OK, so maybe this is more of the same.
@Helpouts hahaha I’m not going to pay money for things that probably won’t even help me.— Nate Morse (@nt4343) May 16, 2014
Cynical, aren’t we?
@Helpouts no, that’s too expensive— Mahalah (@Mahalah_faye) May 16, 2014
And cheap?
@Helpouts im not paying— ononnah❄ (@ononnahh) May 15, 2014
Really cheap. :(
@Helpouts fu I’m not paying for help— Sam (@ttihweimmas) May 14, 2014
And angry?
@Helpouts Chill out!— Atsuro Kihara (@atsuro_enbot) May 15, 2014
And testy?
@Helpouts then why the fuck arent you fixing my mentions if youre such “trusted tech experts”— bianca // KIAN 4/6?? (@lawlorfflower) May 13, 2014
And rude?
@Helpouts no— Abby Williams (@abbywilliams96) May 15, 2014
And curt?
@Helpouts no— spongebobs weave (@googlingcraic) May 15, 2014
Didn’t you already say that???
@Helpouts fuck you— Samuel (@SamuelJones69) May 14, 2014
Didn’t you already say that???
@Helpouts what the heck— hey cal (@holycalum) May 14, 2014
It seems we are having issues communicating here.
@Helpouts ok I fucks with it. Next time try not to spy on my tweets lol— E (@Easy_Eli) May 14, 2014
I’m not sure it is fair to call it spying a half day late.
That was 11 hours ago @Helpouts pic.twitter.com/jRVV7rgtdW— 28 (@5SOSmaryam) May 14, 2014
Better late than never.
@Helpouts I fixed it already maybe? Thank you tho automatic message— Megar (@AH_Megan6) May 15, 2014
Even if automated.
Good catch Megar, as Google has a creepy patent on automating social spam.
@Helpouts Sorry, but I’ll just ask my friends because they are free and I need to do the homework right now.— Emma Raye Mosier (@EmmaRayeMosier) May 15, 2014
Who are your real Google+ friends? Have they all got the bends? Is Google really sinking this low?
@Helpouts Classmates over corporations, Google. I’m not interested.
BTW please don’t censor the European internet.— Hash (@saikorhythm) May 15, 2014
Every journey of a thousand miles begins with a single step.
@Helpouts This is hilarious. If I ever need it I know you guys have this, thanks!— Gabriella (@Gaby441) May 14, 2014
Humorous or sad…depending on your view.
@Helpouts thanks! is there also a Helpout on how to shameless self promote yourself through twitter? just wondering— Etta Grover (@ettafetacheese) May 14, 2014
There’s no wrong way to eat a Reese’s.
Google has THOUSANDS of opportunities available for you to learn how to spam Twitter.
As @Helpouts repeatedly Tweets: “Use the code IFOUNDHELP for $20 off” :D
++++++++
All the above Tweets were from the last few days.
The same sort of anti-social agro spamming campaign has been going on far longer.
@Helpouts fuck off— Dagger Anderson (@daggeranderson) May 8, 2014
When Twitter users said “no thank you”…
@Helpouts THIS SHIT AINT FREE— mariam / 20 (@psychoticamila) May 8, 2014
…Google quickly responded like a Marmaris rug salesman
@Helpouts leave me a lone— mariam / 20 (@psychoticamila) May 9, 2014
Google has a magic chemistry for being able to…
We need to fight spam messages (with MOAR spam messages).
@shelleyamybeth Lets try to get rid of all those Spam messages. Check out #Helpouts by Google for Wordpress help. http://t.co/eTTOMnpIhF— Helpouts by Google (@Helpouts) March 25, 2014
In a recent Youtube video Matt Cutts said: “We got less spam and so it looks like people don’t like the new algorithms as much.” Based on that, perhaps we can presume Helpouts is engaging in a guerrilla marketing campaign to improve user satisfaction with the algorithms.
Or maybe Google is spamming Twitter so they can justify banning Twitter.
Or maybe this is Google’s example of how we should market websites which don’t have the luxury of hard-coding at the top of the search results.
Or maybe Google wasn’t responsible for any of this & once again it was “a contractor.”
What’s Wrong With A/B Testing
A/B testing is an internet marketing standard. In order to optimize response rates, you compare one page against another. You run with the page that gives you the best response rates.
But anyone who has tried A/B testing will know that whilst it sounds simple in concept, it can be problematic in execution. For example, it can be difficult to determine if what you’re seeing is a tangible difference in customer behaviour or simply a result of chance. Is A/B testing an appropriate choice in all cases? Or is it best suited to specific applications? Does A/B testing obscure what customers really want?
In this article, we’ll look at some of the gotchas for those new to A/B testing.
1. Insufficient Sample Size
You set up test. You’ve got one page featuring call to action A and one page featuring call to action B. You enable your PPC campaign and leave it running for a day.
When you stop the test, you’ve found call-to-action A converted at twice the rate of call-to-action B. So call-to-action A is the winner and we should run with it, and eliminate option B.
But this would be a mistake.
The sample size may be insufficient. If we only tested one hundred clicks, we might get a significant difference in results between two pages, but that change doesn’t show up when we get to 1,000 clicks. In fact, the result may even be reversed!
So, how do we determine a sample size that is statistically significant? This excellent article explains the maths. However, there are various online sample size calculators that will do the calculations for you, including Evan’s. Most A/B tracking tools will include sample size calculators, but it’s a good idea to understand what they’re calculating, and how, to ensure the accuracy of your tests.
In short, make sure you’ve tested enough of the audience to determine a trend.
2. Collateral Damage
We might want to test a call to action metric. We want to test the number of people who click on the “find out more” link on a landing page. We find that a lot more people click on this link we use the term “find out more” than if we use the term “buy now”.
Great, right?
But what if the conversion rate for those who actually make a purchase falls as a result? We achieved higher click-thrus on one landing page at the expense of actual sales.
This is why it’s important to be clear about the end goal when designing and executing tests. Also, ensure we look at the process as a whole, especially when we’re chopping the process up into bits for testing purposes. Does a change in one place affect something else further down the line?
In this example, you might A/B test the landing page whilst keeping an eye on your total customer numbers deeming the change effective only if customer numbers also rise. If your aim was only to increase click-thru, say to boost quality scores, then the change was effective.
3. What, Not Why
In the example above, we know the “what”. We changed the wording of a call-to-action link, and we achieved higher click thru’s, although we’re still in the dark as to why. We’re also in the dark as to why the change of wording resulted in fewer sales.
Was it because we attracted more people who were information seekers? Were buyers confused about the nature of the site? Did visitors think they couldn’t buy from us? Were they price shoppers who wanted to compare price information up front?
We don’t really know.
But that’s good, so long as we keep asking questions. These types of questions lead to more ideas for A/B tests. By turning testing into an ongoing process, supported by asking more and hopefully better questions, we’re more likely to discover a whole range of “why’s”.
4. Small Might Be A Problem
If you’re a small company competing directly with big companies, you may already be on the back foot when it comes to A/B testing.
It’s clear that its very modularity can cause problems. But what about in cases where the number of tests that can be run at once is low? While A/B testing makes sense on big websites where you can run hundreds of tests per day and have hundreds of thousands of hits, only a few offers can be tested at one time in cases like direct mail. The variance that these tests reveal is often so low that any meaningful statistical analysis is impossible.
Put simply, you might not have the traffic to generate statistically significant results. There’s no easy way around this problem, but the answer may lay in getting tricky with the maths.
Experimental design massively and deliberately increases the amount of variance in direct marketing campaigns. It lets marketers project the impact of many variables by testing just a few of them. Mathematical formulas use a subset of combinations of variables to represent the complexity of all the original variables. That allows the marketing organization to more quickly adjust messages and offers and, based on the responses, to improve marketing effectiveness and the company’s overall economics
Another thing to consider is that if you’re certain the bigger company is running A/B tests, and achieving good results, then “steal” their landing page*. Take their ideas for landing pages and use that as a test against your existing pages. *Of course, you can’t really steal their landing page, but you can be “influenced by” their approach.
What your competitors do is often a good starting point for your own tests. Try taking their approach and refine it.
5. Might There Be A Better Way?
Are there alternatives to A/B testing?
Some swear by the Multi Armed Bandit methodology:
The multi-armed bandit problem takes its terminology from a casino. You are faced with a wall of slot machines, each with its own lever. You suspect that some slot machines pay out more frequently than others. How can you learn which machine is the best, and get the most coins in the fewest trials?
Like many techniques in machine learning, the simplest strategy is hard to beat. More complicated techniques are worth considering, but they may eke out only a few hundredths of a percentage point of performance.
Then again…..
What multi-armed bandit algorithm does is that it aggressively (and greedily) optimizes for currently best performing variation, so the actual worse performing versions end up receiving very little traffic (mostly in the explorative 10% phase). This little traffic means when you try to calculate statistical significance, there’s still a lot of uncertainty whether the variation is “really” worse performing or the current worse performance is due to random chance. So, in a multi-armed bandit algorithm, it takes a lot more traffic to declare statistical significance as compared to simple randomization of A/B testing. (But, of course, in a multi-armed bandit campaign, the average conversion rate is higher).
Multivariate testing may be suitable if you’re testing a combination of variables, as opposed to just one i.e.
- Product Image: Big vs. Medium vs Small
- Price Text Style: Bold vs Normal
- Price Text Color: Blue vs. Black vs. Red
There would be 3x2x3 different versions to test.
The problem with multivariate tests is they can get complicated pretty quickly and require a lot of traffic to produce statistically significant results. One advantage of multivariate testing over A/B testing is that it can tell you which part of the page is most influential. Was it a graphic? A headline? A video? If you’re testing a page using an A/B test, you won’t know. Multivariate testing will tell you which page sections influence the conversion rate and which don’t.
6. Methodology Is Only One Part Of The Puzzle
So is A/B testing worthwhile? Are the alternatives better?
The methodology we choose will only be as good as the test design. If tests are poorly designed, then the maths, the tests, the data and the software tools won’t be much use.
To construct good tests, you should first take a high level view:
Start the test by first asking yourself a question. Something on the lines of, “Why is the engagement rate of my site lower than that of the competitors…..Collect information about your product from customers before setting up any big test. If you plan to test your tagline, run a quick survey among your customers asking how they would define your product.
Secondly, consider the limits of testing. Testing can be a bit of a heartless exercise. It’s cold. We can’t really test how memorable and how liked one design is over the other, and typically have to go by instinct on some questions. Sometimes, certain designs just work for our audience, and other designs don’t. How do we test if we’re winning not just business, but also hearts and minds?
Does it mean we really understand our customers if they click this version over that one? We might see how they react to an offer, but that doesn’t mean we understand their desires and needs. If we’re getting click-backs most of the time, then it’s pretty clear we don’t understand the visitors. Changing a graphic here, and wording there, isn’t going to help if the underlying offer is not what potential customers want. No amount of testing ad copy will sell a pink train.
The understanding of customers is gained in part by tests, and in part by direct experience with customers and the market we’re in. Understanding comes from empathy. From asking questions. From listening to, and understanding, the answers. From knowing what’s good, and bad, about your competitors. From providing options. From open communication channels. From reassuring people. You’re probably armed with this information already, and that information is highly useful when it comes to constructing effective tests.
Do you really need A/B testing? Used well, it can markedly improve and hone offers. It isn’t a magic bullet. Understanding your audience is the most important thing. Google, a company that uses testing extensively, seem to be most vulnerable when it comes to areas that require a more intuitive understanding of people. Google Glass is a prime example of failing to understand social context. Apple, on the other hand, were driven more by an intuitive approach. Jobs: “We built [the Mac] for ourselves. We were the group of people who were going to judge whether it was great or not. We weren’t going to go out and do market research”
A/B testing is can work wonders, just so long as it isn’t used as a substitute for understanding people.
Learn Local Search Marketing
Last October Vendran Tomic wrote a guide for local SEO which has since become one of the more popular pages on our site, so we decided to follow up with a QnA on some of the latest changes in local search.
Q: Google appears to have settled their monop…