The Truth About Subjective Truths
A few months ago there was an article in New Scientist about Google’s research paper on potentially ranking sites based on how factual their content is. The idea is generally and genuinely absurd.
- You can’t copyright facts, which means that if this were a primary ranking signal & people focused on it then they would be optimizing their site to be scraped-n-displaced into the knowledge graph. Some people may sugar coat the knowledge graph and rich answers as opportunity, but it is Google outsourcing the cost of editorial labor while reaping the rewards.
.@mattcutts I think I have spotted one, Matt. Note the similarities in the content text: pic.twitter.com/uHux3rK57f— dan barker (@danbarker) February 27, 2014
- If Google is going to scrape, displace & monetize data sets, then the only ways to really profit are:
- focus on creating the types of content which can’t be easily scraped-n-displaced, or
- create proprietary metrics of your own, such that if they scrape them (and don’t cheat by hiding the source) they are marketing you
- In some areas (especially religion and politics) certain facts are verboten & people prefer things which provide confirmation bias of their pre-existing beliefs. End user usage data creates a “relevancy” signal out of comfortable false facts and personalization reinforces it.
- In some areas well known “facts” are sponsored falsehoods. In other areas some things slip through the cracks.
- In some areas Google changes what is considered fact based on where you are located.
How Google keeps everyone happy pic.twitter.com/KmBBzpfzdf— Amazing Maps (@Amazing_Maps) March 31, 2015
- Those who have enough money can create their own facts. It might be painting the perception of a landscape, hiring thousands of low waged workers to manipulate public perception on key issues and new technologies, or more sophisticated forms of social network analysis and manipulation to manipulate public perceptions.
- The previously mentioned links were governmental efforts. However such strategies are more common in the commercial market. Consider how Google has sponsored academic conferences while explicitly telling the people who put them on to hide the sponsorship as part of their lobbying efforts.
- Then there is the blurry area where government and commerce fuse, like when Google put about a half-dozen home team players in key governmental positions during the FTC investigation of Google. Google claimed lobbying was disgusting until they experienced the ROI firsthand.
- In some areas “facts” are backward looking views of the market which are framed, distorted & intentionally incomplete. There was a significant gap between internal voices and external messaging in the run up to the recent financial crisis. Even large & generally trustworthy organizations have some serious skeletons in their closets.
@mattcutts I wonder, what sort of impact does http://t.co/vdg3ARGSz2 have on their E-A-T? expertise +1, authority +1, trustworthiness -_?— aaron wall (@aaronwall) April 6, 2015
- In other areas the inconvenient facts get washed away over time by money.
For a search engine to be driven primarily by group think (see unity100’s posts here) is the death of diversity.
Less Diversity, More Consolidation
The problem is rarely attributed to Google, but as ecosystem diversity has declined (and entire segments of the ecosystem are unprofitable to service), more people are writing things like: “The market for helping small businesses maintain a home online isn’t one with growing profits – or, for the most part, any profits. It’s one that’s heading for a bloody period of consolidation.”
As companies grow in power the power gets monetized. If you can manipulate people without appearing to do so you can make a lot of money.
If you don’t think Google wants to disrupt you out of a job, you’ve been asleep at the wheel for the past decade— Michael Gray (@graywolf) March 13, 2015
We Just Listen to the Data (Ish)
As Google sucks up more data, aggregates intent, and scrapes-n-displaces the ecosystem they get air cover for some of their gray area behaviors by claiming things are driven by the data & putting the user first.
Those “data” and altruism claims from Google recently fell flat on their face when the Wall Street Journal published a number of articles about a leaked FTC document.
That PDF has all sorts of goodies in it about things like blocking competition, signing a low margin deal with AOL to keep monopoly marketshare (while also noting the general philosophy outside of a few key deals was to squeeze down on partners), scraping content and ratings from competing sites, Google force inserting itself in certain verticals anytime select competitors ranked in the organic result set, etc.
As damning as the above evidence is, more will soon be brought to light as the EU ramps up their formal statement of objection, as Google is less politically connected in Europe than they are in the United States:
“On Nov. 6, 2012, the night of Mr. Obama’s re-election, Mr. Schmidt was personally overseeing a voter-turnout software system for Mr. Obama. A few weeks later, Ms. Shelton and a senior antitrust lawyer at Google went to the White House to meet with one of Mr. Obama’s technology advisers. … By the end of the month, the FTC had decided not to file an antitrust lawsuit against the company, according to the agency’s internal emails.”
What is wild about the above leaked FTC document is it goes to great lengths to show an anti-competitive pattern of conduct toward the larger players in the ecosystem. Even if you ignore the distasteful political aspects of the FTC non-decision, the other potential out was:
“The distinction between harm to competitors and harm to competition is an important one: according to the modern interpretation of antitrust law, even if a business hurts individual competitors, it isn’t seen as breaking antitrust law unless it has also hurt the competitive process—that is, that it has taken actions that, for instance, raised prices or reduced choices, over all, for consumers.” – Vauhini Vara
Part of the reason the data set was incomplete on that front was for the most part only larger ecosystem players were consulted. Google engineers have went on record stating they aim to break people’s spirits in a game of psychological warfare. If that doesn’t hinder consumer choice, what does?
@aaronwall rofl. Feed the dragon Honestly these G investigations need solid long term SEOs to testify as well as brands.— Rishi Lakhani (@rishil) April 2, 2015
When the EU published their statement of objections Google’s response showed charts with the growth of Amazon and eBay as proof of a healthy ecosystem.
The market has been consolidated down into a few big winners which are still growing, but that in and of itself does not indicate a healthy nor neutral overall ecosystem.
The long tail of smaller e-commerce sites which have been scrubbed from the search results is nowhere to be seen in such charts / graphs / metrics.
The other obvious “untruth” hidden in the above Google chart is there is no way product searches on Google.com are included in Google’s aggregate metrics. They are only counting some subset of them which click through a second vertical ad type while ignoring Google’s broader impact via the combination of PLAs along with text-based AdWords ads and the knowledge graph, or even the recently rolled out rich product answer results.
Who could look at the following search result (during anti-trust competitive review no less) and say “yeah, that looks totally reasonable?”
Google has allegedly spent the last couple years removing “visual clutter” from the search results & yet they manage to product SERPs looking like that – so long as the eye candy leads to clicks monetized directly by Google or other Google hosted pages.
The Search Results Become a Closed App Store
Search was an integral piece of the web which (in the past) put small companies on a level playing field with larger players.
That it no longer is.
WOW. RT @aimclear: 89% of domains that ranked over the last 7 years are now invisible, #SEO extinction. SRSLY, @marcustober #SEJSummit— Jonah Stein (@Jonahstein) April 15, 2015
“What kind of a system do you have when existing, large players are given a head start and other advantages over insurgents? I don’t know. But I do know it’s not the Internet.” – Dave Pell
The above quote was about app stores, but it certainly parallels a rater system which enforces the broken window fallacy against smaller players while looking the other way on larger players, unless they are in a specific vertical Google itself decides to enter.
“That actually proves my point that they use Raters to rate search results. aka: it *is* operated manually in many (how high?) cases. There is a growing body of consensus that a major portion of Googles current “algo” consists of thousands of raters that score results for ranking purposes. The “algorithm” by machine, on the majority of results seen by a high percentage of people, is almost non-existent.” … “what is being implied by the FTC is that Googles criteria was: GoogleBot +10 all Yelp content (strip mine all Yelp reviews to build their database). GoogleSerps -10 all yelp content (downgrade them in the rankings and claim they aren’t showing serps in serps). That is anticompetitive criteria that was manually set.” – Brett Tabke
The remote rater guides were even more explicitly anti-competitive than what was detailed in the FTC report. For instance, requiring hotel affiliate sites rated as spam even if they are helpful, for no reason other than being affiliate sites.
Is Brand the Answer?
About 3 years ago I wrote a blog post about how branding plays into SEO & why it might peak. As much as I have been accused of having a cynical view, the biggest problem with my post was it was naively optimistic. I presumed Google’s consolidation of markets would end up leading Google to alter their ranking approach when they were unable to overcome the established consensus bias which was subsidizing their competitors. The problem with my presumption is Google’s reliance on “data” was a chimera. When convenient (and profitable) data is discarded on an as need basis.
Or, put another way, the visual layout of the search result page trumps the underlying ranking algorithms.
Google has still highly disintermediated brand value, but they did it via vertical search, larger AdWords ad units & allowing competitive bidding on trademark terms.
If Not Illegal, then Scraping is Certainly Morally Deplorable…
As Google scraped Yelp & TripAdvisor reviews & gave them an ultimatum, Google was also scraping Amazon sales rank data and using it to power Google Shopping product rankings.
Around this same time Google pushed through a black PR smear job of Bing for doing a similar, lesser offense to Google on rare, made-up longtail searches which were not used by the general public.
While Google was outright stealing third party content and putting it front & center on core keyword searches, they had to use “about 100 “synthetic queries”—queries that you would never expect a user to type” to smear Bing & even numerous of these queries did not show the alleged signal.
Here are some representative views of that incident:
- “We look forward to competing with genuinely new search algorithms out there—algorithms built on core innovation, and not on recycled search results from a competitor. So to all the users out there looking for the most authentic, relevant search results, we encourage you to come directly to Google. And to those who have asked what we want out of all this, the answer is simple: we’d like for this practice to stop.” – Google’s Amit Singhal
- “It’s cheating to me because we work incredibly hard and have done so for years but they just get there based on our hard work. I don’t know how else to call it but plain and simple cheating. Another analogy is that it’s like running a marathon and carrying someone else on your back, who jumps off just before the finish line.” Amit Singhal, more explicitly.
- “One comment that I’ve heard is that “it’s whiny for Google to complain about this.” I agree that’s a risk, but at the same time I think it’s important to go on the record about this.” – Matt Cutts
- “I’ve got some sympathy for Google’s view that Bing is doing something it shouldn’t.” – Danny Sullivan
What is so crazy about the above quotes is Google engineers knew at the time what Google was doing with Google’s scraping. I mentioned that contrast shortly after the above PR fiasco happened:
when popular vertical websites (that have invested a decade and millions of Dollars into building a community) complain about Google disintermediating them by scraping their reviews, Google responds by telling those webmasters to go pound sand & that if they don’t want Google scraping them then they should just block Googlebot & kill their search rankings
Learning the Rules of the Road
If you get a sense “the rules” are arbitrary, hypocritical & selectively enforced – you may be on to something:
- “The bizrate/nextag/epinions pages are decently good results. They are usually well-format[t]ed, rarely broken, load quickly and usually on-topic. Raters tend to like them” … which is why … “Google repeatedly changed the instructions for raters until raters assessed Google’s services favorably”
- and while claimping down on those services (“business models to avoid“) … “Google elected to show its product search OneBox “regardless of the quality” of that result and despite “pretty terribly embarrassing failures” ”
- and since Google knew their offerings were vastly inferior, “most of us on geo [Google Local] think we won’t win unless we can inject a lot more of local directly into google results” … thus they added “a ‘concurring sites’ signal to bias ourselves toward triggering [display of a Google local service] when a local-oriented aggregator site (i.e. Citysearch) shows up in the web results””
Google’s justification for not being transparent is “spammer” would take advantage of transparency to put inferior results front and center – the exact same thing Google does when it benefits the bottom line!
Around the same time Google hard-codes the self-promotion of their own vertical offerings, they may choose to ban competing business models through “quality” score updates and other similar changes:
The following types of websites are likely to merit low landing page quality scores and may be difficult to advertise affordably. In addition, it’s important for advertisers of these types of websites to adhere to our landing page quality guidelines regarding unique content.
- eBook sites that show frequent ads
- ‘Get rich quick’ sites
- Comparison shopping sites
- Travel aggregators
- Affiliates that don’t comply with our affiliate guidelines
The anti-competitive conspiracy theory is no longer conspiracy, nor theory.
Key points highlighted by the European Commission:
- Google systematically positions and prominently displays its comparison shopping service in its general search results pages, irrespective of its merits. This conduct started in 2008.
- Google does not apply to its own comparison shopping service the system of penalties, which it applies to other comparison shopping services on the basis of defined parameters, and which can lead to the lowering of the rank in which they appear in Google’s general search results pages.
- Froogle, Google’s first comparison shopping service, did not benefit from any favourable treatment, and performed poorly.
- As a result of Google’s systematic favouring of its subsequent comparison shopping services “Google Product Search” and “Google Shopping”, both experienced higher rates of growth, to the detriment of rival comparison shopping services.
- Google’s conduct has a negative impact on consumers and innovation. It means that users do not necessarily see the most relevant comparison shopping results in response to their queries, and that incentives to innovate from rivals are lowered as they know that however good their product, they will not benefit from the same prominence as Google’s product.
Overcoming Consensus Bias
Consensus bias is set to an absurdly high level to block out competition, slow innovation, and make the search ecosystem easier to police. This acts as a tax on newer and lesser-known players and a subsidy toward larger players.
Eventually that subsidy would be a problem to Google if the algorithm was the only thing that matters, however if the entire result set itself can be displaced then that subsidy doesn’t really matter, as it can be retracted overnight.
Whenever Google has a competing offering ready, they put it up top even if they are embarrassed by it and 100% certain it is a vastly inferior option to other options in the marketplace.
That is how Google reinforces, then manages to overcome consensus bias.
How do you overcome consensus bias?