Google's Matt Cutts explained in a blog post yesterday that his company is taking steps to target "content farms" that generate "low-quality" content that's vaguely on topic for various searches. He compared the challenge to Google's domination of "pure webspam," the sort of automatically generated, SEO nonsense of the post-boom, pre-Google Internet.
But I wonder if the challenge ahead of Cutts isn't greater than the one that faced his predecessors. With pure webspam, the giveaway was usually just that text was machine generated. Telling machine writing from human writing isn't that hard. But telling good human writing from not-so-good human writing is a whole different story. Humans, ourselves, don't seem to be so good at it (Cf. Dan Brown), so I wonder how the engineers plan to distill our uncertain/heuristic/malleable rules into an algorithmic weighting system.
As "pure webspam" has decreased over time, attention has shifted instead to "content farms," which are sites with shallow or low-quality content. In 2010, we launched two major algorithmic changes focused on low-quality sites. Nonetheless, we hear the feedback from the web loud and clear: people are asking for even stronger action on content farms and sites that consist primarily of spammy or low-quality content. We take pride in Google search and strive to make each and every search perfect. The fact is that we're not perfect, and combined with users' skyrocketing expectations of Google, these imperfections get magnified in perception. However, we can and should do better.
Read the full story at Google.