How to prepare yourself against Google SEO Penguin update?
Unless you're living in a remote part of the world, undoubtedly you would have heard the rumblings of a coming Google Penguin update of significant proportions.
To paraphrase Google’s web spam lead - Matt Cutts, the algorithm filter will have a "next generation" coming and that undeniably will have a major impact on the Search Engine Results Page (SERPs).
Having watched the initial rollout take many by surprise it make sense this time to at least attempt to prepare for what may be lurking around the corner.
What iClick Media knows about Google Penguin?
We know that the Google Penguin is essentially a link quality filter that sits on top of the search engine results page's core algorithm. It runs sporadically (the last official update was in October 2012), and is designed to take out websites that uses manipulative (or black hat SEO) techniques to improve their customers' search visibility.
While there have been many examples of this being badly executed (just google for "BMW SEO Ban" or "JC Penney SEO Penalise"), with lots of site owners and SEO professionals complaining of injustice, it is clear that the Google web spam engineers have collected a lot of information over recent months and have improved results in many different verticals.
What this means is that Google's team is now on top of the existing data pile and testing output and as a result they are hungry for a major structural change to the way the filter works once again. To all clients who are managing their own SEO - beware!
We know that months of manual resubmissions and disavows have helped the Silicon Valley giant collect an unprecedented amount of data about the "bad neighborhoods" of links that had powered rankings until very recently, for thousands of high profile sites.
They have even been involved in specific and high profile web spam actions against sites like Interflora, working closely with internal teams to understand where links came from and watch closely as they were removed.
In short, Google’s new data pot makes most big data projects look like a school project! All the signs therefore point towards something much more intelligent and all encompassing.
The question is how can you profile your links and understand the probability of being impacted as a result when Penguin hits within the next few weeks or months?
If what you're reading so far sounds geeky and alien, don't worry. Just call 6362 0123 or send us a Request for Quotation here! If not, read on..
The Link Graph – Bad Neighborhoods
Google (and to a certain extent Yahoo) knows a lot about what bad links look like now. They know where a lot of them live and they also understand their DNA. (Ip address)
Once Google start looking it becomes pretty easy to spot the bad from the good and natural.
The link graph is a kind of network graph and is made up of a series of "nodes" or clusters. Clusters form around IPs and as a result it becomes relatively easy to start to build a picture of ownership, or association. An illustrative example of this can be seen below:
Google assigns weight or authority to links using its own PageRank currency, but like any currency it is limited and that means that we all have to work hard to earn it from sites that have, over time, built up enough to go around.
This means that almost all sites that use "manipulative" authority to rank higher will be getting it from an area or areas of the link graph associated with other sites doing the same. PageRank isn't limitless.
These "bad neighborhoods" can be "extracted" by Google, analyzed and dumped relatively easily to leave a graph that looks a little like this:
They won’t disappear, but Google will devalue them and remove them from the PageRank picture, rendering them useless.
Expect this process to accelerate now the search giant has so much data on "spammy links" and swathes of link profiles getting knocked out overnight.
The concern of course is that there will be collateral damage, which is really what this process is, there will be some who are ranked successfully, and some who are not.
Link Building Speed
Another area of interest at present is the rate at which sites acquire links. Many clients have spoken out on the speed of link acquisition. Many clients demand to have the most amount of link, in the shortest amount of time. In recent months there definitely has been a noticeable change in how new links are being treated. While this is very much theory my view is that Google have become very good now at spotting link velocity "spikes" and anything out of the ordinary is immediately devalued.
Whether this is indefinitely or limited by time (in the same way "sandbox" works) I am not sure but there are definite correlations between sites that earn links consistently and good ranking increases. Those that earn lots quickly do not get the same relative effect.
And it would be relatively straightforward to move into the Penguin model, if it isn't there already. The chart below shows an example of a "bumpy" link acquisition profile and as in the example anything above the "normalized" line could be devalued.
The "trust" of a link is also something of interest to Google. Quality is one thing (how much juice the link carries), but trust is entirely another thing.
Majestic SEO has captured this reality best with the launch of its new Citation and Trust flow metrics to help identify untrusted links
How is trust measured? In simple terms it is about good and bad neighborhoods again.
In my view Google uses its Hilltop algorithm, which identifies so-called "expert documents" (websites) across the web, which are seen as shining beacons of trust and delight! The closer your site is to those documents the better the neighborhood. It’s a little like living on the "right" road.
If your link profile contains a good proportion of links from trusted sites then that will act as a "shield" from future updates and allow some slack for other links that are less trustworthy.
Social Signals (Facebook & Twitter especially!)
Many Search Engine Optimisation (SEO
) pros believe that social signals will play a more significant role in the next iteration of Penguin.
While social authority, as it is becoming known, makes a lot of sense in some markets, it also has limitations. Many verticals see little to no social interaction and without big pots of social data a system that qualifies link quality by the number of social shares across site or piece of content can't work effectively.
In the digital marketing industry it would work like a dream but for others it is a non-starter, for now. Google+ is Google’s attempt to fill that void and by forcing as many people as possible to work logged in they are getting everyone closer to Plus and the handing over of that missing data.
In principle it is possible though that social sharing and other signals may well be used in a small way to qualify link quality.
Most SEO professionals will point to anchor text as the key nap shot metric when it comes to identifying spammy link profiles. The first Penguin rollout would undoubtedly have used this data to begin drilling down into link quality.
iClick Media have discussed this issue with many of the other SEO practitioners in Singapore and their opinions on what the key indicator of spam was in researching this post and almost all pointed to anchor text.
Expect this to tighten even more as Google’s understanding of what natural "looks like" improves.
One area that will certainly be under the microscope as Google looks to improve its latent semantic understanding is relevancy. As it builds up a picture of relevant associations that data can be used to assign more weight to relevant links. Penguin will certainly be targeting links with no relevance in future.
- 2013-04-15 18:15:45
- Search Engine Optimisation (SEO)