Understanding the basics of link-based spam detection can improve your understanding of link valuation and help you understand how search engines approach the problem of spam detection, which can lead to better link building practices.
I’d like to talk about a few interesting link spam analysis concepts that search engines may use to evaluate your backlink profile.
I don’t work at a search engine, so I can make no concrete claims about how search engines evaluate links. Engines may use some, or none, of the techniques in this post. They also certainly use more (and more sophisticated) techniques than I can cover in this post. However, I spend a lot of time reading through papers and patents, so I thought it'd be worth sharing some of the interesting techniques.
#1 Truncated PageRank
The basics of Truncated PageRank are covered in the paper
Linked-based Characterization and Detection of Web Spam. Truncated PageRank is a calculation that removes the direct “link juice” contribution provided by the first level(s) of links. So a page boosted by naïve methods (such as article marketing) are receiving a large portion of the PageRank value directly from the first layer. However, a link from a well linked to page will receive “link juice” contribution from additional levels. Spam pages will likely show a Truncated PageRank that is significantly less than the PageRank. The ratio of Truncated PageRank to PageRank can be a signal to indicate the spamminess of a link profile.
#2 Owned / Accessible Contributions
Links can be bucketed into three general buckets.
- Links from owned content – Links from pages that search engines have determined some level of ownership (well-connected co-citation, IP, whois, etc.)
- Links from accessible content – Links from non-owned content that is easily accessible to add links (blogs, forums, article directories, guest books, etc.)
- Links from inaccessible content – Links from independent sources.
A link from any one of these source is neither good nor bad. Links from owned content, via networks and relationships, are perfectly natural. However, a link from inaccessible content could be a paid link, so that bucket doesn’t mean it’s inherently good. However, knowing the bucket a link falls into can change the valuation.
This type of analysis on two sites can show a distinct difference in a link profile, all other factors being equal. The first site is primarily supported on links from content it directly controls or can gain access to. However, the second site has earned links from a substantially larger percentage of unique, independent sources. All things being equal, the second site is less likely to be spam.
#3 Relative Mass
Relative Mass accounts for the percent distribution of a profile for certain types of links. The example of the pie charts above demonstrates the concept of relative massive.
Relative Mass is discussed more broadly in the paper
Link Spam Detection Based on Mass Estimation. Relative Mass analysis can define a threshold at which a page is determined “spam”. In the image above, the red circles have been identified as spam. The target page now has a portion of value attributed to it via “spam” sites. If this value of contribution exceeds a potential threshold, this page could have its rankings suppressed or the value passed through these links minimized. The example above is fairly binary, but there is often a large gradient between not spam and spam.
This type of analysis can be applied to tactics as well, such as distribution of links from comments, directories, articles, hijacked sources, owned pages, paid links, etc. The algorithm may provide a certain degree of “forgiveness” before its relative mass contribution exceeds an acceptable level.
#4 Counting Supporters / Speeds to Nodes
Another method of valuing links is by counting supporters and the speed of discovery of those nodes (and the point at which this discovery peaks).
A histogram distribution of supporting nodes by hops can demonstrate the differences between spam and high quality sites.
Well-connected sites will grow in supporters more rapidly than spam sites and spam sites are likely to peak earlier. Spam sites will grow rapidly and decay quickly as you move away from the target node. This distribution can help signify that a site is using spammy link building practices. Because spam networks have higher degrees of clustering, domains will repeat upon hops, which makes spam profiles bottleneck faster than non-spam profiles.
Protip: I think this is one reason that domain diversity and unique linking root domains is well correlated with rankings. I don’t think the relationship is as naïve as counting linking domains, but an analysis like supporter counting, as well as Truncated PageRank, would make receiving links from a larger set of diverse domains more well correlated with rankings.
#5 TrustRank, Anti-TrustRank, SpamRank, etc.
The model of
TrustRank has been written about several times before and is the basis of metrics like
mozTrust. The basic premise is that seed nodes can have both Trust and Spam scores which can be passed through links. The closer to the seed set, the higher the likelihood you are what that seed set was defined as. Being close to spam, makes you more likely to be spam, being close to trust, makes you more likely to be trusted. These values can be judged inbound and outbound.
I won’t go into much more detail than that, because you can read about it in previous posts, but it comes down to four simple rules.
- Get links from trusted content.
- Don’t get links from spam content.
- Link to trusted content.
- Don’t link to spam content.
This type of analysis has also been used to
use SEO forums against spammers. A search engine can crawl links from top SEO forums to create a seed set of domains to perform analysis. Tinfoil hat time....
#6 Anchor Text vs. Time
Monitoring anchor text over time can give interesting insights that could detect potential manipulation. Let’s look at an example of how a preowned domain that was purchased for link value (and spam) might appear with this type of analysis.
This domain has a historical record of acquiring anchor text including both brand and non-branded targeted terms. Then suddenly that rate drops and after time a new sudden influx of anchor text, never seen before, starts to come in. This type of anchor text analysis, in combination with orthogonal spam detection approaches, can help detect the point in which ownership was changed. Links prior to this point can then be evaluated differently.
#7 Link Growth Thresholds
Sites with rapid link growth could have the impact dampened by applying a threshold of value that can be gained within a unit time. Corroborating signals can help determine if a spike is from a real event or viral content, as opposed to link manipulation.
This threshold can discount the value of links that exceed an assigned threshold. A more paced, natural growth profile is less likely to break a threshold. You can find more information about historical analysis in the paper
Information Retrieval Based on Historical Data.
#8 Robust PageRank
Robust PageRank works by calculating PageRank without the highest contributing nodes.
In the image above, the two strongest links were turned off and effectively reduced the PageRank of a node. Strong sites often have robust profiles and do not heavily depend on a few strong sources (such as links from link farms) to maintain a high PageRank. Robust PageRank calculations is one way the impact of over-influential nodes can be reduced. You can read more about Robust PageRank in the paper
Robust PageRank and Locally Computable Spam Detection Features.
#9 PageRank Variance
The uniformity of PageRank contribution to a node can be used to evaluate spam. Natural link profiles are likely to have a stronger variance in PageRank contribution. Spam profiles tend to be more uniform.
So if you use a tool, marketplace, or service to order 15 PR 4 links for a specific anchor text, it will have a low variance in PR. This is an easy way to detect these sorts of practices.
#10 Diminishing Returns
One way to minimize the value of a tactic is to create diminishing marginal returns on specific types of links. This is easiest to see in sitewide links, such as blogroll links or footer paid links. At one time, link popularity, in volume, was a strong factor which lead to sitewides carrying a disproportionate amount of value.
The first link from a domain carries the first vote and getting additional links from one particular domain will continue to increase the total value from a domain, but only to a point. Eventually inbound links from the same domain will continue to experience diminishing returns. Going from 1 link to 3 links from a domain will have more of an effect than 101 links to 103 links.
Protip: Although it’s easy to see this with sitewide links, I think of most link building tactics in this fashion. In addition to ideas like relative mass, where you don’t want one thing to dominate, I feel tactics lose traction overtime. It is not likely you can earn strong rankings on a limited number of tactics, because many manual tactics tend to hit a point of diminishing returns (sometimes it may be algorithmic, other times it may be due to diminishing returns in the competitive advantage). It's best to avoid
one-dimensional link building.
Link Spam Algorithms
All spam analysis algorithms have some percentage of accuracy and some level of false positives. Through the combination of these detection methods, search engines can maximize the accuracy and minimize the false positives.
Web spam analysis allows for more false positives than email spam detection, because there are often multiple alternatives to replace a pushed down result. It is not like email spam detection, which is binary in nature (inbox or spam box). In addition to this, search engines don’t have to create binary labels of “spam” or “not spam” to effectively improve search results. By using analysis, such as some of those discussed in this post, search engines can simply dampen rankings and minimize effects.
These analysis techniques are also designed to decrease the ROI of specific tactics, which makes spamming harder and more expensive. The goal of this post is not to stress about what links work, and which don’t, because it’s hard to know. The goal is to demonstrate some of the problem solving tactics used by search engines and how this impacts your tactics.
Justin, your disclaimer does not do you honor, it should read "I don't work at search engines, but I am a genius, so listen to me."
This is a great post and my favorite of yours thus far surpassing the Best Tips article.
The thing about links from inaccessible content (independent sources) is of two folds. I agree with you that they could be in many cases from paid links. I however think that the content's industry matters in that a tech firm's website (or blog posts) getting links from tech news outlets and blogs does not necessary have to be a paid scenario, and these are often cited sources or brand references, or even event notifications.
You are also right on the ball when you you conclude that "corroborating signals can help determine if a spike is from a real event or viral content, as opposed to link manipulation." They key is for the algorythm to account for enough variables to detect the nuances between the two. Which leads to the conversation about false positives where once more you are right on the top of it in explaining that "through the combination of these detection methods, search engines can maximize the accuracy and minimize the false positives."
As you know, we play cops and robbers, and as the algorythm is adjusted to better accuracy, black hats optimize to better leverage spammy techniques that will stand the false positive quality control.
Thank you for the post, this was just a great read.
Haha, thank you.
Definately. I think inaccessible content is more likely to be a higher valued link (in my opinion). I was just making sure to note that I don't think the other buckets are "bad" by default. I think sites should aim to get more inacessible links.
The paper talks about that they do use signals to define the threshold, so it's not a set level. If they get other signals, I could imagine them setting it higher. This is one reason I always say to be sure to "have a reason" your links are going up. It just appears more natural then an influx of links, when nothing changed (new content, branded search volume, social shares, etc.)
Thank you for the comment. I think I might do more of these research backed posts in the future.
+1 for the intro -> "I don't work at search engines, but I am a genius, so listen to me."
Wow, Justin. Please , let me totally recover from jet-lag to be able to write a deeper comment,buti wanted to give you my compliments for the amazing post.
While reading iwas wondering if actually exists "indipendent" tools able to reproduce all these kind of link's profile analysis. If not that could a great niche market to investigate.
I agree, I'd love to see some of these done using OSE data. Like supporter counting, Robust PageRank, Truncated Page, and PR variance should all be doable.
Justin, I found your article very refreshing and thoroughly enjoyable... linked research material is fascinating too. I do have a question though. Are you able to clarify what you mean by 'first level links' and 'hops' (in #1 and #4).
If you're looking at a page with inbound links, such as the circle (node/page/site) in the image under 4. By "hop" I mean looking at the pages/nodes that link directly to it. One jump away from where you started to the first "layer".
So if a guest post links to my page, the target is my page and the guest post is in the first level or layer of links. I'd say that guest post is one hop away from my page.
If that guest post gets linked to by a 3rd party blog post, that 3rd party blog post is now the 2nd layer, and 2 hops from my target page/node.
3rd party blog post -> guest post -> my page
2nd layer -> 1st layer -> my page
--
In #1, for example, let's say that you do directory submissions and earn links to the homepage of your site. However, the directory page you're earning a link from has no links to it. That means all the value from that directory link comes from that first layer. It's a supporting page that has no supporters itself.
I hope that helps?
Thanks for answering that, Justin :-)
So;
"Node" is analogous to "webpage" and "Hop" is distance of a backlink to the page (node) it supports.
Thus, in #4 - Spam sites have most links coming from 1-2 layers away, with a fast drop off. Non-spam sites have a peak around the 3rd layer, with a more natural decline.
If these were soundwaves, spam is like a drum, and non-spam is like the voice of an opera singer :-)
That's an awesome post Justin. One of your best so far. Your post is also a good reminder to read Google patents and papers whenever possible. For a moment i though i am reading Bill Slawski post. And for the same reason i highly recommend seobythesea blog. Now back to the post. I am simply amazed how Google algorthmically detect link based spam and even more amazed that still large number of websites rank high on directory, forum, blog comments and other spammy links.
Lol, I had the same feeling of reading a post by Bill :)
Agreed! the style is more like a Bill Slawski posts! :)
Great post with some fantastic graphics, I agree with the statements you make about the links comming from different IP's, varied anchors and the power behind brands.
In regards to setting up varied link profiles to scan what is a bad profile and what is a good one you really need to split it up for various niches and various TLD's. For example I have done a lot of work link building in the US market and then now most activity is in the AUS market. thinks are a bit different down here in terms of what works and what type of links you need to acquire to see the best results.
But overall if you have a quality brand and you work towards quality links and keep up to date with the latestes movements of the market.
I agree. I think some of the link analysis stuff done is keyword / niche dependent. It's how profiles match up relative to the other options. I don't think it's a cut and dry "this works and this doesn't". I've seen everything that's not suppose to work, work.
You know, those grapics realy helped me understand the concepts quite well. It is one thing to verbally explain something - but I like pictures. :)
Anyway, I wanted to thank you for writing this post, as I have been looking for such a clear and consice explanation as to what is or is not considered "spam" now.
+1 for you!
Great post Justin. These patents are some nice "light" reading I bet :)
Great stuff - sorry it took me all day to catch up and read it. Even if the engines don't use all of these tactics, it's still excellent food for thought with good takeaways about what makes for a quality link.
This is a great translation of technical documents into practical SEO techniques. I've recently been thinking about compiling a list of technical resources to gain insight from, and this post certainly contributes to that list. At a very general level, here's what I've got on the list:
Anyone else have sources that could be added to this list?
SEO by the Sea is the best blog on this stuff.
I also used Google scholar to find stuff. When I fund a publisher who does research in the area I'm reading about, I look for their other papers.
Agreed, as an intermediary between original research, patents, etc and SEO end-users, Bill Slawski's blog can hardly be bettered - https://www.seobythesea.com/.
Justin, you're writing in this area is superb too - very accessible, very valuable.
So, in addition to the request for original sources posted by DarioStereo, if anyone can recommend interpretive sources of the same standard as Bill Slawski and Justin Briggs, I'm sure I'd not be alone in welcoming these.
Excellent read, and I thank you for your research on this. Like gfiorelli1, I would be very interested to see some independant toolsets for analyzing this kind of data.
A few of these techniques seem mostly common sense; devalue links from sites that the webmaster owns, removing outliers (oddly high PR links) and measuring standard deviations of linking pages' PR, but I'm very intrigued by some of your other postulated methods, like removing the value from the first level of links.
All in all, a solid and logical post with some exceptional theories. Thank you!
Outstanding Post. I for one would like to see the Big G actually turn up the heat and increase the application on some of these detection methods. We see violations galore that seem to go unchecked. Hopefully the Panda algo evolution will add these checks in a more robust way. This is the evolution of Search at it's best. Link diversity always.
Justin -
Man, this post is phenomenal. Your explanations of some of these last week to me made a lot of sense as well. The ones I find most interesting are Truncated PageRank and Robust PageRank. I had never thought about the idea of clustering as being a spam signal, so they could remove the first set of nodes and use the 2nd tier linking domains. This make a ton of sense.
Robust PageRank is intriguing to me for the same reasons, because it effectually eliminates just a few sites bouncing up your rankings. We are forced to diversify.
Thank you for this post. It's one to come back to for explaining things to high-capacity clients.
Awesome post Justin. Definitely one of the best SEO articles I've read in a long time! Thanks for posting!
Hey Justin,
This is a great post. Thanks for sharing this information. Here are some additional info in addition to your tactics I would like to share with you guys:
I think Google combines all the link spam techniques and there is no black or white with this. In my opinion there is a braking point when 2/3 or more of these spam filters activate when Google decides to take you out of the index. In all the other cases it just sends your pages down the SERPs (it may be just a little bit) and it is hard to detect.
Hey Martin,
Your Link Ghosts are the very same idea I've been trying to sell to SEOs for the past 2 weeks!
My hypothesis is that if they are tracking patterns such as link velocity it's very likely they are tracking Link Ghosts as you call them and it raises a red flag. They can then check out these links via the cache and as well as the rest of your link profile to verify these are paid links and then penalize after the fact.
I spoke at length about it with Bill Slawski over Twitter last night and he doesn't believe there is anything to suggest proof of it especially because what would potentially raise a penalty is now gone and if you come under manual review scrutiny it won't be the removed links that you get penalized for anyway. Do you have any experience that suggests that this Link Ghosts have resulted in a penalty?
I think there is an effect when links drop off. There is obviously the issue associated with the value dropping off, so it's hard to separate changes from the fact there is just a drop in relevancy and popularity. Dropped off links can also impact crawl rates, crawl budgets, and indexation rates. So there are a lot of factors in play.
What may appear like a penalty may likely not be. Might just be a drop in head terms, a slight shake up across thousands of lontail terms, reduced indexed pages, and fewer pages receiving traffic from organic.
However, in one of those papers (I believe the historical one by Google) they discuss using items like churn of SERPs to identify commercial, high value, and likely spammed keywords. These keywords are likely used when buying links.
I think churn rate on links could show a pattern of aggressive targeting. If, when anchors drop off, the company pushes to get them back, either through more whitehat methods, or using another network, agency, or service, this would show a churn for this anchor. They could use this historical view to devalue this anchor text moving forward.
I agree that they likely won't "penalize" - but they can just not count new links for that anchor text for a period, or create a threshold of value that links of that anchor can provide.
This is just my opinion though.
As for Google banning sites:
I suspect there are complicated decision trees they use.
https://en.wikipedia.org/wiki/Decision_tree
It likely has a wide range of on-site, off-site, link, historical, and usage classifiers with various thresholds and conditionals. Like if they trigger the hidden text check, they may then hit a trust classifier and exceed it, so then it remains "safe". Or it hits a hidden text check, so then they check the compressibility of the text (spammy repetitive text compresses better than natural text), and if it compresses really well, it may move down another branch.
Good call and well said.
Also I have seen an MS spam determination white paper about the use of compression ratio as a spam determinant. Might be an interesting read for those interested:
https://research.microsoft.com/pubs/65140/www2006.pdf
It talks about other on-page spam determination methods as well.
Amazing post, amazing comments. Their should be awards for best post of the year and this is definitely one of them.
Hi Justin, I love how you don't just rehash the same old shit everybody else is talking about: you pioneer and that makes your articles (and this one in particular) great.
Epic post sir!
The problem with most link building is that most people didn’t know who to measure the link value so to let client stay with them and let the money coming their why they started to try everything, assuming that some of it will absolutely going to work without knowing which techniques are going to harm…
This post covers the deeper idea of how to go with the link building especially to those who are heavily depends upon link from blog roll, footer links and others…
Thanks Justin!
I knew it would be worth waiting for!
The part that worries me most is that grey area of inaccessible links which could be from diverse sites who love your work, or could actually be paid links! hmmm... perhaps the relative accuracy (or inaccuracy) of this technique might provide some insight on some of the inexplicably bad results we see in the SERPs.
You just keep pushing the bar higher with every post!
I agree.
This is a great post, my boss is teaching a web marketing course at Georgian College and he's having me compete against his students. This is a great post and will help me in my link building practices. If you don't mind I'm going to do a brief blurb in my news section and then link to this article so that all the students can read and learn from it. Our challenge is to get bakermewting, to the top of the search engines. It's an anagram of web marketing. Thanks for all the helpful info, there's a ton of great comments in here too!
Thanks Justin! You expressed multiple difficult concepts so fluently and with graphics that it made understanding those concepts quick and easy.
It's powerful that Google can change values of the algorithm in real time to gain understanding and insight. For instance, eliminating the value of first level links to identify the power of the second and third level links is like real time a/b testing of the SE. Fascinating!
Thank you! This kind of article really enhances my comprehensive understanding of the SE's which filters into every decision I make every day.
Great bit of work justin - something to show the client that thinks just buying random links is effective SEO. Let them see how much can be involved!
great post with deep insights! thank you!
This is the type of information that I would expect to pay for. Thanks for sharing your insight. For FREE!
The linking patterns you have brought to light can be used to improve future projects but I think many will find that it helps them see where they went wrong in the past. This is just the surface of what the algo and filters are capable of but there are still people that think their mass articles and comment SPAM beats scientists, mathematicians, and statisticians. Either that or they enjoy taking advantage of naive customers.
Holy shinkkies, Justin. Thank you for the insights. I've been looking for ways to step up my SEO game recently, and these tactics really gave me some good points to think about, and they have me thinking about new ways to structure my linkbuilding tactics for maximum juice.
Thanks for being awesome, my man.
Justin,
I have to admit that much of what you you had to say is over my head. I am reasonably new at Seo.
But have had some sucess, in building an on-line store and affililate marketing. So have used on-page and off-page seo. I appreciate your work and willingness to share and have bookmarked your pages and will continue to look for more information from you.
Knowledge booster post, i really like how the illustrations support and explain each techniques..Genius way to understand on how to create a link building pattern/method that works.
Fantastic article Justin. Thanks for sharing!
Don't mean to gush, but this is the most informative SEO related article I have read in quite a while. Great to see some very solid data and explanations of things I had suspected but had not done the research to be able to confirm.
Thanks!
Nice Job Justin :)
Excellent post which concisely summarises a number of pretty key points.
I think many SEO's should be aware of most of these factors to an extent anyway, but it's good to have them all presented in the same place and in such a clear manner as a reminder.
I definitely agree that there is likely an element of "variance from the norm for this niche/region/sector" in the analysis.
Thanks for sharing.
Great post!
Not had a chance to read through all the comment so apologies if this has already been bought up but are you guys planning on adding in #6 and #7 to OSE at any point? It would be REALLY cool if we could see link text change over time along with acquistion counts over time. I know Majestic does a bit of #7 but I've yet to see anything which does both.
The challenge has been set....
Thanx for the great post!
Great post really!
What i take from your nice article is that the most important thing in link building is still "DDD" (aka Diversify Diversify Diversify).
Great Job!
As alwyas some very useful information fro SEO Moz. Thanks
The Anchor Text vs. Time post is very revelational, and it makes complete logical sense. It's definitely one of the best ways to find out what sites are gaming the system with purchased links or other types of automated linkbuilding with programs. It kind of shows that if you put all of your eggs in one basket with building anchor text targeted links, when those lose value or just stop working for you, that's all you've got and you're going to drop some rankings. If you choose to purchase anchor text links, that means you'll have to keep it up for the entirety of the existence of your site so that you don't get hit from this filter.
Great revolutionary post, hope there's more to come.
Great post - although some concepts are on a higher plane than my brain can cope with at the moment. Going to have to come back to re-read this and check out a few of the patents you've linked to :)
Very insightful, useful in a practical sense and I am wiser for having read the post several times and al the comments.
I must admit, there were some aspects I had to re-read to understand the meaning of certain expressions e.g. I still don't undertsand what you mean by "domains will repeat upon hops" under #4 but it looks like I'm in a minority of 1 here!
Many thanks Justin! Very informative!
Justin, I agree with Eryck and others...if you don't work at the Search Engines you need too...lol...Great info, it has my creative thinking out of control...Nice job!
Thanx for the great post! But im still not 100% sure which method should i use :d
Exceptional post and special aspects of link valuation and quality. It is complicated to judge how well an inbound link will execute and revolutionize your website
Justin, one of the reasons I sub to SEOmoz was to gleen such content, albeit this content, is above my pay grade. I'll have to revisit and reread to grasp some of the heddy concecpts, but in general, a well laid out piece. - Neil
Many high-quality websites have legitimate link pages with 'more resources' or 'related sites'. However, these linkpages have very few external links pointing to them. Thus, a website that has many links from linkpages will seem spammy? Do you think Google has a special status for 'related websites' pages?
This post is terrific. It would be cool to see these moz version of these metrics like a truncated PA or mozRank. I'd love to build some tools based on around this but it's way too much data to consider without getting into things like MapReduce and building on the cloud...but this ambitious post is quite inspiring....
Thanks! Hope everything is going well.
I was talking with Adam at SEOmoz at Mozcon about some of the items in this post. I imagine some of this stuff is computationally intensive. It'd be nice to see some of these types of ideas work their way into SEO metrics, but I suspect some of these could be built out using the SEOmoz API.
Excellent post on link quality!!! TY
Great article I can't wait to see more from you
A great reminder to diversify your link-building efforts.
Thanks for the article Justin (and the illustrations).
Thanks for the awesome information. The graphics were great as well. Look forward to hearing more from you
I'm having a hard time getting over how awesome this post is. Really good insight into building a successful link profile. Thanks again!
Great Article! It gave me some new insights. I really like all the references you gave.
Tnx
Hey Justin,
Is this part of what drove the recent change in the LinkScape index?
Ian
Hey Ian,
This is independent from that. I'm not certain what drove all the changes to the LinkScape index. This post was just me summarizing stuff I've been reading for a while.
However, I'd love to see this type of stuff work its way into link tools.
A very interesting, insightful, well-written post. Thanks, Justin
As always Justin, brilliant post for a Link Builder (like myself).
I'll be sure to keep these factors in mind when we are conducting our campaigns.
Fantastic post Justin. Thanks for helping us see visually some of the metrics that we need to be aware of.
Another amazing post by Justin Briggs!
I have one question regarding the WHOIS.. How about if the domain is "Privacy Protected" , will it be good or bad if a WHOIS information is protected?
I can't know if it's good or bad. I think open whois could send positive signals, if the site is good. I wouldn't try to hide it if the site is legit.
I use to hide whois when I did affiliate marketing, but that was to hide my sites from snooping competition, not because I was attempting to hide from Google.
I have never myself noticed a minus effect from hiding WHOIS details but as Justin has said I generally encourage the WHOIS to be open and accessible for any I am working on.
Registrars still have acccess to WHOIS records whether you hide them or not. And guess who is a registrar.... Google since 2005.
Useful information. thanks.
Many thanks Justin, this will help us Gray hat guys make our links look more white hat, :p
Brilliantly indepth and insightful for anyone new to old in this area.
Always going to have your Gut instinct on top as well though right?
Definately. I'm fairly cautious about posting my gut instinct ideas, although I have many, I just can't prove them. It's nice to see papers that show search engine's processes, it helps provide some foundation for gut instincts.
Understanding the value of a link and having a natural looking link profile seems to be the key area as to where people are struggling at the moment... :( Great post - thanks.
Great Analysis Justin. Haven't seen many articles on this. Thanks for posting!
Hey Jenny, thanks! Bill at SEO by the Sea has some great patent based articles, but I agree it'd see more articles on topics like this.
Really interesting sharing. Great to see the data out of SEs patents and experimentations. In particular, I dig how you bucket keywords in Owned, Accessibile and Inaccessible.
Wow, amazing article! I loved it, epecially because you cited where you got all your information! All of these papers on spam detection are very interesting and they just reinforce that we need to let links come naturally or the search engines will know that we're trying to artifically increase our rankings.
Thank you so much for compiling all this.
Thanks! I have a handful of other papers I find interesting, so I might try to find some time to work them into posts.
One of the best Moz articles i've read, its nice to finally get some meaningful technical analysis of academic papers - it is areminder that I have been neglecting my continuing self education and need to start diving into the peer reviewed journals etc and not just the blogs and forums...
If you've never checked out SEO by the Sea, I highly recommend it for analysis of academic papers. Bill is basically the man when it comes to this type of stuff.
Great article on the value of the links we work so hard to acquire. Thanks Justin!
Excellent post on different aspects of link valuation and quality. It is difficult to judge how well an inbound link will perform and affect your website and no doubt that having experiance over the years helps. But, the search engine algorithms are always changing and our SERP environments are ever-evolving so, this is why its great to have some domains that you can abuse for experimentation and testing.
Justin-
Thanks for including the graphics in the post. Allows the post to be a lot easier to understand. Great work!
Great article, nice detail and analysis.
Thanks for this great article. Good to see mathematical results and algorthym analyses about SEO.
Great post Justin! You totally rock! Love the post...
thanks for sharing your idea......
Hi there! I simply wish to give an enormous thumbs up for the nice info you’ve here on this post. I will be coming again to your weblog for extra soon. Thank you