Talk Talk Talk

Based in part on Calacanis's recent tirades, Scott Karp recently published a great post about SEO from an outsider's perspective. In his post he runs through how and why some people are biases against SEO. I think a couple big reasons that few people talk about are mis-direction and outsourcing faults onto others.

In other news, what is going on with Goog on Google Finance? People are talking about Cramer. Some are talking about how intelligent he is while others are saying his packaging is bad and he is an idiot. Both are probably increasing his brand value though.

I have to agree. I find Cramer a bit of an idiot. I mean, apparently he has done well in stocks, I am not disputing that, but for an investment adviser he is someone I find....well, comical. I've watched his show, and to me he's like the circus; something to see and laugh at when it comes to town, but not something to take too seriously. His biggest fault, and this is ironically the draw of his show, is how he preys on and encourages the emotions of his followers. Now, he may say that its best not to invest with emotion, but watching him run around on tv with his sleeves rolled up, yelling like some motivational speaker selling a new brand of energy drink, sure sends a different message. In fact, the high quality of his marketing skill, and the poor quality of his advice, kind of reminds me of the Motley Fool.....

In every market people who evoke emotional responses win. Even if they are wrong, you will see them refererenced often just because they are good at marketing and preying on human emotions.

Many popular people create far more controversy than value, but links and trust follow conversation. And so do ad dollars. If people are talking about you, you win, even if you are wrong.

Popular and correct are two different things, and the only way to know who you should trust is to test and then aim.

With the advancement of modern technology, people do not even vote on the content, just the headlines, which somewhat feeds into my belief that search personalization + using links as a proxy for value are going to create a polarized biased web full of recycled garbage. Everything is recycled.

Is it unfair to throw any of that blame toward search engines, or is it just default human nature to outsource our own faults and want to split things up to identify with things that are false but look good at a glance? Are our egos so broken that we have to be part of some minority or fighting for one to feel we have purpose? Must we have outspoken leaders to follow? Do the leaders believe their own words, or is it just self-serving marketing?

As more forms of vertical search come about, subscribing and publishing get easier, and more people vote without reading, you can bet that packaging will become more important than information quality...at least until people get sick of it.

I saw two popular pieces about saving money that explicitly gave money saving tips opposite of each other, both published by a friend, who recently talked up the value of his content. Some days sites like Motley Fool will tell you why a stock is a must buy and then have another article dissing the stock the same day. I think they even have a column based on biased polarized advice called Dueling Fools.

Big claims are remarkable, and worthy of a link. In a sea of rushed judgements and meaningless votes sounding convincing is more important than being correct. The perception of value and actually being of value are two different things. For the next couple years it will be far cheaper and more profitable to cater biased marketing to the ignorant rather than to create meaning with a bit of touch and originality. Or am I wrong?

Yahoo! Pipes Are Cool

Yahoo! Pipes is a visual RSS slicing, dicing, and meshing tool. Basically you can take any feeds you like, add them together, and apply a bit of filtering. It is fairly intuitive and a lot a of fun for a wannabe programmer like me. And then when you create something, someone else can clone your pipes and add more stuff to it.

Here are some cool ways to use Yahoo! Pipes:

  • track the latest news in your industry (filtering by sources, keywords, or both)

  • track domain and marketplace offers like Shoemoney does here
  • track inbound links or mentions by syncing up blog search tools
  • see how fast ideas are spreading by tracking all the major social sites and blog search engines at the same time
  • watch eBay price trends (and just about any other trend which you can subscribe to)
  • see which Yahoo! Pipes are spreading, think of how you can improve those ideas or apply them to other markets
  • if you are dirty or aggressive about monetization ;) create a Pipe that is a core to many other pipes that pulls data from major websites like eBay while using your affiliate links in your Pipes

Tim O'Reilly has a great post about how the concept of Pipes could be highly valuable.

Historical Search Spam Patterns and Link Reciprocation

Some people are wildly speculating that Google and other engines may create historical databases of SEOs and site relationships to identify spam. I have no doubt that some sites that go way too far stay penalized for a long time, and that some penalties may flag related sites for review, but I think search engines have enough data and most people leave enough footprints that search engines do not have to dig too deep into history to connect the dots. And there is little upside in them connecting the dots.

If they did connect the dots manually that would take a long time to do it broadly, and if they did it automatically they would run into problems with false relationships. Some sites I once owned were sold to people who do not use them to spam. If ownership relationships took sites out by proxy I could just create spam sites using a competitors details in the Whois data , or heavily link to their sites from the spam sites.

Where people run into problems with spamming is scalability. If you scale out owning many similar domains you are probably going to leave some sort of footprint: cross linking, affiliate ID codes, AdSense account numbers, analytics tracking scripts, a weird page code, similar site size, similar inlink or outlink ratios, similar page size, or maybe some other footprint that you forgot to think of.

Many of those things can be spoofed too, (what is to prevent me from using your AdSense ID on spam?), so in many cases there has to be a hybrid of automated filtering and flagging and manual review.

And even if you are pretty good at keeping your sites unique on your end, if you outsource anything they are going to have a limited network size, likely a routine procedure with footprints, and if their prices are low they are probably going to be forced to create many obvious footprints to stay profitable. And if you use reciprocal or triangular links associated with those large distributed link farms that puts you in those communities far more than some potential historical relationship of some sort. By linking to it you confirm the relationship.

Search engines do not want to ban false positive, so many spammy link related penalties just suppress rankings until the signs of spam go away. Remove the outbound reciprocal link page that associates you with a bad community, get a few quality links, and watch the rankings shoot up. The thing is, once a site gets to be fairly aggressively spammy it rarely becomes less spammy. If it was created without passion it likely dies then turns into a PPC domainer page with footprints. Hiding low value pages deep in the index until the problem goes away is a fairly safe idea for search engineers, because after a domain has been burned it rarely shifts toward quality unless someone else buys it.

Adjusting Your Marketing

Some marketing fails because it does not use market feedback to help improve the ROI on the next generation of marketing. For example, if I make a couple sites and then take what I learned from making those and apply that to making more sites I will probably be more efficient than if I try to make many sites in parallel without collecting feedback. Many of the best marketers do absolutely stupid stuff that destroys the value of their work, other than what they have learned from testing the boundaries. But after you test them you learn and then you can incorporate that into your next round of marketing. It doesn't matter if you screw up as long as you keep learning from it, and adjusting to the market.

Before making a large commitment see if there are ways you can test the market and gain quicker and cheaper feedback. Build some content and links and see if it ranks. If it ranks build more content and links.

It is smart to emotionally invest into some of your most important projects, but it is a bad call to be so invested into the idea that if that idea doesn't work you keep pushing it against the will of the market until you go bankrupt, especially since there are so many market opportunities out there if you are willing to use market feedback to tweak your ideas to make them more profitable.

Ready. Fire. Aim.
Ready. Fire. Aim.
etc etc etc

Everyone is a Hypocrite and a Spammer

One hates to give Jason Calacanis any additional exposure, but how can a person be so against SEO while selling text links for scuba blackjack online? Is he ahead of the market on global warming?

Grow up.. the only thing you're ever going to prove by trying to game my SeRP is that you're low-class idiots.

True, or maybe we are looking for scuba blackjack customers, and knew that you publish high quality original content and ads for that market.

[Video] What is a Self Reinforcing Authority (and a Self Reinforcing Market Position)?


Video Summary:

Some documents and websites build self reinforcing authority that make them hard to beat for their targeted search terms. This video explains how that works and gives examples of some self reinforcing market authorities, as well as tips on how to make these types of sites and pages.

Resources Mentioned in the Video:

Examples of Self Reinforcing Authorities From This Video:

  • us debt clock

  • xe currency converter
  • search engine history
  • search engine ranking factors
  • black hat seo
  • seo code of ethics
  • seo today / search engine watch

Things I Should Have Mentioned That I Forgot:

  • Your title is important because most people will reference your document by its title.

  • Statistics, standards, and scientific sounding things are easy to become self reinforcing powerhouses, especially if they feed into the ego of the target audience.
  • If you get large media coverage of your idea leverage it to get more coverage. Show it off to seem exceptionally legitimate and trustworthy.
  • US News and world report ranks colleges, and is a great example of a self reinforcing authority.
  • Common ways to undermine authority that may prevent a site or article from becoming authoritative.
  • If someone has an authoritative idea in another market, but nobody has applied it to your market that may present an eay oppurtunity.

Signs of a Low Quality Website

Webmasterworld recently had a good thread about signs of low quality websites. The less a person knows about your topic the more likely they are to rely on general signs of quality (or lack of) when consider if they should link at your site or not.

Common Quality Questions:

Is the design clean? Is the content well organized? Do they have major misspellings on their homepage? Who is behind the site? Is it easy to contact them? Are they referenced by any other credible sources? How unique and useful is the content? How aggressively are ads blended into the content? etc. etc. etc.

Why Proxies for Quality Are Important:

Recently someone spread a God hates fags song website. Friends were instant messaging me about whether it was real or not. Some journalists guessed it wrong. People are getting better at creating fakes. The easier we make it for people to trust us in a snap judgement the more people will trust us (and link to our sites).

These proxies for trust are important, especially when you are new to an established industry, are in a new industry with a small community of support, are in a rapidly growing industry that the media is having a feeding frenzy over, or are the seedy arm of a larger industry.

Example of the Importance of Outside Perception:

If an industry is new, the early leaders of that industry might be determined by mainstream media perception (or other perception outside of that industry). Using blogs as an example, if the media did not constantly pump up the Weblogs Inc. story that company still might be unprofitable today. That media exposure lead to more media exposure, gave the sites the link juice to help them rank, and gave them brand exposure that brought in advertisements.

Relating This to the SEO Industry:

With SEO it is easier to be seen as a SEO expert if you are first seen as an expert on search. It is easier to be trusted as an expert on any topic if your site does not flag common signals of crap.

I just got a link from the WSJ to my keyword research tool, but if I would have scored lower on the proxies for value maybe they never would have linked. And when you get that type of link you can leverage it as an additional signal of trust that makes it easier for others to link at you.

With BlackHatSEO.com, I mentioned as seen in Clickz and Search Engine Watch, but what I didn't mention was that both mentions were brief and in the same syndicated article. When the London Times interviewed me about that site I quickly put up another as seen in at the top of the home page, which will make it easier to get more exposure. You want your press coverage to lead to more press coverage, because those are some of the most trusted links and links that money alone usually can't buy.

But I am Already Doing Well:

Many people who buy consultations are already doing far better than I would expect them to do giving some of the obvious flaws I see with their site structure and marketing methods. Some state that they are already doing well. The point of these sorts of signs of crap is not that you need to fill all the holes to do well, or that you can't do well if you do not fill them, but imagine how much better a site can do after it fixes obvious errors if it was already doing well when it had many errors that undermined its credibility and linkability.

Straight Out of the Andy Hagans Playbook

Andy Hagans reveals virtually everything there is to know about link baiting and social marketing.

Hidden Content Costs & Understanding Your Profit Potential Per Page

Ever since Google has got more selective with what they will index, the model for profitable SEO changed from chucking up pages and hoping some of them are profitable, to where it makes sense to put more strategy into what you are willing to publish.

The Supplemental Index Hates Parasitic SEO:

Each site will only get so many pages indexed given a certain link authority. And each of those pages will rank based on the domain's authority score, and the authority of the individual page, but each page needs a minimum authority score to get indexed and stay out of the supplemental results - this is how Google is trying to fight off parasitic SEO.

Given that many people are leveraging trusted domains, it makes sense that if you have one that you leverage it in a way that makes sense. CNN will rank for a lot of queries, but it does not make sense for Google to return nothing but CNN. It is good for the health of Google to have some variety in their search results. This is why smaller sites can still compete with the bigger ones, Google needs to use the smaller sites to have variety and to have leverage over the larger sites...to keep the larger sites honest if they are too aggressive in leveraging their authority, or have holes that others are exploiting.

Extending a Profitable Website:

If you have a 100 page niche website you may be able to expand it out to 500 pages without seeing too much of a drop in revenue on those first 100 pages, but eventually you will see some drop off where the cost of additional content (via link authority that it pulls from other pages on your site) nearly matches the revenue potential of the new pages. And then at some point, especially if you are not doing good keyword research, have bad information architecture, create pages that compete with other pages on your site, are not actively participating in your market (gaining links and mindshare), or if you are expanding from a higher margin keyword set to a lower margin one, you may see revenues drop as you add more pages.

The solution to fix this problem is build editorial linkage data and stop adding pages unless they have a net positive profit potential.

What are the costs of content?

  • the time and money that went into creating it

  • link equity (and the potential to be indexed) that the page takes from other pages
  • the mindshare and effort that could have been used doing something potentially more productive
  • the time it takes to maintain the content
  • if it is bad or off topic content, anything that causes people to unsubscribe, hurts conversion rates, or lowers your perceived value is a cost

How can a Page Create Profit?

  • anything that leads people toward telling others about you (links or other word of mouth marketing) is a form of profit

  • anything that makes more people pay attention to you or boosts the credibility of your site is a form of profit
  • anything that thickens your margins, increases conversion rates, or increases lifetime value of a customer creates profit
  • anything that reduces the amount of bad customers you have to deal with is a form of profit

Mixing Up Quality for Profit Potential:

I am still a firm believer in creating content of various quality levels and cost levels, using the authoritative content to get the lower quality content indexed, and using the lower quality content earnings to finance the higher quality ideas, but rather than thinking of each page as another chance to profit it helps to weigh the risks and rewards when mapping out a site and site structure.

Increasing Profit:

Rather than covering many fields broadly consider going deeper into the most profitable areas by

  • creating more pages in the expensive niches

  • making articles about the most profitable topics semantically correct with lots of variation and rich unique content
  • highly representing the most valuable content in your navigational scheme and internal link structure
  • creating self reinforcing authority pages in the most profitable verticals
  • requesting visitors add content to the most valuable sections or give you feedback on what content ideas they would like to see covered in your most valuable sections
  • If your site has more authority than you know what to do with consider adding a user generated content area to your site

Take Out the Trash:

If Google is only indexing a portion of your site make sure you make it easy for them to index your most important content. If you have an under-performing section on your site consider:

  • deweighting it's integration in the site's navigational scheme and link structure

  • placing more internal and external link weight on the higher performing sections
  • if it does not have much profit potential and nobody is linking at it you may want to temporarily block Googlebot from indexing that section using robots.txt, or remove the weak content until you have more link authority and/or a better way to monetize it

Google Offers More Link Data

Google's Link: command has been broken forever, but now Google is letting you see a far more representative sample of external links to your site and your internal link structure if you verify that you are the owner of your site by signing up at Google Webmaster Central. They also allow you to export your linkage data in an excel file. Some ways to use this data:

  • look at internal link structure of important pages and make sure they are well represented

  • look at internal external structure of important pages and make sure they are well represented
  • look at which pages on your site are well represented and make sure they link to other key pages
  • download your external linkage data and sort by date to look for new link sources (and why they are linking at your site)
  • run the excel sheet through a duplicate site remover or c class IP range checker to see how diverse your linking profile is

If you have shifty sites obviously there would be little to no upside in verifying those sites with Google, but if your sites are generally above board you might find this tool useful.

Thanks to Adam.

Pages