Advance LSI and SILO for google indexing

Advanced On-page Optimization – LSI & SILO structures The Latent Semantic Indexing

The LSI and SILO structures are regarded presently as the future of SEO on-page optimization. Google is known to have patented different technologies that make extensive use of LSI and results have shown that this technology is already operating on organic results these days.

So what is LSI?

LSI stands for Latent Semantic Indexing.

LSI allows a search engine to determine quickly the relevance of a website outside of the actual search query that a user would type in to search for content.

For example a website about dogs may naturally have content about dog training, canine obedience, dog history and German Shepherds.
LSI adds a new process when indexing a website. Not only are the keywords considered but the entire content of the website is considered as well. The whole site is now considered instead of just single keywords on a single page.

LSI determines whether a site has other terms that are semantically close to the queried keyword. This means that the algorithm is able to determine good content by the natural occurrence of other keywords within the website that are similar (semantic or semantically related) to the queried keyword. LSI lowers the effect of documents that are semantically distant.

There has been a massive amount of research that has gone into this algorithm. It is worthy of note that this simple method of determining semantic closeness to a queried word is surprisingly close to how an actual human would classify a document as being relevant. Yet it is an advanced automated process being conducted by thousands of computer servers.

Documents, which are seen as semantically close, tend to rank surprisingly higher than documents that do not use semantics to cover their theme. LSI is the closest the search engines have come to mimicking human thought.

Why is Google paying much attention to LSI and implementing it? Remember we talked about scrapers earlier?

LSI has the net effect of producing more relevant results for searchers. The days of search engine algorithms using single keyword matches are coming to a fast extinction. The search engines now know that single keyword factors can yield very low quality results by allowing spammers and lazy webmasters to manipulate the rankings.

LSI, in its most basic form, tries to associate words with concepts to ensure relevancy. For example, the words pool and table are associated with the term billiards and not swimming pools and kitchen tables. The words taco and bell are associated with a restaurant chain and not a popular Mexican food or musical instrument.

So what are the implications of LSI for our purposes?

LSI means that by covering the theme thoroughly we will be able to achieve high rankings above most of the sites out there.
So how do we cover our niche thoroughly?

For this purpose, we need to understand what Google thinks of a search term. We need to understand the keywords that Google associates to a search term (correlated keywords, synonyms, etc). By implementing later in our blueprint (unique contents) these keywords all along the website on each corresponding section we will be in an optimal situation. By covering a wide semantic net (covering keywords and related keywords and synonyms of those words) you will rank higher in your niche.

How is the design of the website involved in the process?

A website needs to be properly structured to take full advantage of LSI. Designing your site and your pages using synonyms and correlated keywords will yield higher quality content naturally. You will also be properly communicating the theme of your site to the search engines and will rank higher for your keywords.

What does LSI mean in the practice?

In the practice, LSI will be useful for developing our blue print and keyword-rich contents on each section within our silos structures. By making exhaustive usage of all the keywords that semantically have sense for Google on each case, we will be able to build highly SEO optimized content rewarded by Google.

Let´ s make a semantic search example for the keyword apple. This is achieved by performing the following search on Google: ~apple

Have a look at the related keywords that occur in results:

  • Apple Computer
  • QuickTime
  • Apple’s
  • Windows
  • Computer
  • Mac
  • iMac
  • G4

It is easy to observe that in no moment Google understands in this case the keyword apple as semantically related to the fruit. Instead, it is associating it to computer related stuff.

So LSI enables us to understand the semantic logic that Google is looking for when analyzing our content and assigning a value to it.

What are SILOS?

As we develop the overall theme of the site, as a whole, we will start to creep upwards in rankings for very competitive keywords by concentrating on individual sub-themes that include less competitive keywords within your site. By concentrating on the less broad keywords (long tail) in your sub-themes, you will begin to rank for broader and broader keywords by covering your sites overall theme at a very deep level.

This will be done through strict organization of your site into themes or silos and then including synonymically related keywords within the content of individual pages within each silo structure within your site (LSI).

The most basic explanation of a silo is to keep all related information inside of one folder. By generating different SILOS we will be able to cover our niche with maximum detail for all the relevant keyword variations in different levels.

Developing your Blue Print

Now that we know what LSI means, how we can use it for our benefit on SEO projects? How we are supposed to structure the website based on our SILOS structure? It is time to design our own blue print.

This process will start to create our own specific content for each specific keyword within each sub-theme on its respective silo structure. But this content will not have a random nature. We already know how to identify by usage of LSI, semantically related keywords. So it is time to use this information.

Once you initiate your writing for each specific section, you will have to use semantically related keywords (in Google´ s understanding as explained before) as well as available synonyms. Generating contents this way determines your blue print and you will be able to populate your website with highly optimized & top-notch content from an SEO perspective.

Short-term requirements may not allow a website owner to dedicate so much time to silo structuring and content creation. There are also budget constraints based on the paid services to work on SILO & LSI structures.

Last but not least, though it is the optimal approach regarding SEO on-page factors, the LSI & Silo structuring is highly demanding on unique content volumes. Remember that you must apply LSI on each section so this will be a time consuming process to develop the best possible blueprint.

As we can easily observe, it is by no means a unilateral decision. In the end, it all comes to short-term against long-term requirements as well as budget availability when working with an independent consultant.

Posted in Uncategorized

How to maintain silo structure for google ranking?

Maintaining a Great Silo Structure


Creating fresh content for your site is the key to increasing your value with the search engines and keeping ahead of your competition. If you have a strong silo structure to begin with, it will be easy to determine your ongoing content creation strategy based on that structure. For example, for the site in the example, you have three secondary categories for which you can create articles to add to the content of your site. So an example of your weekly content creation strategy might be:


Monday: Write and post one article on “how to play the blues guitar.”


Tuesday: Write and post one article on “how to play the classical guitar.”


Wednesday: Write and post one article on “how to play the acoustic guitar.”


Thursday: Write and post one article on “how to play the blues guitar.”


Friday: Write and post one article on “how to play the classical guitar.”


Saturday: Write and post one article on “how to play the acoustic guitar.”


After about three months of applying this kind of content creation strategy, you could proceed with a monthly content creation strategy, again based on your overall silo structure. This strategy would be to create one article per week for each of your three categories. By the time six months had passed, you’d have a  well structured site with over 100 unique articles, which (provided that you did good keyword research) would be sure to bring you a good ranking in the search results.
Posted in Uncategorized

SEO Advance lesson

SEO importance post Google’s Panda and Penguin algorithm changes: My assumtions

  • Bleeding PageRank
  • Bottling
  • Content Silos (Siloing)

Bleeding PageRank (PR)
When Google’s founders were pounding beer at local Palo Alto eateries they wrestled with a fundamental problem – how to value websites? Valuing a site is like the Oscars you may HATE movies someone else loves. Movies are a highly individual things. Google had the same problem with websites. What you and I like may not be the same.

The Googlers solved this problem by NOT solving it. Instead of trying to tease out who should win based on impossibly subjective metrics they walked on over to the wisdom of crowds window and opened it wide. “We won’t ‘value’ anything,” Google decided, “we will gather wisdom of crowds in the form a 1 to 10 ordered rank where 10 is US and 0 is someone who doesn’t have any links value their content (more or less)”. Beers slammed to the table and the meeting was adjourned to write the math.

Google’s algorithm is a cool idea with one big problem – fraud and black hatters. Fraud happens when some competitor sets up a autobot to bleed your site’s currency (traffic, PageRank, keywords). Black Hat SEOs attempt to find crevices in Google’s math and then slide in like water, freeze into ice and force the crack apart. When the money is as large as it is some portion of the populace will want to play the game a tad differently wearing “black hats”.

Your site has a rank you can check with a tool like Mikes. Put in your URL and Mikes will ask Google for your homepage’s (or whatever page’s) PR. You bleed rank when you include a lot of links out to non-siloed content you own or, and this kind of link is much worse since Google-juice gallops away, to external sites and pages.

PR is divided by all the links on a page, so each link is valuable. If you have to link out beyond say the most important 5 links on the page, use a “no follow” link. Add a piece of code that tells the spider not to follow the link, not to count pages it would have found. Look up “rel no follow” for how to do this if you are DIYing, but be CAREFUL as you are playing with live ammunition and someone else (Google) holds the gun and will decide when and if to fire.

Use Rel No Follows to stop bleeding rank over pages you never want in the engines such as your password controlled B2B site and all those legalese terms and conditions pages (I might unhook T&C pages out from the navigation too). If I could meet legal requirements by simply having the page up on a server somewhere that is what I would do (don’t let it even walk inside your “stack” of content – the stuff you WANT in Google, the stuff you want your traffic to see. You can also eliminate links on your pages.

Every site lives by a Pareto distribution also known as the 80/20 rule where 80% of sales (or clicks) get done by 20%(or less) of the products. I did extensive testing on this idea in a past life and found there was a Pareto distribution in every segment NO MATTER HOW SMALL “SEGMENTS” were cut. If we selected a random 19 products we would, with slight differences be able to see the most amazing repeating fractal in Internet marketing (read Anderson’s Long Tail for more on this concept).

When you pass rank to someone who doesn’t pass it back you bleed and they become vampires. Don’t send links out from a homepage to any unknown, unheard of page or page cluster EVER!. When you create social content resist creating a bunch of new links OUT or you bleed content from YOU to THEM. Apply this same rule to your internal links. More links = more bleed, so require giving up 2 links for every one new one you add and you will stop your bleed.

Bottling is evaluating every link on a page and deciding if you will pass Google-Juice. Where to pass and when to pass is covered in Siloing, but even if you never get as sophisticated as developing content silos you can make a decision if you want to pass juice to your site’s terms, your order form, your return policy (may want to on your return policy if it is great and buzzed about, if it is standard then keep the spider out). You bottle PR when you evaluate page-by-page whether you will pass it or not.

Content Silos
Now we enter the organic chemistry of content marketing. One thing every Internet marketer learns early is there is always a curtain and more stuff going on back there than we will ever know. One curtain is in front of your IT department. They are in there doing stuff and, trust me, you don’t know the half of it (lol and said with love guys). The other curtain is Google and they would tell you what is going on behind that iron curtain but then they would have to kill you so better if they don’t.

What we can infer after all of this Zoo Roil (Panda and Penguin) is CONTENT IS KING and more king than ever before. If you aren’t watching your site’s heuristic measures such as Time On Site, Pages Viewed and Bounce Rate like a hawk on a wire looking for field mice…..well good luck with that.

You may not have gotten the full HOW GOOGLE HAS CHANGED THE WORLD AGAIN message, so I am here to share it with you in three words:

  • Speed
  • Site Experience (Spiders and People)
  • Money

Bruce Clay is the king of creating content silo themes. I won’t do his explanation justice, but here goes.

Content silos are a process of stacking related and supportive keyword content together. I think in physical terms but programmers, at least good advanced ones, know how to “stack” like this in the abstract. The spider sees a defined world that really only exists the moment the spider is looking at it but all of that is for another post. The key concept of content silos is to figure out the MASTER and SLAVE components.

There is only one MASTER page in a content silo. Master pages RECEIVE but don’t SEND Google-juice. Master’s receive juice from SLAVE pages inside the tower. Slave pages can only EVER send juice to the master. Slaves can’t share juice with each other no matter how thirsty.

This is not to say two slave pages don’t link to each other, they do, but they don’t pass juice. Links between slaves are always “rel no followed” bottling the juice in the slave page except when a link goes to the master page. You can see what this does right? This tactic bottles the rank in a series of pages and sends it to a single source creating a clear and unambiguous signal. The spider can’t mistake what is important.

This tactic fell out of favor just before the UGC (user generated content) explosion created by Facebook et al. Now that we create as much content every two days as from the dawn of man right up to 2003, we must cut through noise or we bleed rank. Bleed rank and your search engine listings are less, traffic is less and so money is less. Even if you have great, well constructed content silos there are other things that will kill your SEO. Here are just a few of things to make sure you do TOO (lol).

Lack Of Speed Kills SEO
Lack of speed will kill your SEO, traffic and conversions now. If you were x fast before Panda and Penguin you should be X * 10 now. If gaining speed requires you upgrade your hosting then UPGRADE YOUR HOSTING. If gaining speed means redoing your graphics then REDO YOUR GRAPHICS. Lacks of speed makes every other SEO thing MOOT.

Site Experience (Spiders and People)
At a conference a few weeks ago the presenter said there was never an audience for more than 100 words on a web page. If you just read that last sentence you are at word 1200 so clearly, for some topics and particular audiences such a statement is wrong on its face. The other goofystupid piece of the speaker’s statement is there is one very important audience for every word on your site – Google’s crawler, their spider, the code they use to index the impossible infinity of the World Wide Web.

I understand what the speaker was attempting to say – that no one reads anymore and less than that read online. She outlined a key core issue to Internet marketing. As soon as you make a declarative statement you are wrong (period full stop). The web is alive and must be treated as such. Here is the only declarative statement I’m comfortable with – When Something Appears To Be X It Isn’t. Live by that rule and you will be fine, crazy but fine (lol).

You must seesaw user experience with spider needs. The answer to both is tell great stories and the only way to know the right answer for your website is to test, test and test more. When you test you must, and I mean MUST, tie engagement metrics to engagement and behavior. If Bounce goes up greater than 2 points then the traffic you just brought in may be hurting more than helping. If your time on site quadruples then your site is crashing since those kinds of leaps don’t happen with real people and in real time (never, you just can’t write or campaign that good). Natural metric changes are usually within 5 standard deviation points. See something greater than 5 SD points positive or negative then you need to know why and NOW.

Posted in Uncategorized

MY SEO Consultancy

Web Analytics

Content Creation

Link Acquisition

Social Media Outreach

Expanding into New Markets

So you see that you don’t have to resort to PageRank sculpting or other questionable SEO techniques advocated by some people in the SEO industry who often fail to embrace a holistic fndability appraoch and instead overtly focus on technical aspects.

This list does not contain all current advanced SEO methods, by far not all of them.

I compiled this list quite quickly from the top of my head.

Posted in Uncategorized

Siloing is a very powerful method to get ranked on Google.

siloing is  an advanced SEO technique that is a heavyweight amongst SEO techniques.

What  is a Silo?
A  Silo  is a vertical page linking design. You have your landing page, or your main page, at the top of the silo and underneath this page you have pages which support your main landing page theme.

Siloing is architectually structuring your website to get the Search Engines to recognize your site is about your targeted keywords. You want the Search Engines to consider you as the subject matter expert.

There are several ways to create silos:
1) Tagging
2) Categories
3) Directories
4) Related pages plugin
5) ONLY link to landing pages using your Target keyword.
6) Create a mini-sitemap on each page within your silo.

Many websites will find the need to increase the amount of their content by creating additional webpages.

Expected Outcomes:
1) Higher rankings
2) Increase in overall traffic
3) More unique visitors
4) More traffic from long tail keywords.

Tagged with: ,
Posted in SEO

Link building technique for SEO

Broken link building is one of the few white-hat link building techniques that could bring good result for SEO in SERP.

But the question is how you would get a good and non-broken link for your website?

Find and find

Find quality resource or links pages for your websites. Let’s say, we are promoting a website on Bonsai tree and have a content marketing campaign for Bonsai trees in a ecommerce site.


Be clever and smart on google search for link building

A clever search on Google – like [bonsai inurl:links] is the best way to find out links.

Start the search phrase with your keyword followed by inurl:links
keyword inurl:links  what does it mean anyway

It tells Google to return pages that are relevant to your keyword, and have the word ‘links’ in the URL.

So if you sold cosmetics, you’d do this search:
cosmetics inurl:links

search on google , you will find lots of broken links of pages resources and other links that’s are related to cosmetics.

But sometimes there should be a problem, the links on the page can be broken – a reader or consumer or customer clicks on one of the resources and gets a 404 page not found message. That’s a very bad for the customer. He would get angry! And would leave your page forever too.

when you find a broken link on a resource webpage, Your first step should be to find other domains that link to the same broken resource.

So when working on broken link building process follow the steps:

Step 1: search for multiple resource pages that are relevant to your site. Each of these will list multiple quality resources – a percentage of which will be broken. The more resource pages you will find, the more broken link opportunities you will uncover.

Step 2: Scrape all the links from these resource pages and check their status to find broken links. The majority of these broken links will point to valuable resources that have moved or no longer exist.

The importance of keywords in broken link building research

To find broken link building you’ve got to conduct multiple searches. For freelancer we can search like this.

For example:

  • [freelancer inurl:resources] or
  • [freelancer inurl:links] or
  • [freelancer inurl:websites] or
  • [freelancer inurl:”recommended resources” and so on.


And the same, lets for robotics :

  • [“robotics” inurl:resources] or
  • [“robotics” inurl:links] or
  • [“robotics” inurl:websites] or
  • [“robotics” inurl:”recommended resources” and so on.


All of these searches will produce relevant resources pages.

The choice of keywords in these searches is important.







Posted in Uncategorized