Sunday, April 09, 2006

Ebay Business - Longer Bidding Time The Bigger The Bank Balance

Optimizing web pages for high rankings in the search engines involved two main processes. Firstly there is the on-page factors which include what keywords you place where on the page itself. The second, and more important process is getting the off-page factors right - incoming links.

This article explores mainly the on-page factors. As the competition for a keyword phrase increases, off-page factors become more important to good rankings and these often mask the effects of on-page factors making it impossible to see what on-page factors are important. For this reason, I am going to look at a high ranking page with low levels of competition in Google.

First, let's consider what we mean by competition.

There are two ways to look at competition in Google. There is the competition a page has when you type the phrase with quotes, and the competition when you type the words without quotes. The number of results returned by Google in each case is YOUR competition.

The main differences between these two types of search are as follows:

Search with Quotes - this returns only those pages that have been "optimized" for the exact phrase.

Search without Quotes - this returns all pages that have been "optimized" for the words making up the phrase.

e.g. (in simple terms)

a) If you search Google for

alsatian dog

Google returns 41,000 competing pages.

b) If you search Google for

"alsatian dog"

Google returns 6,390 competing pages.

In (a) above, there are 41,000 pages that refer to alsatian AND dog, but not necessarily to alsatian dog.

In (b) above, there are 6,390 pages that refer to the exact phrase alsatian dog.

Now, if you want to rank well for the term "alsatian dog" on Google, you only have to compete with 6,390 other pages for this exact term.

However, there are 41,000 - 6,390 = 34,610 other pages that are related to this search, and might still beat you if Google sees them as more relevant than your page.

We have discussed before the importance of link reputation and PR in ranking. It is possible for a high PR page to rank well for a term like alsatian dog, even if it does not have the exact phrase on the page.

This fact clouds the issue somewhat, and so although I recommend searching with quotes to find the real competition, I also recommend that you look at the top few results in Google (as searched without quotes) to determine how important those "partial match" pages are.

A quick search at:

http://www.prsearch.net/

for alsatian dog, shows me that the top pages for this search without quotes have a low PR (0-3) and many of those pages have 0 incoming links.

The same search at PRSearch.net using quotes around the phrase show very similar results. The competing pages for the exact term have low PR and low incoming links.

This phrase should be easy to target and get top rankings if done properly.

A word of warning: Because the PR reported on the Google toolbar is out of date (see earlier), you cannot be 100% sure of the PR of the pages, even using a site like PRSearch. They will use the same formula that the toolbar uses, and so will be equally out of date. Only Google knows the exact PR it is using in its ranking for any one page.

A second check I often do is to check what the PR of the homepage of the site that is ranking well, as this gives me an indication of how important the site as a whole is. For the phrase alsatian dog (with or without quotes), the top page is:

http://www.castleofspirits.com/stories02/alsatian.html

The homepage

http://www.castleofspirits.com

has a PR of 6 - quite an important site.

However, there is no link to the alsatian page on the homepage, so the PR 6 homepage wont directly help towards the high ranking of the alsatian dog web page.

Doing a backward links check on Google does not help since there are no backlinks listed for this top ranking page.

OK, putting on my detective hat, I see a link at the bottom of the Alsatian page called "March 02 Ghost Stories". There is another link to "Ghost Story Page".

Clicking on the link to Ghost Story Page, I am taken to a PR 5 page:

http://www.castleofspirits.com/storypg.html

where I find a link to March 2002 Ghost Stories. Clicking that link takes me to a PR 3 page:

http://www.castleofspirits.com/stories02/mch2002.html

And on this page I find a link to Ghostly Alsatian dog.

So, the top ranking alsatian dog page has one link I know of from a PR 3 page. I might assume that this site also has a sitemap (although I cannot find one) where it contains a second link to the alsatian dog page. That means a total of 2 links, both internal.

I can assume from this that the alsatian page with a PR 2 is probably the correct PR, and the page itself has very few incoming links. I am confident that if I targeted the phrase alsatian dog, I would easily get a top ranking.

The phrase alsatian dog is therefore an EASY phrase to target.

As a final check I went to the searchguild difficulty tool mentioned in section 6 of this newsletter and typed my phrase into that. The Search Guild rates this term as EASY.

With relatively few off-page factors contributing to the high ranking of this page, I can only assume that the on-page factors are what makes this page stand out from the rest and rank at number 1 on Google.

There are a variety of tools available for calculating density, but I use a tool I wrote for myself and is not available for purchase.

Running this URL through my tool tells me a lot of useful information.

Density of the phrase "alsatian dog" on the page is 0.49%

The keyword is found ONCE in the title (11.11%), and TWICE in the main text on the page (a density of just 0.34%).

The keyword is not found in any header or meta tag!

As a second check I always look at what I call the partial density. That is the sum of the densities of all words that make up the phrase.

e.g. the phrase "alsatian dog" is made up of two words - alsatian AND dog. I look at the density of alsatian, and the density of dog, and combine the two densities.

This is useful because it tells me the density on the page of the words that make up the phrase (remember it is possible to rank well without the exact phrase on the page) - a kind of simplified page reputation.

The partial density of this page is 3.09%, made up of 7 occurrences of alsatian, and 12 occurrences of dog. This page is obviously about alsatians and dogs!

Let's look at the prominence of this phrase on the page. First an explanation of what prominence means.

Prominence is a measure of where on the page a word exists.

A prominence of 100 would mean it was the first word on the page.

A prominence of 1 would indicate it was the last word on the page.

A prominence of 50% would indicate it was the middle word on the page.

If the phrase was the first word (100% prominence) and the last word (1% prominence) on the page, the average prominence on the page would be about 50%. That means the keywords are well spread out on the page. As prominence increases, the keyword is found higher up the page, as it decreases, it is found lower down the page.

For analysis of top ranking pages, I look at not only the average prominence of ALL occurrences of the phrase on my page

i.e. how the keywords are spaced out on the page,

but also the prominence of the first occurrence on the page.

i.e. how close to the start of the document is the phrase first found?

The prominence of the first occurrence of the phrase alsatian dog is 99.67%. That means it is almost the first phrase on the page (only the word ghost comes before it).

The average prominence of the whole page for this term is 62.62%. That means that the keywords are distributed more in the upper portion of the page. Haven't I always told you that it was important to get your main keyword in the top one-third of the page?.

This page is a good one to study. It shows a top ranking page for a low competitive keyword phrase. Because of the low competition, incoming links and PR are less important (though if you have both, you could dominate this phrase), while on-page factors will make or break the ranking.

Even though the exact phrase is only found 3 times on the page, the fact that the phrase is in the title of the document and in the body text seems to be enough. This low density is backed up by using the words that make up the phrase several times on the page. Google will be in no doubt what this page is about.

A final help to the ranking of this page is the filename. Notice that part of the keyword phrase is found in the filename - alsatian.html

*****************************************************
Andy Williams is author of the free, ezSEO internet marketing newsletter, offering subscribers up-to-date information on all aspects of internet marketing.
*****************************************************

Article Source: http://EzineArticles.com/?expert=Andrew_Williams

Monday, April 03, 2006

Google Algorithm Problems

Have you noticed anything different with Google lately? The Webmaster community certainly has, and if recent talk on several search engine optimization (SEO) forums is an indicator, Webmasters are very frustrated. For approximately two years Google has introduced a series of algorithm and filter changes that have led to unpredictable search engine results, and many clean (non-spam) websites have been dropped from the rankings. Google updates used to be monthly, and then quarterly. Now with so many servers, there seems to be several different search engine results rolling through the servers at any time during a quarter. Part of this is the recent Big Daddy update, which is a Google infrastructure update as much as an algorithm update. We believe Big Daddy is using a 64 bit architecture. Pages seem to go from a first page ranking to a spot on the 100th page, or worse yet to the Supplemental index. Google algorithm changes started in November 2003 with the Florida update, which now ranks as a legendary event in the Webmaster community. Then came updates named Austin, Brandy, Bourbon, and Jagger. Now we are dealing with the BigDaddy!

The algorithm problems seem to fall into 4 categories. There are canonical issues, duplicate content issues, the Sandbox, and supplemental page issues.

1. Canonical Issues: these occur when a search engine treats www.yourdomain.com, yourdomain.com, and yourdomain.com/index.html all as different websites. When Google does this, it then flags the different copies as duplicate content and penalizes them. Also, if the site not penalized is http://yourdomain.com, but all of the websites link to your website using www.yourdomain.com, then the version left in the index will have no ranking. These are basic issues that other major search engines, such as Yahoo and MSN, have no problem dealing with. Google is possibly the greatest search engine in the world (ranking themselves as a 10 on a scale of 1 to 10). They provide tremendous results for a wide range of topics, and yet they cannot get some basic indexing issues resolved.

2. The Sandbox: this has become one of the legends of the search engine world. It appears that websites, or links to them, are “sandboxed” for a period before they are given full rank in the index, kind of like a maturing time. Some even think it is only applied to a set of competitive keywords, because they were the ones being manipulated the most. The Sandbox existence is debated, and Google has never officially confirmed it. The hypothesis behind the Sandbox is that Google knows that someone cannot create a 100,000 page website overnight, so they have implemented a type of time penalty for new links and sites before fully making the index.

3. Duplicate Content Issues: these have become a major issue on the Internet. Because web pages drive search engine rankings, black hat SEOs (search engine optimizers) started duplicating entire sites’ content under their own domain name, thereby instantly producing a ton of web pages instantly (an example of this would be to download an Encyclopedia onto your website). Due to this abuse, Google aggressively attacked duplicate content abusers with their algorithm updates. But in the process they knocked out many legitimate sites as collateral damage. One example occurs when someone scraps your website. Google sees both sites and may determine the legitimate one to be the duplicate. About the only thing a Webmaster can do is track down these sites as they are scrapped, and submit a spam report to Google. Another big issue with duplicate content is that there are a lot of legitimate uses of duplicate content. News feeds are the most obvious example. A news story is covered by many websites because it is content the viewers want. Any filter will inevitably catch some legitimate uses.

4. Supplemental Page Issues: Webmasters fondly refer to this as Supplemental Hell. This issue has been reported on places like Webmasterworld for over a year, but a major shake up around February 23rd has led to a huge outcry from the Webmaster community. This recent shakeup was part of the ongoing BigDaddy rollout that should finish this month. This issue is still unclear, but here is what we know. Google has 2 indexes: the Main index that you get when you search, and the Supplemental index that contains pages that are old, no longer active, have received errors, etc. The Supplemental index is a type of graveyard where web pages go when they are no longer deemed active. No one disputes the need for a Supplemental index. The problem, though, is that active, recent, and clean pages have been showing up in the Supplemental index. Like a dungeon, once they go in they rarely come out. This issue has been reported with a low noise level for over a year, but the recent February upset has led to a lot of discussion around it. There is not a lot we know about this issue, and no one can seem to find a common cause leading to it.

Google updates were once fairly predictable, with monthly updates that Webmasters anticipated with both joy and angst. Google followed a well published algorithm that gives each website a Page Rank, which is a number given to each webpage based on the number and rank of other web pages pointing to it. When someone searches on a term, all of the web pages deemed relevant are then ordered by their Page Rank. Google uses a number of factors such as keyword density, page titles, meta tags, and header tags to determine which pages are relevant. This original algorithm favored incoming links and the anchor text of them. The more links you got with an anchor text, the better you ranked for that keyword.

As Google gained the bulk of internet searches in the early part of the decade, ranking well in their engine became highly coveted. Add to this the release of Google’s Adsense program, and it became very lucrative. If a website could rank high for a popular keyword, they could run Google ads under Adsense and split the revenue with Google! This combination led to an avalanche of SEO’ing like the Webmaster world had never seen. The whole nature of links between websites changed. Websites used to link to one another because it was good information for their visitors. But now that link to another website could reduce your search engine rankings, and if it is a link to a competitor, it might boost his.

In Google’s algorithm, links coming into your website boost the site’s Page Rank (PR), while links from your web pages to other sites reduce your PR. People started creating link farms, doing reciprocal link partnerships, and buying/selling links. Webmasters started linking to each other for mutual ranking help or money, instead of quality content for their visitors. This also led to the wholesale Scraping of websites. Black hat SEO’s will take the whole content of a website, put Google’s ad on it, get a few high powered incoming links, and the next thing you know they are ranking high in Google and generating revenue from Google’s adsense without providing any unique website content. Worse yet, as Google tries to go after this duplicate content, they sometimes get the real company instead of the scraper.

This is all part of the cat and mouse game that has become the Google algorithm. Once Google realized the manipulation that was happening, they decided to aggressively alter their algorithms to prevent it. After all, their goal is to find the most relevant results for their searchers. At the same time, they also faced huge growth with the internet explosion. This has led to a period of unstable updates, causing many top ranking websites to disappear while many spam and scrapped websites remain. In spite of Google’s efforts, every change seems to catch more quality websites. Many spam sites and websites that violate Google’s guidelines are caught, but there is an endless tide of more spam websites taking their place.

Some people might believe that this is not a problem. Google is there to provide the best relevant listings for what people are searching on, and for the most part the end user has not noticed an issue with Google’s listings. If they only drop thousands of listings out of millions, then the results are still very good. These problems may not be affecting Google’s bottom line now, but having a search engine that cannot be evolved without producing unintended results will hurt them over time in several ways. First, as the competition from MSN and Yahoo grows, having the best results will no longer be a given, and these drops in quality listings will hurt them.

Next, to stay competitive Google will need to continue to change their algorithms. This will be harder if they cannot make changes without producing unintended results. Finally, having the Webmaster community lose faith in them will make them vulnerable to competition. Webmasters provide Google with two things. They are the word of mouth experts. Also, they run the websites that use Google’s Adsense program. Unlike other monopolies, it is easy to switch search engines. People might also criticize Webmasters for relying on a business model that requires free search engine traffic. Fluctuations in ranking are part of the internet business, and most Webmasters realize this. Webmasters are simply asking Google to fix bugs that cause unintended issues with their sites.

Most Webmasters may blame ranking losses on Google and their bugs. But the truth is that many Webmasters do violate some of the guidelines that Google lays out. Most consider it harmless to bend the rules a little, and assume this is not the reason their websites have issues. In some cases, though, Google is right and has just tweaked its algorithm in the right direction. Here is an example: Google seems to be watching the incoming links to your site to make sure they don’t have the same anchor text (this is the text used in the link on the website linking to you). If too many links use the same anchor text, Google discounts these links. This was originally done by some people to inflate their rankings. Other people did it because one anchor text usually makes sense. This is not really a black hat SEO trick, and it is not called out in Google’s guidelines, but it has caused some websites to lose rank.

Webmasters realize that Google needs to fight spam and black hat SEO manipulation. And to their credit, there is a Google Engineer named Matt Cutts who has a Blog site and participates in SEO forums to assist Webmasters. But given the revenue impact that Google rankings have on companies, Webmasters would like to see even more communication around the known issues, and help with identifying future algorithm issues. No one expects Google to reveal their algorithm or what changes they are making. Rumor on the forum boards speculates that Google is currently looking at items like the age of the domain name, websites on the same ip, and frequency of fresh content. It would be nice from a Webmaster standpoint to be able to report potential bugs to Google, and get a response. It is in Google’s best interest to have a bug free algorithm. This will in turn will provide the best search engine results for everyone.

About the Author: Rachel Ringler is Vice President of Advantage1 Web Services, Inc., which owns a network of Web Hosting Informational Websites including Hostchart.com, Resellerconnection.com, and Foundhost.com.

Article Source: http://EzineArticles.com/?expert=Rachel_Ringler