Google Algorithm Problems

Google Algorithm Problems
By Rodney Ringler (c) 2006.
Posted on this blog by Afzal Khan

Have you noticed anything different with Google lately? The
Webmaster community certainly has, and if recent talk on several
search engine optimization (SEO) forums is an indicator,
Webmasters are very frustrated. For approximately two years
Google has introduced a series of algorithm and filter changes
that have led to unpredictable search engine results, and many
clean (non-spam) websites have been dropped from the rankings.
Google updates used to be monthly, and then quarterly. Now with
so many servers, there seems to be several different search
engine results rolling through the servers at any time during a
quarter. Part of this is the recent Big Daddy update, which is a
Google infrastructure update as much as an algorithm update. We
believe Big Daddy is using a 64 bit architecture. Pages seem to
go from a first page ranking to a spot on the 100th page, or
worse yet to the Supplemental index. Google algorithm changes
started in November 2003 with the Florida update, which now
ranks as a legendary event in the Webmaster community. Then came
updates named Austin, Brandy, Bourbon, and Jagger. Now we are
dealing with the BigDaddy!

The algorithm problems seem to fall into 4 categories. There are
canonical issues, duplicate content issues, the Sandbox, and
supplemental page issues.

1. Canonical Issues: These occur when a search engine
treats www.yourdomain.com, yourdomain.com, and yourdomain.com/index.html
all as different websites. When Google does this, it then flags
the different copies as duplicate content and penalizes them.
Also, if the site not penalized is http://yourdomain.com, but
all of the websites link to your website using www.yourdomain.com,
then the version left in the index will have no ranking. These
are basic issues that other major search engines, such as Yahoo
and MSN, have no problem dealing with. Google is possibly the
greatest search engine in the world (ranking themselves as a 10
on a scale of 1 to 10). They provide tremendous results for a
wide range of topics, and yet they cannot get some basic indexing
issues resolved.

2. The Sandbox: This has become one of the legends of
the search engine world. It appears that websites, or links to them,
are "sandboxed" for a period before they are given full rank in the
index, kind of like a maturing time. Some even think it is only
applied to a set of competitive keywords, because they were the
ones being manipulated the most. The Sandbox existence is
debated, and Google has never officially confirmed it. The
hypothesis behind the Sandbox is that Google knows that someone
cannot create a 100,000 page website overnight, so they have
implemented a type of time penalty for new links and sites
before fully making the index.

3. Duplicate Content Issues: These have become a major
issue on the Internet. Because web pages drive search engine rankings,
black hat SEOs (search engine optimizers) started duplicating
entire sites' content under their own domain name, thereby
instantly producing a ton of web pages (an example of this would
be to download an Encyclopedia onto your website). As a result
of this abuse, Google aggressively attacked duplicate content
abusers with their algorithm updates. But in the process they
knocked out many legitimate sites as collateral damage. One
example occurs when someone scrapes your website. Google sees
both sites and may determine the legitimate one to be the
duplicate. About the only thing a Webmaster can do is track down
these sites as they are scraped, and submit a spam report to
Google. Another big issue with duplicate content is that there
are a lot of legitimate uses of duplicate content. News feeds
are the most obvious example. A news story is covered by many
websites because it is content the viewers want. Any filter will
inevitably catch some legitimate uses.

4. Supplemental Page Issues: Webmasters fondly refer to
this as Supplemental Hell. This issue has been reported on places like
WebmasterWorld for over a year, but a major shake up around
February 23rd has led to a huge outcry from the Webmaster
community. This recent shakeup was part of the ongoing BigDaddy
rollout that should finish this month. This issue is still
unclear, but here is what we know. Google has 2 indexes: the
Main index that you get when you search, and the Supplemental
index that contains pages that are old, no longer active, have
received errors, etc. The Supplemental index is a type of
graveyard where web pages go when they are no longer deemed
active. No one disputes the need for a Supplemental index. The
problem, though, is that active, recent, and clean pages have
been showing up in the Supplemental index. Like a dungeon, once
they go in, they rarely come out. This issue has been reported
with a low noise level for over a year, but the recent February
upset has led to a lot of discussion around it. There is not a
lot we know about this issue, and no one can seem to find a
common cause leading to it.

Google updates were once fairly predictable, with monthly
updates that Webmasters anticipated with both joy and angst.
Google followed a well published algorithm that gave each
website a Page Rank, which is a number given to each webpage
based on the number and rank of other web pages pointing to it.
When someone searches on a term, all of the web pages deemed
relevant are then ordered by their Page Rank.

Google uses a number of factors such as keyword density, page
titles, meta tags, and header tags to determine which pages are
relevant. This original algorithm favored incoming links and the
anchor text of them. The more links you got with an anchor text,
the better you ranked for that keyword. As Google gained the
bulk of internet searches in the early part of the decade,
ranking well in their engine became highly coveted. Add to this
the release of Google's Adsense program, and it became very
lucrative. If a website could rank high for a popular keyword,
they could run Google ads under Adsense and split the revenue
with Google!

This combination led to an avalanche of SEO'ing like the
Webmaster world had never seen. The whole nature of links between
websites changed. Websites used to link to one another because
it was good information for their visitors. But now that link to
another website could reduce your search engine rankings, and if
it is a link to a competitor, it might boost his. In Google's
algorithm, links coming into your website boost the site's Page
Rank (PR), while links from your web pages to other sites reduce
your PR. People started creating link farms, doing reciprocal
link partnerships, and buying/selling links. Webmasters started
linking to each other for mutual ranking help or money, instead
of quality content for their visitors. This also led to the
wholesale scraping of websites. Black hat SEO's will take the
whole content of a website, put Google's ad on it, get a few
high powered incoming links, and the next thing you know they
are ranking high in Google and generating revenue from Google's
Adsense without providing any unique website content.

Worse yet, as Google tries to go after this duplicate content,
they sometimes get the real company instead of the scraper. This
is all part of the cat and mouse game that has become the Google
algorithm. Once Google realized the manipulation that was
happening, they decided to aggressively alter their algorithms
to prevent it. After all, their goal is to find the most
relevant results for their searchers. At the same time, they
also faced huge growth with the internet explosion. This has led
to a period of unstable updates, causing many top ranking
websites to disappear while many spam and scraped websites
remain. In spite of Google's efforts, every change seems to
catch more quality websites. Many spam sites and websites that
violate Google's guidelines are caught, but there is an endless
tide of more spam websites taking their place.

Some people might believe that this is not a problem. Google is
there to provide the best relevant listings for what people are
searching on, and for the most part the end user has not noticed
an issue with Google's listings. If they only drop thousands of
listings out of millions, then the results are still very good.
These problems may not be affecting Google's bottom line now,
but having a search engine that cannot be evolved without
producing unintended results will hurt them over time in several
ways.

First, as the competition from MSN and Yahoo grows, having
the best results will no longer be a given, and these drops in
quality listings will hurt them. Next, to stay competitive
Google will need to continue to change their algorithms. This
will be harder if they cannot make changes without producing
unintended results. Finally, having the Webmaster community lose
faith in them will make them vulnerable to competition.
Webmasters provide Google with two things. They are the word of
mouth experts. Also, they run the websites that use Google's
Adsense program. Unlike other monopolies, it is easy to switch
search engines. People might also criticize Webmasters for
relying on a business model that requires free search engine
traffic. Fluctuations in ranking are part of the internet
business, and most Webmasters realize this. Webmasters are
simply asking Google to fix bugs that cause unintended issues
with their sites.

Most Webmasters may blame ranking losses on Google and their
bugs. But the truth is that many Webmasters do violate some of
the guidelines that Google lays out. Most consider it harmless
to bend the rules a little, and assume this is not the reason
their websites have issues. In some cases, though, Google is
right and has just tweaked its algorithm in the right direction.
Here is an example: Google seems to be watching the
incoming links to your site to make sure they don't have the same anchor
text (this is the text used in the link on the website linking
to you). If too many links use the same anchor text, Google
discounts these links. This was originally done by some people
to inflate their rankings. Other people did it because one
anchor text usually makes sense. This is not really a black hat
SEO trick, and it is not called out in Google's guidelines, but
it has caused some websites to lose rank.

Webmasters realize that Google needs to fight spam and black
hat SEO manipulation. And to their credit, there is a Google
Engineer named Matt Cutts who has a Blog site and participates
in SEO forums to assist Webmasters. But given the revenue impact
that Google rankings have on companies, Webmasters would like to
see even more communication around the known issues, and help
with identifying future algorithm issues. No one expects Google
to reveal their algorithm or what changes they are making. Rumor
on the forum boards speculates that Google is currently looking
at items like the age of the domain name, websites on the same
IP, and frequency of fresh content. It would be nice from a
Webmaster standpoint to be able to report potential bugs to
Google, and get a response. It is in Google's best interest to
have a bug free algorithm. This will in turn provide the best
search engine results for everyone.

==============================================
Rodney Ringler is President of Advantage1 Web Services, Inc.,
which owns a network of Web Hosting Informational Websites
including Hostchart.com (http://www.hostchart.com),
Resellerconnection.com (http://www.resellerconnection.com),
Foundhost.com (http://www.foundhost.com) and Resellerforums.com
(http://www.resellerforums.com).
==============================================

"Google: The "Big Daddy' of the Blogging World"

Article Courtesy By Merle http://MCPromotions.com
Posted on this Blog by Afzal Khan

It seems like every time you turn your head these
days you hear something about Google. Definitely
a force to be reckoned with, Google is king online.
Although they're known mostly as a search engine,
they provide many other helpful services to today's
website owner.

Their non-search services include Blogger, Google
Adsense,Adwords, Google Desktop, and more. If you
use one or more of these services, Google probably
publishes a helpful blog to go with it.

Let's examine a few:

1) Google Adsense: http://adsense.blogspot.com

Many website owners use Google's Adsense to display
ads on their websites and generate a monthly income
in exchange for their efforts. This blog is a "look
inside Google Adsense." Who better to learn the ins
and outs of Adsense from than from Google's employees
themselves?

You'll pick up plenty of tips to help improve your
click thrus and learn about many Adsense features
you may not fully understand. If you're an Adsense
user, you need to subscribe to this page.

2) Google SiteMaps: http://sitemaps.blogspot.com

Google created "Sitemaps" for website owners to
supply information about their sites so that Google
can easily crawl through and index their pages. A
sitemap gives Google more complete information about
your website that they may miss when doing a regular
index visit.

This blog covers the "nitty gritty" of sitemaps,
how they work and things you'll need to pay attention
to when creating your files. If you want to fully
understand the program and how to
use it to your best advantage, stop on over.

3) Google Adwords: http://adwords.blogspot.com

Adwords is Google's pay per click ad program. Your
text ads are displayed both on search engine results
pages and other websites when someone searches for
the keywords/phrases that you have placed bids on.

If you use Adwords, you can really learn a lot
about how to improve your conversions and get
more for your dollars here.

4) Google Reader: http://googlereader.blogspot.com

Google Reader was designed to make it easier to
keep up with all of the information you like
to read online. Using it can help you get all of the
latest updates for your favorite websites.

This blog will keep you posted on the latest
reader updates, any known bug problems and tips
for getting more out of Reader.

5) Blogger http://buzz.blogger.com

I love Blogger. It's the fastest, easiest way to
start your own blog at no charge -- and it's owned
by Google, who publish their own blog to help you get
the most out of blogging at Blogger. Read it for
fun stories and news of the latest
add ons and enhancements from Google's Blogger
team.

6) Google Video: http://googlevideo.blogspot.com

Google Video allows anyone to upload videos to
their servers and others can go and watch them.
This blog showcases some of those videos and
upcoming special event videos.

7) Google's Blog: http://googleblog.blogspot.com

Google is like a small country and this blog will
give you a glimpse into their culture. Posts are
done by various employees in various departments.
You'll learn a lot about Google's products and other
technology news. If you're obsessed with Google,
this one should be on your "must see" list.

8) Adwords API Blog: http://adwordsapi.blogspot.com

This free Google service allows developers to
engineer computer programs that interact directly
with the Adwords server. You might say it's for "tech
heads." If you're a "geek" you'll want to hang
your hat here.

9)Google Desktop: http://googledesktop.blogspot.com

Google Desktop is downloadable software you can
use to quickly find files and other information
on your computer. This blog will tell you about
updates, available plug-ins and more.

10) Google Base: http://googlebase.blogspot.com/

Google Base is a place where you can post all types
of content and have it show up on Google.
This Blog will enlighten you on how best to use
it, and even showcases others having success with
Google's latest service.

It can be a full time job just trying
to keep up with them all the delightful
services Google provides. By bookmarking their
blogs, and/or subscribing to their feeds, you too
can become an "expert" on all things Google.


=========================================
By Merle- Want to Know the SECRETS of Article
Promotion? Discover everything you need to know
in this brand New Ebook, "How to Use Articles to
Drive Website Traffic". Get your F-r-e-e Copy now at
http://articleannouncer.mcpromotions.com

An SEO Checklist

Article Courtesy by Google Page Rank.net

Posted on this blog by Afzal Khan

Search engine optimization is on every webmaster's mind these days. Achieving a favorable ranking for the right keywords can mean a steady stream of targeted traffic to your site, and all for free - that's hard to beat. The key to high search engine rankings is structuring your website correctly, including plenty of content that is relevant to your keywords, and making sure your website is spider-friendly. You can use this checklist to make sure all of your Web pages can be found, indexed and ranked correctly:

Your website is themed. Your site deals with an identifiable theme which is obvious from the text on the home page and reinforced by all the other pages on your site. In other words, all the individual Web pages relate to each other and deal with various aspects of some central theme. The text on your home page should state clearly what that theme is and what your website is about, and the other pages should reinforce that.

Your Web pages have enough high quality, relevant content. Spiders come to your website looking for content. If a page doesn't have much content, or the content doesn't appear closely related to the page's title and your website's theme, the page probably won't be indexed or if it is indexed it won't rank well. Search engines love quality content and lots of it - content is what Web searchers are looking for and search engines try to provide.

Your website's navigational structure is relatively flat. You don't want important pages to be too "deep" within your website, meaning it takes several clicks to get there from the home page. Search engines typically index the home page first, then gradually index other pages on a site over time. Many spiders are programmed to only go three layers deep - if some of your important content is buried deeper than that, it may never be found and indexed at all.

You've created a unique "Title" tag for each page. The title is one of the most important aspects of any Web page from an SEO standpoint, especially for Google (which is the most important search engine to optimize for). Don't use a generic title for all your pages, use the keywords your targeting for that page and keep it brief but descriptive.

You use the "Description" meta tag. Contains a highly descriptive sentence about the content and purpose of your page, and contains your most important keyword phrase early in the sentence. Not all of the search engines will display this "canned" description when they list the page in search results, but many of them will, so it's worth getting it right. Check out our Webmaster Tools for an easy way to create great meta tags.

You use the "Keywords" meta tag. As with the meta tag description, not every search engine will use the keywords meta tag. But some will use it and none will penalize you for having it. Also, having a short list of the keywords you're targeting will help you write appropriate content for each page. The keyword tage should contain your targeted keyword phrase and common variations, common misspellings and related terms. Make sure your keywords relate closely to the page content and tie into the overall theme of your site.

Your keywords are included in the visible page content, preferably high up on the page. You have to achieve a balance here - you want to include keyword phrases (and variations) a number of times within your text, but not so many times that you appear to be guilty of "keyword stuffing". The trick is to work the keywords into the text so that it reads as naturally as possible for your site visitors. Remember, you can incorporate keywords into any Web page element that is potentially viewable by site visitors - header text, link text and titles, table captions, the "Alt" attribute of the image tag, the "title" attribute of the link tag, etc.

Every page of your website can be reached by search engine spiders. This is critical - if your pages can't be found, they can't be indexed and included in search results, let alone rank well. Search engines use spiders to explore your website and index the pages, so every page must be accessible by following text links. If pages require a password to view, are generated by a script in response to a query, or have a long and complicated URL, spiders may not be able to read them. You need to have simple text links to the pages you want indexed.

You've included a site map. Unless your site is very small, it's a good idea to create a site map with text links that you link to the site map from your home page. In addition to a link, include descriptive text for containing the relevant keywords for each page.

You link to your most important pages from other pages on your site. Internal links help determine page rank since they show which pages of your site are most important. The more links you have to have to a page, relative to other pages on your site, the more importance search engines will assign to it.

You use keywords in your link text. When you create a text link to another page on your site, use that page's targeted keywords as the text for the link (inside the anchor tags that create the link). Make it as descriptive as possible. For example, a link that says "Premium Customized Widgets" is much better than one that says simply "Product Page", and indicates to search engine spiders what that linked page is about.

Your site doesn't use frames. If possible, don't use frames on any page you want to get indexed by search engines. If you feel you simply must use frames for a page, then also make use of the "noframes" HTML tags to provide alternative text that spiders can read (and make that text descriptive rather than just a notice that "This site uses frames etc. etc.").

You don't use automatic page redirects. Don't make any pages automatically redirect the visitor to another page (the exception is a page you've deleted for good - in which case you should use a "301 redirect", a permanent redirect which is acceptable to search engines).

Your important content is in plain text and not contained in images. Search engine spiders can't "read" content in JPEG, GIF, or PNG files. If you really feel that using an image rather than text is crucial to your design, at least put the same text in the image's "Alt" tag (or in the "title" tag if you're using the image as a hyperlink).

Your important content is not contained in Flash files. Flash is a wonderful technology, but unfortunately spiders don't have the required "plugin" to view Flash files. As a result, Flash content is mostly inaccessible to search engine spiders. Some can find and follow hyperlinks within the Flash file, but unless those links lead to pages with readable HTML content this won't help you much. Don't create all-Flash pages for any content you want to get indexed - instead, put that content in the HTML portion of the page.

Links and keywords are not hidden inside JavaScript code. If your links use JavaScript to direct the user to the appropriate page (for instance, a drop-down list) or important content is contained within JavaScript code (when it's displayed dynamically using DHTML, for instance) search engine spiders won't be able to "see" it. You can, however, use the "noscript" HTML tags to provide an alternative that can be read by spiders.

You've optimized every important page of your website individually. Don't stop at your home page. Take the trouble to optimize any page which has a reasonable chance of being indexed by the major search engines, targeting appropriate keywords for each. If you face a lot of competition it may be nearly impossible to get a top ranking for your home page, but you can still get a lot of search engine traffic to your site from other pages which are focused on very specific keyword phrases.

You didn't duplicate content. Each page of your site should have unique content that distinguishes it from every other page on your site. Duplicating content or having pages that are only slightly different might be seen as "search engine spamming" (trying to manipulate search engine results).

You provide linking instructions for those who want to link to your site. Somewhere on your site state your policies about other people linking to your site and provide the wording you'd like them to use in their link. You want to encourage other people to link to your site, preferably using link text and a description that reflect the keywords for that page. For their convenience provide the ready-made HTML code for the link - not everyone will use it, but most often they will use your preferred text as a courtesy as long as it is truly descriptive of your site and doesn't contain "marketing hype".

You provide linking instructions for those who want to link to your site. Somewhere on your site state your policies about other people linking to your site and provide the wording you'd like them to use in their link. You want to encourage other people to link to your site, preferably using link text and a description that reflect the keywords for that page. For their convenience provide the ready-made HTML code for the link - not everyone will use it, but many will use your preferred text as a courtesy as long as it doesn't contain "marketing hype".

Important hyperlinks are plain text links and not image links or image maps. Text links are better from an SEO standpoint than image links, as spiders can't read text from an image file. If you feel you really must use a graphic as a link, at least include a text description which (including the relevant keywords) by using the "title" attribute of the link tag.

Your website is free of coding errors and broken links. HTML coding errors and non-working links can keep search engine spiders from correctly reading and indexing your pages. For that reason, it's a good idea to use a Web page validation utility to check your HTML code to make sure it's error-free.