Giving Search a Human Touch

Howdy Folk's

Howaz the New Year Celebration's,Enjoyed!!! hmm..... well wish you n your family a prosperous New Year 2007 ahead to bring joy in ur life!!!

Now coming to posting, after having a hang over of new year party n Eid celebration it's all luking very dumbo to me, still managed to read blog's and other intresting articles which I regularly do without fail & Blogging here with an great article found today during search! "Giving Search a Human Touch" How does it sound's, kewl isn't so......But to hear Jimmy Wales, co-founder of Wikipedia and chairman of the for-profit wiki site Wikia, describe his vision of a totally transparent social search engine -- one built with open-source software and inspired by the collaborative spirit of wikis -- you realize that his plan just might work.

Read below the article written by reporter Michael Calore, Editor of Webmonkey and you can find the original copy of this article at Wired.com

However as usual: Article posted on this blog by Afzal Khan

Giving Search a Human Touch

The idea of building a better search engine sounds almost laughable on the surface.
After all, isn't there already a massively successful internet search player with a seemingly insurmountable market share? But to hear Jimmy Wales, co-founder of Wikipedia and chairman of the for-profit wiki site Wikia, describe his vision of a totally transparent social search engine -- one built with open-source software and inspired by the collaborative spirit of wikis -- you realize that his plan just might work.

Wales' plan for the Search Wikia project is to put ordinary users in charge of ranking search results. Heavy lifting such as indexing and raw ranking will still be done by machines, but the more nuanced work of deciding how search results are displayed will be completed by humans.

Google, the current King of Search, ranks search results based on the perceived trust of the web community at large -- the more links a page receives, the more it's trusted as an authoritative source of information, and the higher the rank. However, this method is open to tinkering, trickery and hacks, all of which damage the relevancy of results.

If successful, Wales' project, which launches in early 2007, will be able to filter out such irrelevant results. Operating much the same way as Wales' Wikipedia, both the software algorithms powering Search Wikia and the changes applied by the community will be made transparent on the project's website.

Wired News spoke to Jimmy Wales about Search Wikia. We discussed the ins and outs of how the model will likely work, what it will take to build it, and what sorts of criticisms it will face.

Wired News: Can you describe the new search engine in your own words?

Jimmy Wales: The core of the concept is the open-source nature of everything we're intending to do -- making all of the algorithms public, making all of the data public and trying to achieve the maximum possible transparency. Developers, users, or anyone who wants to can come and see how we're doing things and give us advice and information about how to make things better.

Additionally, we want to bring in some of the concepts of the wiki model -- building a genuine community for discussion and debate to add that human element to the project.

I mention "community" to distinguish us as something different. A lot of times, when people talk about these kinds of (projects), they're not thinking about communities. They're thinking about users randomly voting, and that action turning into something larger. I really don't like the term "crowdsourcing." We're really more about getting lots of people engaged in conversations about how things should be done.

WN: How are the communities going to be managed?

Wales: I don't know! (laughter) If you asked me how the Wikipedia community is managed, I wouldn't know the answer to that, either. I don't think it makes sense to manage a community.

It's about building a space where good people can come in and manage themselves and manage each other. They can have a distinct and clear purpose -- a moral purpose -- that unites people and brings them together to do something useful.

WN: How will the human-powered ranking element work?

Wales: We don't know. That's something that's really very open-ended at this moment. It's really up to the community, and I suspect that there won't be a one-size-fits-all answer. It will depend on the topic and the type of search being conducted.

One of the things that made Wikipedia successful was a really strong avoidance of a priori thinking about exactly "how." We all have a pretty good intuitive sense of what a good search result is. A variety of different factors make a search result "good," qualitatively speaking. How we get those kinds of results for the most possible searches depends on a lot of factors.

A lot of the earlier social search projects fell apart because they were committed a priori to some very specific concept of how it should work. When that worked in some cases but not others, they were too stuck in one mold rather than seeing that a variety of approaches depending on the particular topic is really the way to do it.

WN: I'm envisioning that Wikia Search will incorporate some sort of voting system, and that users will be able to adjust and rank lists of results. Is this the case?

Wales: Yes, but how exactly and under what circumstances that would work is really an empirical question that we'll experiment with. At Wikipedia and in the wiki world, one of the things we've always pushed hard against is voting. Voting is usually not the best way to get a correct answer by consensus. Voting can be gamed, it can be played with. It's a crutch of a tool that you can use when you don't have anything better to use. Sometimes, there is no better way. You have to say, "We've tried to get a consensus and we couldn't, so we took a vote."
In general, envisioning some sort of pre-built algorithm for counting people's votes is just not a good idea.

WN: Speaking of gaming, what methodologies do you think Search Wikia will employ to fight gaming?

Wales: I think the most important thing to use to fight against gaming is genuine human community. Those kinds of gaming behaviors pop up when there is an algorithm that works in some mechanical way, and then people find a way to exploit it. It's pretty hard to do that within a community of people who know each other. Basically, if you're being a jerk, they'll tell you knock it off and you'll be blocked from the site. It's pretty simple for humans to see through that sort of thing. The real way to fight it is to have a group of people who trust each other, with that trust having been built over a period of time.

WN: Will there be some sort of validation that happens when results are ranked by users? Will knowledgeable contributors get the chance to vet changes?

Wales: Yes. The keys of good design here have to do with transparency -- everybody can see what everyone else has done. The communities will have the ability to effect and modify changes as they see fit.

WN: What forms of open-source software are you applying to this search project, and why do you think those would be more successful than proprietary search software?

Wales: Here's the main thing. If we publish all the software -- and we'll be starting with Lucene and Nutch, which are these open source projects that are out there and already quite good -- and do all of our modifications transparently in public, then other programmers can come and check the code. If you see things that aren't working well, you can contribute. People who are coders can contribute in one way, and ordinary people using the site can also contribute in other ways.

It's mostly about the trust that you get from that transparency. You can see for yourself, if you choose to dig into it, how things are ranked and why certain results are ranked the way they are. You can also choose to download the whole thing and do tests or tweak it to make it better in certain areas. That kind of transparency helps if you see a problem with search in some area that you care about, like some technical field for example. There's no good way for you to go and tell Google that their search is broken in this area, or that they need to disambiguate these terms -- or whatever.

By having an up-front commitment to transparency, I think you can do that.

WN: One of the key arguments in favor of a new search model is that traditional search engines like Google are subjected to spam more and more often. How can a wiki-powered search engine better fight search spam?

Wales: Again, I think it's that human element. Humans can recognize that a domain is not returning good results, and if you have a good community of people to discuss it, you can just kick them out of the search engine. It seems pretty simple to me -- it's an editorial judgment. You just have to have a broad base of people who can do that.

WN: How are you going to build this broad base? Will there be an outreach, or are you expecting people to just come to you?

Wales: I think people will come. If we're doing interesting work and people find it fun, then people will come.

WN: When do you expect to see Search Wikia up and running?

Wales: The project to build the community to build the search engine is launching in the first quarter of 2007, not the search engine itself. We may have something up pretty quickly, maybe some sort of demo or test for people to start playing with. But we don't want to build up expectations that people can come in three months and check out this Google-killing search engine that we've written from scratch. It's not going to happen that fast.

What we want to do now is get the community going and get the transparent algorithms going so we can start the real work. It's going to be a couple of years before this really turns into something interesting.

Waiting for your feedbacks about this news. You can mail me by clicking on this link Afzal Khan.

Article to know in depth about how to remove/fix Supplemental Results from Google search results

Steveb of webmasterworld has an excellent posting on how to remove supplement results, I agree 100% with what he says and I recommend his posting to everyone who have supplement results in google and want to remove them, Supplement results are mostly caused when a page of a site once existed and later removed by the site owner of because of any other problem, Supplement results are also caused when a page which is crawled once had links to it then the links dropped off completely.

Article posted on this blog by Afzal Khan.

Here is his posting

"Google's ill-advised Supplemental index is polluting their search results in many ways, but the most obviously stupid one is in refusing to EVER forget a page that has been long deleted from a domain. There are other types of Supplementals in existence, but this post deals specifically with Supplemental listings for pages that have not existed for quite some time.
The current situation: Google refuses to recognize a 301 of a Supplemental listing. Google refuses to delete a Supplemental listing that is now a nonexistent 404 (not a custom 404 page, a literal nothing there) no matter if it is linked to from dozens of pages. In both the above situations, even if Google crawls through links every day for six months, it will not remove the Supplemental listing or obey a 301. Google refuses to obey its own URL removal tool for Supplementals. It only "hides" the supplementals for six months, and then returns them to the index.
As of the past couple days, I have succeeded (using the below tactics) to get some Supplementals removed from about 15% of the datacenters. On the other 85% they have returned to being Supplemental however.
Some folks have hundreds or thousands of this type of Supplemental, which would make this strategy nearly impossible, but if you have less than twenty or so...
1) Place a new, nearly blank page on old/supplemental URL.
2) Put no actual words on it (that it could ever rank for in the future). Only put "PageHasMoved" text plus link text like "MySiteMap" or "GoToNewPage" to appropriate pages on your site for a human should they stumble onto this page.
3) If you have twenty supplementals put links on all of them to all twenty of these new pages. In other words, interlink all the new pages so they all have quite a few links to them.
4) Create a new master "Removed" page which will serve as a permanent sitemap for your problem/supplemental URLs. Link to this page from your main page. (In a month or so you can get rid of the front page link, but continue to link to this Removed page from your site map or other pages, so Google will continually crawl it and be continually reminded that the Supplementals are gone.)
5) Also link from your main page (and others if you want) to some of the other Supplementals, so these new pages and the links on them get crawled daily (or as often as you get crawled).
6) If you are crawled daily, wait ten days.
7) After ten days the old Supplemental pages should show their new "PageHasMoved" caches. If you search for that text restricted to your domain, those pages will show in the results, BUT they will still ALSO continue to show for searches for the text on the ancient Supplemental caches.
8) Now put 301s on all the Supplemental URLs. Redirect them too either the page with the content that used to be on the Supplemental, or to some page you don't care about ranking, like an "About Us" page.
9) Link to some or all of the 301ed Supplementals from your main page, your Removed page and perhaps a few others. In other words, make very sure Google sees these new 301s every day.
10) Wait about ten more days, longer if you aren't crawled much. At that point the 15% datacenters should first show no cache for the 301ed pages, and then hours later the listings will be removed. The 85% datacenters will however simply revert to showing the old Supplemental caches and old Supplemental listings, as if nothing happened.
11) Acting on faith that the 15% datacenters will be what Google chooses in the long run, now use the URL removal tool to remove/hide the Supplementals from the 85% datacenters.
Will the above accomplish anything? Probably not. The 85% of the datacenters may just be reflecting the fact that Google will never under any circumstances allow a Supplemental to be permanently removed. However, the 15% do offer hope that Google might actually obey a 301 if brute forced.
Then, from now on, whenever you remove a page be sure to 301 the old URL to another one, even if just to an "About Us" page. Then add the old URL to your "Removed" page where it will regularly be seen and crawled. An extra safe step could be to first make the old page a "PageHasMoved" page before you redirect it, so if it ever does come back as a Supplemental, at least it will come back with no searchable keywords on the page.
Examples of 15% datacenter: 216.239.59.104 216.239.57.99 64.233.183.99 Examples of 85% datacenter: 216.239.39.104 64.233.161.99 64.233.161.105 "


Regard's

Afzal Khan

Article posted on this blog by Afzal Khan. Actual source of this article has been picked from Search Engine Genie Blog. I recommend all my reader of Toprankseo Blog to visit links at my favourite blogs.

Importance of Sitemap Page in your website

Howdy Folks,

Hope you people are rocking in your life!!! Well it’s being long I have not written any new fresh SEO article. Today I finally decided to come up with new topic, picking some time from my busy schedule.

Today I am going to talk about the hottest topic now a day – Sitemap. About the importance of Sitemap to rank well in search engines there are many SEO tips and tricks that help in optimizing a site but one of those, the importance of which is sometimes underestimated is sitemaps.

Sitemap, as the name simply speaks of is like a map of your website – i.e. on one single page you show the structure of your site, its sections, the links between them, etc. Sitemap helps in making navigation easier for your site and keeping an updated Sitemap on your site is fruitful both for your users and for search engines. It is an important way of communication with search engines. By providing sitemap page to your site you tell search engines where you’d like them to go, while in robots.txt you tell search engine which parts of your site to exclude from indexing.

Sitemap have always been part of best Web design practices but with the adoption of sitemaps by search engines, now they become even more important. However, it is necessary to make a clarification that if you are interested in sitemaps mainly from a SEO point of view; you can't go on with the conventional sitemap only (though currently Yahoo! and MSN still keep to the standard html format). For instance, Google Sitemaps uses a special (XML) format that is different from the ordinary html sitemap for human visitors.

One might ask why two sitemaps are necessary. The answer is obvious - one is for humans, the other is for spiders (for now mainly Googlebot but it is reasonable to expect that other crawlers will join the club shortly). In that relation it is necessary to clarify that having two sitemaps is not regarded as duplicate content. In 'Introduction to Sitemaps', Google explicitly states that using a sitemap will never lead to penalty for your site.

Do check Toprank SEO Blog for other feature articles or mail me at afzal.bpl@gmail.com for other SEO articles which you would love to know in detail.