Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Yes, I agree with you on using a 2-tier search which will increase the relevancy to optimize the quality of search. And Google's search quality is not unbeatable.

I also agree with you by using distributed sites to optimize the results locally. Actually what I proposed is to make the distributed search from both geographic location and vertical market point of view, as opposed of dedicated sites from you. But they are complementary. The dedicated sites definitely will provide better and more relevant results than a global search engine if Google search was not limited to a particular site.

However, the only thing I don't agree with you is when you said it does not have to be distributed though, the search router can integrate the algorithms from the dedicated sites. Then I think it's not quite feasible since it's not possible for Stack Overflow or Wikipedia to share their algorithms with DDG.

Let me know if I misunderstood you. If you'd like to take if offline, I'll be happy to discuss with you via email. See my profile.



Why it's not feasible to share Stack Overflow algorithms with DDG? they just give them the application to run it locally.


What? DDG does not own Stack Overflow but it can run Stack Overflow's algorithm locally? Do you mean that Stack Overflow has it's search algorithm open sourced? I'm sorry, I don't quite get it.

Usually, the router search engine queries data from the second tier websites to get high quality results without having other websites' algorithms. Also there is another problem: how do you know which websites to go for given an arbitrary keywords? For example, when user searches for "cookie" on your search engine, where do you send the query to? How do you know if they are looking for food cookie or browser cookie?


The issue about sharing your algorithms is minimal in this case. Most search engines are using standard frameworks. Google doesn't share their algorithms but StackOverflow doesn't add any specific magic to it.

Regarding how do you know where to route a query, it is an issue but not so great in this case. The article doesn't talk about having a two tiered search for every web site. If it has a two tiered search for the top 100 sites that is enough to challenge Google (the main point of the article) and making 100 searches and filtering them in the 2nd tier it's not difficult.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: