Of course you can fight spam. And the spammers can fight back. I prefer algorithms that don't require an arms race with spammers.
There are other problems with using query frequency. What about all the legitimate users that type "google" or "facebook" into the query box instead of into the location bar? What about the frequent queries that don't match anything on your site? If an algorithm needs that many patches, it is fundamentally a weak approach. wunder On Sep 20, 2011, at 4:11 PM, Markus Jelsma wrote: > A query log parser can be written to detect spam. At first you can use > cookies > (e.g. sessions) and IP-addresses to detect term spam. You can also limit a > popularity spike to a reasonable mean size over a longer period. And you can > limit rates using logarithms. > > There are many ways to deal with spam and maintain decent statistics. > > In practice, it's not a big problem on most sites. > >> Ranking suggestions based on query count would be trivially easy to spam. >> Have a bot make my preferred queries over and over again, and "boom" they >> are the most-preferred.