You might want to also look at Trulia's Thoth project
https://github.com/trulia/thoth-ml/ -- it doesn't supply this feature
out of the box, but it gives you a nice framework for implementing it.

On Sun, Feb 8, 2015 at 5:07 PM, Jorge Luis Betancourt González
<jlbetanco...@uci.cu> wrote:
> For a project I'm working on, what we do is store the user's query in a 
> separated core that we also use to provide an autocomplete query 
> functionality, so far, the frontend app is responsible of sending the query 
> to Solr, meaning: 1. execute the query against our search core and 2. send an 
> update request to store the query in the separated core. We use some 
> deduplication (provided by Solr) to avoid storing the same query several 
> times. We don't do what you're after but it would't be to hard to tag each 
> query with a timestamp field and provide analytics. Thinking from the top of 
> my head we could wrap this logic that is currently done in the frontend app 
> in a custom SearchComponent that automatically send the search query into the 
> other core for storing, abstracting all this logic from the client app. Keep 
> in mind that the considerations regarding volume of data that Shawn has 
> talked keeps being valid.
>
> Hope it helps,
>
> ----- Original Message -----
> From: "Shawn Heisey" <apa...@elyograg.org>
> To: solr-user@lucene.apache.org
> Sent: Sunday, February 8, 2015 11:03:33 AM
> Subject: [MASSMAIL]Re: Trending functionality in Solr
>
> On 2/7/2015 9:26 PM, S.L wrote:
>> Is there a way to implement the trending functionality using Solr , to give
>> the results using a query for say the most searched terms in the past hours
>> or so , if the most searched terms is not possible is it possible to at
>> least the get results for the last 100 terms?
>
> I'm reasonably sure that the only thing Solr has out of the box that can
> record queries is the logging feature that defaults to INFO.  That data
> is not directly available to Solr, and it's not in a good format for
> easy parsing.
>
> Queries are not stored anywhere else by Solr.  From what I understand,
> analysis is a relatively easy part of the equation, but the data must be
> available first, which is the hard part.  Storing it in RAM is pretty
> much a non-starter -- there are installations that see thousands of
> queries every second.
>
> This is an area for improvement, but the infrastructure must be written
> from scratch.  All work on this project is volunteer.  We are highly
> motivated volunteers, but extensive work like this is difficult to fit
> into donated time.
>
> Many people who use Solr are already recording all queries in some other
> system (like a database), so it is far easier to implement analysis on
> that data.
>
> Thanks,
> Shawn
>

Reply via email to