Thanks for stating this in a simple fashion.

On Sun, Feb 8, 2015 at 6:07 PM, Jorge Luis Betancourt González <
jlbetanco...@uci.cu> wrote:

> For a project I'm working on, what we do is store the user's query in a
> separated core that we also use to provide an autocomplete query
> functionality, so far, the frontend app is responsible of sending the query
> to Solr, meaning: 1. execute the query against our search core and 2. send
> an update request to store the query in the separated core. We use some
> deduplication (provided by Solr) to avoid storing the same query several
> times. We don't do what you're after but it would't be to hard to tag each
> query with a timestamp field and provide analytics. Thinking from the top
> of my head we could wrap this logic that is currently done in the frontend
> app in a custom SearchComponent that automatically send the search query
> into the other core for storing, abstracting all this logic from the client
> app. Keep in mind that the considerations regarding volume of data that
> Shawn has talked keeps being valid.
>
> Hope it helps,
>
> ----- Original Message -----
> From: "Shawn Heisey" <apa...@elyograg.org>
> To: solr-user@lucene.apache.org
> Sent: Sunday, February 8, 2015 11:03:33 AM
> Subject: [MASSMAIL]Re: Trending functionality in Solr
>
> On 2/7/2015 9:26 PM, S.L wrote:
> > Is there a way to implement the trending functionality using Solr , to
> give
> > the results using a query for say the most searched terms in the past
> hours
> > or so , if the most searched terms is not possible is it possible to at
> > least the get results for the last 100 terms?
>
> I'm reasonably sure that the only thing Solr has out of the box that can
> record queries is the logging feature that defaults to INFO.  That data
> is not directly available to Solr, and it's not in a good format for
> easy parsing.
>
> Queries are not stored anywhere else by Solr.  From what I understand,
> analysis is a relatively easy part of the equation, but the data must be
> available first, which is the hard part.  Storing it in RAM is pretty
> much a non-starter -- there are installations that see thousands of
> queries every second.
>
> This is an area for improvement, but the infrastructure must be written
> from scratch.  All work on this project is volunteer.  We are highly
> motivated volunteers, but extensive work like this is difficult to fit
> into donated time.
>
> Many people who use Solr are already recording all queries in some other
> system (like a database), so it is far easier to implement analysis on
> that data.
>
> Thanks,
> Shawn
>
>

Reply via email to