According to my knowledge, Solr cannot support this.
In my case, I get data by keyword-matching from Solr and then rank the data
by PageRank after that.
Thanks,
Bing
On Wed, Apr 4, 2012 at 6:37 AM, Manuel Antonio Novoa Proenza <
mano...@estudiantes.uci.cu> wrote:
> Hello,
>
> I have in my Solr
You might want to look into Nutch and its LinkRank instead of Solr for
this. For obtaining such information, you need a crawler to crawl through
the links. Not what Solr is meant for.
Rav
On Wed, Apr 4, 2012 at 8:46 AM, Bing Li wrote:
> According to my knowledge, Solr cannot support this.
>
>
I have finally solved my problem!!
Did the following:
added two lines in the /browse requestHandler
velocity.properties
text/html;charset=UTF-8
Moved velocity.properties from solr/conf/velocity to solr/conf
Not being an expert, I am not 100% sure this is the "best" solution, and
w
Updating a single field is not possible in solr. The whole record has to
be rewritten.
300 MB is still not that big a file. Have you tried doing the indexing (if
its only a one time thing) by giving it ~2 GB or xmx?
A single file with that size is strange! May I ask what is it?
Rav
On Tue, A
Hi,
My index is composed of documents with an "author" field. My system is a
users portal where they can have a friend relationship among each other.
When a user searches for documents, I would like to boost score of docs in
which author is friend of the user doing the search. Note that the list o
Hi,
I have documents in different languages and I want to choose the
tokenizer to use for a document based on the language of the document. The
language of the document is already known and is indexed in a field. What I
want to do is when I index the text in the document, I want to choose
There is https://issues.apache.org/jira/browse/LUCENE-3837 but I suppose
it's too far from completion.
On Wed, Apr 4, 2012 at 2:48 PM, Ravish Bhagdev wrote:
> Updating a single field is not possible in solr. The whole record has to
> be rewritten.
>
> 300 MB is still not that big a file. Have y
Yes, I think there are good reasons why it works like that. Focus of
search system is to be efficient on query side at cost of being not that
efficient on storage.
You must however also note that by default a field's length is limited to
1 words in solrconf.xml which you may also need to modi
Hi,
we want to use the LBHttpSolrServer (4.0/trunk) and specify a preferred
server. Our use case is that for one user request we make several solr
requests with some heavy caching (using a custom request handler with a
special cache) and want to make sure that the subsequent solr requests
are hitt
Hello,
I am trying to use commitwithin in Java but there seams to be no commit
at all with this option.
1. Example Code:
UpdateRequest request = new UpdateRequest();
request.deleteByQuery("fild:value");
request.setCommitWithin(1);
System.out.println(request.getCommitWithin
On Apr 3, 2012, at 10:35 PM, Jamie Johnson wrote:
> I haven't personally seen this issue but I have been told by another
> developer that he ran a deleteByQuery("*:*"). This deleted the index,
> but on restart there was information still in the index. Should this
> be possible? I had planned t
Solr version? I think that for a while now, deletes where not triggering
commitWithin. I think this was recently fixed - if I remember right it will be
part of 3.6 and then 4.
- Mark Miller
lucidimagination.com
On Apr 4, 2012, at 10:12 AM, Jens Ellenberg wrote:
> Hello,
>
> I am trying to us
This is cross posted on Drupal.org: http://drupal.org/node/1515046
Summary: I have a fairly clean install of Drupal 7 with
Apachesolr-1.0-beta18. I have created a content type called document with a
number of fields. I am working with 30k+ records, most of which are related
to "Marion, IA" in some
Did you try to append &debugQuery=on to get more information?
> -Original Message-
> From: Sean Adams-Hiett [mailto:s...@advantage-companies.com]
> Sent: Wednesday, April 04, 2012 10:43 AM
> To: solr-user@lucene.apache.org
> Subject: Search for "library" returns 0 results, but search for "
Yes, can you check if results you get with "marion library" match on marion
or library? By default solr uses OR between words (specified in
solrconfig.xml). You can also easily check this by enabling highlighting.
Ravish
On Wed, Apr 4, 2012 at 4:11 PM, Joshua Sumali wrote:
> Did you try to ap
hi Rav
Thank you for your answer.
In my case I use nutch for crawling the web. Using nutch am a true rookie. How
do I configure nutch to return that information? And how do I make solr to
index that information, or that information is being built with the score of
the indexed documents.
thank
Hi,
Please subscribe to the Nutch mailing list. Scoring is straightforward
and calculated scores can be written to the CrawlDB or as external file
field for Solr.
Cheers
On Wed, 04 Apr 2012 10:22:46 -0500 (COT), Manuel Antonio Novoa Proenza
wrote:
hi Rav
Thank you for your answer.
In my
Here are some of the XML results with the debug on:
library
library
+DisjunctionMaxQuery((content:librari)~0.01)
DisjunctionMaxQuery((content:librari^2.0)~0.01)
+(content:librari)~0.01
(content:librari^2.0)~0.01
DisMaxQParser
0.0
0.0
0.0
0.0
0.0
0.0
0.0
0.0
0.0
0.0
0.0
Hi,
I'm evaluating Solr for use in a project. In the Solr FAQ under "How can I
rebuild my index from scratch if I change my schema?" After restarting the
server, step 5 is to "Re-Index your data" no mention is made of how this
is done.
For more routine changes, are record updates supported with
"Re-Index your data" ~= Reload your data
On Wed, Apr 4, 2012 at 12:46 PM, Joseph Werner wrote:
> Hi,
>
> I'm evaluating Solr for use in a project. In the Solr FAQ under "How can I
> rebuild my index from scratch if I change my schema?" After restarting the
> server, step 5 is to "Re-Index your
On Wed, Apr 4, 2012 at 12:46 PM, Joseph Werner wrote:
> For more routine changes, are record updates supported without the
> necessitity to rebuilt an index? For example if a description field for an
> item needs be changed, am I correct in reading that the recodrd need only
> be resubmitted?
Cor
I am trying to use jndiName attribute in db-data-config.xml. This works great
in tomcat. However having issues in websphere.
Following exception is thrown
"Make sure that a J2EE application does not execute JNDI operations on
"java:" names within static code blocks or in threads created by that
No one knows. But if you ask the devs, they will say 'when its done'.
One clue might be to monitor the bugs/issues scheduled for 4.0. When
they are all resolved, then its ready.
On Wed, 2012-04-04 at 09:41 -0700, srinivas konchada wrote:
> Hello every one
> Does any one know when Solr 4.0 will be
Apologies for not replying sooner on this thread, I just noticed it today...
To add insight into where velocity.properties can reside, it is used this way
in VelocityResponseWriter.java:
SolrVelocityResourceLoader resourceLoader =
new
SolrVelocityResourceLoader(request.getCore().get
Thanks Mark. The delete by query is a very rare operation for us and
I really don't have the liberty to update to current trunk right now.
Do you happen to know about when the fix was made so I can see if we
are before or after that time?
On Wed, Apr 4, 2012 at 10:25 AM, Mark Miller wrote:
>
> O
Thanks.
Increasing max. heap space is not a scalable option as it reduces the
ability of the system to scale with multiple concurrent index requests.
The use case is indexing a set of text files which we have no control over
i.e. could be small or large.
--
View this message in context:
http:/
On Wed, Apr 4, 2012 at 3:04 PM, Jamie Johnson wrote:
> Thanks Mark. The delete by query is a very rare operation for us and
> I really don't have the liberty to update to current trunk right now.
> Do you happen to know about when the fix was made so I can see if we
> are before or after that tim
> Updating a single field is not possible in solr. The whole record has to
> be rewritten.
Unfortunate. Lucene allows it.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Incrementally-updating-a-VERY-LARGE-field-Is-this-possibe-tp3881945p3885253.html
Sent from the Solr -
On Wed, Apr 4, 2012 at 3:14 PM, vybe3142 wrote:
>
>> Updating a single field is not possible in solr. The whole record has to
>> be rewritten.
>
> Unfortunate. Lucene allows it.
I think you're mistaken - the same limitations apply to Lucene.
-Yonik
lucenerevolution.com - Lucene/Solr Open Source
I believe we are talking about two different things. The original question was
about incrementally building up a field during indexing, right?
After a document is committed, a field cannot be separately updated, that is
true in both Lucene and Solr.
wunder
On Apr 4, 2012, at 12:20 PM, Yonik S
Does anyone have a blog, wiki with detailed step by step instructions on
setting up SOLRCloud on multiple JBOSS instances?
Thanks in advance,
Does anyone have any instructions on setting up SOLRCloud on multiple
appservers? Ideally a wiki, blog, step-by-step guide I can follow.
thanks, that will work I think
--
View this message in context:
http://lucene.472066.n3.nabble.com/space-making-it-hard-tu-use-wilcard-with-lucene-parser-tp3882534p3885460.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi again Chris,
I finally manage to find some proper time to test your configuration.
First thing to notice is that it worked for me assuming the following
pre-requisites were satisfied:
- you had the jar containing the AnalysisEngine for the RoomAnnotator.xml
in your libraries section (this is ac
Hello Peter,
I think that is more related to UIMA AlchemyAPIAnnotator [1] or to
AlchemyAPI services themselves [2] because Solr just use the out of the box
UIMA AnalysisEngine for that.
Thus it may make sense to ask on d...@uima.apache.org (or even directly to
AlchemyAPI guys).
HTH,
Tommaso
[1] :
depending on you jvm version, -XX:+UseCompressedStrings would help alleviate
the problem. It did help me before.
xab
--
View this message in context:
http://lucene.472066.n3.nabble.com/Incrementally-updating-a-VERY-LARGE-field-Is-this-possibe-tp3881945p3885493.html
Sent from the Solr - User mail
Hi Martijn,
I created a JIRA issue and attached a test that fails. It seems to exhibit the
same issue that I see on my local box. (If you run it multiple times you can
see that the group value of the top doc changes between runs.)
Also, I had to change add fixShardCount = true; in the construct
Yonik Seeley-2-2 wrote
>
> On Wed, Apr 4, 2012 at 3:14 PM, vybe3142 wrote:
>>
>>> Updating a single field is not possible in solr. The whole record has
>>> to
>>> be rewritten.
>>
>> Unfortunate. Lucene allows it.
>
> I think you're mistaken - the same limitations apply to Lucene.
If you index a set of documents with SolrJ and use
StreamingUpdateSolrServer.add(Collection docs, int
commitWithinMs),
it will perform a commit within the time specified, and it seems to use default
values for waitFlush and waitSearcher.
Is there a place where you can specify different values fo
On Apr 4, 2012, at 6:50 PM, Mike O'Leary wrote:
> If you index a set of documents with SolrJ and use
> StreamingUpdateSolrServer.add(Collection docs, int
> commitWithinMs),
> it will perform a commit within the time specified, and it seems to use
> default values for waitFlush and waitSearcher.
Hi,
I just submitted an issue with patch for this:
https://issues.apache.org/jira/browse/SOLR-3318
Cheers,
Martin
On 04/04/2012 03:53 PM, Martin Grotzke wrote:
> Hi,
>
> we want to use the LBHttpSolrServer (4.0/trunk) and specify a preferred
> server. Our use case is that for one user request
I am indexing some database contents using add(docs, commitWithinMs), and those
add calls are taking over 80% of the time once the database begins returning
results. I was wondering if setting waitSearcher to false would speed this up.
Many of the calls take 1 to 6 seconds, with one outlier that
Hi All,
I am working on an application which makes few solr calls to get the data.
On the high level, We have a requirement like this
- Make first call to Solr, to get the list of products which are
children of a given category
- Make 2nd solr call to get product documents based on a l
I am currently indexing some information and am wondering why I am
getting duplicates in facets. From what I can tell they are the same,
but is there any case that could cause this that I may not be thinking
of? Could this be some non printable character making it's way into
the index?
Sample o
Try using Luke to look at your index and see if there are multiple
similar TFV's. You can browse them easily in Luke.
On Wed, 2012-04-04 at 23:35 -0400, Jamie Johnson wrote:
> I am currently indexing some information and am wondering why I am
> getting duplicates in facets. From what I can tell t
Yes, thanks for the reply. Turns out there is whitespace differences
in these fields, thank you for the quick reply!
On Wed, Apr 4, 2012 at 11:45 PM, Darren Govoni wrote:
> Try using Luke to look at your index and see if there are multiple
> similar TFV's. You can browse them easily in Luke.
>
>
My snapshot was taken 2/27. That would seem to indicate that the
deleteByQuery should be getting versioned, I am not sure if the other
issues that were resolved would change the operation. I'll keep an
eye on it and if it pops up I'll try to push the update. Thanks.
On Wed, Apr 4, 2012 at 3:12
Not sure if this got lost in the shuffle, were there any thoughts on this?
On Wed, Mar 21, 2012 at 11:02 AM, Jamie Johnson wrote:
> Given that in a distributed environment the docids are not guaranteed
> to be the same across shards should the sorting use the uniqueId field
> as the tie breaker b
Hello,
I would like to know the method of extracting from the images that are in html
documents Alt attribute data
10mo. ANIVERSARIO DE LA CREACION DE LA UNIVERSIDAD DE LAS CIENCIAS
INFORMATICAS...
CONECTADOS AL FUTURO, CONECTADOS A LA REVOLUCION
http://www.uci.cu
http://www.
Hi,
Any inputs or experience that others have come across will be really
helpful to know.
Basically, its the same as page ranking but the information used to decide
the rank is much more dynamic in nature..
Appreciate any inputs.
Regards
Monmohan
On Wed, Apr 4, 2012 at 4:22 PM, monmohan wrote:
>
If you have degree of separation (like friend). You could do something like:
...defType=dismax&bq=degree_of_separation:1^100
Thanks.
On Thu, Apr 5, 2012 at 12:55 AM, Monmohan Singh wrote:
> Hi,
> Any inputs or experience that others have come across will be really
> helpful to know.
> Basically
51 matches
Mail list logo