Is there any workaround in Solr/Carrot2 So that we could pass tokens that'd
been filtered with customer tokenizer/filters instead of rawtext that it
currently
uses for clustering ?
I read an issue in following link too .
https://issues.apache.org/jira/browse/SOLR-2917
Is writing our own parsers
On Wed, Jun 6, 2012 at 8:51 PM, Darin Pope wrote:
> When using SolrJ (1.4.1 or 3.5.0) and calling either addBean or
> deleteByQuery, the POST body has numbers before and after the XML (47 and 0
> as noted in the example below):
>
It looks like this is HTTP chunked transfer encoding. As to whethe
I have Solr 4.0 (apache-solr-4.0) and JBoss Application Server 5.1.2
installed in RHEL 6.2
machine. I was successful in integrating solr with JBoss and I am able to
view admin console (single core).
Now I would link to create the Master/Slave configuration for Solr servers ?
can anyone help me?
Thanks Michael and Lance!
I decided to go with an Oracle Pipelined Table function and that took care of
it. I think that's what Michael was referring to below. This enabled us to be
able to make a simple SQL call.
Thanks again.
>
> From: Lance Norskog
>To:
In latest 4.0 release, the addFile() method has a new argument 'contentType':
addFile(File file, String contentType)
In context of Solr Cell how should addFile() method be called? Specifically
I refer to the Wiki example:
ContentStreamUpdateRequest up = new
ContentStreamUpdateRequest("/update/ex
You may want to read the faceting overview:
http://wiki.apache.org/solr/SolrFacetingOverview
-- Jack Krupansky
-Original Message-
From: Swetha Shenoy
Sent: Thursday, June 07, 2012 5:24 PM
To: solr-user@lucene.apache.org
Subject: Filter query vs Facets
Hi,
I had a question regardi
Look at the text_en field type in the Solr 3.6 example schema.
-- Jack Krupansky
-Original Message-
From: Carrie Coy
Sent: Thursday, June 07, 2012 5:04 PM
To: solr-user@lucene.apache.org
Subject: PorterStemmerTokenizerFactory ?
I've read different suggestions on how to handle cases
I've read different suggestions on how to handle cases where synonyms
are used and there are multiple
version of the original word that need to point to the same set of
synonyms (/responsibility, responsibilities, obligation, duty/ ).
The approach that seems most logical is to configure a
Syno
SOLR-1855 has a script that checks replication details:
/solr/${CORE}/replication?command=details
# Get the last time the core replicated correctly.
# Get the last time the core failed to replicate.
# Is this core replicating (aka pulling index from master) right now?
See:
https://issues.apache
Are you requesting a large number of rows? If so, request smaller chunks, like
ten at a time. Then you can show those with a "waiting" note.
wunder
On Jun 7, 2012, at 1:14 PM, Laurent Vaills wrote:
> Hi everyone,
>
> We have some grouping queries that are quite long to execute. Some are too
>
But, check out things like httplib2 and urllib2.
-Original Message-
From: Spadez [mailto:james_will...@hotmail.com]
Sent: Thursday, June 07, 2012 2:09 PM
To: solr-user@lucene.apache.org
Subject: RE: Help! Confused about using Jquery for the Search query - Want to
ditch it
Thank you, that
As far as I know, it is the only way to do this. Look around a bit, Python (or
PHP, or C, etc., etc.) is able to act as an HTTP client...in fact, that is the
most common way that web services are consumed. But, we are definitely beyond
the scope of the Solr list at this point.
-Original Mes
Thank you, that helps. The bit I am still confused about how the server sends
the response to the server though. I get the impression that there are
different ways that this could be done, but is sending an XML response back
to the Python server the best way to do this?
--
View this message in c
Yes (or, at least, I think I understand what you are saying, haha.) Let me
clarify.
1. Client sends GET request to web server
2. Web server (via Python, in your case, if I remember correctly) queries Solr
Server
3. Solr server sends response to web server
4. You take that data and put it into th
+1 on that! If you do want to provide direct results, ALWAYS send
requests through a proxy that can verify that a) all requests are coming
from your web app, and b) only "acceptable" queries are being passed on.
Nick
On 6/7/2012 2:50 PM, Michael Della Bitta wrote:
On Thu, Jun 7, 2012 a
On Thu, Jun 7, 2012 at 1:59 PM, Nick Chase wrote:
> The other option is to create a python page that does the call to Solr and
> spits out just the HTML for your results, then call THAT rather than calling
> Solr directly.
This is the *only* option if you're listening to Walter and I. Don't
giv
Is there a programmatic way or otherwise to become aware when the
replication operation starts? In looking at the source for
ReplicationHandler, there aren't log statements to indicate that it started.
Thanks, Jon
On 6/7/2012 1:53 PM, Spadez wrote:
Hi Ben,
Thank you for the reply. So, If I don't want to use Javascript and I want
the entire page to reload each time, is it being done like this?
1. User submits form via GET
2. Solr server queried via GET
3. Solr server completes query
4. Solr server retur
Hi Ben,
Thank you for the reply. So, If I don't want to use Javascript and I want
the entire page to reload each time, is it being done like this?
1. User submits form via GET
2. Solr server queried via GET
3. Solr server completes query
4. Solr server returns XML output
5. XML data put into resu
I'm new to Solr...but this is more of a web programming question...so I can get
in on this :).
Your only option to get the data from Solr sans-Javascript, is the use python
to pull the results BEFORE the client loads the page.
So, if you are asking if you can get AJAX like results (an already l
For posterity, I think we're going to remove 'preference' data from Solr
indexing and go in the custom Function Query direction with a key-value
store.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Boost-by-Nested-Query-Join-Needed-tp3987818p3988255.html
Sent from the Solr -
Thank you for the reply, but I'm afraid I don't understand :(
This is how things are setup. On my Python website, I have a keyword and
location box. When clicked, it queries the server via a javascript "GET"
request, it then sends back the data via Json.
I'm saying that I dont want to be reliant
I would debug somewhere close to the FuzzyQuery.
Lucene is doing exactly that (just as PrefixQueries are doing): expand a
FuzzyQuery (PrefixQuery) to a disjunction of term-queries for the words that
match that fuzzy or prefix queries.
Maybe it helps you start?
paul
Le 7 juin 2012 à 18:15, Giov
Hi all,
I am wandering if SOLR can return me all words in my text corpus
that have a given levenstein distance with my query word.
Possible?
Difficult?
Cheers,
Giovanni
Sounds like an interesting improvement to propose.
It will also depend on various factors, such as number of unique terms in a
field, field type, etc.
Which field types are giving you the most trouble and how many unique values
do they have? And do you specify a facet.method or just let it de
Is the index otherwise usable for queries? And it is only the optimize that
is failing?
I suppose it is possible that the index could be corrupted, but it is also
possible that there is a bug in Lucene.
I would suggest running Lucene "CheckIndex" next. See what it has to say.
See:
https://bu
And keep Solr behind a firewall or authentication or even better,
both! People *will* find and exploit your Solr installation.
Michael Della Bitta
Appinions, Inc. -- Where Influence Isn’t a Game.
http://www.appinions.com
On Thu, Jun 7, 2012 at 10
Hello, Boris,
If I remember correctly, older versions of Solr report the version of
the as-of-yet uncommitted core in the replication page. So if you did
a commit on the master and then a replication, you'd see that version
on the client.
Michael Della Bitta
-
This is a bad idea. Solr is not designed to be exposed to arbitrary internet
traffic and attacks. The best design is to have a front end server make
requests to Solr, then use those to make HTML pages.
wunder
On Jun 7, 2012, at 4:49 AM, Spadez wrote:
> Final comment from me then Ill let someon
Thanks for your reply.
I think the number could eventually get very large (~1B) as our
customer-base grows, since each customer could possibly have a preference
for each candy, but currently we're looking at around 50M.
I've looked at the Solr-2272 patch for joins, which looks as though it might
Solr (Lucene actually) stores the source form of the data that was fed to
Solr, so it is not yet tokenized and will include all punctuation and
whitespace.
-- Jack Krupansky
-Original Message-
From: Mark , N
Sent: Thursday, June 07, 2012 7:45 AM
To: solr-user@lucene.apache.org
Subjec
What language environment are you using? PHP, Python, Ruby, other?
Each has its own interface.
But ultimately Solr is just another web service with an HTTP and XML or JSON
interface. So, it is mostly a question of how your client environment
accesses web services that have an HTTP and XML or J
Hello!
-Original message-
> From:Dyer, James
> Sent: Wed 06-Jun-2012 17:23
> To: solr-user@lucene.apache.org
> Subject: RE: issues with spellcheck.maxCollationTries and
> spellcheck.collateExtendedResults
>
> Markus,
>
> With "maxCollationTries=0", it is not going out and querying the
Thanks everyone!
--
View this message in context:
http://lucene.472066.n3.nabble.com/Solr-4-0-Clean-Commit-for-production-use-tp3987852p3988183.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi
The search is distributed over all shards. The problem exists locally as well.
Thanks,
-Original message-
> From:Jack Krupansky
> Sent: Wed 06-Jun-2012 17:07
> To: solr-user@lucene.apache.org
> Subject: Re: issues with spellcheck.maxCollationTries and
> spellcheck.collateExtendedRe
Hi,
Best Buy is building new Search Platform/Eco-System powered by Lucene/Solr.
We are hiring multiple Lucene/Solr engineers, tech leads, and architects,
both full-time and consulting based in Minneapolis, MN. This is a long term
project and the team is fun to work with.
Please reach out to me if
Am 07.06.2012 09:55, schrieb sheethal shreedhar:
http://localhost:8983/solr/select/?q=fruit&version=2.2&start=0&rows=10&indent=on
I get
HTTP ERROR 400
Problem accessing /solr/select/. Reason:
undefined field text
Look at your schema.xml. You'll find a line like this:
text
Replace "te
Hi all
can anybosy suggest me how to work with solr in web application
please send the information
Regards
Raja
--
View this message in context:
http://lucene.472066.n3.nabble.com/how-to-work-with-solr-tp3988154.html
Sent from the Solr - User mailing list archive at Nabble.com.
Final comment from me then Ill let someone else speak.
The solution we seem to be looking at is send a GET request to SOLR and then
send back a renderized page, so we are basically creating the results page
on the server rather than the client side.
I would really like to hear what people have to
thanks Jack , I will try updateProcessor
Between does SOLR store tokenized "content" in fields if field have
property stored="true" ?
On Tue, Jun 5, 2012 at 8:23 PM, Jack Krupansky wrote:
> My (very limited) understanding of "boilerpipe" in Tika is that it strips
> out "short text", which
I read someone's question and answer about db connection.
Someone said, db connection is still alive for 10minutes.
But I start to index(dataimport) before 1 hour, all of db connection
remains for 1 hour.
| 88757 | id | localhost:38843 | tmp | Sleep | 3696 | | NULL
|
| 88758 |
During the analysis phase you could add payloads to the terms using
LevensteinDistance and then use that in conjunction with a
PayloadSimilarity class ´See [1] for an example), or just use a custom
Similarity class which uses LevensteinDistance for scoring.
HTH
Tommaso
[1] :
http://www.lucidimagin
I index many tables which are written with entities in data-config.xml.
But after indexing one table, db connection remains
even though I set 'holdability="CLOSE_CURSORS_AT_COMMIT"'.
How can I remove the connection after indexing a table?
You haven't really told us much about what you're doing here. As Lee
hints, we don't know much about the details of *how* you are doing this.
But unless you're doing something odd, Solr shouldn't be the bottleneck
here. Often when a database import is slow, the problem is in the data-
acquisition
Hi Jack,
its the virtual machine running on a VMware vSphere 5 Enterprise Plus.
Machine has 30 GB vRAM, 8 core vCPU 3.0 GHz, 2 TB SATA RAID-10 over iSCSI.
Operation system is CentOS 6.2 64bit.
Here are java infos:
- catalina.base/usr/share/tomcat6
- catalina.home/usr/share/tomcat6
-
Further to my last reply. How about I do the following:
Send the request to the server using the GET method and then return the
results in XML rather than JSON. Does this sound logical?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Help-Confused-about-using-Jquery-for-the-
On Thu, Jun 7, 2012 at 1:18 PM, Ramprakash Ramamoorthy <
youngestachie...@gmail.com> wrote:
> I am porting my app from lucene 2.X(solr 1.3) to lucene 3.X(solr 3.5). The
> following is my issue.
>
> This one was valid in 2.X, but 3.5 throws me an error.
>
> IndexReader reader = IndexReader.open("/h
what is your db schema ? do you need to import all the schema ? (128
joined tables ??)
or are the tables all independant ? (if so dump them out and import
them in using csv)
cheers lee c
On 7 June 2012 02:32, Jihyun Suh wrote:
> Each table has 35,000 rows. (35 thousands).
> I will check the log
Hi,
My current method of searching involes communicating to solr using python.
The clients browser communicates to the search API using jquery/json.
However, although this works, I dont like the dependancy on Javascript.
Either I can keep with this method and have a backup system in place that
wo
I am porting my app from lucene 2.X(solr 1.3) to lucene 3.X(solr 3.5). The
following is my issue.
This one was valid in 2.X, but 3.5 throws me an error.
IndexReader reader = IndexReader.open("/home/path/to/my/dataDir");
2.X accepted a string, but 3.5 strictly wants a Directory object. I find
Dir
50 matches
Mail list logo