Hi All,
I was able to set the cookie value to the Stream_url connection, i was
able to pass the cookie value upto contentstreamBase.URLStream class and i
added
conn.setRequestProperty("Cookie",cookie[0].name"="cookie[0].value) in the
connection setup.. and it is working fine now...
Regards,
s
I'm entering only a query parameter.
I posted a bug description there - http://pecl.php.net/bugs/bug.php?id=22634
2011/4/8 Israel Ekpo
> Hi,
>
> Could you send the enter list of parameters you are ending to solr via the
> SolrClient and SolrQuery object?
>
> Please open a bug request here with
On Fri, Apr 8, 2011 at 6:23 AM, Jens Mueller wrote:
> Hello all,
>
> thanks for your generous help.
>
> I think I now know everything: (What I want to do is to build a web
> crawler
> and index the documents found). I will start with the setup as suggested by
>
>
Write a web crawler from scratch
Hello all,
thanks for your generous help.
I think I now know everything: (What I want to do is to build a web crawler
and index the documents found). I will start with the setup as suggested by
Ephraim (Several sharded masters, each with at least one slave for reads and
some aggregators for quer
Thanks Lance,
I'm using Solr 1.4.
If I want to using TikaEP, need to upgrade to Solr 3.1 or import jar files?
Best Regards,
Roy Liu
On Fri, Apr 8, 2011 at 10:22 AM, Lance Norskog wrote:
> You need the TikaEntityProcessor to unpack the PDF image. You are
> sticking binary blobs into the index.
You need the TikaEntityProcessor to unpack the PDF image. You are
sticking binary blobs into the index. Tika unpacks the text out of the
file.
TikaEP is not in Solr 1.4, but it is in the new Solr 3.1 release.
On Thu, Apr 7, 2011 at 7:14 PM, Roy Liu wrote:
> Hi,
>
> I have a table named *attachme
Hi,
I have a table named *attachment *in MS SQL Server 2008.
COLUMNTYPE
-
id int
titlevarchar(200)
attachment image
I need to index the attachment(store pdf files) column from database via
DIH.
After access this URL, it returns "Ind
Hi,
Could you send the enter list of parameters you are ending to solr via the
SolrClient and SolrQuery object?
Please open a bug request here with the details
http://pecl.php.net/bugs/report.php?package=solr
On Thu, Apr 7, 2011 at 7:59 PM, Haspadar wrote:
> Hello
> I updated Solr to version
Thanks Erick. This looks like it would work... I sent out an update to
my original query, there is another approach that would probably also
work for my case that is being used by SpellCheckerComponent.
I will check out both approaches.
Thanks very much for your help.
-sujit
On Thu, 2011-04-07
I think I found the answer by looking through the code...specifically
SpellCheckComponent.
So my component would have to implement SolrCoreAware and in the
inform() method, register a custom SolrEventListener which will execute
the regeneration code in the postCommit and newSearcher methods.
Woul
I haven't built one myself, but have you considered the Solr UserCache?
See: http://wiki.apache.org/solr/SolrCaching#User.2BAC8-Generic_Caches
It even receives warmup signals I believe...
Best
Erick
On Thu, Apr 7, 2011 at 7:39 PM, Sujit Pal wrote:
> Hi,
>
> I am developing a SearchComponent th
Hello
I updated Solr to version 3.1 on my project. And now when the application
calls getResponse () method (PECL extension) I get the following:
"Fatal error: Uncaught exception 'SolrException' with message 'Error
un-serializing response' in /home/.../Adapter/Solr.php: 78"
How can I fix it?
Than
Hi,
I am developing a SearchComponent that needs to build some initial
DocSets and then intersect with the result DocSet during each query (in
process()).
When the searcher is reopened, I need to regenerate the initial DocSets.
I am on Solr 1.4.1.
My question is, which method in SearchComponent
: Solr 3.1.0 uses different javabin format from 1.4.1
: So if I use Solrj 1.4.1 jar , then i get javabin error while saving to
: 3.1.0
: and if I use Solrj 3.1.0 jar , then I get javabin error while reading the
: document from solr 1.4.1.
you can use the XML format to get portability during the
I think you can specify the in-group sort, and specify a very small number
(perhaps
even one) to go in each group. But you'd have to store the length of each
body and sort by that.
I'm pretty sure grouping is trunk-only.
The problem here is getting something that applies
just within the group and
Would grouping solve this? I'd rather not move to a pre-release solr ...
To clarify the problem:
The data are fine and not duplicated - however, I want to analyze the data, and
summarize one field (kind of like faceting), to understand what the largest
value is.
For example:
Document 1: la
> The jar containing the class is in
> here:
>
> /usr/local/apache-tomcat-6.0.20/webapps/solr/WEB-INF/lib
>
http://wiki.apache.org/solr/SolrPlugins#How_to_Load_Plugins
One possibility is to have just a "groups" field with a positionIncrementGap
of, say, 100.
that is multiValued.
Now, index values like
"group1 foo bar happy joy joy"
"group2 some more words to search"
etc.
Now do phrase queries with a slop of less than 100. Then searches like
groups:"group1 more
The jar containing the class is in here:
/usr/local/apache-tomcat-6.0.20/webapps/solr/WEB-INF/lib
for my setup.
Tri
From: Erick Erickson
To: solr-user@lucene.apache.org
Sent: Thu, April 7, 2011 3:24:14 PM
Subject: Re: class not found
Can you give us some mo
> Date: Friday, April 8, 2011, 1:19 AM
> I have 1 Master, and 3 slaves. The
> master holds the solr index. How do I
> connect the slaves to the master? I have the script in the
> bin folders. I
> have rsyncd installed and snapshooter enabled in the
> master. Thanks, please
HTTP based replication i
yes.
From: Ahmet Arslan
To: solr-user@lucene.apache.org
Sent: Thu, April 7, 2011 3:23:56 PM
Subject: Re: class not found
> I wrote my own parser plugin.
>
> I'm getting a NoClassCefFoundError. Any ideas why?
Did you put jar file - that contains you custom c
Can you give us some more details? I suspect the jar file containing
your plugin isn't in the Solr lib directory and/or you don't have a lib
directive in your solrconfig.xml file pointing to where your jar is.
But that's a guess since you haven't provided any information about
what you did to try
> I wrote my own parser plugin.
>
> I'm getting a NoClassCefFoundError. Any ideas why?
Did you put jar file - that contains you custom code - into /lib directory?
http://wiki.apache.org/solr/SolrPlugins
> Does this contain the
> CollapseComponent?
No, For FieldCollapsing you need trunk.
Did you try the curl commands that Adam suggested as part of this e-mail
thread?
If so, what happened?
Best
Erick
On Wed, Apr 6, 2011 at 7:50 AM, Mike wrote:
> Hi All,
>
> I am new to solr. I have gone through solr documents to index pdf files,
> But
> it was hard to find the exact procedure to
I have 1 Master, and 3 slaves. The master holds the solr index. How do I
connect the slaves to the master? I have the script in the bin folders. I
have rsyncd installed and snapshooter enabled in the master. Thanks, please
help.
Hi,
SIREn [1], a Lucene/Solr plugin, allows you perform queries across an
undetermined number of fields, even if you have hundred of thousands of
fields. It might be helpful for your scenario.
[1] http://siren.sindice.com
--
Renaud Delbru
On 07/04/11 19:18, jisenhart wrote:
I have a questi
:
: Thanks for your response..please find below the schema details corresponding
: to that field..
your message inlcuded nothing but a bunch of blank lines, probably because
your email editor thought you were trying to type in html (instead of xml)
before diving too deeply into your analyser ho
: of the components as well as the flow of data and queries. The result is
: a conceptual architecture diagram, clearly showing how Solr relates to
: the app-server, how cores relate to a Solr instance, how documents enter
: through an UpdateRequestHandler, through an UpdateChain and Analysis a
Hi,
I wrote my own parser plugin.
I'm getting a NoClassCefFoundError. Any ideas why?
Apr 7, 2011 1:12:43 PM org.apache.solr.common.SolrException log
SEVERE: java.lang.NoClassDefFoundError: Could not initialize class
org.apache.solr.search.QParserPlugin
at org.apache.solr.core.SolrCore.
How can one change tomcat specific settings such as tomcat-users.xml? I
can't seem to find any reference to these conf files?
Thanks
I have a question on how to set up queries not having a predetermined
field list to search on.
Here are some sample docs,
1234
hihello
lalachika chika boom
boom
1235
foobarhappy happy
joy joy
some textsome more words to
search
.
.
.
4567
bedrock
memeyou you
Why does solr keeps creating connections for each table, even though they are
in the same database? This happened in the process of full-import. I used
one table as the root entity, and join the other tables. But solr keeps
creating database connections for each table. Any one has any idea or had
s
Hi All,
I am new to solr. I have gone through solr documents to index pdf files, But
it was hard to find the exact procedure to get started.
I need step by step procedure to do this. Could you please let me know the
steps to index pdf files.
Thanks,
Mike
--
View this message in context:
http://
Hey Robert,
Thanks for the quick response. That helps a lot.
--Sid
On Thu, Apr 7, 2011 at 11:19 AM, Robert Muir wrote:
> On Thu, Apr 7, 2011 at 2:13 PM, Siddharth Powar
> wrote:
> > Hey guys,
> >
> > I am in the process of moving to solr3.1 from solr1.4. I am having this
> > issue where solr3
Andrezej,
Thanks for the info. I have a question regarding stability though. How
are you able to guarantee the stability of this release when 4.0 is
still a work in progress? I believe the last version Lucid released was
1.4 so why did you choose to release a 4.x version as opposed to 3.1?
I
On 4/7/11 9:43 PM, Mark wrote:
I noticed that Lucid Works distribution now says is upt to date with 4.X
versions. Does this mean 1.4 or 4.0/trunk?
If its truly 4.0 does that mean it includes the collapse component?
Yes it does.
Also, is the click scoring tools proprietary or was this just a
I noticed that Lucid Works distribution now says is upt to date with 4.X
versions. Does this mean 1.4 or 4.0/trunk?
If its truly 4.0 does that mean it includes the collapse component?
Also, is the click scoring tools proprietary or was this just a
contrib/patch that was applied?
Thanks
Hi Park
I had the same problem. I noticed one of the issues with the blocked
messages are they are HTML/Rich Text.
(FREEMAIL_ENVFROM_END_DIGIT,FREEMAIL_FROM,FS_REPLICA,
HTML_MESSAGE
<-,RCVD_IN_DNSWL_NONE,RFC_ABUSE_POST,SPF_PASS,T_TO_NO_BRKTS_FREEMAIL
In GMail I can switch to plain text. This
This happens almost always because you're sending from a 'free' mail
account (gmail, yahoo, hotmail, etc), and your message contains words
that spam filters don't like.
For me, it was the use of the word 'remplica' (deliberately
mis-spelled so this mail gets sent).
It can also happen from 'non-fre
On Thu, Apr 07, 2011 at 04:21:25PM -0300, Ezequiel Calderara wrote:
> Happened to me a couple of times, couldn't find a way a workaround...
Note that the property "HTML_MESSAGE" has contributed to the email's spam
score:
> > (FREEMAIL_ENVFROM_END_DIGIT,FREEMAIL_FROM,FS_REPLICA,HTML_MESSAGE,RCVD_I
Happened to me a couple of times, couldn't find a way a workaround...
On Thu, Apr 7, 2011 at 4:14 PM, Parker Johnson wrote:
>
> Hello everyone. Does anyone else have problems posting to the list? My
> messages keep getting rejected with this response below. I'll be surprised
> if
> this one m
Hello everyone. Does anyone else have problems posting to the list? My
messages keep getting rejected with this response below. I'll be surprised if
this one makes it through :)
-Park
Sorry, we were unable to deliver your message to the following address.
:
Remote host said: 552 spam scor
Hi Ezequiel,
In Solr the performance of sorting and faceted search is mainly a question of
main memory.
e.g Mike McCandless wrote in s.apache.org/OWK that sorting of 5m wikipedia
documents by title field need 674 MB of RAM.
But again: My main interest is an example of other companies/product wh
Actually, what is the difference between "match" and "response"? It seems
that match always returns one result but I've thrown a few cases at it where
the score of the highest response is higher than the score of match. And
then there are cases where the match score dwarfs the highest response
scor
On Thu, Apr 7, 2011 at 2:13 PM, Siddharth Powar
wrote:
> Hey guys,
>
> I am in the process of moving to solr3.1 from solr1.4. I am having this
> issue where solr3.1 now complains about the synonyms.txt file. I get the
> following error:
> *org.apache.solr.common.SolrException: Error loading resour
Hi,
I'd would like to experiment with the UIMA contrib package, but I have
issues with the OpenCalais service's ToS and would rather not use it.
Is there a way to adapt the UIMA example setup to use only the
AlchemyAPI service? I tried simply leaving out the OpenCalais api key
but i get exceptions
Hey guys,
I am in the process of moving to solr3.1 from solr1.4. I am having this
issue where solr3.1 now complains about the synonyms.txt file. I get the
following error:
*org.apache.solr.common.SolrException: Error loading resource (wrong
encoding?): synonyms.txt*
*
*
This worked fine before in
Try setting a virtual machine and see its performance.
I'm really not a java guy, so i really don't know how to tune it for
performance...
But afaik solr handles pretty well in ram if the index is static...
On Thu, Apr 7, 2011 at 2:48 PM, Karsten Fissmer wrote:
> Hi yonik, Hi Ezequiel,
>
> Jav
Does this contain the CollapseComponent?
Will there be a significant performance boost from 1.4?
Hi yonik, Hi Ezequiel,
Java is no problem for an DVD Version. We already have a DVD version with
Servlet-Container (but this does currently not use Solr).
Some of our customers work in public sector institutions and have less then 1gb
main memory, but they use MS Word and IE and..
But let us s
Hi all,
I've been using MoreLikeThis for a while through select:
http://localhost:8983/solr/select/?q=field:more like
this&mlt=true&mlt.fl=field&rows=100&fl=*,score
I was looking over the wiki page today and saw that you can also do this:
http://localhost:8983/solr/mlt/?q=field:more like
this&m
Hi,
Thanks for your response..please find below the schema details corresponding
to that field..
---
Field type details..
Thanks,
Barani
--
View this message
What version of Solr are you using? And, assuming the version that
has it in, have you seen grouping?
Which is another way of asking why you want to do this, perhaps it's an
XY problem
Best
Erick
On Thu, Apr 7, 2011 at 1:13 AM, Peter Spam wrote:
> Hi,
>
> I have documents with a field that
I'd advise getting a copy of Luke and examining your
indexes. The information you've provided doesn't really
tell us much.
Although I do notice you don't commit in your example code...
Best
Erick
On Thu, Apr 7, 2011 at 10:21 AM, Amel Fraisse wrote:
> Hello every body,
>
> I am using Solr for in
The data are fine and not duplicated - however, I want to analyze the data, and
summarize one field (kind of like faceting), to understand what the largest
value is.
For example:
Document 1: label=1A1A1; body="adfasdfadsfasf"
Document 2: label=5A1B1; body="adfaasdfasdfsdfadsfasf"
Document 3
That's probably an issue of your analyzer. Can you show us the field
definition from the schema.xml file, for the field that you are putting
this text in?
On 4/7/2011 10:37 AM, bbarani wrote:
Hi,
We are trying to index heterogenous data using SOLR, some of the sources
have some unicode chara
On Apr 6, 2011, at 10:29 PM, Jens Mueller wrote:
> Walter, thanks for the advice: Well you are right, mentioning google. My
> question was also to understand how such large systems like google/facebook
> are actually working. So my numbers are just theoretical and made up. My
> system will be smal
Well at this point I'm more dedicated to the Deduplicate issue.
Using a Min_token_len of 4 I'm getting nice comparison results. MLT returns a
lot of similar docs that I don't consider similar - even tuning the parameters.
Finishing this issue, I found out that the signature also contains the fie
On Thu, Apr 7, 2011 at 10:28 AM, Jeffrey Chang wrote:
> Even if you can ship your DVD with a jetty server, you'll still need JAVA
> installed on the customer machine...
Including a JRE on the DVD and a launch script that uses that JRE by
default should be doable as well.
-Yonik
http://www.lucene
Hi,
We are trying to index heterogenous data using SOLR, some of the sources
have some unicode characters like Zone™ but SOLR is converting them to
Zone. Any idea how to resolve this issue?
I am using SOLR on Jetty server...
Thanks,
Barani
--
View this message in context:
http://lucene.4720
Even if you can ship your DVD with a jetty server, you'll still need JAVA
installed on the customer machine...
On Thu, Apr 7, 2011 at 10:18 PM, Ezequiel Calderara wrote:
> Can't you just run a jetty server on the background?
>
> But probably some antivirus or antispyware could take that as an toj
Hi All,
I'd like to make the highlighting work as follows:
length(all snippits) approx. 200 chars
hl.snippits = 2 (2 snippits)
is this possible with the regex fragmenter? or does anyone know of any
contrib fragmenter that might do this?
Many thanks
Dan
Hello every body,
I am using Solr for indexing and searching.
I am using 2 classes for searching document: In the first one I'm
instanciating a SolrServer to search documents as follows :
server = new EmbeddedSolrServer(
coreContainer, "");
server.add(doc);
query.setQuery("id:"+idDoc);
server.qu
I haven't used PostGIS so I can't offer a real comparison. I think if you were
to try out both, you'd be impressed with Solr's performance/scalability thanks
in large part to its sharding. But for "functionality richness" in so far as
geospatial is concerned, that's where Solr currently comes s
Can't you just run a jetty server on the background?
But probably some antivirus or antispyware could take that as an tojan or
something like that.
How many little main memory is? 1gb? less?
I don't think that you are going to have problems above 1gb. The index will
be static, no changes, no opt
Hi folks,
we want to migrate our search-portal to Solr.
But some of our customers search in our informations offline with a DVD-Version.
So we want to estimate the complexity of a Solr DVD-Version.
This means to trim Solr to work on small computers with the opposite of heavy
loads. So no server-o
Hi,
Thank you for this contribution. Such a diagram could be useful in the
official documentation.
David
On Thu, Apr 7, 2011 at 12:15 PM, Jeffrey Chang wrote:
> This is awesome; thank you!
>
> On Thu, Apr 7, 2011 at 6:09 PM, Jan Høydahl wrote:
>
> > Hi,
> >
> > Glad you liked it. You'd like t
Have you looked at solr sharding?
Best
Erick
On Thu, Apr 7, 2011 at 2:24 AM, Sean Bigdatafun
wrote:
> Thanks, David.
>
> I am thinking of a scenario that billions of objects, whose indices are too
> big for a single machine to serve the indexing, to serve the querying. Is
> there any sharding me
Trunk. That's what "next" means in the "fix version" field.
Note that the patch is "as is", it's not guaranteed. The
trunk code may well have moved on so use at your own
risk!
Best
Erick
On Wed, Apr 6, 2011 at 11:44 PM, Isha Garg wrote:
>
>
> Hi,
> Tell me for which solr version does Patch
Hmmm, depends on how you stop the server, I was
assuming you did something radical like 'kill -9' (for SHAME )
or the machine crashed or something else horrible...
Koji was covering graceful shutdown, thanks Koji! I hadn't
even considered that
Erick
On Wed, Apr 6, 2011 at 7:19 PM, Robert Pet
OK, see below.
On Wed, Apr 6, 2011 at 6:22 PM, Preston Marshall wrote:
> Reply Inline:
> On Apr 6, 2011, at 8:12 AM, Erick Erickson wrote:
>
> > Hmmm, this should work just fine. Here are my questions.
> >
> > 1> are you absolutely sure that the new synonym file
> > is available when reindexi
Problem solved. *bangs head on desk*
T
On 7 April 2011 11:33, Tom Mortimer wrote:
> Hi,
>
> I'm having trouble getting highlighting to work for a large text
> field. This field can be in several languages, so I'm sending it to
> one of several fields configured appropriately (e.g. "cv_text_en") a
Hi Stephen,
I looked through the Ingres documentation but I don't see why this will
happen. It seems that the column is not being detected as a Clob by the
transformer and Object.toString is being invoked.
[1] - http://community.ingres.com/wiki/Manipulating_SQL_CLOB_data_with_JDBC
On Thu, Apr 7,
> I guess what I'm asking is - can Solr
> highlight non-indexed fields?
http://wiki.apache.org/solr/FieldOptionsByUseCase
I guess what I'm asking is - can Solr highlight non-indexed fields?
Tom
On 7 April 2011 11:33, Tom Mortimer wrote:
> Hi,
>
> I'm having trouble getting highlighting to work for a large text
> field. This field can be in several languages, so I'm sending it to
> one of several fields configured
Hi,
I'm having trouble getting highlighting to work for a large text
field. This field can be in several languages, so I'm sending it to
one of several fields configured appropriately (e.g. "cv_text_en") and
then copying it to a common field for storage and display ("cv_text").
The relevant fragme
This is awesome; thank you!
On Thu, Apr 7, 2011 at 6:09 PM, Jan Høydahl wrote:
> Hi,
>
> Glad you liked it. You'd like to model the inner architecture of SolrJ as
> well, do you? Perhaps that should be a separate diagram.
>
> --
> Jan Høydahl, search solution architect
> Cominvent AS - www.comin
Hi,
Glad you liked it. You'd like to model the inner architecture of SolrJ as well,
do you? Perhaps that should be a separate diagram.
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
On 6. apr. 2011, at 12.06, Stevo Slavić wrote:
> Nice, thank you!
>
> Wish there wa
Hi,
This is how I have shared schema between several cores. Also you can use ${}
syntax in your solrconfig.xml's to reference shared conf files.
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
On 7. apr. 2011, at 02.13, Mark wrote:
> I
Thanks for the input guys.
I've decided to implement some unit tests for now, although we don't have a
clean data set to work from (sucks, I know).
We're going to keep track of a set of vital queries, and ensure they don't
return 0 results, as we have a pretty decent level of confidence with Solr
Hi all,
When I index a column(image type) of a table via *
http://localhost:8080/solr/dataimport?command=full-import*
*There is a error like this: String length must be a multiple of four.*
Any help?
Thank you very much.
PS. the attachment includes Chinese character.
--
You could use the replication to replicate the configuration files :
http://wiki.apache.org/solr/SolrReplication
What do you want to do with your different cores ?
Ludovic.
-
Jouve
France.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Shared-conf-tp2787771p2789447.htm
You opened the same thread this monday and got two replies.
> Hi,
> Has anyone indexed the data with Trade Mark symbol??...when i tried to
> index, the data appears as below... I want to see the Indexed data with TM
> symbol
>
> Indexed Data:
> 79797 - Siebel Research AI Fund,
> 79797
Dear all,
while having some warnings in Solr 3.1.0 log files
07.04.2011 09:08:50 org.apache.solr.request.SolrQueryResponse
WARNING: org.apache.solr.request.SolrQueryResponse is deprecated.
Please use the corresponding class in org.apache.solr.response
I recommend cleaning up /admin/replication
Hi,
Has anyone indexed the data with Trade Mark symbol??...when i tried to
index, the data appears as below... I want to see the Indexed data with TM
symbol
Indexed Data:
79797 - Siebel Research AI Fund,
79797 - Siebel Research AI Fund,l
Original Data:
79797 - Siebel Researc
You can't view it online, but you should be able to download it from:
https://docs.google.com/leaf?id=0BwOEbnJ7oeOrNmU5ZThjODUtYzM5MS00YjRlLWI
2OTktZTEzNDk1YmVmOWU4&hl=en&authkey=COGel4gP
Enjoy,
Ephraim Ofir
-Original Message-
From: Jens Mueller [mailto:supidupi...@googlemail.com]
Sent:
87 matches
Mail list logo