You can't view it online, but you should be able to download it from:
https://docs.google.com/leaf?id=0BwOEbnJ7oeOrNmU5ZThjODUtYzM5MS00YjRlLWI
2OTktZTEzNDk1YmVmOWU4&hl=en&authkey=COGel4gP
Enjoy,
Ephraim Ofir
-Original Message-
From: Jens Mueller [mailto:supidupi...@googlemail.com]
Sent:
Hi,
Has anyone indexed the data with Trade Mark symbol??...when i tried to
index, the data appears as below... I want to see the Indexed data with TM
symbol
Indexed Data:
79797 - Siebel Research AI Fund,
79797 - Siebel Research AI Fund,l
Original Data:
79797 - Siebel Researc
Dear all,
while having some warnings in Solr 3.1.0 log files
07.04.2011 09:08:50 org.apache.solr.request.SolrQueryResponse
WARNING: org.apache.solr.request.SolrQueryResponse is deprecated.
Please use the corresponding class in org.apache.solr.response
I recommend cleaning up /admin/replication
You opened the same thread this monday and got two replies.
> Hi,
> Has anyone indexed the data with Trade Mark symbol??...when i tried to
> index, the data appears as below... I want to see the Indexed data with TM
> symbol
>
> Indexed Data:
> 79797 - Siebel Research AI Fund,
> 79797
You could use the replication to replicate the configuration files :
http://wiki.apache.org/solr/SolrReplication
What do you want to do with your different cores ?
Ludovic.
-
Jouve
France.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Shared-conf-tp2787771p2789447.htm
Hi all,
When I index a column(image type) of a table via *
http://localhost:8080/solr/dataimport?command=full-import*
*There is a error like this: String length must be a multiple of four.*
Any help?
Thank you very much.
PS. the attachment includes Chinese character.
--
Thanks for the input guys.
I've decided to implement some unit tests for now, although we don't have a
clean data set to work from (sucks, I know).
We're going to keep track of a set of vital queries, and ensure they don't
return 0 results, as we have a pretty decent level of confidence with Solr
Hi,
This is how I have shared schema between several cores. Also you can use ${}
syntax in your solrconfig.xml's to reference shared conf files.
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
On 7. apr. 2011, at 02.13, Mark wrote:
> I
Hi,
Glad you liked it. You'd like to model the inner architecture of SolrJ as well,
do you? Perhaps that should be a separate diagram.
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
On 6. apr. 2011, at 12.06, Stevo Slavić wrote:
> Nice, thank you!
>
> Wish there wa
This is awesome; thank you!
On Thu, Apr 7, 2011 at 6:09 PM, Jan Høydahl wrote:
> Hi,
>
> Glad you liked it. You'd like to model the inner architecture of SolrJ as
> well, do you? Perhaps that should be a separate diagram.
>
> --
> Jan Høydahl, search solution architect
> Cominvent AS - www.comin
Hi,
I'm having trouble getting highlighting to work for a large text
field. This field can be in several languages, so I'm sending it to
one of several fields configured appropriately (e.g. "cv_text_en") and
then copying it to a common field for storage and display ("cv_text").
The relevant fragme
I guess what I'm asking is - can Solr highlight non-indexed fields?
Tom
On 7 April 2011 11:33, Tom Mortimer wrote:
> Hi,
>
> I'm having trouble getting highlighting to work for a large text
> field. This field can be in several languages, so I'm sending it to
> one of several fields configured
> I guess what I'm asking is - can Solr
> highlight non-indexed fields?
http://wiki.apache.org/solr/FieldOptionsByUseCase
Hi Stephen,
I looked through the Ingres documentation but I don't see why this will
happen. It seems that the column is not being detected as a Clob by the
transformer and Object.toString is being invoked.
[1] - http://community.ingres.com/wiki/Manipulating_SQL_CLOB_data_with_JDBC
On Thu, Apr 7,
Problem solved. *bangs head on desk*
T
On 7 April 2011 11:33, Tom Mortimer wrote:
> Hi,
>
> I'm having trouble getting highlighting to work for a large text
> field. This field can be in several languages, so I'm sending it to
> one of several fields configured appropriately (e.g. "cv_text_en") a
OK, see below.
On Wed, Apr 6, 2011 at 6:22 PM, Preston Marshall wrote:
> Reply Inline:
> On Apr 6, 2011, at 8:12 AM, Erick Erickson wrote:
>
> > Hmmm, this should work just fine. Here are my questions.
> >
> > 1> are you absolutely sure that the new synonym file
> > is available when reindexi
Hmmm, depends on how you stop the server, I was
assuming you did something radical like 'kill -9' (for SHAME )
or the machine crashed or something else horrible...
Koji was covering graceful shutdown, thanks Koji! I hadn't
even considered that
Erick
On Wed, Apr 6, 2011 at 7:19 PM, Robert Pet
Trunk. That's what "next" means in the "fix version" field.
Note that the patch is "as is", it's not guaranteed. The
trunk code may well have moved on so use at your own
risk!
Best
Erick
On Wed, Apr 6, 2011 at 11:44 PM, Isha Garg wrote:
>
>
> Hi,
> Tell me for which solr version does Patch
Have you looked at solr sharding?
Best
Erick
On Thu, Apr 7, 2011 at 2:24 AM, Sean Bigdatafun
wrote:
> Thanks, David.
>
> I am thinking of a scenario that billions of objects, whose indices are too
> big for a single machine to serve the indexing, to serve the querying. Is
> there any sharding me
Hi,
Thank you for this contribution. Such a diagram could be useful in the
official documentation.
David
On Thu, Apr 7, 2011 at 12:15 PM, Jeffrey Chang wrote:
> This is awesome; thank you!
>
> On Thu, Apr 7, 2011 at 6:09 PM, Jan Høydahl wrote:
>
> > Hi,
> >
> > Glad you liked it. You'd like t
Hi folks,
we want to migrate our search-portal to Solr.
But some of our customers search in our informations offline with a DVD-Version.
So we want to estimate the complexity of a Solr DVD-Version.
This means to trim Solr to work on small computers with the opposite of heavy
loads. So no server-o
Can't you just run a jetty server on the background?
But probably some antivirus or antispyware could take that as an tojan or
something like that.
How many little main memory is? 1gb? less?
I don't think that you are going to have problems above 1gb. The index will
be static, no changes, no opt
I haven't used PostGIS so I can't offer a real comparison. I think if you were
to try out both, you'd be impressed with Solr's performance/scalability thanks
in large part to its sharding. But for "functionality richness" in so far as
geospatial is concerned, that's where Solr currently comes s
Hello every body,
I am using Solr for indexing and searching.
I am using 2 classes for searching document: In the first one I'm
instanciating a SolrServer to search documents as follows :
server = new EmbeddedSolrServer(
coreContainer, "");
server.add(doc);
query.setQuery("id:"+idDoc);
server.qu
Hi All,
I'd like to make the highlighting work as follows:
length(all snippits) approx. 200 chars
hl.snippits = 2 (2 snippits)
is this possible with the regex fragmenter? or does anyone know of any
contrib fragmenter that might do this?
Many thanks
Dan
Even if you can ship your DVD with a jetty server, you'll still need JAVA
installed on the customer machine...
On Thu, Apr 7, 2011 at 10:18 PM, Ezequiel Calderara wrote:
> Can't you just run a jetty server on the background?
>
> But probably some antivirus or antispyware could take that as an toj
Hi,
We are trying to index heterogenous data using SOLR, some of the sources
have some unicode characters like Zone™ but SOLR is converting them to
Zone. Any idea how to resolve this issue?
I am using SOLR on Jetty server...
Thanks,
Barani
--
View this message in context:
http://lucene.4720
On Thu, Apr 7, 2011 at 10:28 AM, Jeffrey Chang wrote:
> Even if you can ship your DVD with a jetty server, you'll still need JAVA
> installed on the customer machine...
Including a JRE on the DVD and a launch script that uses that JRE by
default should be doable as well.
-Yonik
http://www.lucene
Well at this point I'm more dedicated to the Deduplicate issue.
Using a Min_token_len of 4 I'm getting nice comparison results. MLT returns a
lot of similar docs that I don't consider similar - even tuning the parameters.
Finishing this issue, I found out that the signature also contains the fie
On Apr 6, 2011, at 10:29 PM, Jens Mueller wrote:
> Walter, thanks for the advice: Well you are right, mentioning google. My
> question was also to understand how such large systems like google/facebook
> are actually working. So my numbers are just theoretical and made up. My
> system will be smal
That's probably an issue of your analyzer. Can you show us the field
definition from the schema.xml file, for the field that you are putting
this text in?
On 4/7/2011 10:37 AM, bbarani wrote:
Hi,
We are trying to index heterogenous data using SOLR, some of the sources
have some unicode chara
The data are fine and not duplicated - however, I want to analyze the data, and
summarize one field (kind of like faceting), to understand what the largest
value is.
For example:
Document 1: label=1A1A1; body="adfasdfadsfasf"
Document 2: label=5A1B1; body="adfaasdfasdfsdfadsfasf"
Document 3
I'd advise getting a copy of Luke and examining your
indexes. The information you've provided doesn't really
tell us much.
Although I do notice you don't commit in your example code...
Best
Erick
On Thu, Apr 7, 2011 at 10:21 AM, Amel Fraisse wrote:
> Hello every body,
>
> I am using Solr for in
What version of Solr are you using? And, assuming the version that
has it in, have you seen grouping?
Which is another way of asking why you want to do this, perhaps it's an
XY problem
Best
Erick
On Thu, Apr 7, 2011 at 1:13 AM, Peter Spam wrote:
> Hi,
>
> I have documents with a field that
Hi,
Thanks for your response..please find below the schema details corresponding
to that field..
---
Field type details..
Thanks,
Barani
--
View this message
Hi all,
I've been using MoreLikeThis for a while through select:
http://localhost:8983/solr/select/?q=field:more like
this&mlt=true&mlt.fl=field&rows=100&fl=*,score
I was looking over the wiki page today and saw that you can also do this:
http://localhost:8983/solr/mlt/?q=field:more like
this&m
Hi yonik, Hi Ezequiel,
Java is no problem for an DVD Version. We already have a DVD version with
Servlet-Container (but this does currently not use Solr).
Some of our customers work in public sector institutions and have less then 1gb
main memory, but they use MS Word and IE and..
But let us s
Does this contain the CollapseComponent?
Will there be a significant performance boost from 1.4?
Try setting a virtual machine and see its performance.
I'm really not a java guy, so i really don't know how to tune it for
performance...
But afaik solr handles pretty well in ram if the index is static...
On Thu, Apr 7, 2011 at 2:48 PM, Karsten Fissmer wrote:
> Hi yonik, Hi Ezequiel,
>
> Jav
Hey guys,
I am in the process of moving to solr3.1 from solr1.4. I am having this
issue where solr3.1 now complains about the synonyms.txt file. I get the
following error:
*org.apache.solr.common.SolrException: Error loading resource (wrong
encoding?): synonyms.txt*
*
*
This worked fine before in
Hi,
I'd would like to experiment with the UIMA contrib package, but I have
issues with the OpenCalais service's ToS and would rather not use it.
Is there a way to adapt the UIMA example setup to use only the
AlchemyAPI service? I tried simply leaving out the OpenCalais api key
but i get exceptions
On Thu, Apr 7, 2011 at 2:13 PM, Siddharth Powar
wrote:
> Hey guys,
>
> I am in the process of moving to solr3.1 from solr1.4. I am having this
> issue where solr3.1 now complains about the synonyms.txt file. I get the
> following error:
> *org.apache.solr.common.SolrException: Error loading resour
Actually, what is the difference between "match" and "response"? It seems
that match always returns one result but I've thrown a few cases at it where
the score of the highest response is higher than the score of match. And
then there are cases where the match score dwarfs the highest response
scor
Hi Ezequiel,
In Solr the performance of sorting and faceted search is mainly a question of
main memory.
e.g Mike McCandless wrote in s.apache.org/OWK that sorting of 5m wikipedia
documents by title field need 674 MB of RAM.
But again: My main interest is an example of other companies/product wh
Hello everyone. Does anyone else have problems posting to the list? My
messages keep getting rejected with this response below. I'll be surprised if
this one makes it through :)
-Park
Sorry, we were unable to deliver your message to the following address.
:
Remote host said: 552 spam scor
Happened to me a couple of times, couldn't find a way a workaround...
On Thu, Apr 7, 2011 at 4:14 PM, Parker Johnson wrote:
>
> Hello everyone. Does anyone else have problems posting to the list? My
> messages keep getting rejected with this response below. I'll be surprised
> if
> this one m
On Thu, Apr 07, 2011 at 04:21:25PM -0300, Ezequiel Calderara wrote:
> Happened to me a couple of times, couldn't find a way a workaround...
Note that the property "HTML_MESSAGE" has contributed to the email's spam
score:
> > (FREEMAIL_ENVFROM_END_DIGIT,FREEMAIL_FROM,FS_REPLICA,HTML_MESSAGE,RCVD_I
This happens almost always because you're sending from a 'free' mail
account (gmail, yahoo, hotmail, etc), and your message contains words
that spam filters don't like.
For me, it was the use of the word 'remplica' (deliberately
mis-spelled so this mail gets sent).
It can also happen from 'non-fre
Hi Park
I had the same problem. I noticed one of the issues with the blocked
messages are they are HTML/Rich Text.
(FREEMAIL_ENVFROM_END_DIGIT,FREEMAIL_FROM,FS_REPLICA,
HTML_MESSAGE
<-,RCVD_IN_DNSWL_NONE,RFC_ABUSE_POST,SPF_PASS,T_TO_NO_BRKTS_FREEMAIL
In GMail I can switch to plain text. This
I noticed that Lucid Works distribution now says is upt to date with 4.X
versions. Does this mean 1.4 or 4.0/trunk?
If its truly 4.0 does that mean it includes the collapse component?
Also, is the click scoring tools proprietary or was this just a
contrib/patch that was applied?
Thanks
On 4/7/11 9:43 PM, Mark wrote:
I noticed that Lucid Works distribution now says is upt to date with 4.X
versions. Does this mean 1.4 or 4.0/trunk?
If its truly 4.0 does that mean it includes the collapse component?
Yes it does.
Also, is the click scoring tools proprietary or was this just a
Andrezej,
Thanks for the info. I have a question regarding stability though. How
are you able to guarantee the stability of this release when 4.0 is
still a work in progress? I believe the last version Lucid released was
1.4 so why did you choose to release a 4.x version as opposed to 3.1?
I
Hey Robert,
Thanks for the quick response. That helps a lot.
--Sid
On Thu, Apr 7, 2011 at 11:19 AM, Robert Muir wrote:
> On Thu, Apr 7, 2011 at 2:13 PM, Siddharth Powar
> wrote:
> > Hey guys,
> >
> > I am in the process of moving to solr3.1 from solr1.4. I am having this
> > issue where solr3
Hi All,
I am new to solr. I have gone through solr documents to index pdf files, But
it was hard to find the exact procedure to get started.
I need step by step procedure to do this. Could you please let me know the
steps to index pdf files.
Thanks,
Mike
--
View this message in context:
http://
Why does solr keeps creating connections for each table, even though they are
in the same database? This happened in the process of full-import. I used
one table as the root entity, and join the other tables. But solr keeps
creating database connections for each table. Any one has any idea or had
s
I have a question on how to set up queries not having a predetermined
field list to search on.
Here are some sample docs,
1234
hihello
lalachika chika boom
boom
1235
foobarhappy happy
joy joy
some textsome more words to
search
.
.
.
4567
bedrock
memeyou you
How can one change tomcat specific settings such as tomcat-users.xml? I
can't seem to find any reference to these conf files?
Thanks
Hi,
I wrote my own parser plugin.
I'm getting a NoClassCefFoundError. Any ideas why?
Apr 7, 2011 1:12:43 PM org.apache.solr.common.SolrException log
SEVERE: java.lang.NoClassDefFoundError: Could not initialize class
org.apache.solr.search.QParserPlugin
at org.apache.solr.core.SolrCore.
: of the components as well as the flow of data and queries. The result is
: a conceptual architecture diagram, clearly showing how Solr relates to
: the app-server, how cores relate to a Solr instance, how documents enter
: through an UpdateRequestHandler, through an UpdateChain and Analysis a
:
: Thanks for your response..please find below the schema details corresponding
: to that field..
your message inlcuded nothing but a bunch of blank lines, probably because
your email editor thought you were trying to type in html (instead of xml)
before diving too deeply into your analyser ho
Hi,
SIREn [1], a Lucene/Solr plugin, allows you perform queries across an
undetermined number of fields, even if you have hundred of thousands of
fields. It might be helpful for your scenario.
[1] http://siren.sindice.com
--
Renaud Delbru
On 07/04/11 19:18, jisenhart wrote:
I have a questi
I have 1 Master, and 3 slaves. The master holds the solr index. How do I
connect the slaves to the master? I have the script in the bin folders. I
have rsyncd installed and snapshooter enabled in the master. Thanks, please
help.
Did you try the curl commands that Adam suggested as part of this e-mail
thread?
If so, what happened?
Best
Erick
On Wed, Apr 6, 2011 at 7:50 AM, Mike wrote:
> Hi All,
>
> I am new to solr. I have gone through solr documents to index pdf files,
> But
> it was hard to find the exact procedure to
> Does this contain the
> CollapseComponent?
No, For FieldCollapsing you need trunk.
> I wrote my own parser plugin.
>
> I'm getting a NoClassCefFoundError. Any ideas why?
Did you put jar file - that contains you custom code - into /lib directory?
http://wiki.apache.org/solr/SolrPlugins
Can you give us some more details? I suspect the jar file containing
your plugin isn't in the Solr lib directory and/or you don't have a lib
directive in your solrconfig.xml file pointing to where your jar is.
But that's a guess since you haven't provided any information about
what you did to try
yes.
From: Ahmet Arslan
To: solr-user@lucene.apache.org
Sent: Thu, April 7, 2011 3:23:56 PM
Subject: Re: class not found
> I wrote my own parser plugin.
>
> I'm getting a NoClassCefFoundError. Any ideas why?
Did you put jar file - that contains you custom c
> Date: Friday, April 8, 2011, 1:19 AM
> I have 1 Master, and 3 slaves. The
> master holds the solr index. How do I
> connect the slaves to the master? I have the script in the
> bin folders. I
> have rsyncd installed and snapshooter enabled in the
> master. Thanks, please
HTTP based replication i
The jar containing the class is in here:
/usr/local/apache-tomcat-6.0.20/webapps/solr/WEB-INF/lib
for my setup.
Tri
From: Erick Erickson
To: solr-user@lucene.apache.org
Sent: Thu, April 7, 2011 3:24:14 PM
Subject: Re: class not found
Can you give us some mo
One possibility is to have just a "groups" field with a positionIncrementGap
of, say, 100.
that is multiValued.
Now, index values like
"group1 foo bar happy joy joy"
"group2 some more words to search"
etc.
Now do phrase queries with a slop of less than 100. Then searches like
groups:"group1 more
> The jar containing the class is in
> here:
>
> /usr/local/apache-tomcat-6.0.20/webapps/solr/WEB-INF/lib
>
http://wiki.apache.org/solr/SolrPlugins#How_to_Load_Plugins
Would grouping solve this? I'd rather not move to a pre-release solr ...
To clarify the problem:
The data are fine and not duplicated - however, I want to analyze the data, and
summarize one field (kind of like faceting), to understand what the largest
value is.
For example:
Document 1: la
I think you can specify the in-group sort, and specify a very small number
(perhaps
even one) to go in each group. But you'd have to store the length of each
body and sort by that.
I'm pretty sure grouping is trunk-only.
The problem here is getting something that applies
just within the group and
: Solr 3.1.0 uses different javabin format from 1.4.1
: So if I use Solrj 1.4.1 jar , then i get javabin error while saving to
: 3.1.0
: and if I use Solrj 3.1.0 jar , then I get javabin error while reading the
: document from solr 1.4.1.
you can use the XML format to get portability during the
Hi,
I am developing a SearchComponent that needs to build some initial
DocSets and then intersect with the result DocSet during each query (in
process()).
When the searcher is reopened, I need to regenerate the initial DocSets.
I am on Solr 1.4.1.
My question is, which method in SearchComponent
Hello
I updated Solr to version 3.1 on my project. And now when the application
calls getResponse () method (PECL extension) I get the following:
"Fatal error: Uncaught exception 'SolrException' with message 'Error
un-serializing response' in /home/.../Adapter/Solr.php: 78"
How can I fix it?
Than
I haven't built one myself, but have you considered the Solr UserCache?
See: http://wiki.apache.org/solr/SolrCaching#User.2BAC8-Generic_Caches
It even receives warmup signals I believe...
Best
Erick
On Thu, Apr 7, 2011 at 7:39 PM, Sujit Pal wrote:
> Hi,
>
> I am developing a SearchComponent th
I think I found the answer by looking through the code...specifically
SpellCheckComponent.
So my component would have to implement SolrCoreAware and in the
inform() method, register a custom SolrEventListener which will execute
the regeneration code in the postCommit and newSearcher methods.
Woul
Thanks Erick. This looks like it would work... I sent out an update to
my original query, there is another approach that would probably also
work for my case that is being used by SpellCheckerComponent.
I will check out both approaches.
Thanks very much for your help.
-sujit
On Thu, 2011-04-07
Hi,
Could you send the enter list of parameters you are ending to solr via the
SolrClient and SolrQuery object?
Please open a bug request here with the details
http://pecl.php.net/bugs/report.php?package=solr
On Thu, Apr 7, 2011 at 7:59 PM, Haspadar wrote:
> Hello
> I updated Solr to version
Hi,
I have a table named *attachment *in MS SQL Server 2008.
COLUMNTYPE
-
id int
titlevarchar(200)
attachment image
I need to index the attachment(store pdf files) column from database via
DIH.
After access this URL, it returns "Ind
You need the TikaEntityProcessor to unpack the PDF image. You are
sticking binary blobs into the index. Tika unpacks the text out of the
file.
TikaEP is not in Solr 1.4, but it is in the new Solr 3.1 release.
On Thu, Apr 7, 2011 at 7:14 PM, Roy Liu wrote:
> Hi,
>
> I have a table named *attachme
Thanks Lance,
I'm using Solr 1.4.
If I want to using TikaEP, need to upgrade to Solr 3.1 or import jar files?
Best Regards,
Roy Liu
On Fri, Apr 8, 2011 at 10:22 AM, Lance Norskog wrote:
> You need the TikaEntityProcessor to unpack the PDF image. You are
> sticking binary blobs into the index.
Hello all,
thanks for your generous help.
I think I now know everything: (What I want to do is to build a web crawler
and index the documents found). I will start with the setup as suggested by
Ephraim (Several sharded masters, each with at least one slave for reads and
some aggregators for quer
On Fri, Apr 8, 2011 at 6:23 AM, Jens Mueller wrote:
> Hello all,
>
> thanks for your generous help.
>
> I think I now know everything: (What I want to do is to build a web
> crawler
> and index the documents found). I will start with the setup as suggested by
>
>
Write a web crawler from scratch
I'm entering only a query parameter.
I posted a bug description there - http://pecl.php.net/bugs/bug.php?id=22634
2011/4/8 Israel Ekpo
> Hi,
>
> Could you send the enter list of parameters you are ending to solr via the
> SolrClient and SolrQuery object?
>
> Please open a bug request here with
Hi All,
I was able to set the cookie value to the Stream_url connection, i was
able to pass the cookie value upto contentstreamBase.URLStream class and i
added
conn.setRequestProperty("Cookie",cookie[0].name"="cookie[0].value) in the
connection setup.. and it is working fine now...
Regards,
s
87 matches
Mail list logo