ontent field. Like the default
schema, I'm storing but not indexing the content field, and a lot of the
fields get put into a catchall that is indexed and uses term vectors, but
is not stored.
Thanks,
Mike
On 8/20/2010 9:52 PM, Lance Norskog wrote:
Another way to do this is to set up the "live" core to do replication
from the "standby" core. Replication should work this way, between
different cores in the same Solr instance. This is cleaner than
swapping the two cores.
On Thu, Aug 19, 2010 at 7:2
Hi All,
I am new to solr. I have gone through solr documents to index pdf files, But
it was hard to find the exact procedure to get started.
I need step by step procedure to do this. Could you please let me know the
steps to index pdf files.
Thanks,
Mike
--
View this message in context:
http
application. Can I run solr in the same
application server(Tomcat6) in which my actual application resides.
Thanks,
Mike
--
View this message in context:
http://lucene.472066.n3.nabble.com/Indexing-pdf-files-question-tp2079505p2794841.html
Sent from the Solr - User mailing list archive at Nabble.com.
help me out in fixing the above issue.
Thanks,
Mike
--
View this message in context:
http://lucene.472066.n3.nabble.com/Tika-Solr-running-under-Tomcat-6-on-Debian-tp993295p2805615.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi All,
I have installed solr instance on tomcat6. When i tried to index the PDF
file i was able to see the response:
0
479
Query:
http://localhost:8080/solr/update/extract?stream.file=D:\mike\lucene\apache-solr-1.4.1\example\exampledocs\Struts%202%20Design%20and%20Programming1.pdf
Hi Roy,
Thank you for the quick reply. When i tried to index the PDF file i was able
to see the response:
0
479
Query:
http://localhost:8080/solr/update/extract?stream.file=D:\mike\lucene\apache-solr-1.4.1\example\exampledocs\Struts%202%20Design%20and%20Programming1.pdf&stream.content
-config.xml and
also on schema.xml.
Is it possible to do something like:
Thanks,
Mike
--
View this message in context:
http://lucene.472066.n3.nabble.com/Is-there-a-way-to-create-multiple-doc
y to accomplish this in 1.3?
Mike
day, July 30, 2009 8:05 AM
Subject: Re: Multi select faceting
On Jul 29, 2009, at 2:38 PM, Mike wrote:
Hi,
We're using Lucid Imagination's LucidWorks Solr 1.3 and we have a
requirement to implement multiple-select faceting where the facet cells
show up as checkboxes and despi
y tips on where to look would be greatly appreciated. I'm using the trunk
version of SOLR built on 11/1.
Mike
have solr.data.dir populated via solr.xml when
using Multicore? Maybe you are experiencing the same issue.
Mike
r in this example.
I'm using the dismax and the 1.4 release.
Mike
Mike wrote:
When I do a search using q=*:* and then narrow down the result set
using a filter query, are there rules that are used for the sort order
in the result set? In my results I have a "name" field that appears to
be sorted descending in lexicographical order. For example:
Yonik Seeley wrote:
On Fri, Nov 20, 2009 at 11:15 AM, Mike wrote:
Sorry for the noise - I think I have just answered my own question. The
order in which docs are indexed determine the result sort order unless
overridden via sort query parameters :)
Correct. The internal lucene
Blargy wrote:
Posted a few weeks ago about this but no one seemed to respond. Has anyone
seen this before? Why is this happening and more importantly how can I fix
it? Thanks in advance!
May 11, 2010 12:05:45 PM org.apache.solr.handler.dataimport.DataImporter
doDeltaImport
SEVERE: Delta Import F
Greetings
I've gotten SOLR installed and the admin screens working. At this
point I'm just trying to get my add record to be grabbed by the SOLR
update process, but unfortunately, I'm getting a whitespace error
that I could use some pointers on. I've searched the site and found
similar
I thought that too, I opened it up via vi and nothing was there.
Usually if I have a PC encoding issue ( I use EditPlus as a text
editor) it will show up in vi.
On Apr 26, 2007, at 5:19 PM, Ryan McKinley wrote:
paladin:/data/solr mtorgler1$ curl http://localhost:8080/solr/
update --data-bin
rapping only in this email
or is the source document wrapping like that as well?
/cody
On 4/26/07, Mike <[EMAIL PROTECTED]> wrote:
Greetings
I've gotten SOLR installed and the admin screens working. At this
point I'm just trying to get my add record to be grabbed by the SOLR
up
On Apr 27, 2007, at 4:24 AM, Burkamp, Christian wrote:
curl http://localhost:8080/solr/update --data-binary @
I think the issue was with a bad file in /data/solr/conf. After a
bunch of testing out of the installtion directory, I was able to use
the post.sh script to post correctly into t
:8983
_solr,192.168.56.21:8983_solr,192.168.56.22:8983_solr,192.168.56.23:8983
_solr&collection.configName=default-collection
Thanks,
Mike
file or keep it? We have 4
Solr VMs in our devops environment. They have 2 CPUs and 4GB of RAM. There
are about 7 collections shared between then, but all are negligible (like a
few hundred kb each) except for one which is about 22GB.
Thanks,
Mike
elect (tried changing it to POST and moving query string to the
body with Firefox dev tools, but that failed).
Is there a way to keep doing something like what we were doing or do we
need to limit ourselves to GETs? I think our queries are all small enough
now for that, but it would helpful to know for planning.
Thanks,
Mike
content types?
>
> -Yonik
>
>
> On Wed, Feb 24, 2016 at 8:48 AM, Mike Thomsen
> wrote:
> > With 4.10, we used to post JSON like this example (part of it is Python)
> to
> > /select:
> >
> > {
> > "q": "LONG_QUERY_HERE",
e? I sent updates just fine a day or two ago like
that, now it is acting like the update request is a whole new document.
Thanks,
Mike
I have a three node, one shard SolrCloud cluster.
Last week one of the nodes went out of sync with the other two and I'm
trying to understand why that happened.
After poking through my logs and the solr code here's what I've pieced
together:
1. Leader gets an update request for a batch delete of
I should add that this is on Solr 5.1.0.
On Thu, Apr 28, 2016 at 2:42 PM, Mike Wartes wrote:
> I have a three node, one shard SolrCloud cluster.
>
> Last week one of the nodes went out of sync with the other two and I'm
> trying to understand why that happened.
>
> After
The SVN source is under tags, not branches.
http://svn.apache.org/repos/asf/lucene/dev/tags/lucene_solr_4_10_3/
On Tue, Feb 17, 2015 at 4:39 PM, O. Olson wrote:
> Thank you Hrishikesh. Funny how GitHub is not mentioned on
> http://lucene.apache.org/solr/resources.html
>
> I think common-build.
rward about how to call the API to create a new collection. Is
it just the Solr version I'm using being too old for the current wiki docs
or am I using it incorrectly?
My goal here is to create a process where we can easily programatically
create new collections for users in a web application.
Thanks,
Mike
find the right location to put
it on our system. Sound right?
https://cwiki.apache.org/confluence/display/solr/Config+Sets
Thanks,
Mike
On Sat, Mar 14, 2015 at 2:27 PM, Erick Erickson
wrote:
> I bet you did not push the configuration to Zookeeper before creating
> the collection.
>
ce an update into zookeeper? Or should I
just purge the zookeeper data?
On Sat, Mar 14, 2015 at 3:02 PM, Mike Thomsen
wrote:
> I looked in the tree view and I have only a node called "configs." Nothing
> called "configsets." That's a serious problem, right? So if I&
I tried that with upconfig, and it created it under /configs. Our ZK
configuration data is under /dev-local-solr/configs. Not sure how to
specify that. Also, is "configs" the same thing as "configsets" for the
version of solr that I'm using?
Thanks,
Mike
On Sat, Mar 14
; Now whenever one of the replicas for that collection starts up, it contact
> ZK and reads the config files and starts up. The replica does _not_
> copy the files locally.
>
> HTH,
> Erick
>
> On Sun, Mar 15, 2015 at 6:16 AM, Mike Thomsen
> wrote:
> > I tried that w
Probably a historical artifact.
cz is the country code for the Czech Republic, cs is the language code for
Czech. Once, cs was also the country code for Czechosolvakia, leading some
folks to accidentally conflate the two.
On Tue, Mar 17, 2015 at 12:35 PM, Eduard Moraru
wrote:
> Hi,
>
> First of
I'm trying out schemaless in solr 5.0, but the indexing seems quite a
bit slower than it did in the past on 4.10. Any pointers?
--Mike
I start up solr schemaless and index a bunch of data, and it takes a
lot longer to finish indexing.
No configuration changes, just straight schemaless.
--Mike
On Sun, Mar 22, 2015 at 12:27 PM, Erick Erickson
wrote:
> Please review: http://wiki.apache.org/solr/UsingMailingLists
>
> Yo
That's it!
I hand edited the file that says you are not supposed to edit it and
removed that copyField.
Indexing performance is now back to expected levels.
I created an issue for this, https://issues.apache.org/jira/browse/SOLR-7284
--Mike
On Sun, Mar 22, 2015 at 3:29 PM, Yonik Seeley
tion being used:
Thanks in advance.
Mike
Thanks Jack! That was oversight on my end - I also assumed the
splitOnNumerics="1" and LowerCaseFilterFactory would be breaking out the
tokens. I tried again with generateWordParts="1" generateNumberParts="1" and it
seemed to work. Appreciate it.
Mike
F
the synonym rule?
Thanks in advance!
Mike
Typo: *even when the user delimits with a space. (e.g. base ball should find
baseball).
Thanks,
From: Mike L.
To: "solr-user@lucene.apache.org"
Sent: Tuesday, April 7, 2015 9:05 AM
Subject: DictionaryCompoundWordTokenFilterFactory - Dictionary/Compound-Words
File
to this?
Thanks,Mike
Thanks Jack. I'll give that a whirl.
From: Jack Krupansky
To: solr-user@lucene.apache.org; Mike L.
Sent: Saturday, April 11, 2015 12:04 PM
Subject: Re: Bq Question - Solr 4.10
It all depends on what you want your scores to look like. Or do you care at all
what the scores
ApacheCon is starting tomorrow, so seeing if I pulling up this thread
yields any new replies this time. I'm hanging out in Austin, looking
forward to some good conversations and sessions!
On Wed, Feb 18, 2015 at 9:14 PM, CP Mishra wrote:
> Dmitry, that would be great.
>
> CP
>
> On Thu, Feb 12,
something like this with our schema to even get the
synonyms API working?
I wanted to ask before changing our schema.
Thanks,
Mike
d pretty important dude in the Hive was
discovered alive and well in the hellmouth today!"
]
}
When I use *dummy_text: Crouton*, nothing comes back. I am pretty confident
that I am missing something here. Any ideas?
Thanks,
Mike
On Wed, Apr 15, 2015 at 3:04 PM, Yonik Seeley wr
"initializedOn":"2015-04-14T19:39:55.157Z",
> "managedMap":{
> "GB":["GiB",
> "Gigabyte"],
> "TV":["Television"],
> "happy":["glad",
> &q
gt;
>> http://yonik.com/json-facet-api/
>> http://yonik.com/solr-facet-functions/
>> http://yonik.com/solr-subfacets/
>>
>> I'll also be hanging out more on the #solr-dev IRC channel on freenode
>> if you want to hit me up there about any development ideas.
>>
>> -Yonik
--Mike
This is the content of my autophrases.txt file:
al qaeda in the arabian peninsula
seat belt
I've attached a screenshot showing the analysis view of the index. When I
query for al_qaeda_in_the_arabian_peninsula or
alqaedainthearabianpeninsula, nothing comes back even though at least the
latter app
Is there a way to manage synonyms through Solr's Java API? Google doesn't
turn up any good results, and I didn't see anything in the javadocs that
looked promising.
Thanks,
Mike
What do the various Java IDEs use for indexing classes for
field/type/variable/method usage search? I imagine it's got to be bytecode.
On Fri, May 8, 2015 at 2:40 PM, Tomasz Borek wrote:
> Out of curiosity: why bytecode?
>
> pozdrawiam,
> LAFK
>
> 2015-05-08 21:31 GMT+02:00 Mark :
>
> > I lookin
SHARD
if that helps.
Attached is a screenshot of the Cloud view in the admin console after
running splitshard.
Should it look like that? Do I need to delete shard1 now?
Thanks,
Mike
Thanks. I thought it worked like that, but didn't want to jump to
conclusions.
On Thu, Jun 4, 2015 at 1:42 PM, Anshum Gupta wrote:
> Hi Mike,
>
> Once the SPLITSHARD call completes, it just marks the original shard as
> Inactive i.e. it no longer accepts requests. So yes, you
rch_text:\"baltimore ? police ? ? force\")",
"parsedquery_toString": "search_text:\"baltimore ? police ? ? force\"",
"QParser": "LuceneQParser",
Thanks,
Mike
I'm doing some debugging work on a solr core, and would find it useful to
be able to pretty print the contents of the segments_N file in the index.
Is there already good functionality for this, or will I need to write up my
own utility using SegmentInfos?
Thanks,
Mike
at
org.apache.solr.rest.schema.analysis.ManagedStopFilterFactory.(ManagedStopFilterFactory.java:47)
This is the filter definition I used:
Any ideas?
Thanks,
Mike
I can index
say a 500kb block of text like this?
Thanks,
Mike
ene, the underlying search engine library, imposes this 32K limit for
> individual terms. Use tokenized text instead.
>
> -- Jack Krupansky
>
> On Thu, Jun 25, 2015 at 8:36 PM, Mike Thomsen
> wrote:
>
> > I need to be able to do exact phrase searching on some documents that
> ar
Solr User Group -
Was wondering if anybody had any suggestions/best practices around a
requirement for storing a dynamic category structure that needs to have the
ability to facet on and maintain its hierarchy
Some context:
A product could belong to an undetermined amount of product categor
categories and can maintain the
hierarchy..
I'll take a look at it.
Thanks!
From: Erick Erickson
To: solr-user@lucene.apache.org; Mike L.
Sent: Monday, July 6, 2015 12:42 PM
Subject: Re: Category Hierarchy on Dynamic Fields - Solr 4.10
Hmmm, probably missing something her
Are the clients that are posting updates requesting commits?
On Tue, Jul 7, 2015 at 4:29 PM, Summer Shire wrote:
> HI All,
>
> Can someone help me understand the following behavior.
> I have the following maxTimes on hard and soft commits
>
> yet I see a lot of Opening Searchers in the log
> org
also a result
of our stemmer?
Here's the text field we're using:
Thanks,
Mike
nnection="jdbc:mysql://localhost/users?user=root&password=solr",
sql="SELECT id, name FROM users", sort="id asc",
driver="com.mysql.jdbc.Driver")
)
What's the best way to handle a multivalue field using this API? Is
there a way to tokenize something returned in a database field?
Thanks,
Mike
and I want to use soft
commits to power the alerts feature.
Thanks,
Mike
minutes after every
hour, I'd see stalls.
2. Watching the logs, I saw a flood of queries come through after the line:
Loaded external value source external_pagerank
Some queries were coming through before this line, but I think none of
those queries use the external file field (external_pagerank).
Mike
gher? I guess it's a RAM utilization vs.
speed tradeoff? A low number like 16 seems like it'd have minimal impact on
RAM?
Thanks for all the great replies and for everything you do for Solr. I
truly appreciate your efforts.
Mike
Thanks for the replies. I made the changes so that the external file field
is loaded per:
only written because the result
grouping had such bad performance -- In other words, it doesn't exist to
provide unique features, it exists to be faster than the old way. Maybe we
can get rid of one or the other of these, taking the best parts from each
(syntax from Result Grouping, and performanc
proceed on configuring this because the documentation is a
bit sparse on what some of those values mean in this context. The example
looked geared toward someone using HDFS both to store the index and do
backup/restore.
Thanks,
Mike
x27;m not averse to using the API to dump a copy of each shard to HDFS.
Something like this:
/solr/collection/replication?command=backup&name=shard_1_1&numberToKeep=7
Is that a viable route to achieve this or do we need to do something else?
Thanks,
Mike
v 9, 2016 at 2:18 PM, Hrishikesh Gadre
wrote:
> Hi Mike,
>
> I filed SOLR-9744 <https://issues.apache.org/jira/browse/SOLR-9744> to
> track this work. Please comment on this jira if you have any suggestions.
>
> Thanks
> Hrishikesh
>
>
> On Wed, Nov 9, 20
l you more about the specific error.
Thanks,
Mike
/new has been almost completely deleted by
Solr (this is a bummer, since it takes days to populate).
Is this expected? Am I terribly crazy to try to swap indexes on disk? As
far as I know, the only difference between the indexes is their name.
We're using Solr version 4.10.4.
Thank you,
Mike
a changes, which my first attempt included, but I guess that'll have
to wait.
Thanks for the replies. If anybody can explain why the first strategy
failed, I'd still be interested in learning.
Mike
On Wed, Jun 14, 2017 at 12:09 PM Chris Ulicny wrote:
> Are you physically swapping th
+solr-user
Might get a different audience on this list.
-- Forwarded message --
From: Christine Poerschke (BLOOMBERG/ LONDON)
Date: Fri, Jun 16, 2017 at 11:43 AM
Subject: (how) do folks use the Cloud Graph (Radial) in the Solr Admin UI?
To: d...@lucene.apache.org
Any thoughts o
e SWAP index functionality, so the
above is just in case somebody wants to try to track this down. I'll
probably take those logs offline after a week or two.
Mike
On Tue, Jun 20, 2017 at 7:20 AM Shawn Heisey wrote:
> On 6/14/2017 12:26 PM, Mike Lissner wrote:
> > We are replacing
les. So at that point, the index is gone.
I'd love to prevent this from happening a third time. It's super baffling.
Any ideas?
Mike
On Tue, Jun 20, 2017 at 12:38 PM Mike Lissner <
mliss...@michaeljaylissner.com> wrote:
> Thanks for the suggestions everybody.
>
>
The two collection approach with aliasing is a good approach.
You can also use the backup and restore APIs -
https://lucene.apache.org/solr/guide/6_6/making-and-restoring-backups.html
Mike
On Wed, Jul 12, 2017 at 10:57 AM, Vincenzo D'Amore
wrote:
> Hi,
>
> I'm moving to S
ds to be able to open it first, so we reasoned that we'd
have no way to demonstrate correctness and at that point restoring from a
backup was the best option.
Maybe somebody smarter or more determined has a better experience.
Mike
On Tue, Aug 8, 2017 at 10:21 AM, Shawn Heisey wrote:
> O
When you specify the zk string for a solr instance, you typically include a
chroot in it. I think the default is /solr, but it doesn't have to be, so
you should be able to run with -z zk1:2181/sorl-dev and /solr-prod
https://lucene.apache.org/solr/guide/6_6/setting-up-an-external-zookeeper-ensembl
Given that the already public nature of the disclosure, does it make sense
to make the work being done public prior to release as well?
Normally security fixes are kept private while the vulnerabilities are
private, but that's not the case here...
On Mon, Oct 16, 2017 at 1:20 AM, Shalin Shekhar M
ble? I've been googling about this for a day or so and
can't quite find anything definitive. I'm going to maybe try and dive into
the solr source code, but I'm a c# guy, not java, without a debuggable
environment as unneeded yet, and that could prove pretty painful.
Any help would be appreciated, even if it is just "can't be done", as at
least I could stop chasing my tail.
Mike
need a whitespace and where I
don't! However, I spent a part of the weekend fiddling around with spaces and
+'s and I believe I've got it working as I'd hoped.
Again, many thanks,
Mike
-Original Message-
From: Mikhail Khludnev [mailto:m...@apache.org]
Sent: 18 Nov
orward etc.
Regards,
Alex.
http://www.solr-start.com/ - Resources for Solr users, new and experienced
On 21 November 2016 at 23:57, Mike Allen
wrote:
> Hi Mikhail,
>
> Thanks for your advice, it went a long way towards helping me get the right
> documents in the first place, es
re "+".
Correct encoding seems half the battle to be honest.
So that's what I've got for now, but I wouldn't take it as gospel that it's
working correctly. I'm still validating by hand checking the results I would
expect versus the results I actually get. For
. Any
suggestions on what might be getting us into this situation? Also, is there
a SolrCloud API for forcing those two replicas to sync with the third or do
we have to continue using that manual process?
Thanks,
Mike
Basically, it's an integration with some business logic to make a final
pass at ensuring that certain business rules are followed in the event a
query returns documents a user is not authorized to see.
Thanks,
Mike
iness logic is such that
> you can calculate them all "fast enough", you're golden.
>
> All that said, if there's any way you can build this into tokens in the
> doc and use a standard fq clause it's usually much easier. That may
> take some creative work
Cloudera defaults their Hadoop installation to use HDFS w/ their bundle of
Solr (4.10.3) if that is any indication.
On Tue, Jan 3, 2017 at 7:40 AM, Hendrik Haddorp
wrote:
> Hi,
>
> is the HDFS support in Solr 6.3 considered production ready?
> Any idea how many setups might be using this?
>
> th
D DEVOPS)
If you are using a role-based system that bundles groups of permissions
into a role, all you need to do is decompose the role into a list of
permissions for the user and put all of the required permissions into that
multivalue field.
Mike
On Wed, Jan 4, 2017 at 2:55 AM, wrote:
> I a
I finally got a chance to deep dive into this and have a preliminary
working plugin. I'm starting to look at optimization strategies for how to
speed processing up and am wondering if you can give me some more
information about your "bailout" strategy.
Thanks,
Mike
On Wed, Dec 2
Probably not, but writing your own little Java process to do it would be
trivial with Kafka 0.9.X or 0.10.X. You can also look at the Confluent
Platform as they have tons of connectors for Kafka to directly feed into
other systems.
On Mon, Jan 30, 2017 at 3:05 AM, Mahmoud Almokadem
wrote:
> Hell
Set the fl parameter equal to the fields you want and then query for
id:(SOME_ID OR SOME_ID OR SOME_ID)
On Thu, Feb 9, 2017 at 5:37 AM, Midas A wrote:
> Hi,
>
> i want solr doc partially if unique id exist else we donot want to do any
> thing .
>
> how can i achieve this .
>
> Regards,
> Midas
>
Correct me if I'm wrong, but heavy use of doc values should actually blow
up the size of your index considerably if they are in fields that get sent
a lot of data.
On Tue, Feb 21, 2017 at 10:50 AM, Pratik Patel wrote:
> Thanks for the reply. I can see that in solr 6, more than 50% of the index
>
When you transition to an external zookeeper, you'll need at least 3 ZK
nodes. One is insufficient outside of a development environment. That's a
general requirement for any system that uses ZK.
On Sun, Feb 26, 2017 at 7:14 PM, Satya Marivada
wrote:
> May I ask about the port scanner running? Ca
Just barely skimmed the documentation, but it looks like the tool generates
its own shards and pushes them into the collection by manipulating the
configuration of the cluster.
https://www.cloudera.com/documentation/enterprise/5-8-x/topics/search_mapreduceindexertool.html
If that reading is corre
It's a brittle ZK configuration. A typical ZK quorum is three nodes for
most production systems. One is fine, though, for development provided the
system it's on is not overloaded.
On Mon, Feb 27, 2017 at 6:43 PM, Rick Leir wrote:
> Hi Mike
> We are using a single ZK nod
Found this project and I'd like to know what would be involved with
exposing its RestrictedField type through Solr for indexing and querying as
a Solr field type.
https://github.com/roshanp/lucure-core
Thanks,
Mike
I've only ever used the HDFS support with Cloudera's build, but my
experience turned me off to use HDFS. I'd much rather use the native file
system over HDFS.
On Tue, Mar 14, 2017 at 10:19 AM, Muhammad Imad Qureshi <
imadgr...@yahoo.com.invalid> wrote:
> We have a 30 node Hadoop cluster and each
If Solr is down, then adding through SolrJ would fail as well. Kafka's new
API has some great features for this sort of thing. The new client API is
designed to be run in a long-running loop where you poll for new messages
with a certain amount of defined timeout (ex: consumer.poll(1000) for 1s)
So
1 - 100 of 1080 matches
Mail list logo