Hi,
QUESTION 1
First things first, for the dataimport handler. Is it correct that when I
visit it from the admin panel it takes me to this URL:
*http://x.com:8080/solr/#/collection1/dataimport//dataimport
*
When I visit it on this page, it seems to load my config correctly in the
right panel. A
Thank you both so much for your help. The regex was indeed outdated.
Everything works perfectly now! :)
--
View this message in context:
http://lucene.472066.n3.nabble.com/Enabling-DIH-breaks-Solr4-4-tp4083282p4083286.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
I'm a bit stuck here. I had Solr4.4 working without too many issues. I
wanted to enable the DIH so I firstly added these lines to the
solrconfig.xml:
Restarted and everything was fine. Then I added these lines to the config as
well which broke Solr:
/opt/solr/collection1/conf/
Apparently this is the error:
2013-08-08 09:35:19.994:WARN:oejw.WebAppContext:main: Failed startup of
context
o.e.j.w.WebAppContext@64a20878{/solr,file:/tmp/jetty-0.0.0.0-8080-solr.war-_solr-any-/webapp/,STARTING}{/solr.war}
org.apache.solr.common.SolrException: Could not find necessary SLF4j log
I've been unable to install SOLR into Jetty. Jetty seems to be running fine,
and this is the steps I took to install solr:
# SOLR
cd /opt
wget -O - $SOLR_URL | tar -xzf -
cp solr-4.4.0/dist/solr-4.4.0.war /opt/jetty/webapps/solr.war
cp -R solr-4.4.0/example/solr /opt/
cp -R /opt/solr-4.4.0/dist/ /
For the data import handler I have moved he mysql and postgresql jar files to
the solr lib directory (/opt/solr/lib).
My issue is in the data-config.xml I have put two datasources, however, I am
stuck on what to put for the driver values and the urls.
" url=""
user="" password="" />
"
url="" user
Thank you very much
--
View this message in context:
http://lucene.472066.n3.nabble.com/Adding-Postgres-and-Mysql-JDBC-drivers-to-Solr-tp4082806p4082832.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
I am running Solr4 on Jetty9 and I am trying to include the JDBC drivers for
both MySQL and PostgreSQL. I'm a little confused about how I do this.
I beleive these to be the two files I need:
http://cdn.mysql.com/Downloads/Connector-J/mysql-connector-java-5.1.26.tar.gz
http://jdbc.postgresql.o
Hi,
By the looks of it I have a few options with regards to boosting. I was
wondering from a performance point of view am I better to set the boost of
certain results on import via the DIH or instead is it better to set the
boost when doing queries, by adding it to the default queries?
I have a "
currently have an item which gets imported into solr, lets call it a book
entry. Well that has a single location associated with it as a coordinate
and location name but I am now finding out that a single entry may actually
need to be associated with more than one location, for example "New York"
Hi,
This is the situation, I have two sources of data in my dataimport handler,
one is huge, the other is tiny:
Source A: 10-20 records
Source B: 50,000,000 records
I was wondering what happens if I was to do a DIH just on Source A every 10
mins, and only run the DIH on source B every 24 hours.
Hi,
I was wondering, what setup have people had the most luck with from a
performance point of view?
Tomcat Vs Jetty
Open JDK vs Oracle JDK
I haven't been able to find any information online to backup any sort of
performance claims. I am planning on using Tomcat with Open JDK, has anyone
had any
Hi.
I have recently made the move from SOLR3.6 to SOLR4.0 and so far everything
seems super apart frmo the fact that I had a lot of warnings in my "logging"
section on the solr admin panel. I have tried to work through as many as
possible but I have a few that I am not able to correct. This is the
This is the datetime format SOLR requires as I understand it:
1995-12-31T23:59:59Z
When I try to store this as a datetime field in MySQL it says it isn't
valid. My question is, ideally I would want to keep a datetime in my
database so I can sort by date rather than just making it a varchar, so I
Hi,
I currently have this setup:
Bring in data into the "description" schema and then have this code:
To then truncate the description and move it to "truncated_description".
This works fine.
I was wondering, is it possible so that when I bring in data from another
source I actually bring it
Hi,
I am looking to import entries to my SOLR server by using the DIH,
connecting to an external postgre SQL server using the JDBC driver. I will
be importing about 50,000 entries each time.
Is connecting to an external SQL server for my data unreliable or risky, or
is it instead perfrectly reas
I was wondering, with my current setup I have an ID field which is a number.
If I wanted to then change it so my ID field was actually a mix of numbers
and letters (to do with the backend system), would this cause any sort of
problem?
I would never do any kind of sorting by ID on my search page,
Thank you so much for the help, I really appreciate it.
--
View this message in context:
http://lucene.472066.n3.nabble.com/SQL-DIH-Can-I-have-some-guidance-please-tp4024207p4024250.html
Sent from the Solr - User mailing list archive at Nabble.com.
Thank you so much for your help. Based on the same schema in my first post
and your help I created this, have I implemented it correctly based on your
suggestion? I tried to comment it:
I have two databases (unfortunately they do have to be separate) which get
imported into Solr.
Each database has a primary key for each time but I am concerned that when
it comes to importing the two into SOLR there will be more than one item
with the same ID (one from each DB).
Therefore, in ord
Hi.
I am having a bit of trouble figuruing out the DIH for SQL files. I have
asked around a few different places but havent got any replies so I was
hoping you could help me.
*I have a database schema like this:*
CREATE TABLE company (
id SERIAL PRIMARY KEY,
name varchar(60) NOT NULL
);
Update! Thank you to Lance for the help. Based on your suggestion I have
fixed up a few things.
*My Dataconfig now has the filename pattern fixed and root entity=true*
/
/
*My data.xml has a corrected date format with "T":*
/
123
Delta Import 2
Hi,
These are the exact steps that I have taken to try and get delta import
handler working. If I can provide any more information to help let me know.
I have literally spent the entire friday night and today on this and I throw
in the towel. Where have I gone wrong?
*Added this line to the solrc
Hey guys,
I am after a bash script (or python script) which I can use to trigger a
delta import of XML files via CRON. After a bit of digging and modification
I have this:
Can I get any feedback on this? Is there a better way of doing it? Any
optimisations or improvements would be most welcome.
Hi,
I was wondering if someone could show me an example XML file for use to
import to solr. Bascially I have the following information that I am trying
to import to solr:
Title
Description
Keyword Description
Source
Location Name
Location Co-ordinates
URL
Time
I've never worked with XML before s
Thank you for the reply. The facet range gap loks good but it is too far down
the line to be of use, I wish it was implemented though.
What I want is really a more simple question
http://wiki.apache.org/solr/SimpleFacetParameters#facet.range
Is it correct that to add facets on date and distance
I am using solr and I want to setup two different types of facets on my data.
The first is date, so I would like it like this:
Posted: Today 3 days 7 days All time
Distance: 5 miles 10 miles 30 miles 100 miles
How should I be setting up faceting? It looks like something I need to edit
my solr.xm
I am using Google for location input.
*It often splits out something like this:*
Shorewood, Seattle, Wa
*Since I am using this index analyzer:*
It means that if I search for "Sho" or "Shorew" I get the result I want.
However, if I search for “Sea” or “Seatt” I get no results.
I guess I need
Hi,
Sorry for all the questions today but I paid a third party coder to develop
a schema for me but now that I have more of an understanding myself I have a
questions.
The aim is to do spacial searching so in my schema I have this:
My site doesnt seem to submit via JSON to lat_lng_0_coordina
Hi.
My SQL database assigns a uniqueID to each item. I want to keep this
uniqueID assosiated to the items that are in Solr even though I wont ever
need to display them or have them searchable. I do however what to be able
to target specific items in Solr with it, for updating or deleting the
recor
Hi,
I have been making changes to my schema and unfortauntly I now get a error
when viewing my Solr tomcat admin page.
The error doesnt seem to explain the problem:
HTTP Status 500 - Severe errors in solr configuration. Check your log files
for more detailed information on what may be wrong. If y
Hi,
So I want to enable highlighting on my results. When I run the query like
this:
http://localhost:8080/solr/select?q=book&hl=true
I don't get any highlighted results. I am assuming that more is needed to
actually enable highlighting. Commented out at the bottom of my
solrconfig.xml is this:
Hi,
I was wondering if it is possible to set up highlighting so it is on by
default, and doesnt need to add to the URL. For example:
http://localhost:8080/solr/select?q=book&hl=true
I would like to have it so highlighting is on even if the URL is this:
http://localhost:8080/solr/select?q=boo
Ok, thank you for the reply. I have one more question then I think everything
is cleared up. If I have this code:
The truncated_description is one I need to display in search results. If I
set this to stored=true as above (so it can be displayed in results), does
it mean that I am storing t
Ok, I’ve been doing a bit more research. In order to do the copyfield
technique, I need to store the original full text document within Solr, like
this:
true*">
What about instead if I imported the same fulltext into two seperate fields
for Solr by my Python script:
trucated_description=post.d
d]", but the
> reality is that Solr is not copying from the indexed value of the field,
> but
> from the source value for the field. The idea is that multiple fields can
> be
> based on the same source value even if they analyze and index the value in
> different ways.
>
I'm really confused here. I have a document which is say 4000 words long. I
want to get this put into two fields in Solr without having to save the
original document in its entirety within Solr.
When I import my fulltext (4000 word) document to Solr I was going to put it
straight into keyword_docu
Maybe I dont understand, but if you are copying the keyword description field
and then truncating it then the truncated form will only have keywords too.
That isnt what I want. I want the truncated form to have words like "a"
"the" "it" etc that would have been removed when added to
keyword_descrip
Ok. I can still define GramSize too?
**
--
View this message in context:
http://lucene.472066.n3.nabble.com/Only-exact-match-searches-working-tp4008160p4008361.html
Sent from the Solr - User mailing list archive at Nabble.com.
Thank you for the reply.
The trouble is, I want the truncated desciption to still have the keywords.
If I pass it to the keyword_descipriton and remove words like "and" "i"
"then" "if" etc, then copy it across to truncated_description, my truncated
description will not be a sentance, it will onl
In an attempt to answer my own question, is this a good solution.
Before I was thinking of importing my fulltext description once, then
sorting it into two seperate fields in solr, one truncated, one keyword.
How about instead actually importing my fulltext description twice. Then I
can import it
Purely for searching.
The truncated form is just to show to the user as a preview, and the keyword
form is for the keyword searching.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Taking-a-full-text-then-truncate-and-duplicate-with-stopwords-tp4008269p4008295.html
Sent
I've hit a bit of a wall and would appreciate some guidance. I want to index
a large block of text, like such:
I dont want to store this as it is in Solr, I want to instead have two
versions of it. One as a truncated form, and one as a keyword form.
*Truncated Form:*
*Keyword Form (using stop
Thank you for the reply. I have done a bit of reading and it says I can also
use this one:
This is what I will use I think, as it weeds out words like "at" "I" as a
bonus.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Only-exact-match-searches-working-tp4008160p400826
Hi,
I finally got my Solr working. It indexes posts and I can search it from my
site. I have run into one problem though.
If I have an entry as "England", when I search "England" on my site, or
"england" I get the result returned. When I search "Eng" it tells me there
are no results.
My question
I am sorry to come across as a bit simple, but I would rather fix this by
making sure my default field in my schema.xml is defined, but I cant
determine where my default field is within my schema.xml
I've put it up on here:
https://gist.github.com/3689621
Would anyone be able to give me a bit mor
Hi,
Im trying to do some work on my server but running into problems. When I try
to ping the server through the admin panel I get this error, which I believe
might be causing the problem:
Can anyone give me a bit of guideance as to what might be going wrong? I'm
using Solr 3.6. I think it may b
Hi,
I want to create a very simple geocoder for returning co-ordinates of a
place if a user enters in a town or city. There seems to be very little
information about doing it the way I suggest, so I hope I am on a good path.
My first decision was to divide SOLR into two cores, since I am already
Normalising the data is a good idea, and it would be easy to do since I would
only have around 50,000 entires BUT it is a bit complicated with addresses I
think. Lets say I store the data in this form:
London, England
Swindon, Wiltshire, England
Wiltshire England
England
What happens if someone
I already have a SOLR server up and running which is designed to receive
keywords and co-ordinates and result item results.
The way I see it I have three options:
1. Have one "geocode" solr server to convert place names to co-ordinates and
another seperate Solr server to turn the items results.
2
I am trying to make a system that returns co-ordinates for a given "town",
"city", "country" or "postcode" query.
I want to return the data in the fastest way possible since the server is
going to be hammered and is going to need real concurency.
Is Solr a good software solution to this? Would I
If I wanted to return long/lat from either "town" "city" or "postcode", how
would I structure it?
If I was just searching for towns it would be easy, like this:
> *Town, City, Postcode, Country, Long/Lat*
> Orpington, Kent, Uk, 1.2/1.1
>
However, but there will be "city" entries and "postcode"
Hi!
I am using Solr as my main search system for my site. Currently, I am using
google to turn a place name (such as a postcode or city) into a long / lat
co-ordinate. Then I am supplying this long / lat to Solr so it can perform a
spacial search.
I am really new to this, but I dont like my relia
This seems to suggest you have to reindex Solr in its entirety and cant add a
single document at a time, is this right?
http://stackoverflow.com/questions/11247625/apache-solr-adding-editing-deleting-records-frequently
--
View this message in context:
http://lucene.472066.n3.nabble.com/Importing
Thank you for the reply. Ok, well that brings another question. I dont like
pre-optimisation, but I also dont like inefficiency, so lets see if I can
strike a balance.
It does seem really poor design to reimport 10,000 documents, when only one
needs to be added. I dont like that, can you not inser
Hi,
Lets say I am running an auction site. There are 20,000 entries. 100 entries
come from an on-site SQL database, the rest come from a generated txt file
from scrapped content.
I want to import any new SQL results onto the server as quickly as possible
so they are searchable but I dont want to
Thank you, the query seems to have got through, thats good i guess?
*Jul 4, 2012 6:32:34 PM org.apache.solr.core.SolrCore execute
INFO: [] webapp=/solr path=/select
params={facet=true&facet.query={!key%3Danytime}date:[*+TO+*]&facet.query={!key%3D1day}date:[NOW/DAY-1DAY+TO+NOW/DAY]&facet.query={!ke
Hi,
Sorry for this post, but im having a hard time getting my head around this.
I installed Solr on Tomcat and it seems to work fine. I get the solr admin
page and the "it works" page from tomcat.
When I try to query my solr server I get this message:
*Internal Server Error
The server encounte
Thank you, that helps. The bit I am still confused about how the server sends
the response to the server though. I get the impression that there are
different ways that this could be done, but is sending an XML response back
to the Python server the best way to do this?
--
View this message in c
Hi Ben,
Thank you for the reply. So, If I don't want to use Javascript and I want
the entire page to reload each time, is it being done like this?
1. User submits form via GET
2. Solr server queried via GET
3. Solr server completes query
4. Solr server returns XML output
5. XML data put into resu
Thank you for the reply, but I'm afraid I don't understand :(
This is how things are setup. On my Python website, I have a keyword and
location box. When clicked, it queries the server via a javascript "GET"
request, it then sends back the data via Json.
I'm saying that I dont want to be reliant
Final comment from me then Ill let someone else speak.
The solution we seem to be looking at is send a GET request to SOLR and then
send back a renderized page, so we are basically creating the results page
on the server rather than the client side.
I would really like to hear what people have to
Further to my last reply. How about I do the following:
Send the request to the server using the GET method and then return the
results in XML rather than JSON. Does this sound logical?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Help-Confused-about-using-Jquery-for-the-
Hi,
My current method of searching involes communicating to solr using python.
The clients browser communicates to the search API using jquery/json.
However, although this works, I dont like the dependancy on Javascript.
Either I can keep with this method and have a backup system in place that
wo
Hi,
I installed a fresh copy of Solr 3.6.0 or my server but I get the following
page when I try to access Solr:
http://176.58.103.78:8080/solr/
It says errors to do with my Solr.xml. This is my solr.xml:
I really cant figure out how I am meant to fix this, so if anyone is able to
give some in
Hi,
This is the install process I used in my shell script to try and get Tomcat
running with Solr (debian server):
I swear this used to work, but currently only Tomcat works. The Solr page
just comes up with "The requested resource (/solr/admin) is not available."
Can anyone give me some insig
Hi,
I have the following situation. I have two feeds into my SOLR:
1. XML file containing around 30,000 entries, likely to be changing
significantly each day, only changing once every 24 hours
2. A SQL database containing around 30 entries, only 1-2 new entries each
day, but important to be inde
Right, for Long/Lat I found this information:
<-Long / Lat Field Type->
<-Fields->
Does this look more logical?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Newbie-tries-to-make-a-Schema-xml-tp3974200p3976539.html
Sent from the Solr - User mailing l
Thank you for the feedback. Yes they are uses for geospacial. After doing a
bit of homework I found this correction. Is this how it should be done?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Newbie-tries-to-make-a-Schema-xml-tp3974200p3975458.html
Sent from the S
Hi,
I’m totally out of my depth here but I am trying, so I apologise if this is
a bit of a basic question. I need the following information to be indexed
and then made searchable by Solr:
Title – A title for the company
Company – The name of the company
Description – A description of the company
I discounted geonames to start with but it actually looks pretty good. I may
be stretching the limit of my question here, but say I did go with geonames,
if I go back to my model and add a bit:
Search for "London"->Convert "London to Long/Lat"->Send Query to
Solr>->Return Query>
Since
Hi,
This is quite a challenge. I know there are situations when you can get by
with google maps api or similar, but they limit the number of requests and I
need more than that, unfortunatly for the full service they charge a
fortune!
So, going back to my question, does anyone have any ideas or su
Hi,
I'm finding it a bit hard to get my head around this. Say I am putting items
on a map. This is how I am thinking it would work:
A user submits an "item" and specifies the location as "London"
On submission, I run a process to convert "London" to a Long/Lat which is
stored in the database
How
Hi,
I am designing a custom scrapping solution. I need to store my data, do some
post processing on it and then import it into SOLR.
If I want to import data into SOLR in the quickest, easiest way possible,
what format should I be saving my scrapped data in? I get the impression
that .XML would
Hi,
If I want to do a proximity search and they have provided me with a name of
a city, for example, “London”. How do I search this by proximity within
Solr?
I am assuming I first need a process to convert the city name to a long and
lat, so that Solr can understand where London is. Is this somet
Thank you for the reply. I hate to take more of peoples time but can anyone
elaborate more on the kind of firewall rules I should be looking at?
--
View this message in context:
http://lucene.472066.n3.nabble.com/Securing-Solr-under-Tomcat-IP-best-way-tp3899929p3900040.html
Sent from the Solr -
Hi,
I’m in the process of working how to configure and secure my server running
Nginx, and Nutch and Solr under Tomcat. Is the best security practice for
securing Solr under Tomcat simply to only allow requests only from
127.0.0.1. This way Solr isn’t exposed to the outside world and is only
compr
Hi.
Python is becoming an extremely popular language for web programming but the
official python-solr interface isn't included since 1.3 and a lot of the
ones on the web seem good but not production worthy yet, like mysolr.
Can solr be used with python in a production environment or will I be for
Hi,
I am writing a bash script which will install SOLR. I had two quick
questions:
Is there a way to update SOLR, changing only the files which have changed
from the previous version, or is it a case that I need to fully uninstall
the old version and then reinstall the newest version.
Secondly,
Hi,
Call me crazy, but I don’t like the idea of having a single server which not
only runs my PHP site on Apache, but also runs SOLR and Nutch, inclusive of
Tomcat.
Is it a terrible idea to have one Rackspace VPS account which runs the PHP
site with MYSQL database, and another rackspace account w
A quick, bump, I could really do with some input on this please.
--
View this message in context:
http://lucene.472066.n3.nabble.com/Geolocation-in-SOLR-with-PHP-application-tp3807120p3812364.html
Sent from the Solr - User mailing list archive at Nabble.com.
Hi,
I have two newbie questions. With all my searching I havent been able to
find which would be a better choice to run my SOLR / Nutch install, Tomcat
or Jetty. There seems to be a lot of people on the internet saying Jetty has
better performance but I havent been able to see any proof of that.
Thank you for the information Damien.
Is there a better database to use at the core of the sight which is more
compatible with SOLR than MYSQL, or is hooking MYSQL up with SOLR simple
enough.
--
View this message in context:
http://lucene.472066.n3.nabble.com/SOLR-Just-for-search-or-whole-site-
I am new to this but I wanted to pitch a setup to you. I have a website
being coded at the moment, in the very early stages, but is effectively a
full text scrapper and search engine. We have decided on SOLR for the search
system.
We basically have two sets of data:
One is the content for the se
I am creating what is effectively a search engine. Content is collected via
spiders at
then is inserted into my database and becomes searchable and filterable.
I invision there being around 90K records to be searched at any one time.
The content is
blog posts and forum posts so we are basically lo
85 matches
Mail list logo