Some extra for the pros list:

- Full control over which content to be searchable and not.
- Posibility to make pages searchable almost instant after publication
- Control over when the site is indexed


Friendly

Jan-Eirik

On Fri, Nov 13, 2009 at 10:52 AM, Lukáš Vlček <lukas.vl...@gmail.com> wrote:

> Hi,
>
> I am looking for good arguments to justify implementation a search for
> sites
> which are available on the public internet. There are many sites in
> "powered
> by Solr" section which are indexed by Google and other search engines but
> still they decided to invest resources into building and maintenance of
> their own search functionality and not to go with [user_query site:
> my_site.com] google search. Why?
>
> By no mean I am saying it makes not sense to implement Solr! But I want to
> put together list of reasons and possibly with examples. Your help would be
> much appreciated!
>
> Let's narrow the scope of this discussion to the following:
> - the search should cover several community sites running open source CMSs,
> JIRAs, Bugillas ... and the like
> - all documents use open formats (no need to parse Word or Excel)
> (maybe something close to what LucidImagination does for mailing lists of
> Lucene and Solr)
>
> My initial kick off list would be:
>
> pros:
> - considering we understand the content (we understand the domain scope) we
> can fine tune the search engine to provide more accurate results
> - Solr can give us facets
> - we have user search logs (valuable for analysis)
> - implementing Solr is a fun
>
> cons:
> - requires resources (but the cost is relatively low depending on the query
> traffic, index size and frequency of updates)
>
> Regards,
> Lukas
>
> http://blog.lukas-vlcek.com/
>



-- 
Jan Eirik B. Nævdal
Solutions Engineer | +47 982 65 347
Iterate AS | www.iterate.no
The Lean Software Development Consultancy

Reply via email to