Question regarding empty UUID field

2016-04-12 Thread Susmit Shukla
Hi, I have configured solr schema to generate unique id for a collection using UUIDUpdateProcessorFactory I am seeing a peculiar behavior - if the unique 'id' field is explicitly set as empty string in the SolrInputDocument, the document gets indexed. I can see in the solr query console a good uu

UUID processor handling of empty string

2016-04-14 Thread Susmit Shukla
Hi, I have configured solr schema to generate unique id for a collection using UUIDUpdateProcessorFactory I am seeing a peculiar behavior - if the unique 'id' field is explicitly set as empty string in the SolrInputDocument, the document gets indexed with UUID update processor generating the id.

Re: UUID processor handling of empty string

2016-04-14 Thread Susmit Shukla
are equal, so the tiebreaker is > : the internal Lucene doc ID, which may change as merges > : happen. You can specify secondary sort fields to make the > : sort predictable (the field is popular for this). > : > : Best, > : Erick > : > : On Thu, Apr 14, 2016 at 12:18 PM, Sus

Re: UUID processor handling of empty string

2016-04-16 Thread Susmit Shukla
value not present are not the same > thing. > > So, please clarify your specific situation. > > > -- Jack Krupansky > > On Thu, Apr 14, 2016 at 7:20 PM, Susmit Shukla > wrote: > > > Hi Chris/Erick, > > > > Does not work in the sense the order of documen

Re: UUID processor handling of empty string

2016-04-17 Thread Susmit Shukla
nt out what's not > as you expect. > > You might want to review: > http://wiki.apache.org/solr/UsingMailingLists > > Best, > Erick > > On Sat, Apr 16, 2016 at 9:54 AM, Jack Krupansky > wrote: > > Remove that line of code from your client, or... add the remove blank

Re: Return only parent on child query match (w/o block-join)

2016-04-19 Thread Susmit Shukla
Hi Shamik, you could try solr grouping using group.query construct. you could discard the child match from the result i.e. any doc that has parent_doc_id field and use join to fetch the parent record q=*:*&group=true&group.query=title:title2&group.query={!join from=parent_doc_id to=doc_id}parent_

Re: Cross collection join in Solr 5.x

2016-04-21 Thread Susmit Shukla
I have done it by extending the solr join plugin. Needed to override 2 methods from join plugin and it works out. Thanks, Susmit On Thu, Apr 21, 2016 at 12:01 PM, Mikhail Khludnev < mkhlud...@griddynamics.com> wrote: > Hello, > > There is no much progress on > https://issues.apache.org/jira/brow

Re: The Streaming API (Solrj.io) : id must have DocValues?

2016-04-26 Thread Susmit Shukla
Which solrj version are you using? could you try with solrj 6.0 On Tue, Apr 26, 2016 at 10:36 AM, sudsport s wrote: > @Joel > >Can you describe how you're planning on using Streaming? > > I am mostly using it for distirbuted join case. We were planning to use > similar logic (hash id and join) i

Re: Query String Limit

2016-05-04 Thread Susmit Shukla
Hi Prasanna, What is the exact number you set it to? What error did you get on solr console and in the solr logs? Did you reload the core/restarted solr after bumping up the solrconfig? Thanks, Susmit On Wed, May 4, 2016 at 9:45 PM, Prasanna S. Dhakephalkar < prasann...@merajob.in> wrote: > Hi

Re: fq behavior...

2016-05-05 Thread Susmit Shukla
Please take a look at this blog, specifically "Leapfrog Anyone?" section- http://yonik.com/advanced-filter-caching-in-solr/ Thanks, Susmit On Thu, May 5, 2016 at 10:54 PM, Bastien Latard - MDPI AG < lat...@mdpi.com.invalid> wrote: > Hi guys, > > Just a quick question, that I did not find an easy

Re: Field Definitions Ignored

2016-06-07 Thread Susmit Shukla
Does solr streaming aggregation support pagination? Some documents seem to be skipped if I set "start" parameter for CloudSolrStream for a sharded collection. Thanks, Susmit

start parameter for CloudSolrStream

2016-06-07 Thread Susmit Shukla
*sending with correct subject* Does solr streaming aggregation support pagination? Some documents seem to be skipped if I set "start" parameter for CloudSolrStream for a sharded collection. Thanks, Susmit

Re: start parameter for CloudSolrStream

2016-06-08 Thread Susmit Shukla
port for the > OFFSET SQL clause. > > Joel Bernstein > http://joelsolr.blogspot.com/ > > On Tue, Jun 7, 2016 at 5:08 PM, Susmit Shukla > wrote: > > > *sending with correct subject* > > > > Does solr streaming aggregation support pagination? > > Some doc

export with collapse filter runs into NPE

2016-06-10 Thread Susmit Shukla
Hi, I'm running this export query, it is working fine. f1 is the uniqueKey and running solr 5.3.1 /export?q=f1:term1&sort=f1+desc&fl=f1,f2 if I add collapsing filter, it is giving NullPointerException /export?q=f1:term1&sort=f1+desc&fl=f1,f2&fq={!collapse field=f2} does collapsing filter work

Re: export with collapse filter runs into NPE

2016-06-10 Thread Susmit Shukla
, 2016 at 1:09 PM, Joel Bernstein wrote: > This sounds like a bug. I'm pretty sure there are no tests that use > collapse with the export handler. > > Joel Bernstein > http://joelsolr.blogspot.com/ > > On Fri, Jun 10, 2016 at 3:59 PM, Susmit Shukla > wrote: > > &g

Json Parse Exception in CloudSolrStream class

2016-07-03 Thread Susmit Shukla
Hi, I'm using a string field in sort parameters of a solr query. The query is used with /export handler to stream data using CloudSolrStream. When the data in field contains a double quote, the cloudSolrStream fails to read data and throws this error - field data = "first (alias) last" org.nogg

Re: solr /export handler - behavior during close()

2017-06-25 Thread Susmit Shukla
, Susmit Shukla wrote: > Thanks Joel, will try that. > Binary response would be more performant. > I observed the server sends responses in 32 kb chunks and the client reads > it with 8 kb buffer on inputstream. I don't know if changing that can > impact anything on performance.

Re: solr /export handler - behavior during close()

2017-06-27 Thread Susmit Shukla
u also mentioned that the SolrStream and the SolrClientCache were using > the same approach to create the client. In that case changing the > ParallelStream to set the streamContext shouldn't have any effect on the > close() issue. > > > > > > > > > Joel B

Re: deep paging in parallel sql

2017-09-07 Thread Susmit Shukla
you could use filter clause to create a custom cursor since the results are sorted. I had used the approach with raw cloudsolr stream, not with parallelSQL though. This would be useful- https://lucidworks.com/2013/12/12/coming-soon-to-solr-efficient-cursor-based-iteration-of-large-result-sets/ Th

Re: Streaming and large resultsets

2017-11-11 Thread Susmit Shukla
Hi Lanny, For long running streaming queries with many shards and huge resultsets, solrj's default settings for http max connections/connections per host may not be enough. If you are using the worker collection (/stream), it depends on dispensing http clients using SolrClientCache with default li

solr /export handler - behavior during close()

2017-05-12 Thread Susmit Shukla
Hi, I have a question regarding solr /export handler. Here is the scenario - I want to use the /export handler - I only need sorted data and this is the fastest way to get it. I am doing multiple level joins using streams using /export handler. I know the number of top level records to be retrieve

Re: solr /export handler - behavior during close()

2017-05-12 Thread Susmit Shukla
results until it > encounters a "Broken Pipe" exception. This exception is trapped and ignored > rather then logged as it's not considered an exception if the client > disconnects early. > > Joel Bernstein > http://joelsolr.blogspot.com/ > > On Fri, May 12, 20

Re: solr /export handler - behavior during close()

2017-05-13 Thread Susmit Shukla
If the client closes the connection to the export handler then this > exception will occur automatically on the server. > > Joel Bernstein > http://joelsolr.blogspot.com/ > > On Sat, May 13, 2017 at 1:46 AM, Susmit Shukla > wrote: > > > Hi Joel, > > > > Than

Re: solr /export handler - behavior during close()

2017-05-13 Thread Susmit Shukla
//joelsolr.blogspot.com/ > > On Sat, May 13, 2017 at 12:28 PM, Susmit Shukla > wrote: > > > Hi Joel, > > > > I did not observe that. On calling close() on stream, it cycled through > all > > the hits that /export handler calculated. > > e.g. with a *:*

Re: solr /export handler - behavior during close()

2017-05-13 Thread Susmit Shukla
n a ticket for this? > > Joel Bernstein > http://joelsolr.blogspot.com/ > > On Sat, May 13, 2017 at 2:51 PM, Susmit Shukla > wrote: > > > Hi Joel, > > > > I was using CloudSolrStream for the above test. Below is the call stack. > > > >

Re: solr /export handler - behavior during close()

2017-05-15 Thread Susmit Shukla
SolrStream that > expression will be sent to each shard to be run and each shard will be > duplicating the work and return duplicate results. > > > > > > > > > > > > > > > > > > > > > Joel Bernstein > http://joelsolr.blogspot.

Re: solr /export handler - behavior during close()

2017-05-16 Thread Susmit Shukla
ifferent partitions of the streams will be served by > different replicas. > > If performance doesn't improve with the NullStream after increasing both > workers and replicas then we know the bottleneck is the network. > > Joel Bernstein > http://joelsolr.blogspot.c

Re: solr /export handler - behavior during close()

2017-05-17 Thread Susmit Shukla
> >> I created https://issues.apache.org/jira/browse/SOLR-10698 to track the > >> issue > >> > >> @Susmit looking at the stack trace I see the expression is using > >> JSONTupleStream > >> . I wonder if you tried using JavabinTupleStreamParser co

Re: Performance Issue in Streaming Expressions

2017-06-01 Thread Susmit Shukla
Hi, Which version of solr are you on? Increasing memory may not be useful as streaming API does not keep stuff in memory (except may be hash joins). Increasing replicas (not sharding) and pushing the join computation on worker solr cluster with #workers > 1 would definitely make things faster. Are

fix wiki error

2014-07-08 Thread Susmit Shukla
The url for solr atomic update documentation should contain json in the end. Here is the page - https://wiki.apache.org/solr/UpdateJSON#Solr_4.0_Example curl http://localhost:8983/solr/update/*json* -H 'Content-type:application/json'

Solr cloud setup question

2014-09-22 Thread Susmit Shukla
Hi solr experts, I am building out a solr cluster with this configuration 3 external zookeeprs 15 solr instances (nodes) 3 shards I need to start out with 3 nodes and remaining 12 nodes would be added to cluster. I am able to create a collection with 3 shards. This process works fine using colle

Solr Cloud Default Document Routing

2014-09-24 Thread Susmit Shukla
Hi, I'm building out a multi shard solr collection as the index size is likely to grow fast. I was testing out the setup with 2 shards on 2 nodes with test data. Indexed few documents with "id" as the unique key. collection create command - /solr/admin/collections?action=CREATE&name=multishard&num

Gather Nodes Streaming

2019-03-20 Thread Susmit Shukla
Hi, Trying to use solr streaming 'gatherNodes' function. It is for extracting email graph based on from and to fields. It requires 'to' field to be a single value field with docvalues enabled since it is used internally for sorting and unique streams The 'to' field can contain multiple email addr

solr 8.6.3 and noggit

2020-11-20 Thread Susmit Shukla
Hi, got this error using streaming with solrj 8.6.3 . does it use noggit-0.8. It was not mentioned in dependencies https://github.com/apache/lucene-solr/blob/branch_8_6/solr/solrj/ivy.xml Caused by: java.lang.NoSuchMethodError: 'java.lang.Object org.noggit.ObjectBuilder.getValStrict()' at org.apa

Re: solr 8.6.3 and noggit

2020-11-20 Thread Susmit Shukla
rJ against a newer Solr server (or > vice versa). > > Mike > > On Fri, Nov 20, 2020 at 2:25 PM Susmit Shukla > wrote: > > > Hi, > > got this error using streaming with solrj 8.6.3 . does it use noggit-0.8. > > It was not mentioned in dependencies >