[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689352137


   The import error was fixed already. It was not caused by this PR, but a 
change in master.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14439) Upgrade to Tika 1.24.1

2020-09-09 Thread Andras Salamon (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192690#comment-17192690
 ] 

Andras Salamon commented on SOLR-14439:
---

[~tallison] Can you please change the status to patch available, to start the 
precommit checks.

> Upgrade to Tika 1.24.1
> --
>
> Key: SOLR-14439
> URL: https://issues.apache.org/jira/browse/SOLR-14439
> Project: Solr
>  Issue Type: Task
>  Components: contrib - DataImportHandler
>Reporter: Tim Allison
>Assignee: Tim Allison
>Priority: Major
> Attachments: SOLR-14339.patch
>
>
> We recently released 1.24.1 with several fixes for DoS vulnerabilities we 
> found via fuzzing: CVE-2020-9489 https://seclists.org/oss-sec/2020/q2/69



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Created] (LUCENE-9513) Use seconds instead of millisecs

2020-09-09 Thread Frank Zhu (Jira)
Frank Zhu created LUCENE-9513:
-

 Summary: Use seconds instead of millisecs
 Key: LUCENE-9513
 URL: https://issues.apache.org/jira/browse/LUCENE-9513
 Project: Lucene - Core
  Issue Type: Improvement
  Components: core/other
Affects Versions: 8.6.2, master (9.0)
Reporter: Frank Zhu


In RangeFacetsExample, there is a trivial mistake where 

private final long nowSec is supposed to be secs. 


But it uses as millisecs.

It doesn't affect the output due to the scale change (It will just make 
PAST_HOUR, PAST_SIX_HOURS not as intended). Still it's better to be correct.

 

I've created a pull request on github with the simple fix.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Updated] (LUCENE-9513) Use seconds instead of millisecs

2020-09-09 Thread Frank Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/LUCENE-9513?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Frank Zhu updated LUCENE-9513:
--
Description: 
In RangeFacetsExample, there is a trivial mistake where

private final long nowSec is supposed to be secs.

But it uses as millisecs.

It doesn't affect the output due to the scale change (It will just make 
PAST_HOUR, PAST_SIX_HOURS not as intended). Still it's better to be correct.

 

I've created a pull request on github with the simple fix.

https://github.com/apache/lucene-solr/pull/1846

  was:
In RangeFacetsExample, there is a trivial mistake where 

private final long nowSec is supposed to be secs. 


But it uses as millisecs.

It doesn't affect the output due to the scale change (It will just make 
PAST_HOUR, PAST_SIX_HOURS not as intended). Still it's better to be correct.

 

I've created a pull request on github with the simple fix.


> Use seconds instead of millisecs
> 
>
> Key: LUCENE-9513
> URL: https://issues.apache.org/jira/browse/LUCENE-9513
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/other
>Affects Versions: master (9.0), 8.6.2
>Reporter: Frank Zhu
>Priority: Trivial
>
> In RangeFacetsExample, there is a trivial mistake where
> private final long nowSec is supposed to be secs.
> But it uses as millisecs.
> It doesn't affect the output due to the scale change (It will just make 
> PAST_HOUR, PAST_SIX_HOURS not as intended). Still it's better to be correct.
>  
> I've created a pull request on github with the simple fix.
> https://github.com/apache/lucene-solr/pull/1846



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Updated] (LUCENE-9513) Use seconds instead of millisecs

2020-09-09 Thread Frank Zhu (Jira)


 [ 
https://issues.apache.org/jira/browse/LUCENE-9513?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Frank Zhu updated LUCENE-9513:
--
Fix Version/s: master (9.0)

> Use seconds instead of millisecs
> 
>
> Key: LUCENE-9513
> URL: https://issues.apache.org/jira/browse/LUCENE-9513
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/other
>Affects Versions: master (9.0), 8.6.2
>Reporter: Frank Zhu
>Priority: Trivial
> Fix For: master (9.0)
>
>
> In RangeFacetsExample, there is a trivial mistake where
> private final long nowSec is supposed to be secs.
> But it uses as millisecs.
> It doesn't affect the output due to the scale change (It will just make 
> PAST_HOUR, PAST_SIX_HOURS not as intended). Still it's better to be correct.
>  
> I've created a pull request on github with the simple fix.
> https://github.com/apache/lucene-solr/pull/1846



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Created] (LUCENE-9514) Include TermVectorsWriter in DWPT accounting

2020-09-09 Thread Simon Willnauer (Jira)
Simon Willnauer created LUCENE-9514:
---

 Summary: Include TermVectorsWriter in DWPT accounting
 Key: LUCENE-9514
 URL: https://issues.apache.org/jira/browse/LUCENE-9514
 Project: Lucene - Core
  Issue Type: Improvement
Reporter: Simon Willnauer


As a followup of LUCENE-9511 we should also include the TermVectorsWriter into 
the DWPT accounting since it might hold on to RAM in it's byte buffer recycling 
etc. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9511) Include StoredFieldsWriter in DWPT accounting

2020-09-09 Thread Simon Willnauer (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9511?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192701#comment-17192701
 ] 

Simon Willnauer commented on LUCENE-9511:
-

{quote}I guess this would mean applications using more threads to index would 
have seen RAM consumed but not tracked properly by {{IndexWriter}} until this 
fix.
{quote}
 

 That is true. I think it's insignificant for most of our users since it only 
really comes into play when you are tons of flushes and a large number of 
blocked DWPT that are otherwise flush pending. That situation requires slow 
flushes and slow commits as well as the situation that the DWPTs that are 
flushing or pending are very small by itself ie. only a couple of small docs. 
Otherwise stall control would kick in earlier and prevent adding more docs. 
Still lucene is used in so many environments and having large amounts of 
threads concurrently might add up here.

> Include StoredFieldsWriter in DWPT accounting
> -
>
> Key: LUCENE-9511
> URL: https://issues.apache.org/jira/browse/LUCENE-9511
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Simon Willnauer
>Priority: Major
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> StoredFieldsWriter might consume some heap space memory that can have a 
> significant impact on decisions made in the IW if writers should be stalled 
> or DWPTs should be flushed if memory settings are small in IWC and flushes 
> are frequent. We should add some accounting to the StoredFieldsWriter since 
> it's part of the DWPT lifecycle and not just present during flush.
> Our nightly builds ran into some OOMs due to the large chunk size used in the 
> CompressedStoredFieldsFormat. The reason are very frequent flushes due to 
> small maxBufferedDocs which causes 300+ DWPTs to be blocked for flush causing 
> ultimately an OOM exception.
> {noformat}
>  
>  NOTE: reproduce with: ant test  -Dtestcase=TestIndexingSequenceNumbers 
> -Dtests.method=testStressConcurrentCommit -Dtests.seed=A04943A98C8E2954 
> -Dtests.nightly=true -Dtests.slow=true -Dtests.badapples=true 
> -Dtests.locale=vo-001 -Dtests.timezone=Africa/Ouagadougou 
> -Dtests.asserts=true -Dtests.file.encoding=UTF8*06:06:15*[junit4] ERROR   
>  107s J3 | TestIndexingSequenceNumbers.testStressConcurrentCommit 
> <<<*06:06:15*[junit4]> Throwable #1: 
> org.apache.lucene.store.AlreadyClosedException: this IndexWriter is 
> closed*06:06:15*[junit4]>at 
> org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:876)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:890)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.IndexWriter.commit(IndexWriter.java:3727)*06:06:15*   
>  [junit4]> at 
> org.apache.lucene.index.TestIndexingSequenceNumbers.testStressConcurrentCommit(TestIndexingSequenceNumbers.java:228)*06:06:15*
> [junit4]>  at 
> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)*06:06:15*[junit4]>at 
> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)*06:06:15*
> [junit4]>  at 
> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)*06:06:15*
> [junit4]>  at 
> java.base/java.lang.reflect.Method.invoke(Method.java:566)*06:06:15*
> [junit4]>at 
> java.base/java.lang.Thread.run(Thread.java:834)*06:06:15*[junit4]> 
> Caused by: java.lang.OutOfMemoryError: Java heap space*06:06:15*[junit4]  
>   > at 
> __randomizedtesting.SeedInfo.seed([A04943A98C8E2954]:0)*06:06:15*[junit4] 
>>   at 
> org.apache.lucene.store.GrowableByteArrayDataOutput.(GrowableByteArrayDataOutput.java:46)*06:06:15*
> [junit4]>   at 
> org.apache.lucene.codecs.compressing.CompressingStoredFieldsWriter.(CompressingStoredFieldsWriter.java:111)*06:06:15*
> [junit4]> at 
> org.apache.lucene.codecs.compressing.CompressingStoredFieldsFormat.fieldsWriter(CompressingStoredFieldsFormat.java:130)*06:06:15*
> [junit4]>   at 
> org.apache.lucene.codecs.lucene87.Lucene87StoredFieldsFormat.fieldsWriter(Lucene87StoredFieldsFormat.java:141)*06:06:15*
> [junit4]>at 
> org.apache.lucene.codecs.asserting.AssertingStoredFieldsFormat.fieldsWriter(AssertingStoredFieldsFormat.java:48)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.StoredFieldsConsumer.initStoredFieldsWriter(StoredFieldsConsumer.java:39)*06:06:15*
> [junit4]> at 
> org.apache.lucene.index.StoredFieldsConsumer.startDocument(StoredFieldsConsumer.java:46)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.DefaultIndexingChain.startStoredFields(DefaultIn

[GitHub] [lucene-solr] s1monw opened a new pull request #1847: LUCENE-9514: Include TermVectorsWriter in DWPT accounting

2020-09-09 Thread GitBox


s1monw opened a new pull request #1847:
URL: https://github.com/apache/lucene-solr/pull/1847


   TermVectorsWriter might consume some heap space memory that
   can have a significant impact on decisions made in the IW if
   writers should be stalled or DWPTs should be flushed if memory
   settings are small in IWC and flushes are frequent. This change adds
   RAM accounting to the TermVectorsWriter since it's part of the
   DWPT lifecycle and not just present during flush.
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Comment Edited] (LUCENE-9511) Include StoredFieldsWriter in DWPT accounting

2020-09-09 Thread Simon Willnauer (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9511?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192701#comment-17192701
 ] 

Simon Willnauer edited comment on LUCENE-9511 at 9/9/20, 7:35 AM:
--

{quote}I guess this would mean applications using more threads to index would 
have seen RAM consumed but not tracked properly by {{IndexWriter}} until this 
fix.
{quote}
 

 That is true. I think it's insignificant for most of our users since it only 
really comes into play when you are tons of flushes and a large number of 
blocked DWPT that are otherwise flush pending. That situation requires slow 
flushes and slow commits as well as the situation that the DWPTs that are 
flushing or pending are very small by itself ie. only a couple of small docs. 
Otherwise stall control would kick in earlier and prevent adding more docs. 
Still lucene is used in so many environments and having large amounts of 
threads concurrently might add up here.

What I am more concerned about is that if we have a bug in the code causing 
large amounts of memory allocated will likely be undetected unless we see some 
slowdowns due to flushing and it will help here to detect this, I hope :)


was (Author: simonw):
{quote}I guess this would mean applications using more threads to index would 
have seen RAM consumed but not tracked properly by {{IndexWriter}} until this 
fix.
{quote}
 

 That is true. I think it's insignificant for most of our users since it only 
really comes into play when you are tons of flushes and a large number of 
blocked DWPT that are otherwise flush pending. That situation requires slow 
flushes and slow commits as well as the situation that the DWPTs that are 
flushing or pending are very small by itself ie. only a couple of small docs. 
Otherwise stall control would kick in earlier and prevent adding more docs. 
Still lucene is used in so many environments and having large amounts of 
threads concurrently might add up here.

> Include StoredFieldsWriter in DWPT accounting
> -
>
> Key: LUCENE-9511
> URL: https://issues.apache.org/jira/browse/LUCENE-9511
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Simon Willnauer
>Priority: Major
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> StoredFieldsWriter might consume some heap space memory that can have a 
> significant impact on decisions made in the IW if writers should be stalled 
> or DWPTs should be flushed if memory settings are small in IWC and flushes 
> are frequent. We should add some accounting to the StoredFieldsWriter since 
> it's part of the DWPT lifecycle and not just present during flush.
> Our nightly builds ran into some OOMs due to the large chunk size used in the 
> CompressedStoredFieldsFormat. The reason are very frequent flushes due to 
> small maxBufferedDocs which causes 300+ DWPTs to be blocked for flush causing 
> ultimately an OOM exception.
> {noformat}
>  
>  NOTE: reproduce with: ant test  -Dtestcase=TestIndexingSequenceNumbers 
> -Dtests.method=testStressConcurrentCommit -Dtests.seed=A04943A98C8E2954 
> -Dtests.nightly=true -Dtests.slow=true -Dtests.badapples=true 
> -Dtests.locale=vo-001 -Dtests.timezone=Africa/Ouagadougou 
> -Dtests.asserts=true -Dtests.file.encoding=UTF8*06:06:15*[junit4] ERROR   
>  107s J3 | TestIndexingSequenceNumbers.testStressConcurrentCommit 
> <<<*06:06:15*[junit4]> Throwable #1: 
> org.apache.lucene.store.AlreadyClosedException: this IndexWriter is 
> closed*06:06:15*[junit4]>at 
> org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:876)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:890)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.IndexWriter.commit(IndexWriter.java:3727)*06:06:15*   
>  [junit4]> at 
> org.apache.lucene.index.TestIndexingSequenceNumbers.testStressConcurrentCommit(TestIndexingSequenceNumbers.java:228)*06:06:15*
> [junit4]>  at 
> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)*06:06:15*[junit4]>at 
> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)*06:06:15*
> [junit4]>  at 
> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)*06:06:15*
> [junit4]>  at 
> java.base/java.lang.reflect.Method.invoke(Method.java:566)*06:06:15*
> [junit4]>at 
> java.base/java.lang.Thread.run(Thread.java:834)*06:06:15*[junit4]> 
> Caused by: java.lang.OutOfMemoryError: Java heap space*06:06:15*[junit4]  
>   > at 
> __randomizedtesting.SeedInfo.seed([A04943A98C8E2954]:0)*06:06:15*[junit4] 
>>   at 
> org.apache.lucene.store

[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689369526


   I merged from master to retrigger the check. Clicking on "Rerun checks" did 
not work, as it merged into same master commit as before. Makes no sense to me 
(maybe because of reproducibility).



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14768) Error 500 on PDF extraction: java.lang.NoClassDefFoundError: org/eclipse/jetty/server/MultiParts

2020-09-09 Thread Markus Kalkbrenner (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192711#comment-17192711
 ] 

Markus Kalkbrenner commented on SOLR-14768:
---

I just want to mention that I created this issue because it has been caught by 
automated tests ;)

[https://github.com/solariumphp/solarium/actions] and 
[https://github.com/mkalkbrenner/search_api_solr/actions]

Beside PDF extraction we test things like the Collections API, Streaming 
Expressions, Facets, JSON Facets, ...

Therefore we run Solr docker containers. Maybe we can add a test that includes 
a build of a docker container using the latest Solr dev version?

> Error 500 on PDF extraction: java.lang.NoClassDefFoundError: 
> org/eclipse/jetty/server/MultiParts
> 
>
> Key: SOLR-14768
> URL: https://issues.apache.org/jira/browse/SOLR-14768
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 8.6, 8.6.1
>Reporter: Markus Kalkbrenner
>Assignee: David Smiley
>Priority: Major
> Attachments: Solr v8.6.x fails with multipart MIME in commands.eml, 
> Solr v8.6.x fails with multipart MIME in commands.eml, Solr v8.6.x fails with 
> multipart MIME in commands.eml
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> See [https://www.mail-archive.com/solr-user@lucene.apache.org/msg152182.html]
> The integration tests of the solarium PHP library and the integration tests 
> of the Search API Solr Drupal module both fail on PDF extraction if executed 
> on Solr 8.6.
> They still work on Solr 8.5.1 an earlier versions.
> {quote}2020-08-20 12:30:35.279 INFO (qtp855700733-19) [ x:5f3e6ce2810ef] 
> o.a.s.u.p.LogUpdateProcessorFactory [5f3e6ce2810ef] webapp=/solr 
> path=/update/extract 
> params=\{json.nl=flat&commitWithin=0&omitHeader=false&resource.name=testpdf.pdf&literal.id=extract-test&commit=true&extractOnly=false&uprefix=attr_&wt=json}{add=[extract-test
>  (1675547519474466816)],commit=} 0 957
> solr8_1 | 2020-08-20 12:30:35.280 WARN (qtp855700733-19) [ ] 
> o.e.j.s.HttpChannel /solr/5f3e6ce2810ef/update/extract => 
> java.lang.NoClassDefFoundError: org/eclipse/jetty/server/MultiParts
> solr8_1 | at 
> org.apache.solr.servlet.SolrRequestParsers.cleanupMultipartFiles(SolrRequestParsers.java:624)
> solr8_1 | java.lang.NoClassDefFoundError: org/eclipse/jetty/server/MultiParts
> solr8_1 | at 
> org.apache.solr.servlet.SolrRequestParsers.cleanupMultipartFiles(SolrRequestParsers.java:624)
>  ~[?:?]
> solr8_1 | at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:443)
>  ~[?:?]
> solr8_1 | at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:345)
>  ~[?:?]
> solr8_1 | at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1596)
>  ~[jetty-servlet-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:545) 
> ~[jetty-servlet-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) 
> ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:590) 
> ~[jetty-security-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:235)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:1610)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:233)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.ContextHandler.doHandle(ContextHandler.java:1300)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedHandler.java:188)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.servlet.ServletHandler.doScope(ServletHandler.java:485) 
> ~[jetty-servlet-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.session.SessionHandler.doScope(SessionHandler.java:1580)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.ScopedHandler.nextScope(ScopedH

[jira] [Comment Edited] (SOLR-14768) Error 500 on PDF extraction: java.lang.NoClassDefFoundError: org/eclipse/jetty/server/MultiParts

2020-09-09 Thread Markus Kalkbrenner (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14768?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192711#comment-17192711
 ] 

Markus Kalkbrenner edited comment on SOLR-14768 at 9/9/20, 7:54 AM:


I just want to mention that I created this issue because it has been caught by 
automated tests ;)

[https://github.com/solariumphp/solarium/actions] and 
[https://github.com/mkalkbrenner/search_api_solr/actions]

Beside PDF extraction we test things like the Collections API, Streaming 
Expressions, Facets, JSON Facets, ...

Therefore we run Solr docker containers. Maybe we can add a test that includes 
a build of a docker container using the latest Solr dev version?

BTW both test suites allways use the latest versions of Solr 7 and 8. At the 
moment we had to break this rule and force Solr 8.5 instead of the latest 
version to get the tests to pass again to not interrupt the development of 
these frameworks.


was (Author: mkalkbrenner):
I just want to mention that I created this issue because it has been caught by 
automated tests ;)

[https://github.com/solariumphp/solarium/actions] and 
[https://github.com/mkalkbrenner/search_api_solr/actions]

Beside PDF extraction we test things like the Collections API, Streaming 
Expressions, Facets, JSON Facets, ...

Therefore we run Solr docker containers. Maybe we can add a test that includes 
a build of a docker container using the latest Solr dev version?

> Error 500 on PDF extraction: java.lang.NoClassDefFoundError: 
> org/eclipse/jetty/server/MultiParts
> 
>
> Key: SOLR-14768
> URL: https://issues.apache.org/jira/browse/SOLR-14768
> Project: Solr
>  Issue Type: Bug
>  Security Level: Public(Default Security Level. Issues are Public) 
>  Components: contrib - Solr Cell (Tika extraction)
>Affects Versions: 8.6, 8.6.1
>Reporter: Markus Kalkbrenner
>Assignee: David Smiley
>Priority: Major
> Attachments: Solr v8.6.x fails with multipart MIME in commands.eml, 
> Solr v8.6.x fails with multipart MIME in commands.eml, Solr v8.6.x fails with 
> multipart MIME in commands.eml
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> See [https://www.mail-archive.com/solr-user@lucene.apache.org/msg152182.html]
> The integration tests of the solarium PHP library and the integration tests 
> of the Search API Solr Drupal module both fail on PDF extraction if executed 
> on Solr 8.6.
> They still work on Solr 8.5.1 an earlier versions.
> {quote}2020-08-20 12:30:35.279 INFO (qtp855700733-19) [ x:5f3e6ce2810ef] 
> o.a.s.u.p.LogUpdateProcessorFactory [5f3e6ce2810ef] webapp=/solr 
> path=/update/extract 
> params=\{json.nl=flat&commitWithin=0&omitHeader=false&resource.name=testpdf.pdf&literal.id=extract-test&commit=true&extractOnly=false&uprefix=attr_&wt=json}{add=[extract-test
>  (1675547519474466816)],commit=} 0 957
> solr8_1 | 2020-08-20 12:30:35.280 WARN (qtp855700733-19) [ ] 
> o.e.j.s.HttpChannel /solr/5f3e6ce2810ef/update/extract => 
> java.lang.NoClassDefFoundError: org/eclipse/jetty/server/MultiParts
> solr8_1 | at 
> org.apache.solr.servlet.SolrRequestParsers.cleanupMultipartFiles(SolrRequestParsers.java:624)
> solr8_1 | java.lang.NoClassDefFoundError: org/eclipse/jetty/server/MultiParts
> solr8_1 | at 
> org.apache.solr.servlet.SolrRequestParsers.cleanupMultipartFiles(SolrRequestParsers.java:624)
>  ~[?:?]
> solr8_1 | at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:443)
>  ~[?:?]
> solr8_1 | at 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:345)
>  ~[?:?]
> solr8_1 | at 
> org.eclipse.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1596)
>  ~[jetty-servlet-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.servlet.ServletHandler.doHandle(ServletHandler.java:545) 
> ~[jetty-servlet-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.ScopedHandler.handle(ScopedHandler.java:143) 
> ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.security.SecurityHandler.handle(SecurityHandler.java:590) 
> ~[jetty-security-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.HandlerWrapper.handle(HandlerWrapper.java:127)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.ScopedHandler.nextHandle(ScopedHandler.java:235)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.session.SessionHandler.doHandle(SessionHandler.java:1610)
>  ~[jetty-server-9.4.27.v20200227.jar:9.4.27.v20200227]
> solr8_1 | at 
> org.eclipse.jetty.server.handler.ScopedHand

[GitHub] [lucene-solr] dweiss commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


dweiss commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689410243


   I've tried to clean up things a bit and ended up implementing both forked 
and in-JVM mode. It's not super beautiful (those intermediate interfaces could 
probably be simplified) but it works and allows for some interesting 
combinations of mixed forked-and-local workers. 
   
   I committed this in to your branch, Uwe, but it's really subject to 
criticism, modifications and/or reversal - your decision.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Created] (LUCENE-9515) Detach DWPT from DefaultIndexingChain

2020-09-09 Thread Simon Willnauer (Jira)
Simon Willnauer created LUCENE-9515:
---

 Summary: Detach DWPT from DefaultIndexingChain 
 Key: LUCENE-9515
 URL: https://issues.apache.org/jira/browse/LUCENE-9515
 Project: Lucene - Core
  Issue Type: Improvement
Affects Versions: master (9.0), 8.7
Reporter: Simon Willnauer


DefaultIndexingChain or rather its super class DocConsumer is still tightly 
coupled to DocumentsWriterPerThread. In oder to guarantee better code isolation 
and hopefully independent testing down the road we can expand the IndexingChain 
interface to take all the primitives needed rather than the complex DWPT class. 
It's unclear anyway how this API (DocConsumer and IndexingChain) should be 
implemented with all the classes it depends on being package private. I will 
open a follow-up to discuss if we can remove that abstraction. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689410718


   > I don't want to block the changes here, but just think we should make a 
followup to implement the same test, but with clients in the same JVM.
   >
   > There are two separate cases for NIO locks (held by same VM vs held by 
another process), so we should stress test them both. > Past experience has 
shown that whatever lock testing we have, it is somehow never enough.
   
   I think this could be a separate issue. Basically, this just "restores" the 
test in master, which was done in Ant only. There are no functional changes.
   
   But this new code is easy to adapt, there are several possibilities:
   - We can clone the whole test and just replaces `List processes` by 
a thread pool and instead of spawning a JVM we just call main() in a thread. 
Cleanup is more or less the same.
   - We can modify the current test to not only spawn multiple JVMs, but 
aditionally also add a thread. We can even randomize the number of 
clients/threads (currently fixed to 2 clients, but it's just a change in test 
config).
   
   In both cases, the policy also needs the "connect" permission. An 
alternative would be to allow client/server to alternatively communicate 
through a PipedInput/OutputStream (if its in same JVM).
   
   Some questions before I commit this:
   - Should I rename the test to TestStressLockFactories?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] s1monw opened a new pull request #1848: LUCENE-9515: Detach DWPT from DefaultIndexingChain

2020-09-09 Thread GitBox


s1monw opened a new pull request #1848:
URL: https://github.com/apache/lucene-solr/pull/1848


   This change removes the DWPT dependency from DefaultIndexingChain
   and rather passes on the primitives needed for creating the chain.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Created] (LUCENE-9516) Remove DocConsumer and IndexingChain from Lucene

2020-09-09 Thread Simon Willnauer (Jira)
Simon Willnauer created LUCENE-9516:
---

 Summary: Remove DocConsumer and IndexingChain from Lucene
 Key: LUCENE-9516
 URL: https://issues.apache.org/jira/browse/LUCENE-9516
 Project: Lucene - Core
  Issue Type: Wish
Reporter: Simon Willnauer


Disclaimer: This is a breaking change!

We allow today to replace the entire indexing chain which is a fundamental part 
of our software. I personally don't know if there are any users of this API. 
but given the complexity I personally don't think we should further support 
this. If you are willing to implement this entire thing yourself I really 
wonder if you are better off building lucene from the source. An option like 
this on IWC might make users look into it while I am convinced they shouldn't. 
It's too complex and nothing is made for reuse down there. I wonder what others 
think but from my perspective it's time to remove it in 9.0



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9511) Include StoredFieldsWriter in DWPT accounting

2020-09-09 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9511?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192727#comment-17192727
 ] 

ASF subversion and git services commented on LUCENE-9511:
-

Commit 16a703fa14f6bab8a6852a82dddadd48a63df732 in lucene-solr's branch 
refs/heads/branch_8x from Simon Willnauer
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=16a703f ]

LUCENE-9511: Increase RAM buffer in test since now we account more ram that 
get's allocated


> Include StoredFieldsWriter in DWPT accounting
> -
>
> Key: LUCENE-9511
> URL: https://issues.apache.org/jira/browse/LUCENE-9511
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Simon Willnauer
>Priority: Major
>  Time Spent: 1h 10m
>  Remaining Estimate: 0h
>
> StoredFieldsWriter might consume some heap space memory that can have a 
> significant impact on decisions made in the IW if writers should be stalled 
> or DWPTs should be flushed if memory settings are small in IWC and flushes 
> are frequent. We should add some accounting to the StoredFieldsWriter since 
> it's part of the DWPT lifecycle and not just present during flush.
> Our nightly builds ran into some OOMs due to the large chunk size used in the 
> CompressedStoredFieldsFormat. The reason are very frequent flushes due to 
> small maxBufferedDocs which causes 300+ DWPTs to be blocked for flush causing 
> ultimately an OOM exception.
> {noformat}
>  
>  NOTE: reproduce with: ant test  -Dtestcase=TestIndexingSequenceNumbers 
> -Dtests.method=testStressConcurrentCommit -Dtests.seed=A04943A98C8E2954 
> -Dtests.nightly=true -Dtests.slow=true -Dtests.badapples=true 
> -Dtests.locale=vo-001 -Dtests.timezone=Africa/Ouagadougou 
> -Dtests.asserts=true -Dtests.file.encoding=UTF8*06:06:15*[junit4] ERROR   
>  107s J3 | TestIndexingSequenceNumbers.testStressConcurrentCommit 
> <<<*06:06:15*[junit4]> Throwable #1: 
> org.apache.lucene.store.AlreadyClosedException: this IndexWriter is 
> closed*06:06:15*[junit4]>at 
> org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:876)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.IndexWriter.ensureOpen(IndexWriter.java:890)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.IndexWriter.commit(IndexWriter.java:3727)*06:06:15*   
>  [junit4]> at 
> org.apache.lucene.index.TestIndexingSequenceNumbers.testStressConcurrentCommit(TestIndexingSequenceNumbers.java:228)*06:06:15*
> [junit4]>  at 
> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native 
> Method)*06:06:15*[junit4]>at 
> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)*06:06:15*
> [junit4]>  at 
> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)*06:06:15*
> [junit4]>  at 
> java.base/java.lang.reflect.Method.invoke(Method.java:566)*06:06:15*
> [junit4]>at 
> java.base/java.lang.Thread.run(Thread.java:834)*06:06:15*[junit4]> 
> Caused by: java.lang.OutOfMemoryError: Java heap space*06:06:15*[junit4]  
>   > at 
> __randomizedtesting.SeedInfo.seed([A04943A98C8E2954]:0)*06:06:15*[junit4] 
>>   at 
> org.apache.lucene.store.GrowableByteArrayDataOutput.(GrowableByteArrayDataOutput.java:46)*06:06:15*
> [junit4]>   at 
> org.apache.lucene.codecs.compressing.CompressingStoredFieldsWriter.(CompressingStoredFieldsWriter.java:111)*06:06:15*
> [junit4]> at 
> org.apache.lucene.codecs.compressing.CompressingStoredFieldsFormat.fieldsWriter(CompressingStoredFieldsFormat.java:130)*06:06:15*
> [junit4]>   at 
> org.apache.lucene.codecs.lucene87.Lucene87StoredFieldsFormat.fieldsWriter(Lucene87StoredFieldsFormat.java:141)*06:06:15*
> [junit4]>at 
> org.apache.lucene.codecs.asserting.AssertingStoredFieldsFormat.fieldsWriter(AssertingStoredFieldsFormat.java:48)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.StoredFieldsConsumer.initStoredFieldsWriter(StoredFieldsConsumer.java:39)*06:06:15*
> [junit4]> at 
> org.apache.lucene.index.StoredFieldsConsumer.startDocument(StoredFieldsConsumer.java:46)*06:06:15*
> [junit4]>  at 
> org.apache.lucene.index.DefaultIndexingChain.startStoredFields(DefaultIndexingChain.java:426)*06:06:15*
> [junit4]> at 
> org.apache.lucene.index.DefaultIndexingChain.processDocument(DefaultIndexingChain.java:462)*06:06:15*
> [junit4]>   at 
> org.apache.lucene.index.DocumentsWriterPerThread.updateDocuments(DocumentsWriterPerThread.java:233)*06:06:15*
> [junit4]>   at 
> org.apache.lucene.index.DocumentsWriter.updateDocuments(DocumentsWriter.java:419)*06:06:15*
> [jun

[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689463433


   Hi @dweiss 
   indeed this looks now very complicated (before the test was so simple after 
my changes from last night). Indeed I want to remove some abstraction, the 
factory is unneeded. I always try to use the default functional interfaces 
provided by Java 8, and just creating one for exactly one methods is code 
bloat. A simple if/else in the factory method would be as fine.
   
   We should maybe also change the number of client to 3 instead of 2. Also 
during nightly tests we can hammer with even more clients.
   
   What I committed in: I disabled all custom Mock Filesystems, as this is 
important to run on native OS without any wrapping.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Created] (LUCENE-9517) BugfixDeflater_JDK8252739 causes Java security issues in JDk11

2020-09-09 Thread Ignacio Vera (Jira)
Ignacio Vera created LUCENE-9517:


 Summary: BugfixDeflater_JDK8252739 causes Java security issues in 
JDk11
 Key: LUCENE-9517
 URL: https://issues.apache.org/jira/browse/LUCENE-9517
 Project: Lucene - Core
  Issue Type: Bug
Reporter: Ignacio Vera


We are running into issues when running Elasticsearch CI with java security 
turned on and using JDK11 (only for the ones that contains the jdk bug ).   The 
errors look like:

 

{{}}

{{}}
{code:java}
java.security.AccessControlException: access denied 
("java.lang.RuntimePermission" "accessDeclaredMembers") {code}
{{}}

 

The issue seems to be here:

[http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l989]

As we now have a subclass that wants to run this code. Note that this code has 
been removed in JDK12 and above.

We might need to wrap the creation of this object in a doPriviledged Block or 
find a different solution that does not need to subclass the Deflater class.

 

cc: [~uschindler]

 

 

 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Updated] (LUCENE-9517) BugfixDeflater_JDK8252739 causes Java security issues in JDk11

2020-09-09 Thread Ignacio Vera (Jira)


 [ 
https://issues.apache.org/jira/browse/LUCENE-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ignacio Vera updated LUCENE-9517:
-
Description: 
We are running into issues when running Elasticsearch CI with java security 
turned on and using JDK11 (only for the ones that contains the jdk bug ).   The 
errors look like:

 

 
{code:java}
java.security.AccessControlException: access denied 
("java.lang.RuntimePermission" "accessDeclaredMembers") {code}
 

The issue seems to be here:

[http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l989]

As we now have a subclass that wants to run this code. Note that this code has 
been removed in JDK12 and above.

We might need to wrap the creation of this object in a doPriviledged Block or 
find a different solution that does not need to subclass the Deflater class.

 

cc: [~uschindler]

 

 

 

  was:
We are running into issues when running Elasticsearch CI with java security 
turned on and using JDK11 (only for the ones that contains the jdk bug ).   The 
errors look like:

 

{{}}

{{}}
{code:java}
java.security.AccessControlException: access denied 
("java.lang.RuntimePermission" "accessDeclaredMembers") {code}
{{}}

 

The issue seems to be here:

[http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l989]

As we now have a subclass that wants to run this code. Note that this code has 
been removed in JDK12 and above.

We might need to wrap the creation of this object in a doPriviledged Block or 
find a different solution that does not need to subclass the Deflater class.

 

cc: [~uschindler]

 

 

 


> BugfixDeflater_JDK8252739 causes Java security issues in JDk11
> --
>
> Key: LUCENE-9517
> URL: https://issues.apache.org/jira/browse/LUCENE-9517
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Ignacio Vera
>Priority: Major
>
> We are running into issues when running Elasticsearch CI with java security 
> turned on and using JDK11 (only for the ones that contains the jdk bug ).   
> The errors look like:
>  
>  
> {code:java}
> java.security.AccessControlException: access denied 
> ("java.lang.RuntimePermission" "accessDeclaredMembers") {code}
>  
> The issue seems to be here:
> [http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l989]
> As we now have a subclass that wants to run this code. Note that this code 
> has been removed in JDK12 and above.
> We might need to wrap the creation of this object in a doPriviledged Block or 
> find a different solution that does not need to subclass the Deflater class.
>  
> cc: [~uschindler]
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] mikemccand commented on a change in pull request #1847: LUCENE-9514: Include TermVectorsWriter in DWPT accounting

2020-09-09 Thread GitBox


mikemccand commented on a change in pull request #1847:
URL: https://github.com/apache/lucene-solr/pull/1847#discussion_r485497049



##
File path: lucene/core/src/java/org/apache/lucene/util/Accountable.java
##
@@ -41,4 +41,8 @@
 return Collections.emptyList();
   }
 
+  /**
+   * An accountable that always returns 0
+   */
+  Accountable NULL_ACCOUNTABLE = () -> 0;

Review comment:
   Interface members are always `static final` right?

##
File path: 
lucene/codecs/src/java/org/apache/lucene/codecs/simpletext/SimpleTextTermVectorsWriter.java
##
@@ -193,4 +189,9 @@ private void write(BytesRef bytes) throws IOException {
   private void newLine() throws IOException {
 SimpleTextUtil.writeNewline(out);
   }
+
+  @Override
+  public long ramBytesUsed() {
+return scratch.get().bytes.length;

Review comment:
   Ha.

##
File path: 
lucene/core/src/java/org/apache/lucene/index/StoredFieldsConsumer.java
##
@@ -42,6 +44,7 @@
   protected void initStoredFieldsWriter() throws IOException {
 if (writer == null) { // TODO can we allocate this in the ctor? we call 
start document for every doc anyway
   this.writer = codec.storedFieldsFormat().fieldsWriter(directory, info, 
IOContext.DEFAULT);
+  accountable = writer;

Review comment:
   Hmm what is happening here?  Are we using this new `accountable` member 
somewhere?  I don't see it in the diffs?  Oh, I see, it is returned in 
`DefaultIndexingChain.getChildResources()`, ok!  Maybe add comment above its 
declaration explaining that our parent/owning `DefaultIndexingChain` 
returns/uses it?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] mikemccand commented on a change in pull request #1848: LUCENE-9515: Detach DWPT from DefaultIndexingChain

2020-09-09 Thread GitBox


mikemccand commented on a change in pull request #1848:
URL: https://github.com/apache/lucene-solr/pull/1848#discussion_r485498630



##
File path: 
lucene/core/src/java/org/apache/lucene/index/DefaultIndexingChain.java
##
@@ -73,21 +78,38 @@
   // Holds fields seen in each document
   private PerField[] fields = new PerField[1];
   private final InfoStream infoStream;
-
-  DefaultIndexingChain(DocumentsWriterPerThread docWriter) {
-this.docWriter = docWriter;
-this.fieldInfos = docWriter.getFieldInfosBuilder();
-this.infoStream = docWriter.getIndexWriterConfig().getInfoStream();
+  private final ByteBlockPool.Allocator byteBlockAllocator;
+  private final LiveIndexWriterConfig indexWriterConfig;
+  private final int indexCreatedVersionMajor;
+  private final Consumer abortingExceptionConsumer;
+  private boolean hasHitAbortingException = false;

Review comment:
   Don't need the `= false` -- it is java's default already.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689467186


   I'd also change the test name now to `TestStressLockFactories`.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] mikemccand commented on a change in pull request #1848: LUCENE-9515: Detach DWPT from DefaultIndexingChain

2020-09-09 Thread GitBox


mikemccand commented on a change in pull request #1848:
URL: https://github.com/apache/lucene-solr/pull/1848#discussion_r485499853



##
File path: 
lucene/core/src/java/org/apache/lucene/index/DefaultIndexingChain.java
##
@@ -73,21 +78,38 @@
   // Holds fields seen in each document
   private PerField[] fields = new PerField[1];
   private final InfoStream infoStream;
-
-  DefaultIndexingChain(DocumentsWriterPerThread docWriter) {
-this.docWriter = docWriter;
-this.fieldInfos = docWriter.getFieldInfosBuilder();
-this.infoStream = docWriter.getIndexWriterConfig().getInfoStream();
+  private final ByteBlockPool.Allocator byteBlockAllocator;
+  private final LiveIndexWriterConfig indexWriterConfig;
+  private final int indexCreatedVersionMajor;
+  private final Consumer abortingExceptionConsumer;
+  private boolean hasHitAbortingException = false;
+
+  DefaultIndexingChain(int indexCreatedVersionMajor, SegmentInfo segmentInfo, 
Directory directory, FieldInfos.Builder fieldInfos, LiveIndexWriterConfig 
indexWriterConfig,
+   Consumer abortingExceptionConsumer) {
+this.indexCreatedVersionMajor = indexCreatedVersionMajor;
+byteBlockAllocator = new ByteBlockPool.DirectTrackingAllocator(bytesUsed);
+IntBlockPool.Allocator intBlockAllocator = new 
IntBlockAllocator(bytesUsed);
+this.indexWriterConfig = indexWriterConfig;
+assert segmentInfo.getIndexSort() == indexWriterConfig.getIndexSort();
+this.fieldInfos = fieldInfos;
+this.infoStream = indexWriterConfig.getInfoStream();
+this.abortingExceptionConsumer = abortingExceptionConsumer;
 
 final TermsHash termVectorsWriter;
-if (docWriter.getSegmentInfo().getIndexSort() == null) {
-  storedFieldsConsumer = new StoredFieldsConsumer(docWriter.codec, 
docWriter.directory, docWriter.getSegmentInfo());
-  termVectorsWriter = new TermVectorsConsumer(docWriter);
+if (segmentInfo.getIndexSort() == null) {
+  storedFieldsConsumer = new 
StoredFieldsConsumer(indexWriterConfig.getCodec(), directory, segmentInfo);
+  termVectorsWriter = new TermVectorsConsumer(intBlockAllocator, 
byteBlockAllocator, directory, segmentInfo, indexWriterConfig.getCodec());
 } else {
-  storedFieldsConsumer = new SortingStoredFieldsConsumer(docWriter.codec, 
docWriter.directory, docWriter.getSegmentInfo());
-  termVectorsWriter = new SortingTermVectorsConsumer(docWriter);
+  storedFieldsConsumer = new 
SortingStoredFieldsConsumer(indexWriterConfig.getCodec(), directory, 
segmentInfo);
+  termVectorsWriter = new SortingTermVectorsConsumer(intBlockAllocator, 
byteBlockAllocator, directory, segmentInfo, indexWriterConfig.getCodec());
 }
-termsHash = new FreqProxTermsWriter(docWriter, bytesUsed, 
termVectorsWriter);
+termsHash = new FreqProxTermsWriter(intBlockAllocator, byteBlockAllocator, 
bytesUsed, termVectorsWriter);
+  }
+
+  private void onAbortingException(Throwable th) {

Review comment:
   This was moved from `DWPT`?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] mikemccand commented on a change in pull request #1848: LUCENE-9515: Detach DWPT from DefaultIndexingChain

2020-09-09 Thread GitBox


mikemccand commented on a change in pull request #1848:
URL: https://github.com/apache/lucene-solr/pull/1848#discussion_r485499853



##
File path: 
lucene/core/src/java/org/apache/lucene/index/DefaultIndexingChain.java
##
@@ -73,21 +78,38 @@
   // Holds fields seen in each document
   private PerField[] fields = new PerField[1];
   private final InfoStream infoStream;
-
-  DefaultIndexingChain(DocumentsWriterPerThread docWriter) {
-this.docWriter = docWriter;
-this.fieldInfos = docWriter.getFieldInfosBuilder();
-this.infoStream = docWriter.getIndexWriterConfig().getInfoStream();
+  private final ByteBlockPool.Allocator byteBlockAllocator;
+  private final LiveIndexWriterConfig indexWriterConfig;
+  private final int indexCreatedVersionMajor;
+  private final Consumer abortingExceptionConsumer;
+  private boolean hasHitAbortingException = false;
+
+  DefaultIndexingChain(int indexCreatedVersionMajor, SegmentInfo segmentInfo, 
Directory directory, FieldInfos.Builder fieldInfos, LiveIndexWriterConfig 
indexWriterConfig,
+   Consumer abortingExceptionConsumer) {
+this.indexCreatedVersionMajor = indexCreatedVersionMajor;
+byteBlockAllocator = new ByteBlockPool.DirectTrackingAllocator(bytesUsed);
+IntBlockPool.Allocator intBlockAllocator = new 
IntBlockAllocator(bytesUsed);
+this.indexWriterConfig = indexWriterConfig;
+assert segmentInfo.getIndexSort() == indexWriterConfig.getIndexSort();
+this.fieldInfos = fieldInfos;
+this.infoStream = indexWriterConfig.getInfoStream();
+this.abortingExceptionConsumer = abortingExceptionConsumer;
 
 final TermsHash termVectorsWriter;
-if (docWriter.getSegmentInfo().getIndexSort() == null) {
-  storedFieldsConsumer = new StoredFieldsConsumer(docWriter.codec, 
docWriter.directory, docWriter.getSegmentInfo());
-  termVectorsWriter = new TermVectorsConsumer(docWriter);
+if (segmentInfo.getIndexSort() == null) {
+  storedFieldsConsumer = new 
StoredFieldsConsumer(indexWriterConfig.getCodec(), directory, segmentInfo);
+  termVectorsWriter = new TermVectorsConsumer(intBlockAllocator, 
byteBlockAllocator, directory, segmentInfo, indexWriterConfig.getCodec());
 } else {
-  storedFieldsConsumer = new SortingStoredFieldsConsumer(docWriter.codec, 
docWriter.directory, docWriter.getSegmentInfo());
-  termVectorsWriter = new SortingTermVectorsConsumer(docWriter);
+  storedFieldsConsumer = new 
SortingStoredFieldsConsumer(indexWriterConfig.getCodec(), directory, 
segmentInfo);
+  termVectorsWriter = new SortingTermVectorsConsumer(intBlockAllocator, 
byteBlockAllocator, directory, segmentInfo, indexWriterConfig.getCodec());
 }
-termsHash = new FreqProxTermsWriter(docWriter, bytesUsed, 
termVectorsWriter);
+termsHash = new FreqProxTermsWriter(intBlockAllocator, byteBlockAllocator, 
bytesUsed, termVectorsWriter);
+  }
+
+  private void onAbortingException(Throwable th) {

Review comment:
   This was moved from `DWPT`?
   
   Ahh, no, we forked it from there, ok.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689479947


   Hi, I reordered the methods a bit in the class and applied some changes with 
number of clients:
   - Default is now 3. Actually this does not change the test runtime (on my 
machine)
   - On nightly we create 8 clients
   
   I would still like to get rid of the Supplier. IMHO, we should remove one 
randomization level. As we now have 3 clients (or 8), a randomization could 
always have a 2:1 chance to produce a forked client. What do you think, @dweiss 
?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689480950


   FYI, the limit to 2 clients was caused by ANT's xml. Every client used 
another XML copypaste. But actually the number of clients is not really a 
performance factor.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] dweiss commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


dweiss commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689486069


   I don't mind. Reshape any way you like.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] iverase opened a new pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


iverase opened a new pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849


   see https://issues.apache.org/jira/browse/LUCENE-9517



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] iverase commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


iverase commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689503221


   A couple of notes:
   
   * Not sure how this change can be tested.
   * We know this is only an issue for JDK11, should be add a check and only 
perform privilege action in that case?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14843) Define strongly-typed cluster configuration API

2020-09-09 Thread Andrzej Bialecki (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14843?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192801#comment-17192801
 ] 

Andrzej Bialecki commented on SOLR-14843:
-

{quote}This is not a user facing api
{quote}
Correct, the purpose of this Jira is to define an internal API. Although in 
order to be useful as a read-write facade we will likely need a user-facing 
CRUD API too.

I created SIP-11 to have a place where we outline the proposal in greater 
detail.

 

> Define strongly-typed cluster configuration API
> ---
>
> Key: SOLR-14843
> URL: https://issues.apache.org/jira/browse/SOLR-14843
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Andrzej Bialecki
>Priority: Major
>  Labels: clean-api
> Fix For: master (9.0)
>
>
> Current cluster-level configuration uses a hodgepodge of traditional Solr 
> config sources (solr.xml, system properties) and the new somewhat arbitrary 
> config files kept in ZK ({{/clusterprops.json, /security.json, 
> /packages.json, /autoscaling.json}} etc...). There's no uniform 
> strongly-typed API to access and manage these configs - currently each config 
> source has its own CRUD, often relying on direct access to Zookeeper. There's 
> also no uniform method for monitoring changes to these config sources.
> This issue proposes a uniform config API facade with the following 
> characteristics:
>  * Using a single hierarchical (or at least key-based) facade for accessing 
> any global config.
>  * Using strongly-typed sub-system configs instead of opaque Map-s: 
> components would no longer deal with JSON parsing/writing, instead they would 
> use properly annotated Java objects for config CRUD. Config objects would 
> include versioning information (eg. lastModified timestamp).
>  * Isolating access to the underlying config persistence layer: components 
> would no longer directly interact with Zookeeper or files. Most likely the 
> default implementation would continue using different ZK files per-subsystem 
> in order to limit the complexity of file formats and to reduce the cost of 
> notifications for unmodified parts of the configs.
>  * Providing uniform way to register listeners for monitoring changes in 
> specific configs: components would no longer need to interact with ZK 
> watches, they would instead be notified about modified configs that they are 
> interested in.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] dweiss commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


dweiss commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689507031


   This is a hideous implementation detail inside the JDK but I think the 
privileged action should just be there - I think this is the correct way to 
dodge the problem.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] cpoerschke merged pull request #1790: Rename TestDirectoryFactory to DirectoryFactoriesTest

2020-09-09 Thread GitBox


cpoerschke merged pull request #1790:
URL: https://github.com/apache/lucene-solr/pull/1790


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] cpoerschke merged pull request #1832: SOLR-14831: remove deprecated-and-unused "facet.distrib.mco" constant

2020-09-09 Thread GitBox


cpoerschke merged pull request #1832:
URL: https://github.com/apache/lucene-solr/pull/1832


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14831) remove deprecated-and-unused "facet.distrib.mco" constant

2020-09-09 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192814#comment-17192814
 ] 

ASF subversion and git services commented on SOLR-14831:


Commit 04def9449261d78038e8c34662300791a372bfaa in lucene-solr's branch 
refs/heads/master from Christine Poerschke
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=04def94 ]

SOLR-14831: remove deprecated-and-unused "facet.distrib.mco" constant in 
FacetParams.java (#1832)



> remove deprecated-and-unused "facet.distrib.mco" constant
> -
>
> Key: SOLR-14831
> URL: https://issues.apache.org/jira/browse/SOLR-14831
> Project: Solr
>  Issue Type: Task
>  Components: SolrJ
>Reporter: Christine Poerschke
>Assignee: Christine Poerschke
>Priority: Minor
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> This is ready for removal e.g. as per the 
> https://github.com/apache/lucene-solr/blob/releases/lucene-solr/8.6.2/solr/solrj/src/java/org/apache/solr/common/params/FacetParams.java#L139-L144
>  comment:
> {code}
>* @deprecated
>* This option is no longer used nor will if affect any queries as the fix 
> has been built in. (SOLR-11711)
>* This will be removed entirely in 8.0.0
>*/
>   @Deprecated
>   public static final String FACET_DISTRIB_MCO = FACET_DISTRIB + ".mco";
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14831) remove deprecated-and-unused "facet.distrib.mco" constant

2020-09-09 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192818#comment-17192818
 ] 

ASF subversion and git services commented on SOLR-14831:


Commit b40d6e0509c2352ec6cc44637988a937f070c33a in lucene-solr's branch 
refs/heads/branch_8x from Christine Poerschke
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=b40d6e0 ]

SOLR-14831: remove deprecated-and-unused "facet.distrib.mco" constant in 
FacetParams.java (#1832)



> remove deprecated-and-unused "facet.distrib.mco" constant
> -
>
> Key: SOLR-14831
> URL: https://issues.apache.org/jira/browse/SOLR-14831
> Project: Solr
>  Issue Type: Task
>  Components: SolrJ
>Reporter: Christine Poerschke
>Assignee: Christine Poerschke
>Priority: Minor
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> This is ready for removal e.g. as per the 
> https://github.com/apache/lucene-solr/blob/releases/lucene-solr/8.6.2/solr/solrj/src/java/org/apache/solr/common/params/FacetParams.java#L139-L144
>  comment:
> {code}
>* @deprecated
>* This option is no longer used nor will if affect any queries as the fix 
> has been built in. (SOLR-11711)
>* This will be removed entirely in 8.0.0
>*/
>   @Deprecated
>   public static final String FACET_DISTRIB_MCO = FACET_DISTRIB + ".mco";
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14831) remove deprecated-and-unused "facet.distrib.mco" constant

2020-09-09 Thread Christine Poerschke (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192820#comment-17192820
 ] 

Christine Poerschke commented on SOLR-14831:


The {{FacetParams.FACET_DISTRIB}} constant is now unused then too. I'll go mark 
it as deprecated in branch_8x and remove it on master branch only.

> remove deprecated-and-unused "facet.distrib.mco" constant
> -
>
> Key: SOLR-14831
> URL: https://issues.apache.org/jira/browse/SOLR-14831
> Project: Solr
>  Issue Type: Task
>  Components: SolrJ
>Reporter: Christine Poerschke
>Assignee: Christine Poerschke
>Priority: Minor
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> This is ready for removal e.g. as per the 
> https://github.com/apache/lucene-solr/blob/releases/lucene-solr/8.6.2/solr/solrj/src/java/org/apache/solr/common/params/FacetParams.java#L139-L144
>  comment:
> {code}
>* @deprecated
>* This option is no longer used nor will if affect any queries as the fix 
> has been built in. (SOLR-11711)
>* This will be removed entirely in 8.0.0
>*/
>   @Deprecated
>   public static final String FACET_DISTRIB_MCO = FACET_DISTRIB + ".mco";
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14831) remove deprecated-and-unused "facet.distrib.mco" constant

2020-09-09 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192829#comment-17192829
 ] 

ASF subversion and git services commented on SOLR-14831:


Commit 12dab0f80e76e618c5a086b3664323d85dd0ce00 in lucene-solr's branch 
refs/heads/master from Christine Poerschke
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=12dab0f ]

SOLR-14831: remove deprecated FacetParams.FACET_DISTRIB constant


> remove deprecated-and-unused "facet.distrib.mco" constant
> -
>
> Key: SOLR-14831
> URL: https://issues.apache.org/jira/browse/SOLR-14831
> Project: Solr
>  Issue Type: Task
>  Components: SolrJ
>Reporter: Christine Poerschke
>Assignee: Christine Poerschke
>Priority: Minor
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> This is ready for removal e.g. as per the 
> https://github.com/apache/lucene-solr/blob/releases/lucene-solr/8.6.2/solr/solrj/src/java/org/apache/solr/common/params/FacetParams.java#L139-L144
>  comment:
> {code}
>* @deprecated
>* This option is no longer used nor will if affect any queries as the fix 
> has been built in. (SOLR-11711)
>* This will be removed entirely in 8.0.0
>*/
>   @Deprecated
>   public static final String FACET_DISTRIB_MCO = FACET_DISTRIB + ".mco";
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14831) remove deprecated-and-unused "facet.distrib.mco" constant

2020-09-09 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192828#comment-17192828
 ] 

ASF subversion and git services commented on SOLR-14831:


Commit 4716a0b2fa917365d5c8d7d1d29cda902a9d45a7 in lucene-solr's branch 
refs/heads/master from Christine Poerschke
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=4716a0b ]

SOLR-14831: mark FacetParams.FACET_DISTRIB as deprecated


> remove deprecated-and-unused "facet.distrib.mco" constant
> -
>
> Key: SOLR-14831
> URL: https://issues.apache.org/jira/browse/SOLR-14831
> Project: Solr
>  Issue Type: Task
>  Components: SolrJ
>Reporter: Christine Poerschke
>Assignee: Christine Poerschke
>Priority: Minor
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> This is ready for removal e.g. as per the 
> https://github.com/apache/lucene-solr/blob/releases/lucene-solr/8.6.2/solr/solrj/src/java/org/apache/solr/common/params/FacetParams.java#L139-L144
>  comment:
> {code}
>* @deprecated
>* This option is no longer used nor will if affect any queries as the fix 
> has been built in. (SOLR-11711)
>* This will be removed entirely in 8.0.0
>*/
>   @Deprecated
>   public static final String FACET_DISTRIB_MCO = FACET_DISTRIB + ".mco";
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14831) remove deprecated-and-unused "facet.distrib.mco" constant

2020-09-09 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192831#comment-17192831
 ] 

ASF subversion and git services commented on SOLR-14831:


Commit e07816ed110959009ed728b6bc30d202bf24d807 in lucene-solr's branch 
refs/heads/branch_8x from Christine Poerschke
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=e07816e ]

SOLR-14831: mark FacetParams.FACET_DISTRIB as deprecated


> remove deprecated-and-unused "facet.distrib.mco" constant
> -
>
> Key: SOLR-14831
> URL: https://issues.apache.org/jira/browse/SOLR-14831
> Project: Solr
>  Issue Type: Task
>  Components: SolrJ
>Reporter: Christine Poerschke
>Assignee: Christine Poerschke
>Priority: Minor
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> This is ready for removal e.g. as per the 
> https://github.com/apache/lucene-solr/blob/releases/lucene-solr/8.6.2/solr/solrj/src/java/org/apache/solr/common/params/FacetParams.java#L139-L144
>  comment:
> {code}
>* @deprecated
>* This option is no longer used nor will if affect any queries as the fix 
> has been built in. (SOLR-11711)
>* This will be removed entirely in 8.0.0
>*/
>   @Deprecated
>   public static final String FACET_DISTRIB_MCO = FACET_DISTRIB + ".mco";
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Resolved] (SOLR-14831) remove deprecated-and-unused "facet.distrib.mco" constant

2020-09-09 Thread Christine Poerschke (Jira)


 [ 
https://issues.apache.org/jira/browse/SOLR-14831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Christine Poerschke resolved SOLR-14831.

Fix Version/s: 8.7
   master (9.0)
   Resolution: Fixed

> remove deprecated-and-unused "facet.distrib.mco" constant
> -
>
> Key: SOLR-14831
> URL: https://issues.apache.org/jira/browse/SOLR-14831
> Project: Solr
>  Issue Type: Task
>  Components: SolrJ
>Reporter: Christine Poerschke
>Assignee: Christine Poerschke
>Priority: Minor
> Fix For: master (9.0), 8.7
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> This is ready for removal e.g. as per the 
> https://github.com/apache/lucene-solr/blob/releases/lucene-solr/8.6.2/solr/solrj/src/java/org/apache/solr/common/params/FacetParams.java#L139-L144
>  comment:
> {code}
>* @deprecated
>* This option is no longer used nor will if affect any queries as the fix 
> has been built in. (SOLR-11711)
>* This will be removed entirely in 8.0.0
>*/
>   @Deprecated
>   public static final String FACET_DISTRIB_MCO = FACET_DISTRIB + ".mco";
> {code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Updated] (LUCENE-9484) Allow index sorting to happen after the fact

2020-09-09 Thread Christine Poerschke (Jira)


 [ 
https://issues.apache.org/jira/browse/LUCENE-9484?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Christine Poerschke updated LUCENE-9484:

Fix Version/s: 8.7
   master (9.0)

> Allow index sorting to happen after the fact
> 
>
> Key: LUCENE-9484
> URL: https://issues.apache.org/jira/browse/LUCENE-9484
> Project: Lucene - Core
>  Issue Type: Improvement
>Reporter: Simon Willnauer
>Priority: Major
> Fix For: master (9.0), 8.7
>
>  Time Spent: 2h 10m
>  Remaining Estimate: 0h
>
> I did look into sorting an index after it was created and found that with 
> some smallish modifications we can actually allow that by piggibacking on 
> SortingLeafReader and addIndices in a pretty straight-forward and simple way. 
> With some smallish modifications / fixes to SortingLeafReader we can just 
> merge and unsorted index into a sorted index using a fresh index writer.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689546656


   Hi,
   I refactored the code further and removed the randomization:
   - If uses `(clientId % 4)==0` to produce a thread client and otherwise a 
process client
   - In default mode, it creates 3 clients, so one of them is always in-process
   - In nightly mode, it creates 8 clients, so two of them are in-process
   
   I think that covers all modes.
   
   I think it's ready to commit.
   
   Question: Also backport to 8.x and remove the additional ANT task?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14843) Define strongly-typed cluster configuration API

2020-09-09 Thread Jira


[ 
https://issues.apache.org/jira/browse/SOLR-14843?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192855#comment-17192855
 ] 

Jan Høydahl commented on SOLR-14843:


[~dsmiley] Feel free to adapt the SIP page instructinos to fit our needs. It 
should probably mention that a VOTE is optional but should be held if the 
dev-list discussion is clearly not conclusive. IMO a [VOTE] thread is 
light-weight, and a successful VOTE is a good foundation to start a huge work 
effort from, and likewise a failed VOTE will get you back to the design phase 
instead of wasting days of effort which ends up in a -1 on a Jira or worse a 
revert further down the road.

> Define strongly-typed cluster configuration API
> ---
>
> Key: SOLR-14843
> URL: https://issues.apache.org/jira/browse/SOLR-14843
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Andrzej Bialecki
>Priority: Major
>  Labels: clean-api
> Fix For: master (9.0)
>
>
> Current cluster-level configuration uses a hodgepodge of traditional Solr 
> config sources (solr.xml, system properties) and the new somewhat arbitrary 
> config files kept in ZK ({{/clusterprops.json, /security.json, 
> /packages.json, /autoscaling.json}} etc...). There's no uniform 
> strongly-typed API to access and manage these configs - currently each config 
> source has its own CRUD, often relying on direct access to Zookeeper. There's 
> also no uniform method for monitoring changes to these config sources.
> This issue proposes a uniform config API facade with the following 
> characteristics:
>  * Using a single hierarchical (or at least key-based) facade for accessing 
> any global config.
>  * Using strongly-typed sub-system configs instead of opaque Map-s: 
> components would no longer deal with JSON parsing/writing, instead they would 
> use properly annotated Java objects for config CRUD. Config objects would 
> include versioning information (eg. lastModified timestamp).
>  * Isolating access to the underlying config persistence layer: components 
> would no longer directly interact with Zookeeper or files. Most likely the 
> default implementation would continue using different ZK files per-subsystem 
> in order to limit the complexity of file formats and to reduce the cost of 
> notifications for unmodified parts of the configs.
>  * Providing uniform way to register listeners for monitoring changes in 
> specific configs: components would no longer need to interact with ZK 
> watches, they would instead be notified about modified configs that they are 
> interested in.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14701) Deprecate Schemaless Mode (Discussion)

2020-09-09 Thread Alexandre Rafalovitch (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14701?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192858#comment-17192858
 ] 

Alexandre Rafalovitch commented on SOLR-14701:
--

So, I am doing some quick review of actual code/configuration to better 
understand the issue. And a couple of questions come up:
 # Type widening. Like when the first pass thinks it is a plong and then it 
turns out to be a double. Do we actually need a multi-level type widening or 
just fall-back to default is sufficient if there is mismatch (plong or default, 
pdouble or default, pdate or default). The only example for multi-level is 
plong->pdouble->string if I am not mistaken. 
 # Single/Multi-valued. Currently we only support multiValue. If we are 
tracking, can the field definitions be single-valued (plong, pdouble) and then 
actual fields declare multiValued? What is the value of having double field 
type definition, I find it quite confusing and hard to track actual 
multiplicity.
 # Do we only look at new fields or do we compare with existing schema 
definitions and note the mismatch by name/mulltiplicity?
 # Do we have to split out Managed Schema JSON? Or can we just do structured 
output of name:type:multiplicity and let users compose the commands themselves? 
Or have the output format and add more formats later. Because maybe they want 
to hand-edit the schema or (as it will likely to be) the learning schema with 
this URP chain is not actually the production schema. Or because we are 
detecting mismatches in point above and the precise JSON can get messy. Either 
way, the JSON command generation needs to be in an independent tool I am 
guessing to be reusable. Which is nice, but is it in scope?

My own answers seem to be that the right approach seems to be:
 * Widen to default
 * Use single-value FieldTypes and only assign multiValued on fields themselves
 * Look at only new fields (for now), especially since the learning schema may 
not be matching production one
 * Output structured advice and leave room to do JSON generation for future Jira

Bonus questions:
 * Do the field type definitions actually need to exist in schema if we never 
create the real fields that use them (we check right now). Or does this just 
become an abstract parsing and mapping exercise?
 * Do we still need copyField instructions if we are not creating the fields 
ourselves?

Other notes to self:
 * We could probably track rough cardinality as well and/or field size for 
string (e.g. keep first 10 values/hashes and check length) and give 
recommendation based on that (for Enum, Faceting, Large field attribute, Lazy 
Loading in solrconfig.xml, etc). That's again based on focusing on advice, 
rather than pure schema generation.

 

 

> Deprecate Schemaless Mode (Discussion)
> --
>
> Key: SOLR-14701
> URL: https://issues.apache.org/jira/browse/SOLR-14701
> Project: Solr
>  Issue Type: Improvement
>  Components: Schema and Analysis
>Reporter: Marcus Eagan
>Priority: Major
> Attachments: image-2020-08-04-01-35-03-075.png
>
>
> I know this won't be the most popular ticket out there, but I am growing more 
> and more sympathetic to the idea that we should rip many of the freedoms out 
> that cause users more harm than not. One of the freedoms I saw time and time 
> again to cause issues was schemaless mode. It doesn't work as named or 
> documented, so I think it should be deprecated. 
> If you use it in production reliably and in a way that cannot be accomplished 
> another way, I am happy to hear from more knowledgeable folks as to why 
> deprecation is a bad idea. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler edited a comment on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler edited a comment on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689546656


   Hi,
   *correction*
   
   Actually we need at least two in process clients, otherwise it's not really 
testing anything with in-process (explanation: the in-porcess client is also 
"external" to the spawned process; to actually test that two lock factories in 
same process work, you need 2 local ones. My misunderstanding wa sthat the 
LockVerifyServer does not take place in locking).
   
   I refactored the code further and removed the randomization:
   - If uses `(clientId % 2)==0` to produce a thread client and otherwise a 
process client
   - In default mode, it creates 4 clients, so two of them is always in-process
   - In nightly mode, it creates 12 clients, so six of them are in-process
   
   I think that covers all modes.
   
   I think it's ready to commit.
   
   Question: Also backport to 8.x and remove the additional ANT task?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler edited a comment on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler edited a comment on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689546656


   Hi,
   *correction*
   
   Actually we need at least two in process clients, otherwise it's not really 
testing anything with in-process (explanation: the in-porcess client is also 
"external" to the spawned process; to actually test that two lock factories in 
same process work, you need 2 local ones. My misunderstanding was that the 
LockVerifyServer does not take place in locking).
   
   I refactored the code further and removed the randomization:
   - If uses `(clientId % 2)==0` to produce a thread client and otherwise a 
process client
   - In default mode, it creates 4 clients, so two of them is always in-process
   - In nightly mode, it creates 12 clients, so six of them are in-process
   
   I think that covers all modes.
   
   I think it's ready to commit.
   
   Question: Also backport to 8.x and remove the additional ANT task?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler edited a comment on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler edited a comment on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689562340


   Hi again,
   after reading the code of the LockStressTest: It tries to re-obtain a lock 
in the same JVM based on randomization. So actually the original test did 
everything right and we were also testing "smae JVM case". @rmuir Can you 
verify this?
   
   So IMHO, we can remove the inprocess stuff and rever back to the state from 
this morning.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689562340


   Hi again,
   after reading the code of the LockStressTest: It tries to re-obtain a lock 
in the same JVM based on randomization. So actually the original test did 
everything right. @rmuir Can you verify this?
   
   So IMHO, we can remove the inprocess stuff and rever back to the state from 
this morning.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689563876


   Code is here:
   ```java
   try (final Lock l = verifyLF.obtainLock(lockDir, LOCK_FILE_NAME)) {
 if (rnd.nextInt(10) == 0) {
   if (rnd.nextBoolean()) {
 verifyLF = new 
VerifyingLockFactory(getNewLockFactory(lockFactoryClassName), in, out);
   }
   try (final Lock secondLock = verifyLF.obtainLock(lockDir, 
LOCK_FILE_NAME)) {
 throw new IOException("Double obtain");
   } catch (LockObtainFailedException loe) {
 // pass
   }
 }
 Thread.sleep(sleepTimeMS);
   } catch (LockObtainFailedException loe) {
 // obtain failed
   }
   ```



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler edited a comment on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler edited a comment on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689562340


   Hi again,
   after reading the code of the LockStressTest: It tries to re-obtain a lock 
in the same JVM based on randomization. So actually the original test did 
everything right and we were also testing "same JVM case" since long time. 
@rmuir Can you verify this?
   
   So IMHO, we can remove the inprocess stuff and revert back to the state from 
this morning.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689574638


   I was talking with Robert. Let's revert the in-process stuff. Makes not much 
difference. We actually have a very good test for in-process logs, part of 
BaseLockFactoryTestCase.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] s1monw commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


s1monw commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689577005


   change LGTM



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14701) Deprecate Schemaless Mode (Discussion)

2020-09-09 Thread Jira


[ 
https://issues.apache.org/jira/browse/SOLR-14701?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192884#comment-17192884
 ] 

Jan Høydahl commented on SOLR-14701:


Thanks for doing this Alexandre!
bq. * Widen to default
Sure, but we should support int->float->string

bq. * Use single-value FieldTypes and only assign multiValued on fields 
themselves
Agree

bq. * Look at only new fields (for now), especially since the learning schema 
may not be matching production one
Yep 

bq. * Output structured advice and leave room to do JSON generation for future 
Jira
I'd say update schema at the end. But I'm not fully clear on how you intend 
this tool to be used? Is it a standalone cmd tool, or just an update-chain you 
specify on http POST to your existing collection? Or an implicit handler 
/update/learning that has this update-chain preconfigured?

bq. Do the field type definitions actually need to exist in schema if we never 
create the real fields that use them (we check right now). Or does this just 
become an abstract parsing and mapping exercise?
I have earlier proposed that the major field types be made implicit and need 
not be part of schema. I think I proposed a new tag  or 
something, which would bring 'int', 'long' etc with it. That would let you do 
it explicitly if you don't include that tag...

Wrt the choice between analyzed text or string, it is impossible to know just 
by looking at text length or content whether the intention is to search as 
analyzed text or use for soring/faceting or matching. Likewise whether 
docvalues are needed, vectors, norms etc. I still think the approach from the 
AddFieldsToSchema processor approach makes sense, to index as text and string 
(capped at 256 and with _str suffix) as a default you know will be there.

> Deprecate Schemaless Mode (Discussion)
> --
>
> Key: SOLR-14701
> URL: https://issues.apache.org/jira/browse/SOLR-14701
> Project: Solr
>  Issue Type: Improvement
>  Components: Schema and Analysis
>Reporter: Marcus Eagan
>Priority: Major
> Attachments: image-2020-08-04-01-35-03-075.png
>
>
> I know this won't be the most popular ticket out there, but I am growing more 
> and more sympathetic to the idea that we should rip many of the freedoms out 
> that cause users more harm than not. One of the freedoms I saw time and time 
> again to cause issues was schemaless mode. It doesn't work as named or 
> documented, so I think it should be deprecated. 
> If you use it in production reliably and in a way that cannot be accomplished 
> another way, I am happy to hear from more knowledgeable folks as to why 
> deprecation is a bad idea. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689584918


   OK, reverted most of the changes in the stress tester. I kept the changes 
done by @dweiss about the magic numbers like starting gun "43" or 
access/release in the wire protocol.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9470) TestXYMultiPolygonShapeQueries reproducing failure

2020-09-09 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9470?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192887#comment-17192887
 ] 

ASF subversion and git services commented on LUCENE-9470:
-

Commit 7da15706da6a117ca56fdef36a2e5c3a9a714cd3 in lucene-solr's branch 
refs/heads/master from Ignacio Vera
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=7da1570 ]

LUCENE-9470: make TestXYMultiPolygonShapeQueries more resilient for CONTAINS 
queries (#1776)



> TestXYMultiPolygonShapeQueries reproducing failure
> --
>
> Key: LUCENE-9470
> URL: https://issues.apache.org/jira/browse/LUCENE-9470
> Project: Lucene - Core
>  Issue Type: Improvement
>Affects Versions: master (9.0)
>Reporter: Michael McCandless
>Priority: Major
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> I hit this while beasting tests against [~simonw]'s PR to add "merge segments 
> during getReader":
> It reproduces for me on current 
> ({{77a4d495cc553ec80001346376fd87d6b73a6059}}) mainline:
> {noformat}
>  [junit4:pickseed] Seed property 'tests.seed' already defined: 
> 50C3891F4E0641A4
>  [junit4]  says g'day! Master seed: 50C3891F4E0641A4
>  [junit4] Executing 1 suite with 1 JVM.
>  [junit4]
>  [junit4] Started J0 PID(1993475@localhost).
>  [junit4] Suite: org.apache.lucene.document.TestXYMultiPolygonShapeQueries
>  [junit4] 2> NOTE: reproduce with: ant test 
> -Dtestcase=TestXYMultiPolygonShapeQueries -Dtests.method=testRandomBig 
> -Dtests.seed=50C3891F4E0641A4 -Dtests.nightly=true -Dtests.slow=true 
> -Dtests.badap\
>  ples=true 
> -Dtests.linedocsfile=/l/simon/lucene/test-framework/src/resources/org/apache/lucene/util/2000mb.txt.gz
>  -Dtests.locale=en-IM -Dtests.timezone=America/Lima -Dtests.asserts=true 
> -Dtests.file.enco\
>  ding=UTF-8
>  [junit4] FAILURE 51.3s | TestXYMultiPolygonShapeQueries.testRandomBig <<<
>  [junit4] > Throwable #1: java.lang.AssertionError: wrong hit (first of 
> possibly more):
>  [junit4] > FAIL: id=8233 should match but did not
>  [junit4] > relation=CONTAINS
>  [junit4] > query=XYShapeQuery: field=shape:[XYLINE([3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 3.4028235E38][3.40282\
>  3E38, 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.40282\
>  35E38][3.402823E38, 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4\
>  028229E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229\
>  E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3\
>  .4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.40282\
>  35E38][3.4028229E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028233E38][3.4028229E38, 3.4028233E38]\
>  [3.4028229E38, 3.4028233E38][3.4028229E38, 3.4028233E38][3.4028229E38, 
> 3.4028233E38][3.4028229E38, 3.4028233E38][3.4028229E38, 
> 3.4028233E38][3.4028229E38, 3.4028233E38][3.4028229E38, 3.4028233E38][3.402\
>  8227E38, 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E3\
>  8, 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4\
>  028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233\
>  ...
>  3E38][3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][\
>  3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.402

[GitHub] [lucene-solr] iverase merged pull request #1776: LUCENE-9470: make TestXYMultiPolygonShapeQueries more resilient for CONTAINS queries

2020-09-09 Thread GitBox


iverase merged pull request #1776:
URL: https://github.com/apache/lucene-solr/pull/1776


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9470) TestXYMultiPolygonShapeQueries reproducing failure

2020-09-09 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9470?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192888#comment-17192888
 ] 

ASF subversion and git services commented on LUCENE-9470:
-

Commit 8ac53aa6bc21d460b69866f64dd24f65f304ced6 in lucene-solr's branch 
refs/heads/branch_8x from Ignacio Vera
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=8ac53aa ]

LUCENE-9470: make TestXYMultiPolygonShapeQueries more resilient for CONTAINS 
queries (#1776)



> TestXYMultiPolygonShapeQueries reproducing failure
> --
>
> Key: LUCENE-9470
> URL: https://issues.apache.org/jira/browse/LUCENE-9470
> Project: Lucene - Core
>  Issue Type: Improvement
>Affects Versions: master (9.0)
>Reporter: Michael McCandless
>Priority: Major
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> I hit this while beasting tests against [~simonw]'s PR to add "merge segments 
> during getReader":
> It reproduces for me on current 
> ({{77a4d495cc553ec80001346376fd87d6b73a6059}}) mainline:
> {noformat}
>  [junit4:pickseed] Seed property 'tests.seed' already defined: 
> 50C3891F4E0641A4
>  [junit4]  says g'day! Master seed: 50C3891F4E0641A4
>  [junit4] Executing 1 suite with 1 JVM.
>  [junit4]
>  [junit4] Started J0 PID(1993475@localhost).
>  [junit4] Suite: org.apache.lucene.document.TestXYMultiPolygonShapeQueries
>  [junit4] 2> NOTE: reproduce with: ant test 
> -Dtestcase=TestXYMultiPolygonShapeQueries -Dtests.method=testRandomBig 
> -Dtests.seed=50C3891F4E0641A4 -Dtests.nightly=true -Dtests.slow=true 
> -Dtests.badap\
>  ples=true 
> -Dtests.linedocsfile=/l/simon/lucene/test-framework/src/resources/org/apache/lucene/util/2000mb.txt.gz
>  -Dtests.locale=en-IM -Dtests.timezone=America/Lima -Dtests.asserts=true 
> -Dtests.file.enco\
>  ding=UTF-8
>  [junit4] FAILURE 51.3s | TestXYMultiPolygonShapeQueries.testRandomBig <<<
>  [junit4] > Throwable #1: java.lang.AssertionError: wrong hit (first of 
> possibly more):
>  [junit4] > FAIL: id=8233 should match but did not
>  [junit4] > relation=CONTAINS
>  [junit4] > query=XYShapeQuery: field=shape:[XYLINE([3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 3.4028235E38][3.40282\
>  3E38, 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.40282\
>  35E38][3.402823E38, 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4\
>  028229E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229\
>  E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3\
>  .4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.40282\
>  35E38][3.4028229E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028233E38][3.4028229E38, 3.4028233E38]\
>  [3.4028229E38, 3.4028233E38][3.4028229E38, 3.4028233E38][3.4028229E38, 
> 3.4028233E38][3.4028229E38, 3.4028233E38][3.4028229E38, 
> 3.4028233E38][3.4028229E38, 3.4028233E38][3.4028229E38, 3.4028233E38][3.402\
>  8227E38, 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E3\
>  8, 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4\
>  028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233\
>  ...
>  3E38][3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][\
>  3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.

[jira] [Resolved] (LUCENE-9470) TestXYMultiPolygonShapeQueries reproducing failure

2020-09-09 Thread Ignacio Vera (Jira)


 [ 
https://issues.apache.org/jira/browse/LUCENE-9470?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ignacio Vera resolved LUCENE-9470.
--
Fix Version/s: 8.7
 Assignee: Ignacio Vera
   Resolution: Fixed

> TestXYMultiPolygonShapeQueries reproducing failure
> --
>
> Key: LUCENE-9470
> URL: https://issues.apache.org/jira/browse/LUCENE-9470
> Project: Lucene - Core
>  Issue Type: Improvement
>Affects Versions: master (9.0)
>Reporter: Michael McCandless
>Assignee: Ignacio Vera
>Priority: Major
> Fix For: 8.7
>
>  Time Spent: 20m
>  Remaining Estimate: 0h
>
> I hit this while beasting tests against [~simonw]'s PR to add "merge segments 
> during getReader":
> It reproduces for me on current 
> ({{77a4d495cc553ec80001346376fd87d6b73a6059}}) mainline:
> {noformat}
>  [junit4:pickseed] Seed property 'tests.seed' already defined: 
> 50C3891F4E0641A4
>  [junit4]  says g'day! Master seed: 50C3891F4E0641A4
>  [junit4] Executing 1 suite with 1 JVM.
>  [junit4]
>  [junit4] Started J0 PID(1993475@localhost).
>  [junit4] Suite: org.apache.lucene.document.TestXYMultiPolygonShapeQueries
>  [junit4] 2> NOTE: reproduce with: ant test 
> -Dtestcase=TestXYMultiPolygonShapeQueries -Dtests.method=testRandomBig 
> -Dtests.seed=50C3891F4E0641A4 -Dtests.nightly=true -Dtests.slow=true 
> -Dtests.badap\
>  ples=true 
> -Dtests.linedocsfile=/l/simon/lucene/test-framework/src/resources/org/apache/lucene/util/2000mb.txt.gz
>  -Dtests.locale=en-IM -Dtests.timezone=America/Lima -Dtests.asserts=true 
> -Dtests.file.enco\
>  ding=UTF-8
>  [junit4] FAILURE 51.3s | TestXYMultiPolygonShapeQueries.testRandomBig <<<
>  [junit4] > Throwable #1: java.lang.AssertionError: wrong hit (first of 
> possibly more):
>  [junit4] > FAIL: id=8233 should match but did not
>  [junit4] > relation=CONTAINS
>  [junit4] > query=XYShapeQuery: field=shape:[XYLINE([3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 3.4028235E38][3.40282\
>  3E38, 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.40282\
>  35E38][3.402823E38, 3.4028235E38][3.402823E38, 3.4028235E38][3.402823E38, 
> 3.4028235E38][3.402823E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4\
>  028229E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229\
>  E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3\
>  .4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.40282\
>  35E38][3.4028229E38, 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028235E38][3.4028229E38, 
> 3.4028235E38][3.4028229E38, 3.4028233E38][3.4028229E38, 3.4028233E38]\
>  [3.4028229E38, 3.4028233E38][3.4028229E38, 3.4028233E38][3.4028229E38, 
> 3.4028233E38][3.4028229E38, 3.4028233E38][3.4028229E38, 
> 3.4028233E38][3.4028229E38, 3.4028233E38][3.4028229E38, 3.4028233E38][3.402\
>  8227E38, 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E3\
>  8, 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4\
>  028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233E38][3.4028227E38, 
> 3.4028233E38][3.4028227E38, 3.4028233\
>  ...
>  3E38][3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][\
>  3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028\
>  235E38, 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.4028233E38][3.4028235E38, 3.4028233E38][3.4028235E38, 
> 3.40

[GitHub] [lucene-solr] mikemccand commented on a change in pull request #1848: LUCENE-9515: Detach DWPT from DefaultIndexingChain

2020-09-09 Thread GitBox


mikemccand commented on a change in pull request #1848:
URL: https://github.com/apache/lucene-solr/pull/1848#discussion_r485644701



##
File path: lucene/core/src/java/org/apache/lucene/index/IndexWriter.java
##
@@ -272,7 +273,9 @@ static int getActualMaxDocs() {
* and a message is printed to infoStream, if set (see {@link
* IndexWriterConfig#setInfoStream(InfoStream)}).
*/
-  public final static int MAX_TERM_LENGTH = 
DocumentsWriterPerThread.MAX_TERM_LENGTH_UTF8;
+  // if you increase this, you must fix field cache impl for
+  // getTerms/getTermsIndex requires <= 32768

Review comment:
   Maybe remove this stale comment :)  FieldCache is long gone!

##
File path: lucene/core/src/test/org/apache/lucene/index/TestIndexWriter.java
##
@@ -1533,7 +1533,7 @@ public void testWickedLongTerm() throws IOException {
 Directory dir = newDirectory();
 RandomIndexWriter w = new RandomIndexWriter(random(), dir, new 
StringSplitAnalyzer());
 
-char[] chars = new char[DocumentsWriterPerThread.MAX_TERM_LENGTH_UTF8];
+char[] chars = new char[IndexWriter.MAX_TERM_LENGTH];

Review comment:
   Ooh good catch (`char[]` vs `byte[]`).





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9516) Remove DocConsumer and IndexingChain from Lucene

2020-09-09 Thread Michael McCandless (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9516?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192896#comment-17192896
 ] 

Michael McCandless commented on LUCENE-9516:


+1

> Remove DocConsumer and IndexingChain from Lucene
> 
>
> Key: LUCENE-9516
> URL: https://issues.apache.org/jira/browse/LUCENE-9516
> Project: Lucene - Core
>  Issue Type: Wish
>Reporter: Simon Willnauer
>Priority: Major
>
> Disclaimer: This is a breaking change!
> We allow today to replace the entire indexing chain which is a fundamental 
> part of our software. I personally don't know if there are any users of this 
> API. but given the complexity I personally don't think we should further 
> support this. If you are willing to implement this entire thing yourself I 
> really wonder if you are better off building lucene from the source. An 
> option like this on IWC might make users look into it while I am convinced 
> they shouldn't. It's too complex and nothing is made for reuse down there. I 
> wonder what others think but from my perspective it's time to remove it in 9.0



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] madrob merged pull request #1843: SOLR-14846 Remove Optional.ofNullable.orElse pattern

2020-09-09 Thread GitBox


madrob merged pull request #1843:
URL: https://github.com/apache/lucene-solr/pull/1843


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9513) Use seconds instead of millisecs

2020-09-09 Thread Michael McCandless (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9513?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192898#comment-17192898
 ] 

Michael McCandless commented on LUCENE-9513:


Thanks [~frankmanzhu] – fix looks great; I'll push soon.

> Use seconds instead of millisecs
> 
>
> Key: LUCENE-9513
> URL: https://issues.apache.org/jira/browse/LUCENE-9513
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/other
>Affects Versions: master (9.0), 8.6.2
>Reporter: Frank Zhu
>Priority: Trivial
> Fix For: master (9.0)
>
>
> In RangeFacetsExample, there is a trivial mistake where
> private final long nowSec is supposed to be secs.
> But it uses as millisecs.
> It doesn't affect the output due to the scale change (It will just make 
> PAST_HOUR, PAST_SIX_HOURS not as intended). Still it's better to be correct.
>  
> I've created a pull request on github with the simple fix.
> https://github.com/apache/lucene-solr/pull/1846



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] mikemccand commented on a change in pull request #1846: LUCENE-9513 Use seconds instead of millisecs

2020-09-09 Thread GitBox


mikemccand commented on a change in pull request #1846:
URL: https://github.com/apache/lucene-solr/pull/1846#discussion_r485650485



##
File path: 
lucene/demo/src/java/org/apache/lucene/demo/facet/RangeFacetsExample.java
##
@@ -46,7 +46,7 @@
 
   private final Directory indexDir = new ByteBuffersDirectory();
   private IndexSearcher searcher;
-  private final long nowSec = System.currentTimeMillis();
+  private final long nowSec = System.currentTimeMillis()/1000L;

Review comment:
   Good catch!





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14846) Avoid use of Optional.ofNullable.orElse in the same line

2020-09-09 Thread ASF subversion and git services (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192899#comment-17192899
 ] 

ASF subversion and git services commented on SOLR-14846:


Commit c902837bb237e579f9fdd0db0a8b35680b294ad1 in lucene-solr's branch 
refs/heads/master from Mike Drob
[ https://gitbox.apache.org/repos/asf?p=lucene-solr.git;h=c902837 ]

SOLR-14846 Clean up Optional use (#1843)

* Remove Optional.ofNullable.orElse pattern
* Remove use of Optional as method parameter

> Avoid use of Optional.ofNullable.orElse in the same line
> 
>
> Key: SOLR-14846
> URL: https://issues.apache.org/jira/browse/SOLR-14846
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Mike Drob
>Assignee: Mike Drob
>Priority: Minor
>  Time Spent: 0.5h
>  Remaining Estimate: 0h
>
> Optional was meant to be used in public APIs to convey to the caller that 
> they really need to null-check, not as an internal class to avoid doing our 
> own null checks. There are better methods in {{Objects}} for that if we want 
> them.
> See also Brian Goetz at JavaOne - https://youtu.be/MFzlgAuanU0?t=2719



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9517) BugfixDeflater_JDK8252739 causes Java security issues in JDk11

2020-09-09 Thread Uwe Schindler (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192902#comment-17192902
 ] 

Uwe Schindler commented on LUCENE-9517:
---

Sorry, I don't understand the issue.

Lucene's code does not do any reflection, so what's your issue exactly?

Uwe

> BugfixDeflater_JDK8252739 causes Java security issues in JDk11
> --
>
> Key: LUCENE-9517
> URL: https://issues.apache.org/jira/browse/LUCENE-9517
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Ignacio Vera
>Priority: Major
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> We are running into issues when running Elasticsearch CI with java security 
> turned on and using JDK11 (only for the ones that contains the jdk bug ).   
> The errors look like:
>  
>  
> {code:java}
> java.security.AccessControlException: access denied 
> ("java.lang.RuntimePermission" "accessDeclaredMembers") {code}
>  
> The issue seems to be here:
> [http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l989]
> As we now have a subclass that wants to run this code. Note that this code 
> has been removed in JDK12 and above.
> We might need to wrap the creation of this object in a doPriviledged Block or 
> find a different solution that does not need to subclass the Deflater class.
>  
> cc: [~uschindler]
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9517) BugfixDeflater_JDK8252739 causes Java security issues in JDk11

2020-09-09 Thread Uwe Schindler (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192905#comment-17192905
 ] 

Uwe Schindler commented on LUCENE-9517:
---

This is more a bug in this jdk code. So please report it there.

Lucene is not the only code that subclasses Deflater. Many code is doing this.

> BugfixDeflater_JDK8252739 causes Java security issues in JDk11
> --
>
> Key: LUCENE-9517
> URL: https://issues.apache.org/jira/browse/LUCENE-9517
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Ignacio Vera
>Priority: Major
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> We are running into issues when running Elasticsearch CI with java security 
> turned on and using JDK11 (only for the ones that contains the jdk bug ).   
> The errors look like:
>  
>  
> {code:java}
> java.security.AccessControlException: access denied 
> ("java.lang.RuntimePermission" "accessDeclaredMembers") {code}
>  
> The issue seems to be here:
> [http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l989]
> As we now have a subclass that wants to run this code. Note that this code 
> has been removed in JDK12 and above.
> We might need to wrap the creation of this object in a doPriviledged Block or 
> find a different solution that does not need to subclass the Deflater class.
>  
> cc: [~uschindler]
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9517) BugfixDeflater_JDK8252739 causes Java security issues in JDk11

2020-09-09 Thread Uwe Schindler (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192906#comment-17192906
 ] 

Uwe Schindler commented on LUCENE-9517:
---

Sorry, this is bullshit.

Please don't do this. Fix your CI.

> BugfixDeflater_JDK8252739 causes Java security issues in JDk11
> --
>
> Key: LUCENE-9517
> URL: https://issues.apache.org/jira/browse/LUCENE-9517
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Ignacio Vera
>Priority: Major
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> We are running into issues when running Elasticsearch CI with java security 
> turned on and using JDK11 (only for the ones that contains the jdk bug ).   
> The errors look like:
>  
>  
> {code:java}
> java.security.AccessControlException: access denied 
> ("java.lang.RuntimePermission" "accessDeclaredMembers") {code}
>  
> The issue seems to be here:
> [http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l989]
> As we now have a subclass that wants to run this code. Note that this code 
> has been removed in JDK12 and above.
> We might need to wrap the creation of this object in a doPriviledged Block or 
> find a different solution that does not need to subclass the Deflater class.
>  
> cc: [~uschindler]
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689603748


   Sorry, this is bullshit. Fix your CI and don't apply such stuff in Lucene.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14846) Avoid use of Optional.ofNullable.orElse in the same line

2020-09-09 Thread Erick Erickson (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192908#comment-17192908
 ] 

Erick Erickson commented on SOLR-14846:
---

[~mdrob] Is this something that should never be allowed? If so, WDYT about 
adding it to the forbiddenApis check?

> Avoid use of Optional.ofNullable.orElse in the same line
> 
>
> Key: SOLR-14846
> URL: https://issues.apache.org/jira/browse/SOLR-14846
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Mike Drob
>Assignee: Mike Drob
>Priority: Minor
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Optional was meant to be used in public APIs to convey to the caller that 
> they really need to null-check, not as an internal class to avoid doing our 
> own null checks. There are better methods in {{Objects}} for that if we want 
> them.
> See also Brian Goetz at JavaOne - https://youtu.be/MFzlgAuanU0?t=2719



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689606665


   Actually the bug is in JDK and not in our code.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-9517) BugfixDeflater_JDK8252739 causes Java security issues in JDk11

2020-09-09 Thread Uwe Schindler (Jira)


 [ 
https://issues.apache.org/jira/browse/LUCENE-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler resolved LUCENE-9517.
---
  Assignee: Uwe Schindler
Resolution: Won't Fix

> BugfixDeflater_JDK8252739 causes Java security issues in JDk11
> --
>
> Key: LUCENE-9517
> URL: https://issues.apache.org/jira/browse/LUCENE-9517
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Ignacio Vera
>Assignee: Uwe Schindler
>Priority: Major
>  Time Spent: 1h
>  Remaining Estimate: 0h
>
> We are running into issues when running Elasticsearch CI with java security 
> turned on and using JDK11 (only for the ones that contains the jdk bug ).   
> The errors look like:
>  
>  
> {code:java}
> java.security.AccessControlException: access denied 
> ("java.lang.RuntimePermission" "accessDeclaredMembers") {code}
>  
> The issue seems to be here:
> [http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l989]
> As we now have a subclass that wants to run this code. Note that this code 
> has been removed in JDK12 and above.
> We might need to wrap the creation of this object in a doPriviledged Block or 
> find a different solution that does not need to subclass the Deflater class.
>  
> cc: [~uschindler]
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Resolved] (LUCENE-9513) Use seconds instead of millisecs

2020-09-09 Thread Michael McCandless (Jira)


 [ 
https://issues.apache.org/jira/browse/LUCENE-9513?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Michael McCandless resolved LUCENE-9513.

Fix Version/s: 8.7
   Resolution: Fixed

Thanks [~frankmanzhu]!

> Use seconds instead of millisecs
> 
>
> Key: LUCENE-9513
> URL: https://issues.apache.org/jira/browse/LUCENE-9513
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: core/other
>Affects Versions: master (9.0), 8.6.2
>Reporter: Frank Zhu
>Priority: Trivial
> Fix For: master (9.0), 8.7
>
>  Time Spent: 10m
>  Remaining Estimate: 0h
>
> In RangeFacetsExample, there is a trivial mistake where
> private final long nowSec is supposed to be secs.
> But it uses as millisecs.
> It doesn't affect the output due to the scale change (It will just make 
> PAST_HOUR, PAST_SIX_HOURS not as intended). Still it's better to be correct.
>  
> I've created a pull request on github with the simple fix.
> https://github.com/apache/lucene-solr/pull/1846



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689611169


   Actually I remember that this same issue happened also at other places in 
the JDK, luckily it was fixed on their side, because I complained on the 
mailing list.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689611752


   If this an issue at your CI, just use StackWalker and let this go through.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] sigram commented on pull request #1684: SOLR-14613: strongly typed placement plugin interface and implementation

2020-09-09 Thread GitBox


sigram commented on pull request #1684:
URL: https://github.com/apache/lucene-solr/pull/1684#issuecomment-689612257


   Thanks @murblanc for taking the lead on this issue, and for your patience!



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (LUCENE-9444) Need an API to easily fetch facet labels for a field in a document

2020-09-09 Thread Michael McCandless (Jira)


[ 
https://issues.apache.org/jira/browse/LUCENE-9444?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192918#comment-17192918
 ] 

Michael McCandless commented on LUCENE-9444:


Overall the patch looks good!  Some small feedback:
 * Shouldn't {{FacetLabelReader}} be public so callers can interact with it?  
This is the danger of putting unit tests in the same package as the classes 
they are testing...
 * I think {{FacetLabelReader}} is not thread safe (since {{decodedOrds}} is 
reused)?  So, each thread must pull its own instance?  Can you add javadocs 
explaining the thread safety?
 * Instead of returning {{null}} if caller calls {{next()}} when {{hasNext()}} 
had returned {{false}}, can you throw an exception?
 * I think the caller must provide {{docId}} in non-decreasing order every 
time?  Can you 1) add unit test confirming you get an exception if you break 
that?  And 2) advertise this requirement in the javadocs (for both 
{{lookupLabels}} methods)?
 * Could you add {{@lucene.experimental}} to the class level javadocs?  This 
notifies the caller that these APIs are allowed to suddenly change, even in 
feature release.
 * Shouldn't the comparison be {{== INVALID_ORDINAL}} not {{<=}}?
 * I think {{hasNext}} might incorrectly return {{true}} when there are in fact 
no more labels, for the variant of {{lookupLabels}} taking a 
{{facetDimension}}, when all remaining ordinals are from a different dimension? 
 Maybe 1) add a unit test showing this issue (failing on the current patch), 
then 2) fix it, and 3) confirm the test then passes?  {{Iterator}} is annoying 
because of this {{hasNext}}/{{next}} dynamic!
 * Maybe, instead of {{Iterator}}, we should simply return 
{{List}}?  Or, maybe we just do {{nextFacetLabel}} (and it can 
return {{null}} when done)?

> Need an API to easily fetch facet labels for a field in a document
> --
>
> Key: LUCENE-9444
> URL: https://issues.apache.org/jira/browse/LUCENE-9444
> Project: Lucene - Core
>  Issue Type: Improvement
>  Components: modules/facet
>Affects Versions: 8.6
>Reporter: Ankur
>Priority: Major
>  Labels: facet
> Attachments: LUCENE-9444.patch
>
>
> A facet field may be included in the list of fields whose values are to be 
> returned for each hit.
> In order to get the facet labels for each hit we need to
>  # Create an instance of _DocValuesOrdinalsReader_ and invoke 
> _getReader(LeafReaderContext context)_ method to obtain an instance of 
> _OrdinalsSegmentReader()_
>  # _OrdinalsSegmentReader.get(int docID, IntsRef ordinals)_ method is then 
> used to fetch and decode the binary payload in the document's BinaryDocValues 
> field. This provides the ordinals that refer to facet labels in the 
> taxonomy.**
>  # Lastly TaxonomyReader.getPath(ord) is used to fetch the labels to be 
> returned.
>  
> Ideally there should be a simple API - *String[] getLabels(docId)* that hides 
> all the above details and gives us the string labels. This can be part of 
> *TaxonomyFacets* but that's just one idea.
> I am opening this issue to get community feedback and suggestions.
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] HoustonPutman commented on a change in pull request #1844: SOLR-14847 Create Solr Server TGZ

2020-09-09 Thread GitBox


HoustonPutman commented on a change in pull request #1844:
URL: https://github.com/apache/lucene-solr/pull/1844#discussion_r485683230



##
File path: solr/packaging/build.gradle
##
@@ -109,6 +109,21 @@ task toDir(type: Sync) {
   }
 }
 
+task toTgz(type: Tar) {

Review comment:
   I just ran this in powershell on a fresh install of Solr. It worked fine 
for me. Is there some other way I should test it on windows to see the file 
permission issue?





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14749) Provide a clean API for cluster-level event processing

2020-09-09 Thread Andrzej Bialecki (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14749?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192921#comment-17192921
 ] 

Andrzej Bialecki commented on SOLR-14749:
-

[~ichattopadhyaya] as I explained in the comments that follow yours in the PR, 
we actually need such coordination in Solr core. The concept of cluster 
singleton instances is important, there are quite a few components currently 
that work like this, and they all need this functionality. If Solr core doesn't 
provide this then each plugin author will have to come up with their own hacky 
and messy and buggy scheme.

If the package / plugin infra becomes more robust to support transitive 
dependencies then we can extract the actual implementation of ClusterSingleton 
management into its own plugin - but still, the interface needs to be in the 
core, with a default implementation.

Again, this is not just for autoscaling - autoscaling will be just one user of 
the ClusterSingleton concept.

I removed the Scheduler part from PR-1758. Do you still strongly object to the 
current state of the PR, Ishan? If so, could you please outline technical 
reasons for your objections?

> Provide a clean API for cluster-level event processing
> --
>
> Key: SOLR-14749
> URL: https://issues.apache.org/jira/browse/SOLR-14749
> Project: Solr
>  Issue Type: Improvement
>  Components: AutoScaling
>Reporter: Andrzej Bialecki
>Assignee: Andrzej Bialecki
>Priority: Major
>  Labels: clean-api
>  Time Spent: 9h
>  Remaining Estimate: 0h
>
> This is a companion issue to SOLR-14613 and it aims at providing a clean, 
> strongly typed API for the functionality formerly known as "triggers" - that 
> is, a component for generating cluster-level events corresponding to changes 
> in the cluster state, and a pluggable API for processing these events.
> The 8x triggers have been removed so this functionality is currently missing 
> in 9.0. However, this functionality is crucial for implementing the automatic 
> collection repair and re-balancing as the cluster state changes (nodes going 
> down / up, becoming overloaded / unused / decommissioned, etc).
> For this reason we need this API and a default implementation of triggers 
> that at least can perform automatic collection repair (maintaining the 
> desired replication factor in presence of live node changes).
> As before, the actual changes to the collections will be executed using 
> existing CollectionAdmin API, which in turn may use the placement plugins 
> from SOLR-14613.
> h3. Division of responsibility
>  * built-in Solr components (non-pluggable):
>  ** cluster state monitoring and event generation,
>  ** simple scheduler to periodically generate scheduled events
>  * plugins:
>  ** automatic collection repair on {{nodeLost}} events (provided by default)
>  ** re-balancing of replicas (periodic or on {{nodeAdded}} events)
>  ** reporting (eg. requesting additional node provisioning)
>  ** scheduled maintenance (eg. removing inactive shards after split)
> h3. Other considerations
> These plugins (unlike the placement plugins) need to execute on one 
> designated node in the cluster. Currently the easiest way to implement this 
> is to run them on the Overseer leader node.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14701) Deprecate Schemaless Mode (Discussion)

2020-09-09 Thread Alexandre Rafalovitch (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14701?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192924#comment-17192924
 ] 

Alexandre Rafalovitch commented on SOLR-14701:
--

Thanks Jan, these are good questions. To address it from the middle-out, I 
guess the main question is whether this should end up in the default schema or 
not.

If it ends up in default schema as non-default URP chain, then the usage would 
be:
 * bin/post -params "update-chain=guess-schema" documents (to update 
schema, may require to have commit command to support that)
 * bin/post ..documents (to actually index)

In that case, it makes sense to create schema at the end and to have copyField 
commands and so on.

 

On the other hand, I was envisaging (in general) to have a learning schema and 
a bunch of learning examples that layer on top of that. Schemaless mode could 
be one of the examples. Then, you would create a separate core and it could be 
a default chain there. And then it would echo recommendations at the end back 
to the user. Thinking about it, this would be a bit of a demotion for 
schemaless mode. Perhaps too much of a demotion. And perhaps too much core 
handling. So, maybe it should be a learning/guessing URP chain, not a learning 
schema after all.

 

Or maybe it can be combined somehow (still in main config) with some advice 
given in URP's finish() and schema created in commit(). So, a user could run 
the guess-schema several times, accumulating the changes (with commit off). And 
if they are happy with them, then they run commit. Or rollback. And hope for 
now autoCommit configured in schema... And actually, I am not sure how the 
output/advice actually gets back to the user. So, this may also be in the "too 
hard" category, but I capture it as a thought point.

 

As to text vs. string, I think if we see text entries longer than 256 
characters, we kind of know they make no sense as indexed strings. If we see 
much longer strings, that could trigger a warning about marking fields as long. 
But that's not something that could be created automatically.

 

*int->float->string* requires configuration to recognize that and to decide on 
whether we support just that one level of extra mapping or have to do a full 
tree-walking implementation. Or hardcode the knowledge that Int/Long can widen 
to double.

 

> Deprecate Schemaless Mode (Discussion)
> --
>
> Key: SOLR-14701
> URL: https://issues.apache.org/jira/browse/SOLR-14701
> Project: Solr
>  Issue Type: Improvement
>  Components: Schema and Analysis
>Reporter: Marcus Eagan
>Priority: Major
> Attachments: image-2020-08-04-01-35-03-075.png
>
>
> I know this won't be the most popular ticket out there, but I am growing more 
> and more sympathetic to the idea that we should rip many of the freedoms out 
> that cause users more harm than not. One of the freedoms I saw time and time 
> again to cause issues was schemaless mode. It doesn't work as named or 
> documented, so I think it should be deprecated. 
> If you use it in production reliably and in a way that cannot be accomplished 
> another way, I am happy to hear from more knowledgeable folks as to why 
> deprecation is a bad idea. 



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Updated] (SOLR-14439) Upgrade to Tika 1.24.1

2020-09-09 Thread Tim Allison (Jira)


 [ 
https://issues.apache.org/jira/browse/SOLR-14439?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tim Allison updated SOLR-14439:
---
Status: Patch Available  (was: Open)

> Upgrade to Tika 1.24.1
> --
>
> Key: SOLR-14439
> URL: https://issues.apache.org/jira/browse/SOLR-14439
> Project: Solr
>  Issue Type: Task
>  Components: contrib - DataImportHandler
>Reporter: Tim Allison
>Assignee: Tim Allison
>Priority: Major
> Attachments: SOLR-14339.patch
>
>
> We recently released 1.24.1 with several fixes for DoS vulnerabilities we 
> found via fuzzing: CVE-2020-9489 https://seclists.org/oss-sec/2020/q2/69



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689635578


   The problem with the fix here is:
   - I have no idea how the stack trace looks like. And under which 
circumstances you see this bug. Can you post a link to your CI, to see if it is 
causeed by the CI infrastrcuture? Does it also break with production ready 
software? How?
   - Have you opened a bug report in JDK 10, 11 (it's caused by this change: 
https://bugs.openjdk.java.net/browse/JDK-8185582)? Should I do?
   
   Are there any other ways how to workaround this bug?
   
   Why does this PR help for this bug (I don't know the stack trace)?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689636897


   Does it help, because this code is only called from Deflater's ctor? 
http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l207



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler edited a comment on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler edited a comment on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689635578


   The problem with the fix here is:
   - I have no idea how the stack trace looks like. And under which 
circumstances you see this bug. Can you post a link to your CI, to see if it is 
causeed by the CI infrastrcuture? Does it also break with production ready 
software? How?
   - Have you opened a bug report in JDK 10, 11 (it's caused by this change: 
https://bugs.openjdk.java.net/browse/JDK-8185582)? Should I do?
   - The workaround may only work for Elasticsearch, because lucene-core.jar 
has the accessDeclaredMembers permission given due to the MMapDirectory hack. 
But the problem is that we take now the responsibility for any bad stuff the 
Deflater internals are doing. If you remove the permission from Lucene's code, 
it won't work anymore.
   
   IMHO, as a workaround for this bug, you should add a method to your security 
manager which explicitely allows this access, if it finds the JDK class in the 
stack trace (similarily implemented like the checkExit() method in Lucene's 
test SecurityManager ando also the code that disallows calling System.exit() in 
Elasticsearch). So the workaround should really be in the consuming code.
   
   Are there any other ways how to workaround this bug?
   
   Why does this PR help for this bug (I don't know the stack trace)?



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] HoustonPutman commented on a change in pull request #1844: SOLR-14847: Create Solr Server TGZ

2020-09-09 Thread GitBox


HoustonPutman commented on a change in pull request #1844:
URL: https://github.com/apache/lucene-solr/pull/1844#discussion_r485708813



##
File path: solr/packaging/build.gradle
##
@@ -109,6 +109,21 @@ task toDir(type: Sync) {
   }
 }
 
+task toTgz(type: Tar) {

Review comment:
   To be clear, I tested creating the tgz, not using it. I can try that out 
in a bit and see if the issues arise.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on a change in pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler commented on a change in pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#discussion_r485711771



##
File path: 
lucene/core/src/java/org/apache/lucene/codecs/lucene87/BugfixDeflater_JDK8252739.java
##
@@ -40,7 +42,9 @@ public static Deflater createDeflaterInstance(int level, 
boolean nowrap, int dic
   throw new IllegalArgumentException("dictLength must be >= 0");
 }
 if (IS_BUGGY_JDK) {
-  return new BugfixDeflater_JDK8252739(level, nowrap, dictLength);
+  // Some JDKs need "accessDeclaredMembers" privileges to create a 
Deflater object
+  return AccessController.doPrivileged(
+  (PrivilegedAction) () -> new 
BugfixDeflater_JDK8252739(level, nowrap, dictLength));

Review comment:
   At least this should be `PrivilegedAction`, because that's 
what we return.





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689648275


   Maybe we should do another fix for the original bug: As only setDictionary 
is affected, my proposal would be to not subclass `Deflater` at all and instead 
just return an interface only having the Deflater#setDictionary() method.
   
   In buggy jdks it returns a interface pointing to the original Deflater's 
method, otherwise our own one. I can create a new PR for that. I like this more.
   
   This may also speedup Java 11, because it's not doing crazy reflection on 
every new Deflater instance.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Reopened] (LUCENE-9517) BugfixDeflater_JDK8252739 causes Java security issues in JDk11

2020-09-09 Thread Uwe Schindler (Jira)


 [ 
https://issues.apache.org/jira/browse/LUCENE-9517?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uwe Schindler reopened LUCENE-9517:
---

Maybe we should do a completely different fix for the original issue. I will 
provide a PR soon.

> BugfixDeflater_JDK8252739 causes Java security issues in JDk11
> --
>
> Key: LUCENE-9517
> URL: https://issues.apache.org/jira/browse/LUCENE-9517
> Project: Lucene - Core
>  Issue Type: Bug
>Reporter: Ignacio Vera
>Assignee: Uwe Schindler
>Priority: Major
>  Time Spent: 2h 10m
>  Remaining Estimate: 0h
>
> We are running into issues when running Elasticsearch CI with java security 
> turned on and using JDK11 (only for the ones that contains the jdk bug ).   
> The errors look like:
>  
>  
> {code:java}
> java.security.AccessControlException: access denied 
> ("java.lang.RuntimePermission" "accessDeclaredMembers") {code}
>  
> The issue seems to be here:
> [http://hg.openjdk.java.net/jdk/jdk11/file/1ddf9a99e4ad/src/java.base/share/classes/java/util/zip/Deflater.java#l989]
> As we now have a subclass that wants to run this code. Note that this code 
> has been removed in JDK12 and above.
> We might need to wrap the creation of this object in a doPriviledged Block or 
> find a different solution that does not need to subclass the Deflater class.
>  
> cc: [~uschindler]
>  
>  
>  



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1842: LUCENE-9512: Move LockFactory stress test to be a unit/integration test

2020-09-09 Thread GitBox


uschindler commented on pull request #1842:
URL: https://github.com/apache/lucene-solr/pull/1842#issuecomment-689649444


   @dweiss Are you fine with this? I would like to commit and backport this 
soon.
   
   Unless you have some minor comments about the cleanup process, as I had to 
remove your abstraction on top of that.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14846) Avoid use of Optional.ofNullable.orElse in the same line

2020-09-09 Thread Mike Drob (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192970#comment-17192970
 ] 

Mike Drob commented on SOLR-14846:
--

Individually calling ofNullable and calling orElse* on an Optional is fine, 
it's the immediate combination of the two that is unnecessary. I'm not sure 
that forbiddenAPI is expressive enough to select for that, but I also haven't 
spent a ton of time investigating it and would prefer to defer to an expert in 
the library.

Can forbiddenAPI keep us from declaring methods with Optional parameters? 
IntelliJ has an inspection/warning/error for that, btw.

> Avoid use of Optional.ofNullable.orElse in the same line
> 
>
> Key: SOLR-14846
> URL: https://issues.apache.org/jira/browse/SOLR-14846
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Mike Drob
>Assignee: Mike Drob
>Priority: Minor
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Optional was meant to be used in public APIs to convey to the caller that 
> they really need to null-check, not as an internal class to avoid doing our 
> own null checks. There are better methods in {{Objects}} for that if we want 
> them.
> See also Brian Goetz at JavaOne - https://youtu.be/MFzlgAuanU0?t=2719



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Resolved] (SOLR-14846) Avoid use of Optional.ofNullable.orElse in the same line

2020-09-09 Thread Mike Drob (Jira)


 [ 
https://issues.apache.org/jira/browse/SOLR-14846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mike Drob resolved SOLR-14846.
--
Fix Version/s: master (9.0)
   Resolution: Fixed

> Avoid use of Optional.ofNullable.orElse in the same line
> 
>
> Key: SOLR-14846
> URL: https://issues.apache.org/jira/browse/SOLR-14846
> Project: Solr
>  Issue Type: Improvement
>  Security Level: Public(Default Security Level. Issues are Public) 
>Reporter: Mike Drob
>Assignee: Mike Drob
>Priority: Minor
> Fix For: master (9.0)
>
>  Time Spent: 40m
>  Remaining Estimate: 0h
>
> Optional was meant to be used in public APIs to convey to the caller that 
> they really need to null-check, not as an internal class to avoid doing our 
> own null checks. There are better methods in {{Objects}} for that if we want 
> them.
> See also Brian Goetz at JavaOne - https://youtu.be/MFzlgAuanU0?t=2719



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Commented] (SOLR-14749) Provide a clean API for cluster-level event processing

2020-09-09 Thread Ishan Chattopadhyaya (Jira)


[ 
https://issues.apache.org/jira/browse/SOLR-14749?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17192974#comment-17192974
 ] 

Ishan Chattopadhyaya commented on SOLR-14749:
-

I'd prefer if this were to be completely isolated away from core. Transitive 
dependencies are supported today if all modules are in contrib and their jars 
are in classpath. Anyway, let's refactor it after this is done. I can live with 
this for now. +1 to dealing with triggers totally inside packages and not 
making the concept first class. Thanks for your work on this and patience. 

> Provide a clean API for cluster-level event processing
> --
>
> Key: SOLR-14749
> URL: https://issues.apache.org/jira/browse/SOLR-14749
> Project: Solr
>  Issue Type: Improvement
>  Components: AutoScaling
>Reporter: Andrzej Bialecki
>Assignee: Andrzej Bialecki
>Priority: Major
>  Labels: clean-api
>  Time Spent: 9h
>  Remaining Estimate: 0h
>
> This is a companion issue to SOLR-14613 and it aims at providing a clean, 
> strongly typed API for the functionality formerly known as "triggers" - that 
> is, a component for generating cluster-level events corresponding to changes 
> in the cluster state, and a pluggable API for processing these events.
> The 8x triggers have been removed so this functionality is currently missing 
> in 9.0. However, this functionality is crucial for implementing the automatic 
> collection repair and re-balancing as the cluster state changes (nodes going 
> down / up, becoming overloaded / unused / decommissioned, etc).
> For this reason we need this API and a default implementation of triggers 
> that at least can perform automatic collection repair (maintaining the 
> desired replication factor in presence of live node changes).
> As before, the actual changes to the collections will be executed using 
> existing CollectionAdmin API, which in turn may use the placement plugins 
> from SOLR-14613.
> h3. Division of responsibility
>  * built-in Solr components (non-pluggable):
>  ** cluster state monitoring and event generation,
>  ** simple scheduler to periodically generate scheduled events
>  * plugins:
>  ** automatic collection repair on {{nodeLost}} events (provided by default)
>  ** re-balancing of replicas (periodic or on {{nodeAdded}} events)
>  ** reporting (eg. requesting additional node provisioning)
>  ** scheduled maintenance (eg. removing inactive shards after split)
> h3. Other considerations
> These plugins (unlike the placement plugins) need to execute on one 
> designated node in the cluster. Currently the easiest way to implement this 
> is to run them on the Overseer leader node.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] dweiss commented on a change in pull request #1844: SOLR-14847: Create Solr Server TGZ

2020-09-09 Thread GitBox


dweiss commented on a change in pull request #1844:
URL: https://github.com/apache/lucene-solr/pull/1844#discussion_r485726236



##
File path: solr/packaging/build.gradle
##
@@ -109,6 +109,21 @@ task toDir(type: Sync) {
   }
 }
 
+task toTgz(type: Tar) {

Review comment:
   Create a tgz in Windows and unpack it on Linux. Shell scripts should 
have the executable bit set.
   This can be verified programmatically too, although the above is the 
ultimate check. :)





This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[jira] [Created] (SOLR-14849) Investigate test failures in TestReRankQParserPlugin.testMinExactCount

2020-09-09 Thread Tomas Eduardo Fernandez Lobbe (Jira)
Tomas Eduardo Fernandez Lobbe created SOLR-14849:


 Summary: Investigate test failures in 
TestReRankQParserPlugin.testMinExactCount
 Key: SOLR-14849
 URL: https://issues.apache.org/jira/browse/SOLR-14849
 Project: Solr
  Issue Type: Bug
  Security Level: Public (Default Security Level. Issues are Public)
Reporter: Tomas Eduardo Fernandez Lobbe


I've seen a couple failures in Jenkins, such as:
{noformat}
FAILED:  org.apache.solr.search.TestReRankQParserPlugin.testMinExactCount

Error Message:
Exception during query

Stack Trace:
java.lang.RuntimeException: Exception during query
at 
__randomizedtesting.SeedInfo.seed([D74824A60D443A78:A2C81E3EE299C40F]:0)
at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:1016)
at org.apache.solr.SolrTestCaseJ4.assertQ(SolrTestCaseJ4.java:976)
at 
org.apache.solr.search.TestReRankQParserPlugin.testMinExactCount(TestReRankQParserPlugin.java:675)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817)
at 
com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883)
at 
com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.SystemPropertiesRestoreRule$1.evaluate(SystemPropertiesRestoreRule.java:57)
at 
org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at 
org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47)
at 
org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64)
at 
org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54)
at 
com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at 
com.car

[GitHub] [lucene-solr] iverase commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


iverase commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689663895


   Thanks Uwe for taking the time reviewing this issue. 
   
   >I have no idea how the stack trace looks like. And under which 
circumstances you see this bug. Can you post a link to your CI, to see if it is 
causeed by the CI infrastrcuture? Does it also break with production ready 
software? How?
   
   Stack-trace looks like:
   
   ```
   java.security.AccessControlException: access denied 
("java.lang.RuntimePermission" "accessDeclaredMembers") |  
   
   at __randomizedtesting.SeedInfo.seed([7E21C06936079E14:4A7EB7FED693B2BB]:0) 
|  
   -- | --
 |   | at 
java.security.AccessControlContext.checkPermission(AccessControlContext.java:472)
 |  
 |   | at 
java.security.AccessController.checkPermission(AccessController.java:897) |  
 |   | at 
java.lang.SecurityManager.checkPermission(SecurityManager.java:322) |  
 |   | at java.lang.Class.checkMemberAccess(Class.java:2847) |  
 |   | at java.lang.Class.getDeclaredMethod(Class.java:2471) |  
 |   | at java.util.zip.Deflater$DeflaterZStreamRef.get(Deflater.java:991) 
|  
 |   | at java.util.zip.Deflater.(Deflater.java:207) |  
 |   | at 
org.apache.lucene.codecs.lucene87.BugfixDeflater_JDK8252739.(BugfixDeflater_JDK8252739.java:53)
 |  
 |   | at 
org.apache.lucene.codecs.lucene87.BugfixDeflater_JDK8252739.createDeflaterInstance(BugfixDeflater_JDK8252739.java:43)
 |  
   ```
   
   I believe It happens every time  the runtime environment is set to JDK11 na 
dit is a affected by the Deflator bug. Here is a link to one of the logs:
   
   
https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+7.x+matrix-java-periodic-fips/ES_RUNTIME_JAVA=adoptopenjdk11,nodes=general-purpose/212/consoleFull
   
   This is un-released code so it is not production ready.
   
   > Have you opened a bug report in JDK 10, 11 (it's caused by this change: 
https://bugs.openjdk.java.net/browse/JDK-8185582)? Should I do?
   
   It would be great if you can report it. 
   
   > The workaround may only work for Elasticsearch, because lucene-core.jar 
has the accessDeclaredMembers permission given due to the MMapDirectory hack. 
   
   This is a good point and I agree this fix only works if you give special 
permissions to the jar.
   
   > Maybe we should do another fix for the original bug: As only setDictionary 
is affected, my proposal would be to not subclass Deflater at all and instead 
just return an interface only having the Deflater#setDictionary() method.
   
   That is another option I considered and it now makes more sense than this 
approach. +1 not to subclass Deflater.
   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler opened a new pull request #1850: LUCENE-9517: Don't subclass Deflater and instead create a patch for setDictionary() using a functional interface

2020-09-09 Thread GitBox


uschindler opened a new pull request #1850:
URL: https://github.com/apache/lucene-solr/pull/1850


   Bugfix for LUCENE-9517: Instead of subclassing Deflater like in LUCENE-9500, 
we just provide a method reference (using a functional interface), so code can 
call Deflater#setDictionary.
   
   The issue was caused by a missing AccessController#setPrivileged() in JDK's 
code when Deflater was subclassed. This workaround does not subclass, so the 
missing doPrivileged is not needed.
   
   This also adds the faulty methods to forbiddenapis.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler commented on pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler commented on pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849#issuecomment-689670688


   I have a better fix without hacks, just not subclassing Deflater!
   
   See PR #1850. I will close this one.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] uschindler closed pull request #1849: LUCENE-9517: Add doPrivileged block when creating BugfixDeflater_JDK8252739 objects

2020-09-09 Thread GitBox


uschindler closed pull request #1849:
URL: https://github.com/apache/lucene-solr/pull/1849


   



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene-solr] sigram commented on pull request #1831: SOLR-14749 the scheduler part

2020-09-09 Thread GitBox


sigram commented on pull request #1831:
URL: https://github.com/apache/lucene-solr/pull/1831#issuecomment-689671370


   @noblepaul in this PR the Scheduler depends on ClusterSingleton, that's why 
I imported it into this PR. Whatever the ClusterSingleton becomes it will be 
the version from PR-1758.



This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org



-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



  1   2   >