Hi,
Is it possible I can sampling based on "search result"?
Like run query first, and search result return 1 million documents.
With random sampling, 50% (500K) documents return for facet, and stats.
The sampling need based on "search result".
Thanks,
Yongtao
--
View this message in context:
hi,
We want to provide remove duplicate from search result function.
like we have below documents.
id(uniqueKey) guid
doc1G1
doc2G2
doc3G3
doc4G1
user run one query and hit doc1, doc2 and doc4.
user want to remove
All,
We saw one issue on fnm file.
Looks fnm file size will not be reduced after optimize.
Like we have 1000 documents, and they have field 1 to 1000.
And .fnm file size is 10K.
After delete 999 documents and just keep one document which just has 2
fields.
After run optimize, .fnm file still has
I check implementation.
in SegmentMerger.mergeFieldInfos
public void mergeFieldInfos() throws IOException {
for (AtomicReader reader : mergeState.readers) {
FieldInfos readerFieldInfos = reader.getFieldInfos();
for (FieldInfo fi : readerFieldInfos) {
fieldInfosBuilder.ad
For our setup, the file size is 123M. Internal it has 2.6M fields.
The problem is facet operation. It take a while for facet.
we are stuck in below call stack for 11 second.
java.util.HashMap.transfer(Unknown Source)
java.util.HashMap.resize(Unknown Source)
java.util.HashMap.addEntry(Unknown S
Erick,
Thanks for your update.
The problem this this data is will being until whole document in the section
be deleted.
I understand this is cause optimize double scan index folder in this case.
We may add some logic to check when the file size do this scan when the file
size is too bigger.
Yon
all,
Is it possible I can add new share to existing cloud?
I first call below to create colleciton1
java -DzkRun -Dbootstrap_confdir=./solr/collection1/conf
-Dcollection.configName=myconf -jar start.jar
I did not give "numShards" means it will use implicit router.
After that, how can I add new