Hi Amanda, I did this https://github.com/freedev/solr-import-export-json and works with/without cursormark, in case your index does not have an unique key (primary key).
On Fri, Jan 31, 2020 at 1:18 PM Amanda Shuman <amanda.shu...@gmail.com> wrote: > Thanks all! > > I wasn't familiar with using curl at the command line at all, but I did try > a basic curl yesterday based on this thread, admin console attribute > syntax, and the tutorial in solr documentation ( > https://lucene.apache.org/solr/guide/8_4/solr-tutorial.html) and was able > to produce the file. I basically did what Steve Ge suggested, the command > looks kind of like this for anyone else who needs it in the future: > > curl " > > http://servername.com:8983/solr/collection1/select?indent=on&q=*:*&rows=5000&wt=json > " > > collection1_index.json > > I just set the rows to the number in our index, which I got from the admin > console. > > Amanda > > ------ > Dr. Amanda Shuman > Researcher and Lecturer, Institute of Chinese Studies, University of > Freiburg > Coordinator for the MA program in Modern China Studies > Database Administrator, The Maoist Legacy <https://maoistlegacy.de/> > PhD, University of California, Santa Cruz > http://www.amandashuman.net/ > http://www.prchistoryresources.org/ > Office: +49 (0) 761 203 96748 > > > > On Wed, Jan 29, 2020 at 4:21 PM Emir Arnautović < > emir.arnauto...@sematext.com> wrote: > > > Hi Amanda, > > I assume that you have all the fields stored so you will be able to > export > > full document. > > > > Several thousands records should not be too much to use regular > start+rows > > to paginate results, but the proper way of doing that would be to use > > cursors. Adjust page size to avoid creating huge responses and you can > use > > curl or some similar tool to avoid using admin console. I did a quick > > search and there are several blog posts with scripts that does what you > > need. > > > > HTH, > > Emir > > > > -- > > Monitoring - Log Management - Alerting - Anomaly Detection > > Solr & Elasticsearch Consulting Support Training - http://sematext.com/ > > > > > > > > > On 29 Jan 2020, at 15:43, Amanda Shuman <amanda.shu...@gmail.com> > wrote: > > > > > > Dear all: > > > > > > I've been asked to produce a JSON file of our index so it can be > combined > > > and indexed with other records. (We run solr 5.3.1 on this project; > we're > > > not going to upgrade, in part because funding has ended.) The index has > > > several thousand rows, but nothing too drastic. Unfortunately, this is > > too > > > much to handle for a simple query dump from the admin console. I tried > to > > > follow instructions related to running /export directly but I guess the > > > export handler isn't installed. I tried to divide the query into rows, > > but > > > after a certain amount it freezes, and it also freezes when I try to > > limit > > > rows (e.g., rows 501-551 freezes the console). Is there any other way > to > > > export the index short of having to install the export handler > > considering > > > we're not working on this project anyone? > > > > > > Thanks, > > > Amanda > > > > > > ------ > > > Dr. Amanda Shuman > > > Researcher and Lecturer, Institute of Chinese Studies, University of > > > Freiburg > > > Coordinator for the MA program in Modern China Studies > > > Database Administrator, The Maoist Legacy <https://maoistlegacy.de/> > > > PhD, University of California, Santa Cruz > > > http://www.amandashuman.net/ > > > http://www.prchistoryresources.org/ > > > Office: +49 (0) 761 203 96748 > > > > > -- Vincenzo D'Amore