That's OK, I have determined that caching is not relevant to our use case. However, I have a question about the full import queries that we are using:
Here is the SQL from the top level entity: query="SELECT DISTINCT 'LEAP_PARTY' AS DOCUMENT_TYPE, VPARTY.OWNER AS PARTY_OWNER, VPARTY.PARTY_ID AS PARTY_PARTY_ID, VPARTY.PARTY_ID AS LEAP_PARTY_ID,VPARTY.OWNER AS LEAP_PARTY_OWNER FROM VPARTY" The values from the top level entity are being inserted one by one into the child entity SQL query="SELECT DISTINCT OFFER.REQUEST_NO AS OFFER_REQUEST_NO,OFFER.OWNER AS OFFER_OWNER,OFFER.OFFER_NO AS OFFER_OFFER_NO,OFFER.MODIFY_TS FROM OFFER INNER JOIN (SELECT DISTINCT REQUEST_NO, OWNER, OFFER_NO FROM OFFER WHERE SUPPLIER='${LEAP_PARTY.LEAP_PARTY_ID}' AND OWNER='${LEAP_PARTY.LEAP_PARTY_OWNER}') SQ ON OFFER.REQUEST_NO=SQ.REQUEST_NO AND OFFER.OWNER=SQ.OWNER AND SQ.OFFER_NO = OFFER.OFFER_NO ORDER BY OFFER.MODIFY_TS DESC" Say we have 1000000 party records. Then the child SQL will be run 1000000 times (once for each party record). Isn't there a way to just run the child SQL on all of the party records at once with a join, using a GROUP BY and ORDER BY on the PARTY_ID? Then the results from that query could easily be placed in SOLR according to the primary key (party_id). Is there some part of the Data Import Handler that operates that way? -- View this message in context: http://lucene.472066.n3.nabble.com/cache-implemetation-tp4218825p4218851.html Sent from the Solr - User mailing list archive at Nabble.com.