-----Original Message-----
From: LOPEZ-CORTES Mariano-ext
[mailto:mariano.lopez-cortes-ext@pole-emploi.fr]
Sent: 19 March 2018 21:22
To: 'solr-user@lucene.apache.org' <solr-user@lucene.apache.org>
Subject: RE: Question liste solr
Sorry. Thanks in advance !!
De : LOPEZ-CORTES Mariano-ext
Envoyé : lundi 19 mars 2018 16:50
À : 'solr-user@lucene.apache.org'
Objet : RE: Question liste solr
Hello
We have an index Solr with 3 nodes, 1 shard et 2 replicas.
Our goal is to index 42 millions rows. Indexing time is important. The data
source is an oracle database.
Our indexing strategy is :
· Reading from Oracle to a big CSV file.
· Reading from 4 files (big file chunked) and injection via
ConcurrentUpdateSolrClient
Is it the optimal way of injecting such mass of data into Solr ?
For information, estimated time for our solution is 6h.
No comments:
Post a Comment