lucene-java-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "sandesh.yapuram" <>
Subject [lucene-6.3.0] hit tragic OutOfMemoryError inside getReader
Date Wed, 12 Jul 2017 12:08:03 GMT
I'm trying to index 5M Documents and currently indexing takes 13 hours. My
data source is SQL db and I've verified out of 13 hours only 30 mins is
spent in fetching data so SQL is not the bottleneck. The complete index size
is 40GB.

My application runs on Tomcat with 2GB JVM space.

Just after the indexing process is complete, when I try to run a simple
query, the application hits OutOfMemoryError: Java Heap Space

The InfoReader log reports 'hit exception during NRT Reader'

Also around the same time frame the InfoReader reports following for

I analyzed the logs and found that the merging process is painfully slow.
I'm using the Default IndexWriterConfig settings.

I've tried indexing same 5m Docs on 3GB JVM with a SerialMergeScheduler with
setRAMBufferSizeMB=48 and setMaxMergedSizeMB=2G in TieredMergePolicy which
takes 14Hours.

Increasing the tomcat jvm space may not always be an option, so is there any
way to tune the Merging process to make it faster and also not run out of

View this message in context:
Sent from the Lucene - Java Users mailing list archive at

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message