lucene-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Mark Miller (JIRA)" <>
Subject [jira] [Commented] (SOLR-6056) Zookeeper crash JVM stack OOM because of recover strategy
Date Tue, 13 May 2014 04:26:14 GMT


Mark Miller commented on SOLR-6056:

+1, good stuff. 

Regarding 2, we have to think carefully before making this change. 

What about the sequence where a replica is recovering, and an update to be buffered randomly
fails - what ensures the replica fully recovers?

> Zookeeper crash JVM stack OOM because of recover strategy 
> ----------------------------------------------------------
>                 Key: SOLR-6056
>                 URL:
>             Project: Solr
>          Issue Type: Bug
>          Components: SolrCloud
>    Affects Versions: 4.6
>         Environment: Two linux servers, 65G memory, 16 core cpu
> 20 collections, every collection has one shard two replica 
> one zookeeper
>            Reporter: Raintung Li
>            Assignee: Shalin Shekhar Mangar
>            Priority: Critical
>              Labels: cluster, crash, recover
>         Attachments: patch-6056.txt
> Some errors"org.apache.solr.common.SolrException: Error opening new searcher. exceeded
limit of maxWarmingSearchers=2, try again later", that occur distributedupdateprocessor trig
the core admin recover process.
> That means every update request will send the core admin recover request.
> (see the code doFinish())
> The terrible thing is CoreAdminHandler will start a new thread to publish the recover
status and start recovery. Threads increase very quickly, and stack OOM , Overseer can't handle
a lot of status update , zookeeper node for  /overseer/queue/qn-0000125553 increase more than
40 thousand in two minutes.
> At the last zookeeper crash. 
> The worse thing is queue has too much nodes in the zookeeper, the cluster can't publish
the right status because only one overseer work, I have to start three threads to clear the
queue nodes. The cluster doesn't work normal near 30 minutes...

This message was sent by Atlassian JIRA

To unsubscribe, e-mail:
For additional commands, e-mail:

View raw message