kafka-users mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Maung Than <maung_t...@apple.com>
Subject Re: java.net.SocketTimeoutException in Broker
Date Tue, 27 May 2014 18:43:36 GMT
You meant to say the below parameter; if so we do not override it; so it is the default 10
sec. Should It be larger? It is happening only with Gzip; Snappy or no compression works fine.


request.timeout.ms	10000	The amount of time the broker will wait trying to meet the request.required.acks
requirement before sending back an error to the client.

Thanks,
Mauing

On May 27, 2014, at 9:29 AM, Neha Narkhede <neha.narkhede@gmail.com> wrote:

> I think Guozhang meant to say request.timeout.ms, not session timeout. You
> can try increasing the request timeout through the request-timeout-ms
> command line option.
> 
> Thanks,
> Neha
> 
> 
> On Tue, May 27, 2014 at 8:55 AM, Guozhang Wang <wangguoz@gmail.com> wrote:
> 
>> Maung,
>> 
>> This issue may be due to the session timeout value set too small. With
>> batch size 1000 and message size 420 you are sending 420K of data in each
>> request. What is your time out value?
>> 
>> Guozhang
>> 
>> 
>> On Mon, May 26, 2014 at 11:46 PM, Maung Than <maung_than@apple.com> wrote:
>> 
>>> Hi All,
>>> 
>>> This is what we are running on the Broker:
>>> 
>>> /Users/worun/kafkabuild/kafka-0.8.1-src/bin/kafka-producer-perf-test.sh
>>> --broker-list vp21q12ic-hpaj020921:9092 --messages 10000000 --topic
>>> imessage --threads 10 --message-size 420 --batch-size 1000
>>> --compression-codec 1
>>> 
>>> We are getting the below exception for the test, it occurs only with gzip
>>> in an async mode. In the broker log, I saw Connection reset by peer at
>>> sun.nio.ch.FileDispatcher. exception. Any thoughts?
>>> 
>>> [2014-05-26 22:49:33,361] WARN Failed to send producer request with
>>> correlation id 58510 to broker 3 with data for partitions [imessage,1]
>>> (kafka.producer.async.DefaultEventHandler)
>>> java.net.SocketTimeoutException
>>>        at
>>> sun.nio.ch.SocketAdaptor$SocketInputStream.read(SocketAdaptor.java:201)
>>>        at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:86)
>>>        at
>>> 
>> java.nio.channels.Channels$ReadableByteChannelImpl.read(Channels.java:221)
>>>        at kafka.utils.Utils$.read(Utils.scala:375)
>>>        at
>>> 
>> kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54)
>>>        at
>>> kafka.network.Receive$class.readCompletely(Transmission.scala:56)
>>>        at
>>> 
>> kafka.network.BoundedByteBufferReceive.readCompletely(BoundedByteBufferReceive.scala:29)
>>>        at
>> kafka.network.BlockingChannel.receive(BlockingChannel.scala:100)
>>>        at
>> kafka.producer.SyncProducer.liftedTree1$1(SyncProducer.scala:74)
>>>        at
>>> 
>> kafka.producer.SyncProducer.kafka$producer$SyncProducer$$doSend(SyncProducer.scala:71)
>>>        at
>>> 
>> kafka.producer.SyncProducer$$anonfun$send$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(SyncProducer.scala:102)
>>>        at
>>> 
>> kafka.producer.SyncProducer$$anonfun$send$1$$anonfun$apply$mcV$sp$1.apply(SyncProducer.scala:102)
>>>        at
>>> 
>> kafka.producer.SyncProducer$$anonfun$send$1$$anonfun$apply$mcV$sp$1.apply(SyncProducer.scala:102)
>>>        at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33)
>>>        at
>>> 
>> kafka.producer.SyncProducer$$anonfun$send$1.apply$mcV$sp(SyncProducer.scala:101)
>>>        at
>>> kafka.producer.SyncProducer$$anonfun$send$1.apply(SyncProducer.scala:101)
>>>        at
>>> kafka.producer.SyncProducer$$anonfun$send$1.apply(SyncProducer.scala:101)
>>>        at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33)
>>>        at kafka.producer.SyncProducer.send(SyncProducer.scala:100)
>>>        at
>>> 
>> kafka.producer.async.DefaultEventHandler.kafka$producer$async$DefaultEventHandler$$send(DefaultEventHandler.scala:255)
>>>        at
>>> 
>> kafka.producer.async.DefaultEventHandler$$anonfun$dispatchSerializedData$1.apply(DefaultEventHandler.scala:106)
>>>        at
>>> 
>> kafka.producer.async.DefaultEventHandler$$anonfun$dispatchSerializedData$1.apply(DefaultEventHandler.scala:100)
>>>        at
>>> 
>> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:80)
>>>        at
>>> 
>> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:80)
>>>        at scala.collection.Iterator$class.foreach(Iterator.scala:631)
>>>        at
>>> scala.collection.mutable.HashTable$$anon$1.foreach(HashTable.scala:161)
>>>        at
>>> 
>> scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:194)
>>>        at
>> scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:39)
>>>        at scala.collection.mutable.HashMap.foreach(HashMap.scala:80)
>>>        at
>>> 
>> kafka.producer.async.DefaultEventHandler.dispatchSerializedData(DefaultEventHandler.scala:100)
>>>        at
>>> 
>> kafka.producer.async.DefaultEventHandler.handle(DefaultEventHandler.scala:72)
>>>        at
>>> 
>> kafka.producer.async.ProducerSendThread.tryToHandle(ProducerSendThread.scala:104)
>>>        at
>>> 
>> kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:87)
>>>        at
>>> 
>> kafka.producer.async.ProducerSendThread$$anonfun$processEvents$3.apply(ProducerSendThread.scala:67)
>>>        at scala.collection.immutable.Stream.foreach(Stream.scala:254)
>>>        at
>>> 
>> kafka.producer.async.ProducerSendThread.processEvents(ProducerSendThread.scala:66)
>>>        at
>>> kafka.producer.async.ProducerSendThread.run(ProducerSendThread.scala:44)
>>> 2014-05-26 22:49:02:067, 2014-05-26 22:49:48:306, 1, 420, 1000, 400.54,
>>> 8.6625, 1000000, 21626.7653
>>> 
>>> Connection reset by peer at sun.nio.ch.FileDispatcher.read0(Native
>> Method)
>>> at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:21) at
>>> sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:198) at
>>> sun.nio.ch.IOUtil.read(IOUtil.java:171) at
>>> sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:243) at
>>> kafka.utils.Utils$.read(Utils.scala:375) at
>>> 
>> kafka.network.BoundedByteBufferReceive.readFrom(BoundedByteBufferReceive.scala:54)
>>> at kafka.network.Processor.read(SocketServer.scala:347) at
>>> kafka.network.Processor.run(SocketServer.scala:245) at
>>> java.lang.Thread.run(Thread.java:662)
>>> 
>>> Thanks,
>>> Maung
>>> 
>> 
>> 
>> 
>> --
>> -- Guozhang
>> 


Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message