kafka-users mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Nicolas Phung <nicolas.ph...@gmail.com>
Subject Re: KafkaProducer Retries in .9.0.1
Date Thu, 21 Apr 2016 05:41:16 GMT
Hi Ismail,

Thanks for you reply.

For me, It's happening when I'm doing various breakdown (shutting down
instances / zookeeper) on Kafka brokers on 0.9.0.1 that should simulate a
leader is not available case. The same kind of breakdown on 0.8.2.2
client/broker can retry as expected.

>From my understanding, if the broker (leader) is unavailable, it should
buffer the message in the producer till the broker is available.

Regards,
Nicolas

On Thu, Apr 21, 2016 at 7:33 AM, Ismael Juma <ismael@juma.me.uk> wrote:

> Hi,
>
> This was explained earlier, I think. Retries are only attempted for
> retriable errors. If a message is too large, retrying won't help (it will
> still be too large). However, if a leader is not available, then a retry
> will happen as the leader may be available then.
>
> Ismael
>
> On Wed, Apr 20, 2016 at 1:00 PM, Nicolas Phung <nicolas.phung@gmail.com>
> wrote:
>
> > Hello,
> >
> > Have you solved this ? I'm encountering the same issue with the new
> > Producer on 0.9.0.1 client with a 0.9.0.1 Kafka broker. We tried the same
> > various breakdown (kafka(s), zookeeper) with 0.8.2.2 client and Kafka
> > broker 0.8.2.2 and the retries work as expected on the older version. I'm
> > going to take a look if someone else has filed a related issue about it.
> >
> > Regards,
> > Nicolas PHUNG
> >
> > On Thu, Apr 7, 2016 at 5:15 AM, christopher palm <cpalm3@gmail.com>
> wrote:
> >
> > > Hi Thanks for the suggestion.
> > > I lowered the broker message.max.bytes to be smaller than my payload so
> > > that I now receive an
> > > org.apache.kafka.common.errors.RecordTooLargeException
> > > :
> > >
> > > I still don't see the retries happening, the default back off is 100ms,
> > and
> > > my producer loops for a few seconds, long enough to trigger the retry.
> > >
> > > Is there something else I need to set?
> > >
> > > I have tried this with a sync and async producer both with same results
> > >
> > > Thanks,
> > >
> > > Chris
> > >
> > > On Wed, Apr 6, 2016 at 12:01 AM, Manikumar Reddy <
> > > manikumar.reddy@gmail.com>
> > > wrote:
> > >
> > > > Hi,
> > > >
> > > >  Producer message size validation checks ("buffer.memory",
> > > > "max.request.size" )  happens before
> > > >  batching and sending messages.  Retry mechanism is applicable for
> > broker
> > > > side errors and network errors.
> > > > Try changing "message.max.bytes" broker config property for
> simulating
> > > > broker side error.
> > > >
> > > >
> > > >
> > > >
> > > >
> > > >
> > > > On Wed, Apr 6, 2016 at 9:53 AM, christopher palm <cpalm3@gmail.com>
> > > wrote:
> > > >
> > > > > Hi All,
> > > > >
> > > > > I am working with the KafkaProducer using the properties below,
> > > > > so that the producer keeps trying to send upon failure on Kafka
> > .9.0.1.
> > > > > I am forcing a failure by setting my buffersize smaller than my
> > > > > payload,which causes the expected exception below.
> > > > >
> > > > > I don't see the producer retry to send on receiving this failure.
> > > > >
> > > > > Am I  missing something in the configuration to allow the producer
> to
> > > > retry
> > > > > on failed sends?
> > > > >
> > > > > Thanks,
> > > > > Chris
> > > > >
> > > > > .java.util.concurrent.ExecutionException:
> > > > > org.apache.kafka.common.errors.RecordTooLargeException: The message
> > is
> > > > 8027
> > > > > bytes when serialized which is larger than the total memory buffer
> > you
> > > > have
> > > > > configured with the buffer.memory configuration.
> > > > >
> > > > >  props.put("bootstrap.servers", bootStrapServers);
> > > > >
> > > > > props.put("acks", "all");
> > > > >
> > > > > props.put("retries", 3);//Try for 3 strikes
> > > > >
> > > > > props.put("batch.size", batchSize);//Need to see if this number
> > should
> > > > > increase under load
> > > > >
> > > > > props.put("linger.ms", 1);//After 1 ms fire the batch even if the
> > > batch
> > > > > isn't full.
> > > > >
> > > > > props.put("buffer.memory", buffMemorySize);
> > > > >
> > > > > props.put("max.block.ms",500);
> > > > >
> > > > > props.put("max.in.flight.requests.per.connection", 1);
> > > > >
> > > > > props.put("key.serializer",
> > > > > "org.apache.kafka.common.serialization.StringSerializer");
> > > > >
> > > > > props.put("value.serializer",
> > > > > "org.apache.kafka.common.serialization.ByteArraySerializer");
> > > > >
> > > >
> > >
> >
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message