flink-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "ASF GitHub Bot (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (FLINK-5048) Kafka Consumer (0.9/0.10) threading model leads problematic cancellation behavior
Date Mon, 14 Nov 2016 06:59:59 GMT

    [ https://issues.apache.org/jira/browse/FLINK-5048?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15662977#comment-15662977

ASF GitHub Bot commented on FLINK-5048:

Github user tzulitai commented on a diff in the pull request:

    --- Diff: flink-streaming-connectors/flink-connector-kafka-base/src/test/java/org/apache/flink/streaming/connectors/kafka/KafkaShortRetentionTestBase.java
    @@ -172,8 +173,14 @@ public void cancel() {
     		// ----------- add consumer dataflow ----------
    +		// the consumer should only poll very small chunks
    +		Properties consumerProps = new Properties();
    +		consumerProps.putAll(standardProps);
    +		consumerProps.putAll(secureProps);
    +		consumerProps.setProperty("fetch.message.max.bytes", "100");
    --- End diff --
    I think we shouldn't be setting `fetch.message.max.bytes` here. The config key for this
setting has changed across Kafka versions (for 0.9+ it's `max.partition.fetch.bytes`). The
version-specific `standardProps` already set values for this config.
    So, the original `props` that only contains `standardProps` and `secureProps` should be
enough for the test to work.

> Kafka Consumer (0.9/0.10) threading model leads problematic cancellation behavior
> ---------------------------------------------------------------------------------
>                 Key: FLINK-5048
>                 URL: https://issues.apache.org/jira/browse/FLINK-5048
>             Project: Flink
>          Issue Type: Bug
>          Components: Kafka Connector
>    Affects Versions: 1.1.3
>            Reporter: Stephan Ewen
>            Assignee: Stephan Ewen
>             Fix For: 1.2.0
> The {{FLinkKafkaConsumer}} (0.9 / 0.10) spawns a separate thread that operates the KafkaConsumer.
That thread is shielded from interrupts, because the Kafka Consumer has not been handling
thread interrupts well.
> Since that thread is also the thread that emits records, it may block in the network
stack (backpressure) or in chained operators. The later case leads to situations where cancellations
get very slow unless that thread would be interrupted (which it cannot be).
> I propose to change the thread model as follows:
>   - A spawned consumer thread pull from the KafkaConsumer and pushes its pulled batch
of records into a blocking queue (size one)
>   - The main thread of the task will pull the record batches from the blocking queue
and emit the records.
> This allows actually for some additional I/O overlay while limiting the additional memory
consumption - only two batches are ever held, one being fetched and one being emitted.

This message was sent by Atlassian JIRA

View raw message