flink-issues mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "ASF GitHub Bot (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (FLINK-8240) Create unified interfaces to configure and instatiate TableSources
Date Thu, 25 Jan 2018 12:36:00 GMT

    [ https://issues.apache.org/jira/browse/FLINK-8240?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16339165#comment-16339165
] 

ASF GitHub Bot commented on FLINK-8240:
---------------------------------------

Github user twalthr commented on a diff in the pull request:

    https://github.com/apache/flink/pull/5240#discussion_r163830189
  
    --- Diff: flink-libraries/flink-table/src/main/scala/org/apache/flink/table/api/BatchTableEnvironment.scala
---
    @@ -107,6 +111,16 @@ abstract class BatchTableEnvironment(
         }
       }
     
    +  /**
    +    * Creates a table from a descriptor that describes the resulting table schema, the
source
    +    * connector, source encoding, and other properties.
    +    *
    +    * @param schema schema descriptor describing the table to create
    +    */
    +  def createTable(schema: Schema): BatchTableSourceDescriptor = {
    --- End diff --
    
    I will improve the definition but in general I think it is nice to have more fluent API
without too much nesting. It is also inconvenient to call a method on table environment and
pass it to its parameters.


> Create unified interfaces to configure and instatiate TableSources
> ------------------------------------------------------------------
>
>                 Key: FLINK-8240
>                 URL: https://issues.apache.org/jira/browse/FLINK-8240
>             Project: Flink
>          Issue Type: New Feature
>          Components: Table API &amp; SQL
>            Reporter: Timo Walther
>            Assignee: Timo Walther
>            Priority: Major
>
> At the moment every table source has different ways for configuration and instantiation.
Some table source are tailored to a specific encoding (e.g., {{KafkaAvroTableSource}}, {{KafkaJsonTableSource}})
or only support one encoding for reading (e.g., {{CsvTableSource}}). Each of them might implement
a builder or support table source converters for external catalogs.
> The table sources should have a unified interface for discovery, defining common properties,
and instantiation. The {{TableSourceConverters}} provide a similar functionality but use an
external catalog. We might generialize this interface.
> In general a table source declaration depends on the following parts:
> {code}
> - Source
>   - Type (e.g. Kafka, Custom)
>   - Properties (e.g. topic, connection info)
> - Encoding
>   - Type (e.g. Avro, JSON, CSV)
>   - Schema (e.g. Avro class, JSON field names/types)
> - Rowtime descriptor/Proctime
>   - Watermark strategy and Watermark properties
>   - Time attribute info
> - Bucketization
> {code}
> This issue needs a design document before implementation. Any discussion is very welcome.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Mime
View raw message