sqoop-dev mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From "Richard (JIRA)" <j...@apache.org>
Subject [jira] [Commented] (SQOOP-1393) Import data from database to Hive as Parquet files
Date Mon, 11 Aug 2014 08:01:27 GMT

    [ https://issues.apache.org/jira/browse/SQOOP-1393?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14092544#comment-14092544
] 

Richard commented on SQOOP-1393:
--------------------------------

There are 2 solutions:
1. Use Kite to generate hive table, which is similar with function of import into hdfs as
Parquet file (SQOOP-1390). Set the Kite dataset uri like "dataset:hive?dataset=test&hdfs:host=server-360&hdfs:port:8020"
only to generate hive table directly.
2. use Sqoop to generate hive table, which is similar with function of import into hive as
text file. There are 3 steps. First, import into hdfs as Parquet file (SQOOP-1390). Second,
use Sqoop to create table in hive, like text file. Third, move Parquet file to hive warehouse
(/user/hive/warehouse).


> Import data from database to Hive as Parquet files
> --------------------------------------------------
>
>                 Key: SQOOP-1393
>                 URL: https://issues.apache.org/jira/browse/SQOOP-1393
>             Project: Sqoop
>          Issue Type: Sub-task
>          Components: tools
>            Reporter: Qian Xu
>            Assignee: Richard
>
> Import data to Hive as Parquet file can be separated into two steps:
> 1. Import an individual table from an RDBMS to HDFS as a set of Parquet files.
> 2. Import the data into Hive by generating and executing a CREATE TABLE statement to
define the data's layout in Hive with Parquet format table



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Mime
View raw message