Repository: sqoop
Updated Branches:
refs/heads/trunk e077f9978 -> f19e2a523
SQOOP-2745: Using datetime column as a splitter for Oracle no longer works
(Jarek Jarcec Cecho via Venkat Ranganathan)
Project: http://git-wip-us.apache.org/repos/asf/sqoop/repo
Commit: http://git-wip-us.apache.org/repos/asf/sqoop/commit/9c7638d7
Tree: http://git-wip-us.apache.org/repos/asf/sqoop/tree/9c7638d7
Diff: http://git-wip-us.apache.org/repos/asf/sqoop/diff/9c7638d7
Branch: refs/heads/trunk
Commit: 9c7638d74180cc607ef509b9dd9e6a45ff60c041
Parents: e077f99
Author: Venkat Ranganathan <venkat@hortonworks.com>
Authored: Wed Dec 16 20:47:59 2015 -0800
Committer: Venkat Ranganathan <venkat@hortonworks.com>
Committed: Wed Dec 16 20:47:59 2015 -0800
----------------------------------------------------------------------
.../db/OracleDataDrivenDBInputFormat.java | 4 +-
.../com/cloudera/sqoop/ThirdPartyTests.java | 2 +
.../manager/oracle/OracleSplitterTest.java | 147 +++++++++++++++++++
3 files changed, 151 insertions(+), 2 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/sqoop/blob/9c7638d7/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java
----------------------------------------------------------------------
diff --git a/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java b/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java
index 15c5898..8b5103b 100644
--- a/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java
+++ b/src/java/org/apache/sqoop/mapreduce/db/OracleDataDrivenDBInputFormat.java
@@ -45,7 +45,7 @@ public class OracleDataDrivenDBInputFormat<T extends DBWritable>
* into InputSplits.
*/
@Override
- protected DBSplitter getSplitter(int sqlDataType) {
+ protected DBSplitter getSplitter(int sqlDataType, long splitLimit) {
switch (sqlDataType) {
case Types.DATE:
case Types.TIME:
@@ -53,7 +53,7 @@ public class OracleDataDrivenDBInputFormat<T extends DBWritable>
return new OracleDateSplitter();
default:
- return super.getSplitter(sqlDataType);
+ return super.getSplitter(sqlDataType, splitLimit);
}
}
http://git-wip-us.apache.org/repos/asf/sqoop/blob/9c7638d7/src/test/com/cloudera/sqoop/ThirdPartyTests.java
----------------------------------------------------------------------
diff --git a/src/test/com/cloudera/sqoop/ThirdPartyTests.java b/src/test/com/cloudera/sqoop/ThirdPartyTests.java
index 06b4ae6..50f3192 100644
--- a/src/test/com/cloudera/sqoop/ThirdPartyTests.java
+++ b/src/test/com/cloudera/sqoop/ThirdPartyTests.java
@@ -57,6 +57,7 @@ import org.apache.sqoop.manager.netezza.NetezzaExportManualTest;
import org.apache.sqoop.manager.netezza.NetezzaImportManualTest;
import org.apache.sqoop.manager.oracle.OracleCallExportTest;
import org.apache.sqoop.manager.oracle.OracleIncrementalImportTest;
+import org.apache.sqoop.manager.oracle.OracleSplitterTest;
import org.apache.sqoop.manager.sqlserver.SQLServerDatatypeExportDelimitedFileManualTest;
import org.apache.sqoop.manager.sqlserver.SQLServerDatatypeExportSequenceFileManualTest;
import org.apache.sqoop.manager.sqlserver.SQLServerDatatypeImportDelimitedFileManualTest;
@@ -96,6 +97,7 @@ public final class ThirdPartyTests extends TestCase {
suite.addTestSuite(OracleManagerTest.class);
suite.addTestSuite(OracleCompatTest.class);
suite.addTestSuite(OracleIncrementalImportTest.class);
+ suite.addTestSuite(OracleSplitterTest.class);
// SQL Server
suite.addTestSuite(SQLServerDatatypeExportDelimitedFileManualTest.class);
http://git-wip-us.apache.org/repos/asf/sqoop/blob/9c7638d7/src/test/org/apache/sqoop/manager/oracle/OracleSplitterTest.java
----------------------------------------------------------------------
diff --git a/src/test/org/apache/sqoop/manager/oracle/OracleSplitterTest.java b/src/test/org/apache/sqoop/manager/oracle/OracleSplitterTest.java
new file mode 100644
index 0000000..fc5f56b
--- /dev/null
+++ b/src/test/org/apache/sqoop/manager/oracle/OracleSplitterTest.java
@@ -0,0 +1,147 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.sqoop.manager.oracle;
+
+import com.cloudera.sqoop.SqoopOptions;
+import com.cloudera.sqoop.manager.OracleUtils;
+import com.cloudera.sqoop.testutil.CommonArgs;
+import com.cloudera.sqoop.testutil.ImportJobTestCase;
+import org.apache.commons.logging.Log;
+import org.apache.commons.logging.LogFactory;
+import org.apache.commons.io.FileUtils;
+import org.apache.hadoop.conf.Configuration;
+import org.apache.hadoop.fs.FSDataInputStream;
+import org.apache.hadoop.fs.FileSystem;
+import org.apache.hadoop.fs.Path;
+import org.apache.hadoop.io.IOUtils;
+
+import java.io.BufferedReader;
+import java.io.BufferedWriter;
+import java.io.File;
+import java.io.FileInputStream;
+import java.io.FileWriter;
+import java.io.IOException;
+import java.io.InputStreamReader;
+import java.io.Writer;
+import java.sql.SQLException;
+import java.util.ArrayList;
+import java.util.List;
+
+/**
+ * Test various custom splitters for Oracle.
+ */
+public class OracleSplitterTest extends ImportJobTestCase {
+
+ public static final Log LOG = LogFactory.getLog(
+ OracleSplitterTest.class.getName());
+
+ @Override
+ protected boolean useHsqldbTestServer() {
+ return false;
+ }
+
+ @Override
+ protected String getConnectString() {
+ return OracleUtils.CONNECT_STRING;
+ }
+
+ @Override
+ protected SqoopOptions getSqoopOptions(Configuration conf) {
+ SqoopOptions opts = new SqoopOptions(conf);
+ OracleUtils.setOracleAuth(opts);
+ return opts;
+ }
+
+ @Override
+ protected void dropTableIfExists(String table) throws SQLException {
+ OracleUtils.dropTable(table, getManager());
+ }
+
+ /** the names of the tables we're creating. */
+ private List<String> tableNames;
+
+ @Override
+ public void tearDown() {
+ // Clean up the database on our way out.
+ for (String tableName : tableNames) {
+ try {
+ dropTableIfExists(tableName);
+ } catch (SQLException e) {
+ LOG.warn("Error trying to drop table '" + tableName
+ + "' on tearDown: " + e);
+ }
+ }
+ super.tearDown();
+ }
+
+ protected String [] getArgv(String tableName, String connPropsFileName, String splitByColumn)
{
+ ArrayList<String> args = new ArrayList<String>();
+
+ CommonArgs.addHadoopFlags(args);
+
+ args.add("--connect");
+ args.add(getConnectString());
+ args.add("--target-dir");
+ args.add(getWarehouseDir());
+ args.add("--num-mappers");
+ args.add("2");
+ args.add("--split-by");
+ args.add(splitByColumn);
+ args.add("--table");
+ args.add(tableName);
+ args.add("--connection-param-file");
+ args.add(connPropsFileName);
+
+ return args.toArray(new String[0]);
+ }
+
+ public void testTimestampSplitter() throws IOException {
+ tableNames = new ArrayList<String>();
+ String [] types = { "INT", "VARCHAR(10)", "TIMESTAMP", };
+ String [] vals = {
+ "1", "'old_data'", "TO_TIMESTAMP('1999-01-01 11:11:11', 'YYYY-MM-DD HH24:MI:SS')",
+ "2", "'new_data'", "TO_TIMESTAMP('2000-11-11 23:23:23', 'YYYY-MM-DD HH24:MI:SS')",
+ };
+ String tableName = getTableName();
+ tableNames.add(tableName);
+ createTableWithColTypes(types, vals);
+ // Some version of Oracle's jdbc drivers automatically convert date to
+ // timestamp. Since we don't want this to happen for this test,
+ // we must explicitly use a property file to control this behavior.
+ String connPropsFileName = "connection.properties";
+ FileUtils.writeStringToFile(new File(connPropsFileName), "oracle.jdbc.mapDateToTimestamp=false");
+ String[] args = getArgv(tableName, connPropsFileName, getColName(2));
+ runImport(args);
+
+ File file;
+ List<String> lines;
+
+ // First row should be in the first file
+ file = new File(this.getWarehouseDir(), "part-m-00000");
+ lines = FileUtils.readLines(file, "UTF-8");
+ assertEquals(1, lines.size());
+ assertEquals("1,old_data,1999-01-01 11:11:11.0", lines.get(0));
+
+ // With second line in the second file
+ file = new File(this.getWarehouseDir(), "part-m-00001");
+ lines = FileUtils.readLines(file, "UTF-8");
+ assertEquals(1, lines.size());
+ assertEquals("2,new_data,2000-11-11 23:23:23.0", lines.get(0));
+ }
+}
|