phoenix-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Radha Krishna G <>
Subject Invoking org.apache.phoenix.mapreduce.CsvBulkLoadTool from phoenix- is not working properly
Date Wed, 03 Aug 2016 07:53:49 GMT

Hi All,i am trying to load around 40 GB file using "org.apache.phoenix.mapreduce.CsvBulkLoadTool"
but it is showing the below error message.
INFO mapreduce.Job: Task Id : attempt_1469663368297_56967_m_000042_0, Status : FAILEDError:
java.lang.RuntimeException: java.lang.RuntimeException: (startline 1)
EOF reached before encapsulated token finished        at 
      at        at org.apache.hadoop.mapred.MapTask.runNewMapper( 
      at        at org.apache.hadoop.mapred.YarnChild$ 
      at Method)        at 
      at org.apache.hadoop.mapred.YarnChild.main( by: java.lang.RuntimeException: (startline 1) EOF reached before encapsulated token finished     
  at org.apache.commons.csv.CSVParser$1.getNextRecord(        at org.apache.commons.csv.CSVParser$1.hasNext( 
      at        at        at org.apache.phoenix.mapreduce.CsvToKeyValueMapper$CsvLineParser.parse( 
      ... 9 moreCaused by: (startline 1) EOF reached before encapsulated
token finished        at org.apache.commons.csv.Lexer.parseEncapsulatedToken( 
      at org.apache.commons.csv.Lexer.nextToken(        at org.apache.commons.csv.CSVParser.nextRecord( 
      at org.apache.commons.csv.CSVParser$1.getNextRecord(       
... 14 more

Note : I collected some sample records around(1000) form the same file and able to load using
the same approach, but if i provide full file path its failing, can any one suggest what is
solution for the above issue..
Bellow Command i used==================
hadoop jar phoenix- org.apache.phoenix.mapreduce.CsvBulkLoadTool
--table "Table_Name" --input "HDFS input file path" -d $'\034'

-d $'\034' --> the field separator in the file is FS so we provided the explicitly  

RegardsRadha krishna G
View raw message