sqoop-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From Matthieu Labour <matth...@actionx.com>
Subject Re: Need help and tips for tthe following issue: No data get exported from hadoop to mysql using sqoop.
Date Wed, 17 Oct 2012 22:01:25 GMT
Thank you for the great library !

On Wed, Oct 17, 2012 at 5:58 PM, Jarek Jarcec Cecho <jarcec@apache.org>wrote:

> I'm glad that you've managed to resolve the issue. Happy sqooping!
>
> Jarcec
>
> On Wed, Oct 17, 2012 at 05:53:24PM -0400, Matthieu Labour wrote:
> > Jarceq
> > My bad. My mistake. Amazon attaches the slave and the master to 2
> different
> > security groups and both had to be added to the Amazon RDS
> > It is working now
> > Thank you for your help
> >
> > On Mon, Oct 15, 2012 at 5:40 PM, Jarek Jarcec Cecho <jarcec@apache.org
> >wrote:
> >
> > > Hi Matt,
> > > thanks for getting back to me with actual task log. I'm adding Sqoop
> user
> > > mailing list back in loop in so that others might jump in. I've however
> > > removed entire log to prevent disclosure of any sensitive data.
> > >
> > > The log contained only original Connection exception with no further
> > > wrapped exceptions that would help finding out the cause. I would
> recommend
> > > to do following:
> > >
> > > Connect to each of your slave nodes (ssh) and try to connect to your
> MySQL
> > > box, e.g. something like
> > >
> > > mysql -h mysql.server -u myuser -pmypassword database
> > >
> > > It should be working from each node. If this won't work (and I'm
> expecting
> > > that it won't), then there might be some firewall issues or other
> > > networking problems that you will have to solve.
> > >
> > > Jarcec
> > >
> > > On Mon, Oct 15, 2012 at 12:34:08PM -0400, Matthieu Labour wrote:
> > > > Hi Jarcec
> > > > Please find emclosed the screenshot using the hadoop web interfaces:
> > > >
> > >
> http://docs.amazonwebservices.com/ElasticMapReduce/latest/DeveloperGuide/UsingtheHadoopUserInterface.html
> > > > I am sending an email directly to you as the log might contain some
> info
> > > > that I would rather not have on the web/ mailing list
> > > > The trace is the same as the one I can see when I ssh the master
> node and
> > > > explore the logs under /mnt/var/log/hadoop/steps/ ...
> > > > If you tell me the best place to add some logs in sqoop, then i can
> > > > recompile and rerun
> > > > The bizarre thing is that the select seems to work.
> > > > Cheers
> > > > Matthieu
> > > >
> > >
> > > === SENSITIVE CONTENT REMOVED ===
> > > >
> > > >
> > > > On Thu, Oct 11, 2012 at 11:38 AM, Jarek Jarcec Cecho <
> jarcec@apache.org
> > > >wrote:
> > > >
> > > > > Hi sir,
> > > > > I'm sorry but it's hard to help without the actual task log that
> should
> > > > > contain more details about the exception. I was able to dig
> following
> > > > > Amazon documentation that deals with getting Hadoop Web UI. Would
> you
> > > mind
> > > > > trying it and see if you can reach map task log?
> > > > >
> > > > >
> > > > >
> > >
> http://docs.amazonwebservices.com/ElasticMapReduce/latest/DeveloperGuide/UsingtheHadoopUserInterface.html
> > > > >
> > > > > Jarcec
> > > > >
> > > > > On Thu, Oct 11, 2012 at 10:39:38AM -0400, Matthieu Labour wrote:
> > > > > > Jarceq
> > > > > > Thank you for your reply
> > > > > > I have a hard time to believe that this is a jdbc connection
> issue
> > > > > because
> > > > > > when i execute the sqoop export command, it succesfully executes
> > > > >  Executing
> > > > > > SQL statement: SELECT t.* FROM `ml_ys_log_gmt_test` AS t LIMIT 1
> and
> > > if i
> > > > > > cange the password in the sqoop export command then I
> > > > > > get java.sql.SQLException: Access denied for user
> > > > > > So sqoop export seems to be able to reach the Sql machine with
> that
> > > > > > username and password
> > > > > > I will use the postgresql for now as it works for me!
> > > > > > Thank you for your help
> > > > > >
> > > > > >
> > > > > > On Wed, Oct 10, 2012 at 7:58 PM, Jarek Jarcec Cecho <
> > > jarcec@apache.org
> > > > > >wrote:
> > > > > >
> > > > > > > Hi sir,
> > > > > > > I have actually zero experience with amazon services, so I'm
> afraid
> > > > > that I
> > > > > > > can't much help you navigate to the map tasks logs. Usually on
> > > normal
> > > > > > > hadoop cluster, there is service call "Job Tracker" that is
> > > serving as
> > > > > > > central place for mapreduce jobs. I'm expecting that you
> should be
> > > > > able to
> > > > > > > find this webservice or something similar somehow somewhere.
> You
> > > > > should see
> > > > > > > job executed by hadoop there and you also should be able to
> get to
> > > > > > > individual task logs.
> > > > > > >
> > > > > > > Following my previous blind shoot - How is defined MySQL user
> that
> > > > > you're
> > > > > > > using for Sqoop? I'm very interested to know the host part of
> the
> > > > > user. For
> > > > > > > example usually there are users like root@localhost or jarcec@
> '%'.
> > > If
> > > > > > > your host part (in my examples it's localhost or '%') is
> > > restrictive
> > > > > enough
> > > > > > > your hadoop nodes might not be capable of connecting to that
> MySQL
> > > box
> > > > > and
> > > > > > > thus resulting in connection failures.
> > > > > > >
> > > > > > > Jarcec
> > > > > > >
> > > > > > > On Wed, Oct 10, 2012 at 05:22:14PM -0400, Matthieu Labour
> wrote:
> > > > > > > > Hi Jarcek
> > > > > > > > If i use the postgresql jdbc connector and connect to one of
> our
> > > > > heroku
> > > > > > > > machine then scoop works
> > > > > > > > ~/$SQOOP_ROOT/bin/sqoop export --connect
> > > > > > > > jdbc:postgresql://ec2-XX-XX-XXX-XX.compute-1.amazonaws.com:
> > > database
> > > > > > > > --username username --password password --table
> > > ml_ys_log_gmt_test
> > > > > > > > --export-dir -export-dir
> > > > > > > >
> =hdfs:///mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01
> > > > > > > > --input-fields-terminated-by='\t'
> > > > > > > > --lines-terminated-by='\n' --verbose --batch
> > > > > > > >
> > > > > > > > On Wed, Oct 10, 2012 at 2:06 PM, Matthieu Labour <
> > > > > matthieu@actionx.com
> > > > > > > >wrote:
> > > > > > > >
> > > > > > > > >
> > > > > > > > > Jarcek
> > > > > > > > >
> > > > > > > > > I am quite new to hadoop and amazon EMR. Where are those
> files
> > > > > located?
> > > > > > > > >
> > > > > > > > > Here is what I am doing:
> > > > > > > > >
> > > > > > > > > 1) I am using amazon elastic map reduce and I have created
> a
> > > New
> > > > > Job
> > > > > > > that
> > > > > > > > > does not terminate and whose type is HBase
> > > > > > > > >
> > > > > > > > > 2) I get the job id
> > > > > > > > > myaccount@ubuntu:~/elastic-mapreduce-cli$
> ./elastic-mapreduce
> > > > > --list
> > > > > > > > > --active
> > > > > > > > > j-3EFP15LBJC8R4     RUNNING
> > > > > > > > > ec2-XXX-XX-XXX-XX.compute-1.amazonaws.com         sqooping
> > > > > > > > >    COMPLETED      Setup Hadoop Debugging
> > > > > > > > >    COMPLETED      Start HBase
> > > > > > > > >    COMPLETED      Setup Hive
> > > > > > > > >    RUNNING        Setup Pig
> > > > > > > > >
> > > > > > > > > 3) I attach and run a step:
> > > > > > > > > ./elastic-mapreduce -j j-3EFP15LBJC8R4 --jar
> > > > > > > > > s3://elasticmapreduce/libs/script-runner/script-runner.jar
> > > --arg
> > > > > > > > > s3://mybucket/sqoop/sqoop.sh
> > > > > > > > >
> > > > > > > > > 4) I ssh the machine. ssh -i ~/.ec2/MYKEY.pem
> > > > > > > > > hadoop@ec2-XXX-XX-XXX-XX.compute-1.amazonaws.com
> > > > > > > > >
> > > > > > > > > 5) tail -f /mnt/var/lib/hadoop/steps/6/stderr shows the
> > > mapreduce
> > > > > job
> > > > > > > > > hanging
> > > > > > > > > 12/10/10 17:46:58 DEBUG mapreduce.ExportInputFormat:
> Generated
> > > > > splits:
> > > > > > > > > 12/10/10 17:46:58 DEBUG mapreduce.ExportInputFormat:
> > > > > > > > >
> > > > > > >
> > > > >
> > >
> Paths:/mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01/part-m-00000:0+52
> > > > > > > > > Locations:ip-10-77-70-192.ec2.internal:;
> > > > > > > > > 12/10/10 17:46:58 INFO mapred.JobClient: Running job:
> > > > > > > job_201210101503_0024
> > > > > > > > > 12/10/10 17:46:59 INFO mapred.JobClient:  map 0% reduce 0%
> > > > > > > > >
> > > > > > > > > 6) In /mnt/var/lib/hadoop/steps/6 there is the scoop.sh
> script
> > > file
> > > > > > > with
> > > > > > > > > ~/sqoop-1.4.2.bin__hadoop-1.0.0/bin/sqoop export --connect
> > > > > > > > > jdbc:mysql://hostname:3306/analyticsdb --username username
> > > > > --password
> > > > > > > > > password --table ml_ys_log_gmt_test --export-dir
> > > > > > > > >
> > > =hdfs:///mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01
> > > > > > > > > --input-fields-termi
> > > > > > > > > nated-by='\t' --lines-terminated-by='\n' --verbose --batch
> > > > > > > > >
> > > > > > > > > On that same machine, same location (
> > > > > /mnt/var/lib/hadoop/steps/6), the
> > > > > > > > > following command works
> > > > > > > > > mysql -h hostname -P 3306 -u username -p
> > > > > > > > > password: password
> > > > > > > > > Afterwards I can use the database, describe the table etc
> ....
> > > > > > > > > Please note the mysql machine is running on Amazon RDS and
> I
> > > have
> > > > > > > > > added ElasticMapReduce-master security group to RDS
> > > > > > > > >
> > > > > > > > > Thank you for your help
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > On Wed, Oct 10, 2012 at 1:27 PM, Jarek Jarcec Cecho <
> > > > > jarcec@apache.org
> > > > > > > >wrote:
> > > > > > > > >
> > > > > > > > >> It would be very helpful if you could send us task log
> from
> > > one
> > > > > map
> > > > > > > job
> > > > > > > > >> that Sqoop executes.
> > > > > > > > >>
> > > > > > > > >> Blindly shooting - Sqoop is connecting to your database
> from
> > > map
> > > > > > > tasks.
> > > > > > > > >> Based on the connection issues - are you sure that you can
> > > > > connect to
> > > > > > > your
> > > > > > > > >> database from all nodes in your cluster?
> > > > > > > > >>
> > > > > > > > >> Jarcec
> > > > > > > > >>
> > > > > > > > >> On Wed, Oct 10, 2012 at 01:16:03PM -0400, Matthieu Labour
> > > wrote:
> > > > > > > > >> > Hi Jerek
> > > > > > > > >> >
> > > > > > > > >> > Thank you so much for your help.
> > > > > > > > >> >
> > > > > > > > >> > Following your advice, I run the following command:
> > > > > > > > >> > ~/sqoop-1.4.2.bin__hadoop-1.0.0/bin/sqoop export
> --connect
> > > > > > > > >> > jdbc:mysql://hostname:3306/analyticsdb --username
> username
> > > > > > > --password
> > > > > > > > >> > password --table ml_ys_log_gmt_test --export-dir
> > > > > > > > >> >
> > > hdfs:///mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01
> > > > > > > > >> > --input-fields-terminated-by='\t'
> --lines-terminated-by='\n'
> > > > > > > --verbose
> > > > > > > > >> >
> > > > > > > > >> > It seems to find the file to export. So that is good.
> In the
> > > > > log I
> > > > > > > see
> > > > > > > > >> the
> > > > > > > > >> > following: (I am not sure why :0+52 gets appended)
> > > > > > > > >> > 2/10/10 16:43:41 DEBUG mapreduce.ExportInputFormat:
> > > > > > > > >> >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> Paths:/mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01/part-m-00000:0+52
> > > > > > > > >> > Locations:ip-XX-XX-XX-XXX.ec2.internal:;
> > > > > > > > >> >
> > > > > > > > >> > However it hangs forever after it printed the following:
> > > > > > > > >> > 12/10/10 16:43:42 INFO mapred.JobClient:  map 0% reduce
> 0%
> > > > > > > > >> >
> > > > > > > > >> > Then It seems the JDBC connection is eventually timing
> out.
> > > > > > > > >> > 12/10/10 16:47:07 INFO mapred.JobClient: Task Id :
> > > > > > > > >> > attempt_201210101503_0019_m_000000_0, Status : FAILED
> > > > > > > > >> >
> > > > > > > > >> > Here is the log towards the end:
> > > > > > > > >> >
> > > > > > > > >> > 12/10/10 16:43:40 INFO mapred.JobClient: Default number
> of
> > > map
> > > > > > > tasks: 4
> > > > > > > > >> > 12/10/10 16:43:40 INFO mapred.JobClient: Default number
> of
> > > > > reduce
> > > > > > > > >> tasks: 0
> > > > > > > > >> > 12/10/10 16:43:41 INFO mapred.JobClient: Setting group
> to
> > > hadoop
> > > > > > > > >> > 12/10/10 16:43:41 INFO input.FileInputFormat: Total
> input
> > > paths
> > > > > to
> > > > > > > > >> process
> > > > > > > > >> > : 1
> > > > > > > > >> > 12/10/10 16:43:41 DEBUG mapreduce.ExportInputFormat:
> Target
> > > > > > > > >> numMapTasks=4
> > > > > > > > >> > 12/10/10 16:43:41 DEBUG mapreduce.ExportInputFormat:
> Total
> > > input
> > > > > > > > >> bytes=52
> > > > > > > > >> > 12/10/10 16:43:41 DEBUG mapreduce.ExportInputFormat:
> > > > > maxSplitSize=13
> > > > > > > > >> > 12/10/10 16:43:41 INFO input.FileInputFormat: Total
> input
> > > paths
> > > > > to
> > > > > > > > >> process
> > > > > > > > >> > : 1
> > > > > > > > >> > 12/10/10 16:43:41 DEBUG mapreduce.ExportInputFormat:
> > > Generated
> > > > > > > splits:
> > > > > > > > >> > 12/10/10 16:43:41 DEBUG mapreduce.ExportInputFormat:
> > > > > > > > >> >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> Paths:/mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01/part-m-00000:0+52
> > > > > > > > >> > Locations:ip-XX-XX-XX-XXX.ec2.internal:;
> > > > > > > > >> > 12/10/10 16:43:41 INFO mapred.JobClient: Running job:
> > > > > > > > >> job_201210101503_0019
> > > > > > > > >> > 12/10/10 16:43:42 INFO mapred.JobClient:  map 0% reduce
> 0%
> > > > > > > > >> > 12/10/10 16:47:07 INFO mapred.JobClient: Task Id :
> > > > > > > > >> > attempt_201210101503_0019_m_000000_0, Status : FAILED
> > > > > > > > >> > java.io.IOException:
> > > > > > > > >> > com.mysql.jdbc.exceptions.jdbc4.CommunicationsException:
> > > > > > > Communications
> > > > > > > > >> > link failure
> > > > > > > > >> >
> > > > > > > > >> > The last packet sent successfully to the server was 0
> > > > > milliseconds
> > > > > > > ago.
> > > > > > > > >> The
> > > > > > > > >> > driver has not received any packets from the server.
> > > > > > > > >> >         at
> > > > > > > > >> >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> org.apache.sqoop.mapreduce.ExportOutputFormat.getRecordWriter(ExportOutputFormat.java:79)
> > > > > > > > >> >         at
> > > > > > > > >> >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> org.apache.hadoop.mapred.MapTask$NewDirectOutputCollector.<init>(MapTask.java:635)
> > > > > > > > >> >         at
> > > > > > > > >>
> > > org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:760)
> > > > > > > > >> >         at
> > > > > org.apache.hadoop.mapred.MapTask.run(MapTask.java:375)
> > > > > > > > >> >         at
> > > org.apache.hadoop.mapred.Child$4.run(Child.java:255)
> > > > > > > > >> >         at
> > > java.security.AccessController.doPrivileged(Native
> > > > > > > Method)
> > > > > > > > >> >         at
> > > javax.security.auth.Subject.doAs(Subject.java:396)
> > > > > > > > >> >         at
> > > > > > > > >> >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1132)
> > > > > > > > >> >         at
> > > org.apache.hadoop.mapred.Child.main(Child.java:249)
> > > > > > > > >> > Caused by:
> > > > > com.mysql.jdbc.exceptions.jdbc4.CommunicationsException:
> > > > > > > > >> > Communications link failure
> > > > > > > > >> >
> > > > > > > > >> >
> > > > > > > > >> >
> > > > > > > > >> >
> > > > > > > > >> > On Wed, Oct 10, 2012 at 12:40 PM, Jarek Jarcec Cecho <
> > > > > > > jarcec@apache.org
> > > > > > > > >> >wrote:
> > > > > > > > >> >
> > > > > > > > >> > > Hi sir,
> > > > > > > > >> > > as far as I remember FileInputFormat is not doing
> > > recursive
> > > > > > > descent
> > > > > > > > >> into
> > > > > > > > >> > > subdirectories when looking for input files. Would you
> > > mind
> > > > > > > trying to
> > > > > > > > >> > > export directory
> > > > > > > > >> /mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01
> > > > > > > > >> > > to see if it will help? Something like
> > > > > > > > >> > >
> > > > > > > > >> > > sqoop export ... --export-dir
> > > > > > > > >> > >
> /mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01
> > > ...
> > > > > > > > >> > >
> > > > > > > > >> > > Jarcec
> > > > > > > > >> > >
> > > > > > > > >> > > On Wed, Oct 10, 2012 at 12:30:56PM -0400, Matthieu
> Labour
> > > > > wrote:
> > > > > > > > >> > > > Hi
> > > > > > > > >> > > >
> > > > > > > > >> > > > I want to do the following: Export data stored in
> > > hadoop to
> > > > > > > MySql.
> > > > > > > > >> It is
> > > > > > > > >> > > > not working and I have been pulling my hair. I was
> > > hoping to
> > > > > > > get a
> > > > > > > > >> bit of
> > > > > > > > >> > > > help. Thank you in advance
> > > > > > > > >> > > >
> > > > > > > > >> > > > The command is the following:
> > > > > > > > >> > > >
> > > > > > > > >> > > > ~/sqoop-1.4.2.bin__hadoop-1.0.0/bin/sqoop export
> > > --connect
> > > > > > > > >> > > > jdbc:mysql://hostname:3306/analyticsdb --username
> > > username
> > > > > > > > >> --password
> > > > > > > > >> > > > password --table ml_ys_log_gmt_test --export-dir
> > > > > > > > >> > > > hdfs:///mnt/var/lib/hadoop/dfs/logs_sanitized_test
> > > > > > > > >> > > > --input-fields-terminated-by='\t'
> > > > >  --lines-terminated-by='\n'
> > > > > > > > >> --verbose
> > > > > > > > >> > > >
> > > > > > > > >> > > > On my mysqlserver in the database analyticsdb, I do
> > > have the
> > > > > > > > >> following
> > > > > > > > >> > > > table ml_ys_log_gmt_test
> > > > > > > > >> > > >
> > > > > > > > >> > > > mysql> describe ml_ys_log_gmt_test;
> > > > > > > > >> > > >
> +--------+-------------+------+-----+---------+-------+
> > > > > > > > >> > > > | Field  | Type        | Null | Key | Default |
> Extra |
> > > > > > > > >> > > >
> +--------+-------------+------+-----+---------+-------+
> > > > > > > > >> > > > | mydate | varchar(32) | YES  |     | NULL    |
>   |
> > > > > > > > >> > > > | mydata | varchar(32) | YES  |     | NULL    |
>   |
> > > > > > > > >> > > >
> +--------+-------------+------+-----+---------+-------+
> > > > > > > > >> > > >
> > > > > > > > >> > > > I can see the logs in hdfs
> > > > > > > > >> > > >
> > > > > > > > >> > > > hadoop@ip-XX-XX-XX-XX:/mnt/var/lib/hadoop/steps/5$
> > > hadoop
> > > > > dfs
> > > > > > > -ls
> > > > > > > > >> > > > hdfs:///mnt/var/lib/hadoop/dfs/logs_sanitized_test
> > > > > > > > >> > > > Found 2 items
> > > > > > > > >> > > > drwxr-xr-x   - hadoop supergroup          0
> 2012-10-10
> > > 15:23
> > > > > > > > >> > > >
> > > /mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01
> > > > > > > > >> > > > drwxr-xr-x   - hadoop supergroup          0
> 2012-10-10
> > > 15:23
> > > > > > > > >> > > >
> > > /mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-02
> > > > > > > > >> > > >
> > > > > > > > >> > > > and if i tail one of the file I see the correct data
> > > > > > > > >> > > >
> > > > > > > > >> > > > hadoop@ip-XX-XX-XX-XX:/mnt/var/lib/hadoop/steps/5$
> > > hadoop
> > > > > dfs
> > > > > > > > >> -tail -f
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> hdfs:///mnt/var/lib/hadoop/dfs/logs_sanitized_test/dt=2012-10-01/part-m-00000
> > > > > > > > >> > > > 20121001230101 blablabla1
> > > > > > > > >> > > > 20121001230202 blablabla2
> > > > > > > > >> > > >
> > > > > > > > >> > > >
> > > > > > > > >> > > > Here is the trace when I run the command. Please
> note
> > > that
> > > > > no
> > > > > > > data
> > > > > > > > >> get
> > > > > > > > >> > > > transferred. I would appreciate any tips. Thanks a
> lot!
> > > > > > > > >> > > >
> > > > > > > > >> > > > Warning: /usr/lib/hbase does not exist! HBase
> imports
> > > will
> > > > > fail.
> > > > > > > > >> > > > Please set $HBASE_HOME to the root of your HBase
> > > > > installation.
> > > > > > > > >> > > > 12/10/10 16:25:25 DEBUG tool.BaseSqoopTool: Enabled
> > > debug
> > > > > > > logging.
> > > > > > > > >> > > > 12/10/10 16:25:25 WARN tool.BaseSqoopTool: Setting
> your
> > > > > > > password on
> > > > > > > > >> the
> > > > > > > > >> > > > command-line is insecure. Consider using -P instead.
> > > > > > > > >> > > > 12/10/10 16:25:25 DEBUG sqoop.ConnFactory: Loaded
> > > manager
> > > > > > > factory:
> > > > > > > > >> > > > com.cloudera.sqoop.manager.DefaultManagerFactory
> > > > > > > > >> > > > 12/10/10 16:25:25 DEBUG sqoop.ConnFactory: Trying
> > > > > > > ManagerFactory:
> > > > > > > > >> > > > com.cloudera.sqoop.manager.DefaultManagerFactory
> > > > > > > > >> > > > 12/10/10 16:25:25 DEBUG
> manager.DefaultManagerFactory:
> > > > > Trying
> > > > > > > with
> > > > > > > > >> > > scheme:
> > > > > > > > >> > > > jdbc:mysql:
> > > > > > > > >> > > > 12/10/10 16:25:25 INFO manager.MySQLManager:
> Preparing
> > > to
> > > > > use a
> > > > > > > > >> MySQL
> > > > > > > > >> > > > streaming resultset.
> > > > > > > > >> > > > 12/10/10 16:25:25 DEBUG sqoop.ConnFactory:
> Instantiated
> > > > > > > ConnManager
> > > > > > > > >> > > > org.apache.sqoop.manager.MySQLManager@5ef4f44a
> > > > > > > > >> > > > 12/10/10 16:25:25 INFO tool.CodeGenTool: Beginning
> code
> > > > > > > generation
> > > > > > > > >> > > > 12/10/10 16:25:25 DEBUG manager.SqlManager: No
> > > connection
> > > > > > > > >> paramenters
> > > > > > > > >> > > > specified. Using regular API for making connection.
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG manager.SqlManager: Using
> > > fetchSize
> > > > > for
> > > > > > > next
> > > > > > > > >> > > query:
> > > > > > > > >> > > > -2147483648
> > > > > > > > >> > > > 12/10/10 16:25:26 INFO manager.SqlManager:
> Executing SQL
> > > > > > > statement:
> > > > > > > > >> > > SELECT
> > > > > > > > >> > > > t.* FROM `ml_ys_log_gmt_test` AS t LIMIT 1
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.ClassWriter: selected
> > > columns:
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.ClassWriter:   mydate
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.ClassWriter:   mydata
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG manager.SqlManager: Using
> > > fetchSize
> > > > > for
> > > > > > > next
> > > > > > > > >> > > query:
> > > > > > > > >> > > > -2147483648
> > > > > > > > >> > > > 12/10/10 16:25:26 INFO manager.SqlManager:
> Executing SQL
> > > > > > > statement:
> > > > > > > > >> > > SELECT
> > > > > > > > >> > > > t.* FROM `ml_ys_log_gmt_test` AS t LIMIT 1
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.ClassWriter: Writing
> source
> > > > > file:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/ml_ys_log_gmt_test.java
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.ClassWriter: Table name:
> > > > > > > > >> ml_ys_log_gmt_test
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.ClassWriter: Columns:
> > > mydate:12,
> > > > > > > > >> mydata:12,
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.ClassWriter:
> sourceFilename
> > > is
> > > > > > > > >> > > > ml_ys_log_gmt_test.java
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.CompilationManager:
> Found
> > > > > existing
> > > > > > > > >> > > >
> > > /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/
> > > > > > > > >> > > > 12/10/10 16:25:26 INFO orm.CompilationManager:
> > > HADOOP_HOME
> > > > > is
> > > > > > > > >> > > /home/hadoop
> > > > > > > > >> > > > 12/10/10 16:25:26 INFO orm.CompilationManager: Found
> > > hadoop
> > > > > core
> > > > > > > > >> jar at:
> > > > > > > > >> > > > /home/hadoop/hadoop-core.jar
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.CompilationManager:
> Adding
> > > > > source
> > > > > > > file:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/ml_ys_log_gmt_test.java
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.CompilationManager:
> Invoking
> > > > > javac
> > > > > > > with
> > > > > > > > >> args:
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.CompilationManager:
> > > > > -sourcepath
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.CompilationManager:
> > > > > > > > >> > > >
> > > /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.CompilationManager:   -d
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.CompilationManager:
> > > > > > > > >> > > >
> > > /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.CompilationManager:
> > > -classpath
> > > > > > > > >> > > > 12/10/10 16:25:26 DEBUG orm.CompilationManager:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> /home/hadoop/conf:/usr/lib/jvm/java-6-sun/lib/tools.jar:/home/hadoop:/home/hadoop/hadoop-core-1.0.3.jar:/home/hadoop/lib/activation-1.1.jar:/home/hadoop/lib/annotations.jar:/home/hadoop/lib/ant-1.8.1.jar:/home/hadoop/lib/ant-launcher-1.8.1.jar:/home/hadoop/lib/ant-nodeps-1.8.1.jar:/home/hadoop/lib/apache-jar-resource-bundle-1.4.jar:/home/hadoop/lib/asm-3.1.jar:/home/hadoop/lib/avro-1.5.3.jar:/home/hadoop/lib/avro-compiler-1.5.3.jar:/home/hadoop/lib/avro-ipc-1.5.3.jar:/home/hadoop/lib/avro-maven-plugin-1.5.3.jar:/home/hadoop/lib/aws-java-sdk-1.3.2.jar:/home/hadoop/lib/build-helper-maven-plugin-1.5.jar:/home/hadoop/lib/commons-beanutils-1.7.0.jar:/home/hadoop/lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/lib/commons-cli-1.2.jar:/home/hadoop/lib/commons-codec-1.5.jar:/home/hadoop/lib/commons-collections-3.2.1.jar:/home/hadoop/lib/commons-configuration-1.6.jar:/home/hadoop/lib/commons-daemon-1.0.1.jar:/home/hadoop/lib/commons-digester-1.8.jar:/home/hadoop/lib/commons-el-1.0.jar:/home/hadoop/lib/commons-httpclient-3.1.jar:/home/hadoop/lib/commons-io-2.4.jar:/home/hadoop/lib/commons-lang-2.5.jar:/home/hadoop/lib/commons-logging-1.1.1.jar:/home/hadoop/lib/commons-logging-adapters-1.1.1.jar:/home/hadoop/lib/commons-logging-api-1.1.1.jar:/home/hadoop/lib/commons-math-2.1.jar:/home/hadoop/lib/commons-net-3.1.jar:/home/hadoop/lib/com.sun.el_1.0.0.v201004190952.jar:/home/hadoop/lib/core-3.1.1.jar:/home/hadoop/lib/docbkx-maven-plugin-2.0.13.jar:/home/hadoop/lib/ecj-3.6.jar:/home/hadoop/lib/emr-metrics-1.0.jar:/home/hadoop/lib/emr-s3distcp-1.0.jar:/home/hadoop/lib/file-management-1.2.1.jar:/home/hadoop/lib/ftplet-api-1.0.0.jar:/home/hadoop/lib/ftpserver-core-1.0.0.jar:/home/hadoop/lib/ftpserver-deprecated-1.0.0-M2.jar:/home/hadoop/lib/gson-1.6.jar:/home/hadoop/lib/guava-12.0.jar:/home/hadoop/lib/hadoop-ant-1.0.3.jar:/home/hadoop/lib/hadoop-ant.jar:/home/hadoop/lib/hadoop-capacity-scheduler-1.0.3.jar:/home/hadoop/lib/hadoop-client-1.0.3.jar:/home/hadoop/lib/hadoop-core-1.0.3.jar:/home/hadoop/lib/hadoop-core.jar:/home/hadoop/lib/hadoop-examples-1.0.3.jar:/home/hadoop/lib/hadoop-examples.jar:/home/hadoop/lib/hadoop-fairscheduler-1.0.3.jar:/home/hadoop/lib/hadoop-minicluster-1.0.3.jar:/home/hadoop/lib/hadoop-state-pusher-1.0.jar:/home/hadoop/lib/hadoop-test-1.0.3.jar:/home/hadoop/lib/hadoop-test.jar:/home/hadoop/lib/hadoop-thriftfs-1.0.3.jar:/home/hadoop/lib/hadoop-tools-1.0.3.jar:/home/hadoop/lib/hadoop-tools.jar:/home/hadoop/lib/hamcrest-all-1.1.jar:/home/hadoop/lib/hamcrest-core-1.1.jar:/home/hadoop/lib/hbase-0.92.0.jar:/home/hadoop/lib/high-scale-lib-1.1.1.jar:/home/hadoop/lib/high-scale-lib.jar:/home/hadoop/lib/hsqldb-1.8.0.10.jar:/home/hadoop/lib/httpclient-4.1.1.jar:/home/hadoop/lib/httpclient-cache-4.1.1.jar:/home/hadoop/lib/httpcore-4.1.jar:/home/hadoop/lib/httpcore-nio-4.1.jar:/home/hadoop/lib/httpmime-4.1.1.jar:/home/hadoop/lib/icu4j-4_0_1.jar:/home/hadoop/lib/jackson-core-asl-1.8.8.jar:/home/hadoop/lib/jackson-mapper-asl-1.8.8.jar:/home/hadoop/lib/jamon-anttask-2.4.0.jar:/home/hadoop/lib/jamon-api-2.3.0.jar:/home/hadoop/lib/jamon-maven-plugin-2.3.4.jar:/home/hadoop/lib/jamon-processor-2.4.1.jar:/home/hadoop/lib/jamon-runtime-2.4.0.jar:/home/hadoop/lib/jasper-compiler-5.5.23.jar:/home/hadoop/lib/jasper-runtime-5.5.23.jar:/home/hadoop/lib/java_util_concurrent_chm.jar:/home/hadoop/lib/java_util_hashtable.jar:/home/hadoop/lib/javax.el_2.1.0.v201004190952.jar:/home/hadoop/lib/javax.servlet.jsp_2.1.0.v201004190952.jar:/home/hadoop/lib/jaxb-api-2.1.jar:/home/hadoop/lib/jaxb-impl-2.1.12.jar:/home/hadoop/lib/jcommon-0.9.6.jar:/home/hadoop/lib/jersey-core-1.4.jar:/home/hadoop/lib/jersey-json-1.4.jar:/home/hadoop/lib/jersey-server-1.4.jar:/home/hadoop/lib/jettison-1.1.jar:/home/hadoop/lib/jetty-6.1.26.jar:/home/hadoop/lib/jetty-ajp-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-all-7.5.4.v20111024-javadoc.jar:/home/hadoop/lib/jetty-annotations-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-client-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-continuation-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-deploy-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-http-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-io-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-jmx-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-jndi-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-jsp-2.1-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-overlay-deployer-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-plus-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-policy-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-rewrite-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-security-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-server-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-servlet-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-servlets-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-spring-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-util-6.1.26.jar:/home/hadoop/lib/jetty-util-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-webapp-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-websocket-7.5.4.v20111024.jar:/home/hadoop/lib/jetty-xml-7.5.4.v20111024.jar:/home/hadoop/lib/jfreechart-0.9.21.jar:/home/hadoop/lib/joda-time-2.1.jar:/home/hadoop/lib/jruby-complete-no-joda-1.6.5.jar:/home/hadoop/lib/jsp-2.1-6.1.14.jar:/home/hadoop/lib/jsp-api-2.1-6.1.14.jar:/home/hadoop/lib/jsp-impl-2.1.3-b10.jar:/home/hadoop/lib/jsr305.jar:/home/hadoop/lib/jsr311-api-1.1.1.jar:/home/hadoop/lib/junit-4.8.1.jar:/home/hadoop/lib/junit.jar:/home/hadoop/lib/jython.jar:/home/hadoop/lib/kfs-0.2.2.jar:/home/hadoop/lib/libthrift-0.7.0.jar:/home/hadoop/lib/log4j-1.2.16.jar:/home/hadoop/lib/mail-1.4.3.jar:/home/hadoop/lib/mina-core-2.0.0-M5.jar:/home/hadoop/lib/mockito-all-1.8.5.jar:/home/hadoop/lib/netty-3.2.4.Final.jar:/home/hadoop/lib/opencsv-1.8.jar:/home/hadoop/lib/org.apache.taglibs.standard.glassfish_1.2.0.v201004190952.jar:/home/hadoop/lib/paranamer-2.3.jar:/home/hadoop/lib/plexus-active-collections-1.0-beta-2.jar:/home/hadoop/lib/plexus-build-api-0.0.4.jar:/home/hadoop/lib/plexus-compiler-api-1.8.1.jar:/home/hadoop/lib/plexus-compiler-javac-1.5.3.jar:/home/hadoop/lib/plexus-compiler-manager-1.5.3.jar:/home/hadoop/lib/plexus-digest-1.0.jar:/home/hadoop/lib/plexus-interpolation-1.12.jar:/home/hadoop/lib/plexus-io-1.0-alpha-4.jar:/home/hadoop/lib/plexus-resources-1.0-alpha-5.jar:/home/hadoop/lib/plexus-utils-2.1.jar:/home/hadoop/lib/plexus-velocity-1.1.3.jar:/home/hadoop/lib/protobuf-java-2.4.0a.jar:/home/hadoop/lib/servlet-api-2.5.jar:/home/hadoop/lib/slf4j-api-1.6.1.jar:/home/hadoop/lib/slf4j-log4j12-1.6.1.jar:/home/hadoop/lib/smart-cli-parser.jar:/home/hadoop/lib/snappy-java-1.0.3.2.jar:/home/hadoop/lib/stax-1.2.0.jar:/home/hadoop/lib/stax-api-1.0.1.jar:/home/hadoop/lib/surefire-api-2.4.3.jar:/home/hadoop/lib/surefire-booter-2.4.3.jar:/home/hadoop/lib/surefire-junit4-2.10.jar:/home/hadoop/lib/typica.jar:/home/hadoop/lib/velocity-1.7.jar:/home/hadoop/lib/visualization-datasource-1.1.1.jar:/home/hadoop/lib/xmlenc-0.52.jar:/home/hadoop/lib/xml-maven-plugin-1.0-beta-3.jar:/home/hadoop/lib/zookeeper-3.4.2.jar:/home/hadoop/lib/jsp-2.1/*.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../conf::/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/ant-contrib-1.0b3.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/ant-eclipse-1.0-jvm1.2.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/avro-1.5.3.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/avro-ipc-1.5.3.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/avro-mapred-1.5.3.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/commons-io-1.4.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/hsqldb-1.8.0.10.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/jackson-core-asl-1.7.3.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/jackson-mapper-asl-1.7.3.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/jopt-simple-3.2.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/mysql-connector-java-5.1.22-bin.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/paranamer-2.3.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../lib/snappy-java-1.0.3.2.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../sqoop-1.4.2.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/bin/../sqoop-test-1.4.2.jar::/home/hadoop/hadoop-core.jar:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/sqoop-1.4.2.jar
> > > > > > > > >> > > > Note:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/ml_ys_log_gmt_test.java
> > > > > > > > >> > > > uses or overrides a deprecated API.
> > > > > > > > >> > > > Note: Recompile with -Xlint:deprecation for details.
> > > > > > > > >> > > > 12/10/10 16:25:28 INFO orm.CompilationManager:
> Writing
> > > jar
> > > > > file:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/ml_ys_log_gmt_test.jar
> > > > > > > > >> > > > 12/10/10 16:25:28 INFO orm.CompilationManager:
> Writing
> > > jar
> > > > > file:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/ml_ys_log_gmt_test.jar
> > > > > > > > >> > > > 12/10/10 16:25:28 DEBUG orm.CompilationManager:
> > > Scanning for
> > > > > > > .class
> > > > > > > > >> files
> > > > > > > > >> > > > in directory:
> > > > > > > > >> /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b
> > > > > > > > >> > > > 12/10/10 16:25:28 DEBUG orm.CompilationManager: Got
> > > > > classfile:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/ml_ys_log_gmt_test.class
> > > > > > > > >> > > > -> ml_ys_log_gmt_test.class
> > > > > > > > >> > > > 12/10/10 16:25:28 DEBUG orm.CompilationManager:
> Finished
> > > > > writing
> > > > > > > > >> jar file
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> /tmp/sqoop-hadoop/compile/7f5cd67c0aa5dbf20256f72b30ae922b/ml_ys_log_gmt_test.jar
> > > > > > > > >> > > > 12/10/10 16:25:28 INFO mapreduce.ExportJobBase:
> > > Beginning
> > > > > > > export of
> > > > > > > > >> > > > ml_ys_log_gmt_test
> > > > > > > > >> > > > 12/10/10 16:25:29 WARN mapreduce.ExportJobBase: null
> > > > > FileStatus
> > > > > > > > >> object in
> > > > > > > > >> > > > isSequenceFiles(); assuming false.
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Using
> > > > > InputFormat:
> > > > > > > class
> > > > > > > > >> > > > org.apache.sqoop.mapreduce.ExportInputFormat
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/sqoop-1.4.2.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/mysql-connector-java-5.1.22-bin.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/sqoop-1.4.2.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/sqoop-1.4.2.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/snappy-java-1.0.3.2.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >>
> > > > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/commons-io-1.4.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >>
> > > > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/avro-ipc-1.5.3.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/jackson-core-asl-1.7.3.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/avro-1.5.3.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/ant-eclipse-1.0-jvm1.2.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/jackson-mapper-asl-1.7.3.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/ant-contrib-1.0b3.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/avro-mapred-1.5.3.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >> > >
> > > > > > > > >>
> > > > > > >
> > > > >
> > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/mysql-connector-java-5.1.22-bin.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >>
> > > > >
> file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/paranamer-2.3.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >>
> > > > > > >
> > > file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/jopt-simple-3.2.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 DEBUG mapreduce.JobBase: Adding
> to job
> > > > > > > classpath:
> > > > > > > > >> > > >
> > > > > > > > >>
> > > > > > >
> > > file:/home/hadoop/sqoop-1.4.2.bin__hadoop-1.0.0/lib/hsqldb-1.8.0.10.jar
> > > > > > > > >> > > > 12/10/10 16:25:29 INFO mapred.JobClient: Default
> number
> > > of
> > > > > map
> > > > > > > > >> tasks: 4
> > > > > > > > >> > > > 12/10/10 16:25:29 INFO mapred.JobClient: Default
> number
> > > of
> > > > > > > reduce
> > > > > > > > >> tasks:
> > > > > > > > >> > > 0
> > > > > > > > >> > > > 12/10/10 16:25:30 INFO mapred.JobClient: Setting
> group
> > > to
> > > > > hadoop
> > > > > > > > >> > > > 12/10/10 16:25:30 INFO input.FileInputFormat: Total
> > > input
> > > > > paths
> > > > > > > to
> > > > > > > > >> > > process
> > > > > > > > >> > > > : 2
> > > > > > > > >> > > > 12/10/10 16:25:30 DEBUG mapreduce.ExportInputFormat:
> > > Target
> > > > > > > > >> numMapTasks=4
> > > > > > > > >> > > > 12/10/10 16:25:30 DEBUG mapreduce.ExportInputFormat:
> > > Total
> > > > > input
> > > > > > > > >> bytes=0
> > > > > > > > >> > > > 12/10/10 16:25:30 DEBUG mapreduce.ExportInputFormat:
> > > > > > > maxSplitSize=0
> > > > > > > > >> > > > 12/10/10 16:25:30 INFO input.FileInputFormat: Total
> > > input
> > > > > paths
> > > > > > > to
> > > > > > > > >> > > process
> > > > > > > > >> > > > : 2
> > > > > > > > >> > > > 12/10/10 16:25:30 DEBUG mapreduce.ExportInputFormat:
> > > > > Generated
> > > > > > > > >> splits:
> > > > > > > > >> > > > 12/10/10 16:25:30 INFO mapred.JobClient: Running
> job:
> > > > > > > > >> > > job_201210101503_0017
> > > > > > > > >> > > > 12/10/10 16:25:31 INFO mapred.JobClient:  map 0%
> reduce
> > > 0%
> > > > > > > > >> > > > 12/10/10 16:25:48 INFO mapred.JobClient: Job
> complete:
> > > > > > > > >> > > job_201210101503_0017
> > > > > > > > >> > > > 12/10/10 16:25:48 INFO mapred.JobClient: Counters: 4
> > > > > > > > >> > > > 12/10/10 16:25:48 INFO mapred.JobClient:   Job
> Counters
> > > > > > > > >> > > > 12/10/10 16:25:48 INFO mapred.JobClient:
> > > > > > > SLOTS_MILLIS_MAPS=8498
> > > > > > > > >> > > > 12/10/10 16:25:48 INFO mapred.JobClient:     Total
> time
> > > > > spent
> > > > > > > by all
> > > > > > > > >> > > > reduces waiting after reserving slots (ms)=0
> > > > > > > > >> > > > 12/10/10 16:25:48 INFO mapred.JobClient:     Total
> time
> > > > > spent by
> > > > > > > > >> all maps
> > > > > > > > >> > > > waiting after reserving slots (ms)=0
> > > > > > > > >> > > > 12/10/10 16:25:48 INFO mapred.JobClient:
> > > > > > > SLOTS_MILLIS_REDUCES=0
> > > > > > > > >> > > > 12/10/10 16:25:48 INFO mapreduce.ExportJobBase:
> > > Transferred
> > > > > 0
> > > > > > > bytes
> > > > > > > > >> in
> > > > > > > > >> > > > 19.0791 seconds (0 bytes/sec)
> > > > > > > > >> > > > 12/10/10 16:25:48 INFO mapreduce.ExportJobBase:
> > > Exported 0
> > > > > > > records.
> > > > > > > > >> > >
> > > > > > > > >> >
> > > > > > > > >> >
> > > > > > > > >> >
> > > > > > > > >> > --
> > > > > > > > >> > Matthieu Labour, Engineering | *Action**X* |
> > > > > > > > >> > 584 Broadway, Suite 1002 – NY, NY 10012
> > > > > > > > >> > 415-994-3480 (m)
> > > > > > > > >>
> > > > > > > > >
> > > > > > > > >
> > > > > > > > >
> > > > > > > > > --
> > > > > > > > > Matthieu Labour, Engineering | *Action**X* |
> > > > > > > > > 584 Broadway, Suite 1002 – NY, NY 10012
> > > > > > > > > 415-994-3480 (m)
> > > > > > > > >
> > > > > > > > >
> > > > > > > >
> > > > > > > >
> > > > > > > > --
> > > > > > > > Matthieu Labour, Engineering | *Action**X* |
> > > > > > > > 584 Broadway, Suite 1002 – NY, NY 10012
> > > > > > > > 415-994-3480 (m)
> > > > > > >
> > > > > >
> > > > > >
> > > > > >
> > > > > > --
> > > > > > Matthieu Labour, Engineering | *Action**X* |
> > > > > > 584 Broadway, Suite 1002 – NY, NY 10012
> > > > > > 415-994-3480 (m)
> > > > >
> > > >
> > > >
> > > >
> > > > --
> > > > Matthieu Labour, Engineering | *Action**X* |
> > > > 584 Broadway, Suite 1002 – NY, NY 10012
> > > > 415-994-3480 (m)
> > >
> >
> >
> >
> > --
> > Matthieu Labour, Engineering | *Action**X* |
> > 584 Broadway, Suite 1002 – NY, NY 10012
> > 415-994-3480 (m)
>



-- 
Matthieu Labour, Engineering | *Action**X* |
584 Broadway, Suite 1002 – NY, NY 10012
415-994-3480 (m)

Mime
View raw message