In recent projects, the results of the completion of Hadoop operations (which exist on HDFs) need to be imported into Oracle, but when you import the Date field ' 2016-03-01 ' in HDFs with Sqoop, the Sqoop error says that the date type must be ' yyyy-mm-dd HH:MM:SS.ffffffff .Does Sqoop not support the custom to_date function, so I started looking for answers onli
1. Sqoop installed on Hadoop.client
2. Duplicate a copy of sqoop-env-template.sh, named sqoop-env.sh
3. Modify the contents of sqoop-env.sh:
Export Hadoop_common_home=/home/hadoopuser/hadoop
Export Hadoop_mapred_home=/home/hadoopuser/hadoop/lib
Export Hive_home=/home/hadoopuser/hive
4. Duplicate a copy of
When SQOOP interacts with ORACLE, the user name and table name must be of the same size. Otherwise, no results will be displayed. As follows: [hadoop @ htbbs bin] $. /sqoop list-tables -- connect "jdbc: oracle: thin: @ htbbs: 1521: htdb" -- username YANSP -- password testWarning:/usr/lib/hbase does not exist! HBase imports will fail. please set $ HBASE_HOME to the root of your HBase installation. warning: $
In the hadoop cluster environment, use sqoop to import the data generated by hive into the mysql database. The exception Causedby: java. SQL. SQLException: null, messagefromserver: success; unblockwithmysqladmin
In the hadoop cluster environment, sqoop is used to import the data generated by hive into the mysql database. The exception Caused by: java. SQL. SQLException: null, message from server: Host 'data
1. Modification of Sqoop's capital documents 2. Importing from MySQL to HBase (import)Bin/sqoop import \--connect jdbc:mysql://linux-hadoop3.ibeifeng.com:3306/sqoop \--username root \--password 123456 \--table tohdfs \--hbase-create-table \--hbase-table s1 \--hbase-row-key ID \--column-family info \-M 13. ResultsScan ' s1 '4. Export from HBaseThere is no such usage.Procedure: Integration of hbase with hive
Sqoop exports hive data to oracle and sqoophive
Use sqoop to import data from hive to oracle
1. Create a table in oracle Based on the hive table structure
2. Run the following command:
Sqoop export -- table TABLE_NAME -- connect jdbc: oracle: thin: @ HOST_IP: DATABASE_NAME -- username USERNAME -- password PASSWORD password -- export-dir/user/hive/test/TABLE_N
time of MapReduce is very small.But hive is not all SQL statements mapped to a MapReduce program, the only exception (not configured) is: SELECT * FROM table_name LIMIT, because this is only the first few data in the interception table;(5) Processing cleaning data using hive and exporting to a temporary tableThe data that is really needed will be drawn in this step, that is: the user clicks on the category of the Top20, and the data is counted as a temporary table of hive:item_20141216 table st
Tags: User student tle Use blog--Import target ScoreThere are usually two cases of importing hive table data to MySQL through Sqoop. The first is to import all the data from a table on hive to the table in MySQL. The second is to import some of the data from a table on hive to a table in MySQL. The difference between the two approaches is that the second case requires specifying the name of the column to import the data into. The two cases are importe
Tags: blog file data ar database SQL MySQL appCreate database logs;Use logsCREATE TABLE Weblogs_from_hdfs (MD5 varchar (32),URL varchar (64),Request_date date,Request_time time,IP varchar (15));Sqoop export-m 1--connect jdbc:mysql://hadoop:3306/logs--username root--password root--table weblogs_from_hdfs--expo Rt-dir/data/weblogs/import--input-fields-terminated-by ' \ t 'Null for data imported through Sqoop,
:# Create Hive data table Pms.yhd_categ_prior_userhive-e " set Mapred.job.queue.name=pms; set mapred.job.name=[cis]yhd_categ_prior_user;--Hive DDL DROP TABLE IF EXISTS pms.yhd_categ_prior_user;create TABLE Pms.yhd_categ_prior_user (category_id bigint, category_name string, Categor Y_level int , Default_import_categ_prior int , User_import_categ_prior int , Default_eliminate_categ_prior int , User_eliminate_categ_prior int , upd Ate_time string) row FORMAT delimited fields TERMINATED by ' \
-agent.sinks.hdfs-write.type = HDFs Hdfs-agent.sinks.hdfs-write.hdfs.path = hdfs://namenode/user/usera/test/ Hdfs-agent.sinks.hdfs-write.hdfs.writeformat=text # Bind the source and sink to the channel Hdfs-agent.sources.avro-collect.channels = Ch1 Hdfs-agent.sinks.hdfs-write.channel = Ch1 Start the conf2.conf first, then start conf1.conf agent. Because The Avro source should start first then Avro Sink can connect to it. #when use memory change, issue is: Org.apache.flume.ChannelException:Unabl
Tags: connect dir date overwrite char post arch src 11.2.0.1Due to the increasing volume of business data and the large amount of computing, the traditional number of silos has been unable to meet the computational requirements, so it is basically to put the data on the Hadoop platform to implement the logical computing, then it involves how to migrate Oracle Data Warehouse to the Hadoop platform. Here we have to mention a very useful tool--sqoop, whi
Training Big Data Architecture development!from zero-based to advanced, one-to-one training! [Technical qq:2937765541]--------------------------------------------------------------------------------------------------------------- ----------------------------Course System:get video material and training answer technical support addressCourse Presentation ( Big Data technology is very wide, has been online for you training solutions!) ):get video material and training answer technical support ad
The production of Sqoop is mainly due to the following requirements:1. Most enterprises that use Hadoop technology to handle big data businesses have large amounts of data stored in traditional relational databases (RDBMS).2, due to the lack of tools to support, the Hadoop and traditional database systems in the transmission of data to each other is a very difficult thing.3. Based on the first two considerations, there is a need for a project to trans
[Author]: KwuSqoop handles Clob and Blob fields, Clob is large text in Oracle, andblobs store binary files. This type of field is encountered when importing hive or HDFS requires special handling. 1. Test Tables in OracleCREATE TABLE t_lob ( A INTEGER, B CLOB, C BLOB )Test dataInsert into T_lob (A, B, C) VALUES (1, ' Clob test ', To_blob (' 3456 '));2. Sqoop ScriptImport--append--connectjdbc:oracle:thin: @localhost: 1521/
After the project was completed, we found the tragedy. By default, sqoop was used to list data tables from Oracle databases. If the data accuracy is greater than 15 digits, some fields in the imported table are of the double type by default. As a result, more than 16 fields are imported to hive. The query time is only 15-bit precise. Sorry, remember.
Hadoop cluster-based Hive Installation
Differences between Hive internal tables and external tables
internal three components:A) source: Acquisition source for docking with the data source for data acquisitionb) Sink: sink, collect data for the purpose of transmitting data to the next level agent or transfer data to the final storage systemc) Channel:angent Internal data transfer channel for passing data from source to sinkFlume supports numerous source and sink typesF Installation deployment for Lume1, Flume installation is very simple, only need to decompress, of course, if there is alread
Org.apache.hadoop.mapred.maptask$newtrackingrecordreader.nextkeyvalue (Maptask.java:556) at Org.apache.hadoop.mapreduce.task.MapContextImpl.nextKeyValue (Mapcontextimpl.java: the) at Org.apache.hadoop.mapreduce.lib.map.wrappedmapper$context.nextkeyvalue (Wrappedmapper.java: the) at Org.apache.hadoop.mapreduce.Mapper.run (Mapper.java:145) at Org.apache.sqoop.mapreduce.AutoProgressMapper.run (Autoprogressmapper.java: -) at Org.apache.hadoop.mapred.MapTask.runNewMapper (Maptask.java:787) at Org.ap
Training Big Data architecture development, mining and analysis!from zero-based to advanced, one-to-one training! [Technical qq:2937765541]--------------------------------------------------------------------------------------------------------------- ----------------------------Course System:get video material and training answer technical support addressCourse Presentation ( Big Data technology is very wide, has been online for you training solutions!) ):Get video material and training answer
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.