Discover hadoop commands tutorial, include the articles, news, trends, analysis and practical advice about hadoop commands tutorial on alibabacloud.com
1.hadoopView the directory on HDFs: hadoop fs-ls/ Create a directory on HDFs: -mkdir/jiatest upload the file to HDFs Specify directory: -put test.txt /Jiatest upload jar package to Hadoop run: hadoop jar maven_test-1.0-snapshot.jar org.jiahong.test.WordCount/ jiatest/jiatest/Output View result: -cat/jiatest/output/part-r-000002.linuxU
Hadoop Namenode-format formatted Distributed File systemstart-all.sh Start all Hadoop daemonsstop-all.sh Stop all Hadoop daemonsstart-mapred.sh Start the Map/reduce daemonstop-mapred.sh Stop Map/reduce DaemonStart-dfs.sh starting the HDFs daemonstop-mapred.sh Stop HDFs Daemonstart-balancer.sh HDFS data Block load BalancingFS in the following command can also be w
1.1)vim/etc/udev/rules.d/ --persistent-Net.rulesVI/etc/sysconfig/network-scripts/ifcfg-Eth0type=Ethernetuuid=57d4c2c9-9e9c-48f8-a654-8e5bdbadafb8onboot=yesnm_controlled=YesBootproto = staticDefroute=Yesipv4_failure_fatal=Yesipv6init=NoNAME="System eth0"HWADDR=xx: 0c: in: -: E6:ecipaddr =172.16.53.100PREFIX= -gateway=172.16.53.2Last_connect=1415175123dns1=172.16.53.2The virtual machine's network card is using the virtual network cardSave Exit X or Wq2)Vi/etc/sysconfig/networkNetworking=yesHostnam
Install Hadoop 2.2.0 on Ubuntu Linux 13.04 (Single-node Cluster)This tutorial explains what to install Hadoop 2.2.0/2.3.0/2.4.0/2.4.1 on Ubuntu 13.04/13.10/14.04 (Single-node Cluster) . This is setup does not require a additional user for Hadoop. All files related to Hadoop
From:http://www.2cto.com/database/201303/198460.htmlHadoop HDFs Common CommandsHadoop common commands:Hadoop FSView all commands supported by Hadoop HDFsHadoop fs–lslisting directory and file informationHadoop FS–LSRLoop lists directories, subdirectories, and file informationHadoop fs–put Test.txt/user/sunlightcsCopy the test.txt of the local file system to the/user/sunlightcs directory of the HDFs file sys
, soHDFs has a high degree of fault tolerance.3. High data throughput HDFs uses a "one-time write, multiple read" This simple data consistency model, in HDFS , once a file has been created, written, closed, generally do not need to modify, such a simple consistency model, to improve throughput.4. Streaming data access HDFS has a large scale of data processing, applications need to access a large amount of information at a time, and these applications are generally batch processing, rather than
Modify permissions for a file or directory
Chown
chown [ options ] User [ . Group ] file/dir
Modify the owner of a file
Chgrp
Chgrp [-R] Group name Dir/file
Modify the owning group of a file
Iv. Systems and Networks
Option name
Meaning
passwd xxx
Change Password
Df-ah
View disk space
Ps-ef |grep
View process
Kill-9
Kill the process
mytable from Database mydb to the E:\MySQL\mytable.sql file.
c:\> mysqldump-h localhost-u root-p mydb mytable>e:\mysql\mytable.sql
Example 3: Export the structure of the database mydb to a e:\MySQL\mydb_stru.sql file.
c:\> mysqldump-h localhost-u root-p mydb--add-drop-table >e:\mysql\mydb_stru.sql
Note :-h localhost can be omitted, it is generally used on the virtual host.3) Export data structure onlyformat : mysqldump-u [Database user name]-p-t [the name of the data
localhost-u root-p mydb >e:\mysql\mydb.sql
Then enter the password, wait for an export to succeed, you can check the target file for success. Example 2: Export mytable from Database mydb to the E:\MySQL\mytable.sql file.
c:\> mysqldump-h localhost-u root-p mydb mytable>e:\mysql\mytable.sql
Example 3: Export the structure of the database mydb to a e:\MySQL\mydb_stru.sql file.
c:\> mysqldump-h localhost-u root-p mydb--add-drop-table >e:\mysql\mydb_stru.sql
Cloudera, compilation: importnew-Royce Wong
Hadoop starts from here! Join me in learning the basic knowledge of using hadoop. The following describes how to use hadoop to analyze data with hadoop tutorial!
This topic describes the most important things that users face when u
-distributed mode on a single node, where each Hadoop daemon runs as a standalone Java process.ConfigurationUse the following:Etc/hadoop/core-site.xml:123456Etc/hadoop/hdfs-site.xml:Interested can continue to see the next chapter
Many people know that I have big data training materials, all naïve thought I have a full set of big data development,
Statement
This article is based on CentOS 6.x + CDH 5.x
Zookeeper what to use to see the previous tutorial, you will find multiple occurrences of zookeeper, such as the auto failover Hadoop zookeeper, Hbase Regionserver also have to use zookeeper. In fact, more than Hadoop, including the now small and famous Storm with the zookeeper. So what exactly
processing of batch and interactive data. TEZ is being adopted by other frameworks in Hive, Pig, and Hadoop ecosystems, and can also be used as the underlying execution engine with other commercial software, such as ETL tools, to replace Hadoop MapReduce. ZooKeeper: A high-performance distributed application Coordination Service. (The contents of the ZooKeeper are described in later chapters)
Installing the Hadoop tutorial on WindowsSee 2010.1.6 www.hadoopor.com/[email protected]1. Installing the JDKInstalling the JRE is not recommended, but it is recommended to install the JDK directly because the JRE can be installed at the same time when the JDK is installed. The development of the MapReduce program and the compilation of Hadoop depend on the JDK,
Alex's Hadoop cainiao Tutorial: tutorial 10th Hive getting started, hadoophiveInstall Hive
Compared to many tutorials, I first introduced concepts. I like to install them first, and then use examples to introduce them. Install Hive first.
First confirm whether the corresponding yum source has been installed, if not as written in this
Alex's Hadoop cainiao Tutorial: 7th Sqoop2 export tutorial, hadoopsqoop2
Take over the previous lesson. Now let's talk about the export tutorial.Check connection
First, check whether there are available connection connections. If not, create a connection based on the method of the previous lesson.
sqoop:000> show connector --all1 connector(s) to show: Connector
prompts to entersqoop:000> Create job--xid 1--type exportcreating job for connection with ID 1Please fill following values to create New job Objectname:export to Employeedatabase configurationschema name:table name:employeetable SQL statement:table Co Lumn names:stage table name:clear Stage table:input configurationinput directory:/user/alexthrottling resourcesextract Ors:Loaders:New job is successfully created with validation status FINE and persistent ID 3Perform this tasksqoop:000> Start Jo
Compared with many tutorials, Hive has introduced concepts first. I like to install them first, and then use examples to introduce concepts. Install Hive first. Check whether the corresponding yum source has been installed. If the yum source blog. csdn. netnsrainbowarticledetails42429339hive is not installed according to the yum source file written in this tutorial
Compared with many tutorials, Hive has introduced concepts first. I like to install the
Take over the previous lesson. Now let's talk about exporting the tutorial and check the connection to see if there is any available connection. If not, create a sqoop: 000showconnector -- all1connector (s) toshow according to the method in the previous lesson: connectorwithid1: Name: generic-jdbc-connectorClass: org. apache. sqoop. c
Take over the previous lesson. Now let's talk about exporting the tutorial
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.