needs to be changed to 1.set up HDFs backup mode -Configuration> Property> name>Dfs.replicationname> value>1value> Property>Configuration>
Modify Mapred-site.xml
The file is a MapReduce configuration file that configures the Jobtracker address and port.Configuration> Property> name>Mapred.job.trackername> value>localhost:9001value> Property>Configuration>4. Before starting Hadoop, file system
to the Environment/etc/profile:
Export hadoop_home =/ home/hexianghui/hadoop-0.20.2
Export Path = $ hadoop_home/bin: $ path
7. Configure hadoop
The main configuration of hadoop is under the hadoop-0.20.2/CONF.
(1) configure the Java environment in CONF/hadoop-env.sh (nameno
: // svn.apache.org/repos/asf/hadoop/common/branches/branch-1.2-r1503152
CompiledbymattfonMonJul2215: 23: 09PDT2013
Fromsourcewithchecksum6923c86528809c4e7e6f493b6b413a9a
To start HADOOP, You Need To Format namenode first, and then start all services [html] view plaincopyprint?
Hadoopnamenode-format
Start-all.sh
View the process [html] view plaincopyprint?
Hadoo
the log is retained (in seconds) Configuration item 11, log check time Configuration item 12, directory Configuration item 13, directory prefix Mapred-site.xml does not have mapred-site.xml, enter VI mapred-Press "tab" to find Mapred-site.xml.template, copy the file CP mapred-site.xml.template mapred-site.xml Configuration item 1,mapreduce Framework Configuration item 2,mapreduce communication port Configuration item 3,mapreduce as Industry history Port Configuration item 4,mapreduce
variable under/etc/profile plus Export path= $PA Th:/home/zookeeper/bin Save, run source/etc/profile Copy configuration files to two other node Scp/etc/profile root@ node2:/etc/ Scp/etc/profile root@node3:/etc/
Run Source/etc/profile 6) Start Shutdown Firewall service iptables stopStartzkserver.sh start5. Deployment 1) Start the node2,3,4 journalnodeInto the/home/hadoop-2.5.2/sbin/.Run./hadoop-d
Hadoop cannot be started properly (1)
Failed to start after executing $ bin/hadoop start-all.sh.
Exception 1
Exception in thread "Main" Java. Lang. illegalargumentexception: Invalid URI for namenode address (check fs. defaultfs): file: // has no authority.
Localhost: At org. Apache. hadoop. HDFS. server. namenode. namenode. getaddress (namenode. Java: 214)
Localh
block spread out.
If the version is inconsistent between two clusters, using HDFS may cause an error because the RPC system is incompatible. Then you can use the HFTP protocol based on the HTTP protocol, but the destination address must also be hdfs, like this:
Hadoop distcp hftp://namenode:50070/user/hadoop/input HDFS://NAMENODE:9000/USER/
1. What is a distributed file system?
A file system stored across multiple computers in a management network is called a distributed file system.
2. Why do we need a distributed file system?
The reason is simple. When the data set size exceeds the storage capacity of an independent physical computer, it is necessary to partition it and store it on several independent computers.
3. distributed systems are more complex than traditional file systems
Because the Distributed File System arc
commands:
$ Start-all.sh
In fact, this script calls the above two commands.
The local computer starts three Daemon Processes: One namenode, one secondary namenode, and one datanode. You can view the log files in the logs directory to check whether the daemon is successfully started, or view jobtracker at http: // localhost: 50030/or at http: // localhost: 50070/view namenode. In addition, the JPS command of Java can also check whether the daemon is r
[Linux] [Hadoop] Run hadoop and linuxhadoop
The preceding installation process is to be supplemented. After hadoop installation is complete, run the relevant commands to run hadoop.
Run the following command to start all services:
hadoop@ubuntu:/usr/local/gz/
First explain the configured environmentSystem: Ubuntu14.0.4Ide:eclipse 4.4.1Hadoop:hadoop 2.2.0For older versions of Hadoop, you can directly replicate the Hadoop installation directory/contrib/eclipse-plugin/hadoop-0.20.203.0-eclipse-plugin.jar to the Eclipse installation directory/plugins/ (and not personally verified). For HADOOP2, you need to build the jar f
Why is the eclipse plug-in for compiling Hadoop1.x. x so cumbersome?
In my personal understanding, ant was originally designed to build a localization tool, and the dependency between resources for compiling hadoop plug-ins exceeds this goal. As a result, we need to manually modify the configuration when compiling with ant. Naturally, you need to set environment variables, set classpath, add dependencies, set the main function, javac, and jar configur
-declipse.home=/home/wu/opt/eclipse-dversion=1.0.1 jar
Once the compilation is complete, you can find the Eclipse plugin.3. Installation Steps(1) Pseudo-distributed configuration process is also very simple, only need to modify a few files, in the code of the Conf folder, you can find the following configuration files, the specific process I will not say, here list my configuration:Core-site.xmlHdfs-site.xmlMapred-site.xmlGo to the Conf folder, modify the configuration file:
NameNode at localhost/127.0.0.1************************************************************/An exit status of 0 indicates that the initialization succeeded, and if 1 indicates that the format failed.Then open the daemon for the named node and the data node:After the boot is complete, enter the JPS command to see if the startup is successful, and if successful, you will see 3 processes:Jps、NameNode, DataNode andSecondaryNameNodeAfter successful startup, you can access the Web interface http://lo
/datavalue> Property>Configuration>Run Hadoop after the configuration is complete.Four. Run hadoop4.1 to initialize the HDFS systemExecute the command in the hadop2.7.1 directory:Bin/hdfs Namenode-formatThe following results show that the initialization was successful.4.2 OpenNameNodeAndDataNodeDaemon processExecute the command in the hadop2.7.1 directory:sbin/start-dfs.shSuccess is as follows:4.3 Use the JPS command to view process information:If the
simultaneously:
Mapred. tasktracker. map. tasks. maximum = 8
Mapred. tasktracker. reduce. tasks. maximum = 6
Ssh configuration (so that it can log on via ssh without a password, that is, through certificate authentication)
Sh-keygen-t dsa-p'-f ~ /. Ssh/id_dsa
Cat ~ /. Ssh/id_dsa.pub> ~ /. Ssh/authorized_keys
Note: connect to host localhost port 22: Connection refused
Make sure that ssh runs normally when running hadoop. There may be multiple cause
the following content
Master
Slave1
After the preceding steps are completed, copy the hadoop-2.2.0 directory and content to the same path on the master machine as the hduser using the scp command:
Scp hadoop folder to various machines: scp/home/hduser/hadoop-2.2.0 slave1:/home/hduser/hadoop-2.2.0
7. Format hdfs (usual
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.