file is used to send the past cooperation files.
[[emailprotected] ~]# source ~/.bash_profile [[emailprotected] ~]#
Next, we can start our cluster.Start an HDFS Cluster
HDFS is a distributed file system of hadoop. In short, HDFS stores data. It's massive data!
1. Format namenode
[[emailprotected] ~]# hdfs namenode -format
2. Start the Cluster
Run the JPS command to view the following information:
We can see that a namenode and datanode are
-http://localhost:50070/ (4) Create the HDFs directory required to perform the MapReduce job
$/usr/lib/java/hadoop/bin/hdfs Dfs-mkdir/user $/usr/lib/java/hadoop/bin/hdfs dfs-mkdir/user/(5) Copy the input file to the Distributed File system
$/usr/lib/java/hadoop/bin/hdfs dfs-put Etc/
[Hadoop] how to install Hadoop and install hadoop
Hadoop is a distributed system infrastructure that allows users to develop distributed programs without understanding the details of the distributed underlying layer.
Important core of Hadoop: HDFS and MapReduce. HDFS is res
return.Next, execute:sbin/start-yarn.sh After executing these two commands, Hadoop will start and runBrowser opens http://localhost:50070/, you will see the HDFs administration pageBrowser opens http://localhost:8088, you will see the Hadoop Process Management page7. WordCount TestFirst enter the/usr/local/hadoop/d
This document describes how to operate a hadoop file system through experiments.
Complete release directory of "cloud computing distributed Big Data hadoop hands-on"
Cloud computing distributed Big Data practical technology hadoop exchange group:312494188Cloud computing practices will be released in the group every day. welcome to join us!
First, let's loo
1 Creating Hadoop user groups and Hadoop users STEP1: Create a Hadoop user group:~$ sudo addgroup Hadoop STEP2: Create a Hadoop User:~$ sudo adduser-ingroup Hadoop hadoopEnter the password when prompted, this is the new
step.After the configuration is complete, perform NameNode formatting$./bin/hdfs Namenode-formatStart the Namenode and Datanode processes and view the startup results$./sbin/start-dfs. SH $ JPSAfter the boot is complete, the command JPS can be used to determine whether the startup is successful, and if successful, the following processes are listed: "NameNode", "DataNode" and "Secondarynamenode"At this point, it is also possible to ask for a localhost password, if it is clearly entered the corr
"Note" If you are using Hadoop 2.4.1 64-bit, there may be a series of warn prompts, such as WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable this hint, which can be ignored and will not affect normal use.
View processInput jps canSuccessful startup will list the following processes: NameNode, Datanode, and Secondarynamenode
192.168.1.150
192.168.1.151
4. Copy the configured Hadoop to the corresponding locations of each node, via the SCP
Scp-r/home/hadoop/192.168.1.150:/home/
Scp-r/home/hadoop/192.168.1.151:/home/
5. Start Hadoop at master server, start automatically from node, enter/home/hadoop
=/home/hadoop/hadoop-2.5.1/tmpexport HADOOP_SECURE_DN_PID _dir=/home/hadoop/hadoop-2.5.1/tmp 2.6.yarn-site.xml file 2. TheHadoopAdding environment Variables sudo vim/etc/profile Add the following two lines to export Hadoop_home=/home/hadoop/
. [[emailprotected]~]$hdfsdfsadmin-reportconfiguredcapacity: 2125104381952 (1.93NBSP;TB) presentcapacity:1975826509824 (1.80TB) DFS remaining:1975824982016 (1.80NBSP;TB) dfsused:1527808 (1.46MB) DFSUsed% :0.00%underreplicatedblocks:0blockswithcorruptreplicas: 0missingblocks:0missingblocks (withreplicationfactor1): 0--------------------- ----------------------------livedatanodes (4): View DataNode and are also available through the Web page State of namenode: http://hddcluster2:
Contact Hadoop has been 1.5, during the accumulation of some hadoop operations experience, has always wanted to develop a Hadoop monitoring system, just recently the laboratory has a related project, took the opportunity to study a bit, here to summarize the Hadoop monitoring method.
The HDFs and Jobtracker monitoring
1 operating mode:Stand-alone Mode (standalone): standalone mode is the default mode for Hadoop. When the source package for Hadoop was first decompressed, Hadoop was unable to understand the hardware installation environment and conservatively chose the minimum configuration. In this default mode, all 3 XML files are empty. When the configuration file is empty,
Introduction HDFs is not good at storing small files, because each file at least one block, each block of metadata will occupy memory in the Namenode node, if there are such a large number of small files, they will eat the Namenode node's large amount of memory. Hadoop archives can effectively handle these issues, he can archive multiple files into a file, archived into a file can also be transparent access to each file, and can be used as a mapreduce
address and port of JobTracker.
(7) When hadoop is run for the first time, format the Hadoop file system.
In the hadoop directory, enter:
$ Bin/hadoop namenode-format
(8) Start the Hadoop service:
$ Bin/start-all.sh
If no error is reported, the startup is successful.
(9)
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.