"1.7.0_79"Java (TM) SE Runtime Environment (build 1.7.0_79-b15)Java HotSpot (TM) Client VM (build 24.79-b02, Mixed mode)Indicates that the JDK environment variable is configured successfullyThird, install Hadoop3.1 Download Hadoop, choose Stable version, in fact stable version is 1.2.1, download the site as follows:Http://mirror.esocc.com/apache/hadoop/common/hadoop
Hadoop + Hbase cluster data migration
Data migration or backup is a possible issue for any company. The official website also provides several solutions for hbase data migration. We recommend using Hadoop distcp for migration. It is suitable for data migration between large data volumes or cross-version clusters.
Version
Hadoop2.7.1
Hbase0.98.12
A problem found d
How to makeProgramDistributed running in a hadoop cluster is a headache.
Someone may say that right-click "run on hadoop" in the eclipse class file. Note: by default, "run on hadoop" in Eclipse only runs on a single machine, because in order to make programs run in a distributed manner in a
HDFs cluster, do not use the same version of Hadoop, and avoid Hadoop itself causing data errors.(iii) file system inspectionPeriodically run the HDFs tool on the entire file system to proactively find missing or damaged blocks.It is recommended to do it once a day.[Email protected] ~]$ Hadoop fsck/... Omit the output
Hbase-site.xml3. Exit Safe Mode-safemodeHDFs dfsadmin--safenode Leave4.hadoop cluster boot not successful-format multiple timesClose the cluster, delete the Hadoopdata directory, and delete all the log files in the Logs folder under the Hadoop installation directory. Reformat and start the
[TOC]
Redis Note-taking (ii): Java API use with Redis distributed cluster environment using Redis Java API (i): Standalone version of Redis API usageThe Redis Java API operates through Jedis, so the first Jedis third-party libraries are needed, because the MAVEN project is used, so the Jedis dependency is given first:Basic code exampleThe commands that Redis can provide, Jedis are also provided, and are very similar in use, so here's just some c
Original? Blog. csdn. netyang_bestarticledetails41280553 the following sections describe how to configure a Hadoop cluster. The configuration file's Hadoop configuration is done through two important configuration files under the conf directory: the default configuration for hadoop-default.xml read-only. Configuration
After the accumulation of the front, today finally realized the cluster environment to deploy Hadoop, and successfully ran the official example.
Work as follows:
Two machines:
Namenode: Internet Small, 3G memory, machine name: yp-x100e,ip:192.168.101.130.
Datanode: Virtual machine, Win7 download VMWare10 virtual UBUNTU14, virtual machine name: ph-v370,ip:192.168.101.110
Ensure that you can ping each ot
Apache Ambari is a Web-based open-source project that monitors, manages, and manages Hadoop lifecycles. It is also a project that selects management for the Hortonworks data platform. Ambari supports the following management services:
Apache HBaseApache HCatalogApache Hadoop HDFSApache HiveApache Hadoop MapReduceApache OozieApache PigApache SqoopApache TempletonA
production cluster changeexecuted separately on 5 servers, the following commandvim/app/hadoop/etc/hadoop/hadoop-env.shAdd the following parametersExport hadoop_opts= "-xx:+useparallelgc-xmx4g" In order to facilitate future operation, special record the cluster restart ope
parameter fs. checkpoint. dir;
Copy the file in namesecondary to fs. checkpoint. dir;
./Hadoop namenode-importCheckpoint;
Start NameNode and add-importCheckpoint. (This sentence is plagiarized with hadoop-0.20.2/hadoop-0.20.2/docs/cn/hdfs_user_guide.html # Secondary + NameNode, look at the documentation, There are instructions)
3.
java_home=/path/to/java command in console2 storing the data to be retrieved in HDFs$ bin/hadoop fs-put Urldir UrldirNote: The first Urldir is a local folder, a URL data file is stored, one URL per lineThe second urldir is a storage path for HDFs3 Starting the Nutch commandExecute the following command under the Nutch_hone/runtime/deploy directory$ bin/nutch Crawl Urldir–dir crawl-depth 3–topn 10After the command executes successfully, the crawl dire
requires reboot,Do not want to restart the words in the Code add: System.setproperty ("Hadoop.home.dir", "d:\\soft\\linux\\hadoop-2.4.0");(3) Exception information 3:Exception in thread "main" Org.apache.hadoop.mapred.FileAlreadyExistsException:Output directory hdfs:// 192.168.1.200:9000/user/output already existsWorkaround: The output folder already exists, modify the export folder or between outputs deleted(4) exception information 4:[97;97;98;99;1
After more than a week, I finally set up the latest version of Hadoop2.2 cluster. During this period, I encountered various problems and was really tortured as a cainiao. However, when wordcount gave the results, I was so excited ~~ (If you have any errors or questions, please correct them and learn from each other)
In addition, you are welcome to leave a message when you encounter problems during the configuration process and discuss them with each o
The cluster space has been a little tight recently and is always worried about space shortage and crashes. The recent resizing is not realistic. After communicating with cluster users, we found that the cluster stores a lot of useless historical data and can be deleted, in this way, you can use a crontab script to generate a
comment #) Note: Some blogs write that you need to comment out the next line
export hadoop_opts= "-djava.security.krb5.realm=ox. ac.uk-djava.security.krb5.kdc=kdc0.ox.ac.uk:kdc1.ox.ac.uk "(remove comments) I didn't find this one, so I didn't have this one.
2. Configuration core-site.xml--Specifies the hostname and port of the Namenode
4. Configuration mapred-site.xml--Specifies the hostname and port of the Jobtracker
5.SSH configuration turn on sharing in
configuration file are:
Run the ": WQ" command to save and exit.
Through the above configuration, we have completed the simplest pseudo-distributed configuration.
Next, format the hadoop namenode:
Enter "Y" to complete the formatting process:
Start hadoop!
Start hadoop as follows:
Use the JPS command that comes with Java to query all daemon processes:
Star
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.