\catalina\localhost;Create a new XML file based on the project name you deployed, called Solr.xml if the package is called SOLR.The contents are: 3, Tomcat boot java_opts parameter setting modeUnder the root directory where you installed Tomcat, find Bin\catalina.bat added in the java_opts option,Like Windows, you can add a row of set JAVA_OPTS-DSOLR.SOLR.HOME=C:/EXAMPLE2/SOLR to the frontResources:Http://www.myexception.cn/open-source/745464.html
Copyright NOTICE: This article for Bo Master o
Python Access secured Hadoop Cluster through Thrift APIApache Thrift Python Kerberos Support typical way to connect Kerberos secured Thrift server example-hive example-hbase
Apache Thrift Python Kerberos Support
Both supports are only avaliable in Linux platform Native Support
Dependency: Kerberos (Python package) >> PURE-SASL (python package) >> Thrift (Python package)Source: https://github.com/apache/thr
Fluentd is an open source collection event and log system that currently offers 150 + extensions that let you store big data for log searches, data analysis and storage.
Official address http://fluentd.org/plugin address http://fluentd.org/plugin/
Kibana is a Web UI tool that provides log analysis for ElasticSearch, and it can be used to efficiently search, visualize, analyze, and perform various operations on logs. Official Address http://www.elasticsearch.org/overview/kibana/
Elasticsearch is
Install SSH
Hadoop uses SSH for communication. In this case, we need to set the password to null, that is, no password is required to log on. This eliminates the need to enter a secret during each communication. The installation is as follows:
Enter "Y" for installation and wait for the automatic installation to complete.
Start the service after installing SSH
Run the following command to verify that the service is properly started:
You can see
First, prepareUpload apache-hive-1.2.1.tar.gz and Mysql--connector-java-5.1.6-bin.jar to NODE01Cd/toolsTAR-ZXVF apache-hive-1.2.1.tar.gz-c/ren/Cd/renMV apache-hive-1.2.1 hive-1.2.1This cluster uses MySQL as the hive metadata storeVI Etc/profileExport hive_home=/ren/hive-1.2.1Export path= $PATH: $HIVE _home/binSource/etc/profileSecond, install MySQLYum-y install MySQL mysql-server mysql-develCreating a hive Database Create databases HiveCreate a hive u
When a problem occurs in a single node of a Hadoop cluster, it is generally not necessary to restart the entire system, just restart the node and it will automatically connect to the entire cluster.Enter the following command on the necrotic node:hadoop-daemon.sh Start Datanodehadoop-daemon.sh Start SecondarynamenodeThe cases are as follows:Hadoop node crashes, can ping Pass, SSH Connection not onCase:Time:
mytable from Database mydb to the E:\MySQL\mytable.sql file.
c:\> mysqldump-h localhost-u root-p mydb mytable>e:\mysql\mytable.sql
Example 3: Export the structure of the database mydb to a e:\MySQL\mydb_stru.sql file.
c:\> mysqldump-h localhost-u root-p mydb--add-drop-table >e:\mysql\mydb_stru.sql
Note :-h localhost can be omitted, it is generally used on the virtual host.3) Export data structure onlyformat : mysqldump-u [Database user name]-p-t [the name of the data
localhost-u root-p mydb >e:\mysql\mydb.sql
Then enter the password, wait for an export to succeed, you can check the target file for success. Example 2: Export mytable from Database mydb to the E:\MySQL\mytable.sql file.
c:\> mysqldump-h localhost-u root-p mydb mytable>e:\mysql\mytable.sql
Example 3: Export the structure of the database mydb to a e:\MySQL\mydb_stru.sql file.
c:\> mysqldump-h localhost-u root-p mydb--add-drop-table >e:\mysql\mydb_stru.sql
_ 14
Export Path = $ java_home/bin: $ path
Export classpath =.: $ java_home/lib/dt. jar: $ java_home/lib/tools. Jar
1.5 test JDK
1) create a test. Java file in the text editor, enter the following code and save the file:
Public class test {
Public static void main (string ARGs []) {
System. Out. println ("A New JDK test! ");
}
}
2) Compile:
Run the javac test. Java command on the shell terminal.
3) run:
Run the Java test command on the shell terminal.
When "A New JDK test!" appears
1, add the configuration file to the project source directory (SRC)
mapreduce.framework.name
yarn
Read the contents of the configuration file so that the project knows to submit to the cluster to run2, package the project into the Project source directory (SRC) 3, add a sentence in Java code
Configuration conf = new Configuration();
conf.set("mapreduce.job.jar", "wc.jar");
Reprint: Hadoop Cluster time synchronization
Test environment: 192.168.217.130 Master master.hadoop 192.168.217.131 node1 node1.hadoop 192.168.217.132 node2 node2.hadoopfirst, set the master server timeView local time and time zone [root@master ~]# date Mon Feb 09:54:09 CST 2017 Select time Zone [root@master ~]# tzselect [Root@master ~]# Cp/usr/shar E/zoneinfo/a
Tags: share port number USR data via SQL database my.cnf Chinese garbled problem MySQL installationMySQL installation for Hadoop---clusterOne: Two: Three: Four: Five: Six: Seven: Eight: Modify database character: Solve Chinese garbled problem, mysql default is latin1, we want to change to Utf-81> 2> Then we do modify:--> first we need to build a folder for MySQL at/etc/--and then copy/usr/sharemysql/my-medium.cof to/
nodes4) Check to see if SSH is installedSsh–version/ssh-v5) Client Creation secret keyssh-keygen-t RSA #用rsa算法产生秘钥cd. SSH #进入. SSH directoryLS #查看此目录下的文件: Id_rsa id_rsa.pubIn turn, on other clients.6) Write the master's public key to masterCP Id_rsa.pub Authorized_keysModify Permissions #root用户无需修改SSH Host name #登录验证7) write the slave public key to masterSLAVE1:SCP id_rsa.pub [email protected]:/home/hadoop/id_rsa_01.pubSlave2:scpid_rsa.pub[email prot
command:3) Cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keysThis passage means that the public key is added to the public key file for authentication, where the Authorized_keys is the public key file for authentication. At this point no password login This machine has been set up.4) You can now log in to SSH to confirm that you do not need to enter a password:~$ ssh localhost logout:~$ exit the second time login:~$ ssh localhostLog out:~$ exitThis way, you don't have to enter a password to log in
~/.ssh/id_rsa.pub [email protected]:~/.ssh(User a host operation)⑵ Append to File: Cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys (remote Server B host operation)⑶ Modify the contents of the/etc/ssh/sshd_config:Vi/etc/ssh/sshd_config(#RSAAuthentication Yes#PubkeyAuthentication Yes#AuthorizedKeysFile. Ssh/authorized_keysFind the above 3 statements and take the previous # number off)Note: Some articles say to/root/.ssh this folder and/root/.ssh/authorized_keys this file, you have to modify the p
Just beginning to contact, not very familiar with, make a small record, later revisionGenerate public and private keysSsh-keygen-t Dsa-p "-F ~/.SSH/ID_DSAImport the public key into the Authorized_keys fileCat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keysUnder normal circumstances, SSH login will not need to use the passwordIf prompted: Permission denied, please try againModify SSH configuration, path/etc/ssh/sshd_configPermitrootlogin Without-passwordChange intoPermitrootlogin YesIf the above conf
, theauthorized_keys file appears under SSH. 5. Then copy the Authorized_keys file under the. SSH directory of the first machine to the. SSH directory of the second computer, such as: SCP Authorized_keys [Email protected]:~/.ssh/6. Next to the. SSH directory of the second machine, you will find the file-authorized_keys just transferred, and then execute the command to add the second computer's public key, such as: Cat Id-rsa.pub >> Authorized_keys.7. Transfer the newly generated authorized_keys
Tags: ssh directory gen key cat did not create a public key download installationFirst detect if there is SSH1. If you do not have the download installed, you can create the. ssh folder in your home directorymkdir ~/.ssh2. Generate keySSH-KEYGEN-T RSA3. Write the current public key to the Authorized_keysCat Id-rsa.pub >> Authorized_keys4. After writing, copy the Authorized_keys to the next computer's ~/.ssh folder to overwrite5. Connect to the next computer write the public key of the next compu
assumption that SSH does not exist then it is not installed server through the sudo apt-get install openssh-server command installation can beSixth step: Solve Agent admitted failure to sign using the keyAssuming this error occurs, you need to run Ssh-add ~/.ssh/id_rsa on all nodes to add SSH to the private key. At this point, you are done. You should be able to use SSH masternode/slavenode1/slavenode2 to avoid password with each other on a random machine!!!Copyright notice: This article Bo Mas
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.