1) hostname error:
2011-11-14 11:44:55,497 ERROR com.cloudera.util.NetUtils: Unable to get canonical host name! test: test java.net.UnknownHostException: test: test at java.net.InetAddress.getLocalHost(InetAddress.java:1354) at com.cloudera.util.NetUtils.
Error cause: IP address cannot be obtained from hostname
Solution: add the host name to IP address ing in the/etc/hosts file.
2) Java does not follow
line 234: exec: java: not found
Error cause: the Java command does not exist.
In the distributed system, each machine has the local log that the program runs, sometimes in order to analyze the demand, have to these scattered log summary requirements, I believe many people will choose RSYNC,SCP, but they are not strong in real-time, but also bring the problem of name conflict. The scalability is not satisfactory, not elegant at all.In reality, we are confronted with the need to summarize the Nginx logs of multiple servers on the line in real time.
In a complete large data processing system, in addition to the core of the Hdfs+mapreduce+hive composition Analysis system, data acquisition, result data export, task scheduling and other indispensable auxiliary systems are needed, and these auxiliary tools are There is a convenient open source framework in the Hadoop ecosystem. Log capture framework FlumeFlume is a distributed, reliable, and highly available system for collecting, aggregating, and transmitting large volumes of logs.
Download apache-flume-1.7.0-bin.tar.gz, withTar -zxvfUnzip, add the settings in the/etc/profile file:Export Flume_home=/opt/apache-flume-1.7.0-binexport path= $PATH: $FLUME _home/binModify the two files under $flume_home/conf/and increase the java_home in flume-env.sh:java_home=/opt/jdk1.8.0_121Most importantly, modify
{ //no event, that is Backoffresult =Status.backoff; } //Commit a transactionTransaction.commit (); } Catch(Exception ex) {//rolling back a transactionTransaction.rollback (); Throw NewEventdeliveryexception ("Failed to log event:" +event, ex); } finally { //Close TransactionTransaction.close (); } returnresult; } } 3. Pack and place in/soft/flume/Lib under4, using the custom s
1. Background introduction Many of the company's platforms generate a large number of logs per day (typically streaming data, for example, the search engine PV, query, etc.), the processing of these logs requires a specific log system, in general, these systems need to have the following characteristics: (1) The construction of application systems and analysis systems of the bridge, and the correlation between them decoupling (2) support for near real-time online analysis system and off-line ana
1. Background information
Many of the company's platforms generate a large number of logs (typically streaming data, such as the PV of search engines, queries, etc.), which require a specific log system, which in general requires the following characteristics:
(1) Construct the bridge of application system and analysis system, and decouple the correlation between them;
(2) support the near real-time on-line analysis system and the off-line analysis system similar to Hadoop;
(3) with high scalabi
The recent project team has the need to tap the stream log to collect, learn a bit flume and install successfully. The relevant information to record a bit.1) Download flume1.5 versionwget http://www.apache.org/dyn/closer.cgi/flume/1.5.0.1/apache-flume-1.5.0.1-bin.tar.gz2) Unzip the flume1.5TAR-ZXVF apache-flume-1.5.0.
This article introduces flume data insert hdfs and common directory (), this article continues to introduce flume-ng to insert data into the hbase-0.96.0.
First, modify the flume-node.conf file in the conf directory under the flume folder in node (for the original configuration, refer to the above) and make the followi
Unify the time before building, turn off the firewall, use the jar package version is 1.6.0There are two ways to configure a serviceThe first type: The following steps:1. Pass the jar package to the Node1 and extract it to the root directory2. Change the directory name by using the following command: MV apache-flume-1.6.0-bin/home/install/flume-1.63. After entering the
scribe, Chukwa, Kafka, flume log System comparison1. Background informationMany of the company's platforms generate a large number of logs per day (typically streaming data, such as search engine PV, queries, etc.), processing these logs requires a specific logging system, in general, these systems need to have the following characteristics: (1) Build the bridge of application system and analysis system, and decouple the association between them; (2)
Flume compared with Logstash, the personal experience is as follows:
Logstash more emphasis on the preprocessing of the field, while flume emphasis on data transmission;
Logstash has dozens of plug-ins, flexible configuration, Flume is to emphasize the user's custom development (source and sink kind also has ten or twenty, the channel is relatively s
the high-level interface, which hides the details of the broker, allowing consumer to push data from the broker without having to care about the network topology.
More importantly, for most log systems, the data information that consumer has acquired is saved by the broker, while in Kafka, the data information is maintained by consumer itself.
Cloudera's Flume Flume is Cloudera's Open source log
1, download the latest flume on the official website of Flumewget http://124.205.69.169/files/A1540000011ED5DB/mirror.bit.edu.cn/apache/flume/1.6.0/ apache-flume-1.6.0-bin.tar.gz 2. Solve Flume installation packagecd/export/software/TAR-ZXVF apache-flume-1.6.0-bin.tar.gz-c/e
The project requires C + + code to interface with the Flume, which in turn writes the log to HDFs.Flume native to Java code, the original solution was to invoke the Flume Java method via JNI.But because of the concern about the efficiency of JNI calls, and the fact that the C + + call JNI needs to take care of the local reference and GC issues, the headache has been caused.Rage, rewrite the code, use C + +
How to do integration, in fact, especially simple, online is actually a tutorial.http://blog.csdn.net/fighting_one_piece/article/details/40667035 look here.I'm using the first integration. When you do, there are a variety of problems. Probably from from 2014.12.17 5 o'clock in the morning to 2014.12.17 night 18 o'clock 30 summed up in fact very simple, but do a long time AH Ah!!! This kind of thing, a fall into your wit. Question 1, need to refer to a variety of packages, these packages to bre
Today's meeting to discuss why log processing uses both Flume and Kafka, is it possible to use only Kafka without Flume? The idea was to use only the Flume interface, whether it is the input interface (socket and file) and the output interface (kafka/hdfs/hbase, etc.).Consider a single scenario, and from a simplified system perspective, it might be better to use
Centos6.5 install flume, centos6.5flume
Flume is installed here because it is used for game Business Log collection and analysis.
1. Install the java environmentRpm-ivh jdk-8u51-linux-x64.rpmPreparing... ######################################## ### [100%]1: jdk1.8.0 _ 51 ##################################### ###### [100%]Unpacking JAR files...Rt. jar...Jsse. jar...Charsets. jar...Tools. jar...Localedata. ja
From: http://flume.apache.org/FlumeUserGuide.html#data-flow-model
Learn flume through translation.Introduction
Apache flume is a distributed, highly reliable, and highly available system. It is mainly used to efficiently collect, aggregate, and move a large amount of log data from various data sources.
The collected data is stored in a centralized manner.
The application scenarios of Apache
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.