I haven't written a blog for a long time. We have recently studied storm, flume, and Kafka. Today, I will write down the scenarios and conclusions for testing flume failover and load balance;
The test environment contains five configuration files, that is, five agents.
A main configuration file, that is, the configuration file (flume-sink.properties) for configur
. –CONF-FILE Specifies the configuration of flume actual source, channel, sink, and so on.
#! /bin/bash
function shutdown ()
{
Date
Echo ' shutting down Service '
unset Service_pid # necessary in some cases
Cd/opt/${module_name}
SOURCE stop.sh
}
# # Stop Process
Cd/opt/${module_name}
Echo ' Stopping Service '
SOURCE stop.sh
# # START Process
Echo ' Starting Ser
I. Introduction of FlumeFlume is a distributed, reliable, and highly available mass-log aggregation system that enables the customization of various data senders in the system for data collection, while Flume provides the ability to simply process the data and write to various data-receiving parties (customizable).Design goal:(1) ReliabilityWhen a node fails, the log can be transmitted to other nodes without loss.
http://blog.csdn.net/weijonathan/article/details/18301321Always want to contact storm real-time computing this piece of things, recently in the group to see a brother in Shanghai Luobao wrote Flume+kafka+storm real-time log flow system building documents, oneself also followed the whole, before Luobao some of the articles in some to note not mentioned, some of the wrong points later, In this way I will do the amendment, the content should say that mos
It's been a long time, but it's a very mature architecture.General data flow, from data acquisition-data access-loss calculation-output/Storage1). Data acquisitionresponsible for collecting data in real time from each node and choosing Cloudera Flume to realize2). Data Accessbecause the speed of data acquisition and the speed of data processing are not necessarily synchronous, a message middleware is added as a buffer, using Apache's Kafka3). Flow-bas
Recently learned the use of the next flume, in line with the company will be independent of the development of the log system, the official website user manual: Http://flume.apache.org/FlumeUserGuide.htmlFlume schema A. ComponentFirst move, the structure of the Internet.As you can see from the diagram, the Flume event is defined as a data stream, a data stream consisting of an agent, which is actually a JVM
Personal opinion: Big data we all know about Hadoop, but not all of it. How do we build a large database project. For offline processing, Hadoop is still more appropriate, but for real-time, relatively strong, the amount of data is large, we can use storm, then storm and what technology collocation, to be able to do a suitable project. We can refer to the following.You can read this article with the following questions:1. What are the characteristics of a good project architecture?2. How does th
Http://www.aboutyun.com/thread-6855-1-1.htmlPersonal opinion: Big data we all know about Hadoop, but not all of it. How do we build a large database project. For offline processing, Hadoop is still more appropriate, but for real-time, relatively strong, the amount of data is large, we can use storm, then storm and what technology collocation, to be able to do a suitable project. We can refer to the following.You can read this article with the following questions:1. What are the characteristics o
1, Flume agent installation (using SPOOLDIR mode to obtain system, application and other log information)Note: Install with Jyapp userWhen a single virtual machine deploys multiple Java applications and needs to deploy multiple flume-agent for monitoring,The following configuration files need to be adjusted:The Spool_dir parameter in a flume-agent/conf/app.confJm
flow from a external source to the next D Estination (Hop).-the original sentence of the official website.As you can see, 1 agents need to have three parts: source, channel, sink, system requirements
Java Runtime Environment-Java version 1.8 or later
Memory-source, channel or receiver uses a configuration that has sufficient memory
Disk space-sufficient disk space for the configuration used by the channel or receiver
Directory permissions-read/write permissions for the dire
I am testing HDFs sink, found that the sink side of the file scrolling configuration items do not play any role, configured as follows:a1.sinks.k1.type=hdfsa1.sinks.k1.channel=c1a1.sinks.k1.hdfs.uselocaltimestamp=truea1.sinks.k1.hdfs.path=hdfs:/ /192.168.11.177:9000/flume/events/%y/%m/%d/%h/%ma1.sinks.k1.hdfs.fileprefix=xxxa1.sinks.k1.hdfs.rollinterval= 60a1.sinks.k1.hdfs.rollsize=0a1.sinks.k1.hdfs.rollcount=0a1.sinks.k1.hdfs.idletimeout=0The configur
source code into the installation directory Apache-flume-1.6.0-binTo configure environment variables:[Email protected] ~]$ vim ~/.bash_profileExport flume_home=/home/lan/apache-flume-1.6.0-bin/Export path= $PATH: $FLUME _home/binTo test whether the Flume-ng was installed successfully:Flume-ng version3.Create a new con
," + "increasing capacity, or increasing thread count") ; }
Take before also pre-judgment, if the takelist is full, indicating take operation is too slow, there is an event accumulation phenomenon, you should adjust the transaction capacitywhat happens when a transaction commits, and what does the transaction commit?? Commit is a transaction commitTwo cases:1, put the event submissionwhile (!putlist.isempty ()) { if (!queue.offer (Putlist.removefirst ())) {
From Bin/flume this shell script can see Flume starting from the Org.apache.flume.node.Application class, which is where the main function of Flume is.
The main method first resolves the shell command, throwing an exception if the specified configuration file does not exist.
According to the command contains "no-reload-conf" parameters, decide which way to load t
This article is a simple example of the flume official document in the practice and description of the official example
Http://flume.apache.org/FlumeUserGuide.html#a-simple-example
Flume's Netcat source automatically creates a socket Server that can fetch data simply by sending the data to the netcat source of this socket,flume.
Examples are as follows:
1, first configure the agent: in Flume's conf dire
{ return null; } }4. Return to a usable sinkIf a failure occurs, then look at the execution logic of the first half of the code in the process:Long now = System.currenttimemillis (); while (!failedsinks.isempty () Failedsinks.peek (). Getrefresh () Prerequisites: Failedsinks is not empty and the sink activation time of the team header is less than the current time1, poll out the queue of the first Failedsink2, using the current sink processing, if the processing is successful, then
This paper will take timestampinterceptor as an example to analyze how interceptors work in Flume.First, consider the implementation structure of the Interceptor.1. Interceptor interface is realizedThe method of the interface is defined as follows: public void Initialize (); Public event intercept (event event); Public list public void close ();/** Builder implementations must have a no-arg constructor * * Public Interface Builder extends configurable { Publ IC Interceptor Build (); }2.
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.