OverviewThis time spent part of the time processing the message bus and log docking. Here to share some of the problems encountered in log collection and log parsing and processing scenarios.
Log capture-flumelogstash VS flumeFirst, let's talk about our selection on the log collector. Since we chose to use Elasticsearch as a log of storage with search engines. And based on the Elk (Elasticsearch,logstash,kibana) technology stack in the direction of the log system is so popular, so the Logstash
Recently, an ELK architecture is used for log collection. the intermediate data collection is changed from logstash to flume. The following is the installation of flume: because flume and Elasticsearch are both developed in java, so the java is deployed before installation, ES does not support java1.7, because there is a major bug, so choose jdk-8u51-linux-x64.rp
1. Create a Agent,sink type to be specified as a custom sinkVi/usr/local/flume/conf/agent3.confAgent3.sources=as1Agent3.channels=c1Agent3.sinks=s1Agent3.sources.as1.type=avroagent3.sources.as1.bind=0.0.0.0agent3.sources.as1.port=41414Agent3.sources.as1.channels=c1Agent3.channels.c1.type=memoryAgent3.sinks.s1.type=storm.test.kafka.testkafkasinkAgent3.sinks.s1.channel=c12. Create custom Kafka Sink (custom Kafka sink packaging is the producer of Kafka),
Tag: Connect a storage span through the self-starter installation package StrongOverview
Flume is a distributed, reliable, and highly available system for collecting, aggregating, and transmitting large volumes of logs.
Flume can collect files,socket packets and other forms of source data, but also can export the collected data to HDFS,hbase , Many external storage systems such as Hive, Kafka,
Overview1-flume IntroductionSystem Requirements3-Installation and configuration4-Start and testI. Introduction to FlumeWebsite address: http://flume.apache.org/1-OverviewFlume is a distributed, reliable, and usable service for efficiently collecting, summarizing, and moving large amounts of log data. It has a simple and flexible architecture based on streaming data streams. It has a reliable mechanism of reliability and many failover and recovery me
Flume OutOfMemoryError ErrorRunning flume not long to report the following exception:2016-08-24 17:35:58,927 (Flume Thrift IPC Thread 8) [ERROR- Org.apache.flume.channel.ChannelProcessor.processEventBatch (channelprocessor.java:196)] Error while writing to Required channel:org.apache.flume.channel.memorychannel{name:memorychannel}2016-08-24 17:35:59,332 (sinkrunn
1.installationJdkrefer to the installation of the JDK here. 2.installationFlume2.1. DownloadFlume:http://flume.apache.org/download.html650) this.width=650; "Src=" https://s5.51cto.com/oss/201710/25/ Da9277a9d433278d21a9ccdef349d90a.png-wh_500x0-wm_3-wmp_4-s_3707767358.png "title=" 1.png "alt=" Da9277a9d433278d21a9ccdef349d90a.png-wh_ "/>Click the link: apache-flume-1.7.0-bin.tar.gz download. 2.2. Unpacking the installation package$ tar zxvf apache-
Https://www.ibm.com/developerworks/cn/opensource/os-cn-kafka/index.htmlKafka and Flume Many of the functions are really repetitive. Here are some suggestions for evaluating the two systems:
Kafka is a general-purpose system. You can have many producers and consumers to share multiple themes. Conversely, Flume is designed to work for a specific purpose and is sent specifically to HDFS and HBase.
A/Flume data flow model
Flume event is defined as a data flow unit with byte payload and optional string properties, and the Flume agent is the JVM process that hosts the components of an event from the external source to the next destination. The following figure is the flume agent flowchart
1. Development environment 1.1. Package Download 1.1.1. JDKHttp://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.htmlInstall to the D:\GreenSoftware\Java\Java8X64\jdk1.8.0_91 directory 1.1.2. Mavenhttps://maven.apache.org/download.cgiUnzip to the D:\GreenSoftware\apache-maven-3.3.9 directory 1.1.3. Scalahttps://www.scala-lang.org/download/Unzip to the D:\GreenSoftware\Java\scala-2.12.6 directory 1.1.4. ThriftHttp://thrift.apache.org/downloadPlace the downloaded Thrift-0.
The U.S. mission's log collection system is responsible for the collection of all business logs from the company and provides real-time data streams to the Hadoop platform for offline data and storm platforms. The American mission's log collection system is based on flume design and construction."Flume-based Log collection system" will be divided into two parts for readers to present the United States of Am
The recent project team has the need to tap the stream log to collect, learn a bit flume and install successfully. The relevant information to record a bit.1) Download flume1.5 versionwget http://www.apache.org/dyn/closer.cgi/flume/1.5.0.1/apache-flume-1.5.0.1-bin.tar.gz2) Unzip the flume1.5TAR-ZXVF apache-flume-1.5.0.
Original link: Kee flume-ng some precautionsHere only to consider some of the flume itself, for the JVM, HDFS, HBase and so on are not involved ....First, about Source:1, Spool-source: Suitable for static files, that is, the file itself is not dynamic change;2. Avro source can increase the number of threads appropriately to improve this source performance;3, Thriftsource in the use of a problem to note that
First of all, Flume and Kafka are message systems , but they also have a lot of different places, flume more toward the message acquisition system, and Kafka more toward the message cache system. The difference in "one" designFlume is a message acquisition system, which mainly solves the problem is the multiple collection of messages. As a result, Flume provides
Flume is a real-time message collection system, it defines a variety of source, channel, sink, can be selected according to the actual situation.Flume Download and Documentation:http://flume.apache.org/KafkaKafka is a high-throughput distributed publish-subscribe messaging system that has the following features:
Provides persistence of messages through the disk data structure of O (1), a structure that maintains long-lasting performance even
Personal opinion: Big data we all know about Hadoop, but not all of it. How do we build a large database project. For offline processing, Hadoop is still more appropriate, but for real-time, relatively strong, the amount of data is large, we can use storm, then storm and what technology collocation, to be able to do a suitable project. We can refer to the following.You can read this article with the following questions:1. What are the characteristics of a good project architecture?2. How does th
Http://www.aboutyun.com/thread-6855-1-1.htmlPersonal opinion: Big data we all know about Hadoop, but not all of it. How do we build a large database project. For offline processing, Hadoop is still more appropriate, but for real-time, relatively strong, the amount of data is large, we can use storm, then storm and what technology collocation, to be able to do a suitable project. We can refer to the following.You can read this article with the following questions:1. What are the characteristics o
1. Flume Create configuration file Flume-spark-tail-conf.properties# The configuration file needs to define the sources, # the channels and the sinks.# Sources, channels and sinks are defined per agent, # in this case called ‘agent‘a2.sources = r2a2.channels = c2a2.sinks = k2### define sourcesa2.sources.r2.type = execa2.sources.r2.command = tail -F /opt/datas/spark_word_count.loga2.sources.r2.shell = /bin/b
I blog article if not specifically noted are original! If reproduced please specify the source: http://blog.csdn.net/yanghua_kobe/article/details/46595401Continuing the chat log system, the previous it has mentioned that our selection on the log collection is Flume-ng. The application logs the log to its own log file or to the specified folder (log files are scrolled by day), and then uses the Flume agent t
This article describes Flume (spooling Directory source) + HDFS, and some of the source details in Flume are described in the article http://www.cnblogs.com/cnmenglang/p/6544081.html1. Material Preparation: apache-flume-1.7.0-bin.tar.gz2. Configuration steps:A. Upload to User (LZ user MFZ) directory under ResourcesB. UnzipTAR-XZVF apache-
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.