[TOC]Non-clustered configurationThis situation is not cluster configuration, relatively simple, you can directly refer to my collation of the "Flume notes", the basic structure of the following:Flume multiple agents of a cluster a source structure descriptionThe structure diagram is as follows:The description is as follows:即可以把我们的Agent部署在不同的节点上,上面是两个Agent的情况。其中Agent foo可以部署在日志产生的节点上,比如,可以是我们web服务器例如tomcat或者nginx的节点上,foo的source可以配置为监控日志文件数据的变化,channel则
This article describes the initial process of using flume to transfer data to MongoDB, covering environment deployment and considerations.1 Environment Constructionrequires JDK, flume-ng, MongoDB java driver, Flume-ng-mongodb-sink(1) jdk:http://www.oracle.com/technetwork/java/javase/downloads/jdk8-downloads-2133151.html(2) flune-ng:http://www.apache.org/dyn/close
Flume custom hbasesink class
Reference (to the original author) http://ydt619.blog.51cto.com/316163/1230586Https://blogs.apache.org/flume/entry/streaming_data_into_apache_hbaseSample configuration file of flume 1.5
# Name the components on this agenta1.sources = r1a1. sinks = k1a1. channels = c1 # Describe/configure the sourcea1.sources. r1.type = spooldira1.sour
Flume integrated Kafka:flume capture business log, sent to Kafka installation deployment KafkaDownload1.0.0 is the latest release. The current stable version was 1.0.0.You can verify your download by following these procedures and using these keys.1.0.0
Released November 1, 2017
Source download:kafka-1.0.0-src.tgz (ASC, SHA512)
Binary Downloads:
Scala 2.11-kafka_2.11-1.0.0.tgz (ASC, SHA512)
Scala 2.12-kafka_2
Flume, as a Log collection tool, exhibits a very powerful capability in data collection. Its source, SINK, channel three components of this mode, to complete the data reception, caching, sending this process, has a very perfect fit. But here, we want to say is not flume how good or flume have what merit, we want to talk about is
Label:Original: http://mp.weixin.qq.com/s?__biz=MjM5NzAyNTE0Ng==mid=205526269idx=1sn= 6300502dad3e41a36f9bde8e0ba2284dkey= C468684b929d2be22eb8e183b6f92c75565b8179a9a179662ceb350cf82755209a424771bbc05810db9b7203a62c7a26ascene=0 uin=mjk1odmyntyymg%3d%3ddevicetype=imac+macbookpro9%2c2+osx+osx+10.10.3+build (14D136) version= 11000003pass_ticket=hkr%2bxkpfbrbviwepmb7sozvfydm5cihu8hwlvne78ykusyhcq65xpav9e1w48ts1 Although I have always disapproved of the full use of open source software as a system,
http://blog.csdn.net/weijonathan/article/details/18301321Always want to contact storm real-time computing this piece of things, recently in the group to see a brother in Shanghai Luobao wrote Flume+kafka+storm real-time log flow system building documents, oneself also followed the whole, before Luobao some of the articles in some to note not mentioned, some of the wrong points later, In this way I will do the amendment, the content should say that mos
1) Introduction
Flume is a distributed, reliable, and highly available system for aggregating massive logs. It supports customization of various data senders in the system for data collection. Flume also provides simple data processing, and write the capabilities of various data receivers (customizable.
Design goals:(1) ReliabilityWhen a node fails, logs can be transferred to other nodes without being lost.
It's been a long time, but it's a very mature architecture.General data flow, from data acquisition-data access-loss calculation-output/Storage1). Data acquisitionresponsible for collecting data in real time from each node and choosing Cloudera Flume to realize2). Data Accessbecause the speed of data acquisition and the speed of data processing are not necessarily synchronous, a message middleware is added as a buffer, using Apache's Kafka3). Flow-bas
Flume is an excellent data acquisition component, some heavyweight, its nature is based on the query results of SQL statements assembled into OPENCSV format data, the default separator symbol is a comma (,), you can rewrite opencsv some classes to modify
1, download
[Root@hadoop0 bigdata]# wget http://apache.fayea.com/flume/1.6.0/apache-flume-1.6.0-bin.tar.gz
2
Implementation Architecture
A scenario implementation architecture is shown in the following illustration:
Analysis of 3.1 producer layer
Service assumptions within the PAAs platform are deployed within the Docker container, so to meet non-functional requirements, another process is responsible for collecting logs, thus not intruding into service frameworks and processes. Using flume ng for log collection, this open source component is very powerful
Flume real-time crawl log data and upload to Kafka
1.Linux OK zookeeper is configured, start zookeeper first
sbin/zkserver.sh start
(sbin/zkserver.sh Status View startup state) JPS can check to see Le process as Quorumpeermain
2. Start Kafka,zookeeper need to start before Kafka
bin/kafka-server-start.sh Config/server.properties
3. Start consumer receive log
bin/kafka-console-consumer.sh--zookeeper localhos
flume Installation and configuration:
Download flume, and then unpack:
Tar xvf apache-flume-1.5.2-bin.tar.gz-c./
Configure Flume, under Conf/flume-conf.properties (not created, anyway template):
# example.conf:a Single-node Flume
1, Flume agent installation (using SPOOLDIR mode to obtain system, application and other log information)Note: Install with Jyapp userWhen a single virtual machine deploys multiple Java applications and needs to deploy multiple flume-agent for monitoring,The following configuration files need to be adjusted:The Spool_dir parameter in a flume-agent/conf/app.confJm
Flume as a log acquisition system, has a unique application and advantages, then flume in the actual application and practice in the end what is it? Let us embark on the Flume road together.1. what is Apache Flume(1) Apache Flume is simply a high-performance, distributed l
Questions Guide:1.flume-ng and Scribe, where is the advantage of Flume-ng?2. What issues should be considered in architecture design considerations?3.Agent How can I fix it?Does 4.Collector panic have an impact?What are the measures for 5.flume-ng reliability (reliability)?The U.S. mission's log collection system is responsible for the collection of all business
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.