flume app

Alibabacloud.com offers a wide variety of articles about flume app, easily find your flume app information here online.

Real-time Event statistics Project: Optimizing Flume: Replacing MEM Channel with file channel

Background: Use KAFKA+FLUME+MORPHLINE+SOLR to do real-time statistics.SOLR has no data since December 23. View Log discovery because a colleague added a malformed buried point data, resulting in a lot of error.It is inferred that because the use of MEM channel is full, the message is too late to process, resulting in the loss of new data.Modify flume to use the file channel:Kafka2solr.sources =SOURCE_FROM_K

The ganglia of flume monitoring

For the log, I think the monitoring is not very meaningful, because the speed of writing is generally not particularly fast, but if it is spooldir source, inside a few grams into the data let Flume parse, especially in the combination of Kafka or other framework, monitoring is important, Can analyze the bottleneck of the entire architecture Flume's monitoring is based on JSON, through JMX to generate Metrix data, can be directly accessed through the

flume-ng-1.4.0 spooling Source method adds support for recursive detection of directories

Because Flume Spooldir does not support recursive detection of subdirectory files, and the business needs, so modify the source code, recompile Code modification Reference from: http://blog.csdn.net/yangbutao/article/details/8835563In 1.4, however, the Spoolingfilelinereader class has not been modified, but apache-flume-1.4.0-src\flume-ng-core\src\main\java

Flume 1.7.0 stand-alone version installation

Download unzip to/usr/local/FlumeConfiguring Environment variablesExport flume_home=/usr/local/flumeexport flume_conf_dir= $FLUME _home/confexport PATH=.: $PATH:: $ Flume_home/binConfigure flume-env.sh to add JDK paths in confExport JAVA_HOME=/USR/LIB/JVM/JAVA-8-OPENJDK-AMD64Flume-ng versionVerify that the installation is successful[Email protected]:~# flume-ng

Installation and use of flume

Flume after downloading, unzip, add a configuration file, write the configuration canI wrote the config file under Conf, named Flume-conf-spooldir.propertiesFlume Run Command:Bin/flume-ng agent--conf conf--conf-file conf/flume-conf-spooldir.properties--name logagent-dflume.root.logger= Debug,consolewhich-dflume.root.lo

Flume Monitoring hive log files

Flume Monitoring hive log files One: Flume Monitor hive Log 1.1 case requirements:1. 实时监控某个日志文件,将数据收集到存储hdfs 上面, 此案例使用exec source ,实时监控文件数据,使用Memory Channel 缓存数据,使用HDFS Sink 写入数据2. 此案例实时监控hive 日志文件,放到hdfs 目录当中。hive 的日志目录是hive.log.dir = /home/hadoop/yangyang/hive/logs1.2 Create a collection directory above HDFs:1.3 Copy the jar package required for flumecd /home/hadoop/yangyang/hadoop/cp -p

Flume Custom Source

Hello everyone.The company has a need. Requires Flumne to store the message from MQ to DFS, and writes the flume custom source. , as I was just touching flume. So please forgive me if there is anything wrong with you.See the source code for Flume-ng. are generally based on different scenes extends Abstractsource implements Eventdrivensource, configurableThe Mqsou

Flume Character set error when using spool source

1. Where the error is .-Geneva- + Geneva: at: to,508(pool-3-thread-1) [Error-org.apache.flume.source.spooldirectorysource$spooldirectoryrunnable.run (SpoolDirectorySource.java: the)] Fatal:spool Directory source source1: {spooldir:/home/hadoop_admin/movielog/}: Uncaught exceptioninchSpooldirectorysource Thread. Restart or reconfigure Flume to continue processing.java.nio.charset.MalformedInputException:Input length=1At java.nio.charset.CoderResult.th

[Bigdata] Flume file channel CPU consumption is higher than memory channel reason

https://www.quora.com/ Why-does-flume-take-more-resource-cpu-when-file-channel-is-used-compared-to-when-memory-channel-is-usedIn case of File channel, the CPU would is used for the following serializing/deserializing Events from/to file channel. In memory channel, the is plainly stored in RAM, so no serialization is required. A Small CPU overhead per disk write in determining the disk location where it needs to write. Typically this is ba

[Flume] Channel and sink

The client SDK of the Android log phone was completed last week and started debugging the log server this week.Use flume for log collection, and then go to Kafka. When testing, I always found out some of the event, and later learned that the use of channel and sink is wrong. When multiple sink use the same channel, the event is diverted from the common consumption, not each sink copy. Finally, change to multiple channel, each channel corresponds to a

Flume instance one, listening directory log upload to other server

I. Introduction of FLUME-NGPlease refer to the official documentation: http://flume.apache.org/FlumeUserGuide.htmlSecond, examplesRequirements Description: A directory needs to be monitored and automatically uploaded to the server and encrypted during transmission.Overall solution: N client-agent-->server-agentClient-agent: A1.sources =R1a1.channels=c1a1.sinks=K1#sourcea1.sources.r1.type=Spooldira1.sources.r1.channels=C1a1.sources.r1.basenameHead

Flume Service Management Implementation analysis

The flume can monitor and manage the running state of the component, which can be pulled automatically when the component is closed, by starting a scheduled task thread pool (Monitorservice, the maximum number of threads is 30), running the monitoring thread (monitorrunnable thread), Each 3s determines whether the state of the component (including Channel,sinkrunner) meets the requirements (the available state consists of two start and stop), calls th

Flume reading data from the Kafka

A1.sources =r1a1.sinks=K1a1.channels=C1 #使用内置kafka Sourcea1.sources.r1.type=Org.apache.flume.source.kafka.kafkasource#kafka Connected Zookeepera1.sources.r1.zookeeperConnect= localhost:2181A1.sources.r1.topic= kkt-test-topica1.sources.r1.batchSize= -A1.sources.r1.channels=C1 #这里写到hdfs中a1. Sinks.k1.channel=C1a1.sinks.k1.type=Hdfsa1.sinks.k1.hdfs.path=hdfs://Iz94rak63uyz/user/flumeA1.sinks.k1.hdfs.writeFormat =Texta1.sinks.k1.hdfs.fileType=DataStreama1.sinks.k1.hdfs.rollInterval=0a1.sinks.k1.hdfs.

Spring MVC +mybatis + kafka+flume+zookeeper distributed architecture

management solution, realize the software pipelining production, guarantee the correctness, the reliabilityGuided creation, import of projects, integrated version control (GIT/SVN), project Management (trac/redmine), Code quality (Sonar), continuous integration (Jenkins)Private deployment, unified management, for developersDistributedDistributed services: Dubbo+zookeeper+proxy+restfulDistributed message Middleware: Kafka+flume+zookeeperDistributed ca

Build real-time streaming program based on Flume+kafka+spark streaming

This course is based on the production and flow of real-time data, through the integration of the mainstream distributed Log Collection framework flume, distributed Message Queuing Kafka, distributed column Database HBase, and the current most popular spark streaming to create real-time stream processing project combat, Let you master real-time processing of the entire processing process, to reach the level of big Data intermediate research and develo

Latest version of "Hadoop" Apache Flume 1.7 Practice (unfinished, pending)

Origin: Since Hadoop is used, and because the project is not currently distributed, it is a clustered environment that causes the business log to be moved every time, and then analyzed by Hadoop.In this case, it is not as good as the previous distributed flume to work with out-of-the-box HDFs to avoid unnecessary operations. Preparation Environment: You must have a ready-to-use version of Hadoop. My version is 2.7.3. If you don't know how to install

Big Data Platform Architecture (FLUME+KAFKA+HBASE+ELK+STORM+REDIS+MYSQL)

Last time Flume+kafka+hbase+elk:http://www.cnblogs.com/super-d2/p/5486739.html was implemented.This time we can add storm:storm-0.9.5 simple configuration is as follows:Installation dependencieswget http://download.oracle.com/otn-pub/java/jdk/8u45-b14/jdk-8u45-linux-x64.tar.gztar ZXVF jdk-8u45-linux-x64.tar.gzcd jdk-8u45-linux-/etc/profileAdd the following: Export Java_home =/home/dir/jdk1. 8 . 0_45export CLASSPATH=.: $JAVA _home/jre/lib/rt.jar: $JAVA

Real-time streaming processing complete flow based on flume+kafka+spark-streaming _spark

Real-time streaming processing complete flow based on flume+kafka+spark-streaming 1, environment preparation, four test server Spark Cluster Three, SPARK1,SPARK2,SPARK3 Kafka cluster Three, SPARK1,SPARK2,SPARK3 Zookeeper cluster three, SPARK1,SPARK2,SPARK3 Log Receive server, SPARK1 Log collection server, Redis (this machine is used to do redis development, now used to do log collection test, the hostname does not change) Log collection process: Log

A little experience using flume thrift source

Recently has been using the Flume thrift source, just start to encounter a lot of problems (dependent on other programs more), after the compilation program (with g++, need-l:g++-g-dhave_netr/local/include/thrift-l/ Usr/local/lib flumethriftclient.cpp gen-cpp/flume_constants.cpp gen-cpp/flume_types.cpp gen-cpp/ Thriftsourceprotocol.cpp-o FLUMECLIENT-LTHRIFTNB-LEVENT-LTHRIFT-LRT), these two points have been summarized. After the online test found a app

Liaoliang's most popular one-stop cloud computing big Data and mobile Internet Solution Course V4 Hadoop Enterprise Complete Training: Rocky 16 Lessons (Hdfs&mapreduce&hbase&hive&zookeeper &sqoop&pig&flume&project)

master HBase Enterprise-level development and management• Ability to master pig Enterprise-level development and management• Ability to master hive Enterprise-level development and management• Ability to use Sqoop to freely convert data from traditional relational databases and HDFs• Ability to collect and manage distributed logs using Flume• Ability to master the entire process of analysis, development, and deployment of Hadoop complete projects

Total Pages: 15 1 .... 11 12 13 14 15 Go to: Go

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.