and run the following command:
$./Flume-ng agent -- conf ../conf-f ../conf/flume. conf -- n agent-Dflume. root. logger = INFO, console
You can view the generated log files in the/data/flume directory.
2. combine kafka
Because flume1.5.2 does not have a kafka sink, you need to develop your own kafka sink.
You can refer to the
Kafka is a distributed publish-subscribe messaging system. It is originally developed at LinkedIn and became a Apache project in July, 2011. Today, Kafka is used by LinkedIn, Twitter, and Square for applications including log aggregation, queuing, and real time m Onitoring and event processing.In the upcoming version 0.8 release, Kafka'll support intra-cluster re
Kafka provides two sets of APIs to consumer
The high-level Consumer API
The Simpleconsumer API
the first highly abstracted consumer API, which is simple and convenient to use, but for some special needs we might want to use the second, lower-level API, so let's start by describing what the second API can do to help us do it .
One message read multiple times
Consume only a subset of the messages in a process partition
Secrets of Kafka performance parameters and stress tests
The previous article Kafka high throughput performance secrets introduces how Kafka is designed to ensure high timeliness and high throughput. The main content is focused on the underlying principle and architecture, belongs to the theoretical knowledge category. This time, from the perspective of applicati
Author: Wang, JoshI. Basic overview of Kafka1. What is Kafka?The definition of Kafka on the Kafka website is called: adistributed publish-subscribe messaging System. Publish-subscribe is the meaning of publishing and subscribing, so it is accurate to say that Kafka is a message subscription and release system. Initiall
Kafka based on the 0.8.0 version of the command usage:
See topic Distribution kafka-list-topic.sh# bin/kafka-list-topic.sh--zookeeper 192.168.197.170:2181,192.168.197.171:2181 (List of all topic partitions)
# bin/kafka-list-topic.sh--zookeeper 192.168.197.170:2181,192.168.
Kafka How to read the offset topic content (__consumer_offsets)
As we all know, since zookeeper is not suitable for frequent write operations in large quantities, the new version Kafka has recommended that consumer's displacement information be kept in topic within Kafka, __consumer_offsets topic, and by default
o.a.kafka.common.metrics.metrics-added sensor with name Batch-size09:47:00.699 [main] DEBUG o.a.kafka.common.metrics.metrics-added sensor with name Compression-rate09:47:00.701 [main] DEBUG o.a.kafka.common.metrics.metrics-added sensor with name Queue-time09:47:00.702 [main] DEBUG o.a.kafka.common.metrics.metrics-added sensor with name Request-time09:47:00.702 [main] DEBUG o.a.kafka.common.metrics.metrics-added sensor with name Produce-throttle-time09:47:00.702 [main] DEBUG o.a.kafka.common.met
Introduced
Elk is the industry standard log capture, storage index, display analysis System solutionLogstash provides flexible plug-ins to support a variety of input/outputMainstream use of Redis/kafka as a link between log/messageIf you have a Kafka environment, using Kafka is better than using RedisHere is one of the simplest configurations to make a note, Ela
1. OverviewAt present, the latest version of the Kafka official website [0.10.1.1], has been defaulted to the consumption of offset into the Kafka a topic named __consumer_offsets. In fact, back in the 0.8.2.2 Version, the offset to topic is supported, but the default is to store the offset of consumption in the Zookee
1, Installation Zookeeper
2, Installation Kafka
Step 1: Download Kafka Click to download the latest version and unzip it.
tar-xzf kafka_2.10-0.8.2.1.tgz
CD kafka_2.10-0.8.2.1
Step 2: Start the serviceKafka used to zookeeper, all start Zookper First, the following simple to enable a single-instance Zookkeeper service. You can add a symbol at the end of the comma
Use the kafka-clients operation kafka is always unsuccessful, the reasons are unclear, the following posted related code and configuration, please know how to guide, thank you!Environment and dependenceJDKVersion 1.8, Kafka version 2.12-0.10.2.0 , server use CentOS-7 build.Test code
Testbase.java
public c
In the Kafka core principle of information, there are many online, but if you do not study its source code, always know it but do not know why. Here's how to compile the Kafka source code in the Windows environment, and build the Kafka source environment through the IntelliJ Idea development tool to facilitate local debug debugging to study Kafka's internal imple
Kafka Single-Machine deploymentKafka is a high-throughput distributed publish-subscribe messaging system, Kafka is a distributed message queue for log processing by LinkedIn, with large log data capacity but low reliability requirements, and its log data mainly includes user behaviorEnvironment configuration: CentOS Release 6.3 (Final) JDK version: Jdk-6u31-linux
Import Kafka source code to Scala IDE and kafkascala
After one night of tossing, I finally went to Scala IDE (Eclipse and Sacla plug-in) to view the source code of the Apache Kafka project.
My environment is: win7 32-bit, Scala IDE: 4.0.0, Apache Kafka: 0.8.1.1 (added a gradlew. bat file in version 0.8.2)
After downloa
one consumer within the same consumer group, but multiple consumer group can consume the message simultaneously.
Architecture:A typical Kafka cluster contains a number of producer (either Page view generated from the Web front-end, or server logs, System CPUs, memory, etc.), several broker (Kafka support level extensions, more general broker numbers, The higher the cluster throughput rate, a number of cons
1, preparation work 1.1, machine preparationserver1:10.40.33.11server2:10.40.33.12server3:10.40.33.131.2, port occupancy situationzookeeper:2181,3888,4888kafka:90921.3. Software PreparationJDK1.7.0_51 (latest version of kafka-0.8.2.1 recommended to use 1.7 and later versions of JDK) zookeeper3.4.5 (and above) kafka_2.11-0.8.2.1 (latest version)2, installation 2.1
:\program Files (x86) \java\jre1.8.0_60 (this is the default installation path, if you change the installation directory during installation, fill in the changed path)
PATH: Add after existing value "; %java_home%\bin "
1.3 Open cmd Run "java-version" to view the current system Java version:2. Installing zookeeperThe Kafka
used by the producer. However, after version 0.8.0, the producer no longer connects to the broker through zookeeper, but through brokerlist (192.168.0.1: 9092,192.168 .0.2: 9092,192.168 .0.3: 9092 configuration, directly connected to the broker, as long as it can be connected to a broker, it can get information on other brokers in the cluster, bypassing zookeeper.2. Start the kafka serviceKafka-server-star
Recently used in the project to Kafka, recorded
Kafka role, here do not introduce, please own Baidu. Project Introduction
Briefly introduce the purpose of our project: The project simulates the exchange, carries on the securities and so on the transaction, in the Matchmaking transaction: Adds the delegate, updates the delegate, adds the transaction, adds or updates the position, will carry on the database o
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.