The previous introduction of how to use thrift source production data, today describes how to use Kafka sink consumption data.In fact, in the Flume configuration file has been set up with Kafka sink consumption dataAgent1.sinks.kafkaSink.type =Org.apache.flume.sink.kafka.KafkaSinkagent1.sinks.kafkaSink.topic=TRAFFIC_LO
Flume and Kakfa example (KAKFA as Flume sink output to Kafka topic)To prepare the work:$sudo mkdir-p/flume/web_spooldir$sudo chmod a+w-r/flumeTo edit a flume configuration file:$ cat/home/tester/flafka/spooldir_kafka.conf# Name The components in this agentAgent1.sources = WeblogsrcAgent1.sinks = Kafka-sinkAgent1.channels = Memchannel# Configure The sourceAgent1.s
1. Create a Agent,sink type to be specified as a custom sinkVi/usr/local/flume/conf/agent3.confAgent3.sources=as1Agent3.channels=c1Agent3.sinks=s1Agent3.sources.as1.type=avroagent3.sources.as1.bind=0.0.0.0agent3.sources.as1.port=41414Agent3.sources.as1.channels=c1Agent3.channels.c1.type=memoryAgent3.sinks.s1.type=storm.test.kafka.testkafkasinkAgent3.sinks.s1.channel=c12. Create custom Kafka
序列化方式Agent.sinks.k1.serializer.class=kafka.serializer.stringencoderAgent.sinks.k1.channel=c1Start flumeFlume-ng agent-c conf-f/home/bigdata/flumeconf/log_kafka.log-n agent-dflume.root.logger=info,console
Start Kafka
./kafka-server-start.sh-daemon. /config/server.propertiesIf the topic is not created, the new topicbin/kafka-topics.sh--create--zookeep
final Logger log = Loggerfactory.getlogger (cmcckafkasink.class);p ublic static Final string key_hdr = "KEY";p ublic static final String topic_hdr = "TOPIC";p rivate static final String CHARSET = "UTF-8"; Private Properties kafkaprops;private producerThen mvn clean install to compile the package jar, throw the jar package into the Flume installation directory of the Lib can be, the following is the editing conf fileOf course, the key of the specific attribute in the Conf file is consistent with
Flume acquisition Process:#说明: The case is Flume listening directory/home/hadoop/flume_kafka acquisition to Kafka;Start the clusterStart Kafka,Start the agent,Flume-ng agent-c. -f/home/hadoop/flume-1.7.0/conf/myconf/flume-kafka.conf-n A1-dflume.root.logger=info,consoleOpen Consumerkafka-console-consumer.sh--zookeeper hdp-qm-01:2181--from-beginning--topic mytopicProduction data to KafkaData Catalog:Vi/home/h
concepts.Second, Kafka monitoringI'm going to discuss Kafka's monitoring from five dimensions. The first is to monitor the host where the Kafka cluster resides, and the second is to monitor the performance of the Kafka broker JVM; 3rd, we want to monitor the performance of Kafka broker; and we want to monitor the perf
How to Use CSS to sink the first word and use css to sink
How to Use CSS to sink the first word:Suggestion: writing code as much as possible can effectively improve learning efficiency and depth.The sinking effect of the first word may not be common in practical applications, but it is not because such an application is not used. The effect is indeed quite brill
Basic regular expression character sink, regular expression character sink
^ Word ---- meaning: the string to be searched (word) at the beginning of the lineExample: Search for the row whose first line is # and list the row numberGrep-n' ^ # 'test.txtWord $ ---- meaning: the string to be searched (word) at the end of the rowExample: end the line! Print the row and list the row number.Grep-n '! $ 'Test.txt.
Most people who have used word know that word has a sinking function inArticleSometimes it can add a lot of colors to the article.
Today, we use CSS to simulate the first word sinking function of word, so we don't need to modify it.CodeOnly use the pseudo element "first-letter" of CSS to sink the first character.
Original link: Kafka combat-flume to KAFKA1. OverviewIn front of you to introduce the entire Kafka project development process, today to share Kafka how to get the data source, that is, Kafka production data. Here are the directories to share today:
Data sources
Flume to
Kafka Connector and Debezium
1. Introduce
Kafka Connector is a connector that connects Kafka clusters and other databases, clusters, and other systems. Kafka Connector can be connected to a variety of system types and Kafka, the main tasks include reading from
Recently want to test the performance of Kafka, toss a lot of genius to Kafka installed to the window. The entire process of installation is provided below, which is absolutely usable and complete, while providing complete Kafka Java client code to communicate with Kafka. Here you have to spit, most of the online artic
The Kafka can implement data collection and write to a variety of LOTP databases through custom sink, and the following example is written to a distributed K-v database Aerospike through custom source implementation data. 1. Custom sink code is as follows
Package KAFKA_SINK.ASD;
Import java.io.IOException;
Import java.net.ConnectException;
Import java.util.
Kafka ---- kafka API (java version), kafka ---- kafkaapi
Apache Kafka contains new Java clients that will replace existing Scala clients, but they will remain for a while for compatibility. You can call these clients through some separate jar packages. These packages have little dependencies, and the old Scala client w
The client SDK of the Android log phone was completed last week and started debugging the log server this week.Use flume for log collection, and then go to Kafka. When testing, I always found out some of the event, and later learned that the use of channel and sink is wrong. When multiple sink use the same channel, the event is diverted from the common consumptio
, etc.Sink : Send data to destinations such as HDFs, hbase, etc.Flume The basic unit of the transmitted data is the event, and the transaction is guaranteed at the event level, and the event encapsulates the transmitted dataThe channel will only delete the temporary data after the sink has successfully sent the data in the channel, which guarantees the reliability and security of the data transmission.4. Generalized usage of flumeFlume supports multi-
1. Background information
Many of the company's platforms generate a large number of logs (typically streaming data, such as the PV of search engines, queries, etc.), which require a specific log system, which in general requires the following characteristics:
(1) Construct the bridge of application system and analysis system, and decouple the correlation between them;
(2) support the near real-time on-line analysis system and the off-line analysis system similar to Hadoop;
(3) with high scalabi
A scheme of log acquisition architecture based on Flume+log4j+kafkaThis article will show you how to use Flume, log4j, Kafka for the specification of log capture.Flume Basic ConceptsFlume is a perfect, powerful log collection tool, about its configuration, on the internet there are many examples and information available, here only to do a simple explanation is no longer detailed.The flume contains the three most basic concepts of source, Channel, and
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.