Kafka Common Commands
The following is a summary of Kafka common command line:
1. View topic Details
./kafka-topics.sh-zookeeper 127.0.0.1:2181-describe-topic TestKJ1
2. Add a copy for topic
./kafka-reassign-partitions.sh-zookeeper 127.0.0.1:2181-reassignment-json-file Json/partitions-to-move.json- Execute
3. Create To
Scala, can run on the JVM, so we do not need to separate to build the Scala environment, will be related to programming when we say how to configure the Scala problem, here is not used, Of course you need to know that this is running on Linux. Second, I use the latest version of 0.7.2, you download the Kafka you can open its directory to browse:I'm not going to tell you what's in each package, and I'm going to focus on the fact that you can only find
From: http://doc.okbase.net/QING____/archive/19447.htmlAlso refer to:http://blog.csdn.net/21aspnet/article/details/19325373Http://blog.csdn.net/unix21/article/details/18990123Kafka as a distributed log collection or system monitoring service, it is necessary for us to use it in a suitable situation. The deployment of Kafka includes the Zookeeper environment/kafka environment, along with some configuration o
the Kafka cluster configuration typically has three methods , namely
(1) Single node–single broker cluster;
(2) Single node–multiple broker cluster;(3) Multiple node–multiple broker cluster.
The first two methods of the official network configuration process ((1) (2) To configure the party Judges Network Tutorial), the following will briefly introduce the first two methods, the main introduction to the last method.
preparatory work:
1.
before installing Apache Kafka. Because this article mainly introduces the working principle of Apache Kafka, so how to install and use zookeeper content will no longer repeat, not clear readers can refer to my another article: "Hadoop series: Zookeeper (1)-- Zookeeper single point and cluster Installation ". Here we run zookeeper just using the Zookeeper servic
Background:Various Application Systems in today's society, such as business, social networking, search, and browsing, constantly produce information like information factories. In The Big Data era, we are faced with the following challenges:
How to collect this huge information
How to analyze it
How to implement the above two points in a timely manner
These challenges form a business demand model, that is, information about producer production (produce) and consumer consumption (consume) (pr
Kafka is a distributed data stream platform, which is commonly used as message delivery middleware. This article describes the use of Kafka, with Linux as an example (the Windows system simply changes the following command "bin/" to "bin\windows\", the script extension ". sh" to ". Bat") and is suitable for beginners who have just contacted Kafka and zookeeper. O
The first part constructs the Kafka environment
Install Kafka
Download: http://kafka.apache.org/downloads.html
Tar zxf kafka-
Start Zookeeper
You need to configure config/zookeeper.properties before starting zookeeper:
Next, start zookeeper.
Bin/zookeeper-server-start.sh config/zookeeper.properties
Start
The previous introduction of how to use thrift source production data, today describes how to use Kafka sink consumption data.In fact, in the Flume configuration file has been set up with Kafka sink consumption dataAgent1.sinks.kafkaSink.type =Org.apache.flume.sink.kafka.KafkaSinkagent1.sinks.kafkaSink.topic=TRAFFIC_LOGagent1.sinks.kafkaSink.brokerList=10.208.129.3:9092,10.208.129.4:9092,10.208.129.5:9092ag
1, Kafka is what.
Kafka, a distributed publish/subscribe-based messaging system developed by LinkedIn, is written in Scala and is widely used for horizontal scaling and high throughput rates.
2. Create a background
Kafka is a messaging system that serves as the basis for the activity stream of LinkedIn and the Operational Data Processing pipeline (Pipeline). Act
To start the Kafka service:
bin/kafka-server-start.sh Config/server.properties
To stop the Kafka service:
bin/kafka-server-stop.sh
Create topic:
bin/kafka-topics.sh--create--zookeeper hadoop002.local:2181,hadoop001.local:2181,hadoop003.local:2181-- Replication-facto
Brief introductionThis article describes how to configure and launch Apache Kafka on Windows OS, which will guide you through the installation of Java and Apache Zookeeper.Apache Kafka is a fast and extensible message queue that can handle heavy read-write workloads, such as IO-related work. For more information, see http://kafka.apache.org. Because zookeeper can provide reliable distributed coordination se
The MAVEN components are as follows: org.apache.spark spark-streaming-kafka-0-10_2.11 2.3.0The official website code is as follows:Pasting/** Licensed to the Apache software Foundation (ASF) under one or more* Contributor license agreements. See the NOTICE file distributed with* This work for additional information regarding copyright ownership.* The ASF licenses this file to under the Apache License, Version 2.0* (the "License"); You are no
Build a kafka cluster environment in a docker container
Kafka cluster management and status saving are implemented through zookeeper. Therefore, you must first set up a zookeeper cluster.
Zookeeper cluster Construction
I. software environment:
The zookeeper cluster requires more than half of the nodes to survive for external services. Therefore, the number of servers should be 2 * N + 1. Here, three nodes
ERROR Log event analysis in kafka broker: kafka. common. NotAssignedReplicaException,
The most critical piece of log information in this error log is as follows, and most similar error content is omitted in the middle.
[2017-12-27 18:26:09,267] ERROR [KafkaApi-2] Error when handling request Name: FetchRequest; Version: 2; CorrelationId: 44771537; ClientId: ReplicaFetcherThread-2-2; ReplicaId: 4; MaxWait: 50
First attach the Kafka operation log profile: Log4j.propertiesSet the log according to the appropriate requirements.#日志级别覆盖规则 Priority: All off#1The . Sub-log Log4j.logger overwrites the primary log Log4j.rootlogger, where the log output level is set, threshold sets the Appender log receive level;2. Log4j.logger level below Threshold,appender receive level depends on threshold level;3the Log4j.logger level above the Threshold,appender receive level de
The simplest introduction to Erlang writing Kafka clientsStruggled, finally measured the Erlang to send messages to Kafka, using the Ekaf Library, reference:Kafka producer written in ErlangHttps://github.com/helpshift/ekaf1 Preparing the Kafka clientPrepare 2 machines, one is Ekaf running Kafka client (192.168.191.2),
Learning questions: Does 1.kafka need zookeeper?What is 2.kafka?What concepts does 3.kafka contain?4. How do I simulate a client sending and receiving a message preliminary test? (Kafka installation steps)5.kafka cluster How to interact with zookeeper? 1.
I. Kafka INTRODUCTION
Kafka is a distributed publish-Subscribe messaging System . Originally developed by LinkedIn, it was written in the Scala language and later became part of the Apache project. Kafka is a distributed, partitioned, multi-subscriber, redundant backup of the persistent log service . It is mainly used for the processing of active streaming data
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.