Rsyslog is a log collection tool. Currently, many Linux systems use rsyslog to replace syslog. I will not talk about how to install rsyslog. I will talk about the principle and the configuration of logstash.
Rsyslog itself has a configuration file/etc/rsyslog. conf, which defines the log file and the corresponding storage address. The following statement is used as an example:
local7.* /var/log/boot.log
I
Logstash tcp multihost output (multi-target host output to ensure the stability of the TCP output link), logstashmultihost
When cleaning logs, there is an application scenario, that is, when the TCP output, you need to switch to the next available entry when a host fails, the original tcp output only supports setting a single target host. Therefore, I developed the tcp_multihost output plug-in based on the original tcp to meet this scenario.
The plug-
Logstash is a member of the elk,The Redis plugin is also a handy gadget introduced in the Logstash book.Before, with a smaller cluster deployment, not involved in Redis middleware, so it is not very clear the configuration inside,Later used to find the configuration a bit of a pit.When the first configuration, dead or alive is not connected, always error, said connection refused.But there is no problem with
Recently helped Lei elder brother transplant a set of open source log management software, replace Splunk. Splunk is a powerful log management tool that not only adds logs in a variety of ways, produces graphical reports, but, most of all, its search capabilities-known as "Google for it." Splunk has a free and premium version, the main difference is the size of the index per day (index is the basis of the search function), the free version of the maximum daily 500M. When using the free version,
Use logstash to collect php-fpmslowlog. Currently, the php-fpm service is deployed in docker. the php-fpm log and php error log can be sent through the syslog protocol, however, the slow log of php-fpm cannot be configured as a syslog protocol and can only be output to files, because an slow log consists of multiple lines.
To collect slow logs, you can collect them using tools such as logstash and flume.
Logstash is mainly used for data collection and analysis, with Elasticsearch,kibana easy to use, installation tutorial Google out a lot.Recommended Reading
Elasticsearch Authoritative Guide
Proficient in Elasticsearch
Kibana Chinese Guide
The Logstash Book
ObjectiveEnter the regular Nginx log, filter into the required fields and deposit into the elasticsearch.Log style:115.182.31.1
Types in logstashType array boolean bytes codec hash number password path stringarray in logstash
An array can be a single string value or multiple values. If you specify the same setting multiple times, it appends to the array.Example:
path => [ "/var/log/messages", "/var/log/*.log" ]path => "/data/mysql/mysql.log"
boolean
Boolean, true, falseExample:
ssl_enable => true
bytes
A bytes field is a string field that represents a valid unit of bytes. It i
Write in front: In doing Elk logstash processing MySQL slow query log when the problem: 1, the test database does not have slow log, so there is no log information, resulting in ip:9200/_plugin/head/interface anomalies (suddenly appear log data, deleted the index disappeared) 2, Processing log script Problem 3, the current single-node configuration script file/usr/local/logstash-2.3.0/config/slowlog.conf "V
When elk is deployed, an error is reported when logstash is started.
Sending logstash logs to/var/log/logstash. log.Exception in thread "> output" org. elasticsearch. Discovery. masternotdiscoveredexception: waited for [30 s]At org. elasticsearch. Action. Support. master. transportmasternodeoperationaction $3. ontimeout (ORG/elasticsearch/Action/support/master/t
Grok-patterns contains log parsing rules for regular expressions with many underlying variables, including Apache log parsing (which can also be used for nginx log parsing). Based on Nginx log analysis configuration: 1. Configure the Nginx log format as follows: Log_format main ' $remote _addr [$time _local] "$request" $status $body _bytes_ Sent "" $http _referer "" "$request _time" '; access_log/var/log/nginx/access.log main; The Nginx log is screened, Remove unused logs. At this time, for the
Recently in the project using Logstash do log collection and filtering, feel logstash is still very powerful.Input {file{path = "/xxx/syslog.txt" Start_position = beginning codec = Multilin e{Patterns_dir = ["/xx/logstash-1.5.3/patterns"] pattern = "^%{message}" Nega Te = True what = "previous"}}}filter{mutate{split = ["message", "|"] Add_field = {"tmp" =
Sometimes we need to analyze some server logs and alarm the wrong logs, where we use Logstash to collect these logs and send error log data using our own developed mail delivery system.For example we have several files that need to be monitored (BI logs)We can collect these file logs by configuring Logstash
input{file{Path=> "/diskb/bidir/smartbi_prd_*/apache-tomcat-5.5.25_prd_*/logs/catalina.o
Mutate:http://www.logstash.net/docs/1.4.2/filters/mutateUse Logstash to extract the Ora error from the alter log of Oracle.The log format is as follows:ALTER DATABASE openerrors in file d:\oracle\diag\rdbms\hxw168\hxw168\trace\hxw168_ora_6148.trc:ora-01589: To open a database you must use the Resetlogs or Noresetlogs option ORA-1589 signalled During:alter database Open...alterLogstash content:input{file{codec=>plain{charset=> "CP936" #windows下的编码是cp9
Types in Logstash
Array
Boolean
bytes
Codec
Hash
Number
Password
Path
String
ArrayAn array can is a single string value or multiple values. If you specify the same setting multiple times, it appends to the array.Example"/var/log/messages""/var/log/*.log""/data/mysql/mysql.log"BooleanBull, True,false.ExampletruebytesA bytes field is a string field, that represents a valid unit of bytes. It is a convenient-t
1. Boot automatically start Redis1), copy the Redis_init_script file under Redis directory utils to/ETC/INIT.D and rename it to REDISD, then run chmod u+x REDISD2), modify the redis.conf under the Redis root directory, change the daemonize to Yes, and change the pidfile to/var/run/redis_6379.pid3), copy the redis.conf under the Redis root directory to the/etc/redis/directory and rename it to 6379.conf4), in the console input Chkconfig REDISD on, confi
username[a-za-z0-9_-]+user%{username}int (?: [+]? (?: [0-9]+)] base10num (? Logstash There are many more pattern, please refer toHttps://github.com/logstash-plugins/logstash-patterns-core/tree/master/patternsThis article is from the "Zengestudy" blog, make sure to keep this source http://zengestudy.blog.51cto.com/1702365/1782593Grok pattern in
Elasticsearch + Logstash + Kibana ConfigurationElasticsearch + Logstash + Kibana Configuration
There are many articles about the installation of Elasticsearch + Logstash + Kibana. I will not repeat them here, but I will only record some details here.
Precautions for installing AWS EC2Remember to open the elasticsearch address on ports 9200,9300 and 5601. Do not w
Using spring to consolidate Kafka only supports kafka-2.1.0_0.9.0.0 and above versions
Kafka Configuration
View Topicbin/kafka-topics.sh--list--zookeeper localhost:2181Start a producerbin/kafka-console-producer.sh--broker-list localhost:9092--topic testOpen a consumer (2183)bin/kafka-console-consumer.sh--zookeeper localhost:2181--topic test--from-beginningCreate a Themebin/kafka-topics.sh--create--zookeeper 10.92.1.177:2183--replication-factor 1--partitions 1--topic test
Tags: Logstash elk elasticsearchUse Logstash to fetch a datetime type number from MySQL. In stdout view the data JSON format takes a field value similar to2018-03-23T04:18:33.000Z, because you want to use this field as a @timestamp, use the date of Logstash to match. date { match => ["start_time","ISO8601"] }But the actual discovery of each document wi
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.