Index fields are indexed using automatic detection in ES, such as IP, date auto-detect (default on), Auto-detect (default off) for dynamic mapping to automatically index documents, and when specific types of fields need to be specified, you might use mapping to define mappings in index generation.The settings for the default index in Logstash are template-based.First we need to specify a default mapping file, the contents of the file are as follows:{
Benefits of the unified collection of real-time logs:1. Quickly locate the problem machine in the cluster2, no need to download the entire log file (often relatively large, download time is much)3, the log can be countedA, to find the most frequently occurring anomalies, for tuning processingB, Statistics crawler IPC, Statistical user behavior, do cluster analysis, etc.Based on the above requirements, I adopted the ELK (Elasticsearch + Logstash + kiba
Elasticsearch+logstash+kibana ConfigurationThere are a lot of articles about the installation of Elasticsearch+logstash+kibana, which is not repeated here, only some of the more detailed content.
Considerations for installing in AWS EC2
9200,9300,5601 Port to remember to open
Elasticsearch address do not write external IP, otherwise it will be a waste of data, write internal IP"ip-10-1
Article from Aliyun-yun-Habitat community, the original click here.
The second component of the Logstash three components is also the most complex, logstash component of the entire tool, and, of course, the most useful component.
1, Grok plug-in Grok plug-in has a very powerful function, he can match all the data, but his performance and the loss of resources also let people criticized.
filter{
gro
Install Logstash 2.2.0 and Elasticsearch 2.2.0 on CentOS
This article describes how to install logstash 2.2.0 and elasticsearch 2.2.0. The operating system environment version is CentOS/Linux 2.6.32-504.23.4.el6.x86 _ 64.
JDK installation is required. It is generally available in the operating system. It is only a version issue and will be mentioned later.
Kibana is only a front-end UI written in pure JavaS
little too hard.Open source real-time log analysis Elk platform can perfectly solve our problems above, elk by Elasticsearch, Logstash and Kiabana three open source tools. Official website: https://www.elastic.coElasticsearch is an open source distributed search engine, it features: distributed, 0 configuration, automatic discovery, Index auto-shard, index copy mechanism, RESTful style interface, multi-data source, automatic search load, etc.Logstash
Logstash analysis httpd_logLogstash analysis: httpd_loghttpd or nginx format
Logstash supports two built-in formats: common and combined compatible with httpd.
COMMONAPACHELOG %{IPORHOST:clientip} %{USER:ident} %{USER:auth} \[%{HTTPDATE:timestamp}\] "(?:%{WORD:verb} %{NOTSPACE:request}(?: HTTP/%{NUMBER:httpversion})?|%{DATA:rawrequest})" %{NUMBER:response} (?:%{NUMBER:bytes}|-)COMBINEDAPACHELOG %{COMMONAPAC
Collection process 1nxlog = 2logstash + 3elasticsearch1. Nxlog Use module Im_file to collect log files, turn on location recording function2. Nxlog using the module TCP output log3. Logstash use INPUT-TCP, collect logs, and format, output to ESThe Nxlog configuration file above windowsNxlog.conf
1234567891011121314151617181920212223242526272829303132333435363738394041
##Thisisasampleconfigurationfile.Seethenxlogreferencemanualaboutthe
When using Logstash, some regular expressions are written for finer-grained cutting logs. How to use input { file { type => "billin" path => "/data/logs/product/result.log" } } filter { grok { type => "billin" pattern => "%{BILLINCENTER}" patterns_dir => "/data/logstash/patterns/my_patterns"
Now the mainstream log analysis system has Logstash and flume, combined with a lot of online predecessors, summed up a bit, hope and everyone to share and discuss, there are different ideas welcome message.FlumeCloudera provides a high-availability, high-reliability, distributed mass log collection, aggregation and transmission system;Support the customization of various types of data sender, easy to collect data, general and Kafka subscription messag
Type settings:The Redis plugin in Logstash specifies three ways to read the information in the Redis queue.
List=>blpop ( equivalent to queue )
Channel=>subscribe ( equivalent to a specific channel for publishing subscriptions )
Pattern_channel=>psubscribe ( equivalent to publishing a subscription to a group of channels )
Where list is the equivalent of a queue; a channel is equivalent to a specific channel for a subscription; Pa
Tags: CTE nload. SQL ODI Line SQL ADE JDBC Remove input {stdin {} jdbc {#MySQL JDBC connection string to our backup databseJdbc_connection_string ="Jdbc:mysql://localhost:3306/userdb?useunicode=truecharacterencoding=utf-8usessl=false"
#The user we wish to excute our statement asJdbc_user ="User"Jdbc_password="Pass"
#The path to our downloaded JDBC driverJdbc_driver_library ="Mysql-connector-java-5.1.40-bin.jar"
#The name of the driver class for MySQLJdbc_driver_class ="Com.mysq
Collection process 1nxlog = 2logstash + 3elasticsearch1. Nxlog Use module Im_file to collect log files, turn on location recording function2. Nxlog using the module TCP output log3. Logstash use INPUT-TCP, collect logs, and format, output to ESThe Nxlog configuration file above windowsNxlog.conf##thisisasampleconfigurationfile.seethenxlog referencemanualaboutthe##configurationoptions.itshouldbe installedlocallyandisalsoavailable##onlineathttp://nxlog.
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.