This article is mainly for the Kong installation of the small note, the system environment for CentOS 6.7Please specify the source--xiaoeight the article reproduced
Introduced
Kong is an API gateway that forwards API traffic between the client and (micro) services, extending functionality through plug-ins. There are two main components of Kong:
Kong Server: Nginx-based server used to receive API requests.Apache Cassandra: U
,twitter and digg.com are using Cassandra. The main characteristic of Cassandra is that it is not a database, but a distributed network service composed of a bunch of database nodes, a write operation to Cassandra will be copied to the other nodes, and the read operation to Cassandra will be routed to a node to read. F
Tags: small and medium-sized enterprises big data technology route Selection of big data technology routes for Small and Medium-sized Enterprises
Currently, big data is mainly used in the Internet and e-commerce fields, and is gradually used in the telecom and power industries. For the majority of small and medium-sized enterprises, big data has heard too much. However, the technical threshold for big data is still high. From the technical line, choosing the technical solutions used by large com
Recently, Twitter developed a distributed real-time statistics system, Rainbird.
Usage
Rainbird can be used for real-time data statistics:
1. count the number of clicks on each page and domain name on the website
2. Internal System Operation Monitoring (count the running status of the monitored server)
3. Record the maximum and minimum values
Performance Requirements
As a distributed application of large websites, the following performance is required:
1 High Write Performance, up to 100,000 WP
following way:
All events generated by the original system should be marked with a timestamp.
When the processor in the pipeline processes the stream, it tracks the maximum timestamp, assuming that the persisted global clock is behind, and updates it to the maximum timestamp. Time synchronization of other processors with the global clock
While the data is being played back. Resets the global clock.
Persisted storage RollupWe've discussed how persistent storage can be used
. You might as well build a relational db cluster, but they're using shared storage, which is not the type we want. So there's a nosql era in which Google, Facebook, and Amazon are trying to handle more transmissions.NoSQL eraThere are a lot of NoSQL databases now, such as MongoDB, Redis, Riak, HBase, Cassandra, and so on. Each one has one of the following features:
No longer use the SQL language, such as MongoDB,
require huge overhead and a performance cap, and it's never possible to use a single machine to support all of the load on Google and Facebook. In view of this situation, we need a new database because the relational database does not run well on the cluster. You might as well build a relational db cluster, but they're using shared storage, which is not the type we want. So there's a nosql era in which Google, Facebook, and Amazon are trying to handle more transmissions.NoSQL eraThere are a lot
require huge overhead and a performance cap, and it's never possible to use a single machine to support all of the load on Google and Facebook. In view of this situation, we need a new database because the relational database does not run well on the cluster. You might as well build a relational db cluster, but they're using shared storage, which is not the type we want. So there's a nosql era in which Google, Facebook, and Amazon are trying to handle more transmissions.NoSQL eraThere are a lot
Databases are generally divided into the following types: relational (transactional) databases, represented by Oracle and MySQL, with keyValue databases, represented by redis and memcached dB, there are document-based databases such as MongoDB, and columnar databases represented by hbase, Cassandra, and Dynamo, as well as other graphic databases, object data libraries, and XML databases.
Some databases are distributed database design concepts, such as
on the feasibility of Big Data 3.0 replacing SAP HANA
First, Big Data 3.0 Introduction
In short, Big Data 3.0 is about implementing SQL on the big data and balancing performance, ease of use, and scalability. At present, "search engine + Large data +sql" such a convergence as one of the trend.
Let's take a look at some of my other posts:
1. Agile large data based on Facebookpresto+cassandrahttp://blog.csdn.net/china_world/article/details/39966699
2, small and medium-sized enterprises of large d
. You might as well build a relational db cluster, but they're using shared storage, which is not the type we want. So there's a nosql era in which Google, Facebook, and Amazon are trying to handle more transmissions.NoSQL eraThere are a lot of NoSQL databases now, such as MongoDB, Redis, Riak, HBase, Cassandra, and so on. Each one has one of the following features:
No longer use the SQL language, such as MongoDB,
:
Traditional enterprise database vendors often propose that NoSQL lacks professional monitoring and management tools. Their argument is that enterprise applications need to fine-tune and monitor databases to ensure stable performance and operation. The NoSQL vendor's point of view is that the lack of such degree does not help RDBMS in solutions. Many NoSQL vendors also try to differentiate their monitoring and management software levels, such as Cassandra
database provides several different consistency models:
Atom writes. If the database provides an API, a write operation can only be a single atomic assignment, and the way to avoid writing conflicts is to find the "latest version" of each data. This allows all nodes to obtain the same version at the end of the update, regardless of the order in which the updates are made, and network failures and delays often result in inconsistent sequence of update sequences for each node. The
Nodetool when viewing the local cluster state, an error indicating that an unreachable connection timeout occurred after n long time: $ nodetool Statusnodetool:failed to connect to ' 127.0.0.1:7199 '-connectexception: ' Connection timed out '. After being checked for/etc/hosts file configuration error, the actual host name (not localhost) in the file corresponds to the IP address resolution line set incorrectly, this is due to the host IP address replacement did not update the hosts file in time
Abstract: Almost all software project development needs the support of the database, at present, with the rapid development of open source technology, more and more data suppliers choose Open Source database, contribute to the cause of open source.
As a software developer or DBA, one of the essential tasks is to deal with databases, such as MS SQL Server, MySQL, Oracle, PostgreSQL, MongoDB, and so on. As we all know, MySQL is currently the most widely used and the best free open source database
.
Write consistency. A write conflict often occurs in a partitioned database. The database should be able to handle this conflict and ensure that multiple write requests are not handled by different partitions. In this respect, the database provides several different consistency models:
Atom writes. If the database provides an API, a write operation can only be a single atomic assignment, and the way to avoid writing conflicts is to find the "late
different partitions. In this respect, the database provides several different consistency models:
Atom writes. If the database provides an API, a write operation can only be a single atomic assignment, and the way to avoid writing conflicts is to find the "latest version" of each data. This allows all nodes to obtain the same version at the end of the update, regardless of the order in which the updates are made, and network failures and delays often result in inconsistent seque
Michael Kopp has more than 10 years of architecture and development experience in C ++ and Java/Jee. He is now a compuware technology strategist who specializes in the architecture and performance of large-scale product deployment.
The following is a translation:
Traditional enterprise database vendors often propose that nosql lacks professional monitoring and management tools. Their arguments are: enterprise applicationsProgramThe database needs to be fine-tuned and monitored to ensure stab
In this section, I will test Solr in the Linux environment and check whether the Solr environment is OK by adding, deleting, modifying, and querying the instance demo. I remember some of my friends left a message asking me that I only saw Solr and Cassandra being built in a separate environment, but they didn't see their collaboration. I would like to say that there is no inevitable relationship between the two. In the DSE version of
Data scale BigTable database systems (such as HBase and Cassandra) are designed to meet the needs of massive data storage. The massive data scale mentioned here refers to the size of data stored in a single table in terabytes or petabytes. A single table is composed of hundreds of billions of rows and hundreds of billions of columns. When we mention the data scale, we have to say that we are currently at NoSQ
Data scale BigTable database systems (such
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.