I was fortunate enough to take the MOOC college Hadoop experience class at the academy. This is the little Elephant College hadoop2. X Overview Notes for chapter eighthThe main introduction is HBase, a distributed database application case.Case Overview:1) Time series database (OPENTSDB) Use HBase to store time series data, every moment is resolved, the database is open source 2)
SQL operation is a traditional table.
3, HBase: In layman's words, hbase functions like a database, traditional database management is centralized local data files, and HBase is based on HDFS implementation of distributed data file management, such as adding and deleting. In other words, hbase is just a persistent fil
Zhou Hai Han Wen 2013.4.2 can convert the date 08081620: 56: 29 from hbaselog into a timestamp, the operation is as follows: hbase (main): 021: 0importjava. text. simpleDateFormathbase (main): 022: 0importjava. text. parsePositionhbase (main): 023: 0SimpleDateFormat. new (yyMMdd
Zhou haihan/Wen 2013.4.2 can convert the date 08/08/16 20:56:29 from hbase log to a timestamp. The operation is as follows:
This article will briefly introduce the available data backup mechanism of Apache hbase and the fault recovery/disaster recovery mechanism of massive data.
As hbase is widely used in important commercial systems, many enterprises need to establish robust backup and fault recovery (BDR) for their hbase clusters) mechanism to ensure their enterprise (data) assets.
Tags: server group data structure POP res cannot object data size user reduceFour types of NoSQL The status of NoSQL database in the whole database field is self-evident. In the era of big data, although the RDBMS is excellent, with the rapid growth of data size and increasingly complex data models, the RDBMS is unable to cope with many database processing tasks, when NoSQL relies on easy to expand, Big data volumes and high performance and flexible d
Phoenix 3.1 + HBase 0.94.21 Installation and Use
Apache Phoenix is an HBase SQL driver. Phoenix enables HBase to be accessed through JDBC and converts your SQL query to HBase scan and corresponding actions.
Compatibility:
Phoenix 2.x-HBase 0.94.x
Phoenix 3.x-
cluster. A scenario that is suitable for users to calculate in real time.
In the case of a problem, such as storm cluster restart, the calculation task status is lost, you may need to be able to spout the data back ("Start_timestamp, Stop_timestamp") can meet this demand.
Storm-hbase seeks to be as generic as possible, thus extracting configuration information for Storm and HBase. Configuration options f
Deployment and basic use of the Nutch2.x + Hbase environment
Because the project wants to use Nutch for web crawlers, some research has found that online documents are scattered and difficult to learn. Therefore, I have summarized some of them and put them up to communicate with you.1. Environment deployment
There are 1. x Series and 2. x Series, the main difference is 2. x uses Gora as the persistent layer media to persist data to relational database
Manipulating the HBase database in the shell command line
Shell control
Enter the shell command line interface, execute the hbase command, and attach the SHELL keyword:
[Grid@hdnode3 ~]$ hbase shell
hbase shell; enter¨help¨for list of supported.
Type "Exit" to leave the
HBase is a distributed NoSQL database based on the LSM tree storage model. The LSM tree, compared to the popular B + tree, can achieve high random write performance while maintaining reliable random read performance (refer to here). In the case of a read request, the LSM tree merges multiple subtrees (similar to the B + tree structure) to query, and for HBase, th
:
StringKey; int32_t Timestamp=1352563200;Const Char* PTS = (Const Char*)Timestamp; size_t n=Sizeof(Int32_t); key. append (PTS, N );
Convert the Lexicographic Order to int using the following method:
Const Char* Ts =Key. c_str (); int32_t Timestamp= * (Int32_t *) (TS ));
In PHP, the pack and unpack methods are provided for conversion:
$ Key=Pack("N ",$ Num);$ Num=Unpack("N ",$ Key); 2. Use of tscan traps
In the PHP thrift interface of hbase,
Download: http://mirror.bit.edu.cn/apache/hbase/stable/
Official Guide: http://abloz.com/hbase/book.html
Installation configuration:
Unzip:
TAR-XZVF hbase-0.96.0-hadoop1-bin.tar.gz
Go to $hbase/lib and look at the related Hadoop package to see which version of Hadoop it is.
only the installation of the pseudo-distrib
The previous blog mentions the eclipse operation stand-alone version of HBase is not familiar with friends can go to seeEclipse connects and operates a single version of HBaseThis article describes a Mapreduce that reads data from HBase and computes the column similar to WordCount but the input at this point is read from HBaseFirst you need to create an input sourceStart
not appropriate .
2 Limitations of HBase:
Can only do simple key value query , complex SQL statistics do not .
you can only do quick queries on the row key .
3 row-style storage for traditional databasesIn the case of data analysis, we often use a column as a query condition, and the returned results are often just some columns, not all columns .The I/O performance of the row database is poor in this case ,Oracle, for example,
conditions are also input to find (20120901,20121001, "Good sound in China", "Zongyi", "Zhejiang satellite TV"). At this point we should get records of 1th, 2, 3, 4, 5, 7. 6th because does not belong to "Zhejiang satellite TV" should not be selected. We can do this when we design Rowkey: using the UserID + Createtime + Fileid to make up the rowkey, which can satisfy the multi-condition query and fast query speed.The following points need to be noted:(1) Each record is Rowkey, and each field nee
We often see articles boasting about how fast a product is, how strong it is, and how it is better to test than some of the data described. The reason may be that you don't really understand the internal structure, and are not aware of its performance tuning approach. This article transferred from Taobao's Ken Wu classmate's blog, is currently seeing the more complete hbase tuning article.
Original link:
disk, the file is not only a log file, but also records the operation of the data information, when the machine from the downtime of the state to return to normal operation, The Hlog file is read into memory for data recovery. When the data in the Memstore reaches a threshold, Regionserver initiates a flashcache process that brushes the data from Memestore to Storeflie, each writing forming a separate storefile (hfile).
hbase Read Data flow:
1. Stora
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.