background
We have been using Hive server 1 for a long time, and users Ad-hoc Query,hive-web, wormhole, operations tools, and so on, are submitting statements through Hive Server. But hive server is extremely unstable, often inexplicable mysterious death, causing the client side of all connection are blocked. To this w
Original articles, reproduced please mark from http://blog.csdn.net/lsttoy/article/details/53406710.
The first step :Download the latest hive and go directly to Apache to find hive2.1.0 download on the line.
Step two , unzip to the server
Tar zxvf apache-hive-2.0.0-bin.tar.gz
mv apache-hive-2.0.0-bin/home/hive
The
The kettledesktop version of Pentaho was just started. Here we mainly use its association with hadoop and hive for data processing. Kettle's version is 4.4, and The use process is quite smooth. A conversion task has been successfully established to extract data from hive to a local file. However, once you open it, all utf8 Chinese characters are
The kettle desktop version of Pentaho was just started. Here w
Hive metadata parsing this article is a Hive metadata table prepared by the author. If there is any inaccuracy, please pat it. I will add it later. 1. hive0.11 meta data Table Summary online Hive0.11metastore includes the following 39 tables, mainly divided into the following categories: Database-related Table-related data storage-related SDSCOLUMN-related SERDE-related (serialization) P
Improvements to the Hive Optimizer, improvementshive
LanguageManual JoinOptimization
Improvementsto the Hive Optimizer
Hive can be automatically optimized, and some optimization cases are improved in Hive 0.11.
1. The JOIN side is suitable for storing in memory and there are new optimization solutions
A) read the table
This document describes how to manually install the cloudera hive cdh4.2.0 cluster. For environment setup and hadoop and hbase installation processes, see the previous article.Install hive
Hive is installed on mongotop1. Note that hive saves metadata using the Derby database by default. Replace it with PostgreSQL here.
Reprint Please specify source: https://blog.csdn.net/l1028386804/article/details/80173778I. Hive Overview 1, why hive is used
The birth of the Hadoop ecosystem brings dawn to efficient and fast processing of big data, but requires writing mapreduce or spark tasks, a high threshold for entry, and the need to master a programming language such as Java or Scala.We have long been accustomed to traditional relat
1. Use the local metastore to start directly from the Hive command.
Hive-site.xml files are configured using local MySQL database storage Metastore
Use the following command to turn on
$HIVE _home/bin/hive
The hive command, by default, starts the client service, which i
An error is reported when Sqoop is used to migrate data between Hive and MySQL databases.
An error is reported when Sqoop is used to migrate data between Hive and MySQL databases.
Run./sqoop create-hive-table -- connect jdbc: mysql: // 192.168.1.10: 3306/ekp_11 -- table job_log -- username root -- password 123456 -- hive
Indexes are standard database technologies that support indexing after the hive0.7 version. Hive provides limited indexing functionality, unlike the traditional relational database with the "key" concept, where users can create indexes on some columns to speed up certain operations, and index data created for one table is saved in another table. Hive's indexing function is now relatively late and offers fewer options. However, the index is designed to
How to troubleshoot problems
General error, view error output, follow keyword Google
Exception errors (such as Namenode, Datanode, inexplicably hung): View HADOOP ($HADOOP _home/logs) or hive logs
Hadoop error1.datanode does not start properlyAfter adding Datanode, Datanode does not start normally, the process will somehow hang up, the view Namenode log shows as follows:Text Code2013-06-21 18:53:39,182 FATAL org.apache.hadoop.hdfs.st
Introduction
Hive user-defined aggregate functions (UDAF) are a good function that integrates advanced data processing. Hive has two types of UDAF: simple and general. As the name implies, simple UDAF writes are quite simple, but it causes performance loss due to the use of Java reflection, and some features cannot be used, such as the variable length parameter list. General UDAF can use all functions, bu
Recently, hadoop and hive have been successfully configured on five Linux servers. A hadoop cluster requires a machine as the master node, and the rest of the machines are Server Load balancer nodes (Master nodes can also be configured as Server Load balancer nodes ). You only need to configure and use hive on the master node.
1Configure hadoop
Hadoop configuration is relatively simple, because hadoop does
(1) Hive Start-Up Prerequisites:
Java environment
Hadoop Startup
MySQL Boot
(2) Missing Hive execution Jar:/usr/usr/hive-1.2.2/lib/hive-exec-*.jar
Direct use: Missing Hive execution Jar ...Check Baidu, Google, is not found this answer. I looked for a long time, through thi
Original link: http://blog.ywheel.cn/post/2016/06/12/hive_in_oozie_workflow/
By building and maintaining big data platforms in the company and providing it to other data analysts, Hive is the most (almost unique) service that non-programmers use. Of course, in daily data processing, in order to simplify the coding effort and use the results accumulated by the data analyst, we can use or simply modify the HQL scripts they provide for data processing,
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.