Hadoop real-combat development (HDFS real-combat image, MapReduce, hbase real-combat microblogging, hive application)

Source: Internet
Author: User

What is Hadoop and why do you learn Hadoop?


Hadoop is a distributed system infrastructure developed by the Apache Foundation. Users can develop distributed programs without knowing the underlying details of the distribution. Leverage the power of the cluster for high-speed operations and storage. Hadoop implements a distributed filesystem (Hadoop Distributed File System), referred to as HDFs. HDFs is characterized by high fault tolerance and is designed to be deployed on inexpensive (low-cost) hardware. And it provides high transfer rates (HI throughput) to access application data for applications with very large datasets (large data set). HDFs relaxes (relax) POSIX requirements (requirements) so that data in the form of a stream can be accessed (streaming access) in the file system.
Hadoop is a software framework that enables distributed processing of large amounts of data. But Hadoop is handled in a way that is reliable, efficient, and scalable. Hadoop is reliable because it assumes that compute elements and storage will fail, so it maintains multiple copies of working data, ensuring that it can redistribute processing against failed nodes. Hadoop is efficient because it works in parallel and speeds up processing by parallel processing. Hadoop is also scalable and can handle petabytes of data. In addition, Hadoop relies on community server, so it costs less and can be used by anyone.
Hadoop comes with a framework written in the Java language, so it's ideal to run on a Linux production platform. This course is explained by using a Linux platform for simulation, based on real-world scenarios.

Highlight one: Comprehensive technical point, perfect system

This course takes into account the improvement of knowledge system of Hadoop course, draws out the most applied, deepest and most practical technologies in practical development, and through this course, you will reach the new high point of technology and enter the world of cloud computing. In the technical aspect you will master the basic Hadoop cluster, Hadoop hdfs principle, Hadoop hdfs Basic command, namenode working mechanism, HDFS basic configuration management; MapReduce principle; hbase system architecture; HBase Table Structure HBase How to use mapreduce;mapreduce advanced programming, split implementation details, Hive Primer, hive combined with Mapreduce;hadoop cluster installation and many other knowledge points.

Highlights II: basic + Actual combat = Application, both learning and practice

Each stage of the course has a practical application project, so that students can quickly grasp the application of knowledge points, such as in the first stage, the course combined with the HDFS application, explained the image server design, and how to use the Java API to operate on HDFs, in the second phase; The course combines hbase to realize the various functions of the microblog project so that learners can ingenious. In the third stage: HBase and MapReduce combined with the realization of a single query and statistical system, in the fourth phase, hive combat, through the actual combat data statistics system, so that students in the shortest time to master hive advanced applications.

Highlight three: Lecturer-rich experience in cloud platform operation of Telecom Group

Lecturer Robby has a wealth of experience in the telecommunications group, is currently responsible for all aspects of the cloud platform, and has many years of in-house training experience. The lecture content is completely close to the enterprise demand, not on paper.

For more technical highlights, refer to the course Outline : (This outline is named in Chapter 1 to prevent more than 1 hours of chapter content)

1th Chapter:
> Hadoop Background
> HDFs Design Goals
> HDFs not suitable for the scene
> HDFs Architecture Detailed analysis
> The Fundamentals of MapReduce

2nd Chapter
Introduction to the > Hadoop version
> Installing a standalone version of Hadoop
> Installing Hadoop clusters

3rd Chapter
> HDFs command-line basic operations
The working mechanism of > Namenode
> HDFS Basic Configuration Management

4th Chapter
> HDFS Application Combat: Image Server (1)-System design
> Application Environment Build PHP + bootstrap + Java
> Writing files to HDFs using the Hadoop Java API implementation

5th Chapter
> HDFS Application Combat: Image Server (2)
> read files in HDFs using the Hadoop Java API implementation
> Get a list of HDFs directories using the Hadoop Java API implementation
> Delete files in HDFs using the Hadoop Java API implementation


6th Chapter
> The Fundamentals of MapReduce
> MapReduce's Running Process
> Building a MapReduce Java development environment
> Implementing WordCount using the MapReduce Java interface

7th Chapter
> WordCount Operation Process Analysis
> Combiner of MapReduce
> Using MapReduce to achieve data deduplication
> Using MapReduce for Data sorting
> Using MapReduce to achieve average data score calculation

8th Chapter
> HBase Detailed Introduction
System Architecture for > HBase
> HBASE table structure, RowKey, column family, and timestamp
> Master,region in hbase and region Server


9th Chapter
> Use HBase for Weibo applications (1)
> user registration, Login and logout design
> Build Environment Struts2 + JSP + bootstrap + jquery + HBase Java API
> hbase and user-related table structure design
> Implementation of User Registration

10th Chapter
> Use HBase for Weibo applications (2)
> User logon and logoff using session
> The design of "attention" function
> table structure Design of "attention" function
> The implementation of the "focus" function


11th Chapter
> Use HBase for Weibo applications (3)
> The design of "hair Weibo" function
> The table structure design of "tweet" function
> The implementation of "tweet" function
> show the operation of the entire application

12th Chapter
> HBase and MapReduce Introduction
> How HBase uses MapReduce

13th Chapter

> HBase Application Combat: Single Query and statistics (1)
Overall design for > applications
> Development Environment Construction
> Table Structure Design

14th Chapter
> HBase Application Combat: Single Query and Statistics (2)
Design and implementation of > Single storage
Design and implementation of > Single query

15th Chapter
> HBase Application Combat: Single Query and Statistics (3)
> Statistical function Design
> Statistical Function Realization

16th Chapter
> In-depth mapreduce (1)
An explanation of > Split's implementation
> Implementation of Custom Inputs
> Example Explanation

17th Chapter
> In-depth mapreduce (2)
Partition of > reduce
> Example Explanation

18th Chapter
> Getting Started with hive
> Installing Hive
> Using hive to deposit structured data into HDFs
> Basic use of Hive


19th Chapter
> Using MySQL as Hive's metabase
> Hive combined with MapReduce

20th Chapter
> Hive Application Combat: Data Statistics (1)
> Application design, table structure design

21st Chapter
> Hive Application Combat: Data Statistics (2)
> Data entry and the realization of statistics

Hadoop real-combat development (HDFS real-combat image, MapReduce, hbase real-combat microblogging, hive application)

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.