After more than a year of contact with Docker, Docker was just getting started, and now it is very popular. At that time, we mainly engaged in cloud computing and needed a mechanism to isolate applications between users. I read a lot of information on the internet and found a container technology called LXC (linux container) that can be used for sandbox isolation or allocating different memories and CPUs to containers. Because LXC is relatively low-level and has no image concept, it is inconvenient to manage a large number of containers. Later, I accidentally saw Docker. Compared with LXC, it is more inclined to the application layer, which is equivalent to encapsulating LXC with the concept of image.
What is Docker. Docker is an open-source engine that allows you to easily create a lightweight, portable, and self-sufficient container for any application. Containers compiled and tested on a notebook can be deployed in a production environment in batches, including VMs, OpenStack clusters, and other basic application platforms.
Docker is mainly used: Automated packaging and publishing of web applications; automated testing and continuous integration and publishing; deployment and adjustment of databases or other background applications in a service-type environment; compile or expand the existing OpenShift or Cloud Foundry platform from scratch to build your own PaaS environment.
This section describes how to deploy a hadoop cluster in a Docker container and how to expand a hadoop computing node. Is the entire Hadoop on Docker architecture
The advantage of this architecture is that it can make full use of host resources to prevent namenode from single point of failure.
Hadoop on Docker