Ceph is a fully open source distributed storage solution, a network block device, and a file system with high stability, high performance, and high scalability to handle data volumes from terabyte to exabyte levels.By using innovative scheduling algorithms (CRUSH), active storage nodes, and Peer-to-peer gossip protocols, Ceph avoids the problems of scalability and reliability in traditional centralized cont
First, pre-installation preparation
1.1 Introduction to installation Environment
It is recommended to install a Ceph-deploy management node and a three-node Ceph storage cluster to learn ceph, as shown in the figure.
I installed the Ceph-deploy on the Node1.
First three machines were prepared, the names of which wer
Document directory
1. Design a CEpH Cluster
3. Configure the CEpH Cluster
4. Enable CEpH to work
5. Problems Encountered during setup
Appendix 1 modify hostname
Appendix 2 password-less SSH access
CEpH is a relatively new Distributed File System completed by the USSC storage team. It is a Network File System
Deployment Installation
Regarding the installation ceph the whole process encounters the question, as well as the reliable solution, the personal test is effective, does not represent the general colleague's viewpoint.I am using the server, so I did not engage in any user issues. The machine is centOS7.3. The Ceph version I installed is jewel and currently uses only 3 nodes.Node IP naming role10.0.1.92 e10
85b958c7removing Intermediate container 16dd4e7e3995successfully built 43f885b958c7# Docker imagesrepository TAG IMAGE ID CREATED Sizetest/rbd-rest-api Latest 43f885b958c7 seconds ago 298 MB
To test the boot image, note that we "read-only" mount the local path/etc/ceph:
# docker run
Ceph provides three storage methods: Object Storage, block storage, and file system. The following figure shows the architecture of the Ceph storage cluster:We are mainly concerned about block storage. In the second half of the year, we will gradually transition the virtual machine backend storage from SAN to Ceph. although it is still version 0.94,
This is a creation in
Article, where the information may have evolved or changed.
In the article "using Ceph RBD to provide storage volumes for kubernetes clusters," we learned that one step in the integration process for kubernetes and Ceph is to manually create the RBD image below the Ceph OSD pool. We need to find a way to remove this manual step. The first th
Ceph Client:Most ceph users do not store objects directly into the Ceph storage cluster, and they typically choose one or more of the Ceph block devices, the Ceph file system, and the Ceph object storage;Block device:To practice t
Ceph monitoring Ceph-dash Installation
There are a lot of Ceph monitors, such as calamari or inkscope. When I started to try to install these, they all failed, and then Ceph-dash came into my eyes, according to the official description of Ceph-dash, I personally think it is
Today, Ceph is configured, referencing the official document address of the multiparty document http://docs.ceph.com/docs/master/rados/configuration/ceph-conf/#the-configuration-file
Other great God's blog address http://my.oschina.net/oscfox/blog/217798
Http://www.kissthink.com/archive/c-e-p-h-2.html and so on a case of a.
Overall on the single-node configuration did not encounter any air crashes, but mult
Ceph Cluster Expansion and Ceph Cluster ExpansionThe previous article describes how to create a cluster with the following structure. This article describes how to expand the cluster.
IP
Hostname
Description
192.168.40.106
Dataprovider
Deployment Management Node
192.168.40.107
Mdsnode
MON Node
192.168.40.108
Osdnode1
OSD Node
192.168.40.14
1. Environment and descriptionDeploy ceph-0.87 on ubuntu14.04 server, set Rbdmap to mount/unload RBD block devices automatically, and export the RBD block of iSCSI with a TGT with RBD support. 2. Installing Ceph1) Configure hostname with no password login [Email protected]:/etc/ceph# cat/etc/hosts127.0.0.1localhost192.168.108.4 osd2.osd2osd2192.168.108.3 osd1.osd1osd1192.168.108.2 Mon0.mon0mon0 #示例如下ss
Ceph environment setup (2)1. There are three layout hosts: node1, node2, and node3. Each host has three osdks, as shown in figure. osd1, and 8 are SSD disks, 4 is the SATA disk. Each of the three hosts has a Monitor and an MDS. We use osd1, 3, and 4 to create a pool named ssd, in the form of three copies, osd0, and 2, 4 to build a Pool named sata, in the form of Erasure code, k = 2, m = 1, that is, use two osdks to store data fragments, one osd to sto
As an architect in the storage industry, I have a special liking for file systems. These systems are used to store the user interfaces of the system. Although they tend to provide a series of similar functions, they can also provide significantly different functions. CEpH is no exception. It also provides some of the most interesting features you can find in the file system.
CEpH was initially a PhD resea
Official Document: http://docs.ceph.com/docs/master/start/quick-start-preflight/Chinese version: http://docs.openfans.org/ceph/Principle: Using the Ceph-deploy tool, through the management node Admin-node, using the SSH channel, in order to achieve the control of each distributed node storage sharing function.650) this.width=650; "Src=" http://docs.ceph.com/docs/master/_images/ Ditaa-5d5cab6fc315585e5057a74
Ceph was originally a PhD research project on storage systems, implemented by Sage Weil in University of California, Santa Cruz (UCSC). But by the end of March 2010, you can find Ceph in the mainline Linux kernel (starting with version 2.6.34). Although Ceph may not be suitable for production environments, it is useful for testing purposes. This article explores
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.