will evolve in the future.
References
Learning
The CEpH creator's essay "CEpH: A scalable, High-Performance Distributed File System" (PDF) and sage Weil's PhD paper, "CEpH: reliable, scalable, and high-performance distributed storage (PDF) reveals the
directory, and perform the following operations:
Tar-zxvf itsdangerous-0.24.tar.gzCd itsdangerous-0.24Python setup. py install
After installing itsdangerous, go to the Flask installation directory and try the last step of the previous Flask installation.
Python setup. py develop
Whether the message "itsdangerous" is displayed. If the prompt still does not exist, close the current terminal, re-open the terminal, and re-install it several times.
After Flask is successfully installed, run
metadata server, an object storage server, and a monitor. Ceph fills in the gaps in distributed storage and it will be interesting to see how this open source product evolves in the future.ResourcesLearn
The Ceph creator's paper "Ceph:a Scalable, high-performance distributed File System" (PDF) and the PhD dissertation for Sage Weil, "ceph:reliable, Scal Able
First, pre-installation preparation
1.1 Introduction to installation Environment
It is recommended to install a Ceph-deploy management node and a three-node Ceph storage cluster to learn ceph, as shown in the figure.
I installed the Ceph-deploy on the Node1.
First three machines were prepared, the names of which wer
Synchronize the configuration file and keyring of the admin-node to other nodes:
Ceph-deploy admin-node node1 node2 node3
Sudo chmod + r/etc/ceph. client. admin. keyring
Finally, run the following command to check the cluster health status:
Ceph health
If it succeeds, the following message is displayed: HEALTH_ OK.
Ceph
Document directory
1. Design a CEpH Cluster
3. Configure the CEpH Cluster
4. Enable CEpH to work
5. Problems Encountered during setup
Appendix 1 modify hostname
Appendix 2 password-less SSH access
CEpH is a relatively new Distributed File System completed by the USSC storage team. It is a Network File System
Deployment Installation
Regarding the installation ceph the whole process encounters the question, as well as the reliable solution, the personal test is effective, does not represent the general colleague's viewpoint.I am using the server, so I did not engage in any user issues. The machine is centOS7.3. The Ceph version I installed is jewel and currently uses only 3 nodes.Node IP naming role10.0.1.92 e10
Ceph provides three storage methods: Object Storage, block storage, and file system. The following figure shows the architecture of the Ceph storage cluster:We are mainly concerned about block storage. In the second half of the year, we will gradually transition the virtual machine backend storage from SAN to Ceph. although it is still version 0.94,
This is a creation in
Article, where the information may have evolved or changed.
In the article "using Ceph RBD to provide storage volumes for kubernetes clusters," we learned that one step in the integration process for kubernetes and Ceph is to manually create the RBD image below the Ceph OSD pool. We need to find a way to remove this manual step. The first th
Ceph Client:Most ceph users do not store objects directly into the Ceph storage cluster, and they typically choose one or more of the Ceph block devices, the Ceph file system, and the Ceph object storage;Block device:To practice t
Today, Ceph is configured, referencing the official document address of the multiparty document http://docs.ceph.com/docs/master/rados/configuration/ceph-conf/#the-configuration-file
Other great God's blog address http://my.oschina.net/oscfox/blog/217798
Http://www.kissthink.com/archive/c-e-p-h-2.html and so on a case of a.
Overall on the single-node configuration did not encounter any air crashes, but mult
Ceph Cluster Expansion and Ceph Cluster ExpansionThe previous article describes how to create a cluster with the following structure. This article describes how to expand the cluster.
IP
Hostname
Description
192.168.40.106
Dataprovider
Deployment Management Node
192.168.40.107
Mdsnode
MON Node
192.168.40.108
Osdnode1
OSD Node
192.168.40.14
fallocate:
int FileJournal::_open_file(int64_t oldsize, blksize_t blksize, bool create){... if (create (oldsize
Question 2
When journal is a file and the journal file is opened, the following error is output:
2015-08-19 17:27:48.900894 7f1302791800 -1 journal FileJournal::_open: disabling aio for non-block journal. Use journal_force_aio to force use of aio anyway
That is, ceph does not use aio in this case. Why ???
Int FileJournal: _ o
1. Environment and descriptionDeploy ceph-0.87 on ubuntu14.04 server, set Rbdmap to mount/unload RBD block devices automatically, and export the RBD block of iSCSI with a TGT with RBD support. 2. Installing Ceph1) Configure hostname with no password login [Email protected]:/etc/ceph# cat/etc/hosts127.0.0.1localhost192.168.108.4 osd2.osd2osd2192.168.108.3 osd1.osd1osd1192.168.108.2 Mon0.mon0mon0 #示例如下ss
:
Monitoring: At that time, due to DNS configuration errors during the server configuration process, the monitoring email cannot be sent normally, so no Ceph WARN prompt message is received.
Cloud Platform itself: Due to Ceph mechanism, most of the distribution in the OpenStack platform is ultra-high. From the user's perspective, copying a large amount of data is not inappropriate, however, this problem
Ceph environment setup (2)1. There are three layout hosts: node1, node2, and node3. Each host has three osdks, as shown in figure. osd1, and 8 are SSD disks, 4 is the SATA disk. Each of the three hosts has a Monitor and an MDS. We use osd1, 3, and 4 to create a pool named ssd, in the form of three copies, osd0, and 2, 4 to build a Pool named sata, in the form of Erasure code, k = 2, m = 1, that is, use two osdks to store data fragments, one osd to sto
write time for other disk operations, resulting in a ceph deadlock on osd.
The solution is to disable the rbd persistence of redis.
A long-term solution is to prevent redis from writing data to the ceph partition persistently. In addition, do not write or read high IO images from the ceph Virtual Machine (unreliable ...)
Experience summary:
1.
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.