1. The heartbeat is sent between the OSD
A. OSD every 6 seconds,
you can set the OSD heartbeat interval in [OSD] to modify
B. If you do not receive the heartbeat of the adjacent OSD within 20 seconds, The OSD status is considered down and reported to monitor.
OSD can be set in [OSD] Heartbead grace to modify
2. The OSD reports to monitor
A. OSD is continuously reported 3 times, after a
Ceph provides three storage methods: Object Storage, block storage, and file system. The following figure shows the architecture of the Ceph storage cluster:We are mainly concerned about block storage. In the second half of the year, we will gradually transition the virtual machine backend storage from SAN to Ceph. although it is still version 0.94,
above into ceph.confFsid = d437c823-9d58-43dc-b586-6b36cf286d4f3. Write the IP address of the initial monitor and the initial monitor to the Ceph configuration file, separated by commasMon Initial Members =monMon host =192.168.2.204. Create a key ring for this cluster and generate a monitor key.[[email protected]
node
1. Create a cluster
Ceph-deploy New Ceph-node1
2. Change the default number of copies in the Ceph configuration file from 3 to 2 so that only two OSD can reach active + clean status. Add the following line to the [Global] section:
OSD Pool Default size = 2
echo "OSD Pool Default size = 2" | sudo tee-a ceph.conf
3. If you have multiple network cards, you can
specification
1 About Ceph 1.1 Ceph definition
Ceph is a Linux PB-level Distributed File system. 1.2 ceph origin
Its name is related to the mascot of the UCSC (the birthplace of Ceph), the mascot is "Sammy", a banana-colored slug, a shell-free mollusk in the head-
Document directory
1. Design a CEpH Cluster
3. Configure the CEpH Cluster
4. Enable CEpH to work
5. Problems Encountered during setup
Appendix 1 modify hostname
Appendix 2 password-less SSH access
CEpH is a relatively new Distributed File System completed by the USSC storage team. It is a Network File System
command to synchronize and start the NTP service manually from the server:
# ntpdate 0.cn.pool.ntp.org
# hwclock-w #
Systemctl enable Ntpd.service
# Systemctl Start Ntpd.service
Install SSH service:
# yum Install Openssh-server
The second step, the preparation is done, now began to deploy the Ceph cluster.
Note: The following operations are performed at the Admin-node node, in this article, because Admin-node is shared with e1093, so it can be perf
Ceph environment setup (2)1. There are three layout hosts: node1, node2, and node3. Each host has three osdks, as shown in figure. osd1, and 8 are SSD disks, 4 is the SATA disk. Each of the three hosts has a Monitor and an MDS. We use osd1, 3, and 4 to create a pool named ssd, in the form of three copies, osd0, and 2, 4 to build a Pool named sata, in the form of Erasure code, k = 2, m = 1, that is, use two
ability to adapt to changes and provide optimal performance. It uses POSIX compatibility to complete all these tasks, allowing it to transparently deploy the applications that currently depend on POSIX semantics (through CEpH-oriented improvements. Finally, CEpH is an open-source distributed storage and part of the main Linux kernel (2.6.34.
Back to Top
CEpH
Ceph monitoring Ceph-dash Installation
There are a lot of Ceph monitors, such as calamari or inkscope. When I started to try to install these, they all failed, and then Ceph-dash came into my eyes, according to the official description of Ceph-dash, I personally think it is
the ceph environment.These files are generally available under the Ceph directory:Ceph.bootstrap-mds.keyring ceph.bootstrap-osd.keyring ceph.bootstrap-rgw.keyring ceph.client.admin.keyring ceph.conf Ceph.log ceph.mon.keyring RELEASE.ASC1.Start over:Ceph-deploy purgedata {Ceph-node} [{Ceph-node}] # #清空数据
Ceph Cluster Expansion and Ceph Cluster ExpansionThe previous article describes how to create a cluster with the following structure. This article describes how to expand the cluster.
IP
Hostname
Description
192.168.40.106
Dataprovider
Deployment Management Node
192.168.40.107
Mdsnode
MON Node
192.168.40.108
Osdnode1
OSD Node
192.168.40.14
that currently rely on POSIX semantics (through Ceph-targeted improvements). Finally, Ceph is open source distributed storage and is part of the mainline Linux kernel (2.6.34).Back to top of pageCeph ArchitectureNow, let's explore Ceph's architecture and the high-end core elements. Then I'll expand to another level, explaining some of the key aspects of Ceph and
the top of the page
Ceph Architecture
Now, let's explore the architecture of Ceph and the core elements of the high-end. Then I will extend it to another level to illustrate some of the key aspects of Ceph and provide a more detailed discussion.
The Ceph ecosystem can be roughly divided into four parts (see Figure 1):
3.3 verify that you can log on without a password through SSH
Ssh node1
Ssh node2
Ssh node3
Ssh client
4. Create a Monitor (admin node)
4.1 create a monitor on node1, node2, and node3
Mkdir myceph
Cd myceph
Ceph-deploy new node1 node2 node3
4.2 modify the number of copies of osd and add the default size of osd pool = 2 to the end
Vim/etc/
OverviewDocs:http://docs.ceph.com/docsCeph is a distributed file system that adds replication and fault tolerance while maintaining POSIX compatibility. The most characteristic of Ceph is the distributed meta-data server, which distributes the file location by crush (controlled Replication under scalable Hashing) this quasi-algorithm. The core of Ceph is the Rados (reliableautonomic distributed Object Store
: Ceph Monitor maintains the cluster map status, including monitor map, OSD map, Placement Group (PG) map, and CRUSH map. ceph maintains Ceph Monitors, Ceph OSD Daemons, and historical records of PGs status changes (called an "epo
code refactoring, you will see that there are a lot of mirrored versions. We've been building a separate image for each of the daemon (we'll do this when we integrate the patches). So the monitor, OSD, MDS and RADOSGW each have separate mirrors. This is not the ideal solution, so we are trying to integrate all the components into a mirror called daemon.This image contains all the modules that you can selectively activate from the command line while r
This is a creation in
Article, where the information may have evolved or changed.
In the article "using Ceph RBD to provide storage volumes for kubernetes clusters," we learned that one step in the integration process for kubernetes and Ceph is to manually create the RBD image below the Ceph OSD pool. We need to find a way to remove this manual step. The first th
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.