[TOC]k8s Mount Ceph RBDk8s Mount Ceph RBD There are two ways, one is the traditional way of PVPVC, which means that the administrator needs to pre-create the relevant PV and PVC, and then the corresponding deployment or replication to mount the PVC use. After k8s 1.4, Kubernetes provides a more convenient way to dynamically create PV, that is, Storageclass. Using Storageclass, you do not have to create a fi
Thomas is my pseudonym used by the Ceph China Community Translation team, which was first published in the Ceph China community. Now reproduced to my blog, for everyone to circulateCEPH ObjectStore API IntroductionThis article was translated by the Ceph China Community-thomas, Chen School Draft.English Source: The CEPH
based on the high-speed configuration of the Ceph storage cluster environment, it is possible to do related object operations:1. Set the OSD pool Min_sizefirst look at the pool with the Rados command such as the following:#rados LspoolsDatametadataRBDThe default OSD number of Min_size is configured to 2, where an OSD example is required to set it to 1ceph OSD Pool get {pool-name} {key}ceph OSD Pool set {Poo
Ceph source code analysis: Network Module
Because Ceph has a long history, the network does not use the epoll model that is currently commonly used, but uses a multi-threaded model similar to MySQL. Each connection (socket) has a read thread, read data from the socket continuously. A write thread writes data to the socket. Multi-threaded implementation is simple, but the concurrency performance is not flatt
Install times wrong: Failed to execute goal org.apache.maven.plugins:maven-antrun-plugin:1.7:run (site) on project Hadoop-hdfs:an Ant B Uildexception has occured:input file/usr/local/hadoop-2.6.0-stable/hadoop-2.6.0-src/hadoop-hdfs-project/ Hadoop-hdfs/target/findbugsxml.xml
Application and fault handling of CEPH in KVM virtualizationIn a distributed cluster, the user is provided with object storage, block storage and file storage.Benefits: Unified StorageNo single point of failureData multi-split redundancyScalable storage CapacityAutomatic fault tolerance and fault self-healingThree major role components of ceph and their rolerepresented as 3 daemonsCeph OSDMonitorMdsThere ar
Ceph's data management begins with the Ceph client's write operation, and since Ceph uses multiple replicas and strong consistency policies to ensure data security and integrity, a write request data is written to the primary OSD first and then primary The OSD further copies the data to the secondary and other tertiary OSD and waits for their completion notification before sending the final completion confi
Ceph RADOSGW Object Storage interface, research configuration for a long time, now share the following. The prerequisite for configuring RADOSGW first is that you have successfully configured the Ceph cluster to view the Ceph cluster through Ceph–s, in the health state. Here, the auth configuration of the
In the case of a network, Ubuntu installation software is very convenient, to install Ceph, also on a command to fix, want to install ceph0.72 on ubuntu14.04, because the official source of Ceph Ceph-extra does not contain ubuntu14.04 trusty package, Ceph, which uses 163 of its source, is an unwanted version, so it com
The pool is a logical partition of Ceph's stored data, and it acts as a namespace. Other distributed storage systems, such as MogileFS, Couchbase, and Swift, have the concept of pool, but they are called different. Each pool contains a certain number of objects in the PG,PG are mapped to different OSD, so pool is distributed throughout the cluster.
Apart from isolating data, we can also set different optimization strategies for different pool, such as number of replicas, number of data cleansing
Ceph practice summary: the configuration of the RBD block device client in Centos, cephrbd
Before proceeding to this chapter, you need to complete the basic cluster building, please refer to the http://blog.csdn.net/eric_sunah/article/details/40862215
Ceph Block devices are also calledRBDOrRADOS. Block Device
During the experiment, a virtual machine can be used as the
Dear friends, "the first phase of the Ceph China Community Training Course Open Class" Basics of Ceph Foundation and its principles and basic deployment1. Take you into the Ceph world, from principle to practice, so you quickly build your own ceph cluster.2. Take you step by step to find "object", see RBD Essence, play
First, collocated group status1. CreatingWhen you create a storage pool, it creates a specified number of collocated groups. CEPH displays creating when creating one or more collocated groups, and when created, the OSD in the acting set of its collocated group will be interconnected; Once the interconnect is complete, the Collocated group state should become active+clean, meaning that the Ceph client can wr
Before starting the text, I would like to thank Unitedstack Engineer Zhu Rongze for his great help and careful advice on this blog post. This paper makes a more explicit analysis and elaboration on the calculation method of Ceph reliability (https://www.ustack.com/blog/build-block-storage-service/) for Unitedstack company at the Paris summit. For the interests of this topic friends to discuss, research, the article if there is inappropriate, but also
Ceph practice summary: configuration of the RBD block device client in Centos
Before proceeding to this chapter, you need to complete the basic cluster building, please refer to the http://blog.csdn.net/eric_sunah/article/details/40862215
Ceph Block devices are also calledRBDOrRADOS. Block Device
During the experiment, a virtual machine can be used as the ceph-c
According to http://ovirt-china.org/mediawiki/index.php/%E5% AE %89%E8%A3%85%E9%83%A8%E7%BD%B2Ceph_Calamari
The original article is as follows:
Calamari is a tool for managing and monitoring CEpH clusters and provides rest APIs.
The recommended deployment platform is ubuntu. This article uses centos 6.5.Installation and deployment
Obtain calamari code# git clone https://github.com/ceph/calamari.git# g
Transferred from: https://www.ustack.com/blog/civetweb/The excellent open source project is changing the traditional It,openstack name most loudly, has become the IaaS the fact standard. Ceph is also a great achievement, with its three storage interfaces to meet the diverse needs of the enterprise. Unitedstack has a cloud that combines the benefits of an open source project, such as OpenStack and Ceph, to b
In planning the CEPH distributed storage cluster environment, the choice of hardware is very important, this is related to the performance of the entire Ceph cluster, the following comb to some of the hardware selection criteria, for reference:1) CPU SelectionCeph metadata server will dynamically redistribute the load, which is CPU sensitive, so metadata server should have better processor performance (such
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.