1. Environmental descriptionWith Kolla deployed Ceph, since the OSD 0 occupies the SATA 0 channel, the system disk needs to be swapped with the OSD 0 jumper, after the jumper switch, the OSD 0 does not start properly.2, Reason analysis:Before switching jumpers, the device file of the OSD 0 is/dev/sda2, the switch jumper becomes/dev/sdc2,osd at startup,--osd-journal/dev/sda2, specifies the log device, because the log partition device name becomes/DEV/S
ceph configuration parameters (i)6. Keyvaluestore CONFIG REFERENCEHttp://ceph.com/docs/master/rados/configuration/keyvaluestore-config-ref/KeyValueStore is an alternative OSD Backend compared to Filestore. Currently, it uses LevelDB as backend. Keyvaluestore doesn ' t need journal device. Each operation would flush into the backend directly.
Back end (LEVELDB) used by the Keyvaluestore:keyvaluestore Backend
(1) Queue
Maximum num
Ceph configuration parameters (2)Ceph configuration parameters (a) 6, keyvaluestore config reference http://ceph.com/docs/master/rados/configuration/keyvaluestore-config-ref/ KeyValueStore is an alternative OSD backend compared to FileStore. currently, it uses LevelDB as backend. keyValueStore doesn' t need journal device. each operation will flush into the backend directly. the backend (leveldb) used by Ke
The basic idea of the cache tiering is the separation of hot and cold data, the use of relatively fast/expensive storage devices such as SSD disks to form a pool as the cache layer, the backend with relatively slow/inexpensive devices to form a cold data storage pool.The Ceph cache tiering Agent handles automatic migration of data from the cache layer and storage layer, transparently to client transparent operations. The Cahe layer has two typical mod
IntroductionI need direct programmatic access to Ceph's object storage to see the performance difference between using gateways and without gateways. Examples of access based on Gate-way have gone through. Now the test is not to go to the gateway, with Librados directly with the Ceph cluster.Environment configuration1. Ceph cluster: You have a ceph cluster that i
If you need to add a host name: OSD4 ip:192.168.0.110 OSD1. Create a directory for Mount directories and placement profiles in OSD4SSH 192.168.0.110 (this is from Mon host ssh to OSD4 host)Mkdir/ceph/osd.4Mkdir/etc/ceph2. On OSD4, format the EXT4 Sda3 partition, mount the partition.Mkfs.ext4/dev/sda3Mount-o User_xattr/dev/sda3/ceph/osd.43. The id_dsa.pub of the Mon host is copied to the OSD4 host for passwo
[Analysis] Ceph programming instance interface Librbd (C ++) -- image creation and data read/write, cephlibrbd
Currently, we have two ways to use Ceph Block Storage :?-Use QEMU/KVM to interact with Ceph Block devices through librbd. This mainly provides block storage devices for virtual machines, as shown in ;?
-Use the kernel module to interact with the Host
Because the kubelet itself does not support RBD commands, a kube system plugin is required:Download Plugin Quay.io/external_storage/rbd-provisioner:Https://quay.io/repository/external_storage/rbd-provisioner?tag=latesttab=tagsDownload Docker pull quay.io/external_storage/rbd-provisioner:latest on node of k8s clusterInstall only the plugin itself will error: need to install kube roles and permissions The following are:Https://github.com/kubernetes-incubator/external-storageHttps://github.com/kube
Using the Ceph OSD Tree command to view the Ceph cluster, you will find weight and reweight two values
Weight weight and disk capacity, general 1T, value is 1.000, 500G is 0.5
It is related to the capacity of the disk and does not change due to reduced disk available space
It can be set with the following command
Ceph OSD Crush Reweight
The reweight is a v
, error detection, error recovery will bring great pressure to the client, controller, metadata directory node, and limit the scalability.We have designed and implemented Rados, a reliable, automated distributed object store that seeks to distribute device intelligence to complex thousands of-node-scale clusters involving data-consistent access, redundant storage, error detection, and recovery of logging problems. As part of the Ceph Distributed syste
Cachetier is a ceph server-side caching scheme, simply add a layer of cache layer, the client directly with the cache layer to deal with , improve access speed, the backend has a storage layer, Actually store large amounts of data. The principle of tiered storage is that the access to the stored data is hot, and the data is not evenly accessed. There is a general rule called the 28 principle, that is, 80% 's application only accesses 20% data, this 20
past interval. Last_epoch_started: Last peering after the Osdmap version number epoch. Last_epoch_clean: Last recovery or backfill after the Osdmap version number epoch. (Note: After the peering is finished, the data recovery operation is just beginning, so last_epoch_started and Last_epoch_clean may differ). For example: The current epoch value of the Ceph system is pg1.0, and the acting set and up set are all [0,1,2]
Osd.3 failure resulte
Keyvaluestore is another storage engine that Ceph supports (the first is Filestore), which is in the Emporer version of Add LevelDB support to Ceph cluster backend store Design Sum At MIT, I put forward and implemented the prototype system, and achieved the docking with ObjectStore in the firely version. is now incorporated into Ceph's Master. Keyvaluestore is a lightweight implementation relative to Filest
I recently needed to test the rgw of CEpH in my work, so I learned while testing.
First, the tool uses Intel's open-source tool cosbench, which is also the industry's mainstream object storage testing tool.
1. cosbench installation and startupDownload the latest cosbench packageWget https://github.com/intel-cloud/cosbench/releases/download/v0.4.2.c4/0.4.2.c4.zipExtractUnzip 0.4.2.c4.zip
Install related ToolkitYum install java-1.7.0-openjdk NMAP-ncat
Have no time to write on business trip ...Create 150 OSD Today, find manual write ceph.conf a bit big, researched the increment function of vim.Very simple is a command:: Let I=0|g/reg/s//\=i/|let i=i+1It can match the Reg in your text and then follow your i+n, increasing +n per passThe function of the above command is to find the Reg character in the text, then replace it with 0 from the first, then +1So in the ceph.conf, we can first copy out 150 [OSD.GGGG], and then in the use of the above co
This is a creation in
Article, where the information may have evolved or changed.
All the places involved in storage are very prone to "pits", Kubernetes is no exception.
First, the cause of the problem
The problem began yesterday by upgrading the operation of a stateful service. The pod under the service is mounted with a persistent Volume provided with Ceph RBD. The pod is deployed with normal deployment and does not use the Petset in Alpha state. T
I. OverviewCephfs is a CEpH cluster-based file system that is compatible with POSIX standards.When creating a cephfs file system, you must add the MDS service to the CEpH cluster. This service processes the metadata part in the POSIX file system, and the actual data part is processed by the osds in the CEpH cluster.Cephfs supports loading by using INCORE modules
In centos6.3, to use CEpH block device, you must install a later version of qemu.Install qemu-1.5.2 after CEpH is installed# Tar-xjvf qemu-1.5.2.tar.bz2# Cd qemu-1.5.2#./Configure -- enable-RBDYou must add the -- enable-RBD option so that qemu can support the RBD protocol.In this step, an error may be reported:Error: user requested feature rados block DeviceConfigure was not able to find itThis is because t
The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion;
products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the
content of the page makes you feel confusing, please write us an email, we will handle the problem
within 5 days after receiving your email.
If you find any instances of plagiarism from the community, please send an email to:
info-contact@alibabacloud.com
and provide relevant evidence. A staff member will contact you within 5 working days.