Glusterfs Six-volume mode to say

Source: Internet
Author: User
Tags glusterfs gluster

Glusterfs six volumes to explain

First, the distribution volume

The distributed volume file is randomly distributed across the entire brick volume. With distributed volumes, you need to extend storage, redundancy is important, or provide additional hardware/software layers. (Simplified: Distributed volumes, where files are randomly distributed through hash algorithms to volumes made up of bricks. The resources in the volume are stored only on one server, in non-mirrored or striped mode in the storage pool. )

(In a distributed volumes files is spread randomly across the bricks in the volume. Use distributed volumes where are need to scale storage and redundancy are either not important or are provided by other Har Dware/software layers.

Note

Disk/server failure in distributed volumes can result in a serious loss of data because directory contents is spread rand Omly across the bricks in the volume.)

(pictured, File1 and File2 are stored in Server1, while File3 is stored in server2.) )

Create the distributed Volume:

# Gluster Volume Create New-volname [transport [TCP | rdma | TCP,RDMA]]

New-brick ...

For example, to create a distributed volume with four storage servers using TCP:

# Gluster Volume Create Test-volume server1:/exp1 server2:/exp2 server3:/exp3 server4:/exp4

Creation of Test-volume has been successful

Please start the volume to access data.

(Optional) You can display the volume information:

# Gluster Volume Info

Volume Name:test-volume

Type:distribute

status:created

Number of Bricks:4

Transport-type:tcp

Bricks:

Brick1:server1:/exp1

Brick2:server2:/exp2

Brick3:server3:/exp3

Brick4:server4:/exp4

Second, the volume of the Restoration

The copy volume creates a copy of the file that spans the volume of multiple bricks. You can use replication volumes in environments where high availability and high reliability are critical. (Description: A replicated volume, similar to the number of Raid1,replica must be equal to the number of storage servers contained in the volume brick, high availability.) Create a 22 volumes that are backed up by a single hard disk in the storage pool that does not affect the use of data and requires a minimum of two servers to create a distributed mirrored volume. )

(Replicated volumes create copies of files across multiple bricks in the volume. You can use the replicated volumes in environments where high-availability and high-reliability is critical.

Note

The number of bricks should is equal to of the replica count for a replicated volume. To protect against server and disk failures, it's recommended that the bricks of the volume was from different servers.)

(for example, File1 is stored on Server1 and Server2 at the same time.) The same is true of File2. The file in Server2 is a copy of the file in Server1. )

Create the replicated Volume:

# Gluster Volume Create New-volname [replica COUNT] [Transport [TCP |

RDMA | TCP,RDMA]] New-brick ...

For example, to create a replicated volume with the storage servers:

# gluster Volume Create Test-volume Replica 2 transport TCP SERVER1:/EXP1 SERVER2:/EXP2

Creation of Test-volume has been successful

Please start the volume to access data.

Third, strip reel

The volume of data between striped strips of stripe tiles. For best results, you should use striped volumes and access very large files only in highly concurrent environments. (Simple: Striped volume, similar to the number of raid0,stripe must be equal to the number of storage servers contained in the volume brick, the file is divided into chunks, round robin is stored in bricks, and the granularity is the data block, large file performance is good. )

(striped volumes stripes data across bricks in the volume. For best results, you should the use striped volumes only in high concurrency environments accessing very large files.

Note

The number of bricks should is a equal to the stripe count for a striped volume.)

(For example, file is divided into 6 segments, 1,3,5 is stored in Server1, 2,4,6 is stored in server2.) )

Create the Striped Volume:

# Gluster Volume Create new-volname [stripe COUNT] [Transport [TCP |

RDMA | TCP,RDMA]] New-brick ...

For example, to create a striped volume across, storage servers:

# gluster Volume Create Test-volume Stripe 2 Transport TCP SERVER1:/EXP1 SERVER2:/EXP2

Creation of Test-volume has been successful

Please start the volume to access data.

For example:

IV, Split-type strip reel (retro-fit)

Distributed striped tape files are two or more nodes in a cluster that are two or more. For best results, you should use a distributed stripe volume requirement that is critical for extended storage and high access to very large files in a concurrent environment. (Simplified: Distributed striped volumes, brick in volume must be a multiple of stripe (>=2 times), with both distributed and striped functionality. Each file is distributed across four shared servers and is typically used for large file access processing, requiring a minimum of 4 servers to create distributed striped volumes. )

(Distributed striped volumes stripes files across both or more nodes in the cluster. For best results,you should use distributed striped volumes where the requirement are to scale storage and in high Concurre ncy environments accessing very large files is critical.

Note

The number of bricks should is a multiple of the stripe count for a distributed striped volume.)

Create The Distributed Striped Volume:

# Gluster Volume Create new-volname [stripe COUNT] [Transport [TCP |

RDMA | TCP,RDMA]] New-brick ...

For example, to create a distributed striped volume across eight storage servers:

# gluster Volume Create Test-volume Stripe 4 Transport TCP SERVER1:/EXP1 SERVER2:/EXP2

Server3:/exp3 server4:/exp4 server5:/exp5 server6:/exp6 server7:/exp7 SERVER8:/EXP8

Creation of Test-volume has been successful

Please start the volume to access data.

V, distributed Recovery Volume (recovery type)

Allocate the volume of the file in the copy tile. You can use distributed replication volumes to require a scale environment where storage and high reliability are critical. Distributed replication volumes also provide better read performance in most environments

。 (Description: Distributed replication volumes, brick in volume must be a multiple of replica (>=2 times), with both distributed and replicated functions. )

(distributes files across replicated bricks in the volume. You can use distributed replicated volumes in environments where the requirement are to scale storage and high-reliability Is critical. Distributed replicated volumes also offer improved read performance in most environments.

Note

The number of bricks should is a multiple of the replica count for a distributed replicated volume. Also, the order in which bricks is specified have a great effect on data protection. Each replica_count consecutive bricks in the list you give would form a replica set, with all replica sets combined into a Volume-wide distribute set. To make sure this replica-set members is not placed on the same node, list the first brick on every server and then the SECO nd brick on every server in the same order, and so on.)

Create the distributed replicated Volume:

# Gluster Volume Create New-volname [replica COUNT] [Transport [TCP |

RDMA | TCP,RDMA]] New-brick ...

For example, four node distributed (replicated) volume with a two-way mirror:

# gluster Volume Create Test-volume Replica 2 transport tcp SERVER1:/EXP1 SERVER2:/EXP2 SERVER3:/EXP3 SERVER4:/EXP4

Creation of Test-volume has been successful

Please start the volume to access data.

For example, to create a six node distributed (replicated) volume with a two-way mirror:

# gluster Volume Create Test-volume Replica 2 transport tcp SERVER1:/EXP1 SERVER2:/EXP2 SERVER3:/EXP3 SERVER4:/EXP4 Server 5:/EXP5 SERVER6:/EXP6

Creation of Test-volume has been successful

Please start the volume to access data.

VI, Strip and roll (type)

Strip-and-reel-strip data in a copy of the bricks in the cluster. For best results, it is critical that you use striped replication volumes to access very large files and performance in parallel in high concurrency environments. In this release, this type of volume configuration only supports map reduction effort.

(Striped replicated volumes stripes data across replicated, bricks in the cluster. For best results, should use striped replicated volumes in highly concurrent environments where there is parallel acce SS of very large files and performance is critical. In this release, the configuration of this volume type was supported only for Map Reduce workloads.

Note

The number of bricks should is a multiple of the replicate count and stripe count for a striped replicated volume.)

Create a striped replicated volume:

# Gluster Volume Create New-volname [stripe count] [replica count]

[Transport [TCP | rdma | TCP,RDMA]] New-brick ...

For example, to create a striped replicated volume across four storage servers:

# gluster Volume Create test-volume stripe 2 Replica 2 transport TCP SERVER1:/EXP1

SERVER2:/EXP2 Server3:/exp3 SERVER4:/EXP4

Creation of Test-volume has been successful

Please start the volume to access data.

To create a striped replicated volume across six storage servers:

# gluster Volume Create test-volume stripe 3 replica 2 transport TCP SERVER1:/EXP1

SERVER2:/EXP2 server3:/exp3 server4:/exp4 server5:/exp5 server6:/exp6

Creation of Test-volume has been successful

Please start the volume to access data.

For example:

For example:

VII, distributed stripe-to-reel (three hybrid)

Distributed striped copy volume distribution stripe data in a copy brick cluster. For best results, you should use a highly concurrent striped replicated volume environment to access very large files and performance in parallel is critical. In this release, this type of volume configuration only supports map reduction effort.

(Distributed striped replicated volumes distributes striped data across replicated bricks in the cluster. For best results, should use distributed striped replicated volumes in highly concurrent environments where parallel a Ccess of very large files and performance is critical. In this release,configuration of this volume type are supported only for MAP Reduce workloads.

Note

The number of bricks should is a multiples of number of stripe count and replica count for adistributed striped replicated Volume.)

Create A distributed striped replicated volume using the following command:

# Gluster Volume Create New-volname [stripe count] [replica count]

[Transport [TCP | rdma | TCP,RDMA]] New-brick ...

For example, to create a distributed replicated striped volume across eight storage servers:

# gluster Volume Create test-volume stripe 2 Replica 2 transport TCP SERVER1:/EXP1

SERVER2:/EXP2 server3:/exp3 server4:/exp4 server5:/exp5 server6:/exp6 SERVER7:/EXP7

Server8:/exp8

Creation of Test-volume has been successful

Please start the volume to access data.

RAID technology mainly includes RAID 0 ~ RAID 7 A number of specifications, their focus is different, common norms have the following several :

RAID 0:raid 0 continuously splits data in bits or bytes, reads/writes on multiple disks in parallel, and therefore has a high data transfer rate, but it has no data redundancy and is not considered a real RAID structure. RAID 0 simply improves performance and does not guarantee the reliability of the data, and one of the disk failures will affect all data. Therefore, RAID 0 cannot be applied to situations where data security requirements are high.

RAID 1: It is data redundancy through disk data mirroring, resulting in mutually backed-up data on paired independent disks. When raw data is busy, data can be read directly from the mirrored copy, so RAID 1 can improve read performance. RAID 1 is the highest unit cost in a disk array, but provides high data security and availability. When a disk fails, the system can automatically switch to read and write on the mirrored disk without having to reorganize the failed data.

RAID 0+1: Also known as the raid 10 standard, is actually the product of the combination of RAID 0 and RAID 1 standards, which is used for redundancy of disk mirroring for each disk while continuously splitting the data in bits or bytes and concurrently reading/writing multiple disks. It has the advantage of having RAID 0 at the same time with the extraordinary speed and RAID 1 data high reliability, but the same CPU usage is also higher, and the disk utilization ratio is low.

RAID 2: The data is striped across different hard disks, in bits or bytes, and is used to provide error checking and recovery using coding techniques called "Weighted average error correction code (cleartext)". This coding technique requires multiple disk storage checks and recovery information, making RAID 2 technology more complex to implement and therefore rarely used in a business environment.

RAID 3: It is very similar to Raid 2, where data is striped across different hard disks, except that RAID 3 uses simple parity and holds parity information with a single disk. If a disk fails, the parity disk and other data disks can regenerate the data, and if the parity disk fails, it does not affect data usage. RAID 3 provides a good transfer rate for large amounts of continuous data, but for random data, parity disks can be a bottleneck for write operations.

RAID 4:raid 4 also data is striped and distributed across different disks, but the bars are in blocks or records. RAID 4 uses a single disk as the parity disk, and each write operation requires access to the parity disk, where the parity disk becomes the bottleneck of the write operation, so RAID 4 is rarely used in a commercial environment.

RAID 5:raid 5 does not specify the parity disks separately, but instead accesses data and parity information across all disks. On RAID 5, the read/write pointer can operate against a list of devices at the same time, providing higher data traffic. RAID 5 is more suitable for small data blocks and random read and write data.

The main difference between RAID 3 and RAID 5 is that RAID 3 involves all array disks for each data transfer, whereas for RAID 5, most data transfers operate on only one disk and can be performed in parallel. In RAID 5 There is a "write loss", that is, each write operation will produce four actual read/write operations, two reads the old data and parity information, two times write new data and parity information.

RAID 6: Raid 6 Adds a second independent parity information block compared to RAID 5. Two independent parity systems use different algorithms, and the data is very reliable, even if two disks fail at the same time without affecting the use of the data. However, RAID 6 requires more disk space allocated to parity information and a greater write loss than RAID 5, so "write performance" is very poor. Poor performance and complex implementations allow RAID 6 to be rarely used in practice.

RAID 7: This is a new RAID standard, with its own intelligent real-time operating system and software tools for storage management, can be completely independent of the host running, do not occupy the host CPU resources. RAID 7 can be seen as a storage computer (Storage computer), which differs significantly from other RAID standards. In addition to the various criteria above, we can use RAID 0+1 as a combination of various RAID specifications to build the required RAID array, such as RAID 5+3 (RAID 53) is a more widely used array form. Users typically have the flexibility to configure disk arrays to obtain more disk storage systems that meet their requirements.



Reprint Please specify: Http://blog.163.com/[email protected]/blog/static/62713185201351510303223/

Glusterfs Six-volume mode to say

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.