Tag: Availability
Re: Maintenance mode during Rebuild / Rebalance?
Hello guys,
Can I set Maintenance Mode to an SDS while a Rebuild/Rebalance ?
At ScaleIO 2.0 User guide I see the following:
To invoke maintenance mode, the following conditions are required:
– Only one Fault Unit (or standalone SDS) can be in maintenance mode at any given time.
– No other SDSs can be in degraded or failed state (force override can be used).
– There must be adequate space on other SDSs for the additional backup (force override can be used).
Note
Use of force override options when entering maintenance mode can lead to data unavailability while maintenance mode is activated
At the Managing Dell EMC ScaleIO Ready Nodes using Dell EMC OpenManage Essentials.pdf it mentions:
3.2.1 Meet prerequisites for maintenance
The following prerequisites should be met:
• ESXi servers should be in a cluster with high availability (HA), Distributed Resource Scheduling (DRS) and vMotion enabled. This will allow the administrator to update and reboot the server without impacting the virtual machines’ availability.
• The administrator must verify in the ScaleIO GUI that:
• No SDC or SDS is disconnected.
• No other Fault Unit (standalone SDS) is in Maintenance Mode.
There must be adequate space on other SDSs for additional backup.
• No rebuild or rebalance is running in the background.
• No degraded capacity exists.
• No SDS device is in error state.
Thank you in advance,
Kleanthis
Related:
Maintenance mode during Rebuild / Rebalance?
Hello guys,
Can I set Maintenance Mode to an SDS while a Rebuild/Rebalance ?
At ScaleIO 2.0 User guide I see the following:
To invoke maintenance mode, the following conditions are required:
– Only one Fault Unit (or standalone SDS) can be in maintenance mode at any given time.
– No other SDSs can be in degraded or failed state (force override can be used).
– There must be adequate space on other SDSs for the additional backup (force override can be used).
Note
Use of force override options when entering maintenance mode can lead to data unavailability while maintenance mode is activated
At the Managing Dell EMC ScaleIO Ready Nodes using Dell EMC OpenManage Essentials.pdf it mentions:
3.2.1 Meet prerequisites for maintenance
The following prerequisites should be met:
• ESXi servers should be in a cluster with high availability (HA), Distributed Resource Scheduling (DRS) and vMotion enabled. This will allow the administrator to update and reboot the server without impacting the virtual machines’ availability.
• The administrator must verify in the ScaleIO GUI that:
• No SDC or SDS is disconnected.
• No other Fault Unit (standalone SDS) is in Maintenance Mode.
There must be adequate space on other SDSs for additional backup.
• No rebuild or rebalance is running in the background.
• No degraded capacity exists.
• No SDS device is in error state.
Thank you in advance,
Kleanthis
Related:
Software Defined Storage Availability (Part 2): The Math Behind Availability
As we covered in our previous post ScaleIO can easily be configured to deliver 6-9’s of availability or higher using only 2 replicas that saves 33% of the cost compared to other solutions while providing very high performance. In this blog we will discuss the facts of availability using math and demystify the myth behinds ScaleIO’s high availability. For data loss or data unavailability to occur in a system with two replicas of data (such as ScaleIO) there must be two concurrent failures or a second failure must occur before the system recovers from a first failure. Therefore one of the following four scenarios must occur:
Let us choose two popular ScaleIO configurations and derive the availability of each.
Note: ScaleIO best practices recommend a maximum of 300 drives in a storage pool, therefore for the first configuration we will configure two storage pools with 240 drives in each pool. To calculate the availability of a ScaleIO system we will leverage a couple of well know academic publications:
We will adjust the formulas in the paper to the ScaleIO architecture and model the different failures. Two Drive FailuresWe will use the following formula to calculate the MTBF of ScaleIO system for a two drive failure scenario: Where:
Note: This formula assumes that two drives that fail in the same ScaleIO SDS (server) will not cause DU/DL as the ScaleIO architecture guarantees that replicas of the same data will NEVER reside on the same physical node. Let’s assume two scenarios – in the first scenario the rebuild process is constrained by network bandwidth – in the second scenario the rebuild process is constrained by drive performance bandwidth. Network BoundIn this case we assume that the rebuild time/performance is limited by the availability of network bandwidth. This will be the case if you deploy a dense configuration such as the DELL 740xd servers with a large number of SSDs in a single server. In this case, the MTTR function is: Where:
Plugging in the relevant values in the formula above, we get a MTTR of ~1.5 minutes for the 20 x R740, 24 SSDS @ 1.92TB w/ 4 X 10GbE network connections configuration (two storage pools w/ 240 drives per pool). The 20 x R640, 10SSDs @ 1.92TB w/ 2 X 25GbE network connections config provides MTTR of ~2 minutes. These MTTR values reflect the superiority of ScaleIO’s declustered RAID architecture that result in a very fast rebuild time. In a later post we will show how those MTTR values are critical and how they impact system availability and operational efficiency. SSD Drive BoundIn this case, the rebuild time/performance is bound by the number of SSD drives and the rebuild time is a function of the number of drives available in the system. This will be the case if you deploy less dense configurations such as the 1U Dell EMC PowerEdge R640 servers. In this case, the MTTR function is: Where:
System availability is calculated by dividing the time that the system is available and running, by the total time the system was running added to the restore time. For availability we will use the following formula: Where:
Note: the only purpose of RTO is to translate MTBF to availability. Node and Device FailureNext, let’s discuss the system’s MTBF when a node fails and followed by a drive failure, for this scenario we will be using the followed model: Where:
In a similar way, one can develop the formulas for other failure sequences such as a drive failure after a node failure and a second node failure after a first node failure. Network Bound Rebuild ProcessIn this case we assume that rebuild time/performance is constrained by network bandwidth. We will make similar assumptions as for drive failure. In this case, the MTTR function is: Where:
Plugging the relevant values in the formula above, we get a MTTR of ~30 minutes for the 20 x R740, 24 SSDS @ 1.92TB w/ 4 X 10GbE network connections configuration (two storage pools w/ 240 drives per pool). The 20 x R640, 10SSDs @ 1.92TB w/ 2 x 25GbE Network config provides MTRR of ~20 minutes. During system recovery ScaleIO rebuilt about 48TB of data for the first configuration and about 21TB for the second configuration. SSD Drive BoundIn this case we assume that the Rebuild time/performance is SSD drive bound and the rebuild time is a function of the number of drives available in the system. Using the same assumptions as for drive failures, the MTTR function is: Where:
Based on the provided formulas let’s calculate the availability of ScaleIO system based on the two different configurations: 20 x R740, 24 SSDS @ 1.92TB w/ 4 X 10GbE Network (Deploying 2 storage pools w/ 240 drives per pool)
20 x R640, 10SSDs @ 1.92TB w/ 2 x 25GbE:
Since these calculations are complex, ScaleIO provides its customers with FREE online tools to build HW configurations and obtain availability numbers that includes all possible failure scenarios. We advise customers to use this tool, rather than crunch complex mathematics, to build system configurations based on desired system availability targets. As you can see, yet again, we prove that the ScaleIO system easily exceeds 6-9’s of availability with just 2 replicas of the data. Unlike other vendors, neither extra additional data replicas nor erasure coding is required! So do you have to deploy three replica copies to achieve enterprise availability? No you do not! The myth is BUSTED. |
|||||||||||||||||||||||||||||||||||||
Update your feed preferences | |||||||||||||||||||||||||||||||||||||
Related:
Software Defined Storage Availability (Part 1): Why Do With Three What You Can Do With Two?
Must an enterprise deploy 3 or more replicas of data, or some form of erasure coding, to achieve enterprise class availability (of 99.9999% uptime)? Several software-defined storage (SDS) vendors seem to insist so. But is this true? We will methodically discuss this topic in a series of three blog posts and equip you with better knowledge on this complex and very important topic. When it comes to safety, aircraft manufacturers are a great example of an industry that goes to great lengths to ensure maximum passenger safety. When you boarded your last flight, did you board an aircraft powered by two jet engines or four? Do you generally feel unsafe when you fly aircrafts with two jet engines? Passengers, airliners and airline industry experts all agree that today’s two engine aircrafts provides an equal amount of safety as aircrafts with four jet engines. Why is this so? It is because innovations in engine and aircraft technologies have made a single engine very powerful and efficient – powerful enough to take-off, cruise or land using just a single engine. And once safety is ensured, superior economics determines the winner. As two engines are more economical than four, it makes them hugely popular with airliners (lower costs) and customers (lower ticket prices). The aircraft analogy is very similar to enterprise storage – enterprise customers can enjoy high levels of availability on well architected systems that require fewer copies of data, costing a lot less, but do not require any form of erasure coding. Dell EMC’s industry leading ScaleIO is one such system. Designed for demanding enterprises, ScaleIO is a data center grade software defined storage solution that regularly meets and exceeds six or even seven nines of availability. 5-9’s or 6-9’s? Definition of Enterprise Availability SLAsAvailability of a system is defined in terms how much time a system is up and running throughout the year. A popular measure of availability is in terms of percentages:
For example, 6-9’s of availability means that a system is unavailable for 31.5 seconds in a year. The ScaleIO MagicUsing ScaleIO you can easily build a ScaleIO cluster with 6-9’s of availability or more. ScaleIO’s unique declustered raid technology, its ability to quickly detect failures and perform fast rebuild allow our customers to get predictable and consistent performance and achieve 6-9’s or more availability with 33% lower storage capacity (and proportionate cost) than other vendors’ solutions that demand 3 copies of data. The ScaleIO architecture utilizes multiple mechanisms to achieve enterprise grade availability:
Read more on ScaleIO’s unique architecture here. Since these calculations are complex, ScaleIO provides its customers with FREE online tools to build HW configurations based on ScaleIO Ready Nodes to get comprehensive availability numbers that includes multiple possible failures scenarios. We advise customers to use this tool to build hardware configuration based on desired system availability target. If you are a ScaleIO customer, the tool can be accessed here. Here are some sample configurations each with availability of 6-9’s or higher
As you can see a variety of highly customizable configurations are possible with varying performance and capacity to meet our customer’s needs with every configuration assured of 6-9’s of availability or higher. So like two engine jets, where modern technology is now capable of providing equal safety as four engines, ScaleIO architecture is capable to deliver 6-9’s of availability requiring just 2 replicas of data, and resulting in 33% lower cost. So: why do with three what you can do with two? |
|||||||||||||||||||||||||||||||||||||||||
Update your feed preferences | |||||||||||||||||||||||||||||||||||||||||
Related:
New ECS HA Design white paper published
http://www.emc.com/collateral/whitepaper/h16344-elastic-cloud-storage-ha-design.pdf
Check out this new ECS HA Design white paper. It provides details about the technology behind how ECS provides enterprise availability such as:
• how the distributed infrastructure provides increased system availability
• advanced data protection methods that provide data durability such as erasure coding and triple mirroring
• how data is distributed for optimal availability
• automatic failure detection
• built-in self-healing methods
• disk, node and network failure resolution details
• disaster recovery – how ECS protects against site wide failures
• how consistency is maintained in an active-active multi-site configuration
• how site-wide failures are detected
• access options during a site outage
• how data durability is re-established after a permanent site-wide failure