INTRODUCTION
Cloud Computing may be defined as a model for enabling on-demand network access to a shared pool of configurable computing resources that can be rapidly provisioned and released with minimal management effort or service provider interaction (NIST, 2013). This resource set is typically utilized as a service model where in the client only pays for what is consumed. Video on Demand (VoD) streaming services are run on cloud infrastructure platforms in order to offer cost savings, easy scalability, and high availability to users. To evaluate the availability of the service, hirerarchical analytical models are employed to represent the architecture (Dantas et al., 2012).
In this paper, we propose an availability model of a VoD service based on Eucalyptus cloud environment, for evaluating the sensitivity of VoD service components. For guiding the implementation of system improvements, Design of Experiments (DoE) and percentage difference were used for identifying availability bottlenecks of the VoD service.
The remainder of the paper is organized as follows: Section II introduces related works on system availability and sensitivity analysis, Section III introduces basic concepts of cloud computing technologies, video streaming, dependability models and analysis and sensitivity analysis. Section IV presents the architecture of the system analyzed in this paper. Section V presents the availability model designed for architecture. Section VI conducts a case study about architecture analysis and presents sensitivity analysis. Finally, Section VII shows the conclusions of the study and suggests possible future work.
Recent research have employed hierarchical modeling to represent cloud computing architectures, it is possible to compare the different solutions and estimations of dependability measures (Dantas et al., 2012; Chuob et al., 2011). Moreover, some works have also employed sensitivity analysis to identify the critical system components, and thereby propose infrastructure improvements (Matos et al., 2012). Khazaei et al. (2012) integrated an availability model in overall analytical sub-models of cloud system. Each sub-model captures a specific aspect of cloud centers. The key performance metrics such as task blocking probability and total delay incurred on user tasks are obtained.
Choub et al. (2011) proposed a private cloud, with a modeling of the cloud computing based upon the Eucalyptus architecture. To understand the behavior of Eucalyptus, it was considered Ubuntu Enterprise Cloud (UEC) as reference architecture for our Cloud Test bed environment in Lab. With UEC architecture, it was addressed the availability of each component of the cloud base on Markov chain through the level analysis of hierarchical available model (HAM). Malik et al. (2013) provided a formal analysis modeling and verification of open source state-of-art VM-based cloud management plataforms to model and analyze the structural and behavioral properties of the systems have used high-Level Petri Nets. In (Araujo et al., 2014), an availability model of a digital cloud library through a OpenNebula cloud manager is proposed, using an hierarchical approach to model and evaluate the digital library environment. Measurements were performed to obtain the availability parameters of the library service deployed in a private cloud. In this paper, the authors proposed availability models applied to a cloud environment for VoD streaming service. Design of Experiments (DoE) and percentage difference were applied in order to find the bottlenecks of system availability.
BACKGROUND
This section presents the concepts which provide a background for this paper, including: cloud computing technologies, video streaming, dependability modeling and analysis and sensitivity analysis techniques.
A cloud computing system is comprised of a bundle of resources, such as hardware, software, development platforms and services, readily usable and accessible through the Internet (Armbrust et al., 2010). Services can be provisioned on different levels by cloud computing providers, including Infrastructure as a Service (IaaS), Platform as a Service (PaaS), and Software as a Service (SaaS). In a cloud computing environment, virtual resources can also be dynamically allocated and resized to deal with a varying workload, enabling optimal use of physical resources. Cloud computing services are typically accessed through a pay-per-use model (Pereira, 2010) and the provider offers is determined and guaranteed by means of service level agreements.
Eucalyptus is a software architecture based on Linux that leverage the implementation of private and hybrid IaaS clouds. This means that users can utilize their own collections of resources (hardware, storage and network) through a selfservice interface according to their needs. The Eucalyptus software framework is modular (Eucalyptus Systems, 2010), and consists of five high level components, each with its own Web Service: Cloud Controller (CLC), Cluster Controller (CC), Node Controller (NC), Storage Controller (SC), e Walrus (Eucalyptus Systems, 2010).
Video Streaming
Video streaming is a technology used in the transmission of digital multimedia contenton the Internet (Delgado et al., 2006). Streaming enables data to be sent and flow without the need to wait for content to load completely. This requires smaller network bandwidth and less storage space. As the multimedia data arrives, it is stored in a fast buffer before execution starts. Encoding, protocols and buffering mechanism are factors that might affect transmission of a video streaming (Diaz-Sanchez et al., 2011). The video streaming service works with the protocol Real Time Streaming Protocol (RTSP) which allows the control in the data transfer with real time properties. The RTSP makes the transfer possible on demand of real time data with audio and video (Valeriana and Marcelo, 2008).
Models for Dependability Analysis
In early 1980s Laprie coined the term dependability for encompassing concepts such reliability, availability, safety, confidentiality, maintainability, security and integrity etc. (Laprie, 1992; Laprie, 1995), whereas reliability of system at \(t\) is the probability that the system performs its functions without failing up to time instant \(t\). Availability can be expressed as the ratio of the expected system uptime to the expected system up and downtime: \(A\ = E\lbrack Uptime\rbrack/(E\lbrack Uptime\rbrack + E\lbrack Downtime\rbrack)\). It may also be represented by: \(A\ = \ MTTF/(MTTF\ + MTTR)\) where \(\text{MTTF}\) and \(\text{MTTR}\) is the mean time to failure and to recovery, respectively.
There are several types of models that can be used for analytical evaluation of dependability. Reliability Block Diagrams (RBD), Fault Trees, Stochastic Petri Nets (SPN) and Continuous Time Markov Chains (CTMC) have been used to model fault-tolerant systems and evaluate various dependability measures.
Sensitivity Analysis
The Design of Experiments, known as DoE, is method used to perform sensitivity analysis. Through this method it is possible to assess the importance of each parameter of the system and furthermore it can be used to simultaneously determine the individual and interactive effects of many factors that may affect the output measures. In the DoE method, parameters are called factors and the value that is assigned to each factor is called a level. There are numerous types of DoE and among the most commonly used are: full factorial design, fractional factorial design and simple design (Jain, 1991). Sometimes the number of experiments required for a full factorial design is too large. This may happen if either the number of factors or their levels is large. It may not be possible to use a full factorial design due to the cost or the time required. In such cases, one can only use a fraction of the full factorial design (Mathews, 2004). Fractional factorial designs are commonly used to reduce the number of runs required to build an experiment (Mathews, 2004) In this study we adopted the fractional factorial design, because there are more than five variables that we need to analyze. For \(2^{k}\) designs with five or more variables consider using the fractional factorial designs to reduce the number of runs required by the experiment.
SYSTEM ARCHITECTURE
The system architecture is based on Eucalyptus cloud computing platform. A broad view of the components of the VoD service architecture is seen in Figure 1 (Melo et al., 2017).
The VoD architecture is divided into two sides: client and the server. The physical structure is composed of three machines. One machine is used for frontend and two machines for the nodes. The client connects to the video streaming server through the Internet. A storage volume is allocated in the frontend for storing the collection of videos. A Virtual Machine (VM), running the Apache and VLC applications, is instantiated in the nodes. VLC provides the video streaming features, whereas Apache is responsible for hosting the service on a dedicated Web page. The user issues a request for displaying a video hosted on a specific web page. VLC, in turn, grabs the requested video from the remote storage volume and transmits the stream to the user.
AVAILABILITY MODELS
This section discusses the availability models employed to represent the redundant architecture evaluated by this research, from which the availability values were calculated. Such values were obtained by the hierarchical combinatorial method, which combines the system state representation of Markov chains with RBDs (Sahner and Trivedi, 1987; Kim et al., 2009), and is the method commonly employed to evaluate complex IT systems. The systems were modeled and evaluated with the SHARPE (Silva et al., 2012) and Mercury tools (MoDCS, 2017), which were specifically designed for the analysis of such models.
Model for Architecture
RBD and CTMC models were used to represent the subsystems of the architectures presented in Figure 1. These models are then combined, constituting a hierarchical model. The architecture can be divided into three parts: Frontend, Node and Service, which are represented by the RBD model depicted in Figure 2. In Figure 2, we made a union the service and nodes subsystem to represent the components of the architecture of the Figure 1. This junction was performed because the depedence which exists between the service and nodes subsystem.
Volume subsystem is allocated in the frontend for storing the collection of videos. The Service subsytem is further refined by a CTMC (see Figure 3), which allows to compute availability values to be considered top-level in the RBD. A CTMC was proposed due to the interdependency between the system’s components.
In the top-level model the service as well as the node subsystem infrastructure is represented by the Service RBD block. However the availability of such a sub-system (service + node subsystem infrastructure) cannot be properly represented by an RBD since the node subsystem implements an active redundant mechanism.
Therefore, the Service RBD block is refined by the CTMC depicted in Figure 3, which represents the service availability of the node subsystem infrastructure. The CTMC comprises the states UUW, UDU, UUD, and UWU (service available), and the states DDW, DUW, DDD, DWU, DDU, DWD, and DUD (service unavailable).
The notation for the states is based on the current condition of each component. The three letters represent initialisms of the operating condition of the three components, respectively, the service, the first node, and the second node. The service may be up (U) or down (D). The NCs work by being alternately in warm standby mode, and only one of them should be up (U) at any one time, whilst the other is either in warm standby (W) or down (D). In this model the initial service is represented by UWU, where the service is available, the first node in warm standby, and the second node is running. From this state it is possible to move to DWU (service failure), DWD (second node failure), or UDU (first node failure). From the DWU state (service down, first node in warm standby and second node up), may be reached UWU (representing service repair), DWD (second node failure), or DDU (first node failure).
From state DWD three outcomes are possible; either the failure of all system components (DDD), the initialization of the first node (DUD) or the repair of the second node (DWU). From state UDU (service and second node running), the possible outcomes are DDD (failure of all components), DDU, or UWU. The state UDU can lead to either the failure of all system components (DDD), the repair to waiting state of the first node (DWU), or the instantiation of a new virtual machine with all system applications, making the service available again (UDU).
In state DDD all system components are down; the service is unavailable and the two nodes are unavailable. From this state it is possible to reach two other states; repair of the first node (DUD) and the repair of the second node (DDU). State DUD represents service unavailability, where service and second node are down, but the first node is up. From this state, three other states can be achieved; failure of all components of the system (DDD), repair of the service (UUD), or repair of the warm standby mode of the second node (DUW). Conversely, state UUD indicates system availability, where the service and the first node are up, but the second node is faulty. From here, the following three states can be reached; DUD (service failure), UUW (repair of the node to warm standby mode), and DDD (since failure of the only functional mode will automatically cause service failure too).
In state DUW the system is unavailable due to service application failure, although the first node is up and the second node is in warm standby mode. From DUW the following states can be reached; DUD (failure of the warm standby node), DDW (first node failure), or UUW (service repair). With service and first node operational, and the second node in warm standby, UUW indicates system availability. From this position in the model, the possibilities are warm standby failure (UUD), service failure (DUW) or first node failure, which would cause the service to become unavailable (DDW). The state DDW indicates system unavailability, with service and first node down, and second node in warm standby. From DDW, it is possible to reach three other states; failure of all the components of the system (DDD), initialization of the second node (DDU), or repair of the first node (DUW).
System failure is an event that occurs when the provided service deviates from the intended service (Silva et al., 2012; MoDCS, 2017). The failure rates of the two nodes are represented by \(\lambda_{N}\), whilst \(\mu_{N}\) represents the rate of node repair. A node in warm standby has the failure rate of \(\lambda_{\text{WN}}\), and the repair rate to return it to standby is \(\mu_{N}\). A warm standby node is transformed to available mode at the rate of \(\mu_{\text{WN}}\). The failure rate of the service application is \(\lambda_{V_{0}D}\), while the repair rate is \(\mu_{V_{0}D}\). The \(\lambda_{V_{0}D}\) was obtained from the inverse of the time to failure of the service module. To calculate this result, we used the CTMC model of Figure 3 and the Mercury tool (Silva et al., 2012; MoDCS, 2017). The repair rate of the service is considered as the instantiation of a new virtual machine, including all the applications necessary to its operation (Apache and VLC).
Through the CTMC of Figure 3 it is possible to obtain the closed-form equation for calculating the availability of the redundant model (\(A_{V_{0}D}\)) as shown below in Equation 1. The equation for calculating the availability of the whole infrastructure of the service (\(A_{V_{0}D}\)) can also be obtained from its corresponding RBD model (Dantas et al., 2012).
|
\[A_{V_{0}D} = \frac{\alpha\left( 2\beta^{2}\alpha_{1} + \beta\alpha_{2}\beta_{1} + \alpha_{3} \right)\beta_{1}^{2} + \alpha_{4}\beta_{1}^{3}}{\alpha_{5}\left( \beta^{2}\alpha_{1}\left( \lambda N + 2\mu N \right) + \beta\varphi\beta_{1} + \varphi_{1}\beta_{1}^{2} + \beta_{7}\beta^{3} \right)}\] |
(1) |
where:
\(\beta = \lambda_{\text{WN}}\),
\(\beta_{1} = \mu_{\text{WN}}\),
\(\alpha = \mu N\mu_{V_{0}D}\),
\(\alpha_{1} = \left( \lambda_{N} + \beta + \mu_{N} \right)^{2}\left( \lambda_{N} + \beta + \mu_{N} \right)\),
\(\alpha_{2} = 4\lambda_{N}^{2} + 17\lambda_{N}\beta + 13\beta + 5\lambda\mu_{N} + 12\beta\mu_{N} + 2\mu_{N}^{2}\),
\(\alpha_{3} = \left( 8\beta\left( 2\beta + \mu_{N} \right) + \lambda_{N}\left( 11\beta + \mu_{N} \right) \right)\),
\(\alpha_{3} = \left( 8\beta\left( 2\beta + \mu_{N} \right) + \lambda_{N}\left( 11\beta + \mu_{N} \right) \right)\),
\(\alpha_{4} = \left( 7\beta + 2\mu_{N} \right)\left( \mu_{\text{WN}}^{3} \right)\),
\(\alpha_{5} = \lambda_{N} + \lambda_{V_{0}D} + \mu_{V_{0}D}\),
\(\alpha_{6} = 2\lambda_{N}\left( \lambda_{N} + \beta \right)\left( \lambda_{N} + 3\beta \right)\),
\(\beta_{2} = \left( 8\lambda_{N}^{2} + 24\lambda_{N}\lambda_{\text{WN}} + 13\beta^{2} \right)\mu_{N}\),
\(\beta_{3} = \left( 7\lambda_{N} + 12\beta \right)\mu_{N}^{2} + 2\mu_{N}^{3}\),
\(\beta_{4} = 2\lambda_{N}\beta\left( 2\lambda_{N} + 3\beta \right)\),
\(\beta_{5} = \left( \lambda_{N} + \beta \right)\left( \lambda_{N} + 16\beta \right)\mu_{N}\),
\(\beta_{6} = \left( \lambda_{N} + 8\beta \right)\mu_{N}^{2}\),
\(\beta_{7} = \left( \lambda_{N}\left( \beta + 2\mu_{N} \right) + \mu_{N}\left( 7\beta + 2\mu_{N} \right) \right)\),
\(\varphi = \alpha_{6} + \beta_{2} + \beta_{3}\) and
\(\varphi_{1} = \beta_{4} + \beta_{5} + \beta_{6}\).
A closed-form equation for computing the availability of the complete service (\(A_{\text{Service}}\)) can also be obtained, as demonstrated by Equation 2. \(A_{f}\) (availability of the frontend) and \(A_{v}\) (availability of the volume) can be computed from the RBD of Figure 2, whilst \(A_{V_{0}D}\) is calculated from Equation 1. In this equation, \(A_{f}\), \(A_{v}\), and \(A_{V_{0}D}\) correspond to the availability of the frontend, volume, and service, respectively.
|
\[A_{\text{Service}} = A_{f} \times A_{v} \times A_{V_{0}D}\] |
(2) |
CASE STUDIES
Three case studies were designed to analyze the system availability (Melo et al., 2017). The case studies can be summarized as follows:
i) Case Study I: availability analysis of architecture. ii) Case Study II: sensitivity analysis of all the system components (see Figure 2) to establish a ranking of the most important parameters in the video streaming service. iii) Case Study III: analysis the behavior of the system using other technique of sensitivity analysis, by means of a percentage difference.
Case Study I
Figure 1 depicts the architecture, which has the dedicated frontend machine and two machines for the nodes. The Table 1 shows the values of mean time to failure (MTTF) and mean time to repair (MTTR) used the model of the Frontend. Those values were obtained from (Kim et al., 2009; Dantas et al., 2012) and were used to compute the dependability metrics for the frontend, and subsequently for the whole system. This subsystem has a MTTF of 180.72 h and a MTTR of 0.96999 h.
Table 1. Parameters for RBD of the system
Frontend |
180.72 h |
0.96999h |
Volume |
100000 h |
1 h |
Service |
217.77 h |
0.92633 |
|
The values used of MTTF and MTTR for Volume subsystem are described in the Table 1. Those values were obtained from Melo et al. (2017) and Kim et al. (2009). Table 1 presents the parameters for the blocks of the RBD model for the structure shown in the Figure 2. The values of MTTF and MTTR of the Frontend and volume modules are based on Dantas et al. (2012) and Melo et al. (2017). The availability of the Service module is computed from the CTMC depicted in Figure 3. Table 2 presents all values of the parameters used for computing the availability of the Service module. The values are based on the analyses shown in Dantas et al. (2012) and Melo et al. (2017).
Table 2. Redundant system CTMC parameters
\(\lambda_{N}\) |
Mean Time to Node Failure |
1/481.83 |
\(\lambda_{\text{WN}}\) |
Mean Time to Standby Node Failure |
1/578.196 |
\(\mu_{N}\) |
Mean Time to Node Repair |
1/0.91 |
\(\mu_{\text{WN}}\) |
Mean Time to Standby Node Repair |
1/0.0333 |
\(\lambda_{V_{0}D}\) |
Mean Time to Service Failure |
1/217.779 |
\(\mu_{V_{0}D}\) |
Mean Time to Service Instantiate |
1/0.0275 |
|
For the presented configuration of parameters, we find a value of 0.994401 for the availability of the video streaming system, without redundancy. This availability corresponds to about 49,05 hours of downtime in a year, and therefore highlights the importance of searching for effective solutions to improve this system.
Case Study II
For \(2^{k}\) designs with five or more variables consider using the fractional factorial designs to reduce the number of runs required by the experiment (Mathews, 2005). This way, we performed the analysis of a fractional factorial design of experiments to provide point of view on the sensitivity of Video Stream Service availability with respect to each parameter. This analysis was performed on the 10 parameters shown in the ranking based on partial derivatives. Two levels were considered for each parameter: the minimum and maximum values used in the graphical analysis. This \(2^{k}\) factorial experiment was evaluated according to the individual effects for the system availability, and these values are shown in Table 4. The measures of interest were calculated with the values given in Table 1 and Table 3, and the sensitivity ranking of availability for all parameters of the streaming service were given in Table 4. The tools Minitab was used for calculated sensitivity analysis (Minitab, 2000).
Table 3. Input parameters for the DoE
MTTF |
\[\lambda_{f}\] |
180.72 |
4380 |
|
\[\lambda_{V_{0}D}\] |
217.779 |
4380 |
|
\[\lambda_{V_{0}l}\] |
10000 |
110000 |
|
\[\lambda_{N}\] |
481.83 |
4380 |
|
\[\lambda_{\text{WN}}\] |
578.196 |
4380 |
MTTR |
\[\mu_{V_{0}l}\] |
1 |
2 |
|
\[\mu_{f}\] |
0.96999 |
2 |
|
\[\mu_{V_{0}D}\] |
0.0275 |
0.083333 |
|
\[\mu_{\text{WN}}\] |
0.0333 |
0.083333 |
|
\[\mu_{N}\] |
0.091 |
2 |
|
The results are ordered according to absolute values. Negative values indicate that there is an inverse relationship between parameters and system availability. For example, the sensitivity with respect to failure rates is negative due to the fact that as failure rate increases (that is, the MTTF decreases) availability decreases. The index in Table 4 indicates that the parameters \(\lambda_{f}\), \(\lambda_{\text{WN}}\) and \(\mu_{f}\) have the largest effect values. Based on the results it can be concluded that the Fontend is the critical point of the video streaming system in terms of availability, and should therefore receive priority when improvements to the system are considered. Table 4 also demonstrates that the rate \(\lambda_{N}\), which is the rate of repair the node, is shown to have the second least impact on system availability, with only \(\mu_{N}\) having less. The ranking obtained from design of experiments sensitivity analysis provides a direct view of the order of importance of all parameters.
Table 4. Sensitivity ranking from \(2^{k}\) experiment analysis
\[\lambda_{f}\] |
0.009044 |
\[\lambda_{\text{WN}}\] |
0.003580 |
\[\mu_{f}\] |
−0.002920 |
\[\mu_{\text{WN}}\] |
−0.001342 |
\[\mu_{V_{0}l}\] |
−0.000905 |
\[\lambda_{V_{0}l}\] |
0.000896 |
\[\mu_{V_{0}D}\] |
0.000895 |
\[\lambda_{V_{0}D}\] |
−0.000447 |
\[\lambda_{N}\] |
−0.000447 |
\[\mu_{N}\] |
−0.000001 |
|
Figure 4 is a graphic representation of the system availability where the MTTF parameters for Frontend were altered one at a time, whilst fixing all other parameters to the values given in Table 1 and Table 2. The plot confirms that increasing the failure times of the Frontend modules results in increased availability. The need to implement improvements in the Frontend modules is confirmed by the ranking given in Table 4. An alternative method for reaching this conclusion would be to implement failover mechanisms in one or both of the components and evaluate on impact on system availability.
Figure 5 illustrates the system availability as a function of the MTTF for the Frontend module, and clearly demonstrates the impact that MTTF increases have on system availability. Increasing the MTTF of the Frontend results in a reduction in downtime of 44.80 hours in the year.
Case Study III
In this case study analysis the behavior of the system using another technique of sensitivity analysis. In this case we used the percentage difference with this method we can also identify components that affect the availability of the systems. The percentage difference is calculated using a Equation
3. This equation shows the expression for this approach, where \(\max\left\{ Y(\theta_{i}) \right\}\) and \(\min\left\{ Y(\theta_{i}) \right\}\) are the maximum and minimum output values, respectively, computed when varying the parameter \(\theta\) over the range of its \(n\) possible values of interest. If \(Y(\theta_{i})\) is known to vary monotonically, so only the extreme values of \(\theta\) (i.e., \(\theta_{i}\) and \(\theta_{n}\)) may be used to compute \(\max\left\{ Y(\theta_{i}) \right\}\) and \(\min\left\{ Y(\theta_{i}) \right\}\) subsequently \(S_{\theta_{i}}\left\{ Y \right\}\) citejain.
|
\[S_{\theta_{i}}\left\{ Y \right\} = \frac{\max\left\{ Y\left( \theta_{i} \right) \right\} - \min\left\{ Y\left( \theta_{i} \right) \right\}}{\max\left\{ Y\left( \theta_{i} \right) \right\}}\] |
(3) |
Table 5 presents the ranking of the sensitivity analysis in descending order. In order to generate the results of this ranking we used the input data from Table 3 and applied in Equation (3). Table 5 gives the percentage difference of all system components calculated from the input parameter values given in Table 3.
Table 5. Sensitivity ranking from percentage difference
\[\mu_{f}\] |
0.00563709501 |
\[\lambda_{f}\] |
−0.00514476562 |
\[\mu_{V_{0}D}\] |
0.00037202503 |
\[\mu_{V_{0}l}\] |
0.00018996201 |
\[\lambda_{V_{0}D}\] |
−0.00011998807 |
\[\lambda_{N}\] |
−0.00009585142 |
\[\mu_{\text{WN}}\] |
0.00003967628 |
\[\lambda_{N}\] |
0.00002953291 |
\[\lambda_{V_{0}l}\] |
−0.00000090908 |
\[\mu_{\text{WN}}\] |
−0.00000001800 |
|
The Table 5 indicates that the repair and rate Frontend (\(\mu_{f}\)) and (\(\lambda_{f}\)) module are the most important component with respect to availability. A refined analysis combining the two rankings - \(2^{k}\) DoE and percentage difference indices may provide a reduced list of parameters which deserve the highest priority to improve the system availability. We perform such a combined analysis by checking the parameters which appear among the first three positions of the rankings. The parameters which match such a criterion are: \(\lambda_{f}\), \(\mu_{f}\), \(\mu_{V_{0}D}\) and \(\mu_{V_{0}l}\).
CONCLUSION
This paper presented sensitivity analysis of video streaming service availability based on hierarchical analytical models. Design of Experiments (DoE) and percentage difference were used to assess the impact of each input parameter. The results show that the system availability may be improved effectively by focusing on a reduced set of factors which produce large variation on steady-state availability. Most parameters ranked in the highest positions of sensitivity rankings is related to the frontend component. This component should receive the highest priority to achieve effective improvements on system availability. We also performed a combined analysis from the two techniques which is useful to reduce the list of parameters which an analyst should focus on, and solve possible conflicting results from distinct approaches. For future work, the authors propose to implement further sensitivity techniques, and also to extend the scope of the models considered in different topologies.