A Framework of Human Systems Engineering. Группа авторов

A Framework of Human Systems Engineering - Группа авторов


Скачать книгу
them are described as rx,y where x and y are stakeholders X and Y.

Schematic illustration of an example project social network.

      The relationship between the stakeholders rx,y can be represented as rx,y(d, s, a) where d, s, and a are defined as follows:

       d represents directionality: In Figure 3.5, r1,3 is shown as stakeholder 1 influencing stakeholder 3, but not vice versa. Compare this to r3,4 where this influence is bidirectional. In long‐lived capability development environments, a given relationships rx,y may not persist over time, or new relationships may emerge. Failure to recognize the network structure between relationships can result in unintended consequences that likely inject adverse effects into the development cycle or operations of the ensuing system.

       S indicates strength of influence: Strength of influence is defined as the degree to which the change in one node affects another. The strength of influence can be positive, negative, or neutral. Upon a change in a stakeholder, positive influence will increase the value of the temporal sociotechnical measures, negative influence will decrease them, and neutral influence indicates that a change in one node will have no effect on the other. For example, in Figure 3.5, if N1 was the leader of the organization, it is reasonable to assume that the r1,3(s) would be strongly positive. The diagram above also indicates that the leader is not significantly influenced by stakeholder N3 as arc is unidirectional; in other words, the leader is not listening. Similarly, the discussion with respect to relationships and the strength of influence of a relationship may change over time.

       a is the alignment between stakeholders: Alignment (a) is defined as the difference between the beliefs in both project execution and the underlying project ecosystem between stakeholders. Large differences in beliefs may portend risk as tactical measures may be taken that are not congruent with the success metrics of the parties and larger strategic measures of success may be out of alignment. The alignment is explicitly assessed using the temporal sociotechnical measures, such as the previously discussed belief approach. Whereas the relationship and strength of influence form the underlying substrata for sociotechnical network, risk is directly assessed by alignment (or lack thereof) of the belief structures of the stakeholders.

      Accurate AI modeling of stakeholders using digital twin concepts provides a solid representation of the stakeholders and mechanism to track the evolution of the preferences. These AI‐based risk assessors can look at atomic measures, group measures, and holistic sociotemporal measures to assess risks as shown in Figure 3.5. Structural risk can be assessed using appropriate interpretations of the sociotechnical network, consistently looking for over‐connectiveness as well as sparsity. Structural risks are identified when measures exceed the tolerance of network metrics within a degree of error.

      The canonical definition of risk can be described as a tuple represented as risk{event, likelihood, consequence}. Using this definition, AI‐based models can spot misalignment across the sociotechnical measures and have the added capability of identifying localized risk, enterprise level risks, and emergent risks as misalignment grows over time. Further, with the introduction of evidence, and stakeholders’ interpretation, may result in the identification of hidden risks. For example, a significant reduction in electronic communication between certain stakeholders can be a signal if impending misalignment and the advent of a relational risk. Relational risk occurs when a risk assessor looks at the relationship between two stakeholders and identifies significant incongruence between the sociotemporal measures leading to a lack of alignment. Here again the application of AI can model incidental and emerging risks as the scale of a large enterprise and the rapid pace of change make typical knowledge acquisition efforts across the stakeholders unrealistic. AI‐based models can assist in proactive decision support when coupled with SE to identify measures, establish metrics, and define acceptable deviation limits for the measure.

      A large enterprise may have tens or even hundreds of relevant stakeholders, all of whom are seeing different aspects of the development effort. As evidence is introduced to stakeholders in these complex sociotechnical networks, it is improbable that manual, or mental, methods can adequately track and assess the impact of the evidence. AI‐based models can easily track and learn from these events and provide the practitioner with insight on how best to lower risk and improve value delivery.

      Traditional risks in typical system development projects and their evolution are well documented in literature (Warkentin et al., 2009), but analytical analysis of risk particularly in the sociotemporal space is largely a manual process. An AI‐based risk assessor can monitor trends in the sociotechnical measures to examine if there are emerging risks from a growing misalignment between groups of stakeholders. An AI risk assessor can develop forecasts of emerging risks that can be modeled by deliberate introduction of possible evidence that may occur in the future. Modeling in this fashion provides the capability to proactively assess system risk areas and take preventive steps.

      In addition to localized risk identification, a holistic view of systemic risk across the sociotechnical ecosystem becomes more feasible using AI‐based models. While localized gross individual misalignment may not manifest itself, a general trend toward misalignment across the enterprise can be discovered and identified. Whereas the values for sociotemporal measures typically have been discovered through interviews, AI offers the opportunity to make assessments based upon noninvasive approaches such as analyzing e‐mails and text messages. Applications like the hedonometer (Dodds et al., 2011) have been used for many years; it is suggested here that AI can tap into the results of the hedonometer for both localized and global assessments of risk.

      With an estimate of the likelihood of the risk event, the last step is to assess the consequence(s) if the risk event happens and becomes an issue as mentioned above. This is again an opportunity for the AI to employ uncertain reasoning: taking a risk event and the likelihood that it will happen, what are the anticipated consequences? The assessment can be based upon historical data or generalized rules that categorize the risk and its associated consequences and iterate multiple outcomes based on actions taken to determine enterprise‐level success vice local optimization. Generally, the consequences will manifest themselves as a negative effect along the traditional lines for system development, impact on cost, an extension of the schedule, or a decrease in the capability or value delivered with the system increment. This mapping to the traditional measures of system progress facilitates the movement toward the use of AI to risk amelioration.

      A conceptually straightforward approach for amelioration is to use AI to calculate the expected value of the impact of a given risk if it comes to fruition and then compare the costs of various amelioration


Скачать книгу