Using Predictive Analytics to Improve Healthcare Outcomes. Группа авторов
Being Measured
If there are important variables not being measured, it will be necessary to develop ways to measure them. If influential variables are left out of the study, the model will remain mis‐specified (wrong).
If influential variables are left out of the study, the model will remain mis‐specified (wrong).
Step 8: Select an Analytic Method
This work will likely be overseen by an in‐house or consultant mathematician, statistician, or data analyst. Consider types of analytics beyond linear methods or qualitative descriptive methods, which are the most common methods currently used in healthcare. For example, if the dataset is very large and complex and it is not clear how to sort the predictor variables in a linear method such as regression analysis, try Pareto mathematics where outliers are examined. Pareto mathematics looks at the highs and lows in the dataset to create a profile of success factors. In his book Where Medicine Went Wrong: Rediscovering the Path to Complexity, Dr. Bruce West asserts that Gaussian mathematics throws out the extreme values despite the fact that these extreme values often provide the most valuable information because they provide a profile of the biggest successes and the biggest failures (2007).
Another type of analysis to consider is constructal theory, a method derived from physics, which is the study of “flow” (Bejan & Zane, 2012). If employees are able to talk about what makes their work flow or what makes their work pause, constructal theory will allow for their comments to be themed and addressed operationally. It is the experience of this author that if employees can talk about their workflow, and data can then be arranged for them in themes, the employees get excited about working on productivity because it is readily apparent to them that the overall aim is figuring out how to do more of what works well and less of what does not. Constructal theory makes productivity, or the lack of it, visible. Gaussian mathematics provides insight into linear processes, while analyses like Parato mathematics and constructal theory provide insight into more dynamic/complex and unknown processes, respectively. Pairing the analytic method with the variable of interest is important to achieving insight into the operations of work and associated outcomes.
The overall aim is figuring out how to do more of what works well and less of what does not.
For your most complex models, you may want to consider using a machine learning problem designed to let the computer tell you the rank order of predictor variables as they relate to your variable of interest. This is suggested on the condition that you never allow the machine to have the “final say.” Machines function without regard for theory and context, so they are not able to tell a story capable of deeply resonating with the people whose work is being measured. It is tempting to be lazy and not do the work of carefully building models based on theory and rich in context that will result in data that makes people excited to take action. More on machine learning can be found in Chapter 6.
Step 9: Collect Retrospective Data Right Away and Take Action
If available, use retrospective data (data used for reporting what has already happened) to determine which predictor variables, of those already being measured, most closely relate to your variable of interest. Have a mathematician, statistician, or data analyst run a correlation table of all the predictor variables. Most organizations employ or contract with mathematicians, statisticians, and/or others who know how to run and read a correlation table using statistics software such as SPSS, SAS, or R. The mathematician, statistician, or analyst will be able to identify all the predictor variables found to have a relationship with the variable of interest and rank them in order of the strength of the relationship. (Many examples of these ranked predictor variables will appear in this book.) This will help start a meaningful conversation about what is being discovered while data on the remaining variables from the measurement model is still being collected manually. If actionable information is discovered during this step, operational changes can be implemented immediately. Further examination of the data will continue, but if what you have discovered by this point makes some opportunities for improvements apparent, do not wait to improve operations!
Step 10: Examine Data Before Analysis
Once the data is collected, it is tempting to proceed directly to the fun part and view the results. How do things relate to one another? What are the high and low scores? Viewing the results that answer questions like these is rewarding; however, prior to viewing any results (even in Step 9, where you may have gathered together some pretty compelling retrospective data), it is critical to ensure that all the data is correct. Have the statistician, analyst, or mathematician use statistics software, or even Excel, to examine the distribution of the scores and look for indications that there is missing data. If the distribution of the data for any predictor variable has a prominent leaning toward low or high scores, with a few outliers, the analyst will need to decide whether the outliers should be removed or whether the data should be weighted. If there are patterns of missing data (e.g. one item from a construct has many missing scores), then the group should discuss why the data is missing and what to do about it. Understanding the distribution and missing data will provide additional insight into the population being studied.
Reviewing the data for accuracy can also begin to give the team a feel for the “personality” of the data.
Reviewing the data for accuracy can also begin to give the team a feel for the “personality” of the data. This data all comes from people, but instead of a coherent conversation or observation of the people contained in the data, it is initially just data. As the data is examined, an understanding of the respondents will begin to emerge.
Step 11: Analyze the Data
For this step, it is important to engage a professional who is trained in analytics so the data can be interpreted accurately. There are several options in today's analytic software, such as SPSS, SAS or R, to aid in the examination of data flaws that are not obvious by merely looking at the dataset or its associated graphic representations. Your organization may have software, such as Tableau, which generates graphs automatically and is dependable for graphic visualization. In this step, it is important to engage an analyst or data scientist who can take advantage of the tools and tests contained in analytic software.
Step 12: Present Data to the People Who Work Directly With the Variable of Interest, and Get Their Interpretation of the Data
Ultimately, the most relevant and useful interpretation will be provided by those who live the clinical experience, because they can validate or refine the interpretation of the data. These people are typically clinicians, not analysts, however, so showing complex data in a format that is accessible to the employee not trained in analytics is important. When possible, use bubble graphs that provide a visual review of overall findings.
Figures 1.2 and 1.3 convey two ways to facilitate understanding of complex mathematics. Figure 1.2 provides a visual representation of the traditional graphic for results from a regression equation which examined three predictor variables related to reduction of central line‐associated blood stream infection (CLABSI). The three predictors—location of the central line insertion, the RN assigned, and the phase of the project—have been tested to see the degree to which each of them explains the frequency of CLABSI.
Similarly, Figure 1.3 is a bubble graph which also shows how much of the outcome variable is explained by the same three predictor variables.