Why you shouldn't trust mediation as process evidence

April 11, 2023·
Victor van Pelt
Victor van Pelt
· 9 min read

TL/DR

  • Mediation involves testing the effect of an independent variable on a dependent variable through a mediating variable. Its low cost and easy integration have made it the standard form of process evidence for experimental studies, including those in accounting.
  • However, the inclusion of a mediating variable messes up random assignment and biases the relationship between the mediating variable and the dependent variable, no matter at what point the mediating variable is measured during the experiment.
  • As a result, mediation tends to overestimate the indirect path while underestimating the direct path, leading to invalid conclusions about causal mechanisms.
  • There are alternative experimental techniques, such as moderation-by-process designs and conducting multiple experiments, that outshine mediation in obtaining valid and reliable process evidence. This is primarily because they leverage the USP of experiments: Randomization and causal inference

What is process evidence?

If you are familiar with experimental research, particularly in the field of accounting, then you probably have heard of the term “process evidence.” It is data that helps establish the underlying causal mechanisms through which manipulated independent variables affect measured dependent variables. Process evidence is a highly sought-after (if not a necessary ingredient) of a finished experimental paper. While there are different ways to obtain process evidence in experiments, mediation is the most popular form by far because it is cheap and easy to include in experimental designs. However, there is a hefty, and typically overlooked, price tag, which invalidates it as a strong form of process evidence. In this post, I explain why mediation is the worst type of process evidence for experimental research. I also discuss two other well-known techniques that offer much stronger process evidence.

Mediation at a glance

MTurk Connect

Figure 1: Mediating relationship bbetween XX and YY.

Mediation involves testing whether a manipulated independent variable XX affects a measured dependent variable YY “through” a measured mediating variable MM (see Figure 1). There is evidence for mediation when XX influences MM (relationship aa), which, in turn, influences YY (relationship bb). Unlike XX, MM is endogenous and not manipulated by the experimenter. Instead, MM is either unobtrusively measured during the experiment or elicited interim or ex post. These two approaches to measuring mediating variables each have their merits and problems (Asay et al., 2022), but that is not the focus of this post. We did write a post about the often overlooked limitations of ex post techniques, such as post-experimental questionnaires, in this post I wrote with Farah Arshad.

The mediation trap

Besides being low-cost and immensely popular in fields like accounting, there is a trap when using mediation as process evidence. When we statistically find support for an indirect effect (aa × bb), one may be quick to reason that it provides strong support for the causal mechanism (XXMMYY). Given that the independent variable XX is randomly assigned, one might expect there is no correlated omitted variable bias. That is kind of why we run experiments in the first place. That expectation does apply to the empirical model without mediation, i.e., one that only tests for the causal effect of XX on YY. Specifically, that simpler empirical model typically leads to an unbiased estimation of the causal effect cc (see Figure 2).

MTurk Connect

Figure 2: Causal relationship bbetween XX and YY.

However, this reasoning does not apply to the empirical model that also estimates the mediating path going from XX to YY through MM (see Figure 1). Including MM in the empirical model messes up random assignment, which makes relationship bb (between MM and YY) and cc’ (between XX and YY) generally biased. Relationship bb between MM and YY is biased because, while it picks up the causal effect (if any) of XX, it also picks up the confounded effect of all things that correlate with MM outside the experiment. It does not matter when and how MM is measured. As soon as it is measured, there is a reasonable possibility that MM is subject to classic omitted variable bias. This particular bias in MM is perhaps not new to some because Asay et al. (2022) also highlight it as a weakness of using mediation as a process evidence. However, what is generally overlooked is that the bias in MM spreads to XX, leading to bias in the mediating effect (aa × bb).

The bias also spreads to relationship cc' (the direct path between XX and YY). Since we test mediation by subtracting cc’ from cc, we underestimate (overestimate) cc’ if the sign of the bias in bb is positive (negative). Thus, the sign of the bias in cc’ is the sign of the correlation between MM and YY outside the experiment. In most cases, the model in Figure 1 will overestimate mediation and underestimate cc’. In extreme cases, the positive bias in bb can even lead us to infer there is an indirect “causal” effect via MM, while in reality there is no effect of XX on YY (Rohrer, 2019). It is theoretically possible (but improbable) to underestimate mediation while having a positive bias in cc’. For instance, if, for some strange reason, the correlation between XX and MM is positive in the experiment, but the correlation between MM and XX is negative outside of it.

So what is the intuition?

This dangerous problem with mediation has been covered in the statistical literature and popular outlets outside of accounting (see, for instance, MacKinnon and Pirlott (2015), Rohrer et al. (2021), and Data Colada (2022)). The intuition behind the problem can be illustrated using a simple example (see Figure 3). Let’s say you are interested in how an intervention XX affects employees’ performance Y, and your causal mechanism states that the intervention XX will lead employees to be more productive (MM), which, in turn, leads to higher performance YY. Thanks to randomization, MM includes some productivity free of confounds caused by the intervention XX. However, it also includes productivity related to confounds, like employees being passionate about their job, enjoying the type of tasks they are doing, or caring about being a high-performer.

MTurk Connect

Figure 3: Illustrative example.

This simple example illustrates that the relationship bb will be a combination of the causal and confounded effects of productivity, which, in turn, will bias the mediation effect and the direct effect. The size of the bias will depend on the relative impact of the counfounding factors. If the causal effect of a 1% increase in productivity equals a 4% performance increase and the confounded effect is an 8% increase in performance, then we would estimate the increase in performance to be higher than the causal effect (i.e., 8% > bb < 4%). Thus, we would overpredict the increase in performance due to the intervention via the productivity path. Also, the model in Figure 3 would bias cc’ downwards by the average error arising from bb being inflated toward 8%. Unlike this example, we do not precisely know how large the confounded effect will be, making it difficult to infer how close we are to the true indirect effect and whether this will show up as statistically different from zero in our analyses.

So what? Isn’t everything biased?

Nothing is perfect. Even the magic of randomization can fail. However, the primary point is that mediation is generally biased in an otherwise relatively clean experiment. Experiments’ comparative advantage is identifying causal effects. I.e., their strength is using randomization and implementing interventions to observe how it produces variation in a dependent variable of interest. Its comparative advantage is not making surveyesque path models that infect otherwise relatively clean causal effects. It is a bit like showing up at a funeral in your party outfit. The party outfit may be very impressive, but it is just not befitting a funeral. Thus, as far as a weakness goes for a particular process evidence technique, mediation seems particularly undesirable for experiments. The problem cannot be resolved by unobtrustively measuring mediating variables. Nor does it make sense to start including lots of control variables in some post-experimental questionnaire. Namely, you would basically be changing your methodology from an experiment to a survey.

Alright, but are there alternatives?

There are other ways to obtain process evidence that outshine mediation (Bullock and Green, 2021). These techniques build much more strongly on the experiment’s ability to establish causality. The only weakness these alternative techniques have are weaknesses of the experimental method itself. The first technique is moderation-by-process designs. It involves manipulating a second independent variable in addition to a first. The purpose of the second independent variable is to change the causal mechanism central to the first in a conceptually consistent way. If the data show that the second independent variable indeed modifies the causal relationship generated by the first, we have obtained process evidence of the first.

A second technique is to conduct multiple experiments sequentially and have each build on the others. One approach would be to conduct an initial experiment establishing the causal effect of XX on Y, followed by two more that test for the causal effects of XX on MM and MM on YY. This would be the “experimental” way to lay bare some conceptual path if you want to but it does not make great use of the experiments’ primary strengths. An approach that leverages these strengths better is zooming in or out of the causal mechanism. An initial experiment could identify a rather broad relationship bbetween XX and YY. This broad relationship ccould focus on a bigger picture, while capturing multiple theoretical processes. Followup experiments could try to eliminate some of the theoretical processes at play. For instance, a followup could change one aspect of the experimental design or the independent variable XX to narrow the causal mechanism.

Final note

If we use mediation to obtain process evidence for an experiment, we expect XX to lead to MM and MM to lead to YY. If we expect that indirect path, we should also expect confounding variables to explain the correlation between MM and YY outside the experiment. This type of correlation invalidates mediation because it tends to overestimate the indirect path while underestimating the direct path. This could lead to wrongly inferring that there is evidence for an indirect path while a correlated omitted variable bias drives it. There are ways to obtain process evidence that leverage the strengths of the experimental method much more, like including moderating variables that change the causal process and conducting multiple experiments to zoom in or out of the causal process.

References

Asay, H. S., R. Guggenmos, K. Kadous, L. Koonce, and R. Libby. 2022. Theory testing and process evidence in accounting experiments. The Accounting Review, 97 (6): 341-395. https://doi.org/10.2308/TAR-2019-1001.
Bullock, J. G., D. P. Green. 2021. The failings of conventional mediation analysis and a design-based alternative. Advances in Methods and Practices in Psychological Science, 4 (4): 1-18. https://doi.org/10.1177/25152459211047227.
Data Colada 2022, September 26. Mediation analysis is counterintuitively invalid. http://datacolada.org/103.
Mackinnon, D. P. and A. G. Pirlott 2014. Statistical approaches for enhancing causal interpretation of the M to Y relation in mediation analysis. Personality and Social Psychology Review, 19 (1): 30-43. https://doi.org/10.1177/1088868314542878.
Rohrer, J. 2019, October 3. Indirect Effect Ex Machina. https://www.the100.ci/2019/10/03/indirect-effect-ex-machina/.
Rohrer, J. 2022. That's a lot to process! Pitfalls of popular path models. Advances in Methods and Practices in Psychological Science, 5 (2): 1-14. https://doi.org/10.1177/25152459221095827.
Victor van Pelt
Authors
Assistant Professor of Accounting