Fiveable

📊Causal Inference Unit 3 Review

QR code for Causal Inference practice questions

3.1 Completely randomized designs

📊Causal Inference
Unit 3 Review

3.1 Completely randomized designs

Written by the Fiveable Content Team • Last updated September 2025
Written by the Fiveable Content Team • Last updated September 2025
📊Causal Inference
Unit & Topic Study Guides

Completely randomized designs are a cornerstone of experimental research. They involve randomly assigning units to treatment groups, ensuring each unit has an equal chance of receiving any treatment. This randomization helps balance observed and unobserved covariates across groups, allowing for unbiased estimation of treatment effects.

These designs offer several benefits, including elimination of confounding and balanced covariates on average. By randomly allocating units, researchers can attribute observed differences in outcomes to the treatment itself. This approach strengthens internal validity and enables causal inference, making it a powerful tool in various fields, from clinical trials to social science experiments.

Definition of completely randomized designs

  • Completely randomized designs are experimental designs where units are randomly assigned to treatment and control groups
  • Involves a single randomization step to allocate units to different treatment conditions
  • Ensures that each unit has an equal probability of being assigned to any of the treatment groups
  • Provides a basis for making causal inferences about the effect of the treatment on the outcome of interest

Benefits of randomization

  • Randomization is a key feature of completely randomized designs that helps to ensure the internal validity of the study
  • Allows for the unbiased estimation of treatment effects by balancing both observed and unobserved covariates across treatment groups on average

Elimination of confounding

  • Randomization helps to eliminate confounding by ensuring that treatment assignment is independent of potential outcomes and covariates
  • Balances the distribution of potential confounders across treatment groups, reducing the risk of bias in treatment effect estimates
  • Enables researchers to attribute observed differences in outcomes between treatment groups to the causal effect of the treatment itself

Balance of covariates on average

  • Randomization ensures that, on average, the distribution of covariates is similar across treatment groups
  • Balancing of covariates helps to isolate the effect of the treatment from the influence of other factors
  • Increases the comparability of treatment groups and strengthens the internal validity of the study
  • Allows for the estimation of treatment effects without the need for statistical adjustment for covariates

Mechanics of randomization

  • The process of randomization involves assigning units to treatment groups using a random mechanism
  • Ensures that each unit has an equal chance of being assigned to any of the treatment conditions
  • Can be performed using various methods, such as simple randomization or stratified randomization

Randomization methods

  • Simple randomization: Each unit is independently assigned to a treatment group with equal probability
  • Stratified randomization: Units are first stratified based on important covariates, and then randomization is performed within each stratum
  • Matched-pair randomization: Units are paired based on similarity in covariates, and then one unit from each pair is randomly assigned to each treatment group
  • Blocked randomization: Units are divided into blocks of a fixed size, and randomization is performed within each block to ensure balance

Stratified vs unstratified randomization

  • Unstratified randomization (simple randomization) assigns units to treatment groups without considering any covariates
  • Stratified randomization involves first dividing units into strata based on important covariates and then performing randomization within each stratum
  • Stratified randomization can improve the balance of covariates and increase the precision of treatment effect estimates, especially when the sample size is small
  • Stratification is particularly useful when there are known prognostic factors that could influence the outcome of interest

Estimating causal effects

  • Completely randomized designs allow for the unbiased estimation of causal effects by comparing outcomes between treatment groups
  • The average treatment effect (ATE) is a common measure of the causal effect in completely randomized designs
  • The difference in means estimator is a simple and unbiased estimator of the ATE

Average treatment effect (ATE)

  • The ATE is the expected difference in outcomes between the treatment and control groups
  • Represents the average causal effect of the treatment on the outcome of interest across the entire population
  • Can be estimated by comparing the mean outcomes of the treatment and control groups in a completely randomized design
  • Provides a summary measure of the overall effectiveness of the treatment

Difference in means estimator

  • The difference in means estimator is a simple and unbiased estimator of the ATE in completely randomized designs
  • Calculated by taking the difference between the sample means of the outcomes in the treatment and control groups
  • Assumes that the treatment assignment is independent of potential outcomes and that the sample is representative of the population
  • Provides a straightforward way to estimate the causal effect of the treatment on the outcome of interest

Hypothesis testing and p-values

  • Hypothesis testing is used to assess the statistical significance of the estimated treatment effect
  • Involves specifying a null hypothesis (usually no treatment effect) and an alternative hypothesis (presence of a treatment effect)
  • P-values are used to quantify the strength of evidence against the null hypothesis
  • A small p-value (typically < 0.05) indicates strong evidence against the null hypothesis and supports the presence of a treatment effect
  • Hypothesis testing and p-values help to determine whether the observed difference in outcomes between treatment groups is likely due to chance or a true treatment effect

Statistical inference

  • Statistical inference is the process of drawing conclusions about population parameters based on sample data
  • In completely randomized designs, statistical inference is used to make statements about the true average treatment effect in the population
  • Involves estimating the sampling distribution of the estimator, calculating standard errors, and constructing confidence intervals

Sampling distributions

  • The sampling distribution is the probability distribution of an estimator (e.g., difference in means) over repeated random samples from the same population
  • Describes the variability and central tendency of the estimator under the null hypothesis of no treatment effect
  • The sampling distribution is used to calculate standard errors and construct confidence intervals for the true average treatment effect
  • In completely randomized designs, the sampling distribution of the difference in means estimator is approximately normal for large sample sizes (due to the Central Limit Theorem)

Standard errors and confidence intervals

  • The standard error is a measure of the variability of an estimator (e.g., difference in means) across different samples
  • Calculated as the standard deviation of the sampling distribution of the estimator
  • Used to construct confidence intervals for the true average treatment effect
  • A 95% confidence interval is a range of values that has a 95% probability of containing the true average treatment effect
  • Confidence intervals provide a measure of the precision and uncertainty associated with the estimated treatment effect

Randomization inference

  • Randomization inference is a non-parametric method for testing the null hypothesis of no treatment effect in completely randomized designs
  • Involves comparing the observed test statistic (e.g., difference in means) to the distribution of test statistics under all possible random assignments of units to treatment groups
  • Provides exact p-values and confidence intervals without relying on distributional assumptions
  • Particularly useful when the sample size is small or the assumptions of parametric tests are violated
  • Preserves the validity of the inference by taking into account the actual randomization process used in the study

Efficiency of completely randomized designs

  • The efficiency of a completely randomized design refers to its ability to precisely estimate the average treatment effect
  • Efficiency is influenced by various factors, such as sample size, variability of the outcome, and the presence of covariates
  • Completely randomized designs can be compared to other experimental designs in terms of their efficiency

Comparison to other designs

  • Completely randomized designs are often less efficient than designs that make use of covariates, such as stratified or blocked designs
  • Stratified designs can improve efficiency by reducing the variability of the treatment effect estimator within each stratum
  • Blocked designs can increase efficiency by removing the variability associated with the blocking factors from the error term
  • However, completely randomized designs are simpler to implement and require fewer assumptions about the relationship between covariates and the outcome

Factors affecting efficiency

  • Sample size: Larger sample sizes generally lead to more precise estimates of the treatment effect and increased efficiency
  • Variability of the outcome: Outcomes with high variability require larger sample sizes to achieve the same level of precision as outcomes with low variability
  • Presence of covariates: Adjusting for prognostic covariates can increase the precision of the treatment effect estimator and improve efficiency
  • Allocation ratio: The allocation ratio between treatment and control groups can affect efficiency, with equal allocation typically being the most efficient
  • Heterogeneity of treatment effects: If the treatment effect varies across subgroups, the overall efficiency of the design may be reduced

Limitations of completely randomized designs

  • While completely randomized designs have several advantages, they also have some limitations that researchers should be aware of
  • These limitations can affect the internal and external validity of the study and should be considered when interpreting the results

Lack of covariate balance in small samples

  • In small samples, completely randomized designs may not achieve adequate balance of covariates between treatment groups by chance alone
  • Imbalance of important prognostic factors can lead to biased estimates of the treatment effect and reduced efficiency
  • Stratified or blocked randomization can be used to improve covariate balance in small samples
  • Alternatively, statistical adjustment methods (e.g., regression) can be used to account for covariate imbalances in the analysis

Ethical considerations

  • In some cases, completely randomized designs may not be ethically appropriate, particularly when there is a strong prior belief that one treatment is superior to another
  • Withholding a potentially beneficial treatment from some participants through randomization may be considered unethical
  • Alternative designs, such as adaptive or preference-based designs, may be more appropriate in these situations
  • Researchers should carefully consider the ethical implications of their study design and ensure that the benefits outweigh the risks for all participants

Examples of completely randomized designs

  • Completely randomized designs are widely used in various fields, including medicine, psychology, and social sciences
  • Some common examples include clinical trials and A/B testing in web design

Clinical trials

  • Clinical trials often use completely randomized designs to evaluate the efficacy and safety of new medical interventions (drugs, medical devices)
  • Patients are randomly assigned to receive either the new treatment or a control (placebo or standard treatment)
  • Randomization helps to ensure that treatment groups are comparable and that observed differences in outcomes can be attributed to the treatment effect
  • Example: A randomized controlled trial comparing the effectiveness of a new blood pressure medication to a placebo

A/B testing in web design

  • A/B testing is a form of completely randomized design used to compare two versions of a web page or application
  • Users are randomly assigned to either the "A" version (control) or the "B" version (treatment) of the web page
  • Metrics such as click-through rates, conversion rates, or user engagement are compared between the two versions
  • Randomization ensures that any differences in user behavior can be attributed to the design changes made in the "B" version
  • Example: An e-commerce website randomly assigns visitors to either a standard product page or a redesigned page with new features to assess the impact on sales

Variations on completely randomized designs

  • While the basic completely randomized design is widely used, there are several variations that can be employed to address specific research questions or design considerations
  • These variations include blocked randomized designs and factorial designs

Blocked randomized designs

  • In blocked randomized designs, units are first divided into blocks based on one or more covariates and then randomized within each block
  • Blocking helps to ensure balance of the covariates across treatment groups and can increase the precision of the treatment effect estimator
  • Blocks are typically chosen based on factors that are known or suspected to influence the outcome of interest
  • Example: In an agricultural experiment, fields may be blocked by soil type before randomly assigning different fertilizer treatments within each block

Factorial designs

  • Factorial designs allow researchers to investigate the effects of two or more factors simultaneously
  • Each factor has two or more levels, and treatments are formed by combining the levels of the factors
  • Completely randomized factorial designs involve randomly assigning units to each combination of factor levels
  • Factorial designs enable the estimation of main effects (the effect of each factor averaging over the levels of the other factors) and interaction effects (the extent to which the effect of one factor depends on the level of another factor)
  • Example: A 2x2 factorial design comparing the effects of two different drugs (drug A and drug B) and their combination (drug A + drug B) on patient outcomes

Analyzing completely randomized designs

  • The analysis of completely randomized designs typically involves comparing the outcomes between treatment groups
  • Several statistical methods can be used to estimate treatment effects and assess their significance, including regression analysis and analysis of variance (ANOVA)

Regression analysis

  • Regression analysis is a flexible method for estimating treatment effects in completely randomized designs
  • Involves fitting a regression model with the outcome as the dependent variable and treatment indicators as independent variables
  • Can easily accommodate covariates and assess their influence on the treatment effect
  • Provides estimates of the average treatment effect, standard errors, and confidence intervals
  • Example: Using linear regression to estimate the effect of a new teaching method on student test scores, controlling for baseline performance and demographic factors

Analysis of variance (ANOVA)

  • ANOVA is a common method for analyzing completely randomized designs with a single factor (treatment)
  • Partitions the total variability in the outcome into two components: variability between treatment groups and variability within treatment groups
  • Tests the null hypothesis of no difference in means between treatment groups using an F-test
  • Provides estimates of the average treatment effect and can be extended to include blocking factors (e.g., blocked ANOVA)
  • Example: Using one-way ANOVA to compare the mean weight loss achieved by participants in three different diet groups (low-carb, low-fat, and control) in a randomized trial