Field experiments, explained

Editor’s note: This is part of a series called “The Day Tomorrow Began,” which explores the history of breakthroughs at UChicago.  Learn more here.

A field experiment is a research method that uses some controlled elements of traditional lab experiments, but takes place in natural, real-world settings. This type of experiment can help scientists explore questions like: Why do people vote the way they do? Why do schools fail? Why are certain people hired less often or paid less money?

University of Chicago economists were early pioneers in the modern use of field experiments and conducted innovative research that impacts our everyday lives—from policymaking to marketing to farming and agriculture.  

Jump to a section:

What is a field experiment, why do a field experiment, what are examples of field experiments, when did field experiments become popular in modern economics, what are criticisms of field experiments.

Field experiments bridge the highly controlled lab environment and the messy real world. Social scientists have taken inspiration from traditional medical or physical science lab experiments. In a typical drug trial, for instance, participants are randomly assigned into two groups. The control group gets the placebo—a pill that has no effect. The treatment group will receive the new pill. The scientist can then compare the outcomes for each group.

A field experiment works similarly, just in the setting of real life.

It can be difficult to understand why a person chooses to buy one product over another or how effective a policy is when dozens of variables affect the choices we make each day. “That type of thinking, for centuries, caused economists to believe you can't do field experimentation in economics because the market is really messy,” said Prof. John List, a UChicago economist who has used field experiments to study everything from how people use  Uber and  Lyft to  how to close the achievement gap in Chicago-area schools . “There are a lot of things that are simultaneously moving.”

The key to cleaning up the mess is randomization —or assigning participants randomly to either the control group or the treatment group. “The beauty of randomization is that each group has the same amount of bad stuff, or noise or dirt,” List said. “That gets differenced out if you have large enough samples.”

Though lab experiments are still common in the social sciences, field experiments are now often used by psychologists, sociologists and political scientists. They’ve also become an essential tool in the economist’s toolbox.  

Some issues are too big and too complex to study in a lab or on paper—that’s where field experiments come in.

In a laboratory setting, a researcher wants to control as many variables as possible. These experiments are excellent for testing new medications or measuring brain functions, but they aren’t always great for answering complex questions about attitudes or behavior.

Labs are highly artificial with relatively small sample sizes—it’s difficult to know if results will still apply in the real world. Also, people are aware they are being observed in a lab, which can alter their behavior. This phenomenon, sometimes called the Hawthorne effect, can affect results.

Traditional economics often uses theories or existing data to analyze problems. But, when a researcher wants to study if a policy will be effective or not, field experiments are a useful way to look at how results may play out in real life.

In 2019, UChicago economist Michael Kremer (then at Harvard) was awarded the Nobel Prize alongside Abhijit Banerjee and Esther Duflo of MIT for their groundbreaking work using field experiments to help reduce poverty . In the 1990s and 2000s, Kremer conducted several randomized controlled trials in Kenyan schools testing potential interventions to improve student performance. 

In the 1990s, Kremer worked alongside an NGO to figure out if buying students new textbooks made a difference in academic performance. Half the schools got new textbooks; the other half didn’t. The results were unexpected—textbooks had no impact.

“Things we think are common sense, sometimes they turn out to be right, sometimes they turn out to be wrong,” said Kremer on an episode of  the Big Brains podcast. “And things that we thought would have minimal impact or no impact turn out to have a big impact.”

In the early 2000s, Kremer returned to Kenya to study a school-based deworming program. He and a colleague found that providing deworming pills to all students reduced absenteeism by more than 25%. After the study, the program was scaled nationwide by the Kenyan government. From there it was picked up by multiple Indian states—and then by the Indian national government.

“Experiments are a way to get at causal impact, but they’re also much more than that,” Kremer said in  his Nobel Prize lecture . “They give the researcher a richer sense of context, promote broader collaboration and address specific practical problems.”    

Among many other things, field experiments can be used to:

Study bias and discrimination

A 2004 study published by UChicago economists Marianne Bertrand and Sendhil Mullainathan (then at MIT) examined racial discrimination in the labor market. They sent over 5,000 resumes to real job ads in Chicago and Boston. The resumes were exactly the same in all ways but one—the name at the top. Half the resumes bore white-sounding names like Emily Walsh or Greg Baker. The other half sported African American names like Lakisha Washington or Jamal Jones. The study found that applications with white-sounding names were 50% more likely to receive a callback.

Examine voting behavior

Political scientist Harold Gosnell , PhD 1922, pioneered the use of field experiments to examine voting behavior while at UChicago in the 1920s and ‘30s. In his study “Getting out the vote,” Gosnell sorted 6,000 Chicagoans across 12 districts into groups. One group received voter registration info for the 1924 presidential election and the control group did not. Voter registration jumped substantially among those who received the informational notices. Not only did the study prove that get-out-the-vote mailings could have a substantial effect on voter turnout, but also that field experiments were an effective tool in political science.

Test ways to reduce crime and shape public policy

Researchers at UChicago’s  Crime Lab use field experiments to gather data on crime as well as policies and programs meant to reduce it. For example, Crime Lab director and economist Jens Ludwig co-authored a  2015 study on the effectiveness of the school mentoring program  Becoming a Man . Developed by the non-profit Youth Guidance, Becoming a Man focuses on guiding male students between 7th and 12th grade to help boost school engagement and reduce arrests. In two field experiments, the Crime Lab found that while students participated in the program, total arrests were reduced by 28–35%, violent-crime arrests went down by 45–50% and graduation rates increased by 12–19%.

The earliest field experiments took place—literally—in fields. Starting in the 1800s, European farmers began experimenting with fertilizers to see how they affected crop yields. In the 1920s, two statisticians, Jerzy Neyman and Ronald Fisher, were tasked with assisting with these agricultural experiments. They are credited with identifying randomization as a key element of the method—making sure each plot had the same chance of being treated as the next.

The earliest large-scale field experiments in the U.S. took place in the late 1960s to help evaluate various government programs. Typically, these experiments were used to test minor changes to things like electricity pricing or unemployment programs.

Though field experiments were used in some capacity throughout the 20th century, this method didn’t truly gain popularity in economics until the 2000s. Kremer and List were early pioneers and first began experimenting with the method in the 1990s.

In 2004, List co-authored  a seminal paper defining field experiments and arguing for the importance of the method. In 2008,  he and UChicago economist Steven Levitt published another study tracing the history of field experiments and their impact on economics.

In the past few decades, the use of field experiments has exploded. Today, economists often work alongside NGOs or nonprofit organizations to study the efficacy of programs or policies. They also partner with companies to test products and understand how people use services.  

There are several  ethical discussions happening among scholars as field experiments grow in popularity. Chief among them is the issue of informed consent. All studies that involve human test subjects must be approved by an institutional review board (IRB) to ensure that people are protected.

However, participants in field experiments often don’t know they are in an experiment. While an experiment may be given the stamp of approval in the research community, some argue that taking away peoples’ ability to opt out is inherently unethical. Others advocate for stricter review processes as field experiments continue to evolve.

According to List, another major issue in field experiments is the issue of scale . Many experiments only test small groups—say, dozens to hundreds of people. This may mean the results are not applicable to broader situations. For example, if a scientist runs an experiment at one school and finds their method works there, does that mean it will also work for an entire city? Or an entire country?

List believes that in addition to testing option A and option B, researchers need a third option that accounts for the limitations that come with a larger scale. “Option C is what I call critical scale features. I want you to bring in all of the warts, all of the constraints, whether they're regulatory constraints, or constraints by law,” List said. “Option C is like your reality test, or what I call policy-based evidence.”

This problem isn’t unique to field experiments, but List believes tackling the issue of scale is the next major frontier for a new generation of economists.

Hero photo copyright Shutterstock.com

More Explainers

Illustration of galaxies spreading out from a single luminous point

Dark energy, explained

A chair on stage

Improv, Explained

Get more with UChicago News delivered to your inbox.

Recommended Stories

A hand holding a paper heart, inserting it into a coin slot

An economist illuminates our giving habits—during the pandemic and…

Michael Kremer meeting with officials in Kenya including Dr. Sara Ruto

Collaborating with Kenyan government on development innovations is…

Related Topics

Latest news, uchicago community welcomes students to their new intellectual home.

Patrick Jagoda

Patrick Jagoda to deliver Aims of Education address Sept. 26

Photo of people in a room with monitor showing radar and climate data

UChicago offers new master’s program in environmental science

Silhouettes of young people

Big Brains podcast

Big Brains podcast: What are we getting wrong about young voters?

Inside the Lab

Go 'Inside the Lab' at UChicago

Explore labs through videos and Q&As with UChicago faculty, staff and students

Bats hanging upside down

Collapse of bat populations increased infant mortality rate, study finds

Rows of small white tiles with speech bubble sit on a yellow background. The right upper corner tile is orange with a speech bubble that contains an "X".

AI Insights

AI is biased against speakers of African American English, study finds

Around uchicago.

Ralph A Austen

In Memoriam

Ralph A. Austen, historian of Africa and a ‘scholar’s scholar,’ 1937-2024

Artificial Intelligence

NSF awards $20 million to build AI models that predict scientific discoveries a…

New chicago booth course empowers student entrepreneurs to tackle global issues.

Campus News

Project to improve accessibility, sustainability of Main Quadrangles

Photo of 3 scientists on a stage with The Kavli Prize backdrop behind

UChicago President Paul Alivisatos accepts 2024 Kavli Prize in Nanoscience

Barack Obama addressing a group of young people sitting at tables

Obama Foundation

University of Chicago Obama Foundation Scholars Program includes 18 emerging leaders for 2024-25

Meet A UChicagoan

“The trouble comes from using these immortal materials for disposable products.”

Photo of A scientist with black gloves and blue lab coat and a mustache holding a small item

South Side Science Festival

Third annual South Side Science Festival set to bring a day of discovery for all ages on Oct. 5

Introduction to Field Experiments and Randomized Controlled Trials

Painting of a girl holding a bottle

Have you ever been curious about the methods researchers employ to determine causal relationships among various factors, ultimately leading to significant breakthroughs and progress in numerous fields? In this article, we offer an overview of field experimentation and its importance in discerning cause and effect relationships. We outline how randomized experiments represent an unbiased method for determining what works. Furthermore, we discuss key aspects of experiments, such as intervention, excludability, and non-interference. To illustrate these concepts, we present a hypothetical example of a randomized controlled trial evaluating the efficacy of an experimental drug called Covi-Mapp.

Why experiments?

Every day, we find ourselves faced with questions of cause and effect. Understanding the driving forces behind outcomes is crucial, ranging from personal decisions like parenting strategies to organizational challenges such as effective advertising. This blog aims to provide a systematic introduction to experimentation, igniting enthusiasm for primary research and highlighting the myriad of experimental applications and opportunities available.

The challenge for those who seek to answer causal questions convincingly is to develop a research methodology that doesn't require identifying or measuring all potential confounders. Since no planned design can eliminate every possible systematic difference between treatment and control groups, random assignment emerges as a powerful tool for minimizing bias. In the contentious world of causal claims, randomized experiments represent an unbiased method for determining what works. Random assignment means participants are assigned to different groups or conditions in a study purely by chance. Basically, each participant has an equal chance to be assigned to a control group or a treatment group. 

Field experiments, or randomized studies conducted in real-world settings, can take many forms. While experiments on college campuses are often considered lab studies, certain experiments on campus – such as those examining club participation – may be regarded as field experiments, depending on the experimental design. Ultimately, whether a study is considered a field experiment hinges on the definition of "the field."

Researchers may employ two main scenarios for randomization. The first involves gathering study participants and randomizing them at the time of the experiment. The second capitalizes on naturally occurring randomizations, such as the Vietnam draft lottery. 

Intervention, Excludability, and Non-Interference

Three essential features of any experiment are intervention, excludability, and non-interference. In a general sense, the intervention refers to the treatment or action being tested in an experiment. The excludability principle is satisfied when the only difference between the experimental and control groups is the presence or absence of the intervention. The non-interference principle holds when the outcome of one participant in the study does not influence the outcomes of other participants. Together, these principles ensure that the experiment is designed to provide unbiased and reliable results, isolating the causal effect of the intervention under study.

Omitted Variables and Non-Compliance

To ensure unbiased results, researchers must randomize as much as possible to minimize omitted variable bias. Omitted variables are factors that influence the outcome but are not measured or are difficult to measure. These unmeasured attributes, sometimes called confounding variables or unobserved heterogeneity, must be accounted for to guarantee accurate findings.

Non-compliance can also complicate experiments. One-sided non-compliance occurs when individuals assigned to a treatment group don't receive the treatment (failure to treat), while two-sided non-compliance occurs when some subjects assigned to the treatment group go untreated or individuals assigned to the control group receive the treatment. Addressing these issues at the design level by implementing a blind or double-blind study can help mitigate potential biases.

Achieving Precision through Covariate Balance

To ensure the control and treatment groups are comparatively similar in all relevant aspects, particularly when the sample size (n) is small, it is essential to achieve covariate balance. Covariance measures the association between two variables, while a covariate is a factor that influences the outcome variable. By balancing covariates, we can more accurately isolate the effects of the treatment, leading to improved precision in our findings.

Fictional Example of Randomized Controlled Trial of Covi-Mapp for COVID-19 Management

Let's explore a fictional example to better understand experiments: a one-week randomized controlled trial of the experimental drug Covi-Mapp for managing Covid. In this case, the control group receives the standard care for Covid patients, while the treatment group receives the standard care plus Covi-Mapp. The outcome of interest is whether patients have cough symptoms on day 7, as subsidizing cough symptoms is an encouraging sign in Covid recovery. We'll measure the presence of cough on day 0 and day 7, as well as temperature on day 0 and day 7. Gender is also tracked. The control represents the standard care for COVID-19 patients, while the treatment includes standard care plus the experimental drug.

In this Covi-Mapp example, the intervention is the Covi-Mapp drug, the excludability principle is satisfied if the only difference in patient care between the groups is the drug administration, and the non-interference principle holds if one patient's outcome doesn't affect another's.

First, let's assume we have a dataset containing the relevant information for each patient, including cough status on day 0 and day 7, temperature on day 0 and day 7, treatment assignment, and gender. We'll read the data and explore the dataset:

library(data.table)

d <- fread("../data/COVID_rct.csv")

names(d)


"temperature_day0"  "cough_day0"        "treat_zmapp"       "temperature_day14" "cough_day14"       "male" 

Simple treatment effect of the experimental drug

Without any covariates, let's first look at the estimated effect of the treatment on the presence of cough on day 7. The estimated proportion of patients with a cough on day 7 for the control group (not receiving the experimental drug) is 0.847458. In other words, about 84.7% of patients in the control group are expected to have a cough on day 7, all else being equal. The estimated effect of the experimental drug on the presence of cough on day 7 is -0.23. This means that, on average, receiving the experimental drug reduces the proportion of patients with a cough on day 7 by 23.8% compared to the control group.

covid_1 <- d[ , lm(cough_day7 ~ treat_drug)]

coeftest(covid_1, vcovHC)


                 Estimate Std. Error t value Pr(>|t|)    

(Intercept)       0.847458   0.047616  17.798  < 2e-16 ***

treat_covid_mapp -0.237702   0.091459  -2.599  0.01079 *  

Signif. codes:  0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1

We know that a patient's initial condition would affect the final outcome. If the patient has a cough and a fever on day 0, they might not fare well with the treatment. To better understand the treatment's effect, let's add these covariates:

covid_2 <- d[ , lm(cough_day7 ~ treat_drug +

                   cough_day0 + temperature_day0)]

coeftest(covid_2, vcovHC)


                  Estimate Std. Error t value Pr(>|t|)   

(Intercept)      -19.469655   7.607812 -2.5592 0.012054 * 

treat_covid_mapp  -0.165537   0.081976 -2.0193 0.046242 * 

cough_day0         0.064557   0.178032  0.3626 0.717689   

temperature_day0   0.205548   0.078060  2.6332 0.009859 **

Signif. codes:  0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1

The output shows the results of a linear regression model, estimating the effect of the experimental drug (treat_covid_mapp) on the presence of cough on day 7, adjusting for cough on day 0 and temperature on day 0. The experimental drug significantly reduces the presence of cough on day 7 by approximately 16.6% compared to the control group (p-value = 0.046242). The presence of cough on day 0 does not significantly predict the presence of cough on day 7 (p-value = 0.717689). A one-unit increase in temperature on day 0 is associated with a 20.6% increase in the presence of cough on day 7, and this effect is statistically significant (p-value = 0.009859).

Should we add day 7 temperature as a covariate? By including it, we might find that the treatment is no longer statistically significant since the temperature on day 7 could be affected by the treatment itself. It is a post-treatment variable, and by including it, the experiment loses value as we used something that was affected by intervention as our covariate.

However, we'd like to investigate if the treatment affects men or women differently. Since we collected gender as part of the study, we could check for Heterogeneous Treatment Effect (HTE) for male vs. female. The experimental drug has a marginally significant effect on the outcome variable for females, reducing it by approximately 23.1% (p-value = 0.05391).

covid_4 <- d[ , lm(cough_day7 ~ treat_drug + treat_drug * male +

                   cough_day0 + temperature_day0)]

coeftest(covid_4, vcovHC)


t test of coefficients:


                  Estimate Std. Error  t value  Pr(>|t|)    

(Intercept)      48.712690  10.194000   4.7786 6.499e-06 ***

treat_zmapp      -0.230866   0.118272  -1.9520   0.05391 .  

male              3.085486   0.121773  25.3379 < 2.2e-16 ***

dehydrated_day0   0.041131   0.194539   0.2114   0.83301    

temperature_day0  0.504797   0.104511   4.8301 5.287e-06 ***

treat_zmapp:male -2.076686   0.198386 -10.4679 < 2.2e-16 ***

Signif. codes:  0 ‘***’ 0.001 ‘**’ 0.01 ‘*’ 0.05 ‘.’ 0.1 ‘ ’ 1

Which group, those coded as male == 0 or male == 1, have better health outcomes (cough) in control? What about in treatment? How does this help to contextualize any heterogeneous treatment effect that might have been estimated?

Stargazer is a popular R package that enables users to create well-formatted tables and reports for statistical analysis results.

covid_males <- d[male == 1, lm(temperature_day14 ~ treat_drug)]

covid_females <- d[male == 0, lm(temperature_day14 ~ treat_drug)]


stargazer(covid_males, covid_females,

          title = "",

          type = 'text',

          dep.var.caption = 'Outcome Variable:',

          dep.var.labels = c('Cough on Day 7'),

          se = list(

            sqrt(diag(vcov(covid_males))),

            sqrt(diag(vcovHC(covid_females))))

          )


===============================================================

                                 Outcome Variable:             

                               Temperature on Day 14           

                              (1)                   (2)        

treat_covid_mapp           -2.591***              -0.323*      

                            (0.220)               (0.174)      

Constant                  101.692***             98.487***     

                            (0.153)               (0.102)      

Observations                  37                    63         

R2                           0.798                 0.057       

Adjusted R2                  0.793                 0.041       

Residual Std. Error     0.669 (df = 35)       0.646 (df = 61)  

F Statistic         138.636*** (df = 1; 35) 3.660* (df = 1; 61)

===============================================================

Note:                               *p<0.1; **p<0.05; ***p<0.01

Looking at this regression report, we see that males in control have a temperature of 102; females in control have a temperature of 98.6 (which is very nearly a normal temperature). So, in control, males are worse off. In treatment, males have a temperature of 102 - 2.59 = 99.41. While this is closer to a normal temperature, this is still elevated. Females in treatment have a temperature of 98.5 - .32 = 98.18, which is slightly lower than a normal temperature, and is better than an elevated temperature. It appears that the treatment is able to have a stronger effect among male participants than females because males are *more sick* at baseline.

In conclusion, experimentation offers a fascinating and valuable avenue for primary research, allowing us to address causal questions and enhance our understanding of the world around us. Covariate control helps to isolate the causal effect of the treatment on the outcome variable, ensuring that the observed effect is not driven by confounding factors. Proper control of covariates enhances the internal validity of the study and ensures that the estimated treatment effect is an accurate representation of the true causal relationship. By exploring and accounting for sub groups in data, researchers can identify whether the treatment has different effects on different groups, such as men and women or younger and older individuals. This information can be critical for making informed policy decisions and developing targeted interventions that maximize the benefits for specific groups. The ongoing investigation of experimental methodologies and their potential applications represents a compelling and significant area of inquiry. 

Gerber, A. S., & Green, D. P. (2012). Field Experiments: Design, Analysis, and Interpretation . W. W. Norton.

“DALL·E 2.” OpenAI , https://openai.com/product/dall-e-2

“Data Science 241. Experiments and Causal Inference.” UC Berkeley School of Information , https://www.ischool.berkeley.edu/courses/datasci/241

Have a language expert improve your writing

Run a free plagiarism check in 10 minutes, generate accurate citations for free.

  • Knowledge Base

Methodology

  • Guide to Experimental Design | Overview, Steps, & Examples

Guide to Experimental Design | Overview, 5 steps & Examples

Published on December 3, 2019 by Rebecca Bevans . Revised on June 21, 2023.

Experiments are used to study causal relationships . You manipulate one or more independent variables and measure their effect on one or more dependent variables.

Experimental design create a set of procedures to systematically test a hypothesis . A good experimental design requires a strong understanding of the system you are studying.

There are five key steps in designing an experiment:

  • Consider your variables and how they are related
  • Write a specific, testable hypothesis
  • Design experimental treatments to manipulate your independent variable
  • Assign subjects to groups, either between-subjects or within-subjects
  • Plan how you will measure your dependent variable

For valid conclusions, you also need to select a representative sample and control any  extraneous variables that might influence your results. If random assignment of participants to control and treatment groups is impossible, unethical, or highly difficult, consider an observational study instead. This minimizes several types of research bias, particularly sampling bias , survivorship bias , and attrition bias as time passes.

Table of contents

Step 1: define your variables, step 2: write your hypothesis, step 3: design your experimental treatments, step 4: assign your subjects to treatment groups, step 5: measure your dependent variable, other interesting articles, frequently asked questions about experiments.

You should begin with a specific research question . We will work with two research question examples, one from health sciences and one from ecology:

To translate your research question into an experimental hypothesis, you need to define the main variables and make predictions about how they are related.

Start by simply listing the independent and dependent variables .

Research question Independent variable Dependent variable
Phone use and sleep Minutes of phone use before sleep Hours of sleep per night
Temperature and soil respiration Air temperature just above the soil surface CO2 respired from soil

Then you need to think about possible extraneous and confounding variables and consider how you might control  them in your experiment.

Extraneous variable How to control
Phone use and sleep in sleep patterns among individuals. measure the average difference between sleep with phone use and sleep without phone use rather than the average amount of sleep per treatment group.
Temperature and soil respiration also affects respiration, and moisture can decrease with increasing temperature. monitor soil moisture and add water to make sure that soil moisture is consistent across all treatment plots.

Finally, you can put these variables together into a diagram. Use arrows to show the possible relationships between variables and include signs to show the expected direction of the relationships.

Diagram of the relationship between variables in a sleep experiment

Here we predict that increasing temperature will increase soil respiration and decrease soil moisture, while decreasing soil moisture will lead to decreased soil respiration.

Receive feedback on language, structure, and formatting

Professional editors proofread and edit your paper by focusing on:

  • Academic style
  • Vague sentences
  • Style consistency

See an example

design field experiments

Now that you have a strong conceptual understanding of the system you are studying, you should be able to write a specific, testable hypothesis that addresses your research question.

Null hypothesis (H ) Alternate hypothesis (H )
Phone use and sleep Phone use before sleep does not correlate with the amount of sleep a person gets. Increasing phone use before sleep leads to a decrease in sleep.
Temperature and soil respiration Air temperature does not correlate with soil respiration. Increased air temperature leads to increased soil respiration.

The next steps will describe how to design a controlled experiment . In a controlled experiment, you must be able to:

  • Systematically and precisely manipulate the independent variable(s).
  • Precisely measure the dependent variable(s).
  • Control any potential confounding variables.

If your study system doesn’t match these criteria, there are other types of research you can use to answer your research question.

How you manipulate the independent variable can affect the experiment’s external validity – that is, the extent to which the results can be generalized and applied to the broader world.

First, you may need to decide how widely to vary your independent variable.

  • just slightly above the natural range for your study region.
  • over a wider range of temperatures to mimic future warming.
  • over an extreme range that is beyond any possible natural variation.

Second, you may need to choose how finely to vary your independent variable. Sometimes this choice is made for you by your experimental system, but often you will need to decide, and this will affect how much you can infer from your results.

  • a categorical variable : either as binary (yes/no) or as levels of a factor (no phone use, low phone use, high phone use).
  • a continuous variable (minutes of phone use measured every night).

How you apply your experimental treatments to your test subjects is crucial for obtaining valid and reliable results.

First, you need to consider the study size : how many individuals will be included in the experiment? In general, the more subjects you include, the greater your experiment’s statistical power , which determines how much confidence you can have in your results.

Then you need to randomly assign your subjects to treatment groups . Each group receives a different level of the treatment (e.g. no phone use, low phone use, high phone use).

You should also include a control group , which receives no treatment. The control group tells us what would have happened to your test subjects without any experimental intervention.

When assigning your subjects to groups, there are two main choices you need to make:

  • A completely randomized design vs a randomized block design .
  • A between-subjects design vs a within-subjects design .

Randomization

An experiment can be completely randomized or randomized within blocks (aka strata):

  • In a completely randomized design , every subject is assigned to a treatment group at random.
  • In a randomized block design (aka stratified random design), subjects are first grouped according to a characteristic they share, and then randomly assigned to treatments within those groups.
Completely randomized design Randomized block design
Phone use and sleep Subjects are all randomly assigned a level of phone use using a random number generator. Subjects are first grouped by age, and then phone use treatments are randomly assigned within these groups.
Temperature and soil respiration Warming treatments are assigned to soil plots at random by using a number generator to generate map coordinates within the study area. Soils are first grouped by average rainfall, and then treatment plots are randomly assigned within these groups.

Sometimes randomization isn’t practical or ethical , so researchers create partially-random or even non-random designs. An experimental design where treatments aren’t randomly assigned is called a quasi-experimental design .

Between-subjects vs. within-subjects

In a between-subjects design (also known as an independent measures design or classic ANOVA design), individuals receive only one of the possible levels of an experimental treatment.

In medical or social research, you might also use matched pairs within your between-subjects design to make sure that each treatment group contains the same variety of test subjects in the same proportions.

In a within-subjects design (also known as a repeated measures design), every individual receives each of the experimental treatments consecutively, and their responses to each treatment are measured.

Within-subjects or repeated measures can also refer to an experimental design where an effect emerges over time, and individual responses are measured over time in order to measure this effect as it emerges.

Counterbalancing (randomizing or reversing the order of treatments among subjects) is often used in within-subjects designs to ensure that the order of treatment application doesn’t influence the results of the experiment.

Between-subjects (independent measures) design Within-subjects (repeated measures) design
Phone use and sleep Subjects are randomly assigned a level of phone use (none, low, or high) and follow that level of phone use throughout the experiment. Subjects are assigned consecutively to zero, low, and high levels of phone use throughout the experiment, and the order in which they follow these treatments is randomized.
Temperature and soil respiration Warming treatments are assigned to soil plots at random and the soils are kept at this temperature throughout the experiment. Every plot receives each warming treatment (1, 3, 5, 8, and 10C above ambient temperatures) consecutively over the course of the experiment, and the order in which they receive these treatments is randomized.

Prevent plagiarism. Run a free check.

Finally, you need to decide how you’ll collect data on your dependent variable outcomes. You should aim for reliable and valid measurements that minimize research bias or error.

Some variables, like temperature, can be objectively measured with scientific instruments. Others may need to be operationalized to turn them into measurable observations.

  • Ask participants to record what time they go to sleep and get up each day.
  • Ask participants to wear a sleep tracker.

How precisely you measure your dependent variable also affects the kinds of statistical analysis you can use on your data.

Experiments are always context-dependent, and a good experimental design will take into account all of the unique considerations of your study system to produce information that is both valid and relevant to your research question.

If you want to know more about statistics , methodology , or research bias , make sure to check out some of our other articles with explanations and examples.

  • Student’s  t -distribution
  • Normal distribution
  • Null and Alternative Hypotheses
  • Chi square tests
  • Confidence interval
  • Cluster sampling
  • Stratified sampling
  • Data cleansing
  • Reproducibility vs Replicability
  • Peer review
  • Likert scale

Research bias

  • Implicit bias
  • Framing effect
  • Cognitive bias
  • Placebo effect
  • Hawthorne effect
  • Hindsight bias
  • Affect heuristic

Experimental design means planning a set of procedures to investigate a relationship between variables . To design a controlled experiment, you need:

  • A testable hypothesis
  • At least one independent variable that can be precisely manipulated
  • At least one dependent variable that can be precisely measured

When designing the experiment, you decide:

  • How you will manipulate the variable(s)
  • How you will control for any potential confounding variables
  • How many subjects or samples will be included in the study
  • How subjects will be assigned to treatment levels

Experimental design is essential to the internal and external validity of your experiment.

The key difference between observational studies and experimental designs is that a well-done observational study does not influence the responses of participants, while experiments do have some sort of treatment condition applied to at least some participants by random assignment .

A confounding variable , also called a confounder or confounding factor, is a third variable in a study examining a potential cause-and-effect relationship.

A confounding variable is related to both the supposed cause and the supposed effect of the study. It can be difficult to separate the true effect of the independent variable from the effect of the confounding variable.

In your research design , it’s important to identify potential confounding variables and plan how you will reduce their impact.

In a between-subjects design , every participant experiences only one condition, and researchers assess group differences between participants in various conditions.

In a within-subjects design , each participant experiences all conditions, and researchers test the same participants repeatedly for differences between conditions.

The word “between” means that you’re comparing different conditions between groups, while the word “within” means you’re comparing different conditions within the same group.

An experimental group, also known as a treatment group, receives the treatment whose effect researchers wish to study, whereas a control group does not. They should be identical in all other ways.

Cite this Scribbr article

If you want to cite this source, you can copy and paste the citation or click the “Cite this Scribbr article” button to automatically add the citation to our free Citation Generator.

Bevans, R. (2023, June 21). Guide to Experimental Design | Overview, 5 steps & Examples. Scribbr. Retrieved September 23, 2024, from https://www.scribbr.com/methodology/experimental-design/

Is this article helpful?

Rebecca Bevans

Rebecca Bevans

Other students also liked, random assignment in experiments | introduction & examples, quasi-experimental design | definition, types & examples, how to write a lab report, "i thought ai proofreading was useless but..".

I've been using Scribbr for years now and I know it's a service that won't disappoint. It does a good job spotting mistakes”

Listen-Hard

Conducting Effective Field Experiments in Psychology

design field experiments

Field experiments play a crucial role in psychology research, allowing researchers to study behavior in real-world settings. But what exactly are field experiments, and why are they important? In this article, we will explore the steps involved in designing and conducting effective field experiments, from identifying the research question to collecting and analyzing data. We will also discuss the limitations of field experiments in psychology and provide strategies to overcome challenges such as ethical concerns and unforeseen variables.

Join us as we delve into the world of field experiments in psychology!

  • Field experiments in psychology are valuable for studying behavior in natural settings, allowing for more accurate and generalizable results.
  • To design an effective field experiment, researchers must carefully identify research questions, select appropriate settings, determine sample size, and control for confounding variables.
  • To overcome challenges in conducting field experiments, researchers should address ethical concerns, anticipate and account for unforeseen variables, and strive for improved sample size and randomization.
  • 1 What Are Field Experiments in Psychology?
  • 2 Why Are Field Experiments Important in Psychology Research?
  • 3.1 Identify the Research Question
  • 3.2 Select the Appropriate Setting
  • 3.3 Determine the Sample Size
  • 3.4 Randomize the Participants
  • 3.5 Control for Confounding Variables
  • 3.6 Use Appropriate Measures and Instruments
  • 3.7 Consider Ethical Considerations
  • 4.1 Pre-Experiment Preparation
  • 4.2 Implementation of the Experiment
  • 4.3 Data Collection and Analysis
  • 5 What Are the Limitations of Field Experiments in Psychology?
  • 6.1 Addressing Ethical Concerns
  • 6.2 Dealing with Unforeseen Variables
  • 6.3 Improving Sample Size and Randomization
  • 7.1 What is the purpose of conducting effective field experiments in psychology?
  • 7.2 How do field experiments in psychology differ from laboratory experiments?
  • 7.3 What are some examples of effective field experiments in psychology?
  • 7.4 What are the benefits of conducting field experiments in psychology?
  • 7.5 What are some challenges of conducting field experiments in psychology?
  • 7.6 How can researchers ensure the validity and reliability of field experiments in psychology?

What Are Field Experiments in Psychology?

Field experiments in psychology refer to research studies conducted in real-life settings outside traditional laboratory environments, aiming to observe behavior and study variables in natural contexts.

These experiments play a crucial role in understanding how individuals behave authentically in their daily lives, offering insights that may be missed in controlled lab settings. By manipulating independent variables and observing the subsequent effects on dependent variables in the wild, researchers can uncover more realistic behaviors and responses. This focus on ecological validity enhances the generalizability and applicability of findings to real-world situations, providing a more comprehensive understanding of human behavior.

Why Are Field Experiments Important in Psychology Research?

Field experiments play a crucial role in psychology research by providing insights into human behavior in authentic settings, offering valuable data that enhances the understanding of psychological phenomena and their impact in real-world contexts.

By conducting experiments in natural environments, researchers can observe behaviors as they naturally occur, minimizing the artificiality of controlled lab settings. This method allows for a more ecologically valid study of behavior, leading to results that are more applicable to everyday life. Field experiments facilitate the collection of rich and diverse data, offering a broader perspective on how individuals interact with their surroundings. Researchers can also make direct observations of behavior without the influence of laboratory-induced biases, leading to more accurate interpretations of results.

How To Design an Effective Field Experiment?

Designing an effective field experiment involves careful consideration of variables, conditions, and methodological approaches to ensure the study’s validity and reliability, incorporating statistical methods for data analysis and result interpretation.

One of the key steps in designing a field experiment is selecting the appropriate variables to study, ensuring that they are both relevant to the research question and measurable. Researchers must define independent and dependent variables clearly to establish the cause-and-effect relationship. Establishing the study conditions is crucial, as controlling external factors can help isolate the impact of the manipulated variables. Methodological procedures, such as randomization and control groups, play a vital role in minimizing bias and ensuring the experiment’s internal validity.

Statistical methods such as regression analysis or ANOVA are then utilized to analyze the collected data, allowing researchers to draw meaningful conclusions and make informed decisions based on the results.

Identify the Research Question

The first step in designing a field experiment is to identify a clear research question or hypothesis that guides the study, ensuring that the recruitment process and standardization of procedures align with the null hypothesis for comprehensive analysis.

Formulating a research question serves as the backbone of any scientific investigation, providing a road map for the entire experiment. It helps researchers define the scope of their study, establish objectives, and ultimately draw meaningful conclusions. By recruiting participants that meet specific criteria related to the research question, the experiment’s results become more robust and reliable.

Maintaining standardized procedures ensures consistency in data collection and analysis, enabling researchers to accurately test the null hypothesis and draw valid inferences based on the findings.

Select the Appropriate Setting

Choosing the right setting for a field experiment is paramount, whether in a natural environment or through volunteer participants, to ensure the study’s external validity and relevance to real-world scenarios.

Conducting experiments in natural environments provides researchers with the opportunity to observe behaviors and reactions in settings that closely mirror real-world conditions, enhancing the authenticity of the findings. Involving volunteer participants can lead to a more diverse sample, increasing the generalizability of the study’s results. External validity, crucial for the application of research findings beyond the lab, hinges on how well the experiment’s setting reflects the complexities of the context being studied.

Determine the Sample Size

Determining the sample size for a field experiment involves calculating the number of participants required to achieve statistical significance and effectively represent the targeted population, ensuring the study’s reliability and validity.

One crucial aspect to consider when determining the sample size is the scale of the study. Larger studies often require a proportionally larger sample size to capture the intricacies and variations within the population accurately.

The relationship between the participants in the study can impact the sample size calculation. If the participants are expected to have similar responses or behaviors, a smaller sample size may suffice.

Identifying the level of precision desired in the results is essential. A higher level of precision usually necessitates a larger sample size to reduce the margin of error and enhance the research’s credibility.

Randomize the Participants

Randomizing participants in a field experiment through random allocation is crucial to establish control groups, enable randomization processes, and facilitate study replication for robust and reliable results.

This method ensures that each participant has an equal chance of being assigned to different groups, reducing the risk of bias and increasing the validity of research findings.

By employing random allocation, researchers can minimize the influence of confounding variables and increase the internal validity of the study.

Control groups, which remain untreated or receive a placebo, provide a baseline for comparison, allowing researchers to measure the true effects of the intervention or treatment. This comparison is vital in determining causality and distinguishing between correlation and causation.

Control for Confounding Variables

Controlling for confounding variables in a field experiment is essential to mitigate biases, address selection issues, and ensure that the study’s outcomes are attributed to the independent variables rather than external factors.

One common technique used to control for confounding variables is random assignment, where participants are allocated randomly to experimental groups. This helps ensure that any differences in outcomes are due to the treatment received and not other factors. Researchers can employ matched-pair designs, where participants are matched based on key characteristics. By doing so, the impact of these specific factors can be isolated and accounted for.

Use Appropriate Measures and Instruments

Utilizing appropriate measures and instruments in a field experiment is crucial to gather relevant data, implement study conditions effectively, and conduct thorough analysis using suitable methodologies.

Measures in a field experiment could range from basic tools like tape measures and thermometers to complex devices such as spectrophotometers or geospatial sensors, depending on the nature of the study. These instruments serve the vital role of ensuring accurate data collection, maintaining controlled study conditions, and facilitating the systematic evaluation of results. For instance, a digital spectrometer enables precise measurements of light absorption, crucial in environmental research, while data loggers automate data collection in long-term studies, reducing human error. Methodical analysis heavily relies on the quality and appropriateness of these tools, emphasizing the need for careful selection and calibration.

Consider Ethical Considerations

Ethical considerations are paramount in field experiments, necessitating informed consent from participants, addressing potential risks, and ensuring the study adheres to ethical guidelines to protect participant rights and well-being.

When conducting research involving human subjects, it is crucial to prioritize the ethical implications of the study. Informed consent serves as the cornerstone of ethical research, ensuring that participants are fully aware of the purpose, procedures, and potential risks involved in the study. This transparency allows individuals to make an informed decision about their participation, safeguarding their autonomy and right to self-determination.

Researchers must implement measures to protect the well-being and confidentiality of participants throughout the study. This includes minimizing any potential harm that participants may experience and upholding their privacy and data protection rights.

What Are the Steps Involved in Conducting a Field Experiment?

Conducting a field experiment entails several key steps, including pre-experiment preparation, implementation of the study, data collection, and analysis to examine behavioral patterns and draw meaningful conclusions.

Ahead of the actual experiment, researchers must carefully plan the study design, determine the research questions, and outline the variables to be observed. This initial phase also involves obtaining any necessary permissions, securing equipment, selecting suitable locations for data collection, and addressing any ethical considerations.

Regarding the implementation stage, it is crucial to adhere to the predefined protocols and procedures to ensure the reliability and validity of the experiment. Researchers often conduct pilot studies to fine-tune their approach and make any necessary adjustments before fully launching the field experiment.

Data collection methods in field experiments typically involve direct observation of subjects in their natural environment, recording behaviors, interactions, and responses to various stimuli. Researchers may employ tools such as checklists, behavior coding schemes, or video recordings to capture relevant data accurately.

After gathering the data, the next step is to analyze and interpret the findings. This involves organizing the collected information, identifying patterns, correlations, or trends, and drawing conclusions based on the observed behaviors. Statistical analysis and qualitative coding techniques are often utilized to make sense of the data and provide meaningful insights.

Pre-Experiment Preparation

The pre-experiment preparation phase involves conducting background research, selecting appropriate statistical methods, and establishing the methodology framework to ensure a structured and systematic approach to the field experiment.

Conducting thorough background research is essential as it provides a foundational understanding of the topic under study and helps in identifying gaps in existing literature.

Selecting the most suitable statistical methods is crucial for accurately analyzing and interpreting the data collected during the experiment.

Establishing a robust methodology framework not only ensures reproducibility but also enhances the credibility of the research findings. This comprehensive preparation stage lays the groundwork for a successful and meaningful field study.

Implementation of the Experiment

The implementation phase of a field experiment involves executing the study protocols, observing behavioral patterns, maintaining control over experimental conditions, and ensuring participant compliance to gather valid and reliable data.

During the behavior observation stage, researchers closely monitor how participants interact with the experimental setup and record any deviations from expected responses. Control mechanisms, such as random assignment or placebo groups, are put in place to minimize external influences on the outcomes of the study. Participant interactions play a crucial role as researchers engage with them to explain the study objectives and guidelines, ensuring their understanding and cooperation. For instance, in a study on consumer behavior, participants may be asked to navigate a simulated shopping experience to observe their decision-making processes.

Data Collection and Analysis

Data collection and analysis in field experiments are critical stages that involve interpreting results, assessing statistical significance, conducting thorough analysis, and validating findings through replication for robust conclusions.

During the data collection phase, researchers systematically gather information from the experimental setting, utilizing various tools such as surveys, observations, and sensor data to ensure comprehensive coverage. This raw data is then meticulously processed, organized, and cleaned to prepare it for analysis. Once the data is ready, statistical methods and analytical procedures are applied to identify patterns, relationships, and trends within the dataset.

Interpreting these results requires a deep understanding of the experimental design, the variables involved, and the potential impact of external factors on the outcomes.

What Are the Limitations of Field Experiments in Psychology?

Field experiments in psychology face inherent limitations, including criticisms regarding bias, expectancy effects, and challenges related to controlling extraneous variables that may impact the study outcomes.

One common critique of field experiments is the potential for selection bias, where participants self-select into certain groups, affecting the validity of the results. Due to the naturalistic setting of field experiments, researchers may struggle to isolate the specific variables influencing the outcomes, leading to confounding factors. Expectancy effects, where participants’ behavior is influenced by their expectations of the study’s results, can skew the data, making it challenging to draw accurate conclusions.

How to Overcome Challenges in Conducting Field Experiments?

Overcoming challenges in conducting field experiments requires addressing ethical concerns, implementing standardization protocols, and leveraging laboratory settings to enhance experimental control and reduce potential biases.

One strategy to tackle ethical concerns is to obtain informed consent from participants or stakeholders involved in the study, ensuring transparency and respect for autonomy. Establishing clear guidelines for data collection and analysis helps in standardizing practices across different research sites, leading to consistent and reliable results. Utilizing laboratory environments allows researchers to create controlled conditions, minimizing external influences on the experiment and enhancing the internal validity of the study.

Addressing Ethical Concerns

Addressing ethical concerns in field experiments involves prioritizing participant welfare, obtaining informed consent, and ensuring data confidentiality and integrity to uphold ethical standards and research integrity.

One of the key strategies to safeguard participant welfare in field experiments is by implementing thorough informed consent procedures. This involves clearly outlining the purpose of the study, potential risks and benefits, and the voluntary nature of participation.

Participant protection can also be ensured by minimizing harm through careful study design and monitoring. Another vital aspect is maintaining data privacy measures to safeguard confidential information and maintain the trust of participants. By adhering to these ethical considerations, researchers can uphold the credibility of their work and reinforce respect for participant rights.

Dealing with Unforeseen Variables

Handling unforeseen variables in field experiments necessitates controlling extraneous and confounding variables, implementing rigorous control measures, and addressing selection biases to minimize their impact on the study outcomes.

One crucial aspect of managing unexpected variables in field experiments is the meticulous design of the research protocol. By clearly defining the variables under study and establishing specific criteria for inclusion and exclusion, researchers can maintain clarity and consistency in data collection.

Employing randomization techniques can help distribute potential sources of variability evenly across treatment groups, reducing the influence of uncontrolled factors. Creating a well-structured data collection process and regularly monitoring data quality can enhance the reliability and validity of the findings.

Improving Sample Size and Randomization

Enhancing sample size and randomization techniques in field experiments involves scaling up participant numbers, refining randomization processes, conducting study replications, and employing robust statistical methods to enhance research validity and reliability.

One fundamental strategy to boost sample size in field experiments is by implementing participant scaling . This involves expanding the pool of participants to represent a more diverse population, increasing the generalizability of findings.

Advancements in randomization techniques, such as using cutting-edge algorithms and adaptive designs, can further enhance the randomness of group assignments, reducing bias and increasing the accuracy of results.

Incorporating rigorous replication practices, including multi-site studies and longitudinal designs, can help validate findings across different contexts.

Improving statistical methodologies with advanced techniques like Bayesian analysis and machine learning algorithms can provide more robust and sophisticated insights into the data collected.

Frequently Asked Questions

What is the purpose of conducting effective field experiments in psychology.

The purpose of conducting effective field experiments in psychology is to study behavior and psychological processes in natural settings, as opposed to a controlled laboratory environment, in order to enhance the generalizability of research findings and improve the understanding of real-world behaviors.

How do field experiments in psychology differ from laboratory experiments?

Field experiments in psychology differ from laboratory experiments in that they take place in natural settings, with participants’ behavior and responses being observed in their everyday environment. This allows for a more realistic and applicable understanding of behavior compared to artificial laboratory conditions.

What are some examples of effective field experiments in psychology?

Some examples of effective field experiments in psychology include studying the impact of social norms on behavior in a public setting, observing the effects of natural environments on mood and stress levels, and examining the effectiveness of interventions in real-world settings.

What are the benefits of conducting field experiments in psychology?

Conducting field experiments in psychology has several benefits, including improved external validity, increased ecological validity, and the ability to study real-world behaviors and processes in their natural context. This can lead to a better understanding of human behavior and more accurate applications of research findings.

What are some challenges of conducting field experiments in psychology?

Some challenges of conducting field experiments in psychology include difficulty controlling extraneous variables, potential ethical concerns, and limited control over the research environment. Additionally, field experiments may require more resources and time compared to laboratory experiments.

How can researchers ensure the validity and reliability of field experiments in psychology?

To ensure the validity and reliability of field experiments in psychology, researchers should carefully design their study and control for potential confounds. This may include using random assignment, pre-testing and post-testing, and developing specific protocols for data collection and analysis. It is also important to replicate the study in different settings to ensure consistent results.

' src=

Marcus Wong, a cognitive neuroscientist, explores the mysteries of the human brain and behavior. His work in experimental psychology and brain imaging techniques has contributed to our understanding of memory, decision-making, and neural mechanisms underlying cognitive functions. Marcus is committed to making complex scientific concepts accessible to a broad audience, writing about the latest trends in neuroscience, cognitive enhancement, and the intersection of technology with brain health.

Similar Posts

Exploring the Trichromatic Theory in Psychology: Insights into Color Perception

Exploring the Trichromatic Theory in Psychology: Insights into Color Perception

The article was last updated by Sofia Alvarez on February 9, 2024. Have you ever wondered how we perceive color? The Trichromatic Theory, proposed by…

Understanding the Focus of Social Psychology

Understanding the Focus of Social Psychology

The article was last updated by Alicia Rhodes on February 4, 2024. Social psychology delves into the intricacies of human behavior within the social context,…

Understanding the Significance of ‘Pi’ in Psychological Research

Understanding the Significance of ‘Pi’ in Psychological Research

The article was last updated by Alicia Rhodes on February 9, 2024. Have you ever wondered how psychologists measure and understand the complexities of the…

The Psychology of Love: What Research Says

The Psychology of Love: What Research Says

The article was last updated by Vanessa Patel on February 4, 2024. Love is a complex and multifaceted emotion that plays a significant role in…

The Significance of Music Psychology in the Broad Field of Psychology

The Significance of Music Psychology in the Broad Field of Psychology

The article was last updated by Ethan Clarke on February 8, 2024. Have you ever wondered how music can impact our brain, emotions, and overall…

Exploring Different Perspectives in Psychology

Exploring Different Perspectives in Psychology

The article was last updated by Nicholas Reed on February 5, 2024. Have you ever wondered what makes us think, feel, and behave the way…

China

I want to publish

To find out how to publish or submit your book proposal:

To find a journal or submit your article to a journal:

  • Field Experiment
  • Medicine and Healthcare
  • Medical Statistics & Computing

Experiments

Louis Cohen, Lawrence Manion, Keith Morrison in Research Methods in Education , 2017

The design experiment can be considered as a special case of a field experiment; it has its roots in experimental research, both in ‘true’ and quasi-experiments, and is intended to provide formative feedback on, for example, practical problems in, say, teaching and learning, and to bridge the potential gap between research and practice (Brown, 1992, p. 143; Reinking and Bradley, 2008; Bradley and Reinking, 2011; Engeström, 2011; Seel, 2011, p. 925; Anderson and Shattuck, 2012; Laurillard, 2012), in other words, to enhance the external validity of an experiment. The design experiment strives to avoid the artificial world of the laboratory and the lack of applicability to ‘real-world problems’ that follows from this artificial condition (Bradley and Reinking, 2011; Reinking and Bradley, 2008; Seel, 2011; Laurillard, 2012), and to have direct practical relevance to the complex world of teaching, learning and classrooms. Given their intended direct relevance to classrooms and the field nature – the diverse, complex, ‘real world’ of an actual classroom – design experiments may not be able to fulfil the requirements of a true experiment, for example, in randomization or in the application of controls. In these respects, design experiments are similar to action research (cf. Anderson and Shattuck, 2012).

External validity and public health

Sridhar Venkatapuram, Alex Broadbent in The Routledge Handbook of Philosophy of Public Health , 2023

The other reason they give for doubting the trade-off is because the notion of the “artificiality” of experimental settings, used to justify its existence, is vague and ambiguous (Jimenez-Buedo and Miller 2010: 307). The difficulty in blaming the artificiality of experimental environments for external validity failure is that none of the proponents of this view explain in exactly which respects the environments are meant to differ. In any case, it is clear that there is a lot more to the external validity problem than just a concern about the artificiality of experimental settings. This can be seen by considering the case of so-called field experiments or observational studies, where causes and effects are observed in real-world settings. In these studies, there is nothing artificial about the experimental context; yet, the external validity question is still a legitimate concern. This is because we can have cases where a field experiment exposes a clear causal relation evident in one context, but this causal relation does not hold in a new context because of different confounding factors, for example.

Randomization Tests or Permutation Tests? A Historical and Terminological Clarification

Vance W. Berger in Randomization, Masking, and Allocation Concealment , 2017

Interestingly, another statistical heavyweight, Jerzy Neyman, did something very similar with respect to giving credit to Fisher for the randomization design principle as Edward Pitman did with respect to giving credit to Fisher for developing the test. In his notorious paper, read before the Industrial and Agricultural Research Section of the Royal Statistical Society, Neyman (1935, p. 109) stated:Owing to the work of R. A. Fisher, “Student” and their followers, it is hardly possible to add anything essential to the present knowledge concerning local experiments…. One of the most important achievements of the English School is their method of planning field experiments known as the method of Randomized Blocks and Latin Squares.

Does an economic incentive affect provider behavior? Evidence from a field experiment on different payment mechanisms

Published in Journal of Medical Economics , 2019

Xiaoyu Xi, Ennan Wang, Qianni Lu, Piaopiao Chen, Tian Wo, Kammy Tang

We used a field experiment study design to examine the behaviors of physicians. If a laboratory experiment were performed instead of a field experiment, the conclusions might not be valid due to hyper-abstraction and simplification26. However, the participants in a field study were not restricted to college students, instead, they were adults in society. Moreover, the experimental environment was not confined to a laboratory. A field experiment, as defined by Harrison and List27, was an experiment conducted in multiple locations, including laboratories and actual environments. Its participants included both students and non-college adults. Therefore, under the real social conditional, the experiment subjects could make realistic choices. Above all, because of the differences between the experimental environment and subjects, the field experiment could represent actual conditions in a real environment, and subjects might act instinctively as they do in daily life, increasing the external validity of results28.

Employees’ Improvisational Behavior: Exploring the Role of Leader Grit and Humility

Published in Human Performance , 2022

Arménio Rego, Andreia Vitória, Miguel Pina e Cunha, Bradley P. Owens, Ana Ventura, Susana Leal, Camilo Valverde, Rui Lourenço-Gil

While providing overall support for the proposed causal direction of our model, our research is not without limitations. First, other causalities are possible. For example, employees may develop higher self-efficacy, hope, and optimism after making improvisations that are revealed to be successful. It is also possible that leaders adopt more perseverant efforts in pursuing challenging goals as a consequence of their higher employees’ PsyCap. Although the experiment was designed to enhance realism, which enhances confidence in hypothesized causality, it also suffers from modest external validity and other limitations (Lonati et al., 2018). Future studies should include covariates to rule out confounding and endogeneity effects, should adopt other experimental designs, and should be carried out in real organizational settings (Antonakis, Bendahan, Jacquart, & Lalive, 2014). A field experiment would represent a very important step forward in that endeavor, although abundant obstacles (methodological and practical) may make the endeavor unfeasible. Second, future studies may explore boundary conditions of the PsyCap-improvisation relationship. For example, is the relationship more positive when employees experience psychological safety?

Impact of safety training and interventions on training-transfer: targeting migrant construction workers

Published in International Journal of Occupational Safety and Ergonomics , 2020

Rahat Hussain, Akeem Pedro, Do Yeop Lee, Hai Chien Pham, Chan Sik Park

Notwithstanding, even though the culturally diverse nature of the work crews is a valuable aspect to evaluate migrant worker safety performance, it is hard to engage labourers for such practice. Similarly, the extent of interventions being implemented during and after training sessions is difficult to control. To address these challenges, the experiment design phase of this research has focused more on the reasons for training failures from both the literature and current migrant worker issues in industry. The approach presented in this study measures the combined impact of all interventions in a field experiment; however, in order to improve the reliability and validity in measurement, the individual effects of interventions should also be considered in further studies.

Related Knowledge Centers

  • External Validity
  • Randomization
  • Statistical Inference
  • Random Assignment
  • Rubin Causal Model
  • Standard Deviation
  • Accuracy & Precision
  • Sample Size Determination
  • Stepped-Wedge Trial

Current Research

  • Clinical Trials (United States)
  • Clinical Trials (Europe)
  • Clinical Trials (Australia/New Zealand)
  • Clinical Trials (India)

Knowledge is an evolving asset. Help us improve this page for a future release.

  • Affiliated Professors
  • Invited Researchers
  • J-PAL Scholars
  • Diversity, Equity, and Inclusion
  • Code of Conduct
  • Initiatives
  • Latin America and the Caribbean
  • Middle East and North Africa
  • North America
  • Southeast Asia
  • Agriculture
  • Crime, Violence, and Conflict
  • Environment, Energy, and Climate Change
  • Labor Markets
  • Political Economy and Governance
  • Social Protection
  • Evaluations
  • Research Resources
  • Policy Insights
  • Evidence to Policy
  • For Affiliates
  • Support J-PAL

The Abdul Latif Jameel Poverty Action Lab (J-PAL) is a global research center working to reduce poverty by ensuring that policy is informed by scientific evidence. Anchored by a network of more than 1,000 researchers at universities around the world, J-PAL conducts randomized impact evaluations to answer critical questions in the fight against poverty.

  • Affiliated Professors Our affiliated professors are based at 97 universities and conduct randomized evaluations around the world to design, evaluate, and improve programs and policies aimed at reducing poverty. They set their own research agendas, raise funds to support their evaluations, and work with J-PAL staff on research, policy outreach, and training.
  • Board Our Board of Directors, which is composed of J-PAL affiliated professors and senior management, provides overall strategic guidance to J-PAL, our sector programs, and regional offices.
  • Diversity, Equity, and Inclusion J-PAL recognizes that there is a lack of diversity, equity, and inclusion in the field of economics and in our field of work. Read about what actions we are taking to address this.
  • Initiatives J-PAL initiatives concentrate funding and other resources around priority topics for which rigorous policy-relevant research is urgently needed.
  • Events We host events around the world and online to share results and policy lessons from randomized evaluations, to build new partnerships between researchers and practitioners, and to train organizations on how to design and conduct randomized evaluations, and use evidence from impact evaluations.
  • Blog News, ideas, and analysis from J-PAL staff and affiliated professors.
  • News Browse news articles about J-PAL and our affiliated professors, read our press releases and monthly global and research newsletters, and connect with us for media inquiries.
  • Press Room Based at leading universities around the world, our experts are economists who use randomized evaluations to answer critical questions in the fight against poverty. Connect with us for all media inquiries and we'll help you find the right person to shed insight on your story.
  • Overview J-PAL is based at MIT in Cambridge, MA and has seven regional offices at leading universities in Africa, Europe, Latin America and the Caribbean, Middle East and North Africa, North America, South Asia, and Southeast Asia.
  • Global Our global office is based at the Department of Economics at the Massachusetts Institute of Technology. It serves as the head office for our network of seven independent regional offices.
  • Africa J-PAL Africa is based at the Southern Africa Labour & Development Research Unit (SALDRU) at the University of Cape Town in South Africa.
  • Europe J-PAL Europe is based at the Paris School of Economics in France.
  • Latin America and the Caribbean J-PAL Latin America and the Caribbean is based at the Pontificia Universidad Católica de Chile.
  • Middle East and North Africa J-PAL MENA is based at the American University in Cairo, Egypt.
  • North America J-PAL North America is based at the Massachusetts Institute of Technology in the United States.
  • South Asia J-PAL South Asia is based at the Institute for Financial Management and Research (IFMR) in India.
  • Southeast Asia J-PAL Southeast Asia is based at the Faculty of Economics and Business at the University of Indonesia (FEB UI).
  • Overview Led by affiliated professors, J-PAL sectors guide our research and policy work by conducting literature reviews; by managing research initiatives that promote the rigorous evaluation of innovative interventions by affiliates; and by summarizing findings and lessons from randomized evaluations and producing cost-effectiveness analyses to help inform relevant policy debates.
  • Agriculture How can we encourage small farmers to adopt proven agricultural practices and improve their yields and profitability?
  • Crime, Violence, and Conflict What are the causes and consequences of crime, violence, and conflict and how can policy responses improve outcomes for those affected?
  • Education How can students receive high-quality schooling that will help them, their families, and their communities truly realize the promise of education?
  • Environment, Energy, and Climate Change How can we increase access to energy, reduce pollution, and mitigate and build resilience to climate change?
  • Finance How can financial products and services be more affordable, appropriate, and accessible to underserved households and businesses?
  • Firms How do policies affecting private sector firms impact productivity gaps between higher-income and lower-income countries? How do firms’ own policies impact economic growth and worker welfare?
  • Gender How can we reduce gender inequality and ensure that social programs are sensitive to existing gender dynamics?
  • Health How can we increase access to and delivery of quality health care services and effectively promote healthy behaviors?
  • Labor Markets How can we help people find and keep work, particularly young people entering the workforce?
  • Political Economy and Governance What are the causes and consequences of poor governance and how can policy improve public service delivery?
  • Social Protection How can we identify effective policies and programs in low- and middle-income countries that provide financial assistance to low-income families, insuring against shocks and breaking poverty traps?

Handbook of Field Experiments

The last 15 years have seen an explosion in the number, scope, quality, and creativity of field experiments. To take stock of this remarkable progress, we were invited to edit a Handbook of Field Experiments , published at Elsevier. We were fortunate to assemble a volume made of wonderful papers by the best experts in the field. Some chapters are more methodological, while others are focused on results. All of them provide thoughtful reflections on the advances and issues in the field, useful research tips and insights into what the next steps need to be, all of which should be very useful for graduate students. Taken together, these papers offer an incredibly rich overview of the state of literature. This page collects together all the working paper versions of the chapters, and will also link to the final versions as they become available. We hope you enjoy it.

—Abhijit Banerjee and Esther Duflo

Introduction

An Introduction to the "Handbook of Field Experiments" Abhijit Banerjee and Esther Duflo

Many (though by no means all) of the questions that economists and policymakers ask themselves are causal in nature: What would be the impact of adding computers in classrooms? What is the price elasticity of demand for preventive health products? Would increasing interest rates lead to an increase in default rates? Decades ago, the statistician Fisher (Fisher, 1925) proposed a method to answer such causal questions: Randomized Controlled Trials (RCTs) . In an RCT, the assignment of different units to different treatment groups is chosen randomly. This ensures that no unobservable characteristics of the units are reflected in the assignment, and hence that any difference between treatment and control units reflects the impact of the treatment. While the idea is simple, the implementation in the field can be more involved, and it took some time before randomization was considered to be a practical tool for answering questions in economics.

Some Historical Background

The Politics and Practice of Social Experiments: Seeds of a Revolution Judy Gueron

Between 1970 and the early 2000s, there was a revolution in support for the use of randomized experiments to evaluate social programs. Focusing on the welfare reform studies that helped to speed that transformation in the United States, this chapter describes the major challenges to randomized controlled trials (RCTs), how they emerged and were overcome, and how initial conclusions about conditions necessary to success — strong financial incentives, tight operational control, and small scale — proved to be wrong. The final section discusses lessons from this experience for other fields.

Methodology and Practice of RCTs

The Econometrics of Randomized Experiments Susan Athey and  Guido Imbens

Randomized experiments have a long tradition in agricultural and biomedical settings. In economics they have a much shorter history. Although there have been notable experiments over the years, such as the RAND health care experiment (Manning, Newhouse, Duan, Keeler and Leibowitz, 1987, see the general discussion in Rothstein and von Wachter, 2016) and the Negative Income Tax experiments (e.g., Robins, 1985), it is only recently that there has been a large number of randomized experiments in economics, and development economics in particular. See Duflo, Glennerster, and Kremer (2006) for a survey.  In this chapter we discuss some of the statistical methods that are important for the analysis and design of randomized experiments. A major theme of the chapter is the focus on statistical methods directly justified by randomization, in the spirit of Freedman who wrote “Experiments should be analyzed as experiments, not as observational studies. A simple comparison of rates might be just the right tool, with little value added by ‘sophisticated’ models,” (Freedman, 2006, p. 691) We draw from a variety of literatures. This includes the statistical literature on the analysis and design of experiments, e.g., Wu and Hamada (2009), Cox and Reid (2000), Altman (1991), Cook and DeMets (2008), Kempthorne (1952, 1955), Cochran and Cox (1957), Davies (1954), and Hinkelman and Kempthorne (2005, 2008). We also draw on the literature on causal inference, both in experimental and observational settings, Rosenbaum (1995, 2002, 2009), Rubin (2006), Cox (1992), Morgan and Winship (2007), Morton Williams (2010) and Lee (2005), and Imbens and Rubin (2015). In the economics literature we build on recent guides to practice in randomized experiments in development economics, e.g., Duflo, Glennerster, and Kremer (2006), Glennerster (2016), and Glennerster and Takavarasha (2013) as well as the general empirical micro literature (Angrist and Pischke, 2008).

Decision Theoretic Approaches to Experiment Design and External Validity Abhijit Banerjee, Sylvain Chassang,  and Erik Snowberg

A modern, decision-theoretic framework can help clarify important practical questions of experimental design. Building on our recent work, this chapter begins by summarizing our framework for understanding the goals of experimenters, and applying this to re-randomization.  We then use this framework to shed light on questions related to experimental registries, pre-analysis plans, and most importantly, external validity. Our framework implies that even when large samples can be collected, external decisionmaking remains inherently subjective. We embrace this conclusion, and argue that in order to improve external validity, experimental research needs to create a space for structured speculation.

The Practicalities of Running Randomized Evaluations: Partnerships, Measurement, Ethics, and Transparency Rachel Glennerster

Economists have known for a long time that randomization could help identify causal connections by solving the problem of selection bias. Chapter 1 in this book and Gueron and Rolston (2013) describe the effort in the US to move experiments out of the laboratory into the policy world in the 1960s and 1970s.  This experience was critical in proving the feasibility of field experiments, working through some of the important ethical questions involved, showing how researchers and practitioners could work together, and demonstrating that the results of field experiments were often very different from those generated by observational studies. Interestingly, there was relatively limited academic support for this first wave of field experiments (Gueron and Rolston 2013), most of which were carried out by research groups such as MDRC, Abt, and Mathematica, to evaluate US government programs, and they primarily used individual-level randomization. In contrast, a more recent wave of field experiments starting in the mid-1990s was driven by academics, initially was focused on developing countries, often worked with nongovernmental organizations, and frequently used clustered designs.

The Psychology of Construal in the Design of Field Experiments Elizabeth Levy Paluck and Eldar Shafir

Why might you be interested in this chapter? A fair assumption is that you are reading because you care about good experimental design. To create strong experimental designs that test people’s responses to an intervention, researchers typically consider the classically recognized motivations presumed to drive human behavior.  It does not take extensive psychological training to recognize that several types of motivations could affect an individual’s engagement with and honesty during your experimental paradigm. Such motivations include strategic self-presentation, suspicion, lack of trust, level of education or mastery, and simple utilitarian motives such as least effort and optimization. For example, minimizing the extent to which your findings are attributable to high levels of suspicion among participants, or to their decision to do the least amount possible, is important for increasing the generalizability and reliability of your results.

Understanding Preferences and Preference Change

Field Experiments in Markets Omar Al-Ubaydli and  John List

This is a review of the literature of field experimental studies of markets. The main results covered by the review are as follows: (1) Generally speaking, markets organize the efficient exchange of commodities; (2) There are some behavioral anomalies that impede efficient exchange; (3) Many behavioral anomalies disappear when traders are experienced.

Field Experiments on Discrimination Marianne Bertrand and Esther Duflo

This article reviews the existing field experimentation literature on the prevalence of discrimination, the consequences of such discrimination, and possible approaches to undermine it. We highlight key gaps in the literature and ripe opportunities for future field work.  Section 1 reviews the various experimental methods that have been employed to measure the prevalence of discrimination, most notably audit and correspondence studies; it also describes several other measurement tools commonly used in lab-based work that deserve greater consideration in field research. Section 2 provides an overview of the literature on the costs of being stereotyped or discriminated against, with a focus on self-expectancy effects and self-fulfilling prophecies; section 2 also discusses the thin field-based literature on the consequences of limited diversity in organizations and groups. The final section of the paper, Section 3, reviews the evidence for policies and interventions aimed at weakening discrimination, covering role model and intergroup contact effects, as well as socio-cognitive and technological de-biasing strategies.

Field Experiments on Voter Mobilization: An Overview of a Burgeoning Literature Alan Gerber and Donald Green

In recent years the focus of empirical work in political science has begun to shift from description to an increasing emphasis on the credible estimation of causal effects. A key feature of this change has been the increasing prominence of experimental methods, and especially field experiments. In this chapter we review the use of field experiments to study political participation.  Although several important experiments address political phenomena other than voter participation (Bergan 2009; Butler and Broockman 2015; Butler and Nickerson 2011; Broockman 2013, 2014; Grose 2014), the literature measuring the effect of various interventions on voter turnout is the largest and most fully developed, and it provides a good illustration of how the use of field experiments in political science has proceeded. From an initial focus on the relative effects of different modes of communication, scholars began to explore how theoretical insights from social psychology and behavioral economics might be used to craft messages and how voter mobilization experiments could be employed to test the real world effects of theoretical claims. The existence of a large number of experimental turnout studies was essential, because it provided the background against which unusual and important results could be easily discerned.

Lab in the Field: Measuring Preferences in the Wild Uri Gneezy and Alex Imas

In this chapter, we discuss the “lab-in-the-field” methodology, which combines elements of both lab and field experiments in using standardized, validated paradigms from the lab in targeting relevant populations in naturalistic settings. We begin by examining how the methodology has been used to test economic models with populations of theoretical interest. Next, we outline how lab-in-the-field studies can be used to complement traditional Randomized Control Trials in collecting covariates to test theoretical predictions and explore behavioral mechanisms. We proceed to discuss how the methodology can be utilized to compare behavior across cultures and contexts, and test for the external validity of results obtained in the lab. The chapter concludes with an overview of lessons on how to use the methodology effectively.

Field Experiments in Marketing Duncan Simester

Marketing is a diverse field that draws from a rich array of disciplines and a broad assortment of empirical and theoretical methods. One of those disciplines is economics and one of the methods used to investigate economic questions is field experiments. The history of field experiments in the marketing literature is surprisingly long. Early examples include Curhan (1974) and Eskin and Baron (1977), who vary prices, newspaper advertising, and display variables in grocery stores.  This chapter reviews the recent history of field experiments in marketing by identifying papers published in the last 20 years (between 1995 and 2014). We report how the number of papers published has increased during this period, and evaluate different explanations for this increase. We then group the papers into five topics and review the papers by topic. The chapter concludes by reflecting on the design of field experiments used in marketing, and proposing topics for future research.

The Challenge of Improving Human Capital

Impacts and Determinants of Health Levels in Low-Income Countries Pascaline Dupas and Ted Miguel

Improved health in low-income countries could considerably improve wellbeing and possibly promote economic growth. The last decade has seen a surge in field experiments designed to understand the barriers that households and governments face in investing in health and how these barriers can be overcome, and to assess the impacts of subsequent health gains. This chapter first discusses the methodological pitfalls that field experiments in the health sector are particularly susceptible to, then reviews the evidence that rigorous field experiments have generated so far.  While the link from in utero and child health to later outcomes has increasingly been established, few experiments have estimated the impacts of health on contemporaneous productivity among adults, and few experiments have explored the potential for infrastructural programs to impact health outcomes. Many more studies have examined the determinants of individual health behavior, on the side of consumers as well as among providers of health products and services.

The Production of Human Capital in Developed Countries: Evidence from 196 Randomized Field Experiments Roland Fryer

Randomized field experiments designed to better understand the production of human capital have increased exponentially over the past several decades. This chapter summarizes what we have learned about various partial derivatives of the human capital production function, what important partial derivatives are left to be estimated, and what – together – our collective efforts have taught us about how to produce human capital in developed countries. The chapter concludes with a back of the envelope simulation of how much of the racial wage gap in America might be accounted for if human capital policy focused on best practices gleaned from randomized field experiments.

Field Experiments in Education in Developing Countries Karthik Muralidharan Perhaps no field in development economics in the past decade has benefited as much from the use of experimental methods as the economics of education. The rapid growth in high‐quality studies on education in developing countries (many of which use randomized experiments) is perhaps best highlighted by noting that there have been  several  systematic reviews of this evidence aiming to synthesize findings for research and policy in  just the past three years .   These include Muralidharan 2013 (focused on India), Glewwe et al. 2014 (focused on school inputs), Kremer et al. 2013, Krishnaratne et al. 2013, Conn 2014 (focused on sub‐Saharan Africa), McEwan 2014, Ganimian and Murnane (2016), Evans and Popova (2015), and Glewwe and Muralidharan (2016). While these are not all restricted to experimental studies, they typically provide greater weight to evidence from randomized controlled trials (RCT's).

Designing Effective Social Programs

Social Policy: Mechanism Experiments and Policy Evaluations Bill Congdon,  Jeffrey Kling, Jens Ludwig, and Sendhil Mullainathan

Policymakers and researchers are increasingly interested in using experimental methods to inform the design of social policy. The most common approach, at least in developed countries, is to carry out large-scale randomized trials of the policies of interest, or what we call here policy evaluations. In this chapter we argue that in some circumstances the best way to generate information about the policy of interest may be to test an intervention that is different from the policy being considered, but which can shed light on one or more key mechanisms through which that policy may operate.  What we call mechanism experiments can help address the key external validity challenge that confronts all policy-oriented work in two ways. First, mechanism experiments sometimes generate more policy-relevant information per dollar of research funding than can policy evaluations, which in turn makes it more feasible to test how interventions work in different contexts. Second, mechanism experiments can also help improve our ability to forecast effects by learning more about the way in which local context moderates policy effects, or expand the set of policies for which we can forecast effects. We discuss how mechanism experiments and policy evaluations can complement one another, and provide examples from a range of social policy areas including health insurance, education, labor market policy, savings and retirement, housing, criminal justice, redistribution, and tax policy. Examples focus on the U.S. context.

Field Experiments in Developing Country Agriculture Alain de Janvry, Elisabeth Sadoulet, and Tavneet Suri

This chapter provides a review of the role of field experiments in answering research questions in agriculture that ultimately let us better understand how policy can improve productivity and farmer welfare in developing economies. We first review recent field experiments in this area, highlighting the contributions experiments have already made to this area of research. We then outline areas where experiments can further fill existing gaps in our knowledge on agriculture and how future experiments can address the specific complexities in agriculture.

The Personnel Economics of the State Frederico Finan, Ben Olken, and Rohini Pande

Governments play a central role in facilitating economic development. Yet while economists have long emphasized the importance of government quality, historically they have paid less attention to the internal workings of the state and the individuals who provide the public services. This chapter reviews a nascent but growing body of field experiments that explores the personnel economics of the state.  To place the experimental findings in context, we begin by documenting some stylized facts about how public sector employment differs from that in the private sector. In particular, we show that in most countries throughout the world, public sector employees enjoy a significant wage premium over their private sector counterparts. Moreover, this wage gap is largest among low-income countries, which tends to be precisely where governance issues are most severe. These differences in pay, together with significant information asymmetries within government organizations in low-income countries, provide a prima facie rationale for the emphasis of the recent field experiments on three aspects of the state–employee relationship: selection, incentive structures, and monitoring. We review the findings on all three dimensions and then conclude this survey with directions for future research.

Designing Social Protection Programs: Using Theory and Experimentation to Understand how to Help Combat Poverty Rema Hanna and Dean Karlan

“Anti-poverty” programs come in many varieties, ranging from multi-faceted, complex programs to more simple cash transfers. Articulating and understanding the root problem motivating government and nongovernmental organization intervention is critical for choosing amongst many anti-poverty policies, or combinations thereof. Policies should differ depending on whether the underlying problem is about uninsured shocks, liquidity constraints, information failures, or some combination of all of the above.  Experimental designs and thoughtful data collection can help diagnose the root problems better, thus providing better predictions for what anti-poverty programs to employ in specific conditions and contexts. However, the more complex theories are likewise more challenging to test, requiring larger samples, and often more nuanced experimental designs, as well as detailed data on many aspects of household and community behavior and outcomes. We provide guidance on these design and testing issues for social protection programs, from how to target programs, to who should implement the program, to whether and what conditions to require for program participation. In short, careful experimentation designed testing can help provide a stronger conceptual understanding of why programs do or not work, thereby allowing one to ultimately make stronger policy prescriptions that further the goal of poverty reduction.

Social Experiments in the Labor Market Jesse Rothstein and  Till von Wachter

Large-scale social experiments were pioneered in labor economics, and are the basis for much of what we know about topics ranging from the effect of job training to incentives for job search to labor supply responses to taxation. Random assignment has provided a powerful solution to selection problems that bedevil non- experimental research. Nevertheless, many important questions about these topics require going beyond random assignment.  This applies to questions pertaining to both internal and external validity, and includes effects on endogenously observed outcomes, such as wages and hours; spillover effects; site effects; heterogeneity in treatment effects; multiple and hidden treatments; and the mechanisms producing treatment effects. In this Chapter, we review the value and limitations of randomized social experiments in the labor market, with an emphasis on these design issues and approaches to addressing them. These approaches expand the range of questions that can be answered using experiments by combining experimental variation with econometric or theoretical assumptions. We also discuss efforts to build the means of answering these types of questions into the ex ante design of experiments. Our discussion yields an overview of the expanding toolkit available to experimental researchers.

Princeton University Logo

  • Help & FAQ

Field Experiments

  • Princeton School of Public and International Affairs

Research output : Chapter in Book/Report/Conference proceeding › Chapter

Field experiments are experiments in settings with high degrees of naturalism. This article describes different types of field experiments, including randomized field trials, randomized rollout designs, encouragement designs, downstream field experiments, hybrid lab-field experiments, and covert population experiments, and discusses their intellectual background and benefits. It also lists methodological challenges researchers can encounter when conducting field experiments, including failure to treat, selective attrition, spillover, difficulty of replication, and black box causality, and discusses available solutions. Finally, it provides an overview over current and emerging directions in field experimentation and concludes with a brief history of field experiments.

Original languageEnglish (US)
Title of host publicationInternational Encyclopedia of the Social & Behavioral Sciences: Second Edition
Publisher
Pages128-134
Number of pages7
ISBN (Electronic)9780080970875
ISBN (Print)9780080970868
DOIs
StatePublished - Mar 26 2015

All Science Journal Classification (ASJC) codes

  • General Social Sciences
  • Audit studies
  • Correspondence studies
  • Encouragement designs
  • Experiments
  • External validity
  • Failure to treat
  • Hybrid field experiments
  • Randomized controlled trial
  • Randomized rollout designs
  • Social experiments
  • Stepped wedge designs
  • Waiting list designs

Access to Document

  • 10.1016/B978-0-08-097086-8.10542-2

Other files and links

  • Link to publication in Scopus
  • Link to the citations in Scopus

Fingerprint

  • Hybrid Labs Keyphrases 100%
  • Naturalism Keyphrases 100%
  • Encouragement Design Keyphrases 100%
  • Failure to Treat Keyphrases 100%
  • Selective Attrition Keyphrases 100%
  • Black Box Psychology 100%
  • Field Experimentation Agricultural and Biological Sciences 100%

T1 - Field Experiments

AU - Ditlmann, Ruth

AU - Paluck, Elizabeth Levy

N1 - Publisher Copyright: © 2015 Elsevier Ltd. All rights reserved.

PY - 2015/3/26

Y1 - 2015/3/26

N2 - Field experiments are experiments in settings with high degrees of naturalism. This article describes different types of field experiments, including randomized field trials, randomized rollout designs, encouragement designs, downstream field experiments, hybrid lab-field experiments, and covert population experiments, and discusses their intellectual background and benefits. It also lists methodological challenges researchers can encounter when conducting field experiments, including failure to treat, selective attrition, spillover, difficulty of replication, and black box causality, and discusses available solutions. Finally, it provides an overview over current and emerging directions in field experimentation and concludes with a brief history of field experiments.

AB - Field experiments are experiments in settings with high degrees of naturalism. This article describes different types of field experiments, including randomized field trials, randomized rollout designs, encouragement designs, downstream field experiments, hybrid lab-field experiments, and covert population experiments, and discusses their intellectual background and benefits. It also lists methodological challenges researchers can encounter when conducting field experiments, including failure to treat, selective attrition, spillover, difficulty of replication, and black box causality, and discusses available solutions. Finally, it provides an overview over current and emerging directions in field experimentation and concludes with a brief history of field experiments.

KW - Attrition

KW - Audit studies

KW - Causality

KW - Correspondence studies

KW - Encouragement designs

KW - Experiments

KW - External validity

KW - Failure to treat

KW - Hybrid field experiments

KW - Naturalism

KW - Randomized controlled trial

KW - Randomized rollout designs

KW - Social experiments

KW - Stepped wedge designs

KW - Waiting list designs

UR - http://www.scopus.com/inward/record.url?scp=85043444790&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=85043444790&partnerID=8YFLogxK

U2 - 10.1016/B978-0-08-097086-8.10542-2

DO - 10.1016/B978-0-08-097086-8.10542-2

M3 - Chapter

AN - SCOPUS:85043444790

SN - 9780080970868

BT - International Encyclopedia of the Social & Behavioral Sciences: Second Edition

PB - Elsevier Inc.

design field experiments

  • Science & Math
  • Mathematics

Sorry, there was a problem.

Kindle app logo image

Download the free Kindle app and start reading Kindle books instantly on your smartphone, tablet, or computer - no Kindle device required .

Read instantly on your browser with Kindle for Web.

Using your mobile phone camera - scan the code below and download the Kindle app.

QR code to download the Kindle App

Image Unavailable

Field Experiments: Design, Analysis, and Interpretation

  • To view this video download Flash Player

Follow the authors

Donald P. Green

Field Experiments: Design, Analysis, and Interpretation Illustrated Edition

A brief, authoritative introduction to field experimentation in the social sciences.

  • ISBN-10 0393979954
  • ISBN-13 978-0393979954
  • Edition Illustrated
  • Publisher W. W. Norton & Company
  • Publication date May 29, 2012
  • Language English
  • Dimensions 6.1 x 1 x 9.3 inches
  • Print length 512 pages
  • See all details

Editorial Reviews

About the author, product details.

  • Publisher ‏ : ‎ W. W. Norton & Company; Illustrated edition (May 29, 2012)
  • Language ‏ : ‎ English
  • Paperback ‏ : ‎ 512 pages
  • ISBN-10 ‏ : ‎ 0393979954
  • ISBN-13 ‏ : ‎ 978-0393979954
  • Item Weight ‏ : ‎ 1.58 pounds
  • Dimensions ‏ : ‎ 6.1 x 1 x 9.3 inches
  • #16 in Political History (Books)
  • #264 in Probability & Statistics (Books)
  • #710 in History & Theory of Politics

About the authors

Donald p. green.

Donald P. Green is J.W. Burgess Professor of Political Science at Columbia University, having moved there in 2011 after 22 years at Yale University. The author of four books and more than one hundred essays, Green's research interests span a wide array of topics: voting behavior, partisanship, campaign finance, hate crime, and research methods. Much of his current work uses field experimentation to study the ways in which political campaigns mobilize and persuade voters. He was elected to the American Academy of Arts and Sciences in 2003 and was awarded the Heinz I. Eulau Award for best article published in the American Political Science Review during 2009. In 2010, he helped found the Experimental Research section of the American Political Science Association and served as its first president. Don's hobbies include woodworking and game design.

Alan S. Gerber

Discover more of the author’s books, see similar authors, read book recommendations and more.

Customer reviews

  • 5 star 4 star 3 star 2 star 1 star 5 star 83% 10% 6% 2% 0% 83%
  • 5 star 4 star 3 star 2 star 1 star 4 star 83% 10% 6% 2% 0% 10%
  • 5 star 4 star 3 star 2 star 1 star 3 star 83% 10% 6% 2% 0% 6%
  • 5 star 4 star 3 star 2 star 1 star 2 star 83% 10% 6% 2% 0% 2%
  • 5 star 4 star 3 star 2 star 1 star 1 star 83% 10% 6% 2% 0% 0%

Customer Reviews, including Product Star Ratings help customers to learn more about the product and decide whether it is the right product for them.

To calculate the overall star rating and percentage breakdown by star, we don’t use a simple average. Instead, our system considers things like how recent a review is and if the reviewer bought the item on Amazon. It also analyzed reviews to verify trustworthiness.

Reviews with images

Customer Image

Nice and clean

Customer Image

  • Sort reviews by Top reviews Most recent Top reviews

Top reviews from the United States

There was a problem filtering reviews right now. please try again later..

design field experiments

Top reviews from other countries

design field experiments

  • About Amazon
  • Investor Relations
  • Amazon Devices
  • Amazon Science
  • Sell products on Amazon
  • Sell on Amazon Business
  • Sell apps on Amazon
  • Become an Affiliate
  • Advertise Your Products
  • Self-Publish with Us
  • Host an Amazon Hub
  • › See More Make Money with Us
  • Amazon Business Card
  • Shop with Points
  • Reload Your Balance
  • Amazon Currency Converter
  • Amazon and COVID-19
  • Your Account
  • Your Orders
  • Shipping Rates & Policies
  • Returns & Replacements
  • Manage Your Content and Devices
 
 
 
 
  • Conditions of Use
  • Privacy Notice
  • Consumer Health Data Privacy Disclosure
  • Your Ads Privacy Choices

design field experiments

Experimental Method In Psychology

Saul McLeod, PhD

Editor-in-Chief for Simply Psychology

BSc (Hons) Psychology, MRes, PhD, University of Manchester

Saul McLeod, PhD., is a qualified psychology teacher with over 18 years of experience in further and higher education. He has been published in peer-reviewed journals, including the Journal of Clinical Psychology.

Learn about our Editorial Process

Olivia Guy-Evans, MSc

Associate Editor for Simply Psychology

BSc (Hons) Psychology, MSc Psychology of Education

Olivia Guy-Evans is a writer and associate editor for Simply Psychology. She has previously worked in healthcare and educational sectors.

On This Page:

The experimental method involves the manipulation of variables to establish cause-and-effect relationships. The key features are controlled methods and the random allocation of participants into controlled and experimental groups .

What is an Experiment?

An experiment is an investigation in which a hypothesis is scientifically tested. An independent variable (the cause) is manipulated in an experiment, and the dependent variable (the effect) is measured; any extraneous variables are controlled.

An advantage is that experiments should be objective. The researcher’s views and opinions should not affect a study’s results. This is good as it makes the data more valid  and less biased.

There are three types of experiments you need to know:

1. Lab Experiment

A laboratory experiment in psychology is a research method in which the experimenter manipulates one or more independent variables and measures the effects on the dependent variable under controlled conditions.

A laboratory experiment is conducted under highly controlled conditions (not necessarily a laboratory) where accurate measurements are possible.

The researcher uses a standardized procedure to determine where the experiment will take place, at what time, with which participants, and in what circumstances.

Participants are randomly allocated to each independent variable group.

Examples are Milgram’s experiment on obedience and  Loftus and Palmer’s car crash study .

  • Strength : It is easier to replicate (i.e., copy) a laboratory experiment. This is because a standardized procedure is used.
  • Strength : They allow for precise control of extraneous and independent variables. This allows a cause-and-effect relationship to be established.
  • Limitation : The artificiality of the setting may produce unnatural behavior that does not reflect real life, i.e., low ecological validity. This means it would not be possible to generalize the findings to a real-life setting.
  • Limitation : Demand characteristics or experimenter effects may bias the results and become confounding variables .

2. Field Experiment

A field experiment is a research method in psychology that takes place in a natural, real-world setting. It is similar to a laboratory experiment in that the experimenter manipulates one or more independent variables and measures the effects on the dependent variable.

However, in a field experiment, the participants are unaware they are being studied, and the experimenter has less control over the extraneous variables .

Field experiments are often used to study social phenomena, such as altruism, obedience, and persuasion. They are also used to test the effectiveness of interventions in real-world settings, such as educational programs and public health campaigns.

An example is Holfing’s hospital study on obedience .

  • Strength : behavior in a field experiment is more likely to reflect real life because of its natural setting, i.e., higher ecological validity than a lab experiment.
  • Strength : Demand characteristics are less likely to affect the results, as participants may not know they are being studied. This occurs when the study is covert.
  • Limitation : There is less control over extraneous variables that might bias the results. This makes it difficult for another researcher to replicate the study in exactly the same way.

3. Natural Experiment

A natural experiment in psychology is a research method in which the experimenter observes the effects of a naturally occurring event or situation on the dependent variable without manipulating any variables.

Natural experiments are conducted in the day (i.e., real life) environment of the participants, but here, the experimenter has no control over the independent variable as it occurs naturally in real life.

Natural experiments are often used to study psychological phenomena that would be difficult or unethical to study in a laboratory setting, such as the effects of natural disasters, policy changes, or social movements.

For example, Hodges and Tizard’s attachment research (1989) compared the long-term development of children who have been adopted, fostered, or returned to their mothers with a control group of children who had spent all their lives in their biological families.

Here is a fictional example of a natural experiment in psychology:

Researchers might compare academic achievement rates among students born before and after a major policy change that increased funding for education.

In this case, the independent variable is the timing of the policy change, and the dependent variable is academic achievement. The researchers would not be able to manipulate the independent variable, but they could observe its effects on the dependent variable.

  • Strength : behavior in a natural experiment is more likely to reflect real life because of its natural setting, i.e., very high ecological validity.
  • Strength : Demand characteristics are less likely to affect the results, as participants may not know they are being studied.
  • Strength : It can be used in situations in which it would be ethically unacceptable to manipulate the independent variable, e.g., researching stress .
  • Limitation : They may be more expensive and time-consuming than lab experiments.
  • Limitation : There is no control over extraneous variables that might bias the results. This makes it difficult for another researcher to replicate the study in exactly the same way.

Key Terminology

Ecological validity.

The degree to which an investigation represents real-life experiences.

Experimenter effects

These are the ways that the experimenter can accidentally influence the participant through their appearance or behavior.

Demand characteristics

The clues in an experiment lead the participants to think they know what the researcher is looking for (e.g., the experimenter’s body language).

Independent variable (IV)

The variable the experimenter manipulates (i.e., changes) is assumed to have a direct effect on the dependent variable.

Dependent variable (DV)

Variable the experimenter measures. This is the outcome (i.e., the result) of a study.

Extraneous variables (EV)

All variables which are not independent variables but could affect the results (DV) of the experiment. EVs should be controlled where possible.

Confounding variables

Variable(s) that have affected the results (DV), apart from the IV. A confounding variable could be an extraneous variable that has not been controlled.

Random Allocation

Randomly allocating participants to independent variable conditions means that all participants should have an equal chance of participating in each condition.

The principle of random allocation is to avoid bias in how the experiment is carried out and limit the effects of participant variables.

Order effects

Changes in participants’ performance due to their repeating the same or similar test more than once. Examples of order effects include:

(i) practice effect: an improvement in performance on a task due to repetition, for example, because of familiarity with the task;

(ii) fatigue effect: a decrease in performance of a task due to repetition, for example, because of boredom or tiredness.

Print Friendly, PDF & Email

  • Foundations
  • Write Paper

Search form

  • Experiments
  • Anthropology
  • Self-Esteem
  • Social Anxiety

design field experiments

Field Experiments

For geologists, social scientists and environmental biologists, amongst others, field experiments are an integral part of the discipline.

This article is a part of the guide:

  • Research Designs
  • Quantitative and Qualitative Research
  • Literature Review
  • Quantitative Research Design

Browse Full Outline

  • 1 Research Designs
  • 2.1 Pilot Study
  • 2.2 Quantitative Research Design
  • 2.3 Qualitative Research Design
  • 2.4 Quantitative and Qualitative Research
  • 3.1 Case Study
  • 3.2 Naturalistic Observation
  • 3.3 Survey Research Design
  • 3.4 Observational Study
  • 4.1 Case-Control Study
  • 4.2 Cohort Study
  • 4.3 Longitudinal Study
  • 4.4 Cross Sectional Study
  • 4.5 Correlational Study
  • 5.1 Field Experiments
  • 5.2 Quasi-Experimental Design
  • 5.3 Identical Twins Study
  • 6.1 Experimental Design
  • 6.2 True Experimental Design
  • 6.3 Double Blind Experiment
  • 6.4 Factorial Design
  • 7.1 Literature Review
  • 7.2 Systematic Reviews
  • 7.3 Meta Analysis

As the name suggests, a field study is an experiment performed outside the laboratory, in the 'real' world. Unlike case studies and observational studies , a field experiment still follows all of the steps of the scientific process , addressing research problems and generating hypotheses.

The obvious advantage of a field study is that it is practical and also allows experimentation , without artificially introducing confounding variables .

A population biologist examining an ecosystem could not move the entire environment into the laboratory, so field experiments are the only realistic research method in many fields of science.

In addition, they circumvent the accusation leveled at laboratory experiments of lacking external or ecological validity , or adversely affecting the behavior of the subject.

Social scientists and psychologists often used field experiments to perform blind studies , where the subject was not even aware that they were under scrutiny.

A good example of this is the Piliavin and Piliavin experiment , where the propensity of strangers to help blood covered 'victims' was measured. This is now frowned upon, under the policy of informed consent , and is only used in rare and highly regulated circumstances.

Field experiments can suffer from a lack of a discrete control group and often have many variables to try to eliminate.

For example, if the effects of a medicine are studied, and the subject is instructed not to drink alcohol, there is no guarantee that the subject followed the instructions, so field studies often sacrifice internal validity for external validity .

For fields like biology, geology and environmental science, this is not a problem, and the field experiment can be treated as a sound experimental practice, following the steps of the scientific method .

A major concern shared by all disciplines is the cost of field studies, as they tend to be very expensive.

For example, even a modestly sized research ship costs many thousands of dollars every day, so a long oceanographical research program can run into the millions of dollars.

Pilot studies are often used to test the feasibility of any long term or extensive research program before committing vast amounts of funds and resources. The changeable nature of the external environment and the often-prohibitive investment of time and money mean that field experiments are rarely replicable , so any generalization is always tenuous.

  • Psychology 101
  • Flags and Countries
  • Capitals and Countries

Martyn Shuttleworth (Jun 14, 2010). Field Experiments. Retrieved Sep 26, 2024 from Explorable.com: https://explorable.com/field-experiments

You Are Allowed To Copy The Text

The text in this article is licensed under the Creative Commons-License Attribution 4.0 International (CC BY 4.0) .

This means you're free to copy, share and adapt any parts (or all) of the text in the article, as long as you give appropriate credit and provide a link/reference to this page.

That is it. You don't need our permission to copy the article; just include a link/reference back to this page. You can use it freely (with some kind of link), and we're also okay with people reprinting in publications like books, blogs, newsletters, course-material, papers, wikipedia and presentations (with clear attribution).

design field experiments

Want to stay up to date? Follow us!

Get all these articles in 1 guide.

Want the full version to study at home, take to school or just scribble on?

Whether you are an academic novice, or you simply want to brush up your skills, this book will take your academic writing skills to the next level.

design field experiments

Download electronic versions: - Epub for mobiles and tablets - PDF version here

Save this course for later

Don't have time for it all now? No problem, save it as a course and come back to it later.

Footer bottom

  • Privacy Policy

design field experiments

  • Subscribe to our RSS Feed
  • Like us on Facebook
  • Follow us on Twitter
  • A-Z Publications

Annual Review of Sociology

Volume 43, 2017, review article, field experiments across the social sciences.

  • Delia Baldassarri 1 , and Maria Abascal 2
  • View Affiliations Hide Affiliations Affiliations: 1 Department of Sociology, New York University, New York, New York 10012; email: [email protected] 2 Department of Sociology, Columbia University, New York, New York 10027; email: [email protected]
  • Vol. 43:41-73 (Volume publication date July 2017) https://doi.org/10.1146/annurev-soc-073014-112445
  • First published as a Review in Advance on May 22, 2017
  • © Annual Reviews

Using field experiments, scholars can identify causal effects via randomization while studying people and groups in their naturally occurring contexts. In light of renewed interest in field experimental methods, this review covers a wide range of field experiments from across the social sciences, with an eye to those that adopt virtuous practices, including unobtrusive measurement, naturalistic interventions, attention to realistic outcomes and consequential behaviors, and application to diverse samples and settings. The review covers four broad research areas of substantive and policy interest: first, randomized controlled trials, with a focus on policy interventions in economic development, poverty reduction, and education; second, experiments on the role that norms, motivations, and incentives play in shaping behavior; third, experiments on political mobilization, social influence, and institutional effects; and fourth, experiments on prejudice and discrimination. We discuss methodological issues concerning generalizability and scalability as well as ethical issues related to field experimental methods. We conclude by arguing that field experiments are well equipped to advance the kind of middle-range theorizing that sociologists value.

Article metrics loading...

Full text loading...

Literature Cited

  • Abascal M . 2015 . Us and them: black–white relations in the wake of Hispanic population growth. Am. Sociol. Rev. 80 : 789– 813 [Google Scholar]
  • Adida CL , Laitin DD , Valfort MA . 2016 . Why Muslim Integration Fails in Christian-Heritage Societies Cambridge, MA: Harvard Univ. Press [Google Scholar]
  • Ahmed AM , Hammarstedt M . 2008 . Discrimination in the rental housing market: a field experiment on the Internet. J. Urban Econ. 64 : 362– 72 [Google Scholar]
  • Ahmed AM , Hammarstedt M . 2009 . Detecting discrimination against homosexuals: evidence from a field experiment on the Internet. Economica 76 : 599– 97 [Google Scholar]
  • Arceneaux K , Nickerson DW . 2009 . Who is mobilized to vote? A re-analysis of 11 field experiments. Am. J. Political Sci. 53 : 1– 16 [Google Scholar]
  • Attanasio O , Augsburg B , De Haas R , Fitzsimons E , Harmgart H . 2012 . Group lending or individual lending? Evidence from a randomised field experiment in Mongolia. Work. Pap. No. 136, Eur. Bank Reconstr. Dev. [Google Scholar]
  • Attanasio O , Pellerano L , Reyes SP . 2009 . Building trust? Conditional cash transfer programmes and social capital. Fiscal Stud. 30 : 139– 77 [Google Scholar]
  • Avdeenko A , Gilligan MG . 2015 . International interventions to build social capital: evidence from a field experiment in Sudan. Am. Political Sci. Rev. 109 : 427– 49 [Google Scholar]
  • Ayres I , Siegelman P . 1995 . Race and gender discrimination in bargaining for a new car. Am. Econ. Rev. 85 : 304– 21 [Google Scholar]
  • Baldassarri D . 2015 . Cooperative networks: altruism, group solidarity, and reciprocity in Ugandan farmer organizations. Am. J. Sociol. 121 : 355– 95 [Google Scholar]
  • Baldassarri D . 2016 . Prosocial behavior across communities: evidence from a nationwide lost-letter experiment Presented at Advances with Field Experiments Conf., Sept. 16, Univ Chicago: [Google Scholar]
  • Banerjee A , Bertrand M , Datta S , Mullainathan S . 2009 . Labor market discrimination in Delhi: evidence from a field experiment. J. Comp. Econ. 37 : 14– 27 [Google Scholar]
  • Banerjee A , Duflo E . 2009 . The experimental approach to development economics. Annu. Rev. Econ. 1 : 151– 78 [Google Scholar]
  • Banerjee A , Duflo E . 2011 . Poor Economics: A Radical Rethinking of the Way to Fight Global Poverty. New York: Public Affairs [Google Scholar]
  • Banerjee A , Duflo E , Glennerster R , Kothari D . 2010a . Improving immunization coverage in rural India: Clustered randomized controlled immunisation campaigns with and without incentives. Br. Med. J. 340:c2220 [Google Scholar]
  • Banerjee A , Duflo E , Glennerster R , Kinnan C . 2010b . The miracle of microfinance? Evidence from a randomized evaluation. Work. Pap. No. 13-09, Dep. Econ., MIT [Google Scholar]
  • Barr A . 2003 . Trust and expected trustworthiness: experimental evidence from Zimbabwean villages. Econ. J. 113 : 614– 30 [Google Scholar]
  • Bauchet J , Marshall C , Starita L , Thomas J , Yalouris A . 2011 . Latest findings from randomized evaluations of microfinance. Access Finance Forum Rep. 2 : 1– 27 [Google Scholar]
  • Beath A , Christia F , Enikolopov R . 2013 . Empowering women: evidence from a field experiment in Afghanistan. Am. Political Sci. Rev. 107 : 540– 57 [Google Scholar]
  • Benson PL , Karabenick SA , Lerner RM . 1976 . Pretty pleases: the effects of physical attractiveness, race, and sex on receiving help. J. Exp. Soc. Psychol. 12 : 409– 15 [Google Scholar]
  • Benz M , Meier S . 2008 . Do people behave in experiments as in the field? Evidence from donations. Exp. Econ. 11 : 278– 81 [Google Scholar]
  • Bertrand M , Karlan D , Mullainathan S , Shafir E , Zinman J . 2010 . What's advertising content worth? Evidence from a consumer credit marketing field experiment. Q. J. Econ. 125 : 263– 306 [Google Scholar]
  • Bertrand M , Mullainathan S . 2004 . Are Emily and Greg more employable than Lakisha and Jamal? A field experiment on labor market discrimination. Am. Econ. Rev. 94 : 991– 1013 [Google Scholar]
  • Besbris M , Faber JW , Rich P , Sharkey P . 2015 . Effect of neighborhood stigma on economic transitions. PNAS 112 : 4994– 98 [Google Scholar]
  • Bettinger EP . 2012 . Paying to learn: the effect of financial incentives on elementary school test scores. Rev. Econ. Stat. 94 : 686– 98 [Google Scholar]
  • Bigoni M , Bortolotti S , Casari M , Gambetta D , Pancotto F . 2016 . Amoral familism, social capital, or trust? The behavioural foundations of the Italian north–south divide. Econ. J. 126 : 1318– 41 [Google Scholar]
  • Blommaert L , Coenders M , van Tubergen F . 2014 . Discrimination of Arabic-named applicants in the Netherlands: an Internet-based field experiment examining different phases in online recruitment procedures. Soc. Forces 92 : 957– 82 [Google Scholar]
  • Bond RM , Fariss CJ , Jones JJ , Kramer AD , Marlow C . et al. 2012 . A 61-million-person experiment in social influence and political mobilization. Nature 489 : 295– 98 [Google Scholar]
  • Bosch M , Carnero MA , Farré L . 2010 . Information and discrimination in the rental housing market: evidence from a field experiment. Reg. Sci. Urban Econ. 40 : 11– 19 [Google Scholar]
  • Brearley HC . 1931 . Experimental sociology in the United States. Soc. Forces 10 : 196– 99 [Google Scholar]
  • Butler DM , Broockman DE . 2011 . Do politicians racially discriminate against constituents? A field experiment on state legislators. Am. J. Political Sci. 55 : 463– 77 [Google Scholar]
  • Butler DM , Nickerson DW . 2011 . Can learning constituency opinion affect how legislators vote? Results from a field experiment. Q. J. Political Sci. 6 : 55– 83 [Google Scholar]
  • Camerer C . 2003 . Behavioral Game Theory: Experiments in Strategic Interaction New York, NY: Russell Sage Found. [Google Scholar]
  • Cardenas J , Carpenter J . 2008 . Behavioural development economics: lessons from field labs in the developing world. J. Dev. Stud. 44 : 337– 64 [Google Scholar]
  • Casey K , Glennerster R , Miguel E . 2012 . Reshaping institutions: evidence on external aid and local collective action. Q. J. Econ. 127 : 1755– 812 [Google Scholar]
  • Castilla EJ , Benard S . 2010 . The paradox of meritocracy in organizations. Adm. Sci. Q. 55 : 543– 676 [Google Scholar]
  • Centola D . 2010 . The spread of behavior in an online social network experiment. Science 329 : 1194– 97 [Google Scholar]
  • Charness G , Gneezy U . 2009 . Incentives to exercise. Econometrica 77 : 909– 31 [Google Scholar]
  • Chetty R , Hendren N , Katz LF . 2015 . The effects of exposure to better neighborhoods on children: new evidence from the moving to opportunity experiment. Work. Pap. 21156, NBER, Cambridge, MA [Google Scholar]
  • Chong D , Junn J . 2011 . Politics from the perspective of minority populations. Cambridge Handbook of Experimental Political Science JN Druckman, DP Green, JH Kuklinski, A Lupia, 602– 33 Cambridge, UK: Cambridge Univ. Press [Google Scholar]
  • Cialdini RB , Ascani K . 1976 . Test of a concession procedure for inducing verbal, behavioral, and further compliance with a request to give blood. J. Pers. Soc. Psychol. 61 : 295– 300 [Google Scholar]
  • Cialdini RB , Vincent JE , Lewis SK , Catalan J , Wheeler D , Darby BL . 1975 . Reciprocal concessions procedure for inducing compliance: the door-in-the-face technique. J. Pers. Soc. Psychol. 31 : 206– 15 [Google Scholar]
  • Clampet-Lundquist S , Massey DS . 2008 . Neighborhood effects on economic self-sufficiency: a reconsideration of the Moving to Opportunity experiment. Am. J. Sociol. 114 : 107– 43 [Google Scholar]
  • Cohen J , Dupas P . 2010 . Free distribution or cost-sharing? Evidence from a randomized malaria prevention experiment. Q. J. Econ. 125 : 1– 40 [Google Scholar]
  • Cole S , Giné X , Tobacman J , Topalova P , Townsend R , Vickery J . 2013 . Barriers to household risk management: evidence from India. Am. Econ. J. Appl. Econ. 5 : 104– 35 [Google Scholar]
  • Cook TD , Shadish WR . 1994 . Social experiments: some developments over the past fifteen years. Annu. Rev. Psychol. 45 : 545– 80 [Google Scholar]
  • Correll SJ , Benard S , Paik I . 2007 . Getting a job: is there a motherhood penalty?. Am. J. Sociol. 112 : 1297– 339 [Google Scholar]
  • Cox D . 1958 . Planning of Experiments New York: Wiley [Google Scholar]
  • Crépon B , Devoto F , Duflo E , Parienté W . 2011 . Impact of microcredit in rural areas of Morocco: evidence from a randomized evaluation. Work. Pap., Dep. Econ., MIT [Google Scholar]
  • Cross H , Kenney GM , Mell J , Zimmerman W . 1990 . Employer hiring practices: differential treatment of Hispanic and Anglo job seekers. Tech. rep., Urban Inst., Washington, DC [Google Scholar]
  • Deaton A . 2010 . Instruments, randomization, and learning about development. J. Econ. Lit. 48 : 424– 55 [Google Scholar]
  • Dehejia R , Pop-Eleches C , Samii C . 2015 . From local to global: external validity in a fertility natural experiment. Work. Pap. 21459, NBER, Cambridge, MA [Google Scholar]
  • Doob AN , Gross AE . 1968 . Status as an inhibitor of horn-honking responses. J. Soc. Psychol. 76 : 213– 18 [Google Scholar]
  • Druckman JN , Green DP , Kuklinski JH , Lupia A . 2011 . Cambridge Handbook of Experimental Political Science Cambridge, UK: Cambridge Univ. Press [Google Scholar]
  • Duflo E , Kremer M , Robinson J . 2008 . How high are rates of return to fertilizer? Evidence from field experiments in Kenya. Am. Econ. Rev. 98 : 482– 88 [Google Scholar]
  • Duflo E , Kremer M , Robinson J . 2011 . Nudging farmers to use fertilizer: theory and experimental evidence from Kenya. Am. Econ. Rev. 101 : 2350– 90 [Google Scholar]
  • Dunn EW , Aknin LB , Norton MI . 2008 . Spending money on others promotes happiness. Science 319 : 1687– 88 [Google Scholar]
  • Dunning T . 2012 . Natural Experiments in the Social Sciences: A Design-Based Approach Cambridge, UK: Cambridge Univ. Press [Google Scholar]
  • Dupas P . 2009 . What matters (and what does not) in households’ decision to invest in malaria prevention?. Am. Econ. Rev. 99 : 224– 30 [Google Scholar]
  • Dupas P . 2011 . Do teenagers respond to HIV risk information? Evidence from a field experiment in Kenya. Am. Econ. J. Appl. Econ. 3 : 1– 34 [Google Scholar]
  • Dupas P . 2014 . Short-run subsidies and long-run adoption of new health products: evidence from a field experiment. Econometrica 82 : 197– 228 [Google Scholar]
  • Dupas P , Robinson J . 2011 . Savings constraints and microenterprise development: evidence from a field experiment in Kenya. Work. Pap. 14693, NBER, Cambridge, MA [Google Scholar]
  • Emswiller T , Deaux K , Willits JE . 1971 . Similarity, sex, and requests for small favors. J. Appl. Soc. Psychol. 1 : 284– 91 [Google Scholar]
  • Enos RD . 2014 . Causal effect of intergroup contact on exclusionary attitudes. PNAS 111 : 3699– 704 [Google Scholar]
  • Enos RD , Fowler A , Vavreck L . 2014 . Increasing inequality: the effect of GOTV mobilization on the composition of the electorate. J. Polit. 76 : 273– 88 [Google Scholar]
  • Fearon JD , Humphreys M , Weinstein JM . 2009 . Can development aid contribute to social cohesion after civil war? Evidence from a field experiment in post-conflict Liberia. Am. Econ. Rev. 99 : 287– 91 [Google Scholar]
  • Fearon JD , Humphreys M , Weinstein JM . 2015 . How does development assistance affect collective action capacity? Results from a field experiment in post-conflict Liberia. Am. J. Political Sci. 109 : 450– 69 [Google Scholar]
  • Fershtman C , Gneezy U . 2001 . Discrimination in a segmented society: an experimental approach. Q. J. Econ. 116 : 351– 77 [Google Scholar]
  • Fisher RA . 1935 . The Design of Experiments New York: Hafner [Google Scholar]
  • Fiszbein A , Schady N . 2009 . Conditional cash transfers: reducing present and future poverty. World Bank Policy Res. Rep., World Bank Washington, DC: [Google Scholar]
  • Forbes GB , Gromoll HF . 1971 . The lost letter technique as a measure of social variables: some exploratory findings. Soc. Forces 50 : 113– 15 [Google Scholar]
  • Freedman JL , Fraser SC . 1966 . Compliance without pressure: the foot-in-the-door technique. J. Pers. Soc. Psychol. 4 : 195– 202 [Google Scholar]
  • Freese J , Peterson D . 2017 . Replication in social science. Annu. Rev. Sociol. 43. In press [Google Scholar]
  • Fryer R . 2011 . Financial incentives and student achievement: evidence from randomized trials. Q. J. Econ. 126 : 1755– 98 [Google Scholar]
  • Gaddis SM . 2015 . Discrimination in the credential society: an audit study of race and college selectivity in the labor market. Soc. Forces 93 : 1451– 79 [Google Scholar]
  • Gaddis SM , Ghoshal R . 2015 . Arab American housing discrimination, ethnic competition, and the contact hypothesis. Ann. Am. Acad. Political Soc. Sci. 660 : 282– 99 [Google Scholar]
  • Galster G , Constantine P . 1991 . Discrimination against female-headed households in rental housing: theory and exploratory evidence. Rev. Soc. Econ. 49 : 76– 100 [Google Scholar]
  • Gantner L . 2007 . PROGRESA: An integrated approach to poverty alleviation in Mexico. Case Studies in Food Policy for Developing Countries: Policies for Health, Nutrition, Food Consumption, and Poverty P Pinstrup-Andersen, F Cheng, Vol 1 211– 20 Ithaca, NY: Cornell Univ. Press [Google Scholar]
  • Garfinkel H . 1967 . Studies in Ethnomethodology Englewood Cliffs, NJ: Prentice-Hall [Google Scholar]
  • Gelman A . 2014 . Experimental reasoning in social science. Field Experiments and Their Critics: Essays on the Uses and Abuses of Experimentation in the Social Sciences DL Teele 185– 95 New Haven, CT: Yale Univ. Press [Google Scholar]
  • Gerber AS . 2011 . Field experiments in political science. Cambridge Handbook of Experimental Political Science JN Druckman, DP Green, JH Kuklinski, A Lupia 115– 38 Cambridge, UK: Cambridge Univ. Press [Google Scholar]
  • Gerber AS , Green DP . 2000 . The effects of canvassing, telephone calls, and direct mail on voter turnout: a field experiment. Am. Political Sci. Rev. 94 : 653– 63 [Google Scholar]
  • Gerber AS , Green DP . 2012 . Field Experiments New York: Norton [Google Scholar]
  • Gerber AS , Green DP , Larimer CW . 2008 . Social pressure and voter turnout: evidence from a large scale field experiment. Am. Political Sci. Rev. 102 : 33– 48 [Google Scholar]
  • Gerber AS , Green DP , Shachar R . 2003 . Voting may be habit-forming: evidence from a randomized field experiment. Am. J. Political Sci. 47 : 540– 50 [Google Scholar]
  • Gil-White F . 2004 . Ultimatum game with an ethnicity manipulation: results from Kohvdiin Bulgan Sum, Mongolia. Foundations of Human Sociality: Economic Experiments and Ethnographic Evidence from Fifteen Small-Scale Societies J Henrich, R Boyd, S Bowles, C Camerer, E Fehr, H Gintis, 260– 304 Oxford, UK: Oxford Univ. Press [Google Scholar]
  • Gilligan MJ , Pasquale BJ , Samii C . 2014 . Civil war and social cohesion: lab-in-the-field evidence from Nepal. Am. J. Political Sci. 58 : 604– 19 [Google Scholar]
  • Giné X , Karlan D . 2014 . Group versus individual liability: short and long term evidence from Philippine-microcredit lending groups. J. Dev. Econ. 107 : 65– 83 [Google Scholar]
  • Giné X , Karlan D , Zinman J . 2010 . Put your money where your butt is: a commitment contract for smoking cessation. Am. Econ. J. Appl. Econ. 213– 35 [Google Scholar]
  • Gneezy U , List J , Price MK . 2012 . Toward an understanding of why people discriminate: evidence from a series of natural field experiments. Work. Pap. 17855, NBER, Cambridge, MA [Google Scholar]
  • Gneezy U , Meier S , Rey-Biel P . 2011 . When and why incentives (don't) work to modify behavior. J. Econ. Perspect. 25 : 191– 210 [Google Scholar]
  • Gneezy U , Rey-Biel P . 2014 . On the relative efficiency of performance pay and noncontingent incentives. J. Eur. Econ. Assoc. 12 : 62– 72 [Google Scholar]
  • Gneezy U , Rustichini A . 2000 . A fine is a price. J. Legal Stud. 29 : 1– 17 [Google Scholar]
  • Goel V . 2014 . Facebook tinkers with users’ emotions in news feed experiment, stirring outcry. New York Times , June 30 B1
  • Gosnell HF . 1927 . Getting Out the Vote: An Experiment in the Stimulation of Voting Chicago: Chicago Univ. Press [Google Scholar]
  • Green DP , Gerber A . 2008 . Get Out the Vote: How to Increase Voter Turnout Washington, DC: Brookings Inst. Press. 2nd ed. [Google Scholar]
  • Green DP , Wong J . 2009 . Tolerance and the contact hypothesis: a field experiment. The Political Psychology of Democratic Citizenship 228– 46 Oxford, UK: Oxford Univ. Press [Google Scholar]
  • Greenberg D , Shroder M . 2004 . The Digest of Social Experiments. Washington, DC: Urban Inst. Press [Google Scholar]
  • Grose CR . 2014 . Field experimental work on political institutions. Annu. Rev. Political Sci. 17 : 355– 70 [Google Scholar]
  • Grossman G , Baldassarri D . 2012 . The impact of elections on cooperation: evidence from a lab in the field experiment in Uganda. Am. J. Political Sci. 56 : 964– 85 [Google Scholar]
  • Grossman G , Paler L . 2015 . Using experiments to study political institutions. Handbook of Comparative Political Institutions J Gandhi, R Ruiz-Rufino 84– 97 London: Routledge [Google Scholar]
  • Habyarimana J , Humphreys M , Posner DN , Weinstein JM . 2009 . Coethnicity: Diversity and the Dilemmas of Collective Action New York: Russell Sage Found. [Google Scholar]
  • Harrison GW . 2013 . Field experiments and methodological intolerance. J. Econ. Methodol. 20 : 103– 17 [Google Scholar]
  • Harrison GW , List JA . 2004 . Field experiments. J. Econ. Lit. 42 : 1009– 55 [Google Scholar]
  • Hausman JA , Wise DA . 1985 . Social Experimentation Chicago: Chicago Univ. Press [Google Scholar]
  • Heckman JJ . 1992 . Randomization and social policy evaluation. Evaluating Welfare and Training Programs CF Manski, I Garfinkel 201– 30 Cambridge, MA: Harvard Univ. Press [Google Scholar]
  • Heckman JJ . 1998 . Detecting discrimination. J. Econ. Perspect. 12 : 101– 16 [Google Scholar]
  • Heckman JJ , Siegelman P . 1993 . The Urban Institute audit studies: their methods and findings. Clear and Convincing Evidence: Measurement of Discrimination in America M Fix, RJ Struyk 187– 258 Washington, DC: Urban Inst. Press [Google Scholar]
  • Henrich J , Boyd R , Bowles S , Camerer C , Fehr E . et al. 2001 . In search of homo economicus: behavioral experiments in 15 small-scale societies. Am. Econ. Rev. 91 : 73– 78 [Google Scholar]
  • Henrich J , Ensminger J , McElreath R , Barr A , Barrett C . et al. 2010 . Markets, religion, community size, and the evolution of fairness and punishment. Science 327 : 1480– 84 [Google Scholar]
  • Henrich J , McElreath R , Barr A , Ensminger J , Barrett C . et al. 2006 . Costly punishment across human societies. Science 312 : 1767– 70 [Google Scholar]
  • Henry PJ . 2008 . College sophomores in the laboratory redux: influences of a narrow data base on social psychology's view of the nature of prejudice. Psychol. Inq. 19 : 49– 71 [Google Scholar]
  • Herberich DH , List JA , Price MK . 2011 . How many economists does it take to change a light bulb? A natural field experiment on technology adoption Work. Pap., Univ. Chicago [Google Scholar]
  • Heyman J , Ariely D . 2004 . Effort for payment: a tale of two markets. Psychol. Sci. 15 : 787– 93 [Google Scholar]
  • Holland J , Silva AS , Mace R . 2012 . Lost letter measure of variation in altruistic behaviour in 20 neighbourhoods. PLOS ONE 7 : e43294 [Google Scholar]
  • Houlette MA , Gaertner SL , Johnson KM , Banker BS , Riek BM , Dovidio JF . 2004 . Developing a more inclusive social identity: an elementary school intervention. J. Soc. Issues 60 : 35– 55 [Google Scholar]
  • Humphreys M , Sanchez de la Sierra R , van der Windt P . 2013 . Fishing, commitment, and communication: a proposal for comprehensive nonbinding research registration. Polit. Anal. 21 : 1– 20 [Google Scholar]
  • Imbens G , Wooldridge J . 2009 . Recent developments in the econometrics of program evaluation. J. Econ. Lit. 47 : 5– 86 [Google Scholar]
  • Isen AM , Levin PF . 1972 . Effect of feeling good on helping: cookies and kindness. J. Pers. Soc. Psychol. 21 : 384– 88 [Google Scholar]
  • Jackson M , Cox DR . 2013 . The principles of experimental design and their application in sociology. Annu. Rev. Sociol. 39 : 27– 49 [Google Scholar]
  • Jensen R , Miller N . 2008 . Giffen behavior and subsistence consumption. Am. Econ. Rev. 98 : 1553– 77 [Google Scholar]
  • Kamenica E . 2012 . Behavioral economics and psychology of incentives. Annu. Rev. Econ. 4 : 427– 52 [Google Scholar]
  • Karlan D . 2005 . Using experimental economics to measure social capital and predict financial decisions. Am. Econ. Rev. 95 : 1688– 99 [Google Scholar]
  • Karlan D , Appel J . 2011 . More Than Good Intentions: Improving the Ways the World's Poor Borrow, Save, Farm, Learn, and Stay Healthy New York: Penguin [Google Scholar]
  • Karlan D , Goldberg N . 2011 . Microfinance evaluation strategies: notes on methodology and findings. The Handbook of Microfinance B Armendáriz, M Labie 17– 58 London: World Scientific [Google Scholar]
  • Karlan D , McConnell M , Mullainathan S , Zinman J . 2014 . Getting to the top of mind: how reminders increase saving. Manag. Sci. 62 : 3393– 3411 [Google Scholar]
  • Karlan D , Osei-Akoto I , Osei R , Udry C . 2010 . Examining underinvestment in agriculture: measuring returns to capital and insurance. Work. Pap., Abdul Latif Jameel Poverty Action Lab. https://www.poverty-action.org/sites/default/files/Panel3-3-Farmers-Returns-Capital.pdf [Google Scholar]
  • Karlan D , Zinman J . 2011 . Microcredit in theory and practice: using randomized credit scoring for impact. Science 332 : 1278– 84 [Google Scholar]
  • Keizer K , Lindenberg S , Steg L . 2008 . The spreading of disorder. Science 322 : 1681– 85 [Google Scholar]
  • Kelly E , Moena P , Oakes J , Fan W , Okechukwu C . et al. 2014 . Changing work and work-family conflict: evidence from the work, family, and health network. Am. Sociol. Rev. 79 : 485– 516 [Google Scholar]
  • Kling JR , Liebman JB , Katz LF . 2007 . Experimental analysis of neighborhood effects. Econometrica 75 : 83– 119 [Google Scholar]
  • Kotran A . 2015 . Opower and utility partners save over eight terawatt-hours of energy power and utility partners save over eight terawatt-hours of energy. News release, May 21
  • Kramer ADI , Guillory JE , Hancock JT . 2014 . Experimental evidence of massive-scale emotional contagion through social networks. PNAS 111 : 8788– 90 [Google Scholar]
  • Kremer M . 2003 . Randomized evaluations of educational programs in developing countries: some lessons. Am. Econ. Rev. 93 : 102– 6 [Google Scholar]
  • Kremer M , Brannen C , Glennerster R . 2013 . The challenge of education and learning in the developing world. Science 340 : 297– 300 [Google Scholar]
  • Kremer M , Leino J , Miguel E , Zwane AP . 2011 . Spring cleaning: rural water impacts, valuation, and property rights institutions. Q. J. Econ. 126 : 145– 205 [Google Scholar]
  • Kugelmass H . 2016 . “Sorry, I'm not accepting new patients”: an audit study of access to mental health care. J. Health Soc. Behav. 57 : 168– 83 [Google Scholar]
  • Lacetera N , Macis M . 2010 . Do all material incentives for pro-social activities backfire? The response to cash and non-cash incentives for blood donations. J. Econ. Psychol. 31 : 738– 48 [Google Scholar]
  • Lacetera N , Macis M , Slonim R . 2013 . Economic rewards to motivate blood donations. Science 340 : 927– 28 [Google Scholar]
  • Landry CE , Lange A , List JA , Price MK , Rupp NG . 2010 . Is a donor in hand better than two in the bush? Evidence from a natural field experiment. Am. Econ. Rev. 100 : 958– 83 [Google Scholar]
  • Langer EJ , Rodin J . 1976 . The effects of choice and enhanced responsibility for the aged: a field experiment in an institutional setting. J. Pers. Soc. Psychol. 34 : 191– 98 [Google Scholar]
  • Lauster N , Easterbrook A . 2011 . No room for new families? A field experiment measuring rental discrimination against same-sex couples and single parents. Soc. Probl. 58 : 389– 409 [Google Scholar]
  • Leuven E , Oosterbeek H , van der Klaauw B . 2010 . The effect of financial rewards on students’ achievement: evidence from a randomized experiment. J. Eur. Econ. Assoc. 8 : 1243– 65 [Google Scholar]
  • Levine M , Prosser A , Evans D , Reicher S . 2005 . Identity and emergency intervention: how social group membership and inclusiveness of group boundaries shape helping behavior. Pers. Soc. Psychol. Bull. 31 : 443– 53 [Google Scholar]
  • Levitt SD , List JA . 2009 . Field experiments in economics: the past, the present, and the future. Eur. Econ. Rev. 53 : 1– 18 [Google Scholar]
  • Levitt SD , List JA , Neckerman S , Sadoff S . 2012 . The behavioralist goes to school: leveraging behavioral economics to improve educational performance. Work. Pap. 18165, NBER Cambridge, MA: [Google Scholar]
  • List JA . 2007 . Field experiments: a bridge between lab and naturally occurring data. B.E. J. Econ. Anal. Policy 5 : 2 [Google Scholar]
  • Lucas JW . 2003 . Theory-testing, generalization, and the problem of external validity. Sociol. Theory 21 : 236– 53 [Google Scholar]
  • Ludwig J , Duncan GJ , Gennetian LA , Katz LF , Kessler RC . et al. 2013 . Long-term neighborhood effects on low-income families: evidence from moving to opportunity. Am. Econ. Rev. 103 : 226– 31 [Google Scholar]
  • Ludwig J , Liebman JB , Kling JR , Duncan GJ , Katz LF . et al. 2008 . What can we learn about neighborhood effects from the moving to opportunity experiment?. Am. J. Sociol. 114 : 144– 88 [Google Scholar]
  • Marwell G , Ames RE . 1979 . Experiments on the provision of public goods: resources, interest, group size, and the free-rider problem. Am. J. Sociol. 84 : 1335– 60 [Google Scholar]
  • Massey DS , Lundy G . 2001 . Use of Black English and racial discrimination in urban housing markets: new methods and findings. Urban Aff. Rev. 36 : 452– 69 [Google Scholar]
  • McDermott R . 2011 . Internal and external validity. Cambridge Handbook of Experimental Political Science JN Druckman, DP Green, JH Kuklinski, A Lupia, 27– 40 Cambridge, UK: Cambridge Univ. Press [Google Scholar]
  • McEwan PJ . 2015 . Improving learning in primary schools of developing countries: a meta-analysis of randomized experiments. Rev. Educ. Res. 85 : 353– 94 [Google Scholar]
  • McNutt M . 2015 . Editorial retraction of Lacour & Green. Science 346 : 1366– 69 Science 348 : 1100 [Google Scholar]
  • Merton RK . 1945 . Sociological theory. Am. J. Sociol. 50 : 462– 73 [Google Scholar]
  • Michelson M , Nickerson DW . 2011 . Voter Mobilization Cambridge, UK: Cambridge Univ. Press [Google Scholar]
  • Miguel E , Kremer M . 2004 . Worms: identifying impacts on education and health in the presence of treatment externalities. Econometrica 72 : 159– 217 [Google Scholar]
  • Milgram S , Liberty HJ , Toledo R , Wackenhut J . 1986 . Response to intrusion into waiting lines. J. Pers. Soc. Psychol. 51 : 683– 89 [Google Scholar]
  • Milgram S , Mann L , Hartner S . 1965 . The lost letter technique: a tool of social research. Public Opin. Q. 29 : 437– 38 [Google Scholar]
  • Milkman KL , Akinola M , Chugh D . 2015 . What happens before? A field experiment exploring how pay and representation differentially shape bias on the pathway into organizations. J. Appl. Psychol. 100 : 1678– 712 [Google Scholar]
  • Milkman KL , Beshears J , Choi JJ , Laibson D , Madrian BC . 2011 . Using implementation intentions prompts to enhance influenza vaccination rates. PNAS 108 : 10415– 20 [Google Scholar]
  • Morgan S , Winship C . 2007 . Counterfactuals and Causal Inference Cambridge, UK: Cambridge Univ. Press [Google Scholar]
  • Morton R , Williams K . 2010 . Experimental Political Science and the Study of Causality Cambridge, UK: Cambridge Univ. Press [Google Scholar]
  • Moss-Racusin CA , Dovidio JF , Brescoll V , Graham MJ , Handelsman J . 2012 . Science faculty's subtle gender biases favor male students. PNAS 109 : 16474– 79 [Google Scholar]
  • Munnell AH . 1986 . Lessons from the Income Maintenance Experiments Boston: Fed. Res. Bank of Boston [Google Scholar]
  • Mutz DC . 2011 . Population-Based Survey Experiments Princeton, NJ: Princeton Univ. Press [Google Scholar]
  • Nagda BRA , Tropp LR , Paluck EL . 2006 . Looking back as we look ahead: integrating research, theory, and practice on intergroup relations. J. Soc. Issues 62 : 439– 51 [Google Scholar]
  • Neumark D , Bank RJ , Nort KDV . 1996 . Sex discrimination in restaurant hiring: an audit study. Q. J. Econ. 111 : 915– 41 [Google Scholar]
  • Nickerson DW . 2008 . Is voting contagious? Evidence from two field experiments. Am. Political Sci. Rev. 102 : 49– 57 [Google Scholar]
  • Nolan JM , Kenefick J , Schultz PW . 2011 . Normative messages promoting energy conservation will be underestimated by experts unless you show them the data. Soc. Influence 6 : 169– 80 [Google Scholar]
  • Nolan JM , Schultz PW , Cialdini RB , Goldstein NJ , Griskevicius V . 2008 . Normative social influence is underdetected. Pers. Soc. Psychol. Bull. 34 : 913– 23 [Google Scholar]
  • Nosek B , Aarts A , Anderson J , Anderson C , Attridge P . et al. 2015a . Estimating the reproducibility of psychological science. Science 349 : 943– 51 [Google Scholar]
  • Nosek B , Alter G , Banks G , Borsboom D , Bowman S . et al. 2015b . Promoting an open research culture. Science 348 : 1422– 25 [Google Scholar]
  • Olken B . 2007 . Monitoring corruption: evidence from a field experiment in Indonesia. J. Political Econ. 115 : 200– 49 [Google Scholar]
  • Olken B . 2010 . Direct democracy and local public goods: evidence from a field experiment in Indonesia. Am. Political Sci. Rev. 104 : 243– 67 [Google Scholar]
  • Pager D . 2003 . The mark of a criminal record. Am. J. Sociol. 108 : 937– 75 [Google Scholar]
  • Pager D . 2007 . The use of field experiments for studies of employment discrimination: contributions, critiques, and directions for the future. Ann. Am. Acad. Political Soc. Sci. 609 : 104– 33 [Google Scholar]
  • Pager D , Quillian L . 2005 . Walking the talk: what employers say versus what they do. Am. Sociol. Rev. 70 : 355– 80 [Google Scholar]
  • Pager D , Western B , Bonikowski B . 2009 . Discrimination in a low-wage labor market: a field experiment. Am. Sociol. Rev. 74 : 777– 99 [Google Scholar]
  • Paluck EL . 2009 . Reducing intergroup prejudice and conflict using the media: a field experiment in Rwanda. Interpers. Relat. Group Process. 96 : 574– 87 [Google Scholar]
  • Paluck EL , Cialdini RB . 2014 . Field research methods. Handbook of Research Methods in Social and Personality Psychology HT Reis, CM Judd 81– 97 New York: Cambridge Univ. Press, 2nd ed.. [Google Scholar]
  • Paluck EL , Green DP . 2009 . Prejudice reduction: what works? A review and assessment of research and practice. Annu. Rev. Psychol. 60 : 339– 67 [Google Scholar]
  • Paluck EL , Shepherd H . 2012 . The salience of social referents: a field experiment on collective norms and harassment behavior in a school social network. J. Pers. Soc. Psychol. 103 : 899– 915 [Google Scholar]
  • Paluck EL , Shepherd H , Aronow PM . 2016 . Changing climates of conflict: a social network driven experiment in 56 schools. PNAS 113 : 566– 71 [Google Scholar]
  • Pedulla DS . 2016 . Penalized or protected? Gender and the consequences of non-standard and mismatched employment histories. Am. Sociol. Rev. 81 : 262– 89 [Google Scholar]
  • Pettigrew TF . 1998 . Intergroup contact theory. Annu. Rev. Psychol. 49 : 65– 85 [Google Scholar]
  • Riach PA , Rich J . 2002 . Field experiments of discrimination in the market place. Econ. J. 112 : 480– 518 [Google Scholar]
  • Rodríguez-Planas N . 2012 . Longer-term impacts of mentoring, educational services, and learning incentives: evidence from a randomized trial in the United States. Am. Econ. J. Appl. Econ. 4 : 121– 39 [Google Scholar]
  • Rondeau D , List JA . 2008 . Matching and challenge gifts to charity: evidence from laboratory and natural field experiments. Exp. Econ. 11 : 253– 67 [Google Scholar]
  • Ross SL , Turner MA . 2005 . Housing discrimination in metropolitan America: explaining changes between 1989 and 2000. Soc. Probl. 52 : 152– 80 [Google Scholar]
  • Rossi PH , Berk RA , Lenihan KJ . 1980 . Money, Work, and Crime: Experimental Evidence New York: Academic Press [Google Scholar]
  • Rossi PH , Berk RA , Lenihan KJ . 1982 . Saying it wrong with figures: a comment on Zeisel. Am. J. Sociol. 88 : 390– 93 [Google Scholar]
  • Rossi PH , Lyall KC . 1978 . An overview evaluation of the NIT experiment. Eval. Stud. Rev. 3 : 412– 28 [Google Scholar]
  • Sabin N . 2015 . Modern microfinance: a field in flux. Social Finance Nicholls A, Paton R, Emerson J Oxford, UK: Oxford Univ. Press [Google Scholar]
  • Salganik MJ , Dodds PS , Watts DJ . 2006 . Experimental study of inequality and unpredictability in an artificial cultural market. Science 311 : 854– 56 [Google Scholar]
  • Sampson RJ . 2008 . Moving to inequality: neighborhood effects and experiments meet social structure. Am. J. Sociol. 114 : 189– 231 [Google Scholar]
  • Sampson RJ . 2012 . Great American City: Chicago and the Enduring Neighborhood Effect Chicago, IL: Chicago Univ. Press [Google Scholar]
  • Schuler SR , Hashemi SM , Badal SH . 1998 . Men's violence against women in rural Bangladesh: undermined or exacerbated by microcredit programmes?. Dev. Pract. 8 : 148– 57 [Google Scholar]
  • Schultz P . 2004 . School subsidies for the poor: evaluating the Mexican Progresa poverty program. J. Dev. Econ. 74 : 199– 250 [Google Scholar]
  • Shadish WR , Cook TD . 2009 . The renaissance of field experimentation in evaluating interventions. Annu. Rev. Psychol. 607– 29 [Google Scholar]
  • Shadish WR , Cook TD , Campbell DT . 2002 . Experimental and Quasi-experimental Designs for Generalized Causal Inference. New York: Houghton, Mifflin and Company [Google Scholar]
  • Simpson BT , McGrimmon T , Irwin K . 2007 . Are blacks really less trusting than whites? Revisiting the race and trust question. Soc. Forces 86 : 525– 52 [Google Scholar]
  • Sniderman PM , Grob DB . 1996 . Innovations in experimental design in attitude surveys. Annu. Rev. Sociol. 22 : 377– 99 [Google Scholar]
  • Steinpreis RE , Anders KA , Ritzke D . 1999 . The impact of gender on the review of the curricula vitae of job applicants and tenure candidates: a national empirical study. Sex Roles 41 : 509– 28 [Google Scholar]
  • Stutzer A , Goette L , Zehnder M . 2011 . Active decisions and prosocial behaviour: a field experiment on blood donations. Econ. J. 121 : 476– 93 [Google Scholar]
  • Teele DL . 2014 . Reflections on the ethics of field experiments. Field Experiments and Their Critics: Essays on the Uses and Abuses of Experimentation in the Social Sciences DL Teele 115– 40 New Haven, CT: Yale Univ. Press [Google Scholar]
  • Thornton RL . 2008 . The demand for, and impact of, learning HIV status. Am. Econ. Rev. 98 : 1829– 63 [Google Scholar]
  • Tilcsik A . 2011 . Pride and prejudice: employment discrimination against openly gay men in the United States. Am. J. Sociol. 117 : 586– 626 [Google Scholar]
  • Travers J , Milgram S . 1969 . An experimental study of the small world problem. Sociometry 32 : 425– 43 [Google Scholar]
  • Turner MA , Bednarz BA , Herbig C , Lee SJ . 2003 . Discrimination in metropolitan housing markets phase 2: Asians and Pacific Islanders Tech. rep., Urban Inst., Washington, DC [Google Scholar]
  • Turner MA , Fix M , Struyk RJ . 1991 . Opportunities Denied, Opportunities Diminished: Racial Discrimination in Hiring Washington, DC: Urban Inst. Press [Google Scholar]
  • Turner MA , Ross SL , Galster GC , Yinger J . 2002 . Discrimination in metropolitan housing markets: national results from phase 1 of the Housing Discrimination Study (HDS) Tech. rep., Urban Inst Washington, DC: [Google Scholar]
  • Van Bavel JJ , Mende-Siedlecki P , Brady WJ , Reinero DA . 2016 . Contextual sensitivity in scientific reproducibility. PNAS 113 : 6454– 59 [Google Scholar]
  • Van de Rijt A , Kang SM , Restivo M , Patil A . 2014 . Field experiments of success-breeds-success dynamics. PNAS 111 : 6934– 39 [Google Scholar]
  • Van Der Merwe WG , Burns J . 2008 . What's in a name? Racial identity and altruism in post-apartheid South Africa. South Afr. J. Econ. 76 : 266– 75 [Google Scholar]
  • Vermeersch C , Kremer M . 2005 . School Meals, Educational Achievement, and School Competition: Evidence from a Randomized Evaluation. New York: World Bank [Google Scholar]
  • Volpp KG , Troxel AB , Pauly MV , Glick HA , Puig A . et al. 2009 . A randomized, controlled trial of financial incentives for smoking cessation. N. Engl. J. Med. 360 : 699– 709 [Google Scholar]
  • Whitt S , Wilson RK . 2007 . The dictator game, fairness and ethnicity in postwar Bosnia. Am. J. Political Sci. 51 : 655– 68 [Google Scholar]
  • Wienk RE , Reid CE , Simonson JC , Eggers FJ . 1979 . Measuring racial discrimination in American housing markets: the housing market practices survey. Tech. Rep. HUD-PDR-444(2), Dep. Hous. Urban Dev Washington, DC: [Google Scholar]
  • Williams WM , Ceci SJ . 2015 . National hiring experiments reveal 2:1 faculty preference for women on STEM tenure track. PNAS 112 : 5360– 65 [Google Scholar]
  • Yamagishi T . 2011 . Trust: The Evolutionary Game of Mind and Society New York: Springer [Google Scholar]
  • Yamagishi T , Cook KS , Watabe M . 1998 . Uncertainty, trust, and commitment formation in the United States and Japan. Am. J. Sociol. 104 : 165– 94 [Google Scholar]
  • Zeisel H . 1982 . Disagreement over the evaluation of a controlled experiment. Am. J. Sociol. 88 : 378– 89 [Google Scholar]

Data & Media loading...

  • Article Type: Review Article

Most Read This Month

Most cited most cited rss feed, birds of a feather: homophily in social networks, social capital: its origins and applications in modern sociology, conceptualizing stigma, framing processes and social movements: an overview and assessment, organizational learning, the study of boundaries in the social sciences, assessing “neighborhood effects”: social processes and new directions in research, social exchange theory, culture and cognition, focus groups.

Our systems are now restored following recent technical disruption, and we’re working hard to catch up on publishing. We apologise for the inconvenience caused. Find out more: https://www.cambridge.org/universitypress/about-us/news-and-blogs/cambridge-university-press-publishing-update-following-technical-disruption

We use cookies to distinguish you from other users and to provide you with a better experience on our websites. Close this message to accept cookies or find out how to manage your cookie settings .

Login Alert

  • > Journals
  • > Political Analysis
  • > Volume 25 Issue 4
  • > The Design of Field Experiments With Survey Outcomes:...

design field experiments

Article contents

The design of field experiments with survey outcomes: a framework for selecting more efficient, robust, and ethical designs.

Published online by Cambridge University Press:  18 September 2017

  • Supplementary materials

There is increasing interest in experiments where outcomes are measured by surveys and treatments are delivered by a separate mechanism in the real world, such as by mailers, door-to-door canvasses, phone calls, or online ads. However, common designs for such experiments are often prohibitively expensive, vulnerable to bias, and raise ethical concerns. We show how four methodological practices currently uncommon in such experiments have previously undocumented complementarities that can dramatically relax these constraints when at least two are used in combination: (1) online surveys recruited from a defined sampling frame (2) with at least one baseline wave prior to treatment (3) with multiple items combined into an index to measure outcomes and, (4) when possible, a placebo control. We provide a general and extensible framework that allows researchers to determine the most efficient mix of these practices in diverse applications. Two studies then examine how these practices perform empirically. First, we examine the representativeness of online panel respondents recruited from a defined sampling frame and find that their representativeness compares favorably to phone panel respondents. Second, an original experiment successfully implements all four practices in the context of a door-to-door canvassing experiment. We conclude discussing potential extensions.

Access options

Authors’ note : This paper previously circulated under the title “Testing Theories of Attitude Change With Online Panel Field Experiments.” Software for planning an experiment using all four practices we describe is available at http://experiments.berkeley.edu . Replication data is available as Broockman, Kalla, and Sekhon ( 2017 ), at http://dx.doi.org/10.7910/DVN/EEP5MT . This work was supported by the NARAL Pro-Choice America Foundation, the Signatures Innovations Fellows program at UC Berkeley, UC Berkeley’s Institute for Governmental Studies, and the Office of Naval Research [N00014-15-1-2367]. The studies reported herein were approved by Committee for the Protection of Human Subjects. We thank participants at the 2015 POLMETH meeting and at the University of California, Berkeley’s Research Workshop in American Politics for helpful feedback. Additional feedback was provided by Peter Aronow, Rebecca Barter, Kevin Collins, Alex Coppock, Jamie Druckman, Thad Dunning, Donald Green, Christian Fong, Seth Hill, Dan Hopkins, Gabe Lenz, Winston Lin, Chris Mann, David Nickerson, Kellie Ottoboni, Kevin Quinn, Fredrik Sävje, Yotam Shev-Tom, Bradley Spahn, and Laura Stoker. All remaining errors are our own.

Contributing Editor: R. Michael Alvarez

Broockman et al. supplementary material

Broockman et al. supplementary material 1

Crossref logo

This article has been cited by the following publications. This list is generated based on data provided by Crossref .

  • Google Scholar

View all Google Scholar citations for this article.

Save article to Kindle

To save this article to your Kindle, first ensure [email protected] is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle .

Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

  • Volume 25, Issue 4
  • David E. Broockman (a1) , Joshua L. Kalla (a2) and Jasjeet S. Sekhon (a3)
  • DOI: https://doi.org/10.1017/pan.2017.27

Save article to Dropbox

To save this article to your Dropbox account, please select one or more formats and confirm that you agree to abide by our usage policies. If this is the first time you used this feature, you will be asked to authorise Cambridge Core to connect with your Dropbox account. Find out more about saving content to Dropbox .

Save article to Google Drive

To save this article to your Google Drive account, please select one or more formats and confirm that you agree to abide by our usage policies. If this is the first time you used this feature, you will be asked to authorise Cambridge Core to connect with your Google Drive account. Find out more about saving content to Google Drive .

Reply to: Submit a response

- No HTML tags allowed - Web page URLs will display as text only - Lines and paragraphs break automatically - Attachments, images or tables are not permitted

Your details

Your email address will be used in order to notify you when your comment has been reviewed by the moderator and in case the author(s) of the article or the moderator need to contact you directly.

You have entered the maximum number of contributors

Conflicting interests.

Please list any fees and grants from, employment by, consultancy for, shared ownership in or any close relationship with, at any time over the preceding 36 months, any organisation whose interests may be affected by the publication of the response. Please also list any non-financial associations or interests (personal, professional, political, institutional, religious or other) that a reasonable reader would want to know about in relation to the submitted work. This pertains to all the authors of the piece, their spouses or partners.

tirthajyoti.github.io

Design-of-experiment (doe) matrix generator for engineering and statistics.

design field experiments

Copyright Notice and Code repository

Copyright (c): 2018-2028, Dr. Tirthajyoti Sarkar, Sunnyvale, CA 94086

It uses a MIT License, so although I retain the copyright of this particular code, please feel free to exercise your rights of the free software by using and enhancing it.

Please get the codebase from here .

Table of Contents

What is a scientific experiment, what is experimental design, options for open-source doe builder package in python, limitation of the foundation packages used, simplified user interface, designs available, what supporitng packages are required, eratta for using pydoe, how to run the program, is an installer/python library available, full-factorial design, fractional-factorial design, central-composite design, latin hypercube design, acknowledgements and requirements, introduction.

Design of Experiment (DOE) is an important activity for any scientist, engineer, or statistician planning to conduct experimental analysis. This exercise has become critical in this age of rapidly expanding field of data science and associated statistical modeling and machine learning . A well-planned DOE can give a researcher meaningful data set to act upon with optimal number of experiments preserving critical resources.

After all, aim of Data Science is essentially to conduct highest quality scientific investigation and modeling with real world data. And to do good science with data, one needs to collect it through carefully thought-out experiment to cover all corner cases and reduce any possible bias.

In its simplest form, a scientific experiment aims at predicting the outcome by introducing a change of the preconditions, which is represented by one or more independent variables , also referred to as “input variables” or “predictor variables.” The change in one or more independent variables is generally hypothesized to result in a change in one or more dependent variables , also referred to as “output variables” or “response variables.” The experimental design may also identify control variables that must be held constant to prevent external factors from affecting the results.

Experimental design involves not only the selection of suitable independent, dependent, and control variables, but planning the delivery of the experiment under statistically optimal conditions given the constraints of available resources. There are multiple approaches for determining the set of design points (unique combinations of the settings of the independent variables) to be used in the experiment.

Main concerns in experimental design include the establishment of validity , reliability , and replicability . For example, these concerns can be partially addressed by carefully choosing the independent variable, reducing the risk of measurement error, and ensuring that the documentation of the method is sufficiently detailed. Related concerns include achieving appropriate levels of statistical power and sensitivity .

Need for careful design of experiment arises in all fields of serious scientific, technological, and even social science investigation —  computer science, physics, geology, political science, electrical engineering, psychology, business marketing analysis, financial analytics , etc…

Unfortunately, majority of the state-of-the-art DOE generators are part of commercial statistical software packages like JMP (SAS) or Minitab . However, a researcher will surely be benefited if there exists an open-source code which presents an intuitive user interface for generating an experimental design plan from a simple list of input variables. There are a couple of DOE builder Python packages but individually they don’t cover all the necessary DOE methods and they lack a simplified user API, where one can just input a CSV file of input variables’ range and get back the DOE matrix in another CSV file.

This set of codes is a collection of functions which wrap around the core packages (mentioned below) and generate design-of-experiment (DOE) matrices for a statistician or engineer from an arbitrary range of input variables.

Both the core packages, which act as foundations to this repo, are not complete in the sense that they do not cover all the necessary functions to generate DOE table that a design engineer may need while planning an experiment. Also, they offer only low-level APIs in the sense that the standard output from them are normalized numpy arrays. It was felt that users, who may not be comfortable in dealing with Python objects directly, should be able to take advantage of their functionalities through a simplified user interface.

User just needs to provide a simple CSV file with a single table of variables and their ranges (2-level i.e. min/max or 3-level). Some of the functions work with 2-level min/max range while some others need 3-level ranges from the user (low-mid-high). Intelligence is built into the code to handle the case if the range input is not appropriate and to generate levels by simple linear interpolation from the given input. The code will generate the DOE as per user’s choice and write the matrix in a CSV file on to the disk. In this way, the only API user is exposed to are input and output CSV files. These files then can be used in any engineering simulator, software, process-control module, or fed into process equipments.

  • Full factorial,
  • 2-level fractional factorial,
  • Plackett-Burman,
  • Sukharev grid,
  • Box-Behnken,
  • Box-Wilson (Central-composite) with center-faced option,
  • Box-Wilson (Central-composite) with center-inscribed option,
  • Box-Wilson (Central-composite) with center-circumscribed option,
  • Latin hypercube (simple),
  • Latin hypercube (space-filling),
  • Random k-means cluster,
  • Maximin reconstruction,
  • Halton sequence based,
  • Uniform random matrix

How to use it?

First make sure you have all the necessary packages installed. You can simply run the .bash (Unix/Linux) and .bat (Windows) files provided in the repo, to install those packages from your command line interface. They contain the following commands,

Please note that as installed, PyDOE will throw some error related to type conversion. There are two options

  • I have modified the pyDOE code suitably and included a file with re-written functions in the repo. This is the file called by the program while executing, so you should see no error.
  • If you encounter any error, you could try to modify the PyDOE code by going to the folder where pyDOE files are copied and copying the two files doe_factorial.py and doe_box_behnken.py supplied with this repo.

Note this is just a code repository and not a installer package. For the time being, please clone this repo from GitHub , store all the files in a local directory.

git clone https://github.com/tirthajyoti/Design-of-experiment-Python.git

Then start using the software by simply typing,

python Main.py

After this, a simple menu will be printed on the screen and you will be prompted for a choice of number (a DOE) and name of the input CSV file (containing the names and ranges of your variables).

design field experiments

You must have an input parameters CSV file stored in the same directory that you are running this code from. You should use the supplied generic CSV file as an example. Please put the factors in the columns and the levels in the row (not the other way around). Couple of example CSV files are provided in the repo. Feel free to modify them as per your needs.

At this time, No . I plan to work on turning this into a full-fledged Python library which can be installed from PyPi repository by a PIP command. But I cannot promise any timeline for that :-) If somebody wants to collaborate and work on an installer, please feel free to do so.

Let’s say the input file contains the following table for the parameters range. Imagine this as a generic example of a checmical process in a plant.

Pressure Temperature FlowRate Time
40 290 0.2 5
55 320 0.3 8
70 350 0.4 11

If we build a full-factorial DOE out of this, we will get a table with 81 entries because 4 factors permuted in 3 levels result in 3^4=81 combinations!

Pressure Temperature FlowRate Time
40 290 0.2 5
50 290 0.2 5
70 290 0.2 5
40 320 0.2 5
50 320 0.2 5
70 320 0.2 5
40 290 0.3 8
50 290 0.3 8
70 290 0.3 8
40 320 0.3 8
50 320 0.3 8
70 320 0.3 8
40 320 0.4 11
50 320 0.4 11
70 320 0.4 11
40 350 0.4 11
50 350 0.4 11
70 350 0.4 11

Clearly the full-factorial designs grows quickly! Engineers and scientists therefore often use half-factorial/fractional-factorial designs where they confound one or more factors with other factors and build a reduced DOE. Let’s say we decide to build a 2-level fractional factorial of this set of variables with the 4th variables as the confounding factor (i.e. not an independent variable but as a function of other variables). If the functional relationship is “A B C BC” i.e. the 4th parameter vary depending only on 2nd and 3rd parameter, the output table could look like,

Pressure Temperature FlowRate Time
40 290 0.2 11
70 290 0.2 11
40 350 0.2 5
70 350 0.2 5
40 290 0.4 5
70 290 0.4 5
40 350 0.4 11
70 350 0.4 11

A Box-Wilson Central Composite Design, commonly called ‘a central composite design,’ contains an imbedded factorial or fractional factorial design with center points that is augmented with a group of ‘star points’ that allow estimation of curvature. One central composite design consists of cube points at the corners of a unit cube that is the product of the intervals [-1,1], star points along the axes at or outside the cube, and center points at the origin. Central composite designs are of three types. Circumscribed (CCC) designs are as described above. Inscribed (CCI) designs are as described above, but scaled so the star points take the values -1 and +1, and the cube points lie in the interior of the cube. Faced (CCF) designs have the star points on the faces of the cube. Faced designs have three levels per factor, in contrast with the other types that have five levels per factor. The following figure shows these three types of designs for three factors. [Read this page] (http://blog.minitab.com/blog/understanding-statistics/getting-started-with-factorial-design-of-experiments-doe) for more information about this kind of design philosophy.

Sometimes, a set of randomized design points within a given range could be attractive for the experimenter to asses the impact of the process variables on the output. Monte Carlo simulations are close example of this approach. However, a Latin Hypercube design is better choice for experimental design rather than building a complete random matrix as it tries to subdivide the sample space in smaller cells and choose only one element out of each subcell. This way, a more ‘uniform spreading’ of the random sample points can be obtained. User can choose the density of sample points. For example, if we choose to generate a Latin Hypercube of 12 experiments from the same input files, that could look like,

Pressure Temperature FlowRate Time
63.16 313.32 0.37 10.52
61.16 343.88 0.23 5.04
57.83 327.46 0.35 9.47
68.61 309.81 0.35 8.39
66.01 301.29 0.22 6.34
45.76 347.97 0.27 6.94
40.48 320.72 0.29 9.68
51.46 293.35 0.20 7.11
43.63 334.92 0.30 7.66
47.87 339.68 0.26 8.59
55.28 317.68 0.39 5.61
53.99 297.07 0.32 10.43

Of course, there is no guarantee that you will get the same matrix if you run this function because this are randomly sampled, but you get the idea!

The code was written in Python 3.6. It uses following external packages that needs to be installed on your system to use it,

  • pydoe: A package designed to help the scientist, engineer, statistician, etc., to construct appropriate experimental designs. Check the docs here .
  • diversipy: A collection of algorithms for sampling in hypercubes, selecting diverse subsets, and measuring diversity. Check the docs here .

Information

  • Author Services

Initiatives

You are accessing a machine-readable page. In order to be human-readable, please install an RSS reader.

All articles published by MDPI are made immediately available worldwide under an open access license. No special permission is required to reuse all or part of the article published by MDPI, including figures and tables. For articles published under an open access Creative Common CC BY license, any part of the article may be reused without permission provided that the original article is clearly cited. For more information, please refer to https://www.mdpi.com/openaccess .

Feature papers represent the most advanced research with significant potential for high impact in the field. A Feature Paper should be a substantial original Article that involves several techniques or approaches, provides an outlook for future research directions and describes possible research applications.

Feature papers are submitted upon individual invitation or recommendation by the scientific editors and must receive positive feedback from the reviewers.

Editor’s Choice articles are based on recommendations by the scientific editors of MDPI journals from around the world. Editors select a small number of articles recently published in the journal that they believe will be particularly interesting to readers, or important in the respective research area. The aim is to provide a snapshot of some of the most exciting work published in the various research areas of the journal.

Original Submission Date Received: .

  • Active Journals
  • Find a Journal
  • Journal Proposal
  • Proceedings Series
  • For Authors
  • For Reviewers
  • For Editors
  • For Librarians
  • For Publishers
  • For Societies
  • For Conference Organizers
  • Open Access Policy
  • Institutional Open Access Program
  • Special Issues Guidelines
  • Editorial Process
  • Research and Publication Ethics
  • Article Processing Charges
  • Testimonials
  • Preprints.org
  • SciProfiles
  • Encyclopedia

electronics-logo

Article Menu

design field experiments

  • Subscribe SciFeed
  • Recommended Articles
  • Author Biographies
  • Google Scholar
  • on Google Scholar
  • Table of Contents

Find support for a specific problem in the support section of our website.

Please let us know what you think of our products and services.

Visit our dedicated information section to learn more about MDPI.

JSmol Viewer

Experimental design of steel surface defect detection based on msfe-yolo—an improved yolov5 algorithm with multi-scale feature extraction.

design field experiments

1. Introduction

  • Experimental Design Framework Based on Deep Learning: We propose a deep learning-based experimental design framework that integrates artificial intelligence with industrial applications. This framework not only provides an innovative solution for steel surface defect detection but also serves as a teaching tool aimed at guiding students to learn and master relevant technologies, equipping them with the skills necessary for real-world industrial applications.
  • Introduction of Efficient Multi-Scale Attention (EMA) [ 25 ] Mechanism: By incorporating the EMA mechanism into the Backbone network of the YOLOv5 model, pixel-level relationships are captured through cross-dimensional interactions. Utilizing convolution kernels of varying sizes, the model efficiently fuses multi-scale contextual information, significantly enhancing the feature extraction capabilities and detection accuracy with only a slight increase in the computational cost.
  • Proposed Novel C3DX Module: In the Neck of the network, we introduce the Convolution 3 Dilated Convolution X (C3DX) module. This module uses dilated convolutions with different dilation rates to capture diverse receptive fields and integrate multi-scale contextual information, further improving defect detection precision. In addition to boosting detection performance, this module helps students understand the concept of receptive fields, fostering their innovative thinking skills.
  • Model Validation Across Multiple Datasets: The improved MSFE-YOLOv5 model has been validated on the NEU-DET, GC10-DET, Severstal Steel, and Crack500 datasets, with mean average precision (mAP) increases of 4.7%, 4.5%, 3.1% and 3.0%, respectively. These results demonstrate the model’s excellent performance in detection and generalization, while the experiments help students develop practical skills and the ability to solve real-world problems.

2. Materials and Methods

2.1. yolov5, 2.2. method, 2.2.2. c3dx, 2.2.3. eiou loss, 3.1. dataset preparation and preprocessing, 3.2. experimental setup, 3.3. evaluation metrics, 3.4. performance evaluation of each module, 3.4.1. attention effectiveness experiment, 3.4.2. comparison experiments, 3.4.3. ablation study, 4. discussion, author contributions, data availability statement, conflicts of interest.

NameMeanings
YOLOYOLO (You Only Look Once) is a deep learning model widely used for object detection tasks. Its core idea is to transform the object detection problem into a regression problem, predicting multiple classes and a bounding box
EMAEfficient Multi-Scale Attention.
C3DXConvolution 3 Dilated Convolution X
C3The C3 module is a feature extraction structure in the YOLOv5 that enhances feature extraction and fusion capabilities by incorporating a Cross-Stage Partial (CSP) network. This design further optimizes the model’s ability to capture and integrate features effectively.
mAPmAP (Mean Average Precision) is a comprehensive metric used to evaluate a model’s detection accuracy and localization precision across all categories, with higher values indicating better performance.
  • Vilček, I.; Řehoř, J.; Carou, D.; Zeman, P. Residual stresses evaluation in precision milling of hardened steel based on the deflection-electrochemical etching technique. Robot. Comput.-Integr. Manuf. 2017 , 47 , 112–116. [ Google Scholar ] [ CrossRef ]
  • Abbes, W.; Elleuch, J.F.; Sellami, D. Defect-Net: A new CNN model for steel surface defect classification. In Proceedings of the 2024 IEEE 12th International Symposium on Signal, Image, Video and Communications (ISIVC 2024), Marrakech, Morocco, 21–23 May 2024. [ Google Scholar ] [ CrossRef ]
  • Nguyen, H.-V.; Bae, J.-H.; Lee, Y.-E.; Lee, H.-S.; Kwon, K.-R. Comparison of Pre-Trained YOLO Models on Steel Surface Defects Detector Based on Transfer Learning with GPU-Based Embedded Devices. Sensors 2022 , 22 , 9926. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017 , 39 , 1137–1149. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Cai, Z.; Vasconcelos, N. Cascade R-CNN: High Quality Object Detection and Instance Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2021 , 43 , 1483–1498. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • He, Y.; Jin, Z.; Zhang, J.; Teng, S.; Chen, G.; Sun, X.; Cui, F. Pavement surface defect detection using mask Region-Based convolutional neural networks and transfer learning. Appl. Sci. 2022 , 12 , 7364. [ Google Scholar ] [ CrossRef ]
  • Si, B.; Yasengjiang, M.; Wu, H. Deep learning-based defect detection for hot-rolled strip steel. J. Phys. Conf. Ser. 2022 , 2246 , 012073. [ Google Scholar ] [ CrossRef ]
  • Zhao, W.; Chen, F.; Huang, H.; Li, D.; Cheng, W. A new steel defect detection algorithm based on deep learning. Comput. Intell. Neurosci. 2021 , 2021 , 592878. [ Google Scholar ] [ CrossRef ]
  • Shi, X.; Zhou, S.; Tai, Y.; Wang, J.; Wu, S.; Liu, J.; Xu, K.; Peng, T.; Zhang, Z. An improved faster R-CNN for steel surface defect detection. In Proceedings of the 2022 IEEE 24th International Workshop on Multimedia Signal Processing (MMSP 2022), Shanghai, China, 26–28 September 2022. [ Google Scholar ] [ CrossRef ]
  • Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal loss for dense object detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020 , 42 , 318–327. [ Google Scholar ] [ CrossRef ]
  • Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [ Google Scholar ] [ CrossRef ]
  • Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [ Google Scholar ] [ CrossRef ]
  • Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018 , arXiv:1804.02767. [ Google Scholar ]
  • Jocher, G.R.; Stoken, A.; Borovec, J.; Chaurasia, A.; Changyu, L.; Hogan, A.; Hajek, J.; Diaconu, L.; Kwon, Y.; Defretin, Y.; et al. Ultralytics/Yolov5: V5.0-YOLOv5-P6 1280 Models, AWS, Supervise.Ly and YouTube Integrations ; Zenodo: Geneva, Switzerland, 2021. [ Google Scholar ]
  • Guo, Z.; Wang, C.; Yang, G.; Huang, Z.; Li, G. MSFT-YOLO: Improved YOLOV5 based on transformer for detecting defects of steel surface. Sensors 2022 , 22 , 3467. [ Google Scholar ] [ CrossRef ]
  • Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), Long Beach, CA, USA, 4–9 December 2017; Volume 30, pp. 6000–6010. [ Google Scholar ]
  • Zhu, W.; Zhang, H.; Zhang, C.; Zhu, X.; Guan, Z.; Jia, J. Surface defect detection and classification of steel using an efficient Swin Transformer. Adv. Eng. Inform. 2023 , 57 , 102061. [ Google Scholar ] [ CrossRef ]
  • Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 10–17 October 2021. [ Google Scholar ] [ CrossRef ]
  • Yi, C.; Xu, B.; Chen, J.; Chen, Q.; Zhang, L. An improved YOLOX model for detecting strip surface defects. Steel Res. Int. 2022 , 93 , 2200505. [ Google Scholar ] [ CrossRef ]
  • Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021 , arXiv:2107.08430. [ Google Scholar ]
  • Xikun, X.; Changjiang, L.; Meng, X. Application of attention YOLOV 4 algorithm in metal defect detection. In Proceedings of the 2021 IEEE International Conference on Emergency Science and Information Technology (ICESIT 2021), Chongqing, China, 22–24 November 2021. [ Google Scholar ] [ CrossRef ]
  • Wang, L.; Liu, X.; Ma, J.; Su, W.; Li, H. Real-Time Steel Surface Defect Detection with Improved Multi-Scale YOLO-v5. Processes 2023 , 11 , 1357. [ Google Scholar ] [ CrossRef ]
  • Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable Bag-of-Freebies Sets New State-of-the-Art for Real-Time Object Detectors. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023. [ Google Scholar ] [ CrossRef ]
  • Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLO, Version 8.0.0 ; Ultralytics Inc.: Los Angeles, CA, USA, 2023; Available online: https://github.com/ultralytics/ultralytics (accessed on 10 June 2023).
  • Ouyang, D.; He, S.; Zhang, G.; Luo, M.; Guo, H.; Zhan, J.; Huang, Z. Efficient Multi-Scale Attention Module with Cross-Spatial Learning. In Proceedings of the 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [ Google Scholar ]
  • Wang, C.Y.; Liao, H.Y.M.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A New Backbone that can Enhance Learning Capability of CNN. IEEE Conf. Proc. 2020 , 2020 , 1571–1580. [ Google Scholar ]
  • Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [ Google Scholar ] [ CrossRef ]
  • Brauwers, G.; Frasincar, F. A General Survey on attention Mechanisms in Deep Learning. IEEE Trans. Knowl. Data Eng. 2023 , 35 , 3279–3298. [ Google Scholar ] [ CrossRef ]
  • Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020 , 42 , 2011–2023. [ Google Scholar ] [ CrossRef ]
  • Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Computer Vision—ECCV 2018 ; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2018; pp. 3–19. [ Google Scholar ] [ CrossRef ]
  • Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Montreal, BC, Canada, 10–17 October 2021. [ Google Scholar ] [ CrossRef ]
  • Yao, C.; Tang, Y.; Sun, J.; Gao, Y.; Zhu, C. Multiscale residual fusion network for image denoising. IET Image Process. 2021 , 16 , 878–887. [ Google Scholar ] [ CrossRef ]
  • Zheng, Z.; Wang, P.; Ren, D.; Liu, W.; Ye, R.; Hu, Q.; Zuo, W. Enhancing geometric factors in model learning and inference for object detection and instance segmentation. IEEE Trans. Cybern. 2022 , 52 , 8574–8586. [ Google Scholar ] [ CrossRef ]
  • Zhang, Y.-F.; Ren, W.; Zhang, Z.; Jia, Z.; Wang, L.; Tan, T. Focal and efficient IOU loss for accurate bounding box regression. Neurocomputing 2022 , 506 , 146–157. [ Google Scholar ] [ CrossRef ]
  • He, Y.; Song, K.; Meng, Q.; Yan, Y. An End-to-End steel surface defect detection approach via fusing multiple hierarchical features. IEEE Trans. Instrum. Meas. 2020 , 69 , 1493–1504. [ Google Scholar ] [ CrossRef ]
  • Qilong, W.; Banggu, W.; Pengfei, Z.; Peihua, L.; Wangmeng, Z.; Qinghua, H. ECA-Net: Efficient channel attention for deep convolutional neural networks. IEEE Conf. Proc. 2020 , 2020 , 11531–11539. [ Google Scholar ]
  • Lv, X.; Duan, F.; Jiang, J.-j.; Fu, X.; Gan, L. Deep Metallic Surface Defect Detection: The New Benchmark and Detection Network. Sensors 2020 , 20 , 1562. [ Google Scholar ] [ CrossRef ] [ PubMed ]
  • Severstal: Steel Defect Detection. Available online: https://www.kaggle.com/c/severstal-steel-defect-detection (accessed on 21 May 2021).
  • Yang, F.; Zhang, L.; Yu, S.; Prokhorov, D.; Mei, X.; Ling, H. Feature pyramid and hierarchical boosting network for pavement crack detection. IEEE Trans. Intell. Transp. Syst. 2020 , 21 , 1525–1535. [ Google Scholar ] [ CrossRef ]

Click here to enlarge figure

ModelAdvantagesDisadvantagesImprovements
Traditional Manual InspectionVisual identification, capable of detecting rare defectsRelies on subjective judgment, low efficiency, inaccurate, resource-intensiveNot suitable for large-scale or complex environments, both efficiency and accuracy are insufficient
Traditional Machine VisionAutomation reduces manual interventionManual feature extraction is complex, not suitable for complex environments, computationally intensiveFeature extraction complexity and unsuitability for complex detection environments
R-CNN Series (e.g., Faster R-CNN)High accuracy, widely used in defect detectionSlower speed, challenging to meet real-time requirements, computation–ally complexRegion proposal network limits speed and real-time performance
YOLO Series Real-time detection, fast, suitable for large-scale detectionPrecision for small defects is insufficient, accuracy slightly lowerPrecision issues, especially for small-sized defects
Transformer based YOLO improvementsEnhanced global feature extraction capabilityHigh computational complexity, training difficulties, struggles with real-time requirementsComputational complexity and training challenges limit industrial application
NameParameter
CPUIntel Core i9-7960X
GPURTX 3080Ti 12G
Operating SystemWindows10
Software environmentPython 3.7 + Pytorch 1.8.1 + CUDA 11.3
Method(a)(b)
Size (M) Time (ms)Size (M) Time (ms)
None [ ]13.676.113.413.676.113.4
SE13.778.714.613.778.113.9
CBAM13.778.918.113.778.516.4
ECA13.678.714.413.678.313.9
Transformer17.979.285.117.979.173.8
CA13.779.018.013.778.716.1
EMA (G = 4)13.779.215.713.778.714.8
EMA (G = 8)13.779.416.013.778.915.0
ModelSize
(M)
AP (%) (%)Time
(ms)
CrInPaPSRSSc
Faster R-CNN159.545.584.991.586.168.494.078.4100.4
Cascade R-CNN264.949.384.693.285.769.295.879.6204.2
RetinaNet145.149.082.894.087.966.391.078.554.8
YOLOv3236.548.079.489.379.759.690.274.448.4
YOLOv5s13.642.486.093.981.261.192.376.113.4
YOLOX68.540.885.991.887.861.984.275.418.6
YOLOv7-tiny11.648.782.593.583.553.588.975.111.1
YOLOv771.350.787.092.284.767.594.479.431.9
YOLOv8s22.543.081.492.682.564.394.676.416.3
Our model14.251.986.494.284.971.096.280.818.2
ModelGC10-DETSeverstal SteelCrack500
YOLOv5s-pre69.3%57.5%77.6
MSFE-YOLOv5s-pre72.0% (↑2.7%)59.7% (↑2.2%)79.8% (↑2.2%)
YOLOv5s61.9%55.1%78.1%
MSFE-YOLOv5s66.4% (↑4.5%)58.2% (↑3.1%)81.1% (↑3.0%)
MethodSize
(M)
AP (%)
(%)
Time
(ms)
CrInPaPSRSSc
YOLOv5s(baseline)13.642.486.093.981.261.192.376.113.4
+C3EMA13.748.185.393.786.567.894.979.416.0
+C3DX14.250.885.894.585.569.395.780.318.2
+EIoU14.251.986.494.284.971.096.280.818.2
The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

Li, L.; Zhang, R.; Xie, T.; He, Y.; Zhou, H.; Zhang, Y. Experimental Design of Steel Surface Defect Detection Based on MSFE-YOLO—An Improved YOLOV5 Algorithm with Multi-Scale Feature Extraction. Electronics 2024 , 13 , 3783. https://doi.org/10.3390/electronics13183783

Li L, Zhang R, Xie T, He Y, Zhou H, Zhang Y. Experimental Design of Steel Surface Defect Detection Based on MSFE-YOLO—An Improved YOLOV5 Algorithm with Multi-Scale Feature Extraction. Electronics . 2024; 13(18):3783. https://doi.org/10.3390/electronics13183783

Li, Lin, Ruopeng Zhang, Tunjun Xie, Yushan He, Hao Zhou, and Yongzhong Zhang. 2024. "Experimental Design of Steel Surface Defect Detection Based on MSFE-YOLO—An Improved YOLOV5 Algorithm with Multi-Scale Feature Extraction" Electronics 13, no. 18: 3783. https://doi.org/10.3390/electronics13183783

Article Metrics

Article access statistics, further information, mdpi initiatives, follow mdpi.

MDPI

Subscribe to receive issue release notifications and newsletters from MDPI journals

Analysis and optimization of performance parameters of the 220 Rn chamber in flow-field mode using computational fluid dynamics method

  • Published: 24 September 2024
  • Volume 35 , article number  175 , ( 2024 )

Cite this article

design field experiments

  • Shao-Hua Hu 1 ,
  • Yong-Jun Ye   ORCID: orcid.org/0000-0002-0944-181X 2 ,
  • Zheng-Zhong He   ORCID: orcid.org/0000-0003-4922-0453 1 ,
  • De-Tao Xiao   ORCID: orcid.org/0000-0002-1875-6733 1 ,
  • Xiang-Yu Xu 1 ,
  • Jian-Kai Wang 1 &
  • Qing-Zhi Zhou   ORCID: orcid.org/0009-0005-2179-4991 1  

The impact of the radiation dose produced by \({^{222}\hbox {Rn}}\) / \({^{220}{\hbox {Rn}}}\) and its progeny on human health has garnered increasing interest in the nuclear research field. The establishment of robust, regulatory, and competent \({^{220}{\hbox {Rn}}}\) chambers is crucial for accurately measuring radioactivity levels. However, studying the uniformity of the \({^{220}{\hbox {Rn}}}\) progeny through experimental methods is challenging, because measuring the concentration of \({^{220}{\hbox {Rn}}}\) and its progeny in multiple spatial locations simultaneously and in real time using experimental methods is difficult. Therefore, achieving precise control of the concentration of \({^{220}{\hbox {Rn}}}\) and its progeny as well as the reliable sampling of the progeny pose significant challenges. To solve this problem, this study uses computational fluid dynamics to obtain the flow-field data of the \({^{220}{\hbox {Rn}}}\) chamber under different wind speeds and progeny-replenishment rates. Qualitative analysis of the concentration distribution of the progeny and quantitative analysis of the progeny concentration and uniformity of the progeny concentration are conducted. The research findings indicated that the progeny concentration level is primarily influenced by wind speed and the progeny-complement rate. Wind speed also plays a crucial role in determining progeny concentration uniformity, whereas the progeny-complement rate has minimal impact on uniformity. To ensure the accuracy of \({^{220}{\hbox {Rn}}}\) progeny concentration sampling, we propose a methodology for selecting an appropriate sampling area based on varying progeny concentrations. This study holds immense importance for enhancing the regulation and measurement standards of \({^{220}{\hbox {Rn}}}\) and its progeny.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save.

  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime

Price includes VAT (Russian Federation)

Instant access to the full article PDF.

Rent this article via DeepDyve

Institutional subscriptions

design field experiments

Data availibility statement

The data that support the findings of this study are openly available in Science Data Bank at https://cstr.cn/31253.11.sciencedb.j00186.00107 and https://doi.org/10.57760/sciencedb.j00186.00107.

X. Deng, B. Yu, H. Wu et al., High-efficiency radon adsorption by nickel nanoparticles supported on activated carbon. New J. Chem. 46 , 9222–9228 (2022). https://doi.org/10.1039/d2nj00862a

Article   Google Scholar  

X. Deng, Y. Liao, M. Wang et al., Investigation of the effect of the key pore size on the radon adsorption performance by combining grand canonical Monte Carlo and activated carbon modification experiments. Appl. Surf. Sci. 643 , 158730 (2024). https://doi.org/10.1016/j.apsusc.2023.158730

T.K. Agarwal, B.K. Sahoo, M. Kumar et al., A computational fluid dynamics code for aerosol and decay-product studies in indoor environments. J. Radioanal. Nucl. Ch. 330 , 1347–1355 (2021). https://doi.org/10.1007/s10967-021-07877-8

L. Zhang, W. Zhuo, Q. Guo et al., An instrument for measuring the unattached fraction of radon progeny with etched track detectors. J. Radiol. Prot. 30 , 607 (2010). https://doi.org/10.1088/0952-4746/30/3/014

T. Asano, K. Sato, J.I. Onodera, United Nations scientific committee on the Effects of Atomic Radiation 2000 report. Jpn. J. Health. Phys. 36 , 149–158 (2001). https://doi.org/10.5453/jhps.36.149

Q. Guo, J. Cheng, Indoor thoron and radon concentrations in Zhuhai. China. J. Nucl. Sci. Technol. 42 , 588–591 (2005). https://doi.org/10.1080/18811248.2004.9726425

J. Hu, Y. Wu, M.A. Saputra et al., Radiation exposure due to \({^{222}}\) Rn, \({^{220}}\) Rn and their progenies in three metropolises in China and Japan with different air quality levels. J. Environ. Radioactiv. 244 , 106830 (2022). https://doi.org/10.1016/j.jenvrad.2022.106830

S.D. Kanse, B.K. Sahoo, J.J. Gaware et al., A study of thoron exhalation from monazite-rich beach sands of High Background Radiation Areas of Kerala and Odisha. India. Environ. Earth. Sc. 75 , 1–10 (2016). https://doi.org/10.1007/s12665-016-6279-9

Z.Z. He, D.T. Xiao, L.D. Lv et al., Controlling 212 Bi to 212 Pb activity concentration ratio in thoron chambers. J. Environ. Radioactiv. 178 , 77–83 (2017). https://doi.org/10.1016/j.jenvrad.2017.07.011

J. Lin, D. Xiao, Z. He et al., Regulation and control methods for the unattached fraction of \({^{220}}\) Rn progeny in a \({^{220}}\) Rn progeny chamber. J. Environ. Radioactiv. 235 , 106653 (2021). https://doi.org/10.1016/j.jenvrad.2021.106653

Z. He, D. Xiao, L. Lv et al., Stable control of thoron progeny concentration in a thoron chamber for calibration of active sampling monitors. Radiat. Meas. 102 , 27–33 (2017). https://doi.org/10.1016/j.radmeas.2017.02.013

H. Huang, G. Li, Q. Zhou et al., Study of the cyclic loss rate of 220 Rn progeny in a \({^{220}}\) Rn chamber by an airflow model. J. Radioanal. Nucl. Ch. 332 , 2633–2641 (2023). https://doi.org/10.1007/s10967-023-08930-4

Y. Ye, W. Liu, S. Li et al., A laboratory method for concurrently determining diffusion migration parameters and water saturation effects of thoron in uranium tailings. Chemosphere 249 , 126520 (2020). https://doi.org/10.1016/j.chemosphere.2020.126520

D. Tisha, G. Indranil, Prospective of employing high porosity open-cell metal foams in passive cryogenic radiators for space applications. In IOP Conference Series: Mater. Sci. Eng. 171 , 012048 (2017). https://doi.org/10.1088/1757-899X/171/1/012048

W. Choi, S. Hu, M. He et al., studied the Neighborhood-scale air quality impacts of emissions from motor vehicles and aircrafts. Atmos. Environ. 80 , 310–321 (2013). https://doi.org/10.1016/j.atmosenv.2013.07.043

Article   ADS   Google Scholar  

C. Kim, K. Zhou, Analysis of automotive disc brake squeal considering damping and design modifications for pads and a disc. Int. J. Auto. Tech-kor. 17 , 213–223 (2016). https://doi.org/10.1007/s12239-016-0021-1

K. Tsutsumi, S. Watanabe, S.I. Tsuda et al., Cavitation simulation of automotive torque converter using a homogeneous cavitation model. Eur. J. Mech. B-Fluids  61 , 263–270 (2017). https://doi.org/10.1016/j.euromechflu.2016.09.001

Z.R. Zhang, L. Hui, S.P. Zhu et al., Application of CFD in ship engineering design practice and ship hydrodynamics. J. Hydrodyn. Ser. B 18 , 315–322 (2006). https://doi.org/10.1016/S1001-6058(06)60072-3

S. Song, Y.K. Demirel, M. Atlar et al., Validation of the CFD approach for modelling roughness effect on ship resistance. Ocean Eng. 200 , 107029 (2020). https://doi.org/10.1016/j.oceaneng.2020.107029

D. Kim, S. Song, T. Tezdogan, Free running CFD simulations to investigate ship manoeuvrability in waves. Ocean Eng. 236 , 109567 (2021). https://doi.org/10.1016/j.oceaneng.2021.109567

B. Amblard, R. Singh, E. Gbordzoe et al., CFD modeling of the coke combustion in an industrial FCC regenerator. Chem. Eng. Sci. 170 , 731–742 (2017). https://doi.org/10.1016/j.ces.2016.12.055

N. Anna, D.W. Park, T. Charinpanitkul et al., Numerical analysis on premixed combustion of H 2 -SiCl 4 -Air system to prepare SiO 2 Particles. J. Ind. Eng. Chem. 18 , 509–512 (2012). https://doi.org/10.1016/j.jiec.2011.11.071

K. Ahookhosh, M. Saidi, H. Aminfar et al., Dry powder inhaler aerosol deposition in a model of tracheobronchial airways: validating CFD predictions with in vitro data. Int. J. Pharm. 587 , 119599 (2020). https://doi.org/10.1016/j.ijpharm.2020.119599

L.L.X. Augusto, G.C. Lopes, J.A.S. Gonçalves, A CFD study of deposition of pharmaceutical aerosols under different respiratory conditions. Braz. J. Chem. Eng. 33 , 549–558 (2016). https://doi.org/10.1590/0104-6632.20160333s20150100

Y. Shi, J. Wei, J. Qiu et al., Numerical study of acoustic agglomeration process of droplet aerosol using a three-dimensional CFD-DEM coupled model. Powder Technol. 362 , 37–53 (2020). https://doi.org/10.1016/j.powtec.2019.12.017

T. Zhenbo, Z. Wenqi, Y. Aibing et al., CFD-DEM investigation of the effect of agglomerate-agglomerate collision on dry powder aerosolisation. J. Aerosol Sci. 92 , 109–121 (2016). https://doi.org/10.1016/j.jaerosci.2015.11.005

T.K. Agarwal, B.K. Sahoo, M. Joshi et al., CFD simulations to study the effect of ventilation rate on \({^{220}}\) Rn concentration distribution in a test house. Radiat. Phys. Chem. 162 , 82–89 (2019). https://doi.org/10.1016/j.radphyschem.2019.04.018

T.K. Agarwal, J.J. Gaware, B.K. Sapra, A CFD-based approach to optimize operating parameters of a flow-through scintillation cell for measurement of \({^{220}}\) Rn in indoor environments. Environ. Sci. Pollut. R. 29 , 16404–16417 (2022). https://doi.org/10.1007/s11356-021-16780-4

Y. Ye, L.K. Chung, Q. Zhou et al., Evaluation of \({^{222}}\) Rn and \({^{220}}\) Rn discriminating concentration measurements with pinhole-based twin cup dosimeters using computational fluid dynamics simulations. Radiat. Meas. 134 , 106369 (2020). https://doi.org/10.1016/j.radmeas.2020.106369

T.K. Agarwal, S.D. Kanse, R. Mishra et al., A CFD based approach to assess the effect of environmental parameters on decay product-aerosol attachment coefficient. J. Radioanal. Nucl. Ch. 331 , 3563–3570 (2022). https://doi.org/10.1007/s10967-022-08402-1

K. Akbari, J. Mahmoudi, M. Ghanbari, Influence of indoor air conditions on radon concentration in a detached house. J. Environ. Radioactiv. 116 , 166–173 (2013). https://doi.org/10.1016/j.jenvrad.2012.08.013

W. Zhou, T. Iida, J. Moriizumi et al., Simulation of the concentrations and distributions of indoor radon and thoron. Radiat. Prot. Dosim. 93 , 357–367 (2001). https://doi.org/10.1093/oxfordjournals.rpd.a006448

T.K. Agarwal, B.K. Sahoo, J.J. Gaware et al., CFD based simulation of thoron ( \({^{220}}\) Rn) concentration in a delay chamber for mitigation application. J. Environ. Radioactiv. 136 , 16–21 (2014). https://doi.org/10.1016/j.jenvrad.2014.05.003

P.M. Dieguez-Elizondo, T. Gil-Lopez, P.G. O’Donohoe et al., An analysis of the radioactive contamination due to radon in a granite processing plant and its decontamination by ventilation. J. Environ. Radioactiv. 167 , 26–35 (2017). https://doi.org/10.1016/j.jenvrad.2016.11.016

H. Jun, Y. Guosheng, H. Miklós et al., Numerical modeling of the sources and behaviors of \({^{222}}\) Rn, \({^{220}}\) Rn and their progenies in the indoor environment-A review. J. Environ. Radioactiv. 123 , 114–126 (2018). https://doi.org/10.1016/j.jenvrad.2018.03.006

W. Li, Q. Zhou, Z. He et al., optimized a thoron progeny compensation system in a thoron calibration chamber. J. Radioanal. Nucl. Ch. 324 , 1255–1263 (2020). https://doi.org/10.1007/s10967-020-07180-y

A.C. Lai, W.W. Nazaroff, Modeling indoor particle deposition from turbulent flow onto smooth surfaces. J. Aerosol Sci. 31 , 463–476 (2000). https://doi.org/10.1016/S0021-8502(99)00536-4

Download references

Author information

Authors and affiliations.

School of Nuclear Science and Technology, University of South China, Hengyang, 421001, China

Shao-Hua Hu, Zheng-Zhong He, De-Tao Xiao, Xiang-Yu Xu, Jian-Kai Wang & Qing-Zhi Zhou

School of Resources Environment and Safety Engineering, University of South China, Hengyang, 421001, China

Yong-Jun Ye

You can also search for this author in PubMed   Google Scholar

Contributions

All authors contributed to the study conception and design. Material preparation, data collection, and analysis were performed by Shao-Hua Hu, Xiang-Yu Xu, and Jian-Kai Wang. The first draft of the manuscript was written by Shao-Hua Hu, and all authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.

Corresponding author

Correspondence to Qing-Zhi Zhou .

Ethics declarations

Conflict of interest.

The authors declare that they have no conflict of interest.

Additional information

This work was supported by the National Natural Science Foundation of China (Nos. 12375310, 12175102, and 118750356), Youth Talent Foundation of Hunan Province of China (2022TJ-Q16), and Graduate Research and Innovation Projects of Hunan Province (CX20230964).

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Hu, SH., Ye, YJ., He, ZZ. et al. Analysis and optimization of performance parameters of the 220 Rn chamber in flow-field mode using computational fluid dynamics method. NUCL SCI TECH 35 , 175 (2024). https://doi.org/10.1007/s41365-024-01526-x

Download citation

Received : 26 December 2023

Revised : 09 March 2024

Accepted : 23 March 2024

Published : 24 September 2024

DOI : https://doi.org/10.1007/s41365-024-01526-x

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • CFD simulation
  • Find a journal
  • Publish with us
  • Track your research

VIDEO

  1. Design of Experiments, Lecture 14: 3k Full Factorial Designs

  2. DOE Part 1 : Orthogonality in Design of Experiment

  3. Experiment design (with full sample test answer)

  4. What Is Design of Experiments? Part 2

  5. General Guidelines for the Layout of Field Experiments

  6. Design of Experiments / DOE (What is it and how is it done?)

COMMENTS

  1. Field experiment

    Field experiments allow researchers to collect diverse amounts and types of data. For example, a researcher could design an experiment that uses pre- and post-trial information in an appropriate statistical inference method to see if an intervention has an effect on subject-level changes in outcomes.

  2. What is a field experiment?

    Field experiments, explained. Editor's note: This is part of a series called "The Day Tomorrow Began," which explores the history of breakthroughs at UChicago. Learn more here. A field experiment is a research method that uses some controlled elements of traditional lab experiments, but takes place in natural, real-world settings.

  3. Introduction to Field Experiments and Randomized Controlled Trials

    Field experiments, or randomized studies conducted in real-world settings, can take many forms. While experiments on college campuses are often considered lab studies, certain experiments on campus - such as those examining club participation - may be regarded as field experiments, depending on the experimental design.

  4. Fundamentals of Experimental Design: Guidelines for Designing ...

    Four basic tenets or pillars of experimental design— replication, randomization, blocking, and size of experimental units— can be used creatively, intelligently, and consciously to solve both real and perceived problems in comparative experiments. ... orientation within the field, and basic experimental design characteristics, such as the ...

  5. Guide to Experimental Design

    Table of contents. Step 1: Define your variables. Step 2: Write your hypothesis. Step 3: Design your experimental treatments. Step 4: Assign your subjects to treatment groups. Step 5: Measure your dependent variable. Other interesting articles. Frequently asked questions about experiments.

  6. Conducting Effective Field Experiments in Psychology

    Field experiments in psychology are valuable for studying behavior in natural settings, allowing for more accurate and generalizable results. - To design an effective field experiment, researchers must carefully identify research questions, select appropriate settings, determine sample size, and control for confounding variables. -

  7. Field Experiment

    The design experiment can be considered as a special case of a field experiment; it has its roots in experimental research, both in 'true' and quasi-experiments, and is intended to provide formative feedback on, for example, practical problems in, say, teaching and learning, and to bridge the potential gap between research and practice ...

  8. Handbook of Field Experiments

    The history of field experiments in the marketing literature is surprisingly long. Early examples include Curhan (1974) and Eskin and Baron (1977), who vary prices, newspaper advertising, and display variables in grocery stores. This chapter reviews the recent history of field experiments in marketing by identifying papers published in the last ...

  9. Field Experiments

    Field experiments are experiments in settings with high degrees of naturalism. This article describes different types of field experiments, including randomized field trials, randomized rollout designs, encouragement designs, downstream field experiments, hybrid lab-field experiments, and covert population experiments, and discusses their intellectual background and benefits.

  10. Field Experiments

    field experiments can help one design better lab experiments, and have a methodological role quite apart from their complementarity at a substantive level. In section 2 we offer a typology of field experiments in the literature, identifying the key characteristics defining the species. We suggest some terminology to better identify different ...

  11. Experimental Design: Types, Examples & Methods

    Three types of experimental designs are commonly used: 1. Independent Measures. Independent measures design, also known as between-groups, is an experimental design where different participants are used in each condition of the independent variable. This means that each condition of the experiment includes a different group of participants.

  12. Design of experiments

    The use of a sequence of experiments, where the design of each may depend on the results of previous experiments, including the possible decision to stop experimenting, is within the scope of sequential analysis, a field that was pioneered [12] by Abraham Wald in the context of sequential tests of statistical hypotheses. [13] Herman Chernoff wrote an overview of optimal sequential designs, [14 ...

  13. Field Experiments: Design, Analysis, and Interpretation

    A brief, authoritative introduction to field experimentation in the social sciences. Written by two leading experts on experimental methods, this concise text covers the major aspects of experiment design, analysis, and interpretation in clear language. Students learn how to design randomized experiments, analyze the data, and interpret the ...

  14. Experimental Method In Psychology

    There are three types of experiments you need to know: 1. Lab Experiment. A laboratory experiment in psychology is a research method in which the experimenter manipulates one or more independent variables and measures the effects on the dependent variable under controlled conditions. A laboratory experiment is conducted under highly controlled ...

  15. Field Experiments and Natural Experiments

    It first defines field experimentation and describes the many forms that field experiments take. It also interprets the growth and development of field experimentation. ... Finally, it outlines a list of methodological issues that arise commonly in connection with experimental design and analysis: the role of covariates, planned vs. unplanned ...

  16. Field Experiments

    As the name suggests, a field study is an experiment performed outside the laboratory, in the 'real' world. Unlike case studies and observational studies, a field experiment still follows all of the steps of the scientific process, addressing research problems and generating hypotheses. The obvious advantage of a field study is that it is ...

  17. Field Experiment

    Field Experiment s are experiments that take place in real situations. The researcher, however, manipulates these environments. Some of the best examples of Field Experiment s involve research in preschool settings. The work of Johnson and his colleagues (e.g., J.Johnson & Ershler, 1981), Lawton (e.g., Lawton & Fowler, 1989), and P.K.Smith and Connolly (1980) with preschoolers is illustrative ...

  18. Field Experiments Across the Social Sciences

    Using field experiments, scholars can identify causal effects via randomization while studying people and groups in their naturally occurring contexts. In light of renewed interest in field experimental methods, this review covers a wide range of field experiments from across the social sciences, with an eye to those that adopt virtuous practices, including unobtrusive measurement ...

  19. The Design of Field Experiments With Survey Outcomes: A Framework for

    The Design of Field Experiments With Survey Outcomes: A Framework for Selecting More Efficient, Robust, and Ethical Designs - Volume 25 Issue 4. 22 August 2024: Due to technical disruption, we are experiencing some delays to publication. We are working to restore services and apologise for the inconvenience. For further updates please visit our ...

  20. Design-of-experiment (DOE) matrix generator for engineering and

    Design of Experiment (DOE) is an important activity for any scientist, engineer, or statistician planning to conduct experimental analysis. This exercise has become critical in this age of rapidly expanding field of data science and associated statistical modeling and machine learning. A well-planned DOE can give a researcher meaningful data ...

  21. Laboratories

    This lab has the following facilities for experiments, tests, and data collection and analysis: Oscilloscopes, Logic Analyzers, Thermal Chamber, FPGAs, microcontrollers, hacking tools, and more. The IC Design and Technology Laboratory is dedicated to teaching and research topics on electronic materials and devices, integrated circuit design ...

  22. PDF Multiple-Tuned Millipede Coil for High Field Imaging Applications

    W. H. Wong and S. Sukumar, Varian Inc., 3120 Hansen Way, Palo Alto, CA 94304. USA. Abstract: The "Millipede" RF coil, which was introduced by Wong and Sukumar last year, shows excellent B1 homogeneity and it is a single-tuned resonator. In this paper we shall describe a multiple-tuned coil based on the single- tuned "Millipede" coil design.

  23. PDF Dissertation

    Evidence from a Field Experiment", Management Science, under 3rd-round review.-Awarded by Digital Markets Initiative - Research Award Grant ($11,000), ... Systems Analysis and Design, Class Size: 168, 2019 • ISM 6257 - Intermediate Business Programming (Java), Class Size: 32, 2019

  24. Experimental Design of Steel Surface Defect Detection Based on ...

    Integrating artificial intelligence (AI) technology into student training programs is strategically crucial for developing future professionals with both forward-thinking capabilities and practical skills. This paper uses steel surface defect detection as a case study to propose a simulation-based teaching method grounded in deep learning. The method encompasses the entire process from data ...

  25. Analysis and optimization of performance parameters of the 220Rn

    The impact of the radiation dose produced by $${^{222}\\hbox {Rn}}$$ 222 Rn / $${^{220}{\\hbox {Rn}}}$$ 220 Rn and its progeny on human health has garnered increasing interest in the nuclear research field. The establishment of robust, regulatory, and competent $${^{220}{\\hbox {Rn}}}$$ 220 Rn chambers is crucial for accurately measuring radioactivity levels. However, studying the uniformity ...