Daniel Willingham--Science & Education
Hypothesis non fingo
  • Home
  • About
  • Books
  • Articles
  • Op-eds
  • Videos
  • Learning Styles FAQ
  • Daniel Willingham: Science and Education Blog

A physicist weighs in on the rigor of education research

6/5/2014

 
This post first appeared at RealClearEducation.com on May 6, 2014

As someone who spends most of their time thinking about the application of scientific findings to education, I encounter plenty of opinions about the scientific status of such efforts. Almost always, a comparison is made between the rigor of “hard” sciences and education. What varies is the accompanying judgment: derision for education researchers or despondency about the difficulty of the enterprise. In a recent article, physicist Carl Wieman (2014) offers a different perspective on the issue, suggesting that the difference between research in education and in physics is smaller than you might guess.

In education, Wieman is best known for refining and popularizing techniques to have college students better engage in large lecture courses. He started his career as a physicist, producing work that culminated in a Nobel prize. So he has some credentials in talking about the work of “hard” scientists.

Wieman begins by making clear what he takes to be the outcome of good science: predictive power. Can you use the results of your research to predict with some accuracy what will happen in a new situation? A common mistake is to believe that in education one ought to be able to predict outcomes for individual students; not necessarily so, any more than a physicist must be able to predict the behavior of each atom. Prediction in aggregate—a liter of gas or a school of children—is still an advance.

Wieman’s other points follow from his strong emphasis on prediction.

First, it follows that “rigorous” methods are any that contribute to better prediction. You don’t state in the absolute that randomized controlled trials are better than qualitative research. They provide different types of information in a larger effort to allow prediction.

Second, the emphasis on prediction frames the way one thinks about the messiness of research. Education research is often portrayed as inherently messy because there are so many variables at play. Physics research, in contrast, is portrayed as better controlled and more precise because there are many fewer variables that matter. Wieman argues this view is a misconception.

Physics seems tidy because you’ve probably only studied physics in textbooks, where everything is worked out: in other words, where no extraneous variables are discussed as possibly mattering. When the work that you study was first being conducted, it was plenty messy: false leads were pursued, ideas that (in retrospect) are self-contradictory were taken seriously, and so on. The same is true today: the frontiers of physics research is messy. “Messy” means that you don’t have a very good idea of which variables are important in gaining the predictive power that characterizes good science.

Third, Wieman suggests that bad research is the same in physics and education. Research is bad when the researcher has failed to account for factors that, based on prior research, he or she should have known to include. There is plenty of bad research in the hard sciences. People aren’t stupid; it’s just that science is hard.

I agree with Wieman that differences in “hardness” are mostly illusory. (That’s why I’ve been putting the term in quotation marks.) The fundamentals of the scientific method don’t differ much wherever they are applied. I also agree that people (usually people uninvolved in research) are too quick to conclude that, compared to other fields, a higher proportion of education research is low-quality. Come to a meeting of the Society for Neuroscience and I’ll show you plenty of studies that were poorly conceived, poorly controlled, or were simply wheel-spinning and will be ignored.

Wieman does ignore a difference between physics and education that I take to have important consequences: physics (and other basic sciences) strive to describe the world as it is, and so strive to be value-neutral. Education is an applied science; it is in the business of changing the world, making it more like it ought to be. As such, education is inevitably saturated with values.

Education policy would, I think, benefit, with a greater focus on the true differences between education and other varieties or research, and a reduced focus on the phantom differences in rigor.

Reference:

Wieman, C. E. (2014). The similarities between research in education and research in the hard sciences. Educational Researcher, 43, 12-14.  

Sid K link
6/5/2014 02:32:59 pm

There is one important sense in which physics is "harder" than education research---indeed, harder than most other sciences---which is the presence of very strong and universal theoretical constraints, guiding principles and frameworks.

For example, if your experiment violates the conservation of energy, or the conservation of angular momentum, then you know, with almost absolute certainty that you messed up somewhere. Or if you're conducting an experiment concerning microscopic entities, then surely you must use quantum mechanics. If your experiment is violating some principle of quantum mechanics (like the uncertainty principle), again you know for sure that you're missing something.

There are no such near-universal principles in education research (and indeed most social sciences) and there may never be any.

Another important sense in which sciences like physics, chemistry, biology are "harder" than social sciences is that they deal with stable ontologies: the nature of the entities under study is independent of what we say about them. But in social sciences, there is an interaction.

A concrete example: if you conduct studies and show that students good at chess do well in school and therefore you now create chess as an entrance criteria, then pretty soon students will focus on learning chess to pass the entrance exam and the correlation between chess and academic performance will go away.

Even more subtle is what Ian Hacking calls "dynamic nominalism." This is about how social categories can get created because they're named. For example, he argues that the actual (not just reported) incidence of multiple personality disorder increased enormously after the category of people with multiple personalities got created.

Jane
6/5/2014 05:35:43 pm

As a biologist currently working in curriculum development (math for life sciences majors), I wish I could agree with Wieman. But it was the stunningly low quality of much education research that really got me reading cognitive psychology.

In the higher education research literature, you often see somebody implement a multi-component "active learning" approach including quizzes in one section of a class and stay with traditional teaching in another section. But quizzes themselves improve learning, so is the pedagogical change responsible for higher scores among the active learning group or is it mostly the quizzes? Because of the lack of appropriate controls, we can't tell!

Statistical pseudoreplication is ubiquitous in this literature. When you teach one class one way and another class another way, the CLASS is the experimental unit. You cannot treat individual student scores as independent samples. Yet this is how statistical analysis is typically done in education research.

There are good ways to study complex, messy systems. (Yes, these include qualitative research!) I just wish they were used more often.


Comments are closed.

    Enter your email address:

    Delivered by FeedBurner

    RSS Feed


    Purpose

    The goal of this blog is to provide pointers to scientific findings that are applicable to education that I think ought to receive more attention.

    Archives

    January 2024
    April 2022
    July 2020
    May 2020
    March 2020
    February 2020
    December 2019
    October 2019
    April 2019
    March 2019
    January 2019
    October 2018
    September 2018
    August 2018
    June 2018
    March 2018
    February 2018
    November 2017
    October 2017
    September 2017
    August 2017
    July 2017
    June 2017
    April 2017
    March 2017
    February 2017
    November 2016
    September 2016
    August 2016
    July 2016
    June 2016
    May 2016
    April 2016
    December 2015
    July 2015
    April 2015
    March 2015
    January 2015
    September 2014
    August 2014
    July 2014
    June 2014
    May 2014
    April 2014
    March 2014
    February 2014
    January 2014
    December 2013
    November 2013
    October 2013
    September 2013
    August 2013
    July 2013
    June 2013
    May 2013
    April 2013
    March 2013
    February 2013
    January 2013
    December 2012
    November 2012
    October 2012
    September 2012
    August 2012
    July 2012
    June 2012
    May 2012
    April 2012
    March 2012
    February 2012

    Categories

    All
    21st Century Skills
    Academic Achievement
    Academic Achievement
    Achievement Gap
    Adhd
    Aera
    Animal Subjects
    Attention
    Book Review
    Charter Schools
    Child Development
    Classroom Time
    College
    Consciousness
    Curriculum
    Data Trustworthiness
    Education Schools
    Emotion
    Equality
    Exercise
    Expertise
    Forfun
    Gaming
    Gender
    Grades
    Higher Ed
    Homework
    Instructional Materials
    Intelligence
    International Comparisons
    Interventions
    Low Achievement
    Math
    Memory
    Meta Analysis
    Meta-analysis
    Metacognition
    Morality
    Motor Skill
    Multitasking
    Music
    Neuroscience
    Obituaries
    Parents
    Perception
    Phonological Awareness
    Plagiarism
    Politics
    Poverty
    Preschool
    Principals
    Prior Knowledge
    Problem-solving
    Reading
    Research
    Science
    Self-concept
    Self Control
    Self-control
    Sleep
    Socioeconomic Status
    Spatial Skills
    Standardized Tests
    Stereotypes
    Stress
    Teacher Evaluation
    Teaching
    Technology
    Value-added
    Vocabulary
    Working Memory