Don’t Read Too Much Into It: What Brexit and U.S. Election Surprises Can Teach Us about PISA
The results of the Brexit referendum and U.S. presidential election will go down in history as the biggest surprises of 2016. The final results defied all predictions. The polls were wrong, as were the pundits. Though they predicted that the majority of Brits would vote to remain in the EU, more ended up voting to leave. Though they predicted a win for Clinton, Trump is the one moving into the White House this January. “From the US election to Brexit…the polls set voters and pundits up for a fall,” writes Siobhan Fenton, a correspondent for the UK newspaper The Independent.
There is plenty of head-scratching and hand-wringing over the fact that so many experts got it so wrong, but a generally agreed-upon conclusion is that the data these experts had absolute confidence in somehow fooled them or simply “died.” “I’ve believed in data for 30 years in politics and data died tonight. I could not have been more wrong about this election,” said GOP strategist and NBC analyst Mike Murphy the about the U.S. election on Twitter.
These two back-to-back spectacular failures of data-driven predictions remind us that data can be deceiving, misleading, and sometimes just quits working. Blind faith in data can have disastrous and long-lasting consequences; “…it [the failure of polls] serves as a devastating reminder of the uncertainty in polling, and a warning about being overconfident even when the weight of surveying evidence seems overwhelming,” writes the Economist shortly after the U.S. presidential election.
We are presented with two huge sets of data about education in the world as well as myriad interpretations. We saw the 2015 TIMSS data last week. And the influential international assessment program PISA has announced to release its 2015 results on December 6th. Without any doubt, pundits, journalists, and policy makers around the world will be commenting on the results, attempting to draw conclusions and make recommendations for educational policy and practices. For example, the Alliance for Excellent Education (the Alliance), a D.C.-based national policy and advocacy organization, has already declared December 6th PISA Day and created a website covering the results. It has planned a Deep Dive event the following day to discuss “PISA and the Global Economy.” Bob Wise, Alliance president and former governor of West Virginia, writes on the website:
PISA Day not only provides a look at student performance through an international lens, it focuses on what lessons can be learned from other high-performing nations to ensure U.S. students, especially those who are underserved, are prepared to compete in today’s global economy.
But if the PISA data, like most of the Brexit and presidential election data, is no good, would any conclusions drawn and recommendations made from this data be any good?
PISA is not a political poll, but it does attempt to predict the future. “PISA assesses the extent to which students near the end of compulsory education have acquired key knowledge and skills that are essential for full participation in modern societies,” according to its brochure. These students are 15 years old. In other words, PISA claims that its results predict the future success of individual students and, by association, their nations.
Unlike elections, one cannot definitively prove PISA predictions to be wrong since student success later in life cannot be conclusively reported like final vote counts. But if we think of a student’s success as winning the election, and the skills and knowledge PISA assesses as voters, what the polls missed during Brexit and the 2016 U.S. presidential election provides some interesting cautionary parallels.
The “Shy Tory” factor
British polls failed to predict the results of the 1992 General Election by a large margin. One major factor was the “shy Tories,” Conservative supporters who refused to indicate their voting intentions. Although polling agencies tried to improve their methods after 1992, their efforts might not have been sufficient, as the same phenomenon was observed again in 2015, when opinion polls once again underestimated the Conservative vote. Similarly, “shy Trump voters” were a major contributor to the polling failures in the presidential election. In both cases, “shy voters” drastically compromised the quality of predictive polls.
PISA has a similar problem. Whatever it claims to measure, the range of skills and knowledge PISA can actually assess is very limited and homogeneous for all students in over eighty economies. As a result, it easily misses skills and knowledge that ultimately matter more than what it measures, just like polls missed the opinions of “shy voters.” In other words, what eventually may matter for a student’s success does not show up in the data.
Thus, the first question to ask about PISA is: Does it accurately measure what matters?
“October surprise”
In U.S. presidential elections, pollsters are used to October surprises—unexpected events late in the campaign that could change the trajectory of the election. October surprises can also affect the accuracy of polls. Because they happen so close to Election Day, polls may not be able to fully capture their impact on voters and thereby predict accurately the final outcome.
PISA’s claim to assess skills and knowledge required for future success rests on the assumption that it confidently knows what is required for success in the future. While PISA and its expert sources may have done excellent work imagining the future, the impact of an “October surprise” in education would be much bigger, considering that election polls only attempting to predict the results of a single event (the election), while PISA attempts to predict the future lives of fifteen-year olds across the world. Given the rapidity and scale of changes we have seen in just the last few decades, no one can say how many “October surprises” these fifteen-year olds will experience in their future.
Over the last few decades, automation has already altered the life trajectory of millions of manufacturing workers. E-commerce platforms such as Amazon and eBay have altered the fate of brick-and-mortar store owners and their workforces. Driverless cars are set to change the life of human drivers. And the Fourth Industrial Revolution will further change the nature of work and living in ways we do not and cannot know for certain.
Thus follows a second question to ask about PISA: How does it assess skills needed for an uncertain future with such certainty?
The likely voter problem
Finding the “likely voter” or correctly identifying who will vote has become an increasingly difficult task for poll operators, but it is one of the most important factors affecting polling accuracy. The majority of polls for Brexit and the U.S. presidential election failed to accurately forecast who actually came out to vote, either under- or over estimating the turnout of supporters for each side. Polls make predictions about likely voters using various models, but one thing they rely heavily upon are past elections under the assumption that previous voting patterns will continue. This time, they failed: “pollsters may have incorrectly ruled out the prospect that people who didn’t vote in 2012 would nonetheless cast ballots in 2016, ” according to a story in USA Today. Sam Wang, the Princeton professor who ate a bug as promised if his projection of a Clinton win was proven wrong, said in an interview after the election: “[My] hypothesis was based on [elections from] 2004 to 2012…Now the hypothesis is wrong, and I have to go back and face up to the new data in an honest way.”
PISA’s confidence in the predictive power of its assessment also comes from the past. The subjects it chose to assess–science, reading, and math—have long been believed as important for success in life all over the world. They have been the core subject matter schools teach worldwide with the belief that they are essential for living in the modern age. But will these subjects turn out to help today’s fifteen-year olds ten, twenty, or thirty years later? Are they right candidates for all people in the future, or might different individuals need different sets of skills and knowledge?
This brings us to the third question to ask about PISA: Does its results in reading, math, and science accurately capture the domains of expertise each individual needs for successful participation in the future society anywhere in the world?
PISA not only tries to use its test results in these subjects to predict what skills and knowledge fifteen-year olds will need to succeed in the future, it also disseminates, based on these scores, education policies and practices it believes will equip children with these skills and knowledge. It has the potential to affect the livelihood of hundreds of millions of children, hence the entire world. The consequences are serious. The stakes are so high. Therefore, we must question the quality of the PISA results before eagerly jumping to conclusions. Don’t read too much into it.
p.s. For more about PISA’s problems and dangers, read my 5 part series entitled How PISA puts the world at risk on my blog.
[…] Don’t Read Too Much Into It: What Brexit and U.S. Election Surprises Can Teach Us about PISA Yong Zhao, Education in the Age of Globalization […]
[…] Yong Zhao on the recent US PISA score […]
Feel free to comment:
The views expressed on this site are entirely my own. They do not represent my employer or any other organization/institution. All comments are subject to approval.Archive
07.29.24 AI Has Little to Contribute to Traditional Education: Problems and Possibilities
03.17.24 Focused: Understanding, Negotiating, and Maximizing Your Influence as a School Leader
02.25.24 What Happened to Global Competence?
08.05.23 Rethinking the time spent at school: Could flexibility improve engagement and performance for students and teachers?
01.17.23 Introduction to Improbable Probabilities: The Unlikely Journey of Yong Zhao
01.05.23 How Not to Kill Creativity?
08.19.22 Preface to Improbable Probabilities: The Unlikely Journey of Yong Zhao
02.05.22 Introduction to New Book: Learning for Uncertainty: Teaching Students How to Thrive in a Rapidly Evolving World
09.25.21 Side effects in education: Taxonomy of educational outcomes
07.13.21 Introduction to My New Book: Learners without Borders
03.09.21 New article: Build back better: Avoid the learning loss trap
02.18.21 New article: The changes we need: Education post COVID-19
09.15.20 Watch Ep4 Creativity in Crisis: How well is creativity understood? A Conversation with Barb Kerr, Haiying Long, Ron Beghetto, & Yong Zhao
08.15.20 Can Creativity be Taught? Ep 3 of Creativity in Crisis on August 28th 3:00-4:00pm Pacific Time
07.13.20 Speak a Different Language: Reimagine the Grammar of Schooling
Tag Cloud
Accountability achievement gap CCSSO China/Chinese Commissioner Common Core Standards education Educational Policy Education Reforms national standards New York NGA Singapore standardized testing Standards student performance
WP Cumulus Flash tag cloud by Roy Tanck and Luke Morton requires Flash Player 9 or better.