September, 2020: Opening School Doors to New Opportunities for Universal Success

“Now is the time for all good schools to come to the aid of their country.”

In times of great danger, nations have always called upon their citizens to volunteer to do what is necessary to solve their most pressing problems. Today, our most immediate crisis is, of course, the COVID-19 pandemic. This time, the heroes who have come forward are health care providers, who risk their lives to save the lives of others. The many people who work in essential services, such as grocery stores and pharmacies, also subject themselves to risks so that others can survive. Teachers across the country are working day and night to prepare online lessons, as well as helping get food to hungry students.  Behind the scenes, scientists are working to find cures, expand testing, and determine when it will be safe for our society to return to normal.

In a few months, we will face a new emergency. Schools will open. Hopefully, school opening will not pose major health threats to students and staff, assuming that the danger of infection has passed. But we will without any doubt face a new set of challenges in the education of the more than 50 million children in elementary and secondary schools in the U.S., as well as the billion students in the world as a whole.

blog_5-7-20_backtoschool500x333 2In the U.S., children re-entering our schools will have been out of school since March. Some may have kept up with their school work online, but most will have had little formal schooling for six months. This will be most serious, of course, among the students most at risk. By next September, 2020, millions of children will not only have missed out on schooling, but many will also be traumatized by what they have experienced since they were last in school. Many will have experienced the disease or death of a close relative. Many will have parents who have lost their jobs, and may have lived in fear of lacking food or safety.

This is a predictable crisis. No one can expect that schools and students will just pick up and carry on when schools re-open, as though they’ve just had a few snow days.  No teacher is going to say on Day 1, “Please open your textbook to the page where we left off last March.”

As educators and policy makers, it would be irresponsible to wait until schools re-open and only then take action to solve the entirely predictable problems. Instead, we need to prepare, starting today, to create the schools students will need in September, 2020, or whenever it is deemed safe for schools to open.

Here are a few ideas I would propose to address the problems students are likely to have.

  1. Bring all students up to grade level in reading and mathematics.

In two recent blogs (here and here) I discussed one aspect of this problem, the fact that many students will have fallen behind in basic skills because of their long absence from face-to-face school. I proposed a Marshall Plan for education, including mobilization of tens of thousands of recent college graduates, and others eager to help, to serve as paid tutors to students who are struggling in reading and/or mathematics.  As I noted, research overwhelmingly points to tutoring as the most effective strategy to accelerate the achievement of students who are performing below their capabilities.  According to the evidence, several one-to-one or one-to-small group tutoring models can routinely increase student achievement by an effect size of +0.40 in a year (almost equivalent to the difference between middle class and disadvantaged students). But what if students received effective tutoring for two years, or longer? What if their classroom teachers used teaching methods proven to be effective, contributing further to student success? What if schools could provide services to students with problems with their vision or hearing, or chronic health problems such as asthma? Based on what we already know how to do, a goal of steadily increasing the percentage of students performing at today’s definition of “at grade level” could increase each year, until virtually all students could expect that level of performance.

  1. Schools need to welcome back every child.

When students return to school after the long delay and trauma they may have endured, they need to be welcomed back with enthusiasm by all school staff. The return will create a psychological opportunity.  Students will always remember what happened on the first day, the first weeks, the first months. A big party to welcome students back is a good start, but students will need constant and sincere affirmations of their value and importance to the adults in the school. They need to be told, one at a time and by name, how much they were missed, and how glad everyone is that they are back, safe and healthy. I think the theme of each school should be “a once-in-a-lifetime chance to connect with the school,” not “at last, everything is back to normal.”

  1.  Schools need social emotional and health solutions

In addition to using proven academic approaches, schools need to implement proven social-emotional and health promotion strategies to help all students reconnect and thrive.  Strategies to build self-concept, positive relations with peers, concern for the well-being of others, and a commitment to banish violence and bullying will be especially important.  Cooperative learning can help to build friendships, acceptance, and engagement, in addition to improving achievement.

In light of all that has happened, schools need to enthusiastically welcome their students back, and then provide them the success, respect, and love that they deserve.  They need to give them every reason to believe that they have a new opportunity to achieve success.  Students, parents, and educators alike need to have well-founded confidence that out of the destruction caused by the pandemic, there will come triumph.

 This blog was developed with support from Arnold Ventures. The views expressed here do not necessarily reflect those of Arnold Ventures.

Note: If you would like to subscribe to Robert Slavin’s weekly blogs, just send your email address to thebee@bestevidence.org

New Sections on Social Emotional Learning and Attendance in Evidence for ESSA!

We are proud to announce the launch of two new sections of our Evidence for ESSA website (www.evidenceforessa.org): K-12 social-emotional learning and attendance. Funded by a grant from the Bill and Melinda Gates Foundation, the new sections represent our first foray beyond academic achievement.

blog_2-6-20_evidenceessa_500x333

The social-emotional learning section represents the greatest departure from our prior work. This is due to the nature of SEL, which combines many quite diverse measures. We identified 17 distinct measures, which we grouped in four overarching categories, as follows:

Academic Competence

  • Academic performance
  • Academic engagement

Problem Behaviors

  • Aggression/misconduct
  • Bullying
  • Disruptive behavior
  • Drug/alcohol abuse
  • Sexual/racial harassment or aggression
  • Early/risky sexual behavior

Social Relationships

  • Empathy
  • Interpersonal relationships
  • Pro-social behavior
  • Social skills
  • School climate

Emotional Well-Being

  • Reduction of anxiety/depression
  • Coping skills/stress management
  • Emotional regulation
  • Self-esteem/self-efficacy

Evidence for ESSA reports overall effect sizes and ratings for each of the four categories, as well as the 17 individual measures (which are themselves composed of many measures used by various qualifying studies). So in contrast to reading and math, where programs are rated based on the average of all qualifying  reading or math measures, an SEL program could be rated “strong” in one category, “promising” in another, and “no qualifying evidence” or “qualifying studies found no significant positive effects” on others.

Social-Emotional Learning

The SEL review, led by Sooyeon Byun, Amanda Inns, Cynthia Lake, and Liz Kim at Johns Hopkins University, located 24 SEL programs that both met our inclusion standards and had at least one study that met strong, moderate, or promising standards on at least one of the four categories of outcomes.

There is much more evidence at the elementary and middle school levels than at the high school level. Recognizing that some programs had qualifying outcomes at multiple levels, there were 7 programs with positive evidence for pre-K/K, 10 for 1-2, 13 for 3-6, and 9 for middle school. In contrast, there were only 4 programs with positive effects in senior high schools. Fourteen studies took place in urban locations, 5 in suburbs, and 5 in rural districts.

The outcome variables most often showing positive impacts include social skills (12), school climate (10), academic performance (10), pro-social behavior (8), aggression/misconduct (7), disruptive behavior (7), academic engagement (7), interpersonal relationships (7), anxiety/depression (6), bullying (6), and empathy (5). Fifteen of the programs targeted whole classes or schools, and 9 targeted individual students.

Several programs stood out in terms of the size of the impacts. Take the Lead found effect sizes of +0.88 for social relationships and +0.51 for problem behaviors. Check, Connect, and Expect found effect sizes of +0.51 for emotional well-being, +0.29 for problem behaviors, and +0.28 for academic competence. I Can Problem Solve found effect sizes of +0.57 on school climate. The Incredible Years Classroom and Parent Training Approach reported effect sizes of +.57 for emotional regulation, +0.35 for pro-social behavior, and +0.21 for aggression/misconduct. The related Dinosaur School classroom management model reported effect sizes of +0.31 for aggression/misbehavior. Class-Wide Function-Related Intervention Teams (CW-FIT), an intervention for elementary students with emotional and behavioral disorders, had effect sizes of +0.47 and +0.30 across two studies for academic engagement and +0.38 and +0.21 for disruptive behavior. It also reported effect sizes of +0.37 for interpersonal relationships, +0.28 for social skills, and +0.26 for empathy. Student Success Skills reported effect sizes of +0.30 for problem behaviors, +0.23 for academic competence, and +0.16 for social relationships.

In addition to the 24 highlighted programs, Evidence for ESSA lists 145 programs that were no longer available, had no qualifying studies (e.g., no control group), or had one or more qualifying studies but none that met the ESSA Strong, Moderate, or Promising criteria. These programs can be found by clicking on the “search” bar.

There are many problems inherent to interpreting research on social-emotional skills. One is that some programs may appear more effective than others because they use measures such as self-report, or behavior ratings by the teachers who taught the program. In contrast, studies that used more objective measures, such as independent observations or routinely collected data, may obtain smaller impacts. Also, SEL studies typically measure many outcomes and only a few may have positive impacts.

In the coming months, we will be doing analyses and looking for patterns in the data, and will have more to say about overall generalizations. For now, the new SEL section provides a guide to what we know now about individual programs, but there is much more to learn about this important topic.

Attendance

Our attendance review was led by Chenchen Shi, Cynthia Lake, and Amanda Inns. It located ten attendance programs that met our standards. Only three of these reported on chronic absenteeism, which refers to students missing more than 10% of days. Many more focused on average daily attendance (ADA). Among programs focused on average daily attendance, a Milwaukee elementary school program called SPARK had the largest impact (ES=+0.25). This is not an attendance program per se, but it uses AmeriCorps members to provide tutoring services across the school, as well as involving families. SPARK has been shown to have strong effects on reading, as well as its impressive effects on attendance. Positive Action is another schoolwide approach, in this case focused on SEL. It has been found in two major studies in grades K-8 to improve student reading and math achievement, as well as overall attendance, with a mean effect size of +0.20.

The one program to report data on both ADA and chronic absenteeism is called Attendance and Truancy Intervention and Universal Procedures, or ATI-UP. It reported an effect size in grades K-6 of +0.19 for ADA and +0.08 for chronic attendance. Talent Development High School (TDHS) is a ninth grade intervention program that provides interdisciplinary learning communities and “double dose” English and math classes for students who need them. TDHS reported an effect size of +0.17.

An interesting approach with a modest effect size but very modest cost is now called EveryDay Labs (formerly InClass Today). This program helps schools organize and implement a system to send postcards to parents reminding them of the importance of student attendance. If students start missing school, the postcards include this information as well. The effect size across two studies was a respectable +0.16.

As with SEL, we will be doing further work to draw broader lessons from research on attendance in the coming months. One pattern that seems clear already is that effective attendance improvement models work on building close relationships between at-risk students and concerned adults. None of the effective programs primarily uses punishment to improve attendance, but instead they focus on providing information to parents and students and on making it clear to students that they are welcome in school and missed when they are gone.

Both SEL and attendance are topics of much discussion right now, and we hope these new sections will be useful and timely in helping schools make informed choices about how to improve social-emotional and attendance outcomes for all students.

This blog was developed with support from the Laura and John Arnold Foundation. The views expressed here do not necessarily reflect those of the Foundation.

Measuring Social Emotional Skills in Schools: Return of the MOOSES

Throughout the U. S., there is huge interest in improving students’ social emotional skills and related behaviors. This is indeed important as a means of building tomorrow’s society. However, measuring SEL skills is terribly difficult. Not that measuring reading, math, or science learning is easy, but there are at least accepted measures in those areas. In SEL, almost anything goes, and measures cover an enormous range. Some measures might be fine for theoretical research and some would be all right if they were given independently of the teachers who administered the treatment, but SEL measures are inherently squishy.

A few months ago, I wrote a blog on measurement of social emotional skills. In it, I argued that social emotional skills should be measured in pragmatic school research as objectively as possible, especially to avoid measures that merely reflect having students in experimental groups repeating back attitudes or terminology they learned in the program. I expressed the ideal for social emotional measurement in school experiments as MOOSES: Measurable, Observable, Objective, Social Emotional Skills.

Since that time, our group at Johns Hopkins University has received a generous grant from the Gates Foundation to add research on social emotional skills and attendance to our Evidence for ESSA website. This has enabled our group to dig a lot deeper into measures for social emotional learning. In particular, JHU graduate student Sooyeon Byun created a typology of SEL measures arrayed from least to most MOOSE-like. This is as follows.

  1. Cognitive Skills or Low-Level SEL Skills.

Examples include executive functioning tasks such as pencil tapping, the Stroop test, and other measures of cognitive regulation, as well as recognition of emotions. These skills may be of importance as part of theories of action leading to social emotional skills of importance to schools, but they are not goals of obvious importance to educators in themselves.

  1. Attitudes toward SEL (non-behavioral).

These include agreement with statements such as “bullying is wrong,” and statements about why other students engage in certain behaviors (e.g., “He spilled the milk because he was mean.”).

  1. Intention for SEL behaviors (quasi-behavioral).

Scenario-based measures (e.g., what would you do in this situation?).

  1. SEL behaviors based on self-report (semi-behavioral).

Reports of actual behaviors of self, or observations of others, often with frequencies (e.g., “How often have you seen bullying in this school during this school year?”) or “How often do you feel anxious or afraid in class in this school?”)

This category was divided according to who is reporting:

4a. Interested party (e.g., report by teachers or parents who implemented the program and may have reason to want to give a positive report)

4b. Disinterested party (e.g., report by students or by teachers or parents who did not administer the treatment)

  1. MOOSES (Measurable, Observable, Objective Social Emotional Skills)
  • Behaviors observed by independent observers, either researchers, ideally unaware of treatment assignment, or by school officials reporting on behaviors as they always would, not as part of a study (e.g., regular reports of office referrals for various infractions, suspensions, or expulsions).
  • Standardized tests
  • Other school records

blog_2-21-19_twomoose_500x333

Uses for MOOSES

All other things being equal, school researchers and educators should want to know about measures as high as possible on the MOOSES scale. However, all things are never equal, and in practice, some measures lower on the MOOSES scale may be all that exists or ever could exist. For example, it is unlikely that school officials or independent observers could determine students’ anxiety or fear, so self-report (level 4b) may be essential. MOOSES measures (level 5) may be objectively reported by school officials, but limiting attention to such measures may limit SEL measurement to readily observable behaviors, such as aggression, truancy, and other behaviors of importance to school management, and not on difficult-to-observe behaviors such as bullying.

Still, we expect to find in our ongoing review of the SEL literature that there will be enough research on outcomes measured at level 3 or above to enable us to downplay levels 1 and 2 for school audiences, and in many cases to downplay reports by interested parties in level 4a, where teachers or parents who implement a program then rate the behavior of the children they served.

Social emotional learning is important, and we need measures that reflect their importance, minimizing potential bias and staying as close as possible to independent, meaningful measures of behaviors that are of the greatest importance to educators. In our research team, we have very productive arguments about these measurement issues in the course of reviewing individual articles. I placed a cardboard cutout of a “principal” called “Norm” in our conference room. Whenever things get too theoretical, we consult “Norm” for his advice. For example, “Norm” is not too interested in pencil tapping and Stroop tests, but he sure cares a lot about bullying, aggression, and truancy. Of course, as part of our review we will be discussing our issues and initial decisions with real principals and educators, as well as other experts on SEL.

The growing number of studies of SEL in recent years enables reviewers to set higher standards than would have been feasible even just a few years ago. We still have to maintain a balance in which we can be as rigorous as possible but not end up with too few studies to review.  We can all aspire to be MOOSES, but that is not practical for some measures. Instead, it is useful to have a model of the ideal and what approaches the ideal, so we can make sense of the studies that exist today, with all due recognition of when we are accepting measures that are nearly MOOSES but not quite the real Bullwinkle

This blog was developed with support from the Laura and John Arnold Foundation. The views expressed here do not necessarily reflect those of the Foundation.

Reviewing Social and Emotional Learning for ESSA: MOOSES, not Parrots

This blog was co-authored by Elizabeth Kim

I’m delighted to see all the interest lately in social and emotional skills. These range widely, from kindness and empathy to ability to delay gratification to grit to belief that effort is more important than intelligence to avoidance of bullying, violence, and absenteeism. Social and emotional learning (SEL) has taken on even more importance as the Every Student Succeeds Act (ESSA) allows states to add to their usual reading and math accountability measures, and some are adding measures of SEL. This makes it particularly important to have rigorous research on this topic.

I’ve long been interested in social-emotional development, but I have just started working with a student, Liz Kim, on a systematic review of SEL research. Actually, Liz is doing all the work. Part of the purpose of the SEL review is to add a section on this topic to Evidence for ESSA. In conceptualizing our review, we immediately ran into a problem. While researchers studying achievement mostly use tests, essays, products, and other fairly objective indicators, those studying social-emotional skills and behaviors use a wide variety of measures, many of which are far less objective. For example, studies of social-emotional skills make much use of student self-report, or ratings of students’ behaviors by the teachers who administered the treatment. Researchers in this field are well aware of the importance of objectivity, but they report more and less objective measures within the same studies depending on their research purposes. For academic purposes this is perfectly fine. SEL researchers and the readers of their reports are of course free to emphasize whichever measures they find most meaningful.

The problem arises when SEL measures are used in reviews of research to determine which programs and practices meet the ESSA standards for strong, moderate, or promising levels of evidence. Under ESSA, selecting programs meeting strong, moderate, or promising criteria can have consequences for schools in terms of grant funding, so it could be argued that more objective measures should be required.

In our reviews of K-12 reading and math programs for Evidence for ESSA, we took a hard line on objectivity. For example, we do not accept outcome measures made by the researchers or developers, or those that assess skills taught in the experimental group but not the control group. The reason for this is that effect sizes for such studies are substantially inflated in comparison to independent measures. We also do not accept achievement measures administered individually to students by the students’ own teachers, who implemented the experimental treatment, for the same reason. In the case of achievement studies that use independent measures, at least as one of several measures, we can usually exclude non-independent measures without excluding whole studies.

Now consider measures in studies of social-emotional skills. They are often dependent on behavior ratings by teachers or self-reports by students. For example, in some studies students are taught to recognize emotions in drawings or photos of people. Recognizing emotions accurately may correlate with valuable social-emotional skills, but an experiment whose only outcome is the ability to recognize emotions could just be teaching students to parrot back answers on a task of unknown practical value in life. Many SEL measures used in studies with children are behavior ratings by the very teachers who delivered the treatment. Teacher ratings are sure to be biased (on average) by the normal human desire to look good (called social desirability bias). This is particularly problematic when teachers are trained to use a strategy to improve a particular outcome. For example, some programs are designed to improve students’ empathy. That’s a worthy goal, but empathy is hard to identify in practice. So teachers taught to identify behaviors thought to represent empathy are sure to see those behaviors in their children a lot more than teachers in the control group do, not necessarily because those children are in fact more empathetic, but because teachers and the children themselves may have learned a new vocabulary to recognize, describe, and exhibit empathy. This could be seen as another example of “parroting,” which means that subjects or involved raters (such as teachers or parents) have learned what to say or how to act under observation at the time of rating, instead of truly changing behaviors or attitudes.

For consequential purposes, such as reviews for ESSA evidence standards, it makes sense to ask for independently verified indicators demonstrating that students in an experimental group can and do engage in behaviors that are likely to help them in life. Having independent observers blind to treatments observe students in class or carry out structured tasks indicating empathetic or prosocial or cooperative behavior, for example, is very different from asking them on a questionnaire whether they engage in those behaviors or have beliefs in line with those skills. The problem is not only that attitudes and behaviors are not the same thing, but worse, that participants in the experimental group are likely to respond on a questionnaire in a way influenced by what they have just been taught. Students taught that bullying is bad will probably respond as the experimenters hope on a questionnaire. But will they actually behave differently with regard to bullying? Perhaps, but it is also quite possible that they are only parroting what they were just taught.

To determine ESSA ratings, we’d emphasize indicators we call MOOSES: Measureable, Observable, Objective Social Emotional Skills. MOOSES are quantifiable measures that can be observed in the wild (i.e., the school) objectively, ideally on routinely collected data unlikely to change just because staff or students know there is an experiment going on. For example, reports of disciplinary referrals, suspensions, and expulsions would be indicators of one type of social-emotional learning. Reports of fighting or bullying incidents could be MOOSES indicators.

Another category of MOOSES indicators would include behavioral observations by observers who are blind to experimental/control conditions, or observations of students in structured situations. Intergroup relations could be measured by watching who students play with during recess, for example. Or, if a SEL program focuses on building cooperative behavior, students could be placed in a cooperative activity and observed as they interact and solve problems together.

Self-report measures might serve as MOOSES indicators if they ask about behaviors or attitudes independent of the treatment students received. For instance, if students received a mindfulness intervention in which they were taught to focus on and regulate their own thoughts and feelings, then measures of self-reported or peer-reported prosocial behaviors or attitudes may not be an instance of parroting, because prosocial behavior was not the content of the intervention.

Social-emotional learning is clearly taking on an increasingly important role in school practice, and it is becoming more important in evidence-based reform as well. But reviewers will have to use conservative and rigorous approaches to evaluating SEL outcomes, as we do in evaluating achievement outcomes, if we want to ensure that SEL can be meaningfully incorporated in the ESSA evidence framework. We admit that this will be difficult and that we don’t have all the answers, but we also maintain that there should be some effort to focus on objective measures in reviewing SEL outcomes for ESSA.

This blog is sponsored by the Laura and John Arnold Foundation

Keep Up the Good Work (To Keep Up the Good Outcomes)

I just read an outstanding study that contains a hard but crucially important lesson. The study, by Woodbridge et al. (2014), evaluated a behavior management program for students with behavior problems. The program, First Step to Success, has been successfully evaluated many times. In the Woodbridge et al. study, 200 children in grades 1 to 3 with serious behavior problems were randomly assigned to experimental or control groups. On behavior and achievement measures, students in the experimental group scored much higher, with effect sizes of +0.44 to +0.87. Very impressive.

The researchers came back a year later to see if the outcomes were still there. Despite the substantial impacts seen at posttest, none of three prosocial/adaptive behavior measures, only one of three problem/maladaptive behaviors, and none of four academic achievement measures showed positive outcomes.

These findings were distressing to the researchers, but they contain a message. In this study, students passed from teachers who had been trained in the First Step method to teachers who had not. The treatment is well-established and inexpensive. Why should it ever be seen as a one-year intervention with a follow-up? Instead, imagine that all teachers in the school learned the program and all continued to implement it for many years. In this circumstance, it would be highly likely that the first-year positive impacts would be sustained and most likely improved over time.

Follow-up assessments are always interesting, and for interventions that are very expensive it may be crucial to demonstrate lasting impacts. But so often in education effective treatments can be maintained for many years, creating more effective school-wide environments and lasting impacts over time. Much as we might like to have one-shot treatments with long-lasting impacts, this does not correspond to the nature of children. The personal, family, or community problems that led children to have problems at a given point in time are likely to lead to problems in the future, too. But the solution is clear. Keep up the good work to keep up the good outcomes!

Good Failure/Bad Failure

Evidence junkies (like me) are reacting to the disappointing news on the evaluation of the Adolescent Behavioral Learning Experience (ABLE), a program implemented at Rikers Island to reduce recidivism among adolescent prisoners. Bottom line: The rigorous independent evaluation of the program failed to find any benefits. What makes this experiment especially interesting is that it is the first U.S. application of social impact bonds. Goldman Sachs put up a $7.2 million loan, and Bloomberg Philanthropies committed to a $6 million loan guarantee. Since the program did not produce the expected outcomes, Goldman Sachs lost $1.2 million.

Ironically, New York City administrators are delighted about the outcome because they do not have to pay for the program. They think they learned a great deal from the experience, for free.

It’s unclear what this will do to the social impact bond movement, currently in its infancy. However, I wanted to extend from this fascinating case to a broader issue in evidence-based reform.

The developers and advocates for the ABLE program who expected positive outcomes turned out to be wrong, at least in this implementation. The investors were wrong in expecting to make a profit. But I’d argue that they are all better off because of this experience, just as the N.Y.C. administrators said.

The distinction I want to make is between wrong and wrong-headed. Wrong, as I’m defining it in this context, means that a given outcome was not achieved, but it was entirely reasonable to expect that it might have been achieved. In contrast, wrong-headed means that not only was the desired outcome not achieved, but it was extremely unlikely that it could have been achieved. In many cases, a key component of wrong-headed actions is that the actor does not even know whether the action was effective or ineffective, right or wrong, and therefore continues with the same or similar actions indefinitely.

Wrong, I’d argue, is an honorable and useful outcome. In a recent interview, former White House advisor Gene Sperling noted that when a few cancer drugs fail to cure cancer, you don’t close down NIH. Instead, you take that information and use it to continue the research and development process. “Wrong,” in this view, can be defined as “good failure,” because it is a step on the path to progress.

“Wrong-headed,” on the other hand, is “bad failure.” When you do something wrong-headed, you learn nothing, or you learn the wrong lessons. Wrong-headed decisions tend to lead to more wrong-headed decisions, as you have no systematic guide to what is working and what is not.

The issue of wrong vs. wrong-headed comes up in the current discussions in Congress about continuing the Investing in Innovation (i3) program. By now, committees in both the House and the Senate have recommended ending i3. But this would be the very essence of wrong-headed policy. Sure, it is probable that many i3 programs funded so far will fail to make a difference in achievement, or will fail to go to scale. This just means that these programs have not yet found success. Some of these may still have evidence of promise, and some will not. However, all i3 programs are rigorously evaluated, so we will know a lot about which worked, which did not, and which still seem promising even if they did not work this time. That’s huge progress. The programs that are already showing success can have immediate impact in hundreds or thousands of schools while others greatly enrich understanding of what needs to be done.

Abandoning i3, in contrast, would be wrong-headed, a sure path to bad failure. A tiny slice of education funding, i3 tells us what works and what does not, so we can continually move towards effective strategies and policies. Without i3 and other research and development investments, education policy is just guesswork, and it gets no smarter over time.

No one can honestly argue that American education is as successful as it should be. Our kids, our economy, and our society deserve much better. Policies that seek a mixture of proven success and “good failure” will get us to solid advances in educational practice and policy. Abandoning or cutting programs like i3 is not just wrong. It’s wrong-headed.

School Reform Is Empty Without Love and Support

2014-08-28-Hp45Image.jpg

In a recent article in The New York Times, David Kirp wrote in opposition to “business models” being applied to education: Assessments, accountability, charters, vouchers, technology, and competition. In place of these, he notes that truly effective programs that have stood the test of time focus not on charters or vouchers or technology alone, but on building strong personal bonds between teachers and students. He gives as examples our Success for All modelDiplomas NowBig Brothers/Big Sisters, and YouthBuild. All of these help teachers, administrators, and community members to build positive, caring environments for children in high-poverty elementary and secondary schools. At the same time, all have been rigorously evaluated and found to improve such key outcomes as achievement and high school graduation.

Kirp is not arguing that approaches emphasizing charters or technology cannot work. However, his point is that unless they also intentionally improve teacher-student bonds, they are unlikely to make a lasting difference. We have enough evidence today to indicate that charters and technology are not magic, and that it matters a great deal what programs and practices they incorporate. I believe that Kirp is right in identifying teacher-student relationships as a key component of effective and sustainable reform. Further, building positive teacher-student relationships is valuable in itself.

Children are social beings who want and need to be loved and supported, and want to please teachers who give them love and support. Love and support are not enough to ensure that students can read or succeed in math, science, or other subjects. Teaching skill and effective practices and programs are also essential. One expression of love and support is the use of exciting, engaging teaching methods and communicating high expectations for all students. But even the best of programs and practices are empty and futile if they do not flow from love and support for children. No new form of governance, no new technology, no program of any kind can make a lasting difference without love and support at their core.

Universal Preschool: Use Innovation and Evidence to Make it Effective

In his recent State of the Union Address, President Obama proposed to “make high-quality preschool available to every child in America.” He referred to research that has demonstrated long term positive effects of attending high-quality preschool programs. President Obama’s support has excited the early childhood community. Who could be opposed to expanding high-quality preschool opportunities? Yet this begs the question: What does “high-quality” mean in practice?

“High-quality” preschools are often defined by educators and economists alike as ones in which teachers are adequately paid, facilities are adequate, and the ratio of staff to children is low. These are indeed important elements of quality and they are serious problems, as preschool educators are often very poorly paid, poorly educated themselves, and lack decent facilities. The low salaries received by preschool teachers leads to a high turnover rate, which also reduces quality. So ensuring universal access to high-quality preschools when many current preschoolers are already struggling with quality and funding issues will be a heavy lift.

Leaving aside money issues, however, there is an important question about how preschool programs should be structured. There is lots of research showing the benefits of high-quality preschool in comparison to no preschool (as in the famous Perry Preschool and Abecedarian programs). However, there is far less research showing different benefits of different preschool approaches.

The Preschool Curriculum Effectiveness Research initiative compared a number of promising approaches to each other and to groups using standard preschool teaching methods. The results are summarized in a review on the Best Evidence Encyclopedia. By the end of kindergarten, only a few of the programs showed child outcomes superior to those achieved by other programs. Structured programs that had a very strong focus on language and emergent literacy, giving children many opportunities to use language to work together, solve challenges, and develop positive relationships with each other, had the best outcomes for children.

Technology has so far played a modest role in early childhood education, but this may change as multimedia devices (such as interactive whiteboards) become more commonly used. Technology offers opportunities for teachers to enhance language development by engaging children with brief content that helps them understand how the world works. For example, children learning about health can see videos on how the body works and can be provided with video models of how to stay safe and healthy. Children can make choices and manipulate pictures and videos representing objects and processes. Further, classroom technology allows for linkages with the home, as parents increasingly have computers, DVDs, and other media available. Children can be shown exciting content in school and then take home DVDs or link electronically to specific materials that closely align with the content they learned that day. These electronic activities can be designed to be done with parents and children together, and can then inform parents about what children are learning in school. Also, in high-poverty homes children often have few if any books. Existing DVD or internet technologies can provide children with access to appropriate literature, which can be read to them by narrators or by their parents or older siblings.

Of course, technology will not replace the majority of early childhood teaching. Young children still need to manipulate real objects and learn to work with each other, sing songs, develop coordination and creativity, and practice appropriate behaviors. However, technology may add the capacity for teachers to show anything they want to their children and to link to the home in ways that have not been possible in the past, and this may result in enhanced learning at this critical age.

Expanding preschool access is a terrific idea, but it will take a lot of money and a long time to put into place. The possibility that it may take place should motivate immediate investments in innovation and evaluation, to develop new ways of ensuring that early education leads to enhanced preparation for success, especially for disadvantaged children.

Preschool quality should not just be seen as a question of per-pupil cost. Preschool educators and children need innovative, proven models that use modern teaching strategies and technologies that are appropriate to the developmental needs of four-year-olds. Innovation and research is needed to show the way as we head toward universal preschool.

Classrooms Need More Pizzazz

On a recent trip to London, I visited Cayley Primary School, a high-poverty elementary school that has been using our Success for All* whole-school reform approach for several years. The principal, Lissa Samuel, has been at this same school for many years before and after it adopted Success for All. She is proud of the achievement gains, which include a jump from 30% to 80% of students passing sixth-grade reading assessments. During our conversation, though, she talked more about how disciplinary problems, fights, and stealing had completely disappeared. Success for All has very good approaches to classroom management and social-emotional learning, and Ms. Samuel thought these had helped. But even more powerful, she thought, was the effect of success itself. Kids who feel confident, engaged, and motivated to learn do not act out.

The importance of this observation, which I’ve heard in many, many schools, is profound. Especially at the policy level, I often encounter a belief that the path to improving outcomes on a broad scale is to use test-based accountability that force teachers to align their instruction with desired outcomes. If students are bored or resistant, then teachers should use effective classroom management methods that keep them in control.

Teachers do need a deep understanding of classroom management methods designed to prevent behavior problems, and then they need to be ready with effective responses if students misbehave despite good preventive efforts. Yet using classroom management methods to get students to attend to boring lessons is shoveling against the tide. The key ingredient in effective lessons isn’t alignment, it’s pizzazz: excitement, engagement, challenge.

How do you create pizzazz? Well-structured cooperative learning helps to engage students with each other in jointly learning context. Stimulating video content can add to excitement and understanding. Hands-on experimentation helps a lot when appropriate, as does competition between teams or against the clock.

Cayley Primary was full of pizzazz. Its mostly Bangladeshi students worked eagerly in four-member teams. They took turns reading to each other and helping each other with difficult words. Their teachers called on “random reporters” to represent their teams, and teammates prepared each other, not knowing which of them might be randomly chosen to play this role. Brief, humorous videos introduced letter sounds and sound-blending strategies to first graders. Throughout the school, students were invariably kind and helpful to each other. An observer who did not know the history might think that classroom management was not necessary in such a school, but it was proactive use of pizzazz that got it to where it is, and makes it all look easy.

Classroom management strategies matter, of course, but pizzazz matters more. Motivated, engaged, challenged, and successful students are well-behaved, not because they’ve been threatened but because they are too busy engaged in learning to misbehave. The goal of classroom management is not quiet classrooms, it’s productive students. Using pizzazz to motivate and engage kids in learning valued content is the way to manage classrooms toward accomplishing the real goals of education.

*Robert Slavin is Chairman of the Board of the Success for All Foundation

For the latest on evidence-based education, follow me on twitter: @RobertSlavin