The Top Ten Ways that Educators Make Bad, Large-Scale Programmatic Decisions (Part I of II)

The Hazards of ESEA/ESSA’s Freedom and Flexibility at the State and Local Levels

Dear Colleagues,


As we plan in earnest for the full implementation of the Elementary and Secondary Education Act/Every Student Succeeds Act (ESEA/ESSA), the following “truths” have become self-evident:

  • There is going to be an incredible amount of variability—relative to school and teacher effectiveness, student standards and assessments, multi-tiered academic and behavioral interventions, and school “success” and student “proficiency”—across our states and districts . . . than ever before.
  • States, districts, and schools will be more responsible for selecting their own approaches relative to curriculum, instruction, assessment, intervention, and evaluation . . . than ever before.
And because of this:
  • The impact of frequently-changing superintendents and school administrators, inequitable school staffing and teacher shortages, and a focus on the type of school (e.g., public versus charter schools) rather than the quality of a school . . . will result in more student gaps and failures than ever before.

This is a good thing and a bad thing.

It is a good thing, because many of the U.S. Department of Education’s preferred and pushed No Child Left Behind frameworks (e.g., School Improvement, PBIS, RtI/MTSS, Reading First/Literacy) did not work, and yet, were funded (and continue to be funded relative to PBIS and MTSS) by billions (that’s with a “B”) of your taxpayer dollars.

In essence, across two administrations—one Republican (Bush II) and one Democratic (Obama)—the U.S. Department of Education failed us, and—in their arrogance that they knew best how to educate all students—they left us all behind.

It is a bad thing because, as alluded to above, our districts and schools—who have enough to do with the day-to-day education of our students—will need to make more curriculum, instruction, intervention, and evaluation decisions than ever before.

And, I fear, they are not prepared to make these decisions in scientifically, psychometrically, methodologically, and contextually-sound ways.

It’s not that our educators across the country are trying to be ineffective.

It is just that they do not have the time, people, and resources to be MORE effective, and they often do not know what they do not know.

That is, some educators do not have the sophisticated scientific, psychometric, or methodological in-house expertise to make some critical decisions. And so, they go out-house to the “experts”—some of whom are expert more in marketing themselves, than in recommending or providing true evidence-based, sustainable outcomes.

The “Top Ten” Ways that Educators Make Bad Programmatic Decisions

As a practitioner who has worked in the schools, at two Research I universities, within a state department of education, and as a consultant in every state in the country over 35+ years, I find that educators make important, large-scale (i.e., at the state, district, or school levels) programmatic (curriculum, instruction, intervention, or evaluation) decisions in ways that are incredibly flawed.

And these flawed decisions waste time, money, resources, and energy . . . and they often progressively result in frustrated and resistant staff, and disengaged and negatively-impacted students.

I bring these flaws to your consciousness so that we can all recognize and eliminate these flaws in the future.

At the same time, I fully recognize that, sometimes, a flawed decision actually works. (Remember that even low-probability-of-success events sometimes are successful! Someone is going to win the lottery—against all odds!)

PLEASE NOTE ALSO that the “personal labels” associated with each Reason below are not intended to offend anyone . . . I just want you to think about the implications—to students, colleagues, and schools—of each flaw.

And so: Here—with brief commentary—are the “David Letterman Top Ten Reasons” why educators make (sometimes flawed) programmatic decisions.

Reason #1: “Because I Know Best” (The “Autocrat”)

The Flaw: These are bad decisions made by the educators “in control” of the District or its schools. They are autocratic leaders who make largely unilateral decisions (a) because they “know everything;” (b) they “can;” (c) it’s politically expedient; (d) because of who “has their ear;” or (e) due to some related factor or influence.

Reason #2: “Because My Colleague Says It Works” (The “Daydream Believer”)

The Flaw: These are bad decisions made by educators who depend on the testimonials of “trusted colleagues” who attest that a specific program “works” because either (a) they have tried it; or (b) one of their “trusted colleagues” has recommended it.

Often, these “trusted colleagues” have not themselves objectively validated the program’s efficacy—based on sound research or actual implementation. More critically, for the educator making the decision for their District or school, it appears that “blind trust” has “carried the day.”

Reason #3: “Because It’s On-Line” (The “Connected One”)

The Flaw: These are bad decisions made by educators who believe that anything posted on-line (by the media, marketers, actual researchers, or others) is true—simply because it is on-line. They assume that when something goes “viral”—the “likes,” “retweets,” and “shares” attest to the validity of the program.

Reason #4: “Because It’s Free” (The “Bargain Basement Boss”)

The Flaw: These are bad decisions made by educators who embrace a program because it is provided “at no cost” to the District or its schools. They fail to understand that a “free” program that does not work or actually impedes or implodes the progress of the District has significant negative costs—to students, staff, and schools.

As alluded to above, the U.S. Department of Education (as well as many state departments of education, and many non-profit Foundations) have “pushed” their preferred programs down to districts and schools by providing “free” grants and training. Given restricted and tight budgets, many “accept” the training—assuming that these “experts” know what they are doing (see Reasons #7 and 8).

Relative to the departments of education, these programs are not free—they use your taxpayer dollars. Relative to the Foundations, we need to be aware that social, political, or other agendas may be embedded in the grants. Relative to all, we need to independently validate the programs before we “cash the checks.”

Critically: We need to get away from the “If it’s free, it’s for me” mentality in education.

Reason #5: “Because the Committee Recommended It” (The “Failed Consensus-Builder”)

The Flaw: These are bad decisions made by the educators “in control” of the District or its schools based on the recommendations of a formally-constituted or informally-constituted committee or task force.

The “benign” flaw here is when the Educator embraces the programmatic recommendation assuming that the committee has done its “due diligence,” and that “it knows best.”

The “damaging” flaw is when the Educator accepts a committee’s recommendation (a) knowing that it is flawed, but (b) is afraid of the functional or political consequences of overturning it.

Reason #6: “Because a National Expert Recommended It” (The “Groupie”)

The Flaw: These are bad decisions made by educators because they were recommended by a “national expert”—an author, presenter, program developer, or consultant—who is either working independently, or working for a company, educational resource center, state or federal department of education, etc.

While, clearly, some experts are “true” experts, well-intended, and their recommendations are sound, decision-making educators still need to independently validate their recommendations, and—even if sound—determine whether they can be applied to the students, staff, and schools in question.

Other “experts”— who advocate untested, unreliable, non-transferable, or invalid programs—run the spectrum from those who truly believe that they are doing good, to those who are out to do good only for themselves.

The bottom line? Decision-makers are accountable for their decisions, NOT the recommendations from an expert.

Reason #7: "Because It’s Developed by a Non-Profit" (The “Do-Gooder”)

The Flaw: These are bad decisions made by educators who believe that programs developed and disseminated by non-profit agencies, organizations, or foundations “must be good” because they come from groups with pure, altruistic, or selfless intentions or motivations.

While this may be true, this is not necessarily true. The reality, with all due respect to my non-profit colleagues, is that non-profits are businesses that need to make money to stay in business.

Simplistically, the differences between for-profits and non-profits are that the latter are legally-bound to use their profits in specific, IRS-controlled ways, and they are legally-restricted (depending on their categorization) to certain, limited (or NO) political/lobbying activities.

And so . . . a program from a non-profit should be objectively analyzed and vetted in the same way as a program from a for-profit organization or company.

Significantly, I do not want educators to become cynical relative to non-profit agencies or foundations. However, I do want them to remain vigilant and accountable to the students and staff whom they serve.

Reason #8: "Because It’s Federally- or State-Recommended" (The “Enabler”)

Studies report that up to 35% of school districts nationwide implement department of education programs without any independent analysis or asking any questions.

The Flaw: These are bad decisions made by educators because they believe that their federal and state departments of education (a) are working in everyone’s best educational interests; and that (b) they have objectively analyzed the outcome data of the programs they are recommending (using their own required “Gold Standard” approach); (c) they would not recommend a program that does not have a high probability of success; and (d) they have “no dog in the fight.”

Moreover, some educators make these programmatic decisions believing that adopting a federally- or state-recommended program is a safe bet—or, at least, that they will "be protected" regardless of the outcomes.

Wake up, America!

Let’s remember: That only a third of the districts adopting one of the billion-dollar-supported federal School Improvement models (under No Child Left Behind) had any positive results . . . And that the districts—that did not make progress under the federally-required options—were still held accountable for their failing outcomes.

And, remember that: Less than 10 years ago . . . the Bush administration’s billion-dollar Reading First Program was unfunded by Congress because officials in the U.S. Department of Education were found guilty of (a) favoring specific reading programs, while eliminating a number of exceptional programs from consideration; (b) “stacking” the grant review panels with members biased toward the favored programs; and (c) changing state grant proposals after they were reviewed so that they could be funded with the favored programs.

And while you can say that “this is all in the past” . . . it is not. The federal law (ESEA/ESSA and IDEA) discusses or requires districts to provide “positive behavioral interventions and supports” and “multi-tiered systems of supports” (written in these laws in lower case and without acronyms).

And yet, the U.S. Department of Education, through most state departments of education, continues to singularly advocate their federally-funded (for years) and preferred PBIS and MTSS frameworks, respectively. In fact, the Department-funded national Technical Assistance Centers for these frameworks continually misquote the federal laws to make it appear that their UPPER CASE approaches are legally mandated (through their lower case appearances in federal law).

Relative to Reason #8, the “Enabling” educator making decisions in this area often is simply ignoring the “elephant in the room.”

Reason #9: "Because It’s Federally- or State-Mandated" (The “Abdicator”)

The Flaw: These are bad decisions made by educators who passively comply with federal and state department of education programs that represent these departments’ interpretation or operationalization of federal (e.g., ESEA, IDEA) and/or state laws or statutes.

Instead of abdicating and implementing programs that everyone knows will not positively impact students, staff, and schools, these educators should (a) question the programs and their efficacy; (b) recommend a proven, alternative program that has a higher probability of attaining the desired results; and/or (c) request and defend the need for a waiver.

Unfortunately, educators sometimes abdicate their responsibilities in this area because they feel they can deflect the responsibility for a failed program “because it was mandated.” At other times, they comply for fear of retribution or retaliation—for example, receiving undue scrutiny or sanctions during state department of education audits or compliance visits.

While this is not fair. . . it is real.

Reason #10: "Because It’s “Research-based” (The “Mad Scientist”)

The Flaw: These are bad decisions made by educators who review the “research” that appears to support a specific program, but who do not understand the difference between (a) methodologically sound versus unsound research; (b) research that measures perceptions versus research that objectively measures definitive and replicable outcomes; and (c) research that “works” in a perfectly controlled vacuum versus research that truly works in the real world.

See . . . it’s not about a program’s association with research. It is about the reliability, validity, and generalizability of the research.

More specifically, some “research” is done (a) by convenience; (b) with small, non-representative, and non-random samples; (c) without comparisons to matched “control groups;” and (d) in scientifically unsound ways.

Here, educational decision-makers need to understand (or hire other professionals who understand) that this “research” has a low probability of succeeding when scaled-up in their district.

Moreover, some “research” has not been independently, objectively, or “blindly” reviewed by three or more experts in the field (as when someone publishes their work in a “refereed” professional journal).

Relative to this latter point, educators need to understand that—even when studies are published in a refereed journal—that some research is published because article submissions are low, and the publisher cannot “stop the presses.” Said a different way, journals need to be published in order to attract annual subscriptions and stay in business.

The ultimate point here is that educational decision-makers should not be “Mad Scientists”—assuming that “being published” correlates with “good research.” They need to be “Discriminating Consumers” who recognize sound versus unsound research, and the importance of choosing only the former for programmatic implementation.

Summary and Coming Attractions

As noted in the Introduction, the Elementary and Secondary Education Act/Every Student Succeeds Act (ESEA/ESSA) gives states, districts, and schools more freedom, flexibility, and responsibility for selecting their own approaches to curriculum, instruction, assessment, intervention, and evaluation.

By describing the “Top Ten Ways that Educators Make Bad, Large-Scale Programmatic Decisions,” the hope is that school, district, state, and other educational leaders avoid these hazards, and make their decisions in empirically and functionally sound ways.

There are some great, evidence-based approaches available to schools across the country. Some of them, however, are getting lost in the marketing, promotion, publishing, and competitive “noise” that invades our professional lives each day.

Mark Millar (who, believe it or not, writes comic books for a living) said: “Organizations who win, think deeply, choose wisely, and act decisively.”

Educational decision-makers need to think deeply about the true needs of their students, staff, and schools—identifying what is working and needs to be maintained, as well as the gaps that exist, why they exist, and how they are going to be closed.

They then need to choose wisely from the services, supports, programs, and interventions that are available, and that have demonstrated their ability—from research to practice—to close the gaps. Here, they need to avoid the Top Ten approaches above—sometimes choosing to take “the road less traveled.”

Finally, they need to act decisively—making sure that the training, resources, and mentoring needed for successful implementation is available and sustained.

If we do all this, the “permission” granted by ESEA/ESSA will result in the “promise” of better academic and social, emotional, and behavioral outcomes for all students.

Once again, the goal of this Blog was not to stereotype, offend, or blame anyone relative to their educational decision-making processes. Instead, the goal was to recognize, characterize, and memorialize some of the ways that poor decisions are made—and to emphasize that these decisions have real and sometimes long-lasting impact on students, staff, and schools.

In Part II of this Blog message (in two weeks), we will continue to look at ways to understand sound versus unsound research. While I rarely publicize the topic of my next Blog, I will today.

My next Blog is tentatively titled, Hattie’s Meta-Analysis Madness: The Method is Missing. Stay tuned.

As always, I look forward to your comments. . . whether on-line or via e-mail.

And—with the new school year now upon us: If I can help you in any of the school improvement, school discipline and behavioral intervention, or multi-tiered service and support areas where I specialize, please do not hesitate to contact me.

I am always happy to provide a free one-hour consultation conference call to help you clarify your needs and directions on behalf of your students, staff/colleagues, school(s), and district.

Welcome back! Make it a GREAT YEAR !!!