Hattie's Rankings


RankDomainInfluenced
1StudentSelf-report grades1.44
2StudentPiagetian programs1.28
3TeachingProviding formative evaluation0.90
4TeacherMicro teaching0.88
5SchoolAcceleration0.88
6SchoolClassroom behavioral0.80
7TeachingComprehensive interventions for learning disabled students0.77
8TeacherTeacher clarity0.75
9TeachingReciprocal teaching0.74
10TeachingFeedback0.73
11TeacherTeacher-student relationships0.72
12TeachingSpaced vs. mass practice0.71
13TeachingMeta-cognitive strategies0.69
14StudentPrior achievement0.67
15CurriculaVocabulary programs0.67
16CurriculaRepeated reading programs0.67
17CurriculaCreativity programs0.65
18TeachingSelf-verbalization/self-questioning0.64
19TeacherProfessional development0.62
20TeachingProblem-solving teaching0.61
21TeacherNot Labeling students0.61
22CurriculaPhonics instruction0.60
23TeachingTeaching strategies0.60
24TeachingCooperative vs. individualistic learning0.59
25TeachingStudy skills0.59
26TeachingDirect Instruction0.59
27CurriculaTactile stimulation programs0.58
28CurriculaComprehension programs0.58
29TeachingMastery learning0.58
30TeachingWorked examples0.57
31HomeHome environment0.57
32HomeSocioeconomic status0.57
33TeachingConcept mapping0.57
34TeachingGoals0.56
35CurriculaVisual-perception programs0.55
36TeachingPeer tutoring0.55
37TeachingCooperative vs. competitive learning0.54
38StudentPre-term birth weight0.54
39SchoolClassroom cohesion0.53
40TeachingKeller's PIS0.53
41SchoolPeer influences0.53
42SchoolClassroom management0.52
43CurriculaOutdoor/adventure Programs0.52
44TeachingInteractive video methods0.52
45HomeParental involvement0.51
46CurriculaPlay programs0.50
47CurriculaSecond/third chance programs0.50
48SchoolSmall group learning0.49
49StudentConcentration/persistence/engagement0.48
50SchoolSchool effects0.48
51StudentMotivation0.48
52StudentEarly intervention0.47
53TeachingQuestioning0.46
54CurriculaMathematics0.45
55StudentPreschool programs0.45
56TeacherQuality of Teaching0.44
57CurriculaWriting Programs0.44
58TeacherExpectations0.43
59SchoolSchool size0.43
60StudentSelf-concept0.43
61TeachingBehavioral organizers/Adjunct questions0.41
62TeachingMatching style of learning0.41
63TeachingCooperative learning0.41
64CurriculaScience0.40
65CurriculaSocial skills programs0.40
66StudentReducing anxiety0.40
67CurriculaIntegrated Curriculum Programs0.39
68SchoolEnrichment0.39
69CurriculaCareer Interventions0.38
70TeachingTime on Task0.38
71TeachingComputer assisted instruction0.37
72TeachingAdjunct aids0.37
73CurriculaBilingual programs0.37
74SchoolPrincipals/ school leaders0.36
75StudentAttitude to mathematics/science0.36
76CurriculaExposure to reading0.36
77CurriculaDrama/Arts programs0.35
78StudentCreativity0.35
79TeachingFrequent/effects of testing0.34
80SchoolDecreasing disruptive behavior0.34
81StudentDrugs0.33
82TeachingSimulations0.33
83TeachingInductive teaching0.33
84StudentPositive view of own ethnicity0.32
85TeacherTeacher effects0.32
86TeachingInquiry based teaching0.31
87SchoolAbility grouping for gifted Students0.30
88TeachingHomework0.29
89HomeHome visiting0.29
90StudentExercise/relaxation0.28
91SchoolDesegregation0.28
92SchoolMainstreaming0.28
93CurriculaUse of calculators0.27
94CurriculaValues/moral education programs0.24
95TeachingProgrammed instruction0.24
96TeachingSpecial college programs0.24
97TeachingCompetitive vs. individualistic learning0.24
98SchoolSummer school0.23
99SchoolFinances0.23
100TeachingIndividualized instruction0.23
101SchoolReligious Schools0.23
102StudentLack of Illness0.23
103TeachingTeaching test taking0.22
104TeachingVisual/audio-visual methods0.22
105TeachingComprehensive teaching reforms0.22
106SchoolClass size0.21
107SchoolCharter Schools0.20
108TeachingAptitude/treatment interactions0.19
109StudentPersonality0.19
110TeachingLearning hierarchies0.19
111TeachingCo-team teaching0.19
112TeachingWeb-based learning0.18
113HomeFamily structure0.17
114CurriculaExtra-curricular programs0.17
115TeachingTeacher immediacy0.16
116SchoolWithin class grouping0.16
117TeachingHome-school programs0.16
118TeachingProblem-based learning0.15
119CurriculaSentence combining programs0.15
120TeachingMentoring0.15
121SchoolAbility grouping0.12
122StudentGender0.12
123StudentDiet0.12
124TeacherTeacher training0.11
125TeacherTeacher subject matter knowledge0.09
126TeachingDistance Education0.09
127SchoolOut of school curricula experiences0.09
128CurriculaPerceptual-Motor programs0.08
129CurriculaWhole language0.06
130SchoolCollege halls of residence0.05
131SchoolMulti-grade/age classes0.04
132TeachingStudent control over learning0.04
133SchoolOpen vs. traditional0.01
134SchoolSummer vacation-0.09
135HomeWelfare policies-0.12
136SchoolRetention-0.16
137HomeTelevision-0.18
138SchoolMobility-0.34

Note: in Hattie’s 2016 publication, Learning strategies: a synthesis and conceptual model - he argued against ranking (p. 9)!

"There is much debate about the optimal strategies of learning, and indeed we identified >400 terms used to describe these strategies. Our initial aim was to rank the various strategies in terms of their effectiveness but this soon was abandoned. There was too much variability in the effectiveness of most strategies depending on when they were used during the learning process."

Hattie confirms this in his 2018 podcast with Ollie Lovell. Hattie said,
"it worked then it got misleading so I stopped it"
Hattie's new way of ranking with Corwin (March 2019):




Other Peer Reviews on Hattie's Rankings:

Wecker et al. (2017, p. 34)
"Hattie's ranking suggests that the 138 influencing factors are to be seen as basically alternative options for action, from which the most effective one is to be selected, for example, during lesson design. However, this would require that the influencing factors are in the first place real options for action and that their effect sizes were determined in relation to the same "baseline", ie. that the control conditions in the primary studies in all meta-analyzes included in the first step are to be regarded as an equivalent benchmark. A look at the ranking shows that this is neither the case nor is it possible at all, since the influencing factors belong to completely different types: On the one hand, one finds not only institutional framework conditions such as school size. As well as personal traits such as the self-concept. Such as those classified under "direct instruction". Effect sizes usually relate to a comparison with some form of "traditional" instruction... 
It is doubtful that teachers would ever usefully choose between direct instruction and distributed practice rather than the optimal combination of both to decide. If this is the case, the assessment of the relative effectiveness of these options cannot be based in one case on a comparison with an instructional "standard condition" and on the other hand on a comparison with another design variant of the same instruction form."
Prof Adrian Simpson (2017) is similarly critical of rankings in his detailed analysis, 
"The numerical summaries used to develop the toolkit (or the alternative ‘barometer of influences’: Hattie 2009) are not a measure of educational impact because larger numbers produced from this process are not indicative of larger educational impact. 
Instead, areas which rank highly in Marzano (1998), Hattie (2009) and Higgins et al. (2013) are those in which researchers can design more sensitive experiments. 
As such, using these ranked meta-meta-analyses to drive educational policy is misguided."
Schulmeister & Loviscach (2014),
"If one corrects the errors mentioned above, list positions take big leaps up or down.Even more concerning is the absurd precision this ranking conveys. It only shows the averages of effect sizes but not their considerable variation within every group formed by Hattie and even more so within every individual meta-analysis...
To think that didactics can be presented as a clear ranking order of effect sizes. It is a dangerous illusion. To an extreme degree, the effect of a specific intervention depends on the circumstances. Focusing on the mean effect sizes and ignoring their considerable variations and condensing the data to a seeming exact ranking order, Hattie pulls the wool over his audience’s eyes."
Dr. Jim Thornton (2018),
"To a medical researcher, it seems bonkers that Hattie combines all studies of the same intervention into a single effect size. Why should “sitting in rows”, for example, have the same effect on primary children as on university students, on maths as on art teaching, on behaviour outcomes as on knowledge outcomes? In medicine it would be like combining trials of steroids to treat rheumatoid arthritis, effective, with trials of steroids to treat pneumonia, harmful, and concluding that steroids have no effect! I keep expecting someone to tell me I’ve misread Hattie."
Claes Nilholm (2013) used 2 detailed examples to draw his conclusion (see problem-based learning and effect size),
"...Hattie draws far too far-reaching conclusions... he presents summative measurements in tabular form, and factors are ranked according to their significance. Unfortunately, SKL (Swedish Municipalities and County Council) ,puts great emphasis on these summing measures. I would like to say that the summaries, if used to guide teachers' work, can give utter incorrect implications" (p. 3).
McKnight & Whitburn (2018),
"Teachers will pick and choose from the list of “what works” that forms Visible Learning, even as their guts tell them externally mandated, evidence-based practices will not necessarily work for them (Biesta, 2007)" (p. 16).

No comments:

Post a Comment