NORTHWESTERN UNIVERSITY
Were They Ready To Learn? Short- and Long-Term Effects of Ready To Learn Media on Young Children’s Literacy
A DISSERTATION
SUBMITTED TO THE GRADUATE SCHOOL IN PARTIAL FULFILLMENT OF THE REQUIREMENTS
for the degree
DOCTOR OF PHILOSOPHY
Field of Media, Technology & Society
By
Lisa B. Hurwitz
EVANSTON, ILLINOIS
June 2017 2
© Copyright by Lisa B. Hurwitz 2017
All Rights Reserved 3
Abstract
The U.S. Department of Education’s Ready To Learn (RTL) initiative funds (a) mass media and related community outreach intended to promote school readiness among all children—especially at-risk populations—, and (b) formative and summative research by third party evaluators (Michael Cohen Group, 2012). Today, the majority of American preschoolers have consumed RTL media, and RTL has become one of the largest funders of educational media research (Corporation for Public Broadcasting [CPB] & Public Broadcasting Service
[PBS], 2011). This dissertation primarily aimed to assess RTL’s short- and long-term effectiveness in promoting school readiness, specifically foundational literacy skills. As a secondary goal, this dissertation also sought to shed insight into methodological and theoretical issues concerning children’s learning from media more broadly.
Article 1 provides a meta-analytic review of the accumulated research on the short-term effectiveness of RTL’s literacy-themed media. Results indicate that RTL media resulted in small but positive impacts on children’s early literacy skills, comparable to comprehensive early childcare programs such as Head Start (Kay & Pennucci, 2014), but considerably smaller than targeted nonmediated literacy interventions, such as in-person phonics instruction (Langenberg et al., 2000).
Article 2 outlines how a developmental researcher might re-recruit families who participated in research studies, such as early childhood RTL evaluations, for later research. It highlights newer technological tools that facilitate locating participants who moved between waves of data collection and scheduling them for follow-up sessions (e.g., people-centric search engines built on algorithms that scrape the Internet for publicly available contact information). 4
Researchers conducting non-prospective longitudinal, prospective longitudinal, or even cross- sectional research could benefit from many of these strategies.
Article 3 builds on Articles 1 and 2. Preschool and kindergarten students who participated in one of the evaluations included in the Article 1 meta-analysis were re-recruited for a follow-up study six years later, using recruitment strategies detailed in Article 2. Children were in preschool and kindergarten during the original intervention, and were in 5th and 6th grade at follow-up testing. Children completed age-appropriate literacy assessments, and their parents provided complementary survey data. Findings suggested the effects of early exposure to RTL media sustained into middle childhood, but only for children who had below and above average literacy skills in early childhood prior to the original intervention. These findings are interpreted in light of the Early Learning Hypothesis, and the Traveling Lens, Capacity, and Differential
Susceptibility to Media Effects Models.
Altogether, these results provide relatively positive support in favor of the continued funding of RTL, and speak to larger theoretical debates concerning children’s learning from media. Continued monitoring is warranted, as the specific foci and execution of the RTL initiative has evolved across grant cycles. 5
Acknowledgements
I want to take a moment to thank the very many wonderful people in my life, starting with my brilliant dissertation committee. I give my deepest thanks to my mentor and doctoral advisor Ellen Wartella, who has taught me so much about this field specifically and life in general, always encouraging me to think of big ideas. Further, I thank Alexis Lauricella for providing so much professional guidance and emotional support over the past five years and for always making the time to help me think through both minor and major decisions. Finally, I thank Kelly Schmitt, without whose support this project would not have been possible. From thoughtfully considering how my findings relate to broader educational policy debates to discussing finer details such as contradictory race-ethnicity reporting across surveys, every interaction with Kelly has helped me to elevate the quality of my work and thinking. And of course I am grateful to Kelly and her team for their careful efforts evaluating PBS KIDS Island in 2010, facilitating a relatively straightforward follow-up effort for Articles 2 and 3.
I also would like to thank the many other Northwestern faculty who have shaped my thinking these past five years. I owe particular and immeasurable thanks to Daniel J. O’Keefe, who provided me hours of private tutoring and feedback in support of Article 1. I also thank
Thomas D. Cook; I never would have thought to do this project had I not taken his excellent program evaluation class.
In addition, I want to acknowledge the gifted undergraduate students, graduate students, and postdoctoral fellows who contributed so much to this project. I am very thankful to Meredith
Ford, Dashia Kwok, Zachary Lochmueller, Francesca Pietrantonio, and Mo Ran, who, perhaps unknowingly, risked their lives driving with me around Chicagoland and collecting data for 6
Article 3, and helped to locate participants for Article 2. I also want to thank Shina Aladé,
Leanne Beaudoin-Ryan, and Heather Montague for accompanying me to collect Article 3 data.
Additional thanks to Francesca and Dashia for conducting reliability coding for Article 1. And a big, big thank you to Megan Olsen for keeping me sane and for spending countless hours helping to recruit participants for Articles 2 and 3. Further thanks to Jacob Schauer for his technical assistance during the Article 1 manuscript revision process. I owe still more gratitude to Shina
Aladé, Leanne Beaudoin-Ryan, and Kelly Sheehan for providing feedback on draft manuscripts sharing my dissertation data. I also am grateful to Michael DeVito, Colin Fitzpatrick, Tommy
Rousse, and Madeline Smith for their input on recruitment strategies for Article 2, and to Emily
Ross for her insight into collecting demographic information from low-income populations, crucial to Articles 2 and 3.
Furthermore, I thank the many individuals beyond the Northwestern community who helped to make this research possible. I am very grateful to Barbara Lovitts and all the current and former RTL evaluators who provided me unpublished reports and other data I needed to complete Article 1. Additionally, I thank the families who took the time to participate in the study described in Article 3; the way these families prioritize education is inspiring. I also give thanks to James Alex Bonus and Lana Karasik for recommending background reading for Article
2, and to Daniel Anderson and Laura Sheridan Duel for their feedback on draft articles.
Beyond the scope of this project specifically, I also would like to acknowledge the rest of my Center on Media and Human Development lab family, Sabrina Connell, Courtney Blackwell,
Drew Cingel, Lillian Yi, Silvia Lovato, Sarah Pila, Bri Hightower, Jabari Evans, Eric Morales, and Nadalyn Bangura, and my surviving cohort comrade Jennifer Ihm. 7
Finally, I would like to thank my family and friends from my life beyond Northwestern for encouraging me in and supporting my decision to pursue higher education, and for volunteering your children to be guinea pigs in my studies over the year. You genuinely were interested in my work in moments where I wanted to discuss it, and allowed me to live vicariously through your lives when I didn’t. Thank you; thank you.
8
Table of Contents
Introduction ...... 10
Article 1: Getting a Read on Ready To Learn Media: A Meta-Analytic Review of Effects on Literacy ...... 17
Article 2: “Are You Sure My Child Participated in that Study?” Recruiting and Re-Recruiting for Developmental Research ...... 50
Article 3: Did We Succeed in “Raising Readers”? Effects of Ready To Learn Early Childhood Literacy Computer Games in Middle Childhood...... 78
Conclusion ...... 108
References ...... 118
Appendix A: Academic Search Terms ...... 143
Appendix B: Mean Effect Sizes and Key Codable Characteristics of Included Evaluations .... 144
9
List of Tables and Figures
Tables
Table 1. Literacy Categories ...... 19
Table 2. Descriptive Statistics for Nominal and Ordinal Moderators ...... 28
Table 3. Descriptive Statistics for Continuous Moderators ...... 30
Table 4. Nominal Moderator Effects ...... 34
Table 5. Continuous and Ordinal Moderator Effects ...... 36
Table 6. Techniques for Locating and Scheduling Participants ...... 61
Table 7. Demographics for Children Who Participated in the Middle Childhood Follow-up ..... 89
Table 8. Child Literacy Outcomes in Middle Childhood ...... 97
Table 9. Parent Ratings of Children’s Academic Performance in Middle Childhood...... 100
Figures
Figure 1. Funnel Plot of Each Study’s Mean Effect Size by Study-Average Standard Errors ..... 42
10
Introduction
In 1994, the U.S. Department of Education, Congress, and the Corporation for Public
Broadcasting launched the Ready To Learn (RTL) initiative in response to two public concerns:
(a) America’s children, particularly certain disadvantaged populations, were entering elementary school ill-prepared to succeed in formal educational environments; and (b) these same children were spending hours watching television in a media environment largely characterized by program-length commercials devoid of learning content (Bryant, Bryant, Mullikin, McCollum, &
Love, 2001). RTL provides funding for educational mass media intended to promote school readiness; outreach initiatives complementing and extending this learning; and supportive implementation and effectiveness research (Bryant et al., 2001; Singer & Singer, 1998). Most
U.S. preschool children have consumed RTL media, and RTL has become one of the largest funders of research into children’s learning from media (CPB & PBS, 2011). Originally, RTL funding was used primarily for television, providing seed money for new properties (e.g., Super
WHY!), as well as added support for legacy properties that existed prior to the launch of the grant
(e.g., Sesame Street; CPB & PBS, 2011). RTL producers also have moved to the forefront of experimenting with delivering lessons via newer media platforms, such as computer games and mobile apps (Michael Cohen Group, 2012).
A handful of narrative literature reviews written or commissioned by grant recipients have examined RTL’s media’s overall effectiveness in promoting early literacy (and to a much lesser extent numeracy), highlighting cases for which evaluators noted positive effects for children who consumed RTL media (e.g., Cohen, Hadley, & Marcial, 2016; CPB & PBS, 2011;
Michael Cohen Group, 2012). Typically, evaluators in the studies included in these reviews 11 assessed effects immediately after children complete one- to three-month interventions in which they were exposed to RTL media. However, in a small number of cases, researchers examined if effects persisted up to a year after initial exposure, generally finding supportive evidence in favor of RTL’s longer-term effects, at least for select literacy outcomes (Linebarger, 2010; Neuman,
Newman, & Dwyer, 2011).
Despite this generally supportive evidence in favor of RTL, accountability evidence for the initiative is incomplete at present. Past reviews typically focused on single grant cycles, rather than looking at RTL’s more holistic impact, and did not assess the consistency of positive effects. It is also unclear the extent to which the authors of those reviews searched for and included research conducted by parties not funded by the grant. Taking these steps to consider and fairly weight all existing evidence is necessary for a rigorous, unbiased review (Borenstein,
Hedges, Higgins, & Rothstein, 2009). It is additionally uncertain whether boosting early learning skills through RTL media would have continued downstream positive impacts on children’s academic performance once they enter formal schooling. Exploring the longer-term impacts of early intervention allows for a more compelling policy argument for or against continued funding, particularly for initiatives explicitly attempting to make children better equipped for formal education (e.g., Barnett, 2013). This dissertation address these gaps in accountability evidence by examining the short- and long-term effects of RTL media on young children’s literacy across two research studies described in three articles.
The general theoretical assumption across all articles is that children can learn from media, provided such media is developmentally appropriate and is intended to convey academic lessons (Hirsh-Pasek et al., 2015). Because the effects of educational media are often relatively 12 small, others may question this assumption (e.g., Torgerson, 2007). However, RTL was launched under the belief that educational media could be leveraged to promote school readiness (Bryant et al., 2001), even though the specific theoretical focus adopted by content producers and evaluators has varied over the course of the grant’s history. Thus, the assumption about educational media’s potential is consistent with the theory of the RTL program.
Coloring the overall positive outlook on the potential of educational media espoused in this dissertation, it is likely that certain content, context, and child variables have the potential to positively or negatively moderate educational media’s effects (Guernsey, 2012). The potential impact of many moderators is considered across the articles in this this dissertation.
The word “content” is sometimes used to refer different form of media such as curriculum-based vs. aggressive fare (Guernsey, 2012). For this project, a looser definition of content is used, referring to media product itself, and all affordances, storylines, and characters encompassed therein. Some space in this dissertation is dedicated to assessing whether certain media properties are more effective than others. Additionally, this dissertation explores whether new media (e.g., computer games, mobile apps) are effective at all or as effective as television, a more well-studied platform (Fisch & Truglio, 2000). On the one hand, new media might be able to promote strong learning by automatically “leveling up” to present children with increasingly challenging but developmentally appropriate content (Grant et al., 2012; McManis & Gunnewig,
2012; Roberts, Chung, & Parks, 2016), fostering in children a sense of self-efficacy as they advance through gaming environments (Educational Development Center & SRI International,
2012; Ronimus & Lyytinen, 2015), and using highlighting and other features to draw children’s attention to educational lessons (Guernsey & Levine, 2015). However, the cognitive and 13 perceptual resources needed to operate new media may prevent children from processing intended lessons in a deep enough fashion to facilitate transfer to more distal learning measures
(Aladé, Lauricella, Beaudoin-Ryan, & Wartella, 2016). And new media products are often designed in a manner devoid of plot, such that engaging with new media would not result in transportation – a sense of losing oneself in a storyline in a way that facilitates deeper processing and learning (Lu, Baranowski, Thompson, & Buday, 2012). Likewise, new media often feature distracting hotspots that draw children’s attention away from intended educational lessons
(Guernsey & Levine, 2015).
Regarding potential contextual variables that may moderate learning, this dissertation also examines learning in homes and in schools. This dissertation further explores how various methodological decisions (e.g., employment of a between- or within-subjects design) alter both the context of learning and ability to detect significant learning effects.
In terms of child characteristics, this dissertation tests whether media’s effects vary as a result of measurable socio-demographic characteristics (e.g., socio-economic status) in the short- term, and children’s cognitive skills, independent of family background, in the long-term.
Thanks in part to Valkenburg and Peter (2013), who propose that various child-level factors can predict and moderate media’s effectiveness in their Differential Susceptibility Media Effects
Model, children’s media scholars beginning to shift from primarily asking whether media has an impact on children to asking who experiences the greatest media effects. Extending this theoretical work, Piotrowski and Valkenburg (2015) distinguish between “dandelion children”
(p. 4), who would flourish in any environment and are largely insusceptible to media, and
“orchid children” (p. 4), who flourish under positive media effects but are particularly vulnerable 14 to negative media, such as content modeling aggressive behaviors. Linebarger and Barr (2017) assert that low-income children may be one population of “orchid children”.
In addition to probing into potential content-, context-, and child-level moderators, this dissertation also explore how long-lasting media’s effects may be. As will be explained in more detail in Article 3, Anderson and colleagues (2001) propose that early educational television exposure can spark long-term academic achievement, an idea they refer to as the “Early Learning
Hypothesis” (p. 3). They substantiate this notion with supportive correlational evidence linking preschool television consumption to high school academic achievement, and other scholars have begun to collect causal evidence into whether the effects of media, including new media, are still measurable up to a year after initial exposure (e.g., Reitsma & Wesseling, 1998; Segers &
Verhoeven, 2005). Although the RTL producers and evaluators never explicitly refer to the Early
Learning Hypothesis, this notion is highly resonant with the RTL theory of program.
Nonetheless, the Early Learning Hypothesis may be more polemic than the idea that children can consume educational media and then repeat back what they learned shortly after consumption, especially given that short-term media effects tend to be relatively small (i.e., it is debatable whether small effects could sustain over time).
Across this dissertation, I broadly assess the impact of the RTL initiative as a means of providing accountability evidence speaking to the value of this federal program (Rossi, Lipsey,
& Freeman, 2003). Specifically, I examine whether exposure to RTL literacy-themed media results in robust, positive effects in the short-term, and whether any positive effects sustain into late elementary/early middle school. In the process, I consider how my results shed insight—or 15 fail to shed insight—into some of the potential moderators identified in the preceding paragraphs.
Article 1 systematically assesses the average short-term impact of RTL media exposure on young children’s early literacy skills via a meta-analytic review of the accumulated short-term
RTL evaluations. This article also examines if effects differ as function of content, context, and child variation. This meta-analysis includes both studies funded and not funded by RTL.
In Article 3, I re-contact children who participated in one of the evaluations in my meta- analysis. In this study, I ask whether the effects of playing RTL literacy-themed computer games in early childhood sustain into middle childhood. I also explore whether effects vary as a function of children’s preliteracy skills prior to the RTL intervention. The latter analyses allow me to address whether the media helped to close achievement gaps by supporting the children most in need of intervention.
Article 2 does not RTL initiative specifically but provides a broader reflection on the process of re-recruiting families who participate in research when children are in early childhood
(e.g., in studies such as RTL evaluations) for follow-up waves of research. Re-recruiting families for longitudinal research requires (a) searching across a variety of sources to locate participants and leveraging new online tools designed to yield contact information; (b) writing multifaceted scheduling scripts highlighting a study’s value; and (c) confirming appointments in a way that conveys a casual tone to make parents feel comfortable rescheduling if need be. Rather than burying these insights in the Method section of a traditional dissertation, Article 2 will be published as a standalone piece to help guide other researchers working with families. After all, there have been calls for more longitudinal research specifically in the area of children’s media 16
(Wartella et al., 2016), as well as more broadly in other areas of child development (e.g.,
Karmiloff-Smith, 1998). Demonstrating how to address the formidable challenge of re-recruiting families for follow-up waves of data collection may help to facilitate more longitudinal research with children, and may even inform cross-sectional recruitment protocols.
This dissertation concludes by considering the larger implications these findings have for both policy and scholarship. Overall, the findings described herein positively reflect on the RTL initiative, at least for its impact on children’s early literacy skill as assessed in this dissertation.
However, these findings raise as many new questions as they answer in terms of theory, calling for more nuanced thinking about the factors that moderate learning from media. 17
Article 1: Getting a Read on Ready To Learn Media: A Meta-Analytic Review of Effects on
Literacy
Abstract
Most U.S. preschoolers have consumed media created with funding from the U.S. Department of
Education’s Ready To Learn (RTL) initiative, which was established to promote school readiness. Synthesizing data from 45 evaluations, this meta-analysis examined the short-term effects of RTL media exposure on young U.S. children’s literacy skills. Results indicate positive effects of RTL exposure on children’s literacy outcomes, especially vocabulary and phonological processing. Findings are robust across a variety of research designs and for exposure to both television and new media. Effects are comparable in size to comprehensive early childcare programs such as Head Start. These results are discussed in terms of accountability evidence for
RTL and larger debates in scholarly understanding of educational media effects.
Keywords: Ready To Learn, literacy, meta-analysis, educational media, early childhood
18
Responding to public concerns about growing numbers of U.S. children entering kindergarten ill-prepared for formal schooling, Congress, the U.S. Department of Education
(DoEd), and the Corporation for Public Broadcasting (CPB) launched the Ready To Learn (RTL) initiative in 1994 (Bryant et al., 2001). Through this initiative, DoEd has awarded millions of dollars annually - in recent years over $25 million per year - to facilitate the creation, dissemination, and research into the effectiveness of media products promoting young children’s
(ages 2-8) school readiness, with an emphasis on reaching disadvantaged youth (Michael Cohen
Group, 2012). Funding has been used for the continued production of legacy media properties such as Sesame Street, as well as the creation of newer properties such as Between the Lions and
Super WHY! (CPB & Public Broadcasting Service [a U.S. national public television broadcaster and distributor; hereafter PBS], 2011). Today, the majority of U.S. preschoolers have been exposed to RTL media in some capacity, with millions of children accessing these resources regularly (CPB & PBS, 2011). Historically, early literacy has been one of RTL’s primary focus areas. This article provides a meta-analytic review of the accumulated research on the effectiveness of RTL’s literacy-themed media.
Early Literacy
As young children grow to become literate, they begin to master a variety of related early literacy skills. Young readers must learn to recognize letters of the alphabet (alphabet knowledge), develop increasing mastery over the sounds used in language (phonological processing), recall definitions of words (vocabulary), understand how texts work (e.g., in
English, words are read from left to right and books often begin with a cover and title page; print concepts), and be able to follow story structure (narrative comprehension; see Table 1 for more 19 on each of these literacy skills and Langenberg et al., 2001 for more on literacy fundamentals).
Children typically learn these skills in early childhood, beginning to develop print concepts at around the age of 2.5 years, alphabetic knowledge and phonological processing at around the age of 3 years, narrative comprehension of written texts at around the age of 7 years, and vocabulary across childhood (Grant et al., 2012). To a great extent, young children’s skills in these areas are mutually reinforcing and predictive of adult literacy (Langenberg et al., 2000). For example, early phonological processing (Calfee, Lindamood, & Lindamood, 1973) and vocabulary
(Cunningham & Stanovich, 1997) predict reading at the end of formal schooling. Likewise, early narrative abilities (Griffin, Hemphill, Camp, & Wolf, 2004) and letter identification (Adlof,
Catts, & Lee, 2010) also predict later reading. Of all these skills, phonological processing is arguably the strongest predictor of later reading (Grant et al., 2012).
Table 1. Literacy Categories
Literacy Examples of Relevant Skills Sample Assessments Category
Alphabet Letter identification; letter sequencing Alphabet Knowledge test knowledge from the Phonological Awareness Literacy Screening, Pre-K (PALS- PreK; Invernizzi, Sullivan, Meier, & Swank, 2004)
Print concepts A variety of related skills concerning Conventions subtest knowledge of parts of a book and knowledge from the Test of Early of print conventions Reading Ability, Third Edition (TERA-3; Reid, Hresko, & Hammill, 2001)
Phonological Phonemic awareness (knowledge of language Beginning Sound processing sounds, demonstrated through skills such as Awareness and Letter blending phonemes or joining together Sounds tests from PALS- 20
language sounds to form words or PreK (Invernizzi et al., pseudowords); Phonics (knowledge of the 2004) relation between letters and sounds)
Vocabulary Expressive vocabulary (accurately labeling Peabody Picture objects/states/phenomena); Receptive Vocabulary Test, Third vocabulary (knowing the meaning of words Edition (PPVT-III; Dunn others say aloud) & Dunn, 1997)
Narrative Following and interpreting story plotlines Meanings subtest from comprehension the TERA-3 (Reid et al., 2001)
Note. More information on literacy categories.
Early literacy is often a target for intervention, and promoting literacy has been a major focus of U.S. educational policy, especially over the past two decades (e.g., National Conference of State Legislatures, 2016). Past meta-analyses point to the effectiveness of programs promoting early phonological processing, vocabulary, and reading, particularly for preschool-age children
(e.g., Bus & van Ijzendoorn, 1999; Mol, Bus, de Jong, & Smeets, 2008). Because of such evidence, a National Reading Panel of experts convened by the U.S. Congress in the late 1990’s recommended literacy instruction promote learning in these areas (Langenberg et al., 2000).
Educational Children’s Media
A considerable number of U.S. teachers and parents expose young children to screen media intended to promote one or more of the literacy skills outlined above. Among the 80% of early childcare teachers with access to television, 84% with access to computers, and 29% with access to tablets, most use these devices to promote children’s literacy (62%, 76%, and 79% of teachers respectively per device; Wartella, Blackwell, Lauricella, & Robb, 2013). Many parents complement children’s in-school media use by exposing or allowing children to consume more 21 literacy-themed media at home. Among the families with access to television, computers, and tablets (96%, 76%, and 40% respectively), 61% use television, 46% use computers, and 51% use tablets to expose children to educational content such as literacy-themed media (Rideout &
Saphir, 2013). Television use dominates children’s time with screen media, especially at home, but, as access becomes more widespread, both teachers and parents increasingly are providing children access to newer devices such as tablets (Rideout & Saphir, 2013; Wartella et al., 2013).
A variety of theoretical frameworks suggest children can learn skills such as vocabulary from such media, provided that media products are intended to be educational and are designed in a manner that supports learning (Guernsey, 2012). For the sake of brevity, some of the more recent and widely accepted thinking in this area is reviewed. The Capacity Model posits that children should learn from educational media when the intended educational lessons, plotline, and (when relevant) game mechanics are all mutually reinforcing (Fisch, 2004, 2016). For example, a computer game focused on the /p/ sound might require children to hover their mouses over images of foods like “pepperoni” and “peppers” to hear the words said aloud while they create a menu for a p-themed pizza parlor. This model also suggests well designed media products will (a) base narratives in simple, familiar settings (e.g., a playground) so that children can assimilate lessons into existing schemas, and (b) repeat lessons across multiple contexts
(e.g., include a second game in which children create a shopping list of “pears”, “peaches” and other “p-” foods; Fisch, 2004). Extending the notion of the importance of repetition, others have suggested caregivers should complement media’s lessons with real-world activities promoting the same learning goals - a concept termed Experiential Mediation (Piotrowski, Jennings, &
Linebarger, 2012). More broadly, the idea that other people can engage children in discussions to 22 scaffold their learning from media is known as Joint Media Engagement (Takeuchi & Stevens,
2011). To a lesser extent, the media itself may be able to mimic strong scaffolding behavior (e.g., with characters asking children questions to enhance their comprehension and critical thinking;
Strouse, O'Doherty, & Troseth, 2013).
Scholars currently are debating whether new media such as computer games or mobile applications (apps), and television can embody these production and design principles equally well (Hirsh-Pasek et al., 2015). That is, researchers are asking if new media is more or less effective than television. On the one hand, new media has the potential to facilitate strong learning by automatically becoming more challenging and “leveling up” to enhance engagement
(Hirsh-Pasek et al., 2015) and by incorporating learning aids such as clickable dictionaries or word narration (Guernsey & Levine, 2015). But on the other hand, the interactivity required to engage with these newer platforms may require a fair amount of troubleshooting and technical assistance from caregivers to allow for successful usage, which consequently may hamper learning (Guernsey, 2012).
For most of the 20th century, the bulk of research substantiating the premise that children could learn literacy skills from well-designed media focused on Sesame Street, but, in the past twenty years, there has been an explosion of new literacy-themed media properties and products available across platforms (i.e., television, computer, mobile), as well as related research
(Guernsey & Levine, 2015). A few meta-analyses and narrative literature reviews have explored
Sesame Street and these newer media properties and products consistently enhance literacy (e.g.,
Moses, Linebarger, Wainwright, & Brod, 2010; Pasnik, Penuel, Llorente, Strother, & Schindel,
2007). The literature generally suggests educational media has small but positive effects on early 23 literacy skills, typically as measured by custom assessments designed to reflect the learning goals of the media stimuli (Anderson & Collins, 1988). However, not every media product is equally effective in this regard, and many open questions remain. For example, it is unclear if media benefits all literacy skill areas. Prior reviews provide strong evidence that media can promote vocabulary (e.g., Moses et al., 2010) and some evidence it can promote other skill areas, such as phonological processing (Ehri et al., 2001; Langenberg et al., 2000), narrative comprehension, and print concepts (Pasnik et al., 2007). However, Pasnik and team (2007) failed to find significant gains for alphabet knowledge. Moreover, Gola and colleagues (2012) question if media can promote relatively complex, multifaceted skills such as phonological processing (also see Guernsey, 2012). It also is unclear if effects are equal for all children. Concerningly, some research suggests low-income children learn less from educational media than their more affluent counterparts, which in turn may exacerbate achievement gaps (Cook et al., 1975). Furthermore, these reviews have yet to address whether television or new media typically yield larger effects.
The RTL Initiative
RTL has placed a strong emphasis on promoting early literacy among both general and at-risk populations across every five-year grant cycle, aligned with early literacy’s emphasis in larger U.S. educational policy discussions (e.g., National Conference of State Legislatures,
2016). During the first two cycles (i.e., 1994-2000 and 2000-2005), RTL funding recipients CPB and PBS prioritized early literacy by encouraging parents to read books complementing RTL media (Horowitz et al., 2005), and in subsequent grant cycles, producers increasingly began developing turnkey curricula for teachers and parents to help them supplement RTL’s media content (Llorente, Pasnik, Penuel, & Martin, 2010). Starting in 2005, shortly after the passage of 24 the No Child Left Behind Act (which focused in part on promoting literacy skills such as alphabet knowledge, phonological processing, vocabulary, and comprehension in a manner aligned with National Reading Panel recommendations; National Conference of State
Legislatures, 2016), DoEd explicitly tasked content producers with promoting foundational literacy skills (Michael Cohen Group, 2012). While RTL producers have espoused a variety of different theories of learning across media properties and grant cycles, the initiative’s consistent general underlying assumption is that well-designed, developmentally appropriate media should enhance early literacy and other school readiness skills, especially when complemented with
Joint Media Engagement .
Over the course of the grant’s history, DoEd also has called for increasingly rigorous evaluations into the effectiveness of RTL media. For example, starting in the 2005-2010 grant cycle, DoEd encouraged RTL evaluators to use randomized controlled trials (when feasible) and required content producers to partner with third-party researchers, who, although receiving funding from RTL, presumably would produce more objective results than content producers might find themselves (Michael Cohen Group, 2012). Consequently, RTL became one of the largest funders of research into children’s learning from media (CBP & PBS, 2011).
The Current Study
While extant literature suggests educational media can promote early literacy, systematic exploration of the effectiveness of products created with RTL funding still is needed to fully address whether curriculum-based media in general consistently achieves such effects and whether the RTL products specifically are successful in this regard. That is, it is unclear whether
RTL media embody the strong production and design principles described above. Given the 25 majority of U.S. children are consuming RTL-funded media (CPB & PBS, 2011) and hundreds of millions of taxpayer funds have been dedicated to this initiative since the grant’s inception
(DoEd, 2015), arriving at such an understanding provides accountability evidence for the RTL initiative. Some research has attempted to address this gap in the literature through meta-analytic work that included a mix of RTL and non-RTL media (Moses et al., 2010; Pasnik et al., 2007); however, those studies did not specifically address the effectiveness of the RTL initiative.
Moreover, since the time those reports were prepared, researchers have completed dozens of new evaluations of RTL literacy-themed media, with increasing attention to methodological rigor
(Michael Cohen Group, 2012).
The present exploratory study aimed to update this literature. Specifically, this study asked
(RQ1) whether RTL media exposure positively impacts young children’s early literacy skill development. The study also explored (RQ2) whether these effects varied as a function of participant characteristics (e.g., socio-economic status) or research design (e.g., within- vs. between-subjects design). Finally, this study assessed the relation between grant-specific factors and observed effects, examining if (RQ3a) effects changed over time as content producers might have gained experience or as evaluations became more rigorous, and if (RQ3b) effects varied between researchers who were supported by RTL funding and those who were not. The findings from this study provide accountability evidence for RTL spending and address broader concerns regarding young children’s learning from media.
Method
Identification of Relevant Investigations
Literature search. To locate evaluations funded by RTL, studies referenced in lists of 26 evaluations prepared by RTL grant recipients (e.g., CPB & PBS, 2011) and in compendia of research on educational children’s media (e.g., Fisch, 2004) were identified. Many evaluations were available online as technical reports from RTL research and production websites. To find additional evaluations referenced in RTL reviews but not publicly available, RTL research
Principal Investigators and the Director of Research at CPB were contacted directly. These individuals shared all available reports, including some that failed to support the value of a focal property’s effectiveness. Lastly, on August 8 and 9, 2016, a systematic search was carried out across six academic databases: Communication Source, Eric, ProQuest Dissertations & Theses
Global, PsycINFO, Web of Science, and WorldCat Dissertations & Theses. Search efforts focused on titles, abstracts, and keywords or topics for the titles of every media property identified in outside literature as being funded by RTL, names of major grant recipients, and marketing taglines used in public promotions for RTL (e.g., Bryant et al., 2001; Cohen et al.,
2016; CPB & PBS, 2011; Llorente et al., 2010; Michael Cohen Group, 2012; see Appendix A for the full list of search terms).
Inclusion criteria. For the purposes of this analysis, evaluations had to meet five criteria:
1. Use of final versions of RTL media products without any added content. Studies were
included that used full versions of media content (e.g., Piotrowski et al., 2012), clips from
RTL shows (e.g., Neuman et al., 2011) or versions of episodes that paused videos
occasionally for discussion, so long as no new mediated content was added (e.g., Penuel
et al., 2012).
In contrast, studies were excluded in the case of formative evaluations of media
that would change before commercial distribution (e.g., McCarthy et al., 2012); studies 27
testing versions of episodes with added scenes that did not air nationally (e.g.,
Linebarger, 2006) or using still images of RTL media characters as stimuli (e.g.,
Golinkoff, Jacquet, Hirsh-Pasek, & Nandakumar, 1996); or evaluations of spinoffs of
properties that received RTL funding (e.g., Seseame Beginnings, a Sesame Street spinoff;
Roseberry, Hirsh-Pasek, Parish-Morris, & Golinkoff, 2009), unless there was evidence
producers received RTL funding specifically for the spinoff (e.g., Duck’s Alphabet, an
RTL-funded WordWorld spinoff; Michael Cohen Group, 2012).
If evaluations were conducted outside the U.S., authors needed to indicate they
tested versions of media created exclusively with RTL funds. Some content producers
who received RTL funding also partnered with international groups to create versions of
media tailored for local audiences (e.g., Mares & Pan, 2013). Thus, a property’s content
in a foreign market might feature a mix of content created with and without RTL funds.
Searches did not yield any international studies in which authors specified they used the
RTL-funded versions of media; thus, all studies in this meta-analysis sampled in the U.S.
2. Data collected between 1994 and the present. Studies of legacy properties conducted
before the launch of RTL were excluded (e.g., Wright et al., 2001).
3. Presence of at least one direct assessment of children’s literacy in areas such as alphabet
knowledge, print concepts, phonological processing, vocabulary, and narrative
comprehension (see Table 1 for examples). Studies focused on learning domains other
than literacy (e.g., mathematics; Cohen et al., 2016) and studies where children’s
comprehension of the television show’s plot was the sole cognitive outcome measure
(e.g., Sanchez & Lorch, 1999) were excluded. 28
4. Clear demarcation between the effects of RTL vs. non-RTL media. Studies in which
researchers exposed children to both RTL and other children’s media in a way that
confounded RTL and non-RTL media effects were excluded (e.g., Nathanson &
Rasmussen, 2011).
5. Sufficient statistical information to estimate effect sizes. In cases in which such detail was
not provided in available reports, this information was requested from authors. About two
out of every three contacted research teams provided useable data, with recent RTL grant
recipients tending to provide more data than RTL-funded researchers who conducted
research in the first grant cycle (many of whom had retired) and non-funded researchers.
Final sample. Data for this meta-analysis came from 45 RTL evaluations (33 of which were funded by RTL). In many instances, a combination of an unpublished research report, conference presentation, or published journal article all described the same evaluation. Under these circumstances, all texts were treated as being part of the same evaluation, and available information was consolidated across texts as appropriate to provide a comprehensive portrayal of the evaluation. These 45 evaluations resulted in 783 effect sizes for further analyses. See
Appendix B for the full list of evaluations.
Coding of Study Descriptors
Each evaluation was coded for literacy learning outcomes, participant characteristics, design and protocol decisions, and grant-related factors. Basic descriptive information about each code can be found in Tables 2 and 3 and Appendix B.
Table 2. Descriptive Statistics for Nominal and Ordinal Moderators
Moderator n m k Cohen’s κ Literacy categories .84 29
Alphabet knowledge 18 31 56 Print concepts 8 14 21 Phonological processing 20 35 97 Vocabulary 30 53 536 Narrative comprehension 10 16 20 Multiple skills 16 29 31 Other 12 16 22 Design and procedure-related Study design 1.00 Between subjects 35 65 404 Within subjects 11 14 379 Scope of study .59 One geographic region 30 62 719 Multiple geographic regions 14 16 62 Setting .87 School 36 61 686 Home 11 18 97 Media platform .79 TV only 35 66 563 Incorporates new media 12 13 220 Media property .75 Between the Lions 9 23 129 Martha Speaks 6 9 356 Sesame Street 6 10 45 Super WHY! 5 5 44 Other property 17 23 101 Combination of properties 8 9 108 Extension Activities/Materials .89 Activities/Materials 27 44 553 No Activities/Materials 22 29 220 Control .50 Business as usual 17 33 162 Non-RTL media 12 13 83 Alternate curriculum 7 13 151 Assessment .72 Custom narrow 27 41 528 Custom broad 13 16 30 Standardized and validated 30 54 225 Grant-related Researchers funded by RTL 1.00 Yes 33 55 711 No 12 24 72 Grant Cycle .93 1994-2000 5 8 26 30
2000-2005 5 15 63 2005-2010 26 45 666 2010-2015 9 11 28
Note. Number of studies (n), comparisons (m), and effect sizes (k) associated with each nominal and ordinal moderator, and intercoder reliability (Cohen’s κ) for each set of moderators based on double coding 30% of evaluations.
Literacy learning outcomes. Each evaluation assessed at least one literacy skill. An outcome was treated as a literacy outcome if it concerned any of the following: alphabet knowledge, print concepts, phonological processing, vocabulary, narrative comprehension, or other literacy skill areas such as interest in reading, English conversation skills, or syntax. Each literacy category is described in greater detail in Table 1, with descriptive information in Table 2.
Participant characteristics. Given RTL’s focus on reaching all children, especially ones from low-income households or those who otherwise might be at-risk for poor school outcomes
(CPB & PBS, 2011), data were recorded on each sample’s composition with respect to family income (% low-income with annual household incomes below $30,000 or eligible for free or reduced school lunch), race-ethnicity (% Caucasian, African American, Hispanic or Latino,
Asian, and Native American), gender (% Female), school grade (% school age = Kindergarten and above), and English language learner status (% ELL). See Table 3.
Table 3. Descriptive Statistics for Continuous Moderators
Moderator n m k M SD Median Range ICC Sample-related % Low-income 24 35 514 59 28 62 9 - 100 .99 % Caucasian 33 56 656 40 32 34 0 - 92 .94 % African American 31 50 652 29 25 26 0 - 95 .95 % Hispanic/Latino 26 40 614 29 31 20 0 - 100 .99 % Asian 21 34 572 9 9 6 0 - 29 .77 % Native American 9 14 56 8 27 0 0 - 100 1.00 % Female 38 64 712 51 7 50 27 - 88 .65 % School-age 43 77 771 54 47 69 0 - 100 .97 31
% ELL 17 33 267 35 38 14 0 - 100 1.00 Duration of media consumption Weeks of exposure 43 75 773 14.07 26.34 6 1 - 120 .87
Note. Number of studies (n), comparisons (m), and effect sizes (k) providing information on each moderator, along with means, standard deviations, medians, and ranges (calculated at the comparison level), and ICC(3,1) based on double coding 30% of evaluations.
Design and protocol. The dataset varied along the following dimensions (see Table 2):
1. Study design: Studies varied in whether they used a between- or within-subjects design.
In a between-subjects design, a group exposed to RTL media was compared to a group
not exposed to RTL media. Conversely, in a within-subjects design, children’s
performance was assessed before and after RTL exposure, and there was no comparison
group.
2. Scope: Studies varied in whether data were collected in one geographic region or across
multiple regions. Early into the history of RTL, evaluators concluded that it was
challenging to ensure fidelity of implementation when fielding evaluations across
multiple geographic regions (Horowitz et al., 2005), but this assumption has not been
formally tested.
3. Setting: Some studies examined media consumption in school or another childcare
setting, but others had children consume all or some media at home.
4. Duration of media consumption: Evaluations varied in the number of weeks in which
researchers tasked children with consuming media.
5. Nature of RTL treatment. Stimulus materials presented to children were coded based on
(a) media platform (TV only vs. incorporating some new media), (b) media property
(properties examined in five or more evaluations—Between the Lions, Martha Speaks, 32
Sesame Street, Super WHY!--; other properties examined in fewer than five evaluations;
and combinations of multiple of these properties), and (c) use of supplementary extension
activities or materials.
6. Nature of control. In studies with between-subjects designs, comparison children were
either given an alternate, non-mediated curriculum; exposed to media outside the purview
of RTL; or not given any specific directions (business as usual).
7. Assessment. Generally, evaluations of both literacy interventions (e.g., Langenberg et al.,
2000) and educational media programs (Anderson & Collins, 1988) often measure
learning with custom measures closely reflecting the curricula to which children are
exposed, although occasionally will employ validated measures. Thus, in the present
meta-analysis, outcomes were scored as being assessed by custom measures designed to
either narrowly assess specific items or skills shown in media (custom narrow) or to
broadly assess literacy skills not explicitly modeled in media (coded as custom broad), or
by standardized and nationally validated assessments (see Table 1 for examples).
Grant-related factors. Each evaluation was coded for the grant cycle in which data were collected (1994-2000, 2000-2005, 2005-2010, 2010-2015) and for whether researchers were supported by RTL funding.
Meta-Analytic Procedures
Effect size measure. Cohen’s d was used to calculate effect sizes. For most between- subjects design studies, effect sizes were calculated using procedures outlined by Morris (2007), comparing gains of the treatment groups to gains of control groups based on provided means and standard deviations. When insufficient information was available to use this procedure, effects 33 were calculated by drawing on other study data such as post-test scores (Borenstein et al., 2009) and t-statistics (Rosnow & Rosenthal, 2009). For within-subjects designs, pre- and post-exposure performance was compared, following procedures described by Morris and DeShon (2002).
Structure of dataset. The data reported herein had a nested structure. Evaluations often contained more than one comparison (e.g., comparing a treatment group for whom parents received supplemental extension activities to a control group, and comparing another treatment group for whom parents did not receive supplements to that same control group; Piotrowski et al., 2012). Additionally, evaluators often administered multiple literacy measures to children across these different conditions and comparisons. Thus, measures were nested within comparisons, and comparisons were nested within evaluations. In total, the dataset contained m =
79 comparisons (per evaluation M = 1.76 comparisons, SD = 1.05, range: 1-6) and k = 783 effect sizes (per comparison M = 9.73 effects, SD = 22.68, range: 1-144).
Analytic strategy. Analyses were conducted in R-3.13 (R Core Team, 2016) via the rma.mv function in the metafor package (Viechtbauer, 2010). A precision-weighted, random effects, multi-level model was specified using procedures outlined by van Houwelingen and colleagues (2002). Measures were nested within comparisons, which were nested within evaluations, and it was assumed that effects might vary across comparisons and across evaluations due to coded methodological variation. To avoid making assumptions about the inter-relations of the random effects, a compound symmetric variance structure was used. Others conducting similar meta-analyses on the effects of educational media have specified comparable models (e.g., Mares & Pan, 2013). In the present data, examination of the variance components revealed that 1% of the variance was accounted for at the comparison level, and 40% was 34 accounted for at the evaluation level. To compare the nominal moderators described in the section on Coding of Study Descriptors, meta-analytic procedures by Viechtbauer (2015) were used (see Table 4, Model 1). To explore the effects of the ordinal and continuous moderators, meta-regression was used (Viechtbauer, 2007) (see Table 5, Model 1). To test the robustness of these findings, all analyses were repeated using two more conservative models (Models 2 and 3), described in more detail in a Robustness Check section below. Finally, publication bias (the notion researchers may have refrained from publishing studies with null or negative results;
Borenstein et al., 2009) also was assessed, although the likelihood of this bias was low because so many unpublished RTL evaluations were available due to grant reporting requirements.
Results
Main Effects
Overall, exposure to RTL media significantly predicted an increase in children’s literacy
(d = .21, SE = .04, n = 45, m = 79, k = 783, p < .001). Children who consumed RTL media scored approximately one-fifth of a standard deviation higher on literacy assessments. This result included all effect sizes across studies of varied designs and across all literacy outcomes, although there was evidence of heterogeneity across the dataset (Q(782) = 3751.42, p < .001).
Table 4. Nominal Moderator Effects
Moderator Model 1 Model 2 Model 3 d SE (d) d SE (d) dt SE (dt) Literacy categories † † Alphabet knowledge .09a .05 .12ab .08 .10a .05 † † Print concepts .23ab .12 .10ab .07 .23ab .13 *** ** *** Phonological processing .23b .06 .21b .06 .25b .05 *** ** *** Vocabulary .27b .06 .27b .08 .26b .06 Narrative comprehension -.01a .06 -.07a .11 -.05c .05 † * Multiple skills .15ab .08 .06ab .08 .16ab .07 † Other .14ab .09 .14ab .08 35
Design and procedure-related Study design *** *** Between subjects .17a .04 .17a .03 *** *** Within subjects .37b .09 .36b .08 Scope of study *** ** *** One geographic region .21a .05 .24a .08 .21a .05 *** *** *** Multiple geographic regions .23a .04 .18a .05 .20a .05 Setting *** ** *** School .22a .05 .25a .08 .22a .04 *** *** *** Home .20a .07 .17a .05 .19a .05 Media platform *** ** *** TV only .21a .05 .24a .08 .21a .04 *** *** *** Incorporates new media .20a .03 .20a .04 .20a .03 Media property † Between the Lions .07a .05 .11a .09 .08a .05 † * Martha Speaks .12ab .06 .05a .15 .11a .05 ** *** ** Sesame Street .52c .18 .60b .08 .50b .17 *** *** *** Super WHY! .28bc .06 .27a .07 .30b .05 * † ** Other property .16abc .06 .21a .12 .16ab .05 ** ** Combination of properties .20abc .06 .11a .08 .20ab .06 Extension Activities *** ** *** Activities .25a .06 .28a .10 .24a .05 *** ** *** No Activities .15a .03 .12a .04 .15a .03 Control ** ** ** Business as usual .15a .05 .25a .09 .16a .05 ** * ** Non-RTL media .18a .07 .25a .12 .16a .05 * † Alternate curriculum .19a .09 .26a .19 .17a .10 Assessment *** *** *** Custom narrow .33b .06 .36b .10 .33b .05 ** † * Custom broad .16ab .07 .19ab .10 .16a .07 *** ** *** Standardized and validated .12a .03 .13a .04 .12a .03 Grant-related Researchers funded by RTL *** *** *** Yes .21a .04 .25a .07 .21a .03 † † No .21a .12 .11a .17 .20a .10
Note. Model 1: Effect sizes and corresponding standard errors calculated for the full dataset. Model 2: Examining the 17 highest quality evaluations as per Slavin (1986). Model 3: Data corrected for clustering, as per procedures outlined in Hedges (2007) († p < .10, *p < .05, ** p < .01, *** p < .001). Different letter subscripts denote significant differences between moderators (p < .05).
36
As shown in Table 4, Model 1, effects varied somewhat across literacy outcomes. There were significant positive effects for phonological processing and vocabulary, and nonsignificant but positive effects for alphabet knowledge (p = .07), print concepts (p = .051), and tests of multiple literacy skills (i.e., tests simultaneously assessing more than one literacy skill; p = .052).
Effects for vocabulary and phonological processing were significantly larger than alphabet knowledge and narrative comprehension.
Methodological Moderator Analyses: Model 1
Table 5. Continuous and Ordinal Moderator Effects
Moderator Model 1 Model 2 Model 3 b SE(b) b SE(b) bt SE(bt) Sample-related % Low-income .0030† .0017 .0026 .0020 .0027† .0014 % Caucasian .0017 .0012 .0010 .0022 .0012 .0011 % African American -.0010 .0011 -.0006 .0020 -.0006 .0010 % Hispanic/Latino -.0015 .0012 -.0011 .0017 -.0015 .0012 % Asian .0024 .0040 .0050 .0075 .0036 .0039 % Native American .0016 .0015 .0538 .0473 .0016 .0015 % Female .0009 .0047 .0045 .0083 .0004 .0056 % School-age -.0017* .0008 -.0005 .0013 -.0014* .0007 % ELL .0014 .0013 .0017 .0022 .0013 .0013 Duration of media consumption Weeks of exposure .0002 .0022 .0292† .0165 .0000 .0019 Grant-related Grant period .10034* .0483 .0368 .1083 .0868† .0452
Note. Model 1: Meta-regression beta estimates and corresponding standard errors calculated for the full dataset. Model 2: Examining the 17 highest quality evaluations as per Slavin (1986). Model 3: Data corrected for clustering, as per procedures outlined in Hedges (2007) († p < .10, *p < .05, ** p < .01, *** p < .001).
Sample-related moderators. Effects were relatively stable across sub-populations (see
Table 5, Model 1). Results did not vary based on the racial composition or the proportions of girls or ELLs in a sample. However, evidence suggested media products were more effective for 37 preschool-aged children: A sample of all elementary school students would be expected to score about one-fifth a standard deviation lower than a sample of preschoolers. Interestingly, effects were nonsignificantly larger for samples with a higher proportion of low-income children (p =
.08). A sample with all low-income children would be expected to score about one-third of a standard deviation higher than a sample of all affluent children.
Design and procedure-related moderators.
Study design, scope, setting, and length. Results were fairly homogenous across a variety of different study designs. As shown in Table 4 (Model 1), effects were positive and significant for between- and within-subjects designs; studies conducted in one and multiple geographic regions; and studies conducted in homes and schools. For the most part, results did not vary between these different designs. Effects likewise did not vary as a function of exposure duration, as can be seen in Table 5, Model 1. Nevertheless, effects were larger for within- subjects designs than between-subjects designs.
Nature of treatment.
Media platform. Both studies focused solely on the impact of television and those that included new media components yielded significant and positive effects (see Table 4, Model 1).
Property. Across platforms, all properties, excluding Between the Lions and Martha
Speaks, yielded effects greater than zero. (see Table 4, Model 1). Nonetheless, the effects for
Martha Speaks were still estimated to be positive, with p = .051. Some media properties produced larger impacts than others. Effects for Sesame Street were significantly larger than
Between the Lions and Martha Speaks and nonsignificantly larger than “other” properties examined in fewer than five evaluations (which included properties such as Barney and Friends, 38
The Electric Company, and WordWorld; z = 1.95, p = .051) and combinations of multiple properties (z = 1.77, p = .08). Likewise, the effects for Super WHY! were significantly larger than
Between the Lions and nonsignificantly larger than Martha Speaks (z = 1.90, p = .06).
Supplemental Activities or Materials. Across platforms and properties, effects were significant and positive for both studies with and without supplementary materials or activities, as illustrated in Table 4, Model 1.
Nature of Control. As shown in Table 4, Model 1, when examining the subset of studies with between-subjects designs, effects were significant for evaluations that allowed the control group to go about business as usual, asked control children to consume media created without
RTL funding, and provided control children with an alternate (non-mediated) curriculum.
Nature of assessment. Effects were positive and significant for custom narrow, custom broad, and standardized measures (see Table 4, Model 1). However, effects were larger for custom narrow measures than standardized measures.
Grant-related moderators.
Time trend. As demonstrated in Table 5, Model 1, effects grew over the course of RTL’s history. A child who participated in an intervention during the most recent past grant period
(2010-2015) would be expected to score about two-fifths of a standard deviation higher than a child who participated in the first grant cycle (1994-2000).
Results by funding status. As mentioned previously, nearly three-quarters of the evaluations in the sample were funded by RTL. Only these RTL-funded evaluations produced statistically significant mean effects. Nonetheless, effects for non-funded evaluations were identical in magnitude (i.e., d = .21 for both funded and non-funded evauations); the standard 39 error was larger for non-funded evaluations such that these evaluations did not achieve conventional levels of statistical significance (p = .07).
Robustness Checks
Alternate Models. The robustness of findings was assessed by examining two additional, conservative analytical models (Models 2 and 3). Findings consistent across all three models
(main Model 1 + robustness Models 2 and 3) provide strong evidence supporting patterns in the data, while findings only present in one model should be interpreted with more caution.
For Model 2, evaluations were screened for rigor using criteria suggested by Slavin
(1986). Specifically, evaluations that were in the smallest quartile in terms of sample size (fewer than 51 participants) and duration (shorter than 3.5 weeks), evaluations that failed to use random assignment, and evaluations that failed to specify these three criteria were removed. This left 17 high quality evaluations for subsequent analyses (m = 29, k = 252), which included data for all moderators except measures of multiple literacy skills.
In Model 3, all analyses were rerun on the full dataset, attenuating estimates to account for clustering when relevant. As Hedges (2007) explains, many studies assessing learning outcomes sample by large aggregated units like schools, where children are clustered within classrooms and thus might all achieve similar scores regardless of intervention. Indeed, most studies in this dataset used clustered sampling (n = 30, m = 57, k = 601). To provide more conservative estimates of effects in such cases, Hedges (2007) outlined procedures that attenuate estimates, which were used to estimate Model 3. In doing so, a compendium of K-12 intraclass correlation (ICC) values Hedges and Hedberg (2007) estimated for such purposes was consulted, using kindergarten ICCs for the samples in the present dataset composed of preschoolers. 40
Across both robustness models (see Tables 4 and 5, Models 2 and 3), the general pattern of results was largely unchanged. The overall estimated effect size was significant, positive, and roughly comparable in magnitude for both Models 2 (d = .23, SE = .06, n = 17, m = 29, k = 252, p < .001) and 3 (d = .21, SE = .03, n = 45, m = 79, k = 783, p < .001). However, some particulars varied across models. For literacy skills, in Model 2, there was no longer a significant difference between alphabet knowledge vs. vocabulary and phonological processing, although there was now a nonsignificant difference between vocabulary and tests of multiple skills, with the effects on vocabulary greater than multiple skills (z = 1.86, p = .06). In Model 3, the estimated effect for narrative comprehension was significantly smaller than all other literacy skills.
Regarding sample characteristics, the significant effect for school-age children from
Model 1 disappeared in Model 2. While still nonsignificant, given RTL’s focus on low-income samples, it is worth noting that the beta estimate for low-income samples was similar in magnitude across models (p = .056 in Model 3).
Concerning study design, in Model 2, there was a nonsignificant relation between length of an intervention in weeks and estimated effect size (p = .08), such that estimated effects increased for each week of intervention. Regarding the treatment itself, in Model 2, the effects for Sesame Street were estimated to be greater than any other property. However, in Model 3, the difference between Sesame Street and Martha Speaks observed in the prior two models disappeared. For the control activity, effects for studies in which control children used alternate curricula were nonsignificant in both robustness models (although p = .08 in Model 3). In terms of assessments, the estimated effect for custom broad measures was no longer significant in
Model 2 (p = .054), and in Model 3, custom broad measures yielded significantly smaller effects 41 than custom narrow measures.
Finally, concerning grant-related factors, the estimate for grant period was no longer significant in the robustness models (however, p = .055 in Model 3).
Publication bias. While many evaluations in this meta-analysis were unpublished reports, it is possible the present dataset still was biased by missing data (see Borenstein et al.,
2009 for a discussion of bias in meta-analysis), especially if researchers not accountable to DoEd or content producers opted not to share null findings. However, the average effect size of published (d = .27, SE =.08, n = 18, m =33, k = 217, p = .001) and unpublished studies (d = .24,
SE = .06, n = 30, m = 46, k = 566, p < .001) were comparable, z = .28, p = .78, suggesting publication bias was not an issue for the present dataset. Rosenthal’s (1979) Fail-safe N, which determines how many additional null effects would need to be added to a meta-analytic dataset to “nullify” the estimated effect, suggested it would take 173,218 additional null effects before results would fail to achieve statistical significance. This high number also indicates publication bias or any other bias caused by missing data may not be problematic in this dataset. As shown in Figure 1, a funnel plot of the results provided a visual measure of bias (Light & Pillemer,
1984). The placement of effects in this plot was generally symmetrical and Egger’s test was nonsignificant (p = .15), also suggesting the dataset did not suffer from bias related to missing data (Egger, Smith, Schneider, & Minder, 1997). 42
Figure 1. Funnel Plot of Each Study’s Mean Effect Size by Study-Average Standard Errors
Discussion
Overall, this meta-analysis suggests a positive effect of exposure to RTL media on children’s early literacy (especially their vocabulary and the harder-to-promote skill area of phonological processing). Findings were fairly robust across research designs, although effects were somewhat more pronounced in within-subjects designs and in cases where researchers used custom narrow assessments closely mirroring focal media content. Several different media properties yielded effects on par with Sesame Street.
Interpretation of Overarching Findings
There are multiple ways to interpret the real-world meaning of these findings (see
Cooper, 1981). For education interventions, the What Works Clearinghouse (2014) recommends translating effect sizes into an “improvement index”, or an expected change in percentile rank resulting from intervention. This entails converting an effect size into a z-score, interpreting that 43 z-score as a U3 index (the area under the standard normal curve corresponding to said z-score), and subtracting off 50% (the area under the curve corresponding to “average” performance not impacted by treatment). For the present-analysis, the improvement index would be 8.32. In other words, being exposed to RTL media corresponded to gains in literacy achievement of about 8 percentiles. Judging against normative growth rates set forth by Hill and colleagues (2008), the estimated effects obtained herein are equivalent to about one-and-a-half months of literacy learning above and beyond typical growth.
The results in this study were on par with those obtained in meta-analyses examining computerized reading interventions (Langenberg et al., 2000) and versions of Sesame Street produced in 15 countries outside the U.S. (Mares & Pan, 2013). The present effects also were comparable to meta-analyses probing into the impact of comprehensive U.S. early childcare
(e.g., state preK, Perry Preschool, and Head Start; Kay & Pennucci, 2014). However, the current effects were considerably smaller than those reported in other meta-analyses of sometimes longer, largely nonmediated targeted literacy interventions focusing on programs where teachers or researchers taught children reading strategies or where researchers taught parents coaching strategies like dialogic reading (e.g., Bus & van Ijzendoorn, 1999; Mol et al., 2008).
Implications for RTL Stakeholders
RTL media had small but positive impacts on children’s early literacy, providing fairly positive accountability evidence in favor of RTL. The effects of consuming content from some newer properties (e.g., Super WHY!) for the most part seemed to be about equivalent to the effects of RTL legacy property Sesame Street, which previously had been considered the gold standard for educational children’s media (Fisch, 2004; Guernsey & Levine, 2015). RTL also 44 successfully promoted vocabulary acquisition, replicating prior literature (Moses et al., 2010), and impacted the multifaceted skills of phonological processing, an area previous researchers supposed might not be sensitive to media intervention (Gola et al., 2012). Given the particularly strong link between phonological processing and later reading competence (Grant et al., 2012), this finding is especially noteworthy. Moreover, RTL media appeared to be just as, if not more, effective across grant cycles.
Most of the demographic findings suggested in this meta-analysis also are consistent with
RTL’s goals. RTL media were equally effective for both boys and girls, ELLs and native speakers, and children of a variety of different racial-ethnic backgrounds. There was also evidence that RTL media were as, or more, effective for low-income audiences (an RTL target population; Michael Cohen Group, 2012). In light of past work, which suggested educational children’s media might be less effective for low-income learners (e.g., Cook et al., 1975), this finding is quite encouraging. Although inconsistent with RTL’s aim to reach children age 8 and younger, it is not surprising that media were less impactful for school-age audiences, especially given prior meta-analytic work reporting similar age-based findings for a variety of literacy interventions (Langenberg et al., 2000; Mol et al., 2008). RTL evaluators noted the school-age children in many samples were already scoring at ceiling at pre-test, especially for simpler measures such as assessments of alphabet knowledge (e.g., Penuel et a., 2012). In a similar vein, literacy research more generally suggests that children typically master many skills related to print concepts and phonological processing by kindergarten or first grade (Grant et al., 2012).
Additionally, school-age children might be more likely to receive formal literacy instruction in school and have less to gain from educational media than their younger counterparts. 45
When considering RTL’s overall effectiveness, however, one must consider findings indicating that (a) effects were not consistently significant when the control group engaged with non-mediated alternate curricula; (b) only RTL-funded researchers found significant outcomes on average; and (c) effects were largest for custom narrow assessments mirroring the content presented in media. When considering findings about differences based on control task, it is important to keep in mind that in all three models, there was no significant difference in estimated effects between studies where control groups used alternate curricula or media, or where control groups engaged in business as usual. Given that only 7 evaluations used alternate curricula, there may be an issue of statistical power for this moderator. Regarding the link between funding status and outcomes, it similarly is important to note that the effects for RTL and non-RTL researchers were generally comparable in magnitude, and that the difference between these two groups of researchers never achieved statistical significance. Nonetheless, it is possible that RTL researchers might have conducted studies in qualitatively different ways more likely to yield positive results (e.g., invested more resources creating custom narrow measures) or simply may have been more skilled than some of their non-RTL-funded counterparts
(qualifying them to be awarded competitive RTL research subcontracts). Alternatively, given that only 12 evaluations were authored by teams without RTL funding, it again is possible that the effects for non-RTL researchers might have failed to achieve significance simply due to a lack of power. It would be helpful for more fully independent researchers (i.e., researchers not accountable to RTL grant recipients or in any way benefiting from RTL funding) to elect to use media from some of the properties identified in this review in future studies to help clarify this.
Concerning effects being larger for custom narrow assessments, this is unfortunately a 46 common weakness of intervention research across sub-fields. In evaluations of both mediated and nonmediated literacy interventions, researchers tend to rely heavily on such measures, which in turn yield larger effects than validated measures (Anderson & Collins, 1988; Langenberg et al., 2000). These measures may reflect partial or otherwise biased manifestations of the skillsets they are intended to measure. In this particular dataset, many of the standardized assessments researchers administered or emulated in custom broad measures were criterion-referenced and thus not necessarily sensitive to growth over short interventions (e.g., Invernizzi et al., 2004). It therefore may possible that the effects for standardized and custom broad measures would have been larger if researchers solely relied on norm-referenced tests. Then again, it might be unrealistic and perhaps somewhat frightening if 6-14 weeks of media exposure resulted in large and rather generalized gains (Cohen et al., 2016).
Moving forward, RTL funders and producers may wish to consider continuing to emphasize skills not as well supported by the current body of RTL media, such as narrative comprehension. Although it may be more challenging for an early learning program to effect change in this area, given that it is a more advanced skill than many of the other literacy skills targeted by RTL (Grant et al., 2012), previous research suggests narrative comprehension can be promoted via a variety of narrative storytelling formats (Linebarger & Piotrowski, 2009).
Implications for Developmental Scientists
These findings also speak to broader questions in scholarship on children and media (e.g., the value of Experiential Mediation). Although the findings of this meta-analysis were fairly robust across a variety of research designs, results nonetheless underscore conventional wisdom and theory about designing studies of this nature on some fronts. As might be intuitive, there is 47 weak evidence that longer interventions and strong evidence that interventions without comparison groups result in larger effects. Additionally, consistent with prior scholarship on learning from media (Fisch, 2004) and on learning more generally (Hill et al., 2008), the present results suggest children’s learning might appear greatest when they are given assessments closely mirroring the educational content they have consumed, as it is challenging for them to generalize lessons learned in one context to assessments that do not closely resemble that context.
It may be frustrating to some that these data do not resolve debates into whether television or new media are more effective in promoting literacy. Because RTL producers have been on the forefront of creating content for new platforms, it may be that children and caregivers struggled to fully utilize media products running on newer platforms or that families lacked needed resources such as reliable Internet access (Llorente et al., 2010). Relatedly, RTL producers may have faced a steep learning curve in designing effective new media products, given that designers and scholars alike are only just beginning to explore the new media design principles that best support learning without boring or distracting children (Hirsh-Pasek et al.,
2015). Consequently, device and service penetration may “catch-up” with the set of RTL new media products and the products themselves may continue to improve, eventually providing more of an advantage to new media.
Somewhat surprisingly (see Horowitz et al., 2005), effects were comparable for studies conducted in both one and multiple geographic regions. This suggests that with diligence, evaluators can indeed execute studies in more than one site.
The lack of support for the benefits of extension activities or materials may seem quite unexpected, failing to support literature on Experiential Mediation (Piotrowski et al., 2012) and 48
Joint Media Engagement (Takeuchi & Stevens, 2011). However, some RTL researchers suspected adults facilitating interventions sometimes failed to implement extension activities as intended by media producers (e.g., Horowitz et al., 2005). Also, it is possible the extension materials researchers provided did not always complement mediated lessons. RTL producers have continuously refined their development of extension resources to try to make them more caregiver-friendly (Llorente et al., 2010), but these improvements may not be reflected in the present dataset. Anecdotally, in studies where researchers explicitly laid out curricula with detailed instructions on extending mediated lessons and when facilitators followed these instructions, effects were quite large (e.g., Neuman et al., 2011).
Limitations
These results should be interpreted in light of three limitations. First, because this study focused on a U.S.-based grant, all the studies in this sample were conducted in the U.S. Even though these results are aligned with recent international meta-analyses of similar media products (Mares & Pan, 2013), results obtained here may not generalize to all languages or cultures. Second, there was a limit to questions this meta-analysis could answer. After all, only so much nuance can be captured via nominal moderators. For example, some extension activities or materials might have been stronger than others, but such qualitative variation is not reflected in the present analyses. Likewise, some media properties might be more successful at enacting the production and design principles described in the introduction, but this study did not engage in a content analysis of the media itself to be able to directly address this possibility. Third, the results reported above do not account for dependence between measures (i.e., that two effect sizes from the same comparison are likely correlated). Because 38 evaluations did not provide 49 information on the correlation between measures and because methodological work suggests the impact of ignoring dependence between effects is minor for large datasets like this one (Kim &
Becker, 2010), this decision seemed justified. Indeed, in yet another alternate model (not reported herein), in which covariances between effects within comparisons were imputed from information in the 7 evaluations that did provide some correlational information, the results were nearly identical to those of Model 1. However, it is possible that results may have varied more substantially had fuller correlational information been available.
Conclusion
The results from the present study provide evidence that consuming RTL media is a valuable use of children’s time, especially for preschoolers, and that spending on initiatives such as RTL is a valuable use of taxpayer dollars. These findings demonstrate RTL media can successfully promote early literacy skills (e.g., vocabulary; phonological processing), and that these findings are robust across a variety of different research designs and under a variety of different assumptions. 50
Article 2: “Are You Sure My Child Participated in that Study?” Recruiting and Re-
Recruiting for Developmental Research
Abstract
Recruiting children and families for research studies can be challenging, and re-recruiting former participants for longitudinal research can be even more difficult, especially when a study was not prospectively designed to encompass continuous data collection. This article discusses how researchers can set up initial studies to potentially facilitate later waves of data collection; locate former study participants using newer, often digital, tools; schedule families using recruitment scripts that highlight the many benefits to continued study participation; and confirm appointments using other newer digital tools. This article reflects on my experience conducting a non-prospective longitudinal study with urban parents and their early adolescent children, as well as prior methodological pieces and longitudinal articles. The primary aim is to provide suggestions to others wishing to re-recruit families for longitudinal studies; however, many of the strategies shared in this article also could enhance cross-section recruitment.
Keywords: recruitment, methods, developmental research, longitudinal
51
Across developmental science subfields, researchers have called for more longitudinal research (e.g., Karmiloff-Smith, 1998; Wartella et al., 2016). Such research provides compelling information about developmental trajectories, patterns, sequences, and pathways (Nicholson,
Sanson, Rempel, Smart, & Patton, 2002).
However, conducting longitudinal research presents formidable challenges, chief among which is re-recruiting sizeable samples across waves of research (Cotter, Burke, Loeber, &
Navratil, 2002; Ribisl et al., 1996). Recruiting for cross-sectional studies with children is already quite demanding, requiring researchers to devise thoughtful sampling schemes (Bornstein, Jager,
& Putnick, 2013), set up databases tracking contact with families, convince families to follow through and schedule study appointments, and confirm these appointments in a sensitive manner, all while working around families’ busy schedules (Striano, 2016). In addition to these challenges, re-recruiting for longitudinal work presents the added difficulty of working with a finite pool of families who might become fatigued from participating in multiple waves of research or who might have moved or changed contact information over time (Agrawal, Kellam,
Klein, & Turner, 1978; Barakat-Haddad, Elliott, Eyles, & Pengelly, 2009; Cotter et al., 2002).
In the present paper, I first provide general background information on developmental scientists’ interest and success in re-recruiting for longitudinal work. I then describe my own recent foray re-recruiting families who participated in a brief educational intervention in 2010 for a follow-up study in 2016. I next outline strategies for arranging studies to facilitate later waves of data collection, locating original families, and scheduling them for new testing sessions. I contextualize these strategies in light of previously shared wisdom in this area. My primary aim 52 is to inform the work of researchers conducting longitudinal research; however, many of the suggestions I share also could enhance cross-sectional participant management.
Longitudinal Research in Developmental Science
Numerous developmental scientists are attempting longitudinal research, recruiting participants for a study at one time, and then re-recruiting them for subsequent waves of research at later time points. Large-scale multi-disciplinary cohort and panel studies such as the Dunedin
Multidisciplinary Health and Development Study (e.g., Silva, 1990; Stanton & Silva, 1992) and studies falling under the umbrella of the Early Childhood Longitudinal Study (ECLS) (e.g.,
Heatly, Bachman, & Votruba-Drzal, 2015) are the most renowned sort of longitudinal work in child development. Such studies require extensive and continuous planning and resources
(Nicholson et al., 2002).
In addition to these large-scale efforts, many other scholars have conducted smaller scale follow-up studies, sometimes deciding to re-recruit families who participated in one (often lab- based) study after extensive periods without any contact. Researchers have conducted such longitudinal work across a host of domains, including spatial cognition (e.g., Lauer & Lourenco,
2016), language (e.g., Can, Ginsburg-Block, Golinkoff, & Hirsh-Pasek, 2013), temperament
(e.g., Schwartz, Wright, Shin, Kagan, & Rauch, 2003), personality (e.g., Harris, Brett, Johnson,
& Deary, 2016), self-regulation (e.g., Ayduk et al., 2000), health and physical development (e.g.,
Fein, Li, Chen, Scanlon, & Grummer-Strawn, 2014), mental health (e.g., Agrawal et al., 1978), and media use (e.g., Hanson, 2017). Similarly, researchers who evaluate interventions also have engaged in comparable, sometimes non-prospective, longitudinal research years after the conclusion of interventions. Such studies help assess the long-term impact of programs 53 attempting to directly improve children’s outcomes in areas such as education (e.g., Campbell et al., 2012) and health and fitness (e.g., Lazorick et al., 2014), and indirectly influence child outcomes by providing parents supports such as cash supplements (e.g., Huston et al., 2005) and drug counseling (e.g., Haggerty et al., 2008). Thus, a variety of developmental scholars are responding to calls for longitudinal research, even though in practicality there are often sizeable gaps in time between waves of research in these studies.
Without the resources of large-scale cohort and panel studies, researchers have met the challenges of re-recruiting original study participants with varied success. Some report it to be particularly challenging re-recruiting urban individuals with racial-ethnic minority and low socio-economic status (SES) backgrounds (e.g., Barakat-Haddad et al., 2009; Fein et al., 2014;
Poehlmann-Tynan et al., 2015; Ribisl et al., 1996), older children (e.g., Cotter et al., 2002;
Cotter, Burke, Stouthamer-Loeber, & Loeber, 2005), and individuals with common names (e.g.,
Barakat-Haddad et al., 2009; Masson, Balfe, Hackett, & Phillips, 2013). Conversely, others suggest very high-SES families are especially challenging to re-recruit (Silva, 1990). In the scholarship reviewed in the preceding paragraph, researchers’ re-recruitment rates, when reported or inferable across publications, ranged from less than 15% (Barakat-Haddad et al.,
2009; Schwartz et al., 2003) to nearly 100% (Campbell et al., 2012; Silva, 1990). However, researchers often fail to report retention rates or information that can be used to infer these rates, or to even use consistent definitions of what they consider to be successful retention (e.g., locating a parent AND child versus only locating the parent; Ribisl et al., 1999).
Along these lines, scant guidance exists for re-recruiting families for longitudinal research. A small number of developmental researchers have shared strategies to facilitate 54 multiple rounds of data collection, foremost among which include (a) using contact information participants provided at the beginning of the original study (e.g., Ayduk et al., 2000), (b) employing—often fee-based—databases to find updated contact information (e.g., Haggerty et al., 2008), and (c) providing increasing compensation across waves of data collection (e.g.,
Cotter et al., 2002), among other techniques described in more detail in the following pages.
Nonetheless, typically, published studies in child development focus more on the measures administered across waves of data collection than the recruitment process. Much of the existing re-recruitment guidance exists in articles targeting general populations, with advice not always relevant to children and families (e.g., Ribisl et al., 1996), or in clinical- or practitioner-oriented journals, which often include strategies very specific to small, special populations of children
(e.g., Masson et al., 2013). The primary goal in the present manuscript is to extend this body of literature by outlining strategies for re-recruiting a wide variety of children and families for a wide variety of potential follow-up studies. To do so, I reflect on my own recent experiences re- recruiting urban families for a longitudinal study.
Current Context
I recently re-contacted families (original N = 136) who had participated in an 8 week educational computer game intervention when children were in preschool and kindergarten (M age at Time 1 = 5.24 years, SD = .71). I collected follow-up data from former study participants six years after the original intervention when children were in late elementary school (M age at
Time 2 = 11.28 years, SD = 1.30). This was not a prospective study that was planned from the outset of the original intervention; consequently, the research team had no intermediate contact with families between the intervention and follow-up study. The original sample was American, 55 urban, and racially and ethnically diverse (28% Caucasian, 28% Hispanic, 20% African
American, 24% Other or Mixed race), with approximately 60% of families with incomes below
$40,000 and 45% receiving or eligible for government aid. Given the literature suggesting older children (Cotter et al., 2002), individuals from urban communities and low-income or racial- ethnic minority backgrounds are particularly difficult to re-recruit (Ribisl et al., 1996), I recognized the challenge ahead of me in re-recruiting for this project.
I had modest financial but considerable human resources at my disposal. My budget was a little over US$5,000 to recruit, travel to, and compensate participants. I also had part-time re- recruitment manpower from a faculty member, lab coordinator, doctoral student (who was collecting this data for her dissertation), and five undergraduate research assistants. The faculty member had conducted a similar longitudinal study nearly 20 years ago, before dramatic increases in the popularity of newer interpersonal communication tools, such as text messaging and social networking sites (Duggan, 2013; Purcell, 2011).
Preparing Initial Studies to Facilitate Later Waves of Data Collection
Researchers may not always know at the outset of an initial study if they will be able to conduct subsequent rounds of data collection, for a variety of reasons including tenuous funding
(Nicholson et al., 2002). Likewise, the research literature may prompt new questions that were not under consideration as part of the original research, but which could be addressed by re- recruiting original participants. Indeed, this was the case for us. Even in these situations, there are several steps researchers can take to facilitate potential future rounds of data collection. First, researchers may partner with local organizations, including schools (see Bornstein et al., 2013;
Ribisl et al., 1996). These relationships can assist with both initial recruitment (Striano, 2016) 56 and re-recruitment, depending on the initial language used in establishing said partnership and the nature of the study itself (see the discussion on locating participants below for more information).
Second, researchers can create a database with detailed participant contact information, including information such as participants’ and their family members’ full names and aliases
(e.g., nicknames, maiden names), phone numbers, emails, mailing addresses, educational and employment information (if relevant), contact information for friends, neighbors and/or relatives, birthdates, plans to move or change names, physical descriptions, and favorite hangouts
(Barakat-Haddad et al., 2009; Cotter et al., 2002; Haggerty et al., 2008; Masson et al., 2013;
Ribisl et al., 1996). During original data collection, the research team collected names, phone numbers, email addresses, and employment information for up to two parents; full names, birthdates, and preschool/kindergarten names for children; and mailing addresses for the entire family. This helped ensure high participation over the course of the original intervention, and, as described in more detail below, was invaluable to subsequent re-recruitment efforts.
Third, researchers can adopt a blanket policy of including an optional element on all consent forms seeking permission for future contact about later research opportunities (e.g.,
Masson et al., 2013). This easy addition to consent forms provides an opportunity for future outreach should the need arise. In the absence of this initial consent, follow-up contact unexpectedly may be perceived as a privacy violation, and prevent former participants from choosing to dedicate more time to a research project. Further, university Institutional Review
Boards (IRBs) may have an ethical obligation to prevent non-consensual follow-up contact.
Indeed, this seems to be a common standard across universities in the U.S. (e.g., J. Hecht, 57 personal communication, July 1, 2016). It is important researchers take these initial safeguards, because children are special vulnerable populations (Hartmann, 1992).
Fourth, researchers should attempt to build rapport with both parents and children across research activities (Cotter et al., 2002) and establish clear branding via the use of university or other logos (Haggerty et al., 2008; Ribisl et al., 1996; Striano, 2016). Depending on the level of anonymity of the study, researchers could even establish a Facebook or similar social media group for participants. Having a blog or newsletter to update families about study findings also can be a way to maintain connections with families who have participated in prior studies. In prospective studies, researcher sometimes send participants regular newsletters and birthday/holiday greetings (Ribisl et al., 1996). Such techniques can help ensure the research experience is positive for participants and help them identify with the research study (and perhaps sign-up for other cross-sectional studies with researchers even if a team does not attempt to follow up with a specific study).
Finally, researchers with sufficient foresight in certain sub-domains may wish to gain consent from parents to contact their family members, friends, neighbors or other professionals in their lives (e.g., clergy; case workers) for help locating them (i.e., participating parents and children) at a later point, and to have parents prepare notes for these individuals consenting for them to provide current family contact information (Cotter et al., 2002; Passetti, Godley, Scott,
& Siekmann, 2000; Ribisl et al., 1996). These permissions could later be leveraged to assist with locating study participants for later waves of data collection. As a caveat, this approach may not be appropriate for all topic areas. For example, seeking this information might be reasonable as part of a lengthy intervention but could be intrusive in a one-time lab session. 58
Re-Recruiting for Longitudinal Research
In the discussion of re-recruitment for later waves of longitudinal research, I distinguish between locating, and actually scheduling families and engaging them in study participation. I consider a family to be located if a caregiver responds to re-recruitment attempts verbally or in writing, and acknowledges that I have identified the correct family, regardless of their level of interest in participating in the follow-up study (Haggerty et al., 2008). In contrast, I consider a family to be scheduled and to have participated in the study once they have arranged a time to meet with researchers and followed through with these appointments. Families who are difficult to locate are not necessarily difficult to schedule: In this study, the time I spent searching for a given family was negatively but nonsignificantly correlated with the time it took to schedule said family once located (r = -.20, p =.06).
Although I located, scheduled, and engaged families in study participation simultaneously, others have recommended dedicating some amount of time to locating participants prior to moving to the scheduling phase (Ribisl et al., 1996). That is, researchers may be well served by first locating as many families as they can and then scheduling them for follow-up appointments. Researchers may find certain families more time consuming to recruit than others (Masson et al., 2013), which in turn can prevent researchers from being able to speedily collect all data in a finite time period. When separating the locating and scheduling phases, families might receive a small monetary reward simply for providing updated contact information (followed eventually by a larger reward for participating and contributing data to subsequent waves of data collection; Ribisl et al., 1996). Additionally, more resources can be dedicated to families that researchers suspect might be challenging to locate based on prior 59 interactions (Cotter et al., 2002), especially in light of research suggesting some families are consistently more challenging to re-recruit across multiple waves of research than others (Fein et al., 2014). In hindsight, I think this approach would have served me well and would be valuable for other developmental researchers, in that doing so might help researchers conduct all testing sessions succinctly once they have located most target participants. Testing easier-to-locate children first and testing harder-to-locate children months later might allow significant, potentially confounding developmental differences to arise.
The remainder of this section discusses strategies for locating and scheduling families.
The discussion of the former topic is particularly germane to those conducting follow-up studies, but the discussion of the latter topic includes strategies that also can be applied during the first round of data collection in a longitudinal study and during cross-sectional studies.
Locating Study Participants
The research team and I located 122 (90%) of the original participants in the present study. We waited at least two weeks between attempts to reach the same family, and made efforts to initiate contact during a variety of times of day. Since participating children were in late elementary school at Time 2, many lived in households where all caregivers had fulltime jobs, and accordingly, caregivers sometimes were more receptive to calls made outside of business hours (see Striano, 2016). However, other parents in the sample worked nontraditional jobs or hours, and thus were easier to reach midmorning. On average, it took 22.9 days (SD =
35.04) or about 3 days of outreach attempts from the time the team first tried to locate families via any means to when we successfully located them. As is evident by the large standard deviation, there was considerable variability in the ease of locating families: We located 67 60 families (55% of located families) in less than one week (i.e., following one contact attempt day), 25 families (20% of located families) within one month (i.e., following about two contact attempt days), and the remaining 30 families (25% of located families) after one month. We may have been able to locate some families somewhat more rapidly had I not also been collecting data simultaneously. Located families had higher incomes, were more likely to be Caucasian, and were less likely to be of Other/Mixed race-ethnicity than non-located families (ps < .05).
Otherwise, there were no differences between located and non-located families.
As will be described in more detail below and in Table 1, families were located using the contact information they provided at Time 1, and searching for updated contact information via free and paid tools. Below, I also point to strategies I did not believe would be successful for this study but that other researchers may find useful.
Originally provided contact information. The team used originally provided contact information to call, text, email, and/or mail parents (e.g., Anderson et al., 2001; Ayduk et al.,
2000).. In contrast to other studies, where researchers only saw moderate success leveraging originally provided contact information (e.g., Haggerty et al., 2008), I found that reaching out to caregivers using the information they provided in the original study was by far the most successful location strategy employed. Traditional phone calls were particularly fruitful.
Although many of the participants in the original study had moved from the specific homes they lived in at Time 1, most were still living in the greater metropolitan area and had retained their cell phone number or email address. Had the team exclusively relied on original contact information, we still would have located over 70% of the original participants, which some would argue is an acceptable retention rate (Ribisl et al., 1996). This thus could have saved many
Table 6. Techniques for Locating and Scheduling Participants (N = 136 original study participants)
Contact Method Located Participated Days Taken to Days Between Locate via Location and Successful Participation Method n % % n % % M SD M SD of original of located of original of participants participants participants participating participants Original contact 101 74% 83% 86 63% 87% 11.26 21.54 21.98 28.14 info. Phone 76 56% 62% 62 46% 63% 11.21 23.62 22.92 29.58 Call 72 53% 59% 59 43% 60% 11.78 24.01 23.45 30.18 Text 4 3% 3% 3 2% 3% 1.00 0.00 12.33 8.50 Email 24 18% 20% 24 18% 24% 11.58 14.71 19.39 24.10 Mail 1 1% 1% 0 0% 0% 7.00 Contact info. via 18 13% 15% 13 10% 13% 14.72 13.81 19.38 28.69 free sources Phone (calls) 8 6% 7% 7 5% 7% 13.63 14.60 13.29 10.75 Email 6 4% 5% 4 3% 4% 14.00 11.10 38.00 48.39 New email 4 3% 3% 2 2% 2% 22.33 5.86 17.50 6.36 New Gmail 2 2% 2% 2 2% 2% 1.00 0.00 58.50 72.83 Social media 4 3% 3% 2 2% 2% 18.00 18.96 3.50 2.12 (Facebook) Contact info. via 3 2% 3% 2 2% 2% 5.33 3.79 1.50 0.71 paid sources
Phone (calls) 1 1% 1% 0 0% 0% 1.00 61
Mail 2 2% 2% 2 2% 2% 7.00 9.00 1.50 0.71 62
Note. Number of families who first responded to recruitment attempts via each contact method, percentage located of the original sample (N =136) and percentage of located sample (n = 122) responsive to each method, number of families who participated in the follow-up after responding to each contact method, percentage of the original sample (N = 136) and percentage of participating sample (n = 101) who participated after responding to each contact method, days taken to locate via successful method (the average number of days between when researchers first attempted to contact families and when they finally located them, categorized by the method that ultimately provided to be successful in locating them (where the day initially contacted = day 1)), the standard deviation on that number, days between location and participation (the average number of days between when researchers first located a family and when the family participated in the study per method), and the standard deviation on that number.
manpower hours attempting the strategies outlined below. Growing reliance on cell phones means that individuals, especially in urban areas such as where I collected my data and especially for higher-SES populations, are increasingly maintaining the same phone number even after they relocate (Dost & McGeeney, 2016). This may in part explain some of my team’s success using several-years-old phone numbers to re-recruit relative to prior research conducted before increased cell phone penetration.
Searching for participants via free online tools. Multiple free strategies were employed to ascertain updated contact information in cases where originally provided information was outdated. First, a variety of free search engines and databases were leveraged (e.g., Cotter et al.,
2002; Passetti et al., 2000). Again somewhat contrasting with prior work, where free resources only yielded a small amount of accurate contact information (e.g., Haggerty et al., 2008; Masson et al., 2013), this was the most successful search avenue for in the present study after exhausting original contact information. Mirroring recent prior similar studies (Barakat-Haddad et al., 2009;
Cotter et al., 2002; Fein et al., 2014; Haggerty et al., 2008), general search engines such as
Google, online databases known for contact information (e.g., the National Change of Address database) and people-specific search engines (i.e., search engines designed to scrape the Internet 63 just for publicly available contact information) were all leveraged. Whitepages.com, pipl.com, zabasearch.com, , and RefereneceUSA.com yieldied the most fruitful results, although be cautioned that there is high turnover across such websites (for a similar warning, see Masson et al., 2013). Several websites recommended in prior, relatively recent methods pieces were no longer live during the present data data collection. I suspect my team’s success with free resources may in part be attributable to the large volumes of contact information some people now post online through personal websites, public social network profiles, and the like (Rainie,
Kiesler, Kang, & Madden, 2013), and to newer blogs and articles showcasing high quality free people-centric search engines (e.g., Boswell, 2007). It also may be that the algorithms these services use have improved over time.
Second, my team tested plausible alternate email addresses. We located a small number of families by pairing the local part of email addresses parents had provided at Time 1 with
Gmail domains (e.g., If a mother indicated her email address was something like [email protected] at Time 1, we attempted to contact her at Time 2 using [email protected]). I noticed early into recruitment, before we began implementing this strategy, that several parents had changed their emails along these lines. Additionally, I know from outside reports that Gmail grew in popularity in between the time we originally collected data and when I conducted the second wave of data collection, while Yahoo, Hotmail and AOL declined in use (Creager, 2011; Dupre, 2014; Khan, 2015). More broadly, because people frequently use the same usernames repeatedly across accounts (Jacobsson Purewal,
2015), researchers may succeed in locating participants by pairing any usernames on file with the latest or most popular email or social networking providers. 64
Third, in cases in which the team were unable to find active contact information for families by the means described above, we searched for and attempted to contact parents through
Facebook, Twitter, and LinkedIn. As per prior research (Masson et al., 2013), team members occasionally searched through publicly available lists of parents’ contacts on these networks to ensure they had identified the correct family (e.g., searching if a potential parent was connected to his/her spouse as identified on the Time 1 contact forms).
The team debated creating clean study accounts vs. using personal accounts to message participants, but ultimately decided to use personal accounts for enhanced credibility (i.e., to come across like real people and not spammers; T. Rousse, personal communication, June 28,
2016). Accordingly, I suspected recruiters who were similar to the sample parents (i.e., fellow parents of school-age children) might be more successful recruiting in this manner (see Ribisl et al., 1996 for a similar discussion about recruiter-participant similarities in the context of more traditional location methods). Nonetheless, the data do not provide clear evidence for or against this supposition: The youngest undergraduate researcher on the team received a response from 0 out of the 3 families she contacted via social media, the faculty member (who had children) received a response from 1 out of 8 families she contacted, but I received a response and scheduled 1 out of 3 families she contacted.
Despite searching across multiple social networking sites, the team only successfully recruited (and scheduled) through Facebook, paralleling other recent non-prospective longitudinal studies that relied on social media for re-recruitment (Masson et al., 2013). These findings may reflect Facebook’s popularity among middle-aged adults relative to other social networking sites at the time of data collection (Duggan, Ellison, Lampe, Lenhart, & Madden, 65
2015). Interestingly, in other work, participants reported preferring to be contacted through
Facebook over traditional mail or telephone because they felt Facebook was private and could conveniently be accessed by both smartphone or computer (Masson et al., 2013).
Searching for participants via paid online databases. Like others (e.g., Cotter et al.,
2002), after exhausting the aforementioned free options, I paid to access fee-based contact information databases. There are mixed opinions on the usefulness of these resources. Some prior studies have utilized such search engines early into their re-recruitment efforts and found them to be quite useful (e.g., Haggerty et al., 2008). Others, however, are more skeptical about their credibility (Cotter et al., 2002). Some paid search engines use marketing language that might give users unrealistic expectations about the quality of the results they can yield; these paid search engines may not provide much, if any, contact information that users could not find for free (Boswell, 2007). I did not begin using paid search engines until late into the recruitment process, after the team had already drawn heavily from available free resources. I located one participant via a $40 paid database (PhoneDetective.com), who ultimately requested not to participate in the Time 2 study, and two participants via a $30 paid database (Intelius.com), who did opt to participate. Altogether, this suggests that researchers with limited resources may be successful re-recruiting with free resources and should wait to resort to paid databases until they are confident that have exhausted the data available via free searches.
Additional location strategies. Articles also recommended additional search strategies, which I did not believe would be feasible or sensible for my team. However, other developmental researchers may find these approaches useful.
• Leveraging relationships with partner organizations (e.g., schools, after school 66
programs) to facilitate continued data collection (e.g., Agrawal et al., 1978; Lazorick
et al., 2014). In some cases, organizations such as school districts may help re-locate
original study participants (Tourangeau, Le, & Nord, 2005) or provide researchers
useful contact information (Agrawal et al., 1978). These sorts of researcher-
organization partnerships are common in prospective longitudinal studies with
children (e.g., Tourangeau et al., 2005). Although policies differ from one school
district to the next, the general consensus seems to be that for a school or similar
organization to connect researchers with former participants, they (i.e., families) must
have initially consented to future contact, as well as for the school district sharing
their contact information (Ribisl et al., 1996). Nonetheless, even with these conditions
in place, schools still would likely require submission to the school district research
review board, which is a time-consuming process with no guarantee of additional
participant recruitment information. Researchers also may need to budget to facilitate
this form of collaboration. In some cases, it may be possible to obtain contact data
from schools free-of-charge if a study benefits the district, for example, by providing
data about a school-based program’s effectiveness (P. Godard, personal
communication, August 10, 2016). However, in other cases, school districts may
charge researchers a fee regardless of educational relevance to cover the staff time
invested in pulling the data (S. Dickson, personal communication, August 12, 2016).
Relatedly, some school districts may be more helpful if they have received a grant
related to the study.
• Visiting the neighborhoods participants lived in at the time of the original study (e.g., 67
Haggerty et al., 2008; Ribisl et al., 1996). Because families with young children can
develop attachments to their neighbors and first home, cumulative inertia (i.e.,
resistance to moving the longer a family stays in one location) sometimes sets in until
additional life events impact the probability of moving (such as growing family size,
change in jobs, desire for more space; Huff & Clark, 1978). As such, prior researchers
have had modest success locating original study participants by visiting their former
homes and neighborhoods, and reaching out to their friends and neighbors (e.g.,
Anderson, Huston, Schmitt, Linebarger, & Wright, 2001; Haggerty et al., 2008).
Studies with a primarily middle-SES sample may note particularly high levels of
cumulative inertia and residential stability, as prior research indicates that those with
average income or education levels are less likely to move than persons at the
extremes (Abu-Lughod & Foley, 1960). Accordingly, visiting the neighborhoods
participants lived in at Time 1 could be fruitful for teams working with certain
populations, but should only be attempted in cases where researchers gained
appropriate consent as described above.
• Posting local advertisements. Barakat-Haddad and colleagues (2009) located a small
number (less than 1%) of children in their longitudinal study through advertisements
in local newspapers. In rural, small, or tight-knit communities, local and grassroots
outreach could provide another avenue for locating former participants.
Concluding notes on location. I believe all the location approaches described above complement one another, and indeed, others have suggested multiple recruitment outreach methods often work in tandem (Ribisl et al., 1996). Several parents who previously ignored 68 telephone calls responded positively to calls after receiving a written letter describing the study.
One father emailed after receiving a re-recruitment letter and explained that it piqued his daughter’s interest, at which point I began addressing letters to both parents and children.
Likewise, all the parents who responded to recruitment attempts via text message had received calls and emails from the research team first. Consequently, Table 6 might under-express the value of certain recruitment channels (i.e., parents may have been primed to be receptive to calls from researchers after receiving re-recruitment messages in other forms).
Experienced or committed research staff may be able to brainstorm additional ways to search for contact information specific to local communities or more efficiently leverage the latest search tools. As prior research suggests, more experienced researchers are often stronger recruiters in general (Sugden & Moulson, 2015). I found undergraduate volunteers often needed quite a bit of direction when searching to yield usable location data.
Throughout the location process, the research team maintained a detailed log of communication attempts across platforms, somewhat aligned with recommendations from Cotter and colleagues (2002) and Ribisl and team (1996). For efficiency’s sake, we categorized phone and email contact information according to whether we had (a) confirmed its connection to a participant, (b) denied its connection to the participant (i.e., wrong or inactive phone number or email), (c) not yet tested it for connection, or (d) tested it but not received a definitive confirmation or denial. Though retaining this level of detail did not result in the most streamlined database, this was necessary to avoid wasting time retesting communication avenues already deemed unhelpful. In additional columns in the database, researchers recorded the date and time of previous communication attempts, successful or not, keeping the most recent attempt on top 69 for ease of determining when someone had last been contacted. This allowed for the diversification of timing of attempts.
Scheduling Families for Study Participation
After (or in my case during) the location process, researchers schedule participants for study appointments. In this follow-up study, I completed appointments with 101 families (83% of the located sample and 74% of the full original sample). The remaining located families either refused to participate in the Time 2 study (6 families; 5% of located families), missed or cancelled appointments and were unresponsive to attempts to reschedule (6 families; 5% of located families), or failed to ever schedule appointments before the conclusion of data collection
(8 families; 7% of located families). Of the families I did test, it took on average 21.20 days (SD
= 28.07) from first locating them to completing research sessions with them. Again, I encountered a great deal of variation in ease of scheduling, holding sessions with 40 families
(40% of participating families) within one week of location, 41 (41%) within one month, and 20
(20%) after one month had past. This data mirrors prior work; for example, Cotter and team
(2002) reported that about a third of their pediatric mental health clinic sample was easy to schedule, a sixth required multiple contact attempts before scheduling, and 7-8% refused to participate in some follow-up sessions. Table 6 provides a more detailed breakdown of how long it took to schedule participants after locating them via each method described in the previous section. Among the 122 located families, those who scheduled and attended Time 2 appointments were nonsignificantly less affluent than those located but not scheduled (p = .09).
However, no other fully or marginally significant differences emerged between these groups.
When reaching out to schedule appointments, I included language I thought might 70 motivate busy families to make time for this study (Striano, 2016). For credibility, I referenced my university towards the beginning of most recruitment communications (Haggerty et al., 2008;
Silva, 1990; Sugden & Moulson, 2015). As a potential appeal, I also explained the overarching study goal (Ribisl et al., 1996; Silva, 1990), which was to evaluate the long-term effectiveness of educational computer games created with funding from the U.S. government. I assumed this approach would speak to the parents in the sample, who had enrolled their children in an optional educational computer game intervention when children were in preschool or kindergarten. Later in the recruitment process, I began describing to parents roughly how many families had already participated in the Time 2 study. I intended for this to both legitimize the study and to make parents feel as if they were part of something large and important. I also mentioned that participation would help me conclude my dissertation project and graduate, to associate participating with an additional positive outcome. This personal connection may explain why I had slightly more success recruiting via social media than the other team members.
Furthermore, parents and children were provided $20 for participating, and this was highlighted in recruitment communications. This compensation seemed highly motivating to children but less so to parents. Minimum wage was about $10 per hour in the city where I collected my data, and the cost of living was higher than the national average. The relatively low incentive for this area may explain why some of the affluent located families ultimately did not schedule and follow through with appointments. Other longitudinal work providing larger compensation yielded somewhat higher scheduling rates (e.g., Haggerty et al., 2008), although over-compensating families, especially lower-income families like many in this sample could be considered unethical or coercive (American Psychological Association, 2010; Hartmann, 1992). 71
Initially, I described to parents what I believed were salient aspects of the original study, attempting to trigger positive memories or loyalty to the intervention (see Cotter et al., 2002;
Ribisl et al., 1996 for a discussion on the value of study loyalty and affiliation). However, it became clear over the course of re-recruitment that many parents struggled to recall the Time 1 intervention: Thirteen parents wrote in questionnaires that they did not recall the intervention or wrote comments clearly confusing this study with ones conducted by other groups on other topics, and additional parents verbally mentioned having forgotten the original study. Similar difficulties also have been suggested in other longitudinal studies (Barakat-Haddad et al., 2009).
This may explain in part why the faculty member, who was the only Time 2 study team member involved in the original data collection, was not more successful recruiting via social media – weak personal connection. To further jog memory and enhance credibility, Ibegan taking care to name the preschool or kindergarten where families were initially recruited.
Across all recruitment messaging, some parents seemed to extrapolate a sense that participating in the study would somehow provide children an academic enrichment opportunity, aid them as parents in better facilitating children’s continued education, or otherwise abstractly improve children’s education. Similarly, some parents said that they ultimately decided to participate in the study hoping it would help inspire their children to attend the university where the research team was based. Since different appeals may speak to different families, it may be best to position a study as accruing a variety of tangible benefits (Striano, 2016; Sugden &
Moulson, 2015).
Adhering to recommendations in prior methodological pieces, participants were given a great deal of flexibility in terms of scheduling (Cotter et al., 2002; Ribisl et al., 1996). For 72 convenience and comfort, families were given the choice of participating in their homes (an option chosen by 45 families, 45% of participants), university lab space in a suburb just outside of the urban city families lived in at Time 1 (22 families, 22% of participants), or local libraries with private study rooms (23 families, 23% of participants). It is likely that this flexibility increased participation rates, as logistical barriers such as inflexible work schedules, lack of transportation, or need for child care for younger children are known to impact the ability to retain low-income families in research (Duch, 2005). Additionally, 11 families (11%) who had moved out of the metropolitan area completed online surveys instead of participating in person
(another strategy also recommended in other recent methodological reviews; e.g., Kalkhoff,
Youngreen, Nath, & Lovaglia, 2014). This helped to address physical barriers found in other studies, such as where some participants were unable to be interviewed by phone due to physical constraints (e.g., SEAL training in Alaska, being in prison, and having re-located to Mexico;
Anderson et al., 2001). Most appointment time requests were honred, excluding cases, for example, when families wanted to meet late at night or at local libraries during hours they were not open. As alluded to above, I tried to schedule appointments as soon after reaching participants as possible, ideally within one week of initially locating families (Kalkhoff et al.,
2014; Ribisl et al., 1996). If I did not have much availability over the course of the following week, I held off attempting to contact parents until my schedule was more open rather than seeking appointments weeks in advance. This practice sought to minimize the window in which families could forget about appointments or experience other schedule changes.
Online and digital tools can further help with appointment management. I manually entered study appointments into a digital calendar shared by the research team, emailed parents 73 event appointments that could be added to any personal digital calendars they maintained, and programmed these event appointments to send parents reminders the night before or morning of their scheduled study sessions. Others have reported using digital interfaces such as
YouCanBook.Me to allow participants to privately and relatively independently sign-up for and, if need be, reschedule study appointments, choosing among several session times researchers make available (Kalkhoff et al., 2014). These interfaces also can send automated appointment reminders in advance of sessions (Kalkhoff et al., 2014).
Confirming Appointments
As with all developmental research, the last step in the re-recruitment process was to confirm appointments with parents, making it clear I would be happy to reschedule if need be
(Striano, 2016). Initially, I confirmed appointments exclusively by phone call and email, aligned with methodological recommendations elsewhere (Kalkhoff et al., 2014), but later into recruitment, began calling, emailing, or texting parents, depending on their preferences.
Even though only a small number of parents responded to early locating/scheduling efforts by text message (see Table 1), this communication channel worked well for confirmation, perhaps because families perceived text messaging to be a less formal means of communicating.
Four dyads (11% of in-person appointments scheduled at the time time) missed their appointments before I started texting parents, but this only occurred once (2% of in-person appointments) after I introduced texting (and this one appointment was missed due to an unfortunate family emergency). Parents who needed to cancel or reschedule appointments seemed more comfortable doing so via text rather than over the phone or through email.
Moreover, if researchers suspected a participant might forget about the appointment after official 74 confirmation, they would send casual “on my way” text messages to prompt parents’ memories. I used my personal phone for such purposes, although it might have been wiser to buy a study- specific cell phone or set up a Google Voice account that researchers could share without compromising privacy (M. Smith, personal communication, April 27, 2016). Some of the digital study appointment management tools described in the previous section can automatically send confirmation emails or text messages, and may allow parents to reschedule appointments without needing to interface with a researcher (Kalkhoff et al., 2014).
Conclusion
Across developmental sub-fields – from basic cognition (e.g., Lauer & Lourenco, 2016) and language development (e.g., Can et al., 2013) to mental (e.g., Agrawal et al., 1978) and physical health (e.g., Fein et al., 2014) to applied interventions targeting children (e.g., Campbell et al., 2012) and their families (e.g., Huston et al., 2005) – researchers are engaging in longitudinal work, re-recruiting families who participated in one study to gain a better understanding of children’s developmental trajectories. Frequently, research teams do not decide to begin embarking on this work until after an original study has concluded. My work demonstrates the feasibility of re-recruiting sizeable numbers of urban families after an extended gap in communication with limited financial resources. Future developmental research teams should be able to achieve high follow-up rates by (a) setting up initial studies in which parents provide detailed contact information, including contact information for multiple caregivers, and consent for later waves of research, (b) search across a variety of sources to locate participants, e.g., social media, people-centric search engines, etc., (c) write multifaceted scheduling scripts highlighting the study’s value, and (d) confirm appointments in a way that conveys a casual tone 75 that makes parents feel comfortable, even if they need to reschedule. Many of these strategies may likewise enhance the recruitment process even for cross-sectional research. Developmental researchers also may wish to consult Table 1 to help inform timelines as they plan.
Relative to prior research, I had more success re-recruiting with free tools such as people- centric search engines, and less success using paid and other tools. Differences between my study and others may be attributable in part to advances in modern technology and my team’s efforts to leverage popular technological services. Because parents are increasingly maintaining the same cell phone numbers, especially among high- and middle-income populations, even after they move (Dost & McGeeney, 2016) and because of the existence of a plethora of free people- centric search engines, I had more success using originally provided telephone numbers and free search engines than researchers reported previously. I similarly found text messaging, which is currently very popular in the U.S. (Duggan, 2013), to be helpful in ensuring participants maintained their appointments or felt comfortable rescheduling if necessary. Indeed, even researchers conducting cross-sectional work may wish to consider incorporating text messages into their appointment confirmation protocols. I only saw limited success locating participants via social media or paid databases, somewhat contrasting prior research where paid databases were more effective (e.g., Haggerty et al., 2008). However, I suspect my team’s location findings would have varied had we engaged the various search methods in a different order.
Technology-related advances aside, this study also reinforces the value of strategies others have suggested to schedule participants and calls into question assumptions about the ease of scheduling particular groups of participants. My experiences underscore the importance of collecting detailed contact information during an initial study (Ribisl et al., 1996) and using 76 recruitment (or re-recruitment) scripts that make the university affiliation and study goals clear
(Sugden & Moulson, 2015). Such re-recruitment scripts may be especially valuable in cases where researchers are focused on academic, prosocial, or other potentially positive outcomes.
Moreover, emphasizing the university and study aims may even be more worthwhile than reminding families of the particulars of the original study. Like others (see Ribisl et al., 1996), I also believe my re-recruitment success is in part attributable to the fact that I planned Time 2 study activities in a way that allowed me to conduct research in a variety of settings (i.e., lab, library, home). However, I recognize such flexibility may not be feasible for all development sub-domains, such as when researchers are interested in collecting neurological data (e.g.,
Schwartz et al., 2003).
Like some studies, I had somewhat more success locating families with higher-incomes, as well as Caucasian families (e.g., Fein et al., 2014). At least for this study, I expect these two seemingly different findings may both trace back to stability issues related to family income, rather than anything cultural. Caucasian families in this study were more affluent than the rest of the sample (p < .001), a trend that bears out nationally in the U.S. (Wilson, 2015). In general and across racial-ethnic groups, low-income families are more likely to experience disruptions in phone service, changes to cell phone numbers (Ahlers-Schmidt et al., 2012), and physical relocation over time (Abu-Lughod & Foley, 1960), factors which in turn might make them particularly challenging to locate. However, mirroring findings that contradict my location results (e.g., Silva, 1990), I was arguably less successful scheduling higher-income participants, who may have been unmotivated to find time to participate. To increase scheduling rates, those with greater resources may wish to provide larger monetary compensation or even more fervent 77 appeals to the study’s mission to help attract more affluent potential participants.
As is typical in longitudinal research (Ribisl et al., 1996), I attempted to re-recruit all of the original families and used all of the allotted appointment time during each research session to gain information about children’s present, study-specific functioning. Consequently, I did not formally survey parents about their perceptions of the re-recruitment experience or experimentally compare the effectiveness of different recruitment strategies (which might have resulted in losing participants assigned to less successful recruitment strategy conditions).
Moreover, I refined my recruitment approach over the course of data collection, as is common in research of this nature (see Ribisl et al., 1996). Future researchers with larger initial samples or more time and financial resources should consider formally testing some of the assumptions in this article; such work would address important gaps in the methodological literature.
Despite the aforementioned limitations, I hope my efforts can guide other developmental scientists interested in conducting longitudinal research. Moreover, some of these strategies may even positively impact recruitment for cross-sectional studies. Given the power longitudinal studies have to clarify developmental trajectories (Nicholson et al., 2002) and provide compelling accountability evidence for interventions (e.g., Barnett, 2013), and given the growing interest in work of this nature (e.g., Wartella et al., 2016) it is important developmental scientists feel capable of re-recruiting sufficiently large samples, even with limited resources and even when they decide to begin such work long after the conclusion of a particular study. 78
Article 3: Did We Succeed in “Raising Readers”? Effects of Ready To Learn Early
Childhood Literacy Computer Games in Middle Childhood
Abstract
The majority of young children in the U.S. have consumed media created with funding from the
Department of Education’s Ready To Learn (RTL) initiative. RTL aims to promote foundational early learning skills (e.g., basic literacy) so that children are prepared to succeed in formal schooling. The Early Learning Hypothesis predicts early exposure to such educational media can catalyze long-term academic success. This study assessed whether the effects of RTL media in early childhood sustain into middle childhood. One-hundred-and-one youth who had participated in an evaluation of an RTL literacy-themed computer game in early childhood were re-contacted in middle childhood. Their present-day literacy skills were assessed, and parents provided complementary data. A curvilinear relationship between children’s early childhood pretest scores and middle childhood outcomes was detected. The positive effects of the games sustained into middle childhood, but only for children with below and above average literacy skill prior to the original intervention. These findings are interpreted in light of the Traveling Lens and Capacity
Models, two theoretical frameworks outlining the relationship between children’s skillsets and their receptivity to educational media.
Keywords: Ready To Learn, educational media, longitudinal, early intervention
79
The U.S. Department of Education’s (DoEd) Ready To Learn (RTL) initiative provides
$25 million annually for the development of public mass media intended to promote school readiness (Bryant et al., 2001; DoEd, 2015; Singer & Singer, 1998). RTL has provided an added stream of funding for some longstanding educational media properties that premiered prior to the launch of the grant, such as Sesame Street, along with seed money for new properties that did not exist before RTL, including Between the Lions, Super WHY!, and WordWorld (Corporation for
Public Broadcasting & Public Broadcasting Service [hereafter CPB and PBS, respectively],
2011). Today, the majority of young children in the U.S. have been exposed to media paid for in part by RTL (CPB & PBS, 2011). Research suggests this RTL-funded media successfully has promoted basic early academic skills (e.g., alphabet knowledge) in the short-term (see Article 1), but whether this expenditure has resulted in more children truly being ready to succeed in school is an open question. Theoreticians believe early educational media exposure can catalyze long- term learning, an idea known as the Early Learning Hypothesis (Anderson et al., 2001). To date, this hypothesis largely has been substantiated by correlational or indirect measures of the long- term effects of educational media (e.g., Kearney & Levine, 2015; Rosser et al., 2007).
In this study, children who participated in an evaluation of RTL-funded literacy-themed computer games in early childhood were re-contacted to determine if the positive effects of exposure to these games sustained into middle childhood. These findings provide accountability evidence for the RTL initiative, validate the Early Learning Hypothesis with causal data, and speak to other scholarly debates on children’s learning from media.
Ready To Learn Initiative 80
RTL launched in 1994, a time when the American public was concerned about the dearth of high quality children’s television programming and reports of large numbers of young children entering elementary school ill-prepared for formal schooling (Bryant et al., 2001). The intention of the initiative was to fund mass media that would help America’s school children, especially at-risk populations, gain foundational early learning skills necessary for success in elementary school and beyond (Singer & Singer, 1998). At around the same time, commercial media providers in the U.S. concurrently began providing an increasing volume of educational fare (Bryant et al., 2001), and numerous international content producers have created similar media (e.g., Mares & Pan, 2013).
Unlike many of these other content providers, RTL grant recipients uniquely have made most of their content publicly available for free or at a low cost (CPB & PBS, 2011), have promoted positive parenting practices and to a lesser extent the RTL media via social marketing campaigns (with taglines like “Raising Readers” or “Anytime is learning time"; Hurtado, Galdo,
Agin, & Heil, 2010), and have created outreach programming (e.g., media-themed summer camps) intended to complement and extend the lessons present in the media (Llorente et al.,
2010). Together, these factors have helped to make RTL media wide-reaching (CPB & PBS,
2011). RTL media producers also have been at the forefront of experimenting with newer forms of media; starting in 2005, the grant explicitly tasked content producers with creating new media tools, such as educational computer games (Michael Cohen Group, 2012). The present study aims in part to evaluate the long-term success of one such tool: an educational computer gaming suite that promoted foundational reading skills (Schmitt, Sheridan Duel, & Linebarger, 2017).
Educational Media as a Catalyst for Long-Term Learning and Growth 81
With their Early Learning Hypothesis, Anderson and colleagues (2001) propose that early exposure to educational television, such as the programs funded by RTL, can spark growth and learning that persists throughout children’s time in school. They suggest children who acquire key school readiness skills from educational television in early childhood may be better prepared for elementary school, initially placed in higher ability groups in school, and thus set on a trajectory of continued success. The present study explores whether exposure to interactive media such as RTL computer games also leads to similar long-term growth.
Some empirical evidence substantiates the Early Learning Hypothesis. A host of studies following up with educational media evaluations provide causal evidence that positive effects from educational media are detectable several months after initial exposure. For instance, in an evaluation of an RTL-funded vocabulary intervention, program participants retained vocabulary gains six months after the intervention had concluded (Neuman et al., 2011). Likewise, in another RTL intervention, young children who watched episodes of literacy-themed Super WHY! towards the beginning of the school year outperformed their peers at the end of the year on a measure of letter sound knowledge, and children who watched the show and played complementary online games outperformed peers on measures of lower case letter knowledge and rhyme awareness (Linebarger, 2010). However, contrary to the Early Learning Hypothesis, children in the study’s control group who did not consume any media had the strongest long- term performance on a measure of beginning word sound awareness (i.e., understanding of the sounds at the beginning of words; Linebarger, 2010).
Similar studies looking at other media also found results consistent with the Early
Learning Hypothesis. For example, the positive effects of a noncommercial, literacy-themed 82 computer game that promoted letter-sound knowledge sustained several months after an initial effectiveness evaluation (Segers & Verhoeven, 2005). Likewise, kindergarten students who played computer games that promoted blending (the ability to combine letter sounds into words) demonstrated stronger reading skills mid-way through first grade (Reitsma & Wesseling, 1998).
And kindergartners who played a mix of literacy-themed computer games and apps evinced stronger literacy on composite measures assessing competence across several early literacy skills at multiple time points across first grade (Ponciano & Thai, 2016).
A small number of scholars have examined even longer-term impacts from educational media (mostly television) prior to the launch of the RTL initiative using correlational or indirect methods. The evidence these studies provide is mixed. To illustrate, in a correlational study, preschoolers who watched child-targeted television as in the early 1980’s (including educational programs such as Sesame Street, which at the time had not yet received RTL funding) read more books and achieved better grades in high school English, math, and science (Anderson et al.,
2001). In contrast, in another quasi-experimental study, there were no differences in high school reading, vocabulary, or math performance between preschoolers who lived in and outside communities with access to Sesame Street in the late 1960’s (again many years before Sesame
Street received RTL funding; Kearney & Levine, 2015). However, in the same study, those living in communities with Sesame Street demonstrated stronger labor force outcomes despite the lack of evidence of differential school performance (Kearney & Levine, 2015). Differences between these two studies may be driven in part by changes in Sesame Street’s curriculum over the past half-decade (Fisch & Truglio, 2000) or as a result of methodological differences.
Differential Susceptibility to Media Effects 83
Per other media theories, it may be overly simplistic to assume all children would realize comparable long-term benefits from early educational media exposure, and indeed, a focus on individual differences between children may help explain why different studies yield seemingly conflicting findings (Piotrowski & Valkenburg, 2015; Valkenburg & Peter, 2013). Turning to theories originally conceived to explain short-term learning from media may shed insight into which children are most likely to benefit from such exposure in the long-term.
The Traveling Lens Model posits children gain the most from media when they perceive its content as being moderately difficult – not too easy or too challenging (Huston & Wright,
1989; Rice, Huston, & Wright, 1982). Specifically, the model predicts children’s interest in and attention to media will be strongest for media content they find moderately difficult, and this interest and attention will facilitate greater learning. The model goes on to suggest that as children’s thinking becomes more advanced or their familiarity with the mediated content increases, their interest, attention, and learning from a given media stimulus may decline. In such situations, children might begin attending to more challenging aspects of favored media products or seek out more challenging media content entirely. In today’s new media environment, replete with active and interactive games, the media itself might “level up” to continuously present children with content aligned to their evolving skillsets (Guernsey, 2012; Hirsh-Pasek et al.,
2015; Walker, 2011). Extending this line of thinking, perhaps educational media may have more pronounced long-term effects for children who found content moderately difficult at initial exposure and in cases where media present children increasingly challenging content over time.
The evidence is mixed as to whether, in the short-term, RTL-funded media were most effective for children with below or above average abilities. That is, it is unclear if RTL media 84 content aligned best with the abilities of weaker or stronger early readers. For instance, in an evaluation of the RTL television show WordWorld, children with higher initial literacy skills prior to WordWorld exposure scored higher on posttest measures of word recognition, while children with lower initial literacy demonstrated greater gains in phonemic awareness (i.e., mastery of language sounds; Michael Cohen Group, 2009). In other studies focused on literacy summer camps using curricula and media from Super WHY! and on Between the Lions television episodes, children with average and above average literacy at pretest benefited most from media exposure (Jennings, 2013; Linebarger, Kosanic, Greenwood, & Doku, 2004). Conversely, in another evaluation, children under the age of 4.5 only benefited from RTL-funded Pocoyo apps if their English language skills were poor pre-exposure (Michael Cohen Group, 2013). This substantial variation makes it difficult to say whether the body of RTL short-term evaluations substantiate the Traveling Lens Model. It may be that media’s alignment to children’s skillsets varies across different RTL media products and/or across different samples of children.
The Capacity Model proposes somewhat different mechanisms undergirding learning from media (Fisch, 2000, 2004, 2016). This model is most well-known for its nuanced discussion of the characteristics of media that lead to learning, but, especially pertinent to the current discussion, it also outlines individual differences that may make certain children especially receptive to educational media (Aladé & Nathanson, 2016). The model posits children with stronger media-related subject matter knowledge, interest, and, for some metrics, verbal ability should gain more from educational media because they should be able to assimilate the narrative and/or educational messages presented in the media more efficiently, easily, and readily (Fisch,
2004). That is, the model implies a linear relationship between subject matter knowledge, 85 interest, and verbal ability, and learning (however, Fisch never uses the word “linear” to describe these relationships, and in a footnote focused on the characteristics of media that facilitate learning, acknowledges the possibility of nonlinear relationships in certain cases; Fisch, 2004).
In a validation study, Aladé and Nathanson (2016) examined these assumptions by asking preschool-age children to view a science-themed television episode and by administering a battery of cognitive and interest assessments pre- and post-exposure. Consistent with the
Capacity Model, they found children with stronger verbal skills and higher levels of science subject matter knowledge learned more from the episode. Although Aladé (2013) noted children’s interest in science was correlated with their verbal ability, children’s scientific interest did not predict learning of science content like verbal ability did (Aladé & Nathanson, 2016).
That said, Aladé and Nathanson (2016) questioned whether their pattern of results would have looked the same had they operationalized “interest” differently, perhaps asking children how motivated they were to attend to and learn from the science program.
Despite these generally supportive data from Aladé and Nathanson (2016), the RTL literature reviewed above calls into question the assumption that prior skill, verbal ability, and subject matter interest linearly predict learning (e.g., Michael Cohen Group, 2009, 2013). Also, similar to the Traveling Lens Model, to my knowledge, the Capacity Model’s tenets about child characteristics have only been examined in the short-term, and it therefore is difficult to say if and how these individual differences would manifest in the long-term.
The Current Study
This investigation followed up with children who participated in an RTL literacy evaluation in early childhood, re-assessing their abilities in middle childhood, six years after the 86 original intervention. The primary aim was to better understand whether the RTL initiative has resulted in the creation of media capable of stimulating meaningful and long lasting impacts on children’s school performance. The study intended to provide data extending scholarly understanding of the Early Learning Hypothesis, testing this hypothesis with causal data focused on the long-term effects of educational computer games intended to support literacy learning and exploring whether learning varies as a function of children’s skillsets prior to media exposure.
Original early childhood RTL intervention. In 2010, Schmitt and colleagues (2017) evaluated an RTL literacy computer game with a diverse sample of preschool and kindergarten children living in a city in the Midwestern U.S. Researchers randomly assigned about one third of the sample to play a control suite of computer games created by a commercial company, which featured non-educational puzzles and arts activities. They assigned the remaining two- thirds of the sample to play a leveled RTL literacy computer game suite (PBS KIDS Island) featuring the media properties Between the Lions, Martha Speaks, Sesame Street, Super WHY!, and WordWorld.
The RTL literacy gaming suite presented eight sequenced levels that each contained four mini-games focused on different literacy skills. Once children successfully completed all four mini-games in one level, they advanced to a new, more challenging level. Early levels promoted relatively basic literacy skills (see Grant et al., 2012 for more on literacy learning sequences).
For example, in levels one and two, children played mini-games in which they put together jigsaw puzzles featuring various letters of the alphabet as a means of enhancing letter recognition. Later levels promoted more advanced skills. For instance, mini-games in levels four
- six presented children with challenges they solved by correctly choosing words that rhymed 87 with a target. All children completed the first three levels before the conclusion of the study, and
65% advanced all the way to the final level (i.e., level 8).
Relative to the control group, children who played the RTL literacy games for 6-8 weeks scored higher on three assessments measuring mastery of English language sounds (i.e., beginning sound awareness, letter-sound knowledge, letter-sound fluency, and phonological awareness), as well as on measures of letter sequencing and vocabulary. Parents also indicated the games promoted children’s letter recognition, word recognition, spelling, and self-esteem.
Middle childhood follow-up. In the present evaluation, families who participated in the above early childhood RTL intervention were re-contacted. The aim was to assess whether the positive effects of early exposure to the RTL literacy games sustained into middle childhood, six years after the intervention. More specifically, this study asked (RQ1a) whether children who played the RTL literacy games still evinced stronger literacy skills in middle childhood, and
(RQ1b) whether these effects varied as a function of children's literacy skill prior to the intervention.
Additionally, the study asked (RQ2a) whether long-term effects generalized to other learning domains beyond literacy. Even though RTL media often focus on specific early learning domains such as literacy (see Article 1), the intention of the initiative is to broadly prepare children for success in school (Singer & Singer, 1998). Additionally, other research suggests literacy undergirds success in a variety of school subjects, because so many school courses require children to read texts and process written and verbal directions (Schiefele, Schaffner,
Möller, & Wigfield, 2012). Moreover, prior research also indicates that children might be able to apply or transfer the skills they learn in media to other contexts in cases in which they encode 88 lessons in a manner abstracted from the original media (e.g., consuming literacy-themed media focused on the /L/ sound and later sounding out the word “log” during story time in school or
“lung” during a science lesson on the body; for a longer discussion, see Fisch, 2004; Fisch,
Kirkorian, & Anderson, 2005). The study further questioned (RQ2ab) whether these long-term transfer effects also varied as a function of children’s literacy prior to intervention.
Method
Sample
Families in the original study were recruited from 15 schools in a major city in the
Midwestern U.S. in Winter and Spring 2010. One-hundred-thirty-six children (48% male) completed the full intervention. Of these, 93 were assigned to the RTL literacy group, and 43 were assigned to the control group. Across the full sample, 94 children were in preschool during the original study, and 42 were in kindergarten. The sample was racially and ethnically diverse
(28% Caucasian, 28% Hispanic, 20% African American, 7% Asian, and 17% other/mixed background). An effort was made to recruit families across the socio-economic spectrum, with an average sample income-to-needs ratio1 of 1.96 (SD = 1.42).
1 Income-to-needs ratios were calculated via a three step process. First, researchers asked parents in early childhood to report their household income using a 9-point scale ranging from under $20,000 (1) to over $125,000 (9). Second, researchers identified the midpoint for each of these nine income ranges, and divided them by a correction factor to adjust for the cost of living in the city where the research originally took place, which was above the U.S. national average (American Chamber of Commerce Research, 2007). Finally, for each family, researchers divided these adjusted estimates of income by the federal poverty threshold for a family of that size, based on poverty data provided by the U.S. Census (Institute of Research on Poverty, 2010). 89
After receiving approval for the follow-up study from my university Institutional Review
Board (IRB), families from the original intervention were re-contacted in Spring and Summer
2016. Of the 136 original participating families, 101 participated in the follow-up study (74% of original participants). This follow-up rate is on par with rates achieved in comparable longitudinal intervention evaluations (e.g., Huston et al., 2005). Of the 35 families who did not participate in the follow-up, six refused to participate, 15 were located but did not attend a study session before the conclusion of data collection, and 14 were unlocatable. Relative to the families who did participate in the follow-up study, children in the missing group were more likely to be of other/mixed race (p < .05). Besides race-ethnicity, there were no other significant differences between the families who participated in this follow-up and those who did not.
Table 7 provides full demographic information for the sample who participated in the middle childhood follow-up. Among the full follow-up sample, those who were in the RTL literacy group during the original intervention were slightly more affluent than their control group counterparts (a difference that was nonsignificant in the original evaluation). But otherwise, the makeup of families across both conditions was comparable.
Table 7. Demographics for Children Who Participated in the Middle Childhood Follow-up
All Treatment Control (N = 101) (n = 65) (n = 36) Age in Years in Middle Childhood 11.28 (1.30) 11.35 (0.77) 11 .14 (1.93) Grade in Middle Childhood 4th or below 17% 17% 18% 5th 52% 46% 62% 6th 27% 32% 18% 7th or above 4% 5% 3% 90
Female 55% 54% 56% Child Race-Ethnicity Caucasian 31% 37% 19% African American/Black 21% 15% 31% Hispanic 29% 28% 31% Other or Mixed 20% 20% 19% Out-of-Area Residence 11% 12% 8% in Middle Childhood Maternal Level of Education at Middle Childhood Follow-up Less than a Bachelor’s 47% 42% 56% Bachelor’s or Higher 53% 58% 44% Income-to-Needs Ratio 1.94 (1.37) 2.12 (1.39) 1.60 (1.29) in Early Childhood† Income-to-Needs Ratio 2.30 (1.44) 2.55 (1.41) 1.80 (1.40) in Middle Childhood* Receipt of Public Assistance 28% 24% 37% Pretest Composite Scores -0.03 (1.00) 0.01 (1.04) -0.11 (0.95) in Early Childhood
Note. Means with standard deviations in parentheses for continuous variables and frequencies for categorical variables. Listwise deletion was used for item-level missing data. † p < .1, * p < .05.
Procedure
The protocol for the middle childhood data collection varied depending on families’ residence at follow-up. Families still living in the area where the original study was conducted (n
= 90) were given the option of completing an in-person research session in their home, at my university lab facility, or at a neighborhood library. During these sessions, parents completed a
25-minute questionnaire, while children completed a 40-60-minute researcher-led battery of assessments. Parents of families who had moved (n = 11) were emailed an online version of the 91 parent questionnaire. These parents also were given the option to complete the questionnaire verbally over the phone, although no parents opted to do so. As part of both in-person and online data collection, parents consented to allow researchers to access the data they and their children provided in early childhood and link it to the data provided at follow-up. All children who participated in the study received $20 cash. Parents who completed the questionnaire during an in-person session also received $20 cash, and parents who completed the online version of the survey received $20 Amazon e-gift cards.
Measures
Early childhood direct measures of child literacy. In the original early childhood intervention, children completed a series of standardized and custom literacy assessments at pretest and posttest. At both time points, children completed the Phonological Awareness
Literacy Screening for Preschool (PALS-PreK; Invernizzi et al., 2004) Alphabet Knowledge,
Letter Sounds, Beginning Sound Awareness, and Rhyme Awareness subtests, as well as the Get
Ready to Read! Screener, a composite measure of early literacy covering a variety of literacy sub-skills (Whitehurst & Lonigan, 2001). Children also completed researcher-developed measures of letter sequencing, phonemic awareness, and vocabulary. See Schmitt and team
(2017) for a fuller description of these measures. Using the psych package in R (Revelle, 2016), an item cluster analysis suggested all of these measures clustered together at pretest (Cluster fit =
.94, Pattern fit = .99, RMSR = .08, Cronbach’s α = .86). For the current study, a pretest composite score for each child was created using regression weights based on the cluster analysis. The average early childhood pretest composite score for the full original sample was 0
(SD = .98, Range: -2.8 to 1.12), and for the subset of children who participated in the middle 92 childhood follow-up was -.03 (SD = 1.00, Range: -2.8 to 1.12).
Middle childhood direct measures of child literacy. Schmitt and colleagues’ (2017) original data provided fairly convincing evidence that the RTL literacy games promoted children’s phonological awareness (i.e., the ability to identify and manipulate language sounds), with statistically significant findings at the .05 level for three measures of phonological awareness, and an average effect size of Cohen’s d = .2 across all measures of phonological awareness. The present study focused on children’s present day rhyming (a skill falling under the broader umbrella of phonological awareness), deletion (the ability to remove sounds from the beginning, end, or middle of words; another phonological awareness subskill), segmentation (the ability to break up words into smaller linguistic units such as syllables or letter sounds; a third phonological awareness subskill), decoding (applying letter-sound knowledge to decipher words and pseudowords), spelling, and reading. Previous research suggests all the aforementioned literacy skills are predicted by early phonological awareness (e.g., Juel, 1988; Scarborough, Ehri,
Olson, & Fowler, 1998; Wagner et al., 1997). To assess these skills, several sub-tests from the
Kaufman Test of Educational Achievement, II (KTEA-II; Kaufman & Kaufman, 2004a) were administered. The KTEA-II was normed with 200 school-age children, and split-half reliability coefficients for most subtests were .90 or above (Bonner & Carpenter, 2005).
Rhyming. Children’s rhyming was assessed with the Rhyming section of the KTEA-II
Phonological Awareness sub-test. For this measure, children were asked to aurally discriminate between rhyming and non-rhyming words. A researcher read children six sets of words with four words each. Within each four-word set, three words rhymed, and one word did not. Children earned one point for correctly identifying the non-rhyming word in each set, for a maximum 93 score of 6 points (M = 5.78, SD = .67, Range: 2 to 6).
Deletion. Children’s deletion was assessed with the Deletion section of the KTEA-II
Phonological Awareness sub-test, which required children to delete sounds from target words.
For example, a researcher asked children to “Say make, but without the /k/ sound”, and children earned a point if they correctly deleted the /k/ sound to verbally form the word “may”. Children completed six of these deletion items, again for a maximum potential score of 6 points (M =
5.19, SD = 1.13, Range: 1 to 6).
Segmentation. Children’s segmentation was assessed with the Segmentation section of the KTEA-II Phonological sub-test, which required children to separate words into smaller linguistic units. In the first part of this task, children broke four words into distinct syllables (e.g., spa…ghe...ti), earning one point for each word correctly segmented. Two of these four target words contained two syllables, and the other two contained three. In the second part of the task, children had to segment five words into distinct sounds or phonemes (e.g., say “crust” as
“/k/…/r/…/uh/…/s/…/t/”), again earning one point for each correctly segmented word. Thus, altogether children could earn 9 points for Segmentation (M = 4.88, SD = 1.92, Range: 2 to 9).
Decoding. Children’s decoding was assessed via the KTEA-II Nonsense Word Decoding sub-test. In this test, children read aloud an increasingly challenging set of pseudowords (e.g.,
“kimp”, “sprewful”). Children could earn up to 50 points on this sub-test for correctly decoding all pseuodowords; however, administration of the sub-test terminated once children incorrectly decoded four words in a row (M = 26.01, SD = 9.04, Range: 0 to 42).
Spelling. Children’s spelling was assessed via the KTEA-II Spelling sub-test. In this test, a researcher read a target word, contextualized it by reading it in a sentence, and read it one more 94 time. Children then had to spell (i.e., write out) each word, earning one point for each correct spelling. Administration of the sub-test terminated after children spelled four items incorrectly in a row. Children could earn up to 41 points on this test (M = 25.51, SD = 7.73, Range: 2 to 38).
Reading. Children’s word-level reading was assessed with the KTEA-II Word
Recognition sub-test. Similar to the Nonsense Word Decoding sub-test, this battery required children to read aloud a series of increasingly difficult words (real words this time), awarded one point for each word children read correctly, and terminated after children read four in a row incorrectly. Children could receive scores of up to 75 (M = 40.78, SD = 10.04, Range: 6 to 66).
Parent measures. Parents rated children’s school performance across seven subject areas: reading, writing, science, social studies, mathematics, speaking/listening, and following directions. They rated how well children performed in each area relative to other children that age, using a five-point scale ranging from Poor (1) to Excellent (5). There was excellent internal consistency across parents’ ratings in all seven subject areas (Cronbach’s α = .92). A composite measure of parents’ ratings across all subject areas (averaging across their seven ratings2; M =
3.82, SD = 0.90, Range: 1.14 to 5.00) was examined, along with each indicator independently to investigate long-term effects on children’s literacy and downstream transfer effects on other learning domains beyond literacy.
2 a composite score averaging across all parent ratings (i.e., a mean) was used so that this score would be on the same scale as each individual item, thereby facilitating ease of interpretation. However, results were similar in alternate models that used a composite based on the same procedure used for the early childhood pretest composite. 95
Parents also provided updated demographic information, including information on family size and income used to calculate family income-to-needs ratios3 at follow-up.
Results
Data were analyzed in R-3.31 (R Core Team, 2016) using multiple regression to test whether assignment to the RTL literacy group predicted either children’s middle childhood literacy outcomes or children’s parental ratings of academic performance, and controlling for child sex, age at follow-up, income-to-needs ratio, change in income-to-needs ratio, residence, pretest score, and condition assignment (RTL literacy v. control). Additionally, to test RQ1b and
RQ2b (whether effects varied based on children’s skillsets prior to the intervention), models included a Condition x Pretest interaction term. Moreover, because it was unclear whether the media would be more effective for children with below average, average, or above average pretest scores in early childhood, models also included a squared pretest term and a Condition x
3 Similar procedures to Schmitt and team (2017) were used to calculate updated income-to-needs ratios in the present study. First, parents indicated their family income using a 14-point scale ranging from under $5,000 (1) to over $200,000 (14). The midpoint for each range on this scale was then divided by a correction factor to account for the cost of living in the area where data collection activities occurred at the time of the follow-up (Council for Community and Economic Research, 2010). For each family, this adjusted income was divided by the latest estimates of the poverty threshold for a family of that size (United States Census Bureau, 2016). To avoid multicollinearity, models included families’ income-to-needs ratios during the original testing, as well as a change in income-to-needs term, which subtracted families’ income-to-needs ratios in early childhood from their income-to-needs ratios at follow-up. This term controlled for any relationship between increases or decreases in family wealth over the six years between data collection periods, and children’s middle childhood outcomes. 96
Pretest2 interaction term in the models. Cubic pretest and Condition x Pretest3 interaction terms were initially added, but ultimately omitted these from the final models because neither were significant in any model (Stock & Watson, 2006).
In conducting these analyses, item-level missing data were imputed using predictive mean matching implemented via multiple imputation by chained equations with the mice package in R (van Buuren & Groothuis-Oudshoorn, 2011). One or more items included in the models were missing for 23% of the sample: Family income information was missing for 13 children, and one or more middle childhood outcome measures were missing for 12 children
(with two children missing both family income and one or more outcomes). Gender, age, pretest, and intervention condition assignment information were available for all children. To impute the missing items, a subset of the data that included all variables in the statistical models, along with variables that might be related to those in the model (e.g., whether a family received welfare, which might predict family income) was isolated. For each variable, the software replaced missing cells with values provided by other families who seemed to be similar based on available data. The software repeated this process five times to create five imputed datasets. For a relatively small dataset like this one, five imputations were sufficient (Allison, 2015). When conducting regressions, the software pooled across these five plausible imputed datasets.
Long-Term Literacy Development
In the first set of analyses, children’s middle childhood literacy assessment scores were set as dependent variables to test RQ1a and RQ1b. The beta estimate for the Condition term was nonsignificant across all models. However, there was consistent evidence across most outcomes for a positive Condition x Pretest2 interaction. As shown in Table 8, this interaction term was
Table 8. Child Literacy Outcomes in Middle Childhood (N = 101)
Rhyming Deletion Segmentation Decoding Spelling Reading b SE b SE b SE b SE b SE b SE Female 0.05 0.15 0.27 0.23 -0.25 0.42 1.48 1.74 1.86 1.45 -0.76 4.10 Child Age -0.05 0.06 0.11 0.09 -0.12 0.17 0.13 0.73 0.75 0.62 -0.66 0.71 at Follow-up Income-to-Needs 0.05 0.06 0.17† 0.09 0.20 0.18 1.23 0.72 0.94 0.63 1.74* 0.66 in Early Childhood Change 0.04 0.10 -0.24 0.15 -0.44 0.31 1.01 1.24 -0.07 1.07 2.79* 1.19 in Income-to-Needs Out-of-Area Residence -0.46 0.40 -0.78 0.54 -0.06 0.71 -2.48 3.15 -1.13 3.25 -0.76 4.10 Pretest in Early Childhood 0.18 0.14 0.22 0.19 0.73* 0.36 3.11* 1.51 3.60** 1.28 5.80*** 1.47 Pretest in Early Childhood2 -0.11 0.08 -0.14 0.11 0.06 0.21 -0.24 0.91 -0.19 0.75 0.50 0.84 Condition -0.26 0.21 -0.35 0.29 0.29 0.54 -2.84 2.38 -2.43 2.00 0.68 2.24 Condition x Pretest 0.09 0.23 -0.18 0.33 -1.18† 0.63 2.76 2.82 1.09 2.31 1.50 2.52 Condition x Pretest2 0.38** 0.14 0.38† 0.21 -0.48 0.39 4.32* 1.71 2.72† 1.42 3.18* 1.60 R2 0.30 0.28 0.18 0.26 0.35 0.46
Note. Missing cases were imputed using predictive mean matching. † p < .1, * p < .05, ** p < .01, *** p < .001
97
98 significant for Rhyming, Decoding, and Reading, and approached significance for Deletion and
Spelling. This means that RTL literacy group children with low and high pretest scores in early childhood outperformed counterparts in the control group with comparably low and high pretest scores, but that children with average pretest scores tended to score similarly regardless of condition assignment.
To illustrate this more concretely, a child in the RTL literacy group whose early childhood pretest score was one standard deviation below the mean would be expected to score
2.51 points higher in Reading in middle childhood than a peer in the control group who also scored one standard deviation below the mean before the intervention. Put differently, an RTL literacy group participant whose pretest score was one standard deviation below the mean would be expected to be reading approximately one grade level higher than his/her peer in the control group with the same early childhood pretest score (i.e., 4th grade vs. 3rd grade reading level;
Kaufman & Kaufman, 2004b). Likewise, an RTL literacy group participant whose early childhood pretest score was one standard deviation above the mean would be expected to out- perform a comparable control group peer by 5.13 points or two grade levels (i.e., 9th vs. 7th grade performance) in middle childhood (Kaufman & Kaufman, 2004b). Children, however, who scored “average” at early childhood pretest would be expected to score similarly in middle childhood regardless of condition (children in the RTL literacy group only demonstrated a .64 point advantage, with both groups scoring at a 6th grade level; Kaufman & Kaufman, 2004b).
Thus, addressing RQ1a and RQ1b, the RTL literacy games did seem to have affected children’s middle childhood literacy, but effects were strongest for children who scored on the lower and higher end at pretest prior to early childhood media exposure. 99
The only exception to the general pattern found for the literacy skills assessed (i.e.,
Rhyming, Decoding, Reading, as well as the trends for Deletion and Spelling) was for
Segmentation. For this outcome, the Condition x Pretest term was negative and trended toward significance. This meant that for Segmentation, there is some evidence that the RTL literacy games had long-term benefits only for children with below average early childhood pretest scores, but led to lower scores for children with above average scores.
Parental Ratings of Children’s Performance
Next, parents’ ratings of children’s school performance were examined as a means of determining if the effects of the intervention generalized/transferred beyond literacy (addressing
RQ2a and RQ2b). As with the literacy results, the beta estimate for Condition was nonsignificant overall and for each individual skill area. However, again mirroring the literacy-specific findings, the key Condition x Pretest2 interaction for parents’ ratings of children’s overall school performance was significant. Interestingly, the Condition x Pretest2 interaction also was positive and significant for parents’ ratings of children’s performance specifically in science and social studies. Thus, based on parents’ ratings, there is some evidence the RTL literacy games’ long- term effects extended to school areas beyond English Language Arts, although only for children whose early childhood pretest scores were lower or higher than average. The Condition x
Pretest2 interaction term also trended towards significance for parents’ ratings of children’s reading and writing, consistent with the direct measure findings reported earlier.
Table 9. Parent Ratings of Children’s Academic Performance in Middle Childhood (N = 101)
Overall Reading Writing Science Social Math Speaking/ Following Performance Studies Listening Directions b SE b SE b SE b SE b SE b SE b SE b SE Female 0.53** 0.16 0.60** 0.20 0.94*** 0.22 0.42* 0.19 0.41† 0.21 0.27 0.24 0.43* 0.17 0.65** 0.20 Child Age -0.05 0.07 -0.11 0.09 -0.07 0.09 -0.06 0.08 0.02 0.11 -0.09 0.10 -0.04 0.07 -0.02 0.08 at Follow- up Income-to- 0.02 0.06 0.05 0.08 0.02 0.09 0.08 0.07 0.06 0.08 0.02 0.10 -0.05 0.07 -0.01 0.08 Needs Early Childhood Change in 0.14 0.12 0.39* 0.16 0.17 0.17 -0.02 0.14 0.10 0.16 0.13 0.17 0.15 0.13 0.07 0.15 Income-to- Needs Out-of- -0.33 0.26 -0.08 0.33 -0.29 0.35 -0.22 0.30 -0.55 0.34 -0.44 0.39 -0.36 0.28 -0.34 0.32 Area Residence Pretest 0.36** 0.14 0.54** 0.17 0.31† 0.18 0.44** 0.16 0.36* 0.18 0.45* 0.20 0.24† 0.15 0.17 0.17 Early Childhood Pretest 0.00 0.08 0.06 0.10 -0.03 0.11 0.05 0.10 0.02 0.11 -0.08 0.12 0.00 0.09 0.00 0.10 Early Childhood2 Condition -0.06 0.22 -0.19 0.28 -0.11 0.29 -0.24 0.26 -0.33 0.29 0.05 0.32 0.14 0.23 0.28 0.27
100 Condition x 0.03 0.25 0.05 0.32 0.13 0.34 0.07 0.29 0.00 0.33 -0.02 0.37 -0.16 0.27 0.14 0.31
Pretest Condition x 0.32* 0.16 0.35† 0.20 0.36† 0.21 0.47* 0.19 0.42* 0.21 0.31 0.24 0.15 0.17 0.20 0.20 Pretest2 R2 0.33 0.34 0.30 0.28 0.22 0.23 0.22 0.21
Note. Missing cases were imputed using predictive mean matching. Overall performance averages parents’ rating across the individual academic categories listed in the remaining columns in this table (Cronbach’s α = .92; † p < .1, * p < .05, ** p < .01, *** p < .001).
101
102
Discussion
The RTL-funded early childhood literacy computer game had long-lasting impacts on children’s literacy as assessed by standardized measures, as well as their performance in other school subjects as reported by parents. Even though the original intervention was fairly short
(i.e., 6-8 weeks) and primarily focused on literacy (Schmitt et al., 2017), this early learning experience seems to have been powerful enough to have enhanced some children’s learning trajectories in English Language Arts and in other subject areas throughout elementary school.
To an extent, these findings substantiate the main principle of the Early Learning Hypothesis
(i.e., that early educational media exposure can catalyze long-term scholastic success; Anderson et al., 2001) with strong, causal, longitudinal evidence, and indicate that the Early Learning
Hypothesis extends beyond television to early educational computer play. As intended (Singer &
Singer, 1998), the RTL media resulted in some children being more prepared to succeed in elementary school. However, effects were only detectable for children with below and above average literacy skills prior to exposure to the intervention in early childhood, with negligible effects for children with average pretest scores. Although there is some ambiguity as to the exact mechanisms and pathways that yielded these results due in large part to the large time gap between waves of data collection, the Traveling Lens (Huston & Wright, 1989; Rice et al., 1982) and Capacity Models (Fisch, 2000, 2004, 2016) can provide insights as to why this pattern of results have arisen.
It is possible that the learning content in the games may have initially aligned well with the skillsets of the weaker readers in the sample. Before the intervention, these children scored low on measures of very basic literacy such as alphabet letter knowledge. The RTL games, 103 which began by promoting this basic literacy content and leveled up to more advanced content
(see Grant et al., 2012), might have engaged these children and scaffolded their learning in a manner consistent with the Traveling Lens Model (Huston & Wright, 1989; Rice et al., 1982). In other words, the RTL literacy games may have begun at just the right degree of difficulty for the weakest readers in the sample and slowly became more challenging in a way that continued to align with their growing literacy (Huston & Wright, 1989; Rice et al., 1982). Furthermore, the strong learning and engagement the games fostered for the children most in need of literacy remediation might have provided them a toolkit of literacy strategies they could draw from to excel in first grade reading classes or in other classes requiring reading and helped to improve their attitudes towards reading and literacy (Fisch, 2004; Fisch et al., 2005). This in turn may have helped them place into more advanced groups for reading or for other school subjects requiring skill in reading (Anderson et al., 2001). A combination of these circumstances may have initiated a chain of continued scholastic success as per the Early Learning Hypothesis
(Anderson et al., 2001).
In contrast, children with above average literacy skills at early childhood pretest might have found the RTL gaming suite engaging and reinforcing despite less-than-perfect skill alignment at first. Initially, they may have been interested and motivated to play the beginning, easy games just because they enjoyed engaging with literacy-themed content, somewhat consistent with the Capacity Model (Fisch, 2004). That is, these children may have liked literacy and voluntarily allocated attentional resources to the literacy-themed games, even though they may have already known some of the lessons promoted in the early levels (Fisch, 2004; Hirsh-
Pasek et al., 2015). Doing so may have helped them rehearse and strengthen foundational skills 104
(Fisch, 2004; Rice et al., 1982). They later may have continued to approach the gaming experience with comparable focus (Hirsh-Pasek et al., 2015), eventually reaching more advanced levels that introduced newer content. They also may have uniquely benefited from certain games with text elements that might have been incomprehensible to other children or from games with more linguistically complex instructions administered orally, again consistent with Capacity
Model principles (Fisch, 2000; 2004). Altogether, the games may have strengthened, reinforced, and extended these children’s literacy skills and/or their motivation to engage or interest in reading at a crucial time before they began formal schooling, setting them on an even more advanced academic trajectory than they were on prior to the intervention (Anderson et al., 2001).
Why might the findings for Segmentation not mirror the findings for the other literacy measures? Children’s early childhood pretest scores significantly predicted their middle childhood Segmentation scores (see Table 8), and Segmentation scores were significantly correlated with all other middle childhood literacy measures except Rhyming and Deletion (.11 ≤ rs ≤ .28, ps ≥ .30). However, children may not have fully understood the Segmentation task instructions. The training items for this task asked children to practice segmenting words with two syllables or phonemes (sounds), and many children consistently segmented words into two parts, regardless of the correct number of syllables or phonemes in a given test item (i.e., even if test items contained three or more syllables or phonemes). To address these sorts of issues,
Pearson recently released a newer streamlined version of the KTEA (Scheller, 2014). The version of the Segmentation test used in the present study may not have provided as accurate a depiction of children’s literacy as the other literacy measures.
Limitations 105
The findings from this study should be interpreted in light of four limitations. First and foremost, plausible pathways through which the games may or may not have influenced children’s long-term outcomes were suggested. Although these proposed explanations align with the data collected and theory, additional data could have provided even stronger evidence for the arguments herein. For example, researchers did not directly measure engagement, motivation, or related constructs in a standardized, quantifiable fashion in the original study. Thus, statements about children’s potential level of focus during gameplay cannot be verified. Future scholars conducting similar early learning evaluations could measure motivation via direct verbal measures or by monitoring biomarkers such as heart rate (Aladé & Nathanson, 2016). Likewise, data on children’s performance in first grade or earlier in elementary school were not collected, and therefore one can only surmise that the pathways Anderson and colleagues (2001) proposed are correct. Future RTL evaluators or other scholars conducting similar research should consider prospectively planning follow-up studies at regular intervals to provide an even richer understanding of educational media’s long-term effects.
Second, families were recruited to the original evaluation by convenience sampling, which presents a threat to the external validity of the games’ impact in both the short- and long- term (Bornstein et al., 2013). The sample was diverse in terms of race and socio-economic status.
However, these families fairly universally reported highly valuing education. All participating families chose to enroll children in an optional literacy intervention in preschool or kindergarten, and many families in both conditions anecdotally reported continuing to enroll children in optional scholastic enrichment programs. As such, the pattern of results may have been different for families who place less emphasis on education, especially in light of other research showing 106 parent beliefs about education predict children’s educational attainment (e.g., Wu & Qi, 2006).
Third, the original intervention focused only on one suite of early learning computer games created with RTL funding and intended to promote literacy. Meta-analytic work suggests there is considerable variability in the effectiveness of RTL media properties in the short-term
(Article 1). Although these particular games featured multiple media properties (Between the
Lions, Martha Speaks, Sesame Street, Super WHY!, and WordWorld), it is possible different long-term results would have emerged had children played games focused on different media properties. Likewise, differing results may have arisen had the focal media content (a) initially presented more challenging content, (b) had the lessons been delivered via a television show or app or (c) drawn from other, non-RTL-funded educational media properties.
Finally, parent reports of children’s school performance were used to address RQ2a and
RQ2b. Measuring children’s school performance directly would have allowed a more definitive determination of the degree to which results transferred beyond literacy.
Conclusion
The results from this longitudinal evaluation provide causal evidence that even relatively brief exposure to educational computer games in early childhood can have long-lasting effects detectable six years after initial media exposure for some children both in the learning domain directly promoted (in this case literacy), with additional downstream effects on other domains
(e.g., science and social studies). These findings support expenditure on such educational tools through programs like the U.S. RTL initiative. Moreover, these findings, to a certain extent, substantiate and help add nuance to scholarly understanding of the Early Learning Hypothesis
(Anderson et al., 2001): Early educational computer game exposure can have long-term effects 107 on children’s school performance throughout elementary school. However, media may not be a silver bullet that equally affects all children (Piotrowski & Valkenburg, 2015; Valkenburg &
Peter, 2013). Aligned with the Traveling Lens Model, early media’s long-term effects may be particularly strong when mediated content closely reflects children current skillsets (Huston &
Wright, 1989; Rice et al., 1982), or, somewhat in accordance with the Capacity Model, when children themselves are able to bring stronger skills and interest to the mediated experience
(Fisch, 2000, 2004). To help ensure comparable long-term effects are realized by all children, it may be necessary to create games that can quickly adapt educational lessons to children’s skillsets to make the gaming experience more universally engaging (Guernsey, 2012; Roberts et al., 2016).
108
Conclusion
While RTL media are not a magic bullet that ameliorate all achievement gaps at school entry, overall this dissertation suggests the RTL initiative is effective. RTL media have modest but positive effects on children’s early literacy skills, and this early boost lasts for some children at least through elementary school and generalizes to other subject domains beyond literacy.
Below, broader implications for the RTL initiative and for children’s media theory are discussed, cutting across insights gleaned from all three articles.
Implications for the RTL Initiative
Even though school readiness, especially among low-income populations, is still a public policy concern in the U.S. (Bradbury, Corak, Waldfogel, & Washbrook, 2015), federal funding for public media and many educational programs/services like RTL is at risk as of the time of this writing. Yet, as stated above, this dissertation provides fairly positive accountability evidence in favor of the RTL initiative and its role in promoting early literacy and school readiness among both general and at-risk populations. RTL media yield effects on par with Head
Start (Kay & Pennucci, 2014). Only $25 million in national taxpayer dollars are allocated for
RTL each year (DoEd, 2015). While this may initially seem like a considerable sum, and indeed
I note that this expenditure is nontrivial in Article 1, it is miniscule compared to other federal budget line items. For instance, Congress allocated upwards of $9 billion for Head Start last year
(National Head Start Association, 2017). And other developed countries allocate considerably larger proportions of taxpayer dollars for public media (Lee, 2012). In this context, eliminating or significantly reducing such a small line item like RTL seems unfounded. Indeed, Kearney and 109
Levine (2015) previously have made similar arguments about the benefit of educational public media following an examination of the effects of Sesame Street prior to the launch of RTL.
Despite the overall positive findings of this dissertation, there is a need for continued monitoring of the RTL initiative (assuming it remains funded), especially as the program has continuously evolved with a sizable number of newer properties dedicated to promoting STEM learning and with increasing emphasis on various forms of new media (Pasnik, Llorente, Hupert,
& Moorthy, 2016). It is difficult to say whether the effects found in this dissertation would generalize to the current RTL context. That is, it is hard to predict whether a comparable set of studies focused on STEM-themed properties would have found similar results. Article 1 implies that future reviews may note larger effects as RTL producers continue to refine their approach to community outreach (Llorente et al., 2010) and improve technology designed to automatically level play experiences to align with children’s existing skillsets (Roberts et al., 2016).
Additionally, the new media platforms producers currently are favoring might particularly lend themselves to promoting certain types of inquiry-based STEM learning (Fisch, Damashek, &
Aladé, 2016). On the other hand, also in Article 1, RTL media’s effects on literacy became larger across grant cycles, perhaps as producers became more skilled at promoting learning in that domain. It is possible that RTL’s effects on STEM learning might similarly be small now but might grow over time. Relatedly, it may be harder for researchers to adequately measure STEM learning in preschool-age children due to a lack of relevant, validated measures (Pasnik,
Llorente, Hupert, Dominguez, & Silander, 2015), such that evaluators may be less well-equipped to assess learning in STEM than literacy at this point. 110
It likewise is impossible to surmise how the effective older RTL literacy-themed media might be if lifted out of their original contexts and re-evaluated today, given that the specific focus of each RTL grant cycle has been very much a reflection of broader and concurrent educational policy discussions. The 2005-2010 grant cycle focused on literacy shortly after the dissemination of the conclusions by the National Reading Panel (Langenberg et al., 2000) and prompted media adhering to those recommendations (Michael Cohen Group, 2012). The 2010-
2015 cycle focused on both literacy and mathematics aligned with the then new Common Core
State Standards, which also centered on those topic areas (PBS, 2011). The current cycle concerns science and literacy, consistent with the new and popular Next Generation Science
Standards (Educational Development Center & SRI International, 2016). Preschool and early elementary school teachers around the country likely have been drawing from curricula that emphasize and promote learning in a manner aligned with the same policy shifts. Thus, lessons children received in school may have complemented the media they consumed as part of RTL evaluations. Even though there was not support for the benefit of Joint Media Engagement as operationalized in Article 1 (i.e., parents/teachers doing any sort of extension activity), educators and parents still may be able to extend media’s lessons through in-person lessons that mirror media’s content but risk mitigating media’s effects by delivering contradictory lessons (Savage et al., 2013).
Accordingly, it is possible some of the literacy media products examined in the present dissertation might be less effective today or in the future as learning science and related policy continue to evolve, and as teachers adopt new instruction methods that may reflect the latest pedagogical knowledge but may become increasingly dissimilar to the way older RTL media 111 delivered lessons. Therefore, it may be wise for content producers to take care to remove products created with dated pedagogy from circulation. It seems as if PBS may already be doing this through its PBS LearningMedia hub (https://www.pbslearningmedia.org/), but older products often remain accessible elsewhere online (Webster, 2014) for potential use by parents, teachers, or researchers. This is not an issue for evaluators testing or teachers using recently produced media, just a potential concern when considering whether a media product produced in a different educational policy context and deemed effective at that time would have the same effects today.
In sum, it seems unjustified that RTL and programs like it are at risk for elimination based on the present positive findings. However, that is not to say that RTL and related programs should be immune from continued monitoring. Moreover, teachers and caregivers should critically consider whether a given RTL media product is appropriate for a child based on his/her broader educational experiences, keeping in mind that an older product once judged effective may cease to mirror the most recent learning science.
Implications for Children’s Media Theory
In addition to having practical policy ramifications, this dissertation also extends theoretical thinking concerning children’s learning from media in some areas, and raises new questions in others. As discussed at length in Article 3, these findings provide causal evidence substantiating the Early Learning Hypothesis and indicate that this hypothesis applies to new media. Moreover, these results also suggest that it is necessary to consider individual differences when assessing media’s long-term effects. Bridging both latter points, these findings further 112 signal that some children’s media theories originally intended to describe short-term individual difference phenomena may continue to bear out in the longer-term.
Future research beyond this dissertation is needed, however, to further integrate these patterns of individual differences into a more coherent theoretical framework and to definitively differentiate between the potential benefits and detriments of new media. Each of these topics – the value of new media, and theory regarding individual differences – is discussed in greater detail below.
A content-based approach to considering new media. In the larger children’s media research community, scholars recently and frequently have asked whether curriculum-based new media can promote learning at all (e.g., Blackwell, 2015) or whether it can do so as well as television (e.g., Aladé et al., 2016). In Article 1, the meta-analytic results suggest that on average, television and new media proved equally effective in promoting children’s early literacy, but I speculate that new media may have not fully realized its potential as an educational tool and may become increasingly effective over time (Hirsh-Pasek et al., 2015).
It could be more sophisticated, however, to ask under what circumstances new media has the potential to be effective at all, or to be more effective than television. Just as children’s media scholars are increasingly considering the content, context, and individual child when discussing television’s potential effects on children (Guernsey, 2012), so too might it be necessary to consider such factors when determining whether an educational lesson would be delivered more effectively via television or new media. The overall null difference between television and new media observed in Article 1 may be attributable to content producers or evaluators either instinctively choosing to test properties via the platform that is best suited to a given content, 113 context, child combination, or to mistakes in choice of platform cancelling each other out (e.g., an evaluation of a highly effective app averaged with an evaluation of an ineffective app). In that vein, additional analyses of the Article 1 data not reported above reveal that researchers tended to use new media more often in home-based studies and television more often in school-based studies (χ2(1, k = 783) = 41.65, p =.01). Evaluators or producers may have intuited that new media could work well in the home context, where children could easily play independently or with the support of other family members (Plowman, McPake, & Stephen, 2008), whereas television viewing might lend itself better to a whole class activity.
In additional exploratory analyses also not included in Article 1, results indicated that new media was more effective than television at promoting alphabet knowledge (z = 2.10, p =
.04), although there were no other significant differences between television and new media for any other literacy outcomes. Alphabet knowledge is a more basic early literacy skill particularly germane to children on the younger end of RTL’s target age range (Grant et al., 2012). Younger preschoolers tend to consume television in a less attentive fashion than their slightly older peers
(Anderson & Lorch, 1983), but perhaps would be more focused playing a simple interactive game (Michael Cohen Group, 2011). In other words, child-level developmental state and the content of the media could explain this finding. Alphabet knowledge also hypothetically may receive less of a boost from the stronger narratives typically provided in television (unlike vocabulary, for example, where contextual information aids learning; Gola et al., 2012).
Although others similarly argued for the importance of considering a given platform’s fit to an educational context (e.g., Fisch et al., 2016), this line of thinking has not permeated the way scholars currently are designing new media research or conceptualizing related theoretical 114 frameworks. When researchers find mixed results as to new media’s effectiveness at all or effectiveness relative to television, frequently, they provide posthoc explanations for such results. The findings of this dissertation suggest a need for more theoretical work to help prospectively guide decisions to use or test new media or television, moving beyond whether a given platform is appropriate to when it might be.
Alternatively, the larger children’s media research community may be overly fixated on the differences in affordances between television and new media (Wartella, 2015). Television and new media products may truly lead to roughly equal potential effects. Perhaps the potential boost well-designed interactivity could provide (Hirsh-Pasek et al., 2015) is equivalent to the benefit of a stronger narrative that television is more likely to offer (Lu et al., 2012).
Interchangeably referring to both computer games and mobile apps as new media was necessary for the present line of inquiry (there were not enough evaluations to examine computer games and apps separately in Article 1), but would be problematic going forward, especially in light of RTL’s very young target audience. Children on the younger end of RTL’s target age range have more difficulty using a mouse than children in early elementary school (Crook,
1992), and consequential navigational difficulties could limit children’s learning from computers
(Guernsey, 2012). Although similar age-related patterns are noted in studies of children’s tablet engagement (Michael Cohen Group, 2011), difficulties may be less pronounced. Thus, younger preschool children might learn more from apps than similar computer games. In Article 3, the learning content of the 2010 RTL computer game began with a focus on relatively basic literacy skills that children typically begin learning around age 3 (Grant et al., 2012), even though children in the original sample ranged in age from 4-6 years. In app form, the game could have 115 been “aged down” to be developmentally appropriate for 3-year-old children, provided they had background familiarity with at least some letters of the alphabet to allow them to assimilate the in-game lessons into their existing schemas about literacy (Schmitt & Linebarger, 2009). This design choice could have led to stronger effects in both the short- and long-term.
Growing the garden. Elsewhere in Article 3, results provide evidence substantiating both the Traveling Lens Model’s notion of media helping children whose skillsets align with its content (Huston & Wright, 1989; Rice et al., 1982) and the Capacity Model’s notion of media benefiting the children who have the strongest verbal skills (Fisch, 2004). Reconsidering and reconceptualizing the Differential Susceptibility to Media Effects Model and its notions of orchid children who are susceptible to media vs. dandelion children who are impervious (Piotrowski &
Valkenburg, 2015) in light of these findings may help to reconcile this apparent discrepancy.
Initially considering these results, one might conclude that the low-income children in
Article 1 (aligned with Linebarger & Barr, 2017) and the below average early readers in Article
3 are orchid children, while children with average early literacy skills in Article 3 are dandelions.
After all, the former groups of children particularly benefited from early RTL media exposure, while the latter group seemed more impervious. However, as argued above, I am not convinced that the children with average literacy skills were hopeless cases. According to the Traveling
Lens Model (Huston & Wright, 1989; Rice et al., 1982), these children could have exhibited comparable effects had they played a game better aligned with their current abilities. That is, with a game that began with more challenging content or that automatically re-calibrated to better meet their needs, these children too may have grown like orchids. The right temperature, level of humidity, etc. may have been in place in 2010 for the below average children who 116 participated in the original evaluation, but the children with average preliteracy skills may have flourished in slightly warmer conditions. Future empirical work is needed with harder games or with games using item response theory to automatically level content to more definitively determine if there really are dandelion children, at least under circumstances mirroring the present study (i.e., when families opt-in to engage in an educational media intervention).
One could also make a case that the children with above average early literacy skills in
Article 3 were more orchids. These children were already strong readers for their age prior to media exposure (a beautiful characteristic, reminiscent of a beautiful flower) and seemed to benefit from the intervention. However, the data does not support the notion that these children might have withered without the early boost provided by the game. Children in the control condition with comparable above average pretest scores were still scoring above average in middle childhood; their scores were just slightly lower than their RTL peers. It may be more apt to compare these gifted children to tulips. They would grow well in many conditions with minimal interference, but blossom even more with the right intervention.
Altogether, I am arguing that this intervention grew a garden with multiple genera of orchids (children who are not precocious but who can benefit from developmentally appropriate media, as per the Traveling Lens Model; Huston & Wright, 1989; Rice et al., 1992) and tulips
(above average children who would still be gifted even without intervention but who benefit from extra exposure to engaging educational content, as per the Capacity Model; Fisch, 2004). I do not see any dandelions in this garden. But perhaps because the original intervention in Article
3 used convenience sampling, with families electing to engage in the optional intervention, only flowers with the potential to grow were planted (i.e., participated). 117
As a caveat, Piotrowski and Valkenburg (2015) specifically call for identifying child- level factors that simultaneously make children more susceptible to both positive and negative media effects – that these factors should help identify orchid children. Because the RTL evaluations explored herein did not expose children to any negative forms of media, I can only partially expand up their model and am unsure the extent to which these ideas would translate to violent media effects.
Closing Thoughts
Overall, these findings provide fairly supportive accountability evidence in favor of
RTL’s power to positively influence children’s school readiness and school performance, and point to a need for more nuanced examination of undergirding causal mechanisms. RTL media have positive short-term effects on children’s early literacy and lasting impacts on some children’s performance in literacy and other school subjects throughout elementary school. The findings of this dissertation therefore support the continued funding of this initiative.
Nonetheless, it is important not to overgeneralize these results. More research is needed to verify that the initiative continues to induce comparable change under evolving conditions and to help scholars better understand when television or new media are more appropriate tools to induce learning and when individual children may benefit the most from media exposure. 118
References
Abu-Lughod, J., & Foley, M. M. (1960). Consumer strategies. In N. N. Foote, J. Abu-Lughod,
M. M. Foley & L. Winnick (Eds.), Housing choices and housing constraints (pp. 71-
271). New York: McGraw Hill.
Adlof, S. M., Catts, H. W., & Lee, J. (2010). Kindergarten predictors of second versus eighth
grade reading comprehension impairments. Journal of Learning Disabilities, 43, 332-
345. doi:10.1177/0022219410369067
Agrawal, K. C., Kellam, S. G., Klein, Z. E., & Turner, J. (1978). The Woodlawn mental health
studies: Tracking children and families for long-term follow-up. American Journal of
Public Health, 68, 139-142. doi:10.2105/AJPH.68.2.139
Ahlers-Schmidt, C. R., Chesser, A. K., Nguyen, T., Brannon, J., Hart, T. A., Williams, K. S., &
Wittler, R. R. (2012). Feasibility of a randomized controlled trial to evaluate Text
Reminders for Immunization Compliance in Kids (TRICKs). Vaccine, 30, 5305-5309.
doi:10.1016/j.vaccine.2012.06.058
Aladé, F. (2013). What preschoolers bring to the show: The effects of cognitive abilities and
viewer characteristics on children's learning from educational television (Master's
thesis). Retrieved from http://rave.ohiolink.edu/etdc
Aladé, F., Lauricella, A. R., Beaudoin-Ryan, L., & Wartella, E. (2016). Measuring with Murray:
Touchscreen technology and preschoolers' STEM learning. Computers in Human
Behavior, 62, 433-441. doi:10.1016/j.chb.2016.03.080 119
Aladé, F., & Nathanson, A. I. (2016). What preschoolers bring to the show: The relation between
vewer characteristics and children’s learning from educational television. Media
Psychology, 19, 406-430. doi:10.1080/15213269.2015.1054945
Allison, P. (2015). Imputation by predictive mean matching: Promise and peril. Retrieved from
http://statisticalhorizons.com/predictive-mean-matching
American Chamber of Commerce Research. (2007). Cost of living index for selected U.S. cities,
2005. Retrieved from http://www.infoplease.com/ipa/a0883960.html
American Psychological Association. (2010). Publication manual of the American Psychological
Association (6th ed.). Washington, DC: American Psychological Association.
Anderson, D. R., & Collins, P. A. (1988). The impact on children's education: Television's
influence on cognitive development (Working paper no. 2). Washington, D.C.: Office of
Education Research and Improvement.
Anderson, D. R., Huston, A. C., Schmitt, K. L., Linebarger, D. L., & Wright, J. C. (2001). Early
childhood television viewing and adolescent behavior: The recontact study. Monographs
of the Society for Research in Child Development, 66, vii-147. doi:10.1111/1540-
5834.00120
Anderson, D. R., & Lorch, E. P. (1983). Looking at television: Action or reaction? In J. Bryant &
D. R. Anderson (Eds.), Children's understanding of television (pp. 1-33). New York, NY:
Academic Press.
Ayduk, O., Mendoza-Denton, R., Mischel, W., Downey, G., Peake, P. K., & Rodriguez, M.
(2000). Regulating the interpersonal self: Strategic self-regulation for coping with 120
rejection sensitivity. Journal of Personality and Social Psychology, 79, 776-792.
doi:10.1037/0022-3514.79.5.776
Barakat-Haddad, C. P., Elliott, S. J., Eyles, J., & Pengelly, D. (2009). Predictors of locating
children participants in epidemiological studies 20 years after last contact: Internet
resources and longitudinal research. European Journal of Epidemiology, 24, 397-405.
doi:10.1007/s10654-009-9364-5
Barnett, W. S. (2013). Getting the facts right on pre-k and the President's pre-k proposal. New
Brunswick, NJ: National Institute for Early Education Research.
Blackwell, C. K. (2015). Technology use in early childhood education: Investigating teacher
access and attitudes toward technology and the effect of iPads on student achievement
(Doctoral dissertation). Retrieved from ProQuest Dissertations & Theses Global
database. (3705215)
Bonner, M., & Carpenter, D. (2005). Kaufman Test of Educational Achievement, Second
edition, Comprehensive form. In R. A. Spies & B. S. Plake (Eds.), The sixteenth mental
measurements yearbook: University of Nebraska Press & the Buros Center for Testing.
Retrieved from http://web.a.ebscohost.com.turing.library.northwestern.edu/.
Borenstein, M., Hedges, L. V., Higgins, J. P. T., & Rothstein, H. R. (2009). Introduction to meta-
analysis. Hoboken, NJ: Wiley.
Bornstein, M. H., Jager, J., & Putnick, D. L. (2013). Sampling in developmental science:
Situations, shortcomings, solutions, and standards. Developmental Review, 33, 357-370.
doi:10.1016/j.dr.2013.08.003 121
Boswell, W. (2007, July 23). Where to find public records online, Lifehacker. Retrieved from
http://lifehacker.com/
Bradbury, B., Corak, M., Waldfogel, J., & Washbrook, E. (2015). Too many children left behind:
The U.S. achievement gap in comparative perspective. New York, NY: Russell Sage
Foundation.
Bryant, J. A., Bryant, J., Mullikin, L., McCollum, J. F., & Love, C. C. (2001). Curriculum-based
preschool television programming and the American family: Historical development,
impact of public policy, and social and educational effects. In J. Bryant & A. Bryant, J.
(Eds.), Television and the American family (2nd ed., pp. 415-434). Mahwah, NJ:
Lawrence Erlbaum Associates.
Bus, A. G., & van Ijzendoorn, M. H. (1999). Phonological awareness and early reading: A meta-
analysis of experimental training studies. Journal of Educational Psychology, 91, 403-
414. doi:10.1037/0022-0663.91.3.403
Calfee, R. C., Lindamood, P., & Lindamood, C. (1973). Acoustic-phonetic skills and reading:
Kindergarten through twelfth grade. Journal of Educational Psychology, 64, 293-298.
doi:10.1037/h0034586
Campbell, F. A., Pungello, E. P., Burchinal, M., Kainz, K., Pan, Y., Wasik, B. H., . . . Ramey, C.
T. (2012). Adult outcomes as a function of an early childhood educational program: An
Abecedarian Project follow-up. Developmental Psychology, 48, 1033-1043.
doi:10.1037/a0026644
Can, D. D., Ginsburg-Block, M., Golinkoff, R. M., & Hirsh-Pasek, K. (2013). A long-term
predictive validity study: Can the CDI Short Form be used to predict language and early 122
literacy skills four years later? Journal of Child Language, 40, 821-835.
doi:10.1017/S030500091200030X
Cohen, M., Hadley, M., & Marcial, C. (2016). The role of research and evaluation in Ready to
Learn transmedia development. Journal of Children and Media, 10, 248-256.
doi:10.1080/17482798.2016.1140488
Cook, T. D., Appleton, H., Conner, R. F., Shaffer, A., Tamkin, G., & Weber, S. J. (1975).
Sesame Street revisited. New York, NY: Russell Sage Foundation.
Cooper, H. M. (1981). On the significance of effects and the effects of significance. Journal of
Personality and Social Psychology, 41, 1013-1018. doi:10.1037/0022-3514.41.5.1013
Cotter, R. B., Burke, J. D., Loeber, R., & Navratil, J. L. (2002). Innovative retention methods in
longitudinal research: A case study of the developmental trends study. Journal of Child
and Family Studies, 11, 485-498. doi:10.1023/a:1020939626243
Cotter, R. B., Burke, J. D., Stouthamer-Loeber, M., & Loeber, R. (2005). Contacting participants
for follow-up: How much effort is required to retain participants in longitudinal studies?
Evaluation and Program Planning, 28, 15-21. doi:10.1016/j.evalprogplan.2004.10.002
Council for Community and Economic Research. (2010). Cost of living index for selected U.S.
cities. Retrieved from http://www.infoplease.com/business/economy/cost-living-index-
us-cities.html
CPB, & PBS. (2011). Findings from Ready To Learn 2005-2010. Arlington, VA: Corporation for
Public Broadcasting and Public Broadcasting Service. 123
Creager, K. (2011, July 5). Is your email address preventing you from getting the job you want?
[Blog post], Brazen Blog. Retrieved from http://www.brazen.com/blog/archive/job-
search/is-your-email-address-preventing-you-from-getting-the-job-you-want/
Crook, C. (1992). Young children's skill in using a mouse to control a graphical computer
interface. Computers & Education, 19, 199-207. doi:10.1016/0360-1315(92)90113-J
Cunningham, A. E., & Stanovich, K. E. (1997). Early reading acquisition and its relation to
reading experience and ability 10 years later. Developmental Psychology, 33, 934-945.
doi:10.1037/0012-1649.33.6.934
DoEd. (2015). Funding status. Retrieved from http://www2.ed.gov/programs/rtltv/funding.html
Dost, M., & McGeeney, K. (2016). Moving without changing your cellphone number: A
predicament for pollsters. Washington, D.C.: Pew Research Center's Internet &
American Life Project.
Duch, H. (2005). Redefining parent involvement in Head Start: A two‐generation approach.
Early Child Development and Care, 175, 23-35. doi:10.1080/0300443042000206237
Duggan, M. (2013). Cell phone activities 2013: 50% of cell owners download apps to their
phones; 48% listen to music services; video calling has tripled since 2011; texting
remains a popular activity. Washington, D.C.: Pew Research Center's Internet &
American Life Project.
Duggan, M., Ellison, N. B., Lampe, C., Lenhart, A., & Madden, M. (2015). Socail media update
2014: While Facebook remains the most popular site, other platforms see higher rates of
growth. Washington, D.C.: Pew Research Center's Internet & American Life Project. 124
Dunn, L. M., & Dunn, L. M. (1997). Peabody Picture Vocabulary Test-Third Edition (PPVT-
III). Circle Pines, MN: American Guidance Service.
Dupre, E. (2014, April 1). Growing up Gmail: 10 years of email marketing lessons, DMN.
Retrieved from http://dmnews.com
Educational Development Center, & SRI International. (2012). PBS KIDS transmedia suites
gaming study. Education Development Center, Inc. & SRI International. New York and
Menlo Park, CA.
Educational Development Center, & SRI International. (2016). Science + literacy: A formative
research tool for CPB-PBS Ready To Learn producers. Education Development Center,
Inc. & SRI International. New York and Menlo Park, CA.
Egger, M., Smith, G. D., Schneider, M., & Minder, C. (1997). Bias in meta-analysis detected by
a simple, graphical test. British Medical Journal, 315, 629-634.
doi:10.1136/bmj.315.7109.629
Ehri, L. C., Nunes, S. R., Willows, D. M., Schuster, B. V., Yaghoub-Zadeh, Z., & Shanahan, T.
(2001). Phonemic awareness instruction helps children learn to read: Evidence from the
National Reading Panel's meta-analysis. Reading Research Quarterly, 36, 250-287.
doi:10.1598/RRQ.36.3.2
Fein, S. B., Li, R., Chen, J., Scanlon, K. S., & Grummer-Strawn, L. M. (2014). Methods for the
year 6 follow-up study of children in the Infant Feeding Practices Study II. Pediatrics,
134, S4.
Fisch, S. M. (2000). A Capacity Model of children's comprehension of educational content on
television. Media Psychology, 2, 63-91. doi:10.1207/S1532785XMEP0201_4 125
Fisch, S. M. (2004). Children's learning from educational television: Sesame Street and beyond.
Mahwah, NJ: Lawrence Erlbaum.
Fisch, S. M. (2016, October). The Capacity Model, 2.0: Cognitive processing in children's
comprehension of educational games. In J. T. Piotrowski (Chair), Digital media and early
literacy. Symposium conducted at the Society for Research on Child Development
Special Topics Meeting: Technology and Media in Children’s Development, Irvine, CA.
Fisch, S. M., Damashek, S., & Aladé, F. (2016). Designing media for cross-platform learning:
Developing models for production and instructional design. Journal of Children and
Media, 10, 238-247. doi:10.1080/17482798.2016.1140485
Fisch, S. M., Kirkorian, H. L., & Anderson, D. R. (2005). Transfer of learning in informal
education. In J. P. Mestre (Ed.), Transfer of learning from a modern multidisciplinary
perspective (pp. 371-393). Greenwich, CT: Information Age Publishing.
Fisch, S. M., & Truglio, R. (Eds.). (2000). G is for "growing": Thirty years of research on
children and Sesame Street. Mahwah, NJ, USA: Lawrence Erlbaum
Gola, A. A. H., Mayeux, L., & Naigles, L. R. (2012). Electronic media as incidental language
teachers. In D. G. Singer & J. L. Singer (Eds.), Handbook of children and the media (2nd
ed., pp. 139-156). Thousand Oaks, CA: SAGE.
Golinkoff, R. M., Jacquet, R. C., Hirsh-Pasek, K., & Nandakumar, R. (1996). Lexical principles
may underlie the learning of verbs. Child Development, 67, 3101-3119.
doi:10.1111/j.1467-8624.1996.tb01905.x
Grant, A., Wood, E., Gottardo, A., Evans, M. A., Phillips, L., & Savage, R. (2012). Assessing
the content and quality of commercially available reading software programs: Do they 126
have the fundamental structures to promote the development of early reading skills in
children? NHSA Dialog, 15, 319-342. doi:10.1080/15240754.2012.725487
Griffin, T. M., Hemphill, L., Camp, L., & Wolf, D. P. (2004). Oral discourse in the preschool
years and later literacy skills. First Language, 24, 123-147.
doi:10.1177/0142723704042369
Guernsey, L. (2012). Screen time: How electronic media - from baby videos to educational
software - affects your young child. Philadelphia, PA: Basic Books.
Guernsey, L., & Levine, M. H. (2015). Tap, click, read: Growing readers in a world of screens.
San Francisco, CA: Jossey-Bass.
Haggerty, K. P., Fleming, C. B., Catalano, R. F., Petrie, R. S., Rubin, R. J., & Grassley, M. H.
(2008). Ten years later: Locating and interviewing children of drug abusers. Evaluation
and Program Planning, 31, 1-9. doi:10.1016/j.evalprogplan.2007.10.003
Hanson, K. G. (2017). The influence of early media exposure on children's development and
learning (Unpublished doctoral dissertation). University of Massachusetts Amherst,
Amherst, MA. Retrieved from
Harris, M. A., Brett, C. E., Johnson, W., & Deary, I. J. (2016). Personality stability from age 14
to age 77 years. Psychology and Aging, 31, 862-874. doi:10.1037/pag0000133
Hartmann, D. P. (1992). Design, measurement, and analysis: Technical issues in developmental
research. In M. H. Bornstein & M. E. Lamb (Eds.), Developmental psychology: An
advanced textbook (3rd ed., pp. 132-154). Hillsdale, NJ: Lawrence Erlbaum.
Heatly, M. C., Bachman, H. J., & Votruba-Drzal, E. (2015). Developmental patterns in the
associations between instructional practices and children's math trajectories in elementary 127
school. Journal of Applied Developmental Psychology, 41, 46-59.
doi:10.1016/j.appdev.2015.06.002
Hedges, L. V. (2007). Effect sizes in cluster-randomized designs. Journal of Educational and
Behavioral Statistics, 32, 341-370. doi:10.3102/1076998606298043
Hedges, L. V., & Hedberg, E. C. (2007). Intraclass correlation values for planning group-
randomized trials in education. Educational Evaluation and Policy Analysis, 29, 60-87.
doi:10.3102/0162373707299706
Hill, C. J., Bloom, H. S., Black, A. R., & Lipsey, M. W. (2008). Empirical benchmarks for
interpreting effect sizes in research. Child Development Perspectives, 2, 172-177.
doi:10.1111/j.1750-8606.2008.00061.x
Hirsh-Pasek, K., Zosh, J. M., Golinkoff, R. M., Gray, J. H., Robb, M. B., & Kaufman, J. (2015).
Putting education in “educational” apps: Lessons from the science of learning.
Psychological Science in the Public Interest, 16, 3-34. doi:10.1177/1529100615569721
Horowitz, J. E., Juffer, K., Davis Sosenko, L., Stout Hoffman, J. L., Bojorquez, J. C., Dailey, K.,
& Wingren, G. (2005). Data collection of the federal performance indicators for PBS
Ready To Learn: Year five summary. Los Alamitos, CA: WestEd.
Huff, J. O., & Clark, W. A. V. (1978). Cumulative stress and cumulative inertia: A behavioral
model of the decision to move. Environment and Planning A, 10, 1101-1119.
doi:10.1068/a101101
Hurtado, M., Galdo, J., Agin, L., & Heil, S. (2010). A social marketing approach to promoting
literacy-related behaviors among low-income families. Paper presented at the 2010
Annual Meeting of the American Educational Research Association, Denver, Colorado. 128
Huston, A. C., Duncan, G. J., McLoyd, V. C., Crosby, D. A., Ripke, M. N., Weisner, T. S., &
Eldred, C. A. (2005). Impacts on children of a policy to promote employment and reduce
poverty for low-income parents: New Hope after 5 years. Developmental Psychology, 41,
902-918. doi:10.1037/0012-1649.41.6.902
Huston, A. C., & Wright, J. C. (1989). Forms of television and the child viewer. In G. A.
Comstock (Ed.), Public communication and behavior (Vol. 2, pp. 103-159). Orlando,
FL.: Academic Press.
Institute of Research on Poverty. (2010). What are the poverty thresholds and poverty
guidelines? Retrieved from http://www.irp.wisc.edu/faqs/faq1.htm#year2000
Invernizzi, M., Sullivan, A., Meier, J., & Swank, L. (2004). PreK teacher manual: PALS
Phonological Awareness Literacy Screening. Richmond, VA: University of Virginia.
Jacobsson Purewal, S. (2015, January 22). 5 tips for finding anything, about anyone, online,
CNET. Retrieved from http://dmnews.com
Jennings, N. (2013). Super readers at CET/ThinkTV: An evaluation of Super WHY! reading
camps: Final report. Children’s Education and Entertainment Research (CHEER) Lab,
University of Cincinnati. Cincinnati, OH.
Juel, C. (1988). Learning to read and write: A longitudinal study of 54 children from first
through fourth grades. Journal of Educational Psychology, 80, 437-447.
doi:10.1037/0022-0663.80.4.437
Kalkhoff, W., Youngreen, R., Nath, L., & Lovaglia, M. J. (2014). Human participants in
laboratory experiments in the social sciences. In M. J. Webster & J. Sell (Eds.), 129
Laboratory experiments in the social sciences (2nd ed., pp. 103-126). Waltham, MA:
Elsevier.
Karmiloff-Smith, A. (1998). Development itself is the key to understanding developmental
disorders. Trends in Cognitive Sciences, 2, 389-398. doi:10.1016/S1364-6613(98)01230-
3
Kaufman, A. S., & Kaufman, N. L. (2004a). Kaufman test of educational achievement (2nd ed.).
San Antonio, TX: Pearson.
Kaufman, A. S., & Kaufman, N. L. (2004b). Kaufman test of educational achievement:
Comprehensive form norm book (2nd ed.). San Antonio, TX: Pearson.
Kay, N., & Pennucci, A. (2014). Early childhood education for low-income students: A review of
the evidence and benefit-cost analysis (Doc. No. 14-01-2201). Olympia, WA:
Washington State Institute for Public Policy.
Kearney, M. S., & Levine, P. B. (2015). Early childhood education by MOOC: Lessons from
Sesame Street. (Working Paper 21229), Retrieved from the National Bureau of Economic
Research website: http://www.nber.org/papers/w21229
Khan, O. (2015, July 15). Major email provider trends in 2015: Gmail's lead increases [Blog
post], MailChimp. Retrieved from http://blog.mailchimp.com/major-email-provider-
trends-in-2015-gmail-takes-a-really-big-lead/
Kim, R.-S., & Becker, B. J. (2010). The degree of dependence between multiple-treatment effect
sizes. Multivariate Behavioral Research, 45, 213-238. doi:10.1080/00273171003680104
Langenberg, D. N., Correro, G., Ehri, L., Ferguson, G., Garza, N., Kamil, M. L., . . . Yatvin, J.
(2000). Report of the National Reading Panel: Teaching children to read: An evidence- 130
based assessment of the scientific research literature on reading and its implications for
reading instruction: Reports of the subgroups. Bethesda, MD: National Institute of Child
Health and Human Development, National Institutes of Health.
Lauer, J. E., & Lourenco, S. F. (2016). Spatial processing in infancy predicts both spatial and
mathematical aptitude in childhood. Psychological Science, 27, 1291-1298.
doi:10.1177/0956797616655977
Lazorick, S., Crawford, Y., Gilbird, A., Fang, X., Burr, V., Moore, V., & Hardison Jr, G. T.
(2014). Long-term obesity prevention and the Motivating Adolescents with Technology
to CHOOSE Health™ program. Childhood Obesity, 10, 25-33.
doi:10.1089/chi.2013.0049
Lee, S. (2012, October 11). Big Bird debate: How much does federal funding matter in public
broadcasting?, ProPublica. Retrieved from http://www.propublica.org
Light, R. J., & Pillemer, D. B. (1984). Summing up: The science of reviewing research.
Cambridge, MA: Harvard University Press.
Linebarger, D. L. (2006). The Between the Lions American Indian literacy iniative research
component: Report prepared for the United States Department of Education.
Philadelphia, PA: Annenberg School for Communication, University of Pennsylvania.
Linebarger, D. L. (2010). Television's impact on children's reading skills: A longitudinal study.
Philadelphia, PA: Annenberg School for Communication, University of Pennsylvania.
Linebarger, D. L., & Barr, R. (2017, April). Differntial susceptibility to media effects: Who is
affected by what media content under which circumstances. Poster presented at the
Biennial Meeting of the Society for Research on Child Development, Austin, TX. 131
Linebarger, D. L., Kosanic, A. Z., Greenwood, C. R., & Doku, N. S. (2004). Effects of viewing
the television program Between the Lions on the emergent literacy skills of young
children. Journal of Educational Psychology, 96, 297-308. doi:10.1037/0022-
0663.96.2.297
Linebarger, D. L., & Piotrowski, J. T. (2009). TV as storyteller: How exposure to television
narratives impacts at-risk preschoolers' story knowledge and narrative skills. British
Journal of Developmental Psychology, 27, 47-69. doi:10.1348/026151008X400445
Llorente, C., Pasnik, S., Penuel, W. R., & Martin, W. (2010). Organizing for literacy: How
public media stations are raising readers in their communities: A case study report to the
Ready to Learn Initiative. Education Development Center, Inc. & SRI International. New
York and Menlo Park, CA.
Lu, A. S., Baranowski, T., Thompson, D., & Buday, R. (2012). Story immersion of videogames
for youth health promotion: A review of literature. Games for Health Journal, 1, 199-
204. doi:10.1089/g4h.2011.0012
Mares, M.-L., & Pan, Z. (2013). Effects of Sesame Street: A meta-analysis of children's learning
in 15 countries. Journal of Applied Developmental Psychology, 34, 140-151.
doi:10.1016/j.appdev.2013.01.001
Masson, H., Balfe, M., Hackett, S., & Phillips, J. (2013). Lost without a trace? Social networking
and social research with a hard-to-reach population. British Journal of Social Work, 43,
24-40. doi:10.1093/bjsw/bcr168 132
McCarthy, B., Atienza, S., Kao, Y., Tiu, M., Banes, L., & Hubbard, A. (2012). Alpha testing of
PBS KIDS transmedia suites: A report to the CPB-PBS Ready to Learn initiative. Los
Alamitos, CA: WestEd.
McManis, L. D., & Gunnewig, S. B. (2012). Finding the education in educational technology
with early learners. YC Young Children, 67, 14-24.
Michael Cohen Group. (2009). The effects of WordWorld viewing on preschool children's
acquisition of pre-literacy skills and emergent literacy: A cluster-randomized controlled
trial. New York, NY: Michael Cohen Group.
Michael Cohen Group. (2011). Your children, apps, & iPad. New York, NY: Michael Cohen
Group, LLC.
Michael Cohen Group. (2012). Learning from Ready To Learn. New York, NY: Michael Cohen
Group, LLC.
Michael Cohen Group. (2013). Effects of interaction with Pocoyo playsets on preschool
(Spanish) ELL children's English Language Learning: A randomized controlled trial.
New York, NY: Michael Cohen Group, LLC.
Mol, S. E., Bus, A. G., de Jong, M. T., & Smeets, D. J. H. (2008). Added value of dialogic
parent–child book readings: A meta-analysis. Early Education and Development, 19, 7-
26. doi:10.1080/10409280701838603
Morris, S. B. (2007). Estimating effect sizes from the pretest-posttest-control group designs.
Organizational Research Methods. doi:10.1177/1094428106291059 133
Morris, S. B., & DeShon, R. P. (2002). Combining effect size estimates in meta-analysis with
repeated measures and independent-groups designs. Psychological Methods, 7, 105-125.
doi:10.1037/1082-989X.7.1.105
Moses, A. M., Linebarger, D. L., Wainwright, D. K., & Brod, R. (2010). Television’s impact on
children’s vocabulary knowledge: A meta-analysis. Philadelphia, PA: Annenberg School
for Communication, University of Pennsylvania.
Nathanson, A. I., & Rasmussen, E. E. (2011). TV Viewing compared to book reading and toy
playing reduces responsive maternal communication with toddlers and preschoolers.
Human Communication Research, 37, 465-487. doi:10.1111/j.1468-2958.2011.01413.x
National Conference of State Legislatures. (2016). Literacy & No Child Left Behind (NCLB).
Retrieved from http://www.ncsl.org/research/education/literacy-no-child-left-behind.aspx
National Head Start Association. (2017). Top issues. Retrieved from
http://www.nhsa.org/knowledge-center/center-advocacy/top-issues
Neuman, S. B., Newman, E. H., & Dwyer, J. (2011). Educational effects of a vocabulary
intervention on preschoolers' word knowledge and conceptual development: A cluster-
randomized trial. Reading Research Quarterly, 46, 249-272. doi:10.1598/RRQ.46.3.3
Nicholson, J., Sanson, A., Rempel, L., Smart, D., & Patton, G. (2002). Longitudinal studies of
children and youth: Implications for future studies. In A. Sanson (Ed.), Children's health
and development: New research directions for Australia (pp. 38-62). Melbourne,
Australia: Australian Institute of Family Studies.
Pasnik, S., Llorente, C., Hupert, N., Dominguez, X., & Silander, M. (2015). Supporting parent-
child experiences with Peg+Cat early math concepts: Report to the CPB-PBS Ready To 134
Learn initiative. Education Development Center, Inc. & SRI International. New York and
Menlo Park, CA.
Pasnik, S., Llorente, C., Hupert, N., & Moorthy, S. (2016). Reflections on the Ready To Learn
Initiative, 2010 to 2015: How a federal program in parternship with public media
supported young children's equitable learning during a time of great change. Education
Development Center, Inc. & SRI International. New York and Menlo Park, CA.
Pasnik, S., Penuel, W. R., Llorente, C., Strother, S., & Schindel, J. (2007). Review of research on
media and young children’s literacy: Report to the Ready to Learn Initiative. Education
Development Center, Inc. & SRI International. New York and Menlo Park, CA.
Passetti, L. L., Godley, S. H., Scott, C. K., & Siekmann, M. (2000). A low-cost follow-up
resource: Using the World Wide Web to maximize client location efforts. American
Journal of Evaluation, 21, 195-203. doi:10.1177/109821400002100206
PBS. (2011). Ready To Learn. Retrieved from http://pbskids.org/readytolearn/
Penuel, W. R., Bates, L., Gallagher, L. P., Pasnik, S., Llorente, C., Townsend, E., . . .
VanderBorght, M. (2012). Supplementing literacy instruction with a media-rich
intervention: Results of a randomized controlled trial. Early Childhood Research
Quarterly, 27, 115-127. doi:10.1016/j.ecresq.2011.07.002
Piotrowski, J. T., Jennings, N. A., & Linebarger, D. L. (2012). Extending the lessons of
educational television with young American children. Journal of Children and Media, 7,
216-234. doi:10.1080/17482798.2012.693053 135
Piotrowski, J. T., & Valkenburg, P. M. (2015). Finding orchids in a field of dandelions:
Understanding children’s differential susceptibility to media effects. American
Behavioral Scientist, 59, 1776-1789. doi:10.1177/0002764215596552
Plowman, L., McPake, J., & Stephen, C. (2008). Just picking it up? Young children learning with
technology at home. Cambridge Journal of Education, 38, 303-319.
doi:10.1080/03057640802287564
Poehlmann-Tynan, J., Gerstein, E. D., Burnson, C., Weymouth, L., Bolt, D. M., Maleck, S., &
Schwichtenberg, A. J. (2015). Risk and resilience in preterm children at age 6.
Development and Psychopathology, 27, 843-858. doi:10.1017/S095457941400087X
Ponciano, L., & Thai, K. P. (2016, October). Reducing the risk of failure in kindergarten with
ABCMouse. Poster presented at the Special Topics Meeting: Technology and Media in
Children’s Development, Irvine, CA.
Purcell, K. (2011). Search and email still top the list of most popular online activities: Two
activities nearly universal among adult internet users. Washington, D.C.: Pew Research
Center's Internet & American Life Project.
R Core Team. (2016). R: A language and environment for statistical computing [Computer
software]. Retrieved from http://www.r-project.org
Reid, K., Hresko, W. P., & Hammill, D. D. (2001). Test of Early Reading Ability (TERA-3) (3rd
ed.). Austin, TX: Pro-Ed.
Reitsma, P., & Wesseling, R. (1998). Effects of computer-assisted training of blending skills in
kindergartners. Scientific Studies of Reading, 2, 301-320.
doi:10.1207/s1532799xssr0204_1 136
Revelle, W. (2016). psych: Procedures for personality and psychological research. Evanston, IL:
Northwestern University. Retrieved from http://CRAN.R-project.org/package=psych
Ribisl, K. M., Walton, M. A., Mowbray, C. T., Luke, D. A., Davidson, W. S., & Bootsmiller, B.
J. (1996). Minimizing participant attrition in panel studies through the use of effective
retention and tracking strategies: Review and recommendations. Evaluation and Program
Planning, 19, 1-25. doi:10.1016/0149-7189(95)00037-2
Rice, M. L., Huston, A. C., & Wright, J. C. (1982). The forms and codes of television: Effects of
children's attention, comprehension, and social behavior. In D. Pearl, L. Bouthilet & J.
Lazar (Eds.), Television and behavior: Ten years of scientific progress and implications
for the eighties (pp. 24-38). Washington, DC: Government Printing Office.
Rideout, V., & Saphir, M. (2013). Zero to eight: Children's media use in America 2013. San
Francisco, CA: Common Sense Media.
Roberts, J. D., Chung, G. K. W. K., & Parks, C. B. (2016). Supporting children’s progress
through the PBS KIDS learning analytics platform. Journal of Children and Media, 10,
257-266. doi:10.1080/17482798.2016.1140489
Ronimus, M., & Lyytinen, H. (2015). Is school a better environment than home for digital game-
based learning? The case of GraphoGame. Human Technology: An Interdisciplinary
Journal on Humans in ICT Environments, 11, 123-147.
doi:10.17011/ht/urn.201511113637
Roseberry, S., Hirsh-Pasek, K., Parish-Morris, J., & Golinkoff, R. M. (2009). Live action: Can
young children learn verbs from video? Child Development, 80, 1360-1375.
doi:10.1111/j.1467-8624.2009.01338.x 137
Rosenthal, R. (1979). The file drawer problem and tolerance for null results. Psychological
Bulletin, 86, 638-641. doi:10.1037/0033-2909.86.3.638
Rosnow, R. L., & Rosenthal, R. (2009). Effect sizes: Why, when, and how to use them.
Zeitschrift für Psychologie / Journal of Psychology, 217, 6-14. doi:10.1027/0044-
3409.217.1.6
Rosser, J. C. J., Lynch, P. J., Cuddihy, L., Gentile, D. A., Klonsky, J., & Merrell, R. (2007). The
impact of video games on training surgeons in the 21st century. Archives of Surgery, 142,
181-186. doi:10.1001/archsurg.142.2.181
Rossi, P. H., Lipsey, M. W., & Freeman, H. E. (2003). Evaluation: A systematic approach (7th
ed.). Thousand Oaks, CA: Sage Publications.
Sanchez, R. P., & Lorch, E. P. (1999). Comprehension of televised stories by preschool children
with ADHD. Journal of Clinical Child Psychology, 28, 376.
Savage, R. S., Abrami, P. C., Piquette, N., Wood, E., Deleveaux, G., Sanghera-Sidhu, S., &
Burgos, G. (2013). A (Pan-Canadian) cluster randomized control effectiveness trial of the
ABRACADABRA web-based literacy program. Journal of Educational Psychology, 105,
310-328. doi:10.1037/a0031025
Scarborough, H. S., Ehri, L. C., Olson, R. K., & Fowler, A. E. (1998). The fate of phonemic
awareness beyond the elementary school years. Scientific Studies of Reading, 2, 115-142.
doi:10.1207/s1532799xssr0202_2
Scheller, A. (2014). KTEA-3: Overview: Part 1. Retrieved from
http://downloads.pearsonclinical.com/videos/KTEA3-Overview-
021414/KTEA3OverviewPart1WebianrHandout-02142014.pdf 138
Schiefele, U., Schaffner, E., Möller, J., & Wigfield, A. (2012). Dimensions of reading motivation
and their relation to reading behavior and competence. Reading Research Quarterly, 47,
427-463. doi:10.1002/RRQ.030
Schmitt, K. L., & Linebarger, D. L. (2009). Summative evaluation of PBS KIDS Island: Home
study. KLMedia Research & Children's Media Lab, University of Pennsylvania. Chicago,
IL.
Schmitt, K. L., Sheridan Duel, L., & Linebarger, D. L. (2017, April). Can young children learn
pre-literacy skills from online educational games? Poster presented at the Biennial
Meeting of the Society for Research on Child Development, Austin, TX.
Schwartz, C. E., Wright, C. I., Shin, L. M., Kagan, J., & Rauch, S. L. (2003). Inhibited and
uninhibited infants "grown up": Adult amygdalar response to novelty. Science, 300,
1952.
Segers, E., & Verhoeven, L. (2005). Long-term effects of computer training of phonological
awareness in kindergarten. Journal of Computer Assisted Learning, 21, 17-27.
doi:10.1111/j.1365-2729.2005.00107.x
Silva, P. A. (1990). The Dunedin Multidisciplinary Health and Development Study: A 15 year
longitudinal study. Paediatric and Perinatal Epidemiology, 4, 76-107.
doi:10.1111/j.1365-3016.1990.tb00621.x
Singer, J. L., & Singer, D. G. (1998). Barney & Friends as education and entertainment:
Evaluating the quality and effectiveness of a television series for preschool children. In J.
K. Asamen & G. L. Berry (Eds.), Research paradigms, television, and social behavior
(pp. 305-367). Thousand Oaks, CA: SAGE Publications. 139
Slavin, R. E. (1986). Best-evidence synthesis: An alternative to meta-analytic and traditional
reviews. Educational Researcher, 15, 5-11. doi:10.3102/0013189X015009005
Stanton, W. R., & Silva, P. A. (1992). A longitudinal study of the influence of parents and
friends on children's initiation of smoking. Journal of Applied Developmental
Psychology, 13, 423-434. doi:10.1016/0193-3973(92)90010-F
Stock, J. J., & Watson, M. W. (2006). Introduction to econometrics (2nd ed.). Boston, MA:
Addison Wesley.
Striano, T. (2016). Recruitment and access Doing developmental research: A practical guide
(pp. 92-109). New York, NY: Guilford Press.
Strouse, G. A., O'Doherty, K., & Troseth, G. L. (2013). Effective coviewing: Preschoolers’
learning from video after a dialogic questioning intervention. Developmental Psychology,
49, 2368-2382. doi:10.1037/a0032463
Sugden, N. A., & Moulson, M. C. (2015). Recruitment strategies should not be randomly
selected: Empirically improving recruitment success and diversity in developmental
psychology research. Frontiers in Psychology, 6. doi:10.3389/fpsyg.2015.00523
Takeuchi, L., & Stevens, R. (2011). The new coviewing: Designing for learning through joint
media engagement. Retrieved from http://www.joanganzcooneycenter.org/wp-
content/uploads/2011/12/jgc_coviewing_desktop.pdf
Torgerson, C. J. (2007). The quality of systematic reviews of effectiveness in literacy learning in
English: a ‘tertiary’ review. Journal of Research in Reading, 30, 287-315.
doi:10.1111/j.1467-9817.2006.00318.x 140
Tourangeau, K., Le, T., & Nord, C. (2005). Early Childhood Longitudinal Study, Kindergarten
class of 1998-99 (ECLS-K): Fifth-grade methodology report (NCES 2006-037).
Washington, DC: National Center for Education Statistics, U.S. Department of
Education.
United States Census Bureau. (2016). Poverty thresholds by size of family and number of
children. Retrieved from http://www.census.gov/data/tables/time-series/demo/income-
poverty/historical-poverty-thresholds.html
Valkenburg, P. M., & Peter, J. (2013). The differential susceptibility to media effects model.
Journal of Communication, 63, 221-243. doi:10.1111/jcom.12024 van Buuren, S., & Groothuis-Oudshoorn, K. (2011). mice: Multivariate Imputation by Chained
Equations in R. Journal of Statistical Software, 45, 1-67. URL:
http://www.jstatsoft.org/v45/i03/. van Houwelingen, H. C., Arends, L. R., & Stijnen, T. (2002). Advanced methods in meta-
analysis: multivariate approach and meta-regression. Statistics in Medicine, 21, 589-624.
doi:10.1002/sim.1040
Viechtbauer, W. (2007). Accounting for heterogeneity via random-effects models and moderator
analyses in meta-analysis. Zeitschrift für Psychologie/Journal of Psychology, 215, 104-
121. doi:10.1027/0044-3409.215.2.104
Viechtbauer, W. (2010). Conducting meta-analyses in R with the metafor Package. Journal of
Statistical Software, 36, 1-48. URL: http://www.jstatsoft.org/v36/i03/. 141
Viechtbauer, W. (2015). Comparing estimates of independent meta-analyses or subgroups.
Retrieved from http://www.metafor-
project.org/doku.php/tips:comp_two_independent_estimates
Wagner, R. K., Torgesen, J. K., Rashotte, C. A., Hecht, S. A., Barker, T. A., Burgess, S. R., . . .
Garon, T. (1997). Changing relations between phonological processing abilities and
word-level reading as children develop from beginning to skilled readers: A 5-year
longitudinal study. Developmental Psychology, 33, 468-479. doi:10.1037/0012-
1649.33.3.468
Walker, H. (2011). Evaluating the effectiveness of apps for mobile devices. Journal of Special
Education Technology, 26, 59-63.
Wartella, E. (2015). Educational apps: What we do and do not know. Psychological Science in
the Public Interest, 16, 1-2. doi:10.1177/1529100615578662
Wartella, E., Beaudoin-Ryan, L., Blackwell, C. K., Cingel, D. P., Hurwitz, L. B., & Lauricella,
A. R. (2016). What kind of adults will our children become? The impact of growing up in
a media-saturated world. Journal of Children and Media, 10, 13-20.
doi:10.1080/17482798.2015.1124796
Wartella, E., Blackwell, C. K., Lauricella, A. R., & Robb, M. B. (2013). Technology in the lives
of educators and early childhood programs: 2012 survey of early childhood educators.
Latrobe, PA: Fred Rogers Center at Saint Vincent's College.
Webster, J. G. (2014). The marketplace of attention: How audiences take shape in a digital age.
Cambridge, MA: MIT Press. 142
What Works Clearinghouse. (2014). Procedures and standards handbook (3rd ed.). Washington,
D.C.: What Works Clearinghouse, Institute of Education Sciences, DoEd.
Whitehurst, G. J., & Lonigan, C. J. (2001). Get ready to read. Columbus, OH: Pearson Early
Learning.
Wilson, V. (2015, September 16). New Census data shows no progress in closing stubborn racial
income gaps [Blog post], Economics Policy Institute. Retrieved from
http://www.epi.org/blog/new-census-data-show-no-progress-in-closing-stubborn-racial-
income-gaps/
Wright, J. C., Huston, A. C., Murphy, K. C., St. Peters, M., Piñon, M., Scantlin, R., & Kotler, J.
(2001). The relations of early television viewing to school readiness and vocabulary of
children from low-income families: The early window project. Child Development, 72,
1347-1366. doi:10.1111/1467-8624.t01-1-00352
Wu, F., & Qi, S. (2006). Longitudinal effects of parenting on children's academic achievement in
African American families. Journal of Negro Education, 75, 415-429.
143
Appendix A:
Academic Search Terms
• Ready To Learn • Dragon Tales • Sesame Street
• RTL • Duck’s Alphabet • Sid the Science Kid
• Corporation for • Fizzy’s Lunch Lab • Super WHY
Public • Great Word Quest • SuperWHY
Broadcasting • Martha Speaks • The Electric
• CPB • Miss Spider’s Company
• Public Sunny Patch • UMIGO
Broadcasting Friends • Word Girl
Service • Mission to Planet • WordGirl
• PBS 429 • Word World
• Arthur • Nova the Robot • WordWorld
• Barney and Friends • Odd Squad • World of Words
• Between the Lions • Peg + Cat • Any time is
• Cat in the Hat • Pocoyo learning time
Knows a Lot About • Postcards from • Anytime is learning
That Buster time
• Clifford the Big • R U There • Learning Triangle
Red Dog • RU There • Raising Readers
• Curious George • Reading Rainbow
• Dinosaur Train • Ruff Ruffman
Appendix B:
Mean Effect Sizes and Key Codable Characteristics of Included Evaluations
Evaluation d Texts Literacy Sample Study Scope Setting Weeks Nature of Nature of Assessments Grant RTL Available Categories Characteristics Design of RTL Control Cycle Funding Tested Media Treatment Status
Chiong and 0.31 Unpublished Alphabet 64% Within Multiple Home 2 Martha N/A Mix of 2005- Funded Shuler (2010); reports knowledge; Caucasian; regions Speaks custom 2010 Rockman et al. Phonological 16% African and Super broad and (2010) processing; American; 6% WHY! standardized Vocabulary; Hispanic; 3% apps Narrative Asian; 44% without comprehension Female; 72% extension School-age; activities 39% Low- income
Dwyer (2010); 0.57 Journal Vocabulary 58% Within One School 8 Sesame Business Mix of 2005- Funded Neuman and article; Caucasian; and region Street TV as usual custom 2010 Dwyer (2011) Dissertation 27% African Between with narrow, American; 0% extension custom School-age activities broad, and standardized
Educational 0.01 Unpublished Alphabet 93% African Between One Home 8 Between Non-RTL Mix of 2005- Funded Development reports knowledge; American; region the Lions Media custom 2010 Center and SRI Print concepts; 49% Female; and broad and International Phonological 0% School-age Sesame standardized (2008); Penuel, processing Street TV Clements, with Pasnik, and extension Llorente (2008) activities
Ferrell (2002) -0.2 Thesis Alphabet 56% Between Multiple Home Sesame Non-RTL All custom 1994- Not knowledge; Caucasian; regions Street TV Media and broad 2000 Funded Other 15% African Alternate American; 6% curriculum Asian; 2% Native American; 50% Female; 100% School- 144 age;
Garrity, 0.14 Unpublished Phonological 18% Between One Home 10 The Business Mix of 2005- Funded Piotrowski, report processing; Caucasian; region and Electric as usual custom 2010 McMenamin, Vocabulary 71% African school Company narrow and and Linebarger American; TV, with standardized (2010) 51% Female; and 100% School- without age; 81% Low- computer income; 7% games, ELL with and without extension activities
Godfrey (2015) 0.55 Dissertation Vocabulary; 65% Within One School 6 WordWor N/A All 2010- Not Narrative Caucasian; 5% region ld TV standardized 2015 Funded comprehension African with and American; without 20% Asian; extension 57% Female; activities 0% School-age
Harper et al. 0.23 Unpublished Print concepts; 0% School-age Between One School 20 Between Business All 2005- Funded (2006) report Narrative region the Lions as usual standardized 2010 comprehension; TV with Multiple skills extension activities
Jennings -0.1 Unpublished Alphabet 39% Female Between One School 1 Super Business Mix of 2010- Funded (2013) report knowledge; region WHY! TV as usual custom 2015 Phonological with narrow, processing; extension custom Other activities broad, and standardized
Kaefer and 0.46 Journal Vocabulary 36% Between School 2 Sesame Business Mix of 2010- Not Neuman (2013) article Caucasian; Street TV as usual custom 2015 Funded 34% African with narrow and American; 4% extension custom Hispanic; 11% activities broad Asian; 54% Female; 0% School-age; 77% Low- income; 100% ELL
145
Kelly (2003) -0.3 Dissertation Multiple skills 60% Between One Home 19 Between Business All 2000- Not Caucasian; 4% region and the Lions as usual standardized 2005 Funded African school TV with American; extension 27% Asian; activities 47% Female; 100% School- age; 9% ELL
Linebarger 0.09 Journal Alphabet 81% Between One School 3.5 Between Business Mix of 1994- Funded (2000); article; knowledge; Caucasian; 6% region the Lions as usual custom 2000 Linebarger, Unpublished Phonological African TV narrow, Kosanic, report processing; American; 7% without custom Greenwood, Vocabulary; Hispanic; 1% extension broad, and and Doku Multiple skills Native activities standardized (2004) American; 49% Female; 100% School- age; 34% Low- income; 7% ELL
Linebarger 0.02 Unpublished Alphabet 46% Female; Between One School 30 Between Business All 2005- Funded (2009) report knowledge; 0% School-age region the Lions as usual standardized 2010 Print concepts; TV with Phonological extension processing; activities Vocabulary; Narrative comprehension; Multiple skills
Linebarger and -0.1 Journal Narrative 52% Female; Between One School 8 Clifford Non-RTL Mix of 2005- Funded Piotrowski article; comprehension; 0% School-age region the Big Media custom 2010 (2007, 2009) Unpublished Multiple skill Red Dog broad and report TV standardized without extension activities
Linebarger and 0.28 Journal Vocabulary 90% Between One Home 120 Arthur, Mix of All 2000- Not Walker (2005) article Caucasian; region Clifford, custom standardized 2005 Funded 55% Female; Barney, narrow 0% School- Dragon and age; 9% Low- Tales, and custom
income Sesame broad 146 Street TV
Linebarger, 0.27 Journal Alphabet 68% Between One Home 8 Super Non-RTL Mix of 2005- Funded McMenamin, article; knowledge; Caucasian; 1% region WHY! TV Media custom 2010 and Unpublished Print concepts; African without narrow, Wainwright report Phonological American; 2% extension custom (2009) processing; Hispanic; 44% activities broad, and Vocabulary; Female; 0% standardized Narrative School-age; comprehension; 22% Low- Multiple skills income
Linebarger, 0.2 Journal Vocabulary 56% Female; Between Multiple Home 4 Martha Business All custom 2005- Funded Moses, article; 46% School- regions Speaks as usual narrow 2010 Jennings, and Conference age; 95% Low- TV McMenamin presentation; income without (2010); Unpublished extension Linebarger, reports activities Moses, and McMenamin (2010a, 2010b); Linebarger, Moses, Garrity Liebeskind, and McMenamin (2013); Moses, Linebarger, McMenamin, and Liss- Mariño (2009)
Linebarger, 1.06 Conference Vocabulary 100% School- Between One School 12 Postcards Business All custom 2005- Funded Piotrowski, and presentation; age; 100% region from as usual narrow 2010 Vaala (2007); Unpublished ELL Buster and Non- Piotrowski, report TV with RTL Vaala, and extension Media Linebarger activities (2009)
Marshall, Lapp, 0.39 Unpublished Alphabet 25% Within Multiple School 1 Super N/A All custom 2005- Funded and Cavoto report knowledge; Caucasian; regions WHY! TV narrow 2010 (2009) Phonological 55% African with processing; American; extension Vocabulary; 16% Hispanic; activities Other 48% Female; 59% School- age
147
McCarthy et al. 0.23 Unpublished Phonological 100% School- Within Multiple School 5.5 The N/A All 2010- Funded (2011) report processing age regions Electric standardized 2015 Company TV and computer games with extension activities
Meyer and 0.26 Unpublished Alphabet 53% Between Multiple School 1 Super Alternate Mix of 2005- Funded Sroka (2010) report knowledge; Caucasian; regions WHY! TV curriculum custom 2010 Phonological 26% African with narrow and processing; American; extension standardized Vocabulary; 19% Hispanic; activities Other; Multiple 2% Asian; skills 53% Female; 14% ELL
Michael Cohen 0.15 Unpublished Alphabet 32% Between Multiple School 6 WordWor Business Mix of 2005- Funded Group (2009) report knowledge; Caucasian; regions ld TV as usual custon 2010 Print concepts; 30% African without narrow and Phonological American; extension standardized processing; 21% Hispanic; activities Vocabulary 6% Asian; 1% Native American; 0% School-age; 33% Low- income
Michael Cohen 0.33 Unpublished Alphabet 6% Caucasian; Between Multiple Home 2 Duck's Non-RTL Mix of 2005- Funded Group (2010a) report knowledge; 57% African regions Alphabet Media custom 2010 Phonological American; computer broad and processing; 23% Hispanic; games standardized Multiple skills 47% Female; without 0% School- extension age; 60% Low- activities income
Michael Cohen 0.19 Unpublished Vocabulary; 5% Caucasian; Between Multiple School 2 Mission Non-RTL Mix of 2005- Funded Group (2010b) report Narrative 35% African regions to Planet Media custom 2010 comprehension; American; 429 narrow and Other 57% Hispanic; computer standardized 47% Female; games
78% Low- without 148
income; 100% extension School-age activities
Michael Cohen - Unpublished Vocabulary; 14% Between Multiple School 2 R U Non-RTL Mix of 2005- Funded Group (2010c) 0.02 report Narrative Caucasian; regions There? Media custom 2010 comprehension; 51% African TV and narrow, Other American; computer custom 21% Hispanic; games broad, and 6% Asian; without standardized 56% Female; extension 100% School- activities age; 25% Low- income
Michael Cohen 0.12 Unpublished Vocabulary; 100% Between Multiple School 3 Pocoyo Non-RTL Mix of 2010- Funded Group (2013) report Other Hispanic; 52% regions apps Media custom 2015 Female; 0% without narrow and School-age; extension standardized 56% Low- activities income; 94% ELL
Michael Cohen 0.16 Unpublished Vocabulary; 6% Caucasian; Between Multiple School 4 Pocoyo Non-RTL Mix of 2010- Funded Group (2015) report Multiple skills 25% African regions apps with Media custom 2015 American; extension narrow and 59% Hispanic; activities standardized 13% Asian; 55% Female; 0% School- age; 59% Low- income; 69% ELL
Naigles (2000) 0.24 Book Vocabulary 85% Between One School 2 Barney All 1994- Not chapter Caucasian; region TV standardized 2000 Funded 59% Female; without 0% School- extension age; 0% ELL activities
Naigles et al. 0.15 Book Other 88% Between One School 2.5 Barney Business All custom 1994- Funded (1997); Singer chapter; Caucasian; region TV as usual narrow 2000 and Singer Unpublished 50% Female; without (1998) report 0% School- extension age; 0% ELL activities
Neuman and 1.12 Journal Vocabulary 92% Within One School 8 Sesame N/A All custom 2010- Not Kaefer (2013) article Caucasian; 3% region Street TV narrow 2015 Funded 149 African with
American; 4% extension Hispanic; 52% activities Female; 0% School-age; 100% Low- income
Neuman, 0.46 Journal Vocabulary 37% Between One School 8 Sesame Alternate All custom 2005- Funded Newman, and article; Caucasian; region Street TV curriculum narrow 2010 Dwyer (2010, Unpublished 43% African with 2011) report American; 3% extension Hispanic; 12% activities Asian; 52% Female; 0% School-age; 100% Low- income; 4% ELL
Penuel et al. 0.32 Journal Alphabet 6% Caucasian; Between One School 10 Between Non-RTL Mix of 2005- Funded (2012); Penuel article; knowledge; 28% African region the Lions, Media custom 2010 et al. (2009) Unpublished Phonological American; Super broad and report processing; 53% Hispanic; WHY! and standardized Print concepts 10% Asian; Sesame 3% Native Street TV American; and 51% Female; computer 0% School- games age; 68% Low- income
Phillips (2008) 0.41 Unpublished Alphabet 26% Within Multiple School 1 Super N/A All custom 2005- Funded report knowledge; Caucasian; regions WHY! TV narrow 2010 Phonological 44% African with processing; American; extension Vocabulary; 24% Hispanic; activities Other 46% Female; 68% School- age
Piotrowski, -0.1 Journal Alphabet 49% Female; Between One School 4 Between Business All 2005- Funded Jennings, and article; knowledge; 65% School- region the Lions as usual standardized 2010 Linebarger Unpublished Phonological age TV with (2012); report processing; and Piotrowski, Vocabulary; without
Linebarger, and Multiple skills extension 150 activities
Jennings (2009)
Prince, Grace, 0.26 Unpublished Print concepts; 20% Between One School 26 Between Business All 2000- Funded Linebarger, report Vocabulary; Caucasian; region the Lions as usual standardized 2005 Atkinson, and Narrative 48% African TV with Huffman comprehension; American in extension (2002) Other; Multiple one activities skills comparison; 100% Native American in another comparison; 71% School- age; 80% Low- income
Register (2003, 0.05 Journal Alphabet 100% School- Between One School 3.4 Between Business All 2000- Not 2004) article; knowledge; age region the Lions as usual standardized 2005 Funded Dissertation Print concepts; TV with and Phonological and Alternate processing; without curriculum Narrative extension (in comprehension; activities separate Multiple skills compariso ns)
Rollins (2000) -0.3 Thesis Other 67% Female; Within One School 3 Reading N/A All custom 1994- Not 100% School- region Rainbow narrow 2000 Funded age TV without extension activities
Schmitt and 0.43 Unpublished Alphabet 43% Female; Within One Home 4 Between N/A All 2005- Funded Linebarger, report knowledge; 0% School-age region the Lions, standardized 2010 (2009) Phonological Martha processing; Speaks, Other Super WHY!, Sesame Street, and WordWor ld
computer 151 games
without extension activities
Schmitt, 0.17 Conference Alphabet 31% Between One Home 6.8 Between Non-RTL Mix of 2005- Funded Sheridan Duel, presentation; knowledge; Caucasian; region the Lions, Media custom 2010 and Linebarger Unpublished Phonological 18% African Martha narrow, (2016, 2017); report processing; American; Speaks, custom Schmitt, Vocabulary; 29% Hispanic; Super broad, and Sheridan, Multiple skills 8% Asian; WHY!, standardized McMenamin, 52% Female; Sesame and Linebarger 31% School- Street, (2010) age; 62% Low- and income; 15% WordWor ELL ld computer games with and without extension activities
Silverman 0.18 Journal Vocabulary 33% African Within One School 4 Martha N/A All custom 2005- Funded (2009a, 2013) article; American; region Speaks narrow 2010 Unpublished 63% Hispanic; TV report 1% Asian; without 55% Female; extension 100% School- activities age; 80% Low- income; 67% ELL
Silverman 0.14 Unpublished Vocabulary 6% Caucasian; Within One School 4 Martha N/A All custom 2005- Funded (2009b) report 19% African region Speaks narrow 2010 American; TV and 61% Hispanic; computer 8% Asian; games (in 49% Female; separate 100% School- compariso age; 94% Low- ns) with income extension activities
Silverman - Journal Vocabulary 10% Within One School 4 Martha Alternate All custom 2005- Funded (2009c, 2013) 0.02 article; Caucasian; region Speaks curriculum narrow 2010
Unpublished 24% African TV with 152 report American; and
53% Hispanic; without 2% Asian; extension 49% Female; activities 100% School- age; 51% Low- income; 59% ELL
Silverman 0.27 Conference Vocabulary 12% Between Multiple School 8 Martha Alternate All custom 2005- Funded (2009d); presentation; Caucasian; regions Speaks curriculum narrow 2010 Silverman and Unpublished 63% African TV with Carlis (2010) report American; extension 10% Hispanic; activities 8% Asian; 55% Female; 100% School- age; 15% ELL
Silverman, 0.17 Journal Vocabulary; 62% Between One School Martha Business Mix of 2010- Funded Kim, Hartranft, article Narrative Caucasian; region Speaks as usual custom 2015 Nunn, and comprehension 20% African TV with narrow and McNeish American; extension standardized (2016) 50% Female; activities 100% School- age; 61% Low- income; 8% ELL
Tollefson 0.41 Thesis Multiple skills 92% Between One School 30 Between Alternate All 2010- Not (2013) Caucasian; 5% region the Lions curriculum standardized 2015 Funded African TV with American; 3% extension Hispanic; 45% activities Female; 0% School-age; 31% Low- income
Uchikoshi 0.25 Journal Phonological 100% Between One School 18 Arthur or Business All 2000- Not (2004, 2005, articles; awareness; Hispanic; 47% region Between as usual standardized 2005 Funded 2006a, 2006b) Dissertation Vocabulary; Female; 100% the Lions Multiple skills School-age; TV 80% Low- without income extension activities
153
Note. This table reflects the information that was coded from each evaluation, collapsing across comparisons. It only references measures for which there was sufficient statistical information to calculate effects. For the purposes of this meta-analysis, cases in which authors compared two different versions of RTL treatment to one another without a comparison group were coded as within- subjects designs in the present dataset, with each version of the RTL treatment considered a separate comparison.
154