Skip to main content
  • Research article
  • Open access
  • Published:

Shared reading quality assessment by parental report: preliminary validation of the DialogPR

Abstract

Background

The American Academy of Pediatrics (AAP) recommends shared reading beginning as soon as possible after birth to promote healthy development. Shared reading quality can strongly influence outcomes, especially in children from low-SES households. Dialogic reading is a method developed to enhance verbal interactivity and engagement through book sharing, advocated by the AAP and clinic-based programs such as Reach Out and Read. There is no brief, validated, caregiver report measure of dialogic reading or shared reading quality currently available.

Methods

This cross-sectional study involved 49 healthy mother-child dyads (mean child age 4.5 yrs., SD = 0.6 yrs.) from 2 separate MRI-based studies. The DialogPR was administered by trained research coordinators following MRI, along with the READ subscale of the validated StimQ-P measure of home cognitive environment. The DialogPR consists of eight items developed in consultation with experts in early literacy, based on the PEER/CROWD dialogic reading conceptual model. Estimated reading level is 6th grade. Descriptive statistics were computed at both the item and scale levels. Modern theory Rasch methods were used to analyze all eight DialogPR items along with preliminary estimates of reliability and validity.

Results

Our combined sample involved 15 boys and 34 girls, and was diverse in terms of age, household income, and maternal education. DialogPR administration time was less than 2 min, with no problems reported. The DialogPR demonstrated strong internal consistency and reliability (Cronbach’s alpha = 0.82), and criterion-related validity with the StimQ-P READ (Spearman’s rho coefficient = 0.53). Rasch analysis revealed strong psychometric properties in terms of reliability, variability in item difficulty, and inter-item and item-measure correlations.

Conclusions

Preliminary evidence suggests that the DialogPR may be an efficient means to assess shared reading quality and dialogic reading via caregiver report for clinical and research purposes, warranting further investigation.

Peer Review reports

Background

The American Academy of Pediatrics (AAP) recommends literacy promotion in primary care beginning as soon as possible after birth [1]. Literacy is a major social and public health issue, the cost of low achievement estimated at over $350 billion per year in the United States, and over $1.2 trillion worldwide [2]. As of 2015, 64% of US 4th graders scored below proficient in reading, largely unchanged from prior reports [3], and lower still for children from minority and low-socioeconomic status (SES) households [3, 4]. While 5 to 14% of reading difficulties have an organic cause (e.g. dyslexia) [5], the majority are environmentally based and largely preventable, a consequence of inadequate resources, motivation and/or stimulation required to learn to read [2]. Many children arrive at school at a substantial disadvantage in readiness, unlikely to catch up with peers as academic demands accelerate [4]. Thus, early screening and intervention offer large potential savings in terms of productivity and health [6,7,8].

Parents are considered to be a child’s “first and most important teachers [9].” Cognitive stimulation in the home, exemplified by shared reading [10,11,12], greatly influences educational and health outcomes [13, 14]. Literacy promotion programs based in pediatric clinics (notably Reach Out and Read [15]), preschool [16], and home visitation [17] share a goal of enhancing home literacy environment, a composite of quantitative and qualitative factors [1, 18]. Quantitative factors typically include number of children’s books in the home, frequency of shared reading (e.g. days per week, minutes per day), approximate age when shared reading was initiated, and variety/type of books read [19, 20]. Qualitative factors typically include parent and child interest in and enjoyment of reading, and verbal and social-emotional interactivity during shared reading. Given the ease of assessment via parental report, quantitative factors are most often screened and addressed [21,22,23]. However, qualitative factors such as verbal interactivity during shared reading may be even more important and have significant impact [14, 24], though are often overlooked.

Dialogic reading is a method of shared (usually parent-child) reading developed to promote reciprocal dialogue between a caregiver and child during story sharing [25], advocated by Reach Out and Read and the AAP [1, 15]. The acronym PEER is used to reflect the dialogic process [26], as follows: 1) Prompt the child to say something about the story, 2) Evaluate what the child says, 3) Expand on what the child says, and 4) Repeat and reinforce associations. Similarly, the acronym CROWD is used to reflect evocative caregiver prompts: 1) Completion (of a sentence), 2) Recall earlier aspects of the story, 3) Open-ended questions, 4) Wh- questions, and 5) Distancing (relate the story to the child’s experience). Behavioral evidence suggests that this qualitative aspect of shared reading may confer moderate to large cognitive and social-emotional benefits beginning in infancy [27], especially for children from low-SES backgrounds [28, 29]. However, there is currently no validated measure of dialogic reading or shared reading quality currently available that is feasible for clinical use.

The purpose of this study was to develop and pilot test a brief caregiver report measure of shared reading quality (DialogPR) based on a dialogic reading conceptual model. Our eight-item measure was reviewed by experts in measure design and child development, pilot tested for clarity, and then administered as an exploratory aim in 2 unrelated, MRI-based studies involving healthy, preschool-age children and their mothers: one comprised exclusively of low-SES (n = 22) and the other of largely higher-SES (n = 27) dyads. The validated StimQ-P measure of cognitive stimulation in the home [30, 31], which includes a subscale of home reading practices, was administered as an external standard. Psychometric analyses, including modern-theory Rasch modeling, were performed. Our hypothesis was that the DialogPR would be feasible to administer, reliable, and valid in this combined sample, attesting to the value of a cohesive conceptual model of dialogic reading, warranting further investigation.

Methods

Sample

This study involved 49 healthy mother-child dyads enrolled in two recent MRI-based studies of cognitive and brain development at our institution, which were combined for the present analysis. Inclusion criteria for both studies were: preschool-age (3–5 years), full-term gestation, native English-speaking household, no history of brain injury, developmental delay or stimulant use, and no contraindications to MRI. The first sample (n = 22) was drawn from an ongoing home-injury prevention trial involving low-SES families at-risk for poor health and social outcomes [32]. Girls were exclusively sampled for this study due to time/budget constraints and higher MRI success rates for girls at this age [33]. The second sample (n = 27) involved mother-child dyads recruited via advertisement from employee families at a large academic medical center, with no gender constraint [34]. Families were compensated for time and travel, and each study was approved by the Cincinnati Children’s Hospital Institutional Review Board.

Instrument

The conceptual model for our DialogPR instrument was the PEER/CROWD dialogic reading construct developed by Whitehurst, et al. [25]. Our DialogPR instrument included a scripted introduction: “When deciding your answer, try to think about how you and [CHILD’S FIRST NAME] have read children’s books together over the past month.” It was comprised of eight questions: 1) frequency of discussing what the book might be about before reading, 2) frequency of discussion during story sharing, 3) five questions referencing frequency of respective CROWD prompts during story sharing, and 4) frequency of discussion of what the book was about after reading. Categorical responses for questions 1, 2, and 8 were: “always,” “usually,” “sometimes,” and “never.” Responses for CROWD questions 3 to 7 were anchored to a hypothetical book (“When you are reading a children’s book with [CHILD’S FIRST NAME], how often do you stop reading to do the following things?”): “on most pages,” “on around half of the pages,” “on a few pages,” or “never.” An ordinal scoring system was used for each question, with response options ranging from 0 to 3 points for each item, with higher scores reflecting greater frequency of dialogic behaviors.

Wording for the DialogPR was refined in consultation with experts in measure development at our institution, and pilot tested for clarity among colleagues and families attending a hospital-based primary care clinic. Estimated Flesch-Kincaid reading level was 6th grade. Research coordinators practiced administration with the principal investigator and were instructed to adhere to instrument wording verbatim. Research coordinators administered the DialogPR to mothers at the study visit following the child’s MRI, and transcribed response data into a secure REDCap database [35].

Reference measure

The StimQ-P served as the criterion-referenced standard for this study [30, 31], and was administered to mothers following the DialogPR. The StimQ-P is a validated parental report measure of cognitive stimulation in the home for children 36 to 72 months of age, and consists of 4 subscales involving mostly “yes/no” questions: 1) Availability of learning materials (ALM); 2) Reading (READ), reflecting access to books, frequency of shared reading, variety of books read, and interactivity/quality of reading; 3) Parental Involvement in Developmental Advance (PIDA); and 4) Parental Verbal Responsivity (PVR). StimQ has been found to have excellent internal consistency (Cronbach’s alpha 0.88 to 0.93), convergent validity with the HOME inventory, and predictive/concurrent validity with language scores, including for its subscales [30]. For the aims of this study, and to keep our assessment brief, only the READ subscale was administered.’

Statistical analyses

Data analysis proceeded in three distinct steps. First, demographic characteristics were computed for the entire sample of 49 children. Second, descriptive statistics were computed for all variables in the data set, at both the scale and item levels. All eight DialogPR items were evaluated for smoothness, modality, difficulty, polarity, and sufficiency of observations across levels. Modern theory Rasch rating scale methods were used for analysis due to the identical, ordered categorical nature of response options across all items [36, 37]. Model fit was tested for each item to identify any that were markedly or unnecessarily influencing the scale-level distributions. Third, preliminary estimates of DialogPR’s reliability and validity were computed, beginning with Cronbach’s coefficient alpha (αCr) and standard error of measurement (SEM) as our measures of reliability, and a Spearman-rho (rρ) correlation coefficient between DialogPR total score and StimQ-P READ subscale score as our measure of criterion-related validity. Spearman-rho correlation was chosen given relatively small sample size warranting a conservative, non-parametric approach. The criterion for statistical significance was set at the unadjusted α = 0.05 level due to the preliminary nature of the study. All analyses were conducted using SAS v9.4 and Winsteps v4.0 software.

Results

Demographic characteristics

This study involved two sample populations from separate studies involving mothers and their preschool-age children. These were combined in the present analyses, and are summarized in Table 1.

Table 1 Demographic Characteristics of Participants (n = 49)

Descriptive statistics for the DialogPR

Research coordinators reported no difficulty administering the DialogPR, with all subjects completing the survey in less than 2 min. Mean DialogPR score was 12.6 (SD = 4.8), and ranged from a minimum of 5 to a maximum of 24. Mean STIMQ-READ score was 14.1 (SD = 2.4) and ranged from a minimum of 9 to a maximum of 19 across 49 participants. Histograms of DialogPR and StimQ READ score distributions are provided in Fig. 1.

Fig. 1
figure 1

Distributions of DialogPR and StimQ-P READ Scores. Histograms for DialogPR and StimQ-P total scores (n = 49). Shapiro-Wilk tests revealed non-normality for each (p = 0.04 for each), and non-parametric analyses were conducted

Item analysis

Item level information for the DialogPR is provided in Table 2. Rasch estimates of item difficulty ranged from − 2.00 (less difficult) to 1.06 (more difficult). Point-measure correlations ranged from 0.44 (item 2) to 0.73 (item 4), suggesting a moderate relationship between each of the DialogPR items and the entire scale. Item fit statistics using empirically-derived z-values were all well below the traditional + 2 standard deviations, suggesting no outliers likely to influence the distributions [38]. With respect to inter-item correlations of the DialogPR, significant correlations ranged from rρ = 0.29 (Q2-Q6; low) to 0.62 (Q3-Q4; moderate), shown in Table 3. There was no significant correlation between child gender (male = 1) and DialogPR item or total scores, with the exception if item 2, which was negatively correlated (p < 0.05). Household income was negatively correlated with items 2, 4, 5, 6, and total DialogPR score (p < 0.05). Maternal education level was negatively correlated with item 6 only (p < 0.05). These demographic correlations are summarized in Table 3. StimQ-P READ score was negatively correlated with household income and maternal education (p < 0.05), but not with child gender.

Table 2 Item Analysis and Summary Statistics for DialogPR (Rasch Analysis)
Table 3 Intercorrelation Table for DialogPR Items, StimQ-P READ Score and Selected Demographic Characteristics

Reliability and validity

For reliability, internal consistency was acceptable to good at rCo-α = 0.82. The DialogPR’s standard error of measurement, a measure of reproducibility of test scores, was estimated at SEM = 2.0. This means that the DialogPR “true” score for participants is expected to fall, on average, within the range of 8.7 to 16.5, 95% of the time. For criterion-related validity, the correlation between the DialogPR and the STIMQ-READ subscale score was rρ = 0.53 (p < 0.001). There was significant, positive correlation between individual DialogPR items and StimQ-READ (p < 0.05) with the exception on items Q1, Q2, and Q3, which were non-significant (Table 3).

Discussion

Literacy is a major predictor of educational, occupational and health outcomes [2]. While causality has not definitively been proven [39], important drivers of reading difficulties include deficient or absent reading role models in the home and consequently impaired abilities, attitudes and routines [40, 41]. Substantial resources are devoted to initiatives to enhance home literacy environment [14, 15, 42], efficacy and improvement dependent on the ability to collect data efficiently and reliably. Research to date has largely relied on aspects of home literacy environment that are straightforward to assess via parental report, such as access to books and reading frequency, potentially neglecting critical behaviors such as verbal interactivity. These “dialogic” qualities may be particularly at-risk in parents lacking confidence or experience with shared reading, such as those from low-SES households [43], and can influence reading outcomes [24, 28].

The DialogPR instrument performed remarkably well in preliminary psychometric analysis. With only 8 items and brief administration time, this performance suggests potential value in clinical and research settings, though more expansive validation studies are needed. We attribute this strong performance to an evidence-based conceptual model of dialogic reading [26], which guided item development. In general, DialogPR items showed low to moderate inter-item correlation and good response variability, suggesting that each contributed uniquely to the overall score, and that parents could identify their own, shared reading behavior in a range of options. The five items corresponding to specific CROWD prompts (items 3–7) each performed well, with moderate correlations between them suggesting cohesion as reading behaviors. Interestingly, items 3 and 4 were less strongly correlated with the other CROWD items than with each other, which we suspect may be attributable to completion and recall prompts seeming more abstract or unnatural to some parents at this age, compared to the other types of questioning. This may also be why items 3 and 4 were the behaviors parents reported least often. By contrast, more straightforward items 5 and 6 (open-ended and wh- questions) were more strongly correlated with each other than with the other CROWD items, yet were among the easiest to endorse. Items 1 and 8, reflecting discussion before and after reading, respectively, and not core PEER/CROWD components, were also among the most highly correlated item pairs, yet discussion before reading was reported far less frequently.

The item performing the weakest (though still respectably), item 2, concerns frequency of pauses to answer a child’s questions, intended to reflect dialogic evaluation and expansion (Es in PEER). Possibly due to its general wording, all but 7 parents responded that they “always” or “usually” do so, the other 7 responding “sometimes,” and none responding “never.” It is intriguing that the only significant inter-item correlations involving item 2 were with items 5 (open-ended questions) and 6 (“wh-” questions). While speculative, this finding seems intuitive and suggests that this sample of mothers associated these types of relatively straightforward prompts with further dialogue. Variability in this item may have been higher if it instead asked how often specific types of responses (e.g. evaluations and expansions) were made. Such refinement in future versions of DialogPR to more explicitly assess evaluation and expansion during shared reading, may be worthwhile.

Initial evidence suggests that the most frequently endorsed items for these parents, reflecting the most common or least “difficult” shared reading behaviors, were item 2 (frequency of pauses during reading to respond to a child’s questions or comments, discussed above), item 8 (discussion after reading), item 6 (“wh-” questions), and item 5 (open-ended questions). The least commonly endorsed, or most “difficult” shared reading behaviors were item 4 (recall prompts), item 1 (discussion before reading), and item 3 (completion prompts). Overall, this item performance seems highly intuitive, the more straightforward, or “natural” behaviors reported most often, with the possible exception of discussion before reading to generate interest. Interestingly, this finding also concurs with shared reading observations conducted for a separate MRI study involving our low-SES population (n = 22) [32] where wh- prompts and open-ended questions were used almost exclusively.

Variability in item responses (with the marginal exception of item 2), and even slight skew towards lower scores, suggests that the DialogPR was not overly influenced by social desirability bias, a universal concern in parental report measures [44]. While both DialogPR and StimQ READ scores were not technically normally distributed (Shapiro-Wilk test, p = 0.04), strong correlation between DialogPR and this validated instrument is also reassuring in this respect. Interestingly, DialogPR total scores were negatively correlated with household income, though not with child gender or maternal education. StimQ-P READ scores were similarly negatively correlated with income, and also with maternal education. While speculative, these paradoxical findings may be attributable to a more nuanced type of reporting bias, where mothers from low-SES backgrounds may be more likely to over-report desirable reading behaviors, particularly ones that are more straightforward (in this sample, notably ‘wh-‘questions and responding to the child), while mothers from higher-SES backgrounds may be more critical of their reading behaviors. Comparison between DialogPR scores and direct observation of shared reading in the home would be useful to quantify potential reporting effects, though this may be difficult in practice with a large sample.

This study has several important strengths. The DialogPR was developed and refined referencing an evidence-based conceptual model of dialogic reading, which afforded clarity in item content and organization. Despite a small sample size, it exhibited remarkably strong psychometric properties using advanced analytic modeling techniques. DialogPR scores were highly correlated with a validated parental report measure of reading behaviors in the home (StimQ-P READ). Administration time was brief with no concerns reported in two samples of parents from diverse backgrounds, including low-SES families who are in greatest need of effective assessment and intervention. The DialogPR addresses an important evidence gap and need for an efficient assessment of shared reading quality via caregiver report, with potential research and clinical application in programs advocating dialogic reading such as Reach Out and Read.

Our study also has a number of limitations that should be noted. Our sample population was a combination from 2 smaller studies, which can also be viewed as a strength, and efficient, innovative use of resources. The first exclusively involved 4-year-old girls from low-SES households by design, while the second was diverse in age (3 to 5) and gender from largely higher-SES employee families at an academic medical center. Together, this provided a remarkably diverse sample, showing that DialogPR may be generalizable, though a larger, inherently diverse sample would help affirm this. Girls are marginally over-represented, though it is reasonable to assume that shared reading quality should not be overly dependent on a child’s gender, especially in the preschool age range. For budgetary reasons, we were not able to conduct follow-up visits, and thus unable to assess test-retest reliability. We were unable to compare directly these findings to direct observation, and it is unclear whether reported behaviors reflect a long-term pattern, especially in oft-chaotic home environments. Most importantly, while this study offers a respectable first step, our relatively small sample size is inadequate for definitive validation, and future research with a larger sample will be needed to corroborate the findings here, including test-retest reliability and, ideally, concurrent validity with observational data. Further refinement is needed, perhaps including more explicit means to assess parental responses to the child (i.e. evaluations and expansions) during shared reading, in a parsimonious way. Overall, at this preliminary stage, the DialogPR offers a conceptually- and psychometrically-sound step toward improved, efficient insight into dialogic reading and shared reading quality in the home, important catalysts for healthy cognitive- and social-emotional development.

Conclusion

In this pilot study involving a relatively small (n = 49) yet diverse sample of mothers of preschool-age children, the 8-item DialogPR exhibited strong and promising psychometric properties, including internal consistency, reliability and validity referenced to an external standard of home literacy environment. The DialogPR is founded on a conceptual model of dialogic reading, which is advocated by the AAP and programs such as Reach Out and Read to improve verbal interactivity and engagement through book sharing, and may be an efficient, valid means of assessment warranting further investigation.

Abbreviations

AAP:

American Academy of Pediatrics

MRI:

Magnetic resonance imaging

SES:

Socioeconomic status

References

  1. High PC, Klass P. Literacy promotion: an essential component of primary care pediatric practice. Pediatrics. 2014;134(2):404–9.

    Article  Google Scholar 

  2. Cree A, Kay A, Stewart J. The economic and social cost of illiteracy: a snapshot of illiteracy in a global context. Melbourne, Australia: World Literacy Foundation; 2015.

    Google Scholar 

  3. National Center for Education Statistics, criterion-referenced standard for this Grades 4 and 8. Mathematics and Reading., Institute for Education Sciences, editor. Washington, DC: National Center for Education Statistics; 2015.

    Google Scholar 

  4. The Annie E. Casey Foundation, Double Jeopardy. How third grade Reading skills and poverty influence high school graduation. Baltimore, MD: The Annie E. Casey Foundation; 2012.

    Google Scholar 

  5. Ozernov-Palchik, O., et al., Longitudinal stability of pre-reading skill profiles of kindergarten children: implications for early screening and theories of reading. Dev Sci, 2016.

    Google Scholar 

  6. Pool JL and Johnson ES, Screening for Reading Problems in Preschool and Kindergarten: An Overview of Select Measures. 2016, Boise State University: Boise, ID.

  7. Justice LM, Invernizzi MA, Meier JD. Designing and implementing an early literacy screening protocol: suggestions for the speech-language pathologist. Language, Speech, and Hearing Services in Schools. 2002;33:84–101.

    Article  Google Scholar 

  8. Snow CE, Burns MS, Griffin P. Preventing reading difficulties in young children. Washington, DC: National Academy Press; 1998.

    Google Scholar 

  9. Ramey CT, Ramey SL. Early intervention and early experience. Am Psychol. 1998;53(2):109–20.

    Article  CAS  Google Scholar 

  10. Bus AG, Van Ijzendoorn MH, Pellegrini A. Joint Book Reading Makes for Success in Learning to Read: A meta-analysis on intergenerational transmission of literacy. Rev Educ Res. 1995;65(1):1–21.

    Article  Google Scholar 

  11. Swanson E, et al. A synthesis of read-aloud interventions on early reading outcomes among preschool through third graders at risk for reading difficulties. J Learn Disabil. 2011;44(3):258–75.

    Article  Google Scholar 

  12. Anderson RC, et al. Becoming a nation of readers: the report of the commission on Reading. Washington, DC: National Institute of Education; 1985.

    Google Scholar 

  13. National Early Literacy Panel, Developing early literacy. Report of the National Early Literacy Panel. Washington, DC: National Institute for Literacy; 2008.

    Google Scholar 

  14. Niklas F, Schneider W. Intervention in the home literacy environment and kindergarten children’s vocabulary and phonological awareness. First Language. 2017;37(5).

    Article  Google Scholar 

  15. Zuckerman B, Khandekar A. Reach out and read: evidence based approach to promoting early child development. Curr Opin Pediatr. 2010;22(4):539–44.

    Article  Google Scholar 

  16. Zhai F, Waldfogel J, Brooks-Gunn J. Head start, pre-kindergarten, and academic school readiness: a comparison among regions in the U.S. J Soc Serv Res. 2013;39(3):345–64.

    Article  Google Scholar 

  17. Every Child Succeeds. Every Child Succeeds Home Visitation Program. 2016; Available from: https://www.everychildsucceeds.org/.

  18. Tambyraja SR, et al. Home literacy environment profiles of children with language impairment: associations with caregiver- and child-specific factors. Int J Lang Commun Disord. 2017;52(2):238–49.

    Article  Google Scholar 

  19. Horowitz-Kraus T, Hutton JS. From emergent literacy to reading: how learning to read changes a child's brain. Acta Paediatr. 2015;104(7):648–56.

    Article  Google Scholar 

  20. Roberts J, Jurgens J, Burchinal M. The role of home literacy practices in preschool children's language and emergent literacy skills. J Speech Lang Hear Res. 2005;48(2):345–59.

    Article  Google Scholar 

  21. Needlman R, Silverstein M. Pediatric interventions to support reading aloud: how good is the evidence? J Dev Behav Pediatr. 2004;25(5):352–63.

    Article  Google Scholar 

  22. Al Otaiba S, et al. Home literacy environments of young children with Down syndrome: findings from a web-based survey. Remedial Spec Educ. 2009;30(2):96–107.

    Article  Google Scholar 

  23. Bitetti D, Hammer CS, The Home Literacy Environment and the English Narrative Development of Spanish-English Bilingual Children. J Speech Lang Hear Res. 2016:1–13.

  24. Lever R, Senechal M. Discussing stories: on how a dialogic reading intervention improves kindergartners' oral narrative construction. J Exp Child Psychol. 2011;108(1):1–24.

    Article  Google Scholar 

  25. Whitehurst G, et al. Accelerating language development through picture book Reading. Dev Psychol. 1988;24(4):552–9.

    Article  Google Scholar 

  26. Whitehurst, G. Dialogic Reading: an effective way to read to preschoolers. 2013; Available from: http://www.readingrockets.org/article/400/.

    Google Scholar 

  27. Vally, Z., et al., The impact of dialogic book-sharing training on infant language and attention: a randomized controlled trial in a deprived south African community. J child Psychol psychiatry, 2014.

  28. U.S. Department of Education, Dialogic Reading. In What Works Clearinghouse. Washington, DC: Institute of Education Sciences; 2007.

    Google Scholar 

  29. Mol S, et al. Added value of dialogic parent–child book readings: a meta-analysis. Early Educ Dev. 2008;19(1):7–26.

    Article  Google Scholar 

  30. Bellevue Project for Early Language Literacy and Education Success. STIMQ Home Cognitive Environment. 2017 [cited 2017; Available from: https://med.nyu.edu/pediatrics/developmental/research/belle-project/stimq-cognitive-home-environment. Accessed 6 Oct 2018.

  31. Dreyer BP, Mendelsohn AL, Tamis-LeMonda CS. Assessing the child's cognitive home environment through parental report: reliability and validity. Early Development and Parenting. 1996;5:271–87.

    Article  Google Scholar 

  32. Hutton JS, et al. Shared Reading Quality and Brain Activation during Story Listening in Preschool-Age Children. J Pediatr. 2017;191:204–211.e1.

    Article  Google Scholar 

  33. Vannest J, et al. Factors determining success of awake and asleep magnetic resonance imaging scans in nonsedated children. In: Neuropediatrics; 2014.

    Google Scholar 

  34. Hutton JS, et al. Goldilocks Effect? Illustrated Story Format Seems “Just Right” and Animation “Too Hot” for Integration of Functional Brain Networks in Preschool-Age Children, in Pediatric Academic Societies Annual Meeting. Canada: Toronto; 2018.

    Google Scholar 

  35. Harris P, et al. Research electronic data capture (REDCap) - a metadata-driven methodology and workflow process for providing translational research informatics support. J Biomed Inform. 2009;42(2):377–81.

    Article  Google Scholar 

  36. Smith EV, Smith RM. Rasch measurement : advanced and specialized applications. Maple Grove, MN: JAM Press; 2007.

    Google Scholar 

  37. Rasch G, On general laws and the meaning of measurement in psychology, in Proceedings of the fourth Berkeley symposium on mathematical statistics and probability. 1961: Berkeley, CA.

  38. Smith RM. Detecting item bias with the Rasch model. J Appl Meas. 2004;5(4):430–49.

    PubMed  Google Scholar 

  39. Puglisi ML, et al. The home literacy environment is a correlate, but perhaps not a cause, of variations in Children’s language and literacy development. Sci Stud Read. 2017;21(6):498–514.

    Article  Google Scholar 

  40. Berkule SB, et al. Attitudes about shared reading among at-risk mothers of newborn babies. Ambul Pediatr. 2007;7(1):45–50.

    Article  Google Scholar 

  41. Green CM, et al. Maternal literacy and associations between education and the cognitive home environment in low-income families. Arch Pediatr Adolesc Med. 2009;163(9):832–7.

    Article  Google Scholar 

  42. Read Aloud 15 Minutes. Read Aloud 15 Minutes Website. 2015. Available from: http://www.readaloud.org. Accessed 6 Oct 2018.

  43. Whitehurst G, et al. A picture book reading intervention in day care and home for children from low-income families. Dev Psychol. 1994;30(5):679–89.

    Article  Google Scholar 

  44. Hofferth SL, Response Bias in a Popular Indicator of Reading to Chilldren, in Workshop on the National Children’s Study. 2004, University of Maryland: Crystal City, VA.

Download references

Acknowledgements

The authors would like to thank research coordinators Stacey Woeste, Rachel Gruber, and Amy Kerr for overseeing data collection, data entry, and quality control.

Ethical approval and consent to participate

This study was approved by the Cincinnati Children’s Hospital Institutional Review Board. All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki declaration and its later amendments or comparable ethical standards. Written informed consent was obtained from a custodial parent of all individual participants included in the study, who were all under 6 years old.

Funding

The clinical trial from which subjects for the first group involved in this analysis (n = 22) were sampled was supported by grant ##1R01HD066115-01A1 from the Eunice Kennedy Schriver National Institutes for Child Health and Human Development (Phelan). The study from which subjects for the second group involved in this analysis (n = 27) were enrolled was supported by grant from the Thrasher Research Fund (Hutton). Study design, measure development and data analysis were also supported by an Arnold W. Strauss Fellow Award from the Cincinnati Children’s Research Foundation (Hutton).

Availability of data and materials

The datasets used and/or analyzed during the current study and DialogPR measure are available from the corresponding author on reasonable request.

Author information

Authors and Affiliations

Authors

Contributions

JH researched and drafted the DialogPR screening instrument used in this study, designed the study protocol, supervised data collection, collaborated in data analysis and measure refinement, and drafted the initial manuscript and subsequent revisions. KP is the principal investigator of the clinical trial from which the first group of subjects (n = 22) for this study was sampled, collaborated in protocol design for this study, and reviewed and revised the manuscript. TD provided guidance on the study protocol design, analysis, and measure refinement, and reviewed and revised the manuscript. GH performed all data analysis for this study, and reviewed and revised the manuscript under the supervision of RI. RI guided DialogPR measure assessment and refinement, and reviewed and revised the manuscript. All authors read and approved the final manuscript as submitted.

Corresponding author

Correspondence to John S Hutton.

Ethics declarations

Consent for publication

Not applicable.

Competing interests

The authors declare that they have no competing interests.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Hutton, J.S., Huang, G., Phelan, K.J. et al. Shared reading quality assessment by parental report: preliminary validation of the DialogPR. BMC Pediatr 18, 330 (2018). https://doi.org/10.1186/s12887-018-1298-1

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1186/s12887-018-1298-1

Keywords