Assessing the Kindergarten Readiness of Dual Language Learners
This is the fourth post in a five-part blog series,听DLL Data Gaps.听Click here听to learn more about this project and access the other blogs in the series.
Inadequate data on young dual language learners (DLLs) hampers efforts to support these children in early care and education (ECE). This blog series has explored gaps that exist in tracking the enrollment of DLLs and rating the quality of services for these learners. Finally, as this post will address, there is also a lack of meaningful assessment data to validly capture the full range of DLLs鈥 development in ECE.
Age-appropriate testing of students鈥 proficiencies can serve many in ECE, including formative assessment for instruction, screening for special needs, or program-wide research or evaluations. State policy leaders are increasingly focused on student outcomes through more standardized assessment data, collected and aggregated at the systems level, to inform decision-making and the allocation of ECE resources.听听
Kindergarten readiness assessments (KRAs), in particular, have gained traction as a strategy to provide educators, families, and district and state leaders with more standardized data on the status of children’s abilities when they enter kindergarten鈥攁 鈥渟napshot on development,鈥 to the BUILD Initiative. KRAs are intended to both support instruction in the early elementary years and provide information that can help policy leaders support school readiness, not to prevent children from enrolling in school.
More than 40 states are currently developing or implementing KRAs, up from just in 2010. The Obama administration promoted states鈥 adoption of KRAs through the federal Race to the Top 鈥 Early Learning Challenge, a discretionary grant program launched in 2011. In part, the grant encouraged states to measure children鈥檚 outcomes in a range of developmental domains in tandem with 鈥渋mplementing comprehensive data systems and using data to improve instruction, practices, services, and policies.鈥 At least have now mandated KRA use by state law.
States are rolling out their KRA systems in a of ways. Some use a commercial assessment tool, such as Teaching Strategies GOLD庐, while others created their own state tool or participated in one of three interstate consortia supported by federal grants. The tools direct assessment (requiring a direct interaction between the test administrator and the child), observation of the child in authentic classroom activities, or a combination of these two approaches.
KRAs have weathered a fair amount of concerns and in their roll-out. Some teachers experienced the testing as an extra . Researchers cautioned against drawing inferences from a one-time assessment of young children when development is in great flux. As the National Education Goals Panel , 鈥渢he younger the child, the more difficult it is to obtain reliable and valid assessment data. It is particularly difficult to assess children’s cognitive abilities accurately before age 6.鈥 Due to questions of validity and reliability, policy , including my colleagues at 麻豆果冻传媒, have also stressed that KRAs should not be used punitively as an accountability measure for ECE providers.
These broader concerns over KRAs have implications for all children鈥攊ncluding DLLs. But states also need to specifically consider how to incorporate DLLs in KRAs as a special population. The development of bilingual children looks fundamentally different than their monolingual peers given that DLLs鈥 knowledge and skills are spread across two languages. For example, research suggests that DLLs have smaller vocabularies in English and their home language when taken separately, but their total vocabulary sizes鈥攖he sum of what they know in both languages鈥攁re to monolingual peers. As such, it is critical that young DLLs have an opportunity to show what they know and can do in their home language.
And yet, at present, almost all state KRAs assess children only in English. 鈥淚f we want equitable assessments for DLLs, we have to assess across both languages,鈥 said DLL expert Linda Espinosa, who works with states and districts across the country and served on the National Technical Advisory for KRAs. She said nearly all states are failing 鈥渢o take on the challenge of creating truly equivalent [bilingual] forms of these test items,鈥 a complicated, costly in the context of tight state budgets. 听to the Migration Policy Institute, New Jersey, Oregon, and have developed KRAs entirely in Spanish while Illinois and state allow DLLs to be assessed in their home language for some test items. Bilingual KRA testing also goes hand-in-hand with a need for more bilingual, bicultural assessors, which requires additional investments.
Moreover, while most states have not even attempted to do so, even those pursuing bilingual KRA assessments tend to use direct translation of the English version, Espinosa said. This method can produce tests that are psychometrically (e.g. not normed or validated on DLLs) and/or irrelevant in content for linguistically and culturally diverse children (e.g. asking a DLL in Alaska to recognize a beach umbrella on a vocabulary test).听For these reasons, 鈥榯ruly鈥 bilingual KRA testing for DLLs will likely remain a long-term challenge.听
Still, KRA implementation needs major improvements even in English-only testing contexts to collect higher-quality DLL data. The U.S. Department of Education highlighted this reality in a 2016 case study of four states鈥 initial implementation of KRAs. Teachers that they did not fully understand guidelines for assessing DLLs, and a majority were unsure about testing procedures for DLLs. For example, some states allow certain testing for DLLs, such as accepting correct answers in non-verbal forms like pointing or gesturing. Teachers also voiced a desire for greater support, such as more explicit training on administering KRAs with DLLs and on-site assistance from bilingual staff.听听
In addition to serving DLLs more equitably at the stage of assessment, state leaders should consider how to share and disseminate data on DLLs鈥 KRA results. Decision-making around how to publish KRA results often reflects the different ways states view KRAs and their purpose. 鈥淲hether to publicly report [KRA] data and what should be included…is often a complicated discussion, involving many stakeholders,鈥 ECE researcher GG Weisenfeld in a 2017 report.
Some states, like and , publish KRA data publicly on state websites. state also decided to use the KRA data on its state report card, but 鈥渞ecognized the paradox of reporting formative assessment data in a summative presentation, and landed on multiple ways to accurately portray the data…opt[ing] not to use a single, composite 鈥榬eadiness鈥 score.鈥 Others states have resisted aggregating or reporting out results. , for example, emphasizes that its voluntary KRA is a formative tool to inform instruction and professional development鈥攏ot to publicize achievement gaps and trends. Michigan similarly states that it finds such summative use 鈥.鈥
For DLLs in states that do decide to publicly report, state leaders must also decide whether to disaggregate the data results by DLL status for public users. In the K-12 context, federal law requires the disaggregation of academic data by English learner (EL) status in grades 3-8, as well as by race and ethnicity, family income, and disability status. stress the importance of these mandates for ensuring that disadvantaged students do not get ignored or masked in data systems. At the same time, for ELs at lower proficiencies, language barriers will definitionally interfere with their academic performance and thereby drive down the subgroups鈥 results.
Similarly, KRA systems leaders can use subgroups to expose achievement gaps in ECE, but they also must consider if data is reliable and valid enough to do so. As 麻豆果冻传媒 stressed in a 2017 report on K-12 EL data, below a certain threshold of English proficiency, it is impossible to make valid claims about academic proficiencies in English. In the case of English-only KRAs, DLLs鈥 scores in literacy and math development may reflect English proficiencies rather than true knowledge of concepts and skills. If states publicly report KRA data without this context, and rely on 鈥渘ative English speakers as the norm against which all students are compared, the unique characteristics of DLLs are likely to be misinterpreted, or worse, determined to be delays,鈥 according to DLL expert Espinosa. Again, this is why testing young DLLs bilingually is so critical鈥攖o capture an accurate, complete picture of their development.
As shown above, Maryland disaggregates by DLL/EL status in its of KRA data. The gaps between EL and English proficient kindergarteners are significantly greater in literacy and math than in physical and social domains (which are less language-dependent), perhaps indicating the extent to which language interferes with DLL data鈥檚 validity. As the report notes, “Because the KRA is not given in the student鈥檚 home language, the knowledge and skills of ELs may not be fully captured.鈥
Recommendations for State Leaders:
- Assess DLLs/ELs in their home language on kindergarten readiness assessments (KRAs).听听
听 –听 Invest in the development of valid bilingual assessment tools in home languages.
听 –听 Invest in expanding access to bilingual assessors. - Improve and increase professional development and guidance for teachers on administering KRAs with DLLs/ELs.
- If publicly reporting data by DLL/EL status for KRAs, provide guidance and explain limitations of these data to users.