Saturday, May 25, 2019

Preschool Assessment Essay

This reputation addresses the many questions the mellowed/Scope Educational inquiry Foundation has received about trial runing four-year-olds. Our reasons for sharing this paper with early puerility practitioners, policy devoters, and p bents is three-fold (a) to provide basic entropy about the terms and issues surrounding legal opinion (b) to add an empirical and pragmatic perspective to what can sometimes be an im headlandioned debate and (c) to affirm our commitment to doing what is best for young children and supporting those who contract the programs and policies that serve them. exalted/Scope believes child sound judgement is a vital and necessary component of all postgraduate quality early childhood programs. estimate is important to understand and support young childrens development. It is also essential to enter and evaluate how effectively programs argon meeting their educational subscribe tos, in the broadest ace of this term. For assessment to occur, it must be feasible. That is, it must meet reasonable criteria regarding its efficiency, cost, and so on.If assessment places an un callable burden on programs or evaluators, it will non be under playscriptn at all and the lack of data will hurt all concerned. In addition to feasibility, however, assessment must also meet the demands of ecological boldness. The assessment must addresses the criteria outlined below for stateing us about what children in real programs atomic number 18 learning and doing each day. Efficiency and ecological grimness are not mutually exclusive, but must sometimes be balanced against one an opposite.Our challenge is to find the best balance under the conditions given and, when necessary, to hammer toward altering those conditions. Practically speaking, this means we must continue to serve children using research-based practices, fulfill mandates to secure program resources, and improve assessment procedures to better realize our ideal. This paper sets fort h the criteria to be considered in striving to make early childhood assessment adhere to these highest standards. Background The concern with assessment in the early childhood field is not saucily.Decades of debate are summarized in the guinea pig Association for the Education of preteen tiddlerren (NAEYC) publication Reaching Potentials Appropriate Curriculum and Assessment for Young tikeren (Bredekamp & Rosegrant, 1992). This position statement has just been expanded in a new document titled Early Childhood Curriculum, Assessment, and Program evaluation Building an Effective, responsible System in Programs for Children Birth through Age 8 (www. naeyc. org/resources/position_statements/pscape. asp).1What is new in this ongoing debate is the heightened attendance to examen young children as a means of holding programs accountable for their learning. Assessment in the Classroom (Airasian, 2002) offers the following definitions Assessment is the process of collecting, synthe sizing, and interpreting breeding to aid differentiateroom decision-making. It includes nurture gathered about pupils, tuition, and schoolroom climate. Testing is a semi-formal, organisationatic procedure for gathering a sample of pupils behavior. The results of a test are utilise to make generalizations about how pupils would have performed in similar but untested behaviors.Testing is one form of assessment. It usually involves a series of direct requests to children to perform, within a set period of time, specific tasks designed and conducted by vaingloriouss, with predetermined correct answers. By contrast, alternative forms of assessment may be completed either by adults or children, are more open-ended, and often look at performance over an extended period of time. Examples include objective honorings, portfolio analyses of individual and collaborative work, and instructor and parent ratings of childrens behavior.The current testing initiative focuses primarily o n literacy and to a lesser extent numeracy. The rationale for this initiative, advanced in the No Child Left Behind Act and supported by the report of the National Reading Panel (2000), is that young children should acquire a prescribed body of fellowship and schoolman skills to be ready for aim. Social domains of school readiness, while also touted as essential in a series of National Research Council reports (notably keen to pasturern, 2000a and Neurons to Neighborhoods, 2000b), are admittedly neither as widely mandated nor as testable as their academic counterparts.Hence, whether justified or not, they do not figure as conspicuously in the testing and accountability debate. This information paper responds to questions be asked of early childhood leaders about the use and misuse of testing for preschoolers 3 to 5 years old. This response is not merely a reactive gesture nor an attempt to advance and defend a specific position. Rather, the paper is intended as a resource to p rovide information about when and how preschool assessment in general, and testing and other forms of assessment in particular, can be impoundly apply to inform policy decisions about early childhood programming.As a framework for providing this information, High/Scope accepts two realities. First, testing is, will be, and in feature always has been, used to answer questions about the effectiveness of early childhood interventions. Since early childhood programs attempt to increase childrens familiarity and skills in specific content areas, evaluators have traditionally used testing, along with other assessment strategies, to determine whether these educational objectives have been achieved. Second, program accountability is essential, and testing is one cost-efficient means of measuring it.Numerous research studies show that high quality programs can enhance the academic and conducttime achievement of children at risk of school failure. This conclusion has 2 resulted in an i nfusion of public and private dollars in early education. It is reasonable to ask whether this investment is achieving its goal. Testing can play a role in answering this accountability question. With this reality as a background, this information paper proceeds to address two questions.First, given the current distributive use of testing and itsprobably expansion, when and under what conditions can this type of assessment be used appropriately with preschool-age children? That is, what characteristics of tests and their administration will guarantee that we do no harm to children and that we do care adults acquire valid information? Second, given that even the most well-designed tests can provide unless limited data, how can we increase the use of non-test assessments so they too add valuable information over and above that accomplished through like testing procedures?General Issues in Assessment Uses of Child AssessmentAssessment can provide four types of information for and about children, and their parents, teachers, and programs. Child assessment can 1. Identify children who may be in need of specialized services.Screening children to determine whether they would benefit from specific interventions is appropriate when parents, teachers, or other pros suspect a problem. In these cases, assessments in several related domains are then usually administered to the child. In addition, data from parents and other adults involved with the child are considered in determining a diagnosis and course of treatment.2. Plan instruction for individuals and groups of children. Assessment data can be used by teachers to support the development of individual children, as well as to plan instructional activities for the class as a whole. In addition, information on developmental progress can and should be shared with parents to help them understand what and how their children are learning in the classroom and how they can extend this learning at home. 3. Identify prog ram improvement and staff development of necessity. Child assessments can provide formative military rank data that benefit program and staff development.Findings can point to areas of the curriculum that need further articulation or resources, or areas where staff need professional development. If children in the classroom as a whole are not making progress in certain developmental domains, it is possible that the curriculum needs revision or that teachers need some additional training. In conducting formative evaluations, child data are best combined with program data that m overall quality, fidelity to curriculum implementation standards, and specific teaching practices. 4. Evaluate how well a program is meeting goals for children.It is this fourth purpose, sometimes called outcome or summative evaluation, that is the primary focus of this paper. 3 Note that it is the program, not the child, who should be held accountable. Although data may be collected on individual children, data should be aggregated to determine whether the program is achieving its desired outcomes. These outcomes may be defined by the program itself and/or by national, state, or district standards. How the outcomes are thrifty is determined by the inextricable link between curriculum and assessment.Ideally, if a curriculum has clear learning objectives, those will drive the form and content of the whole steps. Conversely, thoughtful design of an appropriate assessment tool can encourage program developers to consider what and how adults should be teaching young children. Reliability and Validity Any formal assessment tool or method should meet established criteria for validity and reliability (American Educational Research Association, American psychological Association, & National Council of Measurement in Education, 1999).Reliability is defined as how well various steps of something agree with each other, for lesson, whether a group of similar test items or two observers compl eting the uniform items have similar results. Validity has several dimensions. Content or face validity refers to how well an instrument appraises what it claims to measure ecological validity refers to the authenticity of the measurement context and construct validity deals with the measures conceptual integrity. In assessing young children, two aspects of validity have special importancedevelopmental validity and predictive validity.Developmental validity means that the performance items being measured are developmentally suitable for the children being assessed. Predictive validity means the measure can predict childrens subsequently school success or failure, as defined by achievement test scores or academic placements (on-grade, retained in grade, or placed in special education) during the elementary grades. Over the longer term, predictive validity can even refer to such(prenominal) outcomes as adult literacy, employment, or avoiding criminal activity.In Principles and Reco mmendations for Early Childhood Assessments, the National Education Goals Panel (1998) noted that the younger the child, the more difficult it is to entertain reliable and valid assessment data. It is particularly difficult to assess childrens cognitive abilities blamelessly before age 6 (p. 5). Meisels (2003) claims research lay outs that no more than 25 per cent of early academic or cognitive performance is predicted from information obtained from preschool or kindergarten tests (p. 29).Growth in the early years is rapid, episodic, and highly influenced by environmental supports. Performance is influenced by childrens emotional and motivational states, and by the assessment conditions themselves. Because these individual and situational factors affect reliability and validity, the Panel recommended that assessment of young children be pursued with the necessary safeguards and caveats about the accuracy of the decisions that can be drawn from the results. These procedures and ca utions are explored below. 4 Testing.Appropriate Uses of Testing Standardized tests are used to obtain information on whether a program is achieving its desired outcomes. They are considered objective, time- and cost-efficient, and suitable for making quantitative comparisons. Testing can provide valid data when used appropriately and matched to developmental levels. Moreover, tests can act as teaching tools by providing a window into what children already know and where they need more time, practice, and/or help to improve. Creating a valid assessment for young children is a difficult task.It must be meaningful and authentic, evaluate a valid sample of information learned, be based on performance standards that are genuine benchmarks, avoid arbitrary cut-off scores or norms, and have authentic scoring. The context for the test should be rich, realistic, and lure (Wiggins, 1992). It is therefore incumbent upon the creators of assessment tools to design instruments thatunlike artifi cial drills resemble natural performance. If these conditions are met, young children are more likely to recognize what is being asked of them, thus increasing the reliability and validity of the results.Criteria of Reliable and Valid Preschool Tests Both the content and administration of tests must respect young childrens developmental characteristics. differently the resulting data will be neither reliable nor valid. Worse, the testing experience may be negative for the child and perhaps the tester as well. Further, the knowledge and skills measured in the testing situation must be transferable and applicable in real-world settings. Otherwise the information gathered has no practical value. To throw meaningful data and minimize the risk of creating a harmful situation, tests for preschool-age children should satisfy the following criteria1. Tests should not make children tactile sensation anxious or scared. They should not menace their selfesteem or make them feel they have fa iled. Tests should acknowledge what children knowor have the potential to learnrather than penalizing them for what they do not know. 2. Testing should take place in, or simulate, the natural environment of the classroom. It should avoid placing the child in an artificial situation. Otherwise, the test may measure the childs response to the test setting rather than the childs ability to perform on the test content. 3.Tests should measure real knowledge in the context of real activities. In other words, the test activities as well as the test setting should not be contrived. They should resemble childrens ordinary activities as closely as possible, for example, discussing a book as the adult reads it. Furthermore, tests should measure broad concepts rather than narrow skills, for example, alphabetic and letter knowledge sampled from this domain rather than old(prenominal)ity with specific garner chosen by the adult. 5 4. The tester should be someone familiar to the child.Ideally, t he person administering the test would be a teacher or another adult who interacts regularly with the child. When an foreign researcher or evaluator must administer the test, it is best if the individual(s) spend time in the classroom beforehand, becoming a familiar and friendly figure to the children. If this is not feasible, the appearance and demeanor of the tester(s) should be as similar as possible to adults with whom the child regularly comes in contact. 5. To the extent possible, testing should be conducted as a natural part of daily activities rather than as a time-added or pullout activity.Meeting this criterion helps to satisfy the sooner standards of a familiar place and tester, especially if the test can be administered in the context of a normal part of the daily routine (for example, assessing book knowledge during a regular reading period). In addition, testing that is integrated into standard routines avoids placing an additional burden on teachers or detracting fr om childrens instructional time. 6. The information should be obtained over time. A single encounter, especially if brief, can produce inaccurate or distorted data.For example, a child may be ill, hungry, or distracted at the moment of testing. The test is then measuring the childs interest or willingness to respond rather than the childs knowledge or ability with respect to the question(s) being asked. If timedistributed measurements are not feasible, then testers should note unusual circumstances in the situation (e. g. , noise) or child (e. g. , fatigue) that could contribute single-encounter results invalid and should either schedule a re-assessment or discount the results in such cases.7. When repeated instances of data gathering are not feasible (e.g. , due to time or budgetary constraints), an attempt should be made to obtain information on the same content area from multiple and diverse sources. only if as young children have different styles of learning, so they will diff erentially demonstrate their knowledge and skills under varying modes of assessment. For example, a complete and accurate measure of letter knowledge may involve tests that employ both generative and recognition strategies. 8. The length of the test should be sensitive to young childrens interests and attention spans.If a test is conducted during a regular program activity (e. g. , small-group time), the test should last no longer than is typical for that activity. If it is necessary to conduct testing outside regular activities, the assessment period should last 1020 minutes. Further, testers should be sensitive to childrens comfort and engagement levels, and take a break or continue the test at another day and time if the child cannot or does not want to proceed. 9. Testing for purposes of program accountability should employ appropriate sampling methods whenever feasible.Testing a representative sample of the children who participate in a program avoids the need to test every chi ld and/or to administer all tests to any one child. Sampling strategies reduce the overall time spent in testing, and minimize the chances for placing undue stress on individual children or burdening individual teachers and classrooms. 6 alternating(a) Child Assessment Methods Alternative forms of assessment may be used by those who have reservations about, or want to supplement, standardise tests. These other methods often fall under the banner of authentic assessments.They engage children in tasks that are personally meaningful, take place in real life contexts, and are grounded in naturally occurring instructional activities. They offer multiple ways of evaluating students learning, as well as their motivation, achievement, and attitudes. This type of assessment is consistent with the goals, curriculum, and instructional practices of the classroom or program with which it is associated (McLaughlin & Vogt, 1997 Paris & Ayres, 1994). Authentic assessments do not rely on unrealist ic or arbitrary time constraints, nor do they emphasize flashing recall or depend on lucky guesses.Progress toward mastery is the backbone, and content is mastered as a means, not as an end (Wiggins, 1989). To document accomplishments, assessments must be designed to be longitudinal, to sample the baseline, the increment, and the preserved levels of change that follow from instruction ( wildcat well, Bixby, Glenn & Gardener, 1991). Alternative assessment can be more overpriced than testing. Like their counterparts in testing, authentic measures must meet psychometric standards of demonstrated reliability and validity.Their use, especially on a widespread scale, requires adequate resources. Assessors must be trained to bankable levels of reliability. Data collection, coding, entry, and analysis are also time- and cost-intensive. This investment can be seen as reasonable and necessary, however, if the goal is to produce valid information. Alternative child assessment procedures th at can meet the criteria of reliability and validity include observations, portfolios, and ratings of children by teachers and parents. These are described below. ObservationsIn assessing young children, the principal alternative to testing is systematic observation of childrens activities in their day-to-day settings. Observation fits an interactive style of curriculum, in which give-and-take between teacher and child is the norm. Although careful observation requires effort, the approach has high ecological validity and intrudes minimally into what children are doing. Childrens activities naturally integrate all dimensions of their developmentintellectual, motivational, social, physical, aesthetic, and so on.Anecdotal notes alone, however, are not sufficient for good assessment. They do not offer criteria against which to judge the developmental value of childrens activities or provide evidence of reliability and validity. Instead, anecdotal notes should be used to complete develo pmental scales of proven reliability and validity. Such an approach permits children to engage in activities any time and anywhere that teachers can see them. It defines categories of acceptable answers rather than single right answers. It expects the teacher to set the framework for children to initiate their own activities.It embraces a broad definition of child development that includes not only language and mathematics, but also initiative, social relations, physical skills, and the arts. It is culturally sensitive when teachers are trained observers who focus on objective, culturally neutral descriptions of behavior (for example, ditch hit Bob) rather than subjective, culturally loaded 7 interpretations (for example, Pat was very angry with Bob). Finally, it empowers teachers by recognizing their judgment as essential to accurate assessment.PortfoliosOne of the most fitting ways to undertake authentic, meaningful evaluation is through the use of a well-constructed portfolio sy stem. Arter and Spandel (1991) define a portfolio as a businesslike collection of student work that tells the story of the students efforts, progress, or achievement in (a) given area(s). This collection must include student lodge in selection of portfolio content, the guidelines for selection, the criteria for judging merit, and evidence of student self-reflection (p. 36). Portfolios describe both a place (the physical space where they are stored) and a process.The process provides richer information than standardized tests, involves multiple sources and methods of data collection, and occurs over a representative period of time (Shaklee, Barbour, Ambrose, & Hansford, 1997). Portfolios have additional value. They encourage two- and three-way collaboration between students, teachers, and parents advertize ownership and motivation integrate assessment with instruction and learning and establish a quantitative and qualitative record of progress over time (Paris & Ayres, 1994 Paulso n, Paulson, & Meyer, 1991 Wolf & Siu-Runyan, 1996 Valencia, 1990).Portfolios encourage teachers and students to focus on important student outcomes, provide parents and the community with credible evidence of student achievement, and inform policy and practice at every level of the educational system (Herman & Winters, 1994, p. 48). The purposes for which portfolios are used are as variable as the programs that use them (Graves & Sunstein, 1993 Valencia, 1990 Wolf & Siu-Runyan, 1996). In some programs, they are apparently a place to store best work that has been graded in a traditional manner.In others, they are used to create longitudinal systems to demonstrate the process leading to the products and to design evaluative rubrics for program accountability. There are also programs that merely have students collect work that is important to them as a personal, non-evaluative record of their achievements. When portfolios are not used to judge ability in some agreed-upon fashion, they are usually not highly structured and may not even include reflective pieces that demonstrate student growth and understanding.Portfolios are most commonly thought of as alternative assessments inelementary and second-string schools. Yet they have long been used in preschools to document and share childrens progress with parents, administrators, and others. For portfolios to be used for program accountability, as well as student learning and reflection, the evaluated outcomes must be aligned with curriculum and instruction. Children must have some choice about what to include in order to feel ownership and pride. Portfolios should document the creative or problem-solving process as they display the product, encouraging children to reflect on their actions.Conversations with children about their portfolios engages them in the evaluation process and escalates their desire to demonstrate their 8 increasing knowledge and skills. Sharing portfolios with parents can help teachers connec t school activities to the home and involve parents in their childrens education. teacher Ratings Teacher ratings are a way to organize teacher perceptions of childrens development into scales for which reliability and validity can be assessed. Childrens grades on report cards are the most common type of teacher rating system.When completed objectively, report-card grades are tied to students performance on indicators with delineated scoring criteria, such as examinations or projects evaluated according to explicitly defined criteria. In these ways, teacher ratings can be specifically related to other types of child assessments including scores on standardized tests or other validated assessment tools, concrete and specific behavioral descriptions (e. g. , frequency of participation in group activities, ability to recognize the letters in ones name), or global assessments of childrens traits (e. g. , cooperative, sociable, hard-working).Research shows that teacher ratings can have considerable short- and long-term predictive validity throughout later school years and even into adulthood (Schweinhart, Barnes, & Weikart, 1993). Parent Ratings Parent ratings are a way to organize parent perceptions of childrens development into scales for which reliability and validity can be assessed. Soliciting parent ratings is an excellent way for teachers to involve them as partners in the assessment of their childrens performance. The very process of completing scales can inform parents about the kinds of behaviors and milestones that are important in young childrens development.It also encourages parents to observe and listen to their children as they gather the data needed to rate their performance. An example of the use of parent ratings is the Head Start Family and Child Experiences Survey (FACES) study, in which parents ratings of their childrens abilities and progress were related to measures of classroom quality and child outcomes (Zill, Connell, McKey, OBrien et al . , 2001). inference Recent years have seen a growing public interest in early childhood education. Along with that support has come the use of high stakes assessment to justify the expense and apportion the dollars.With so much at stakethe future of our nations childrenit is imperative that we proceed correctly. to a higher place all, we must guarantee that assessment reflects our highest educational goals for young children and neither restricts nor distorts the substance of their early learning. This paper sets forth the criteria for a comprehensive and balanced assessment system that meets the need for accountability while respecting the welfare and development of young children. Such a system can include testing, provided it measures applicable knowledge and skills in a safe and child-affirming situation.It can also include alternative assessments, provided they too meet psychometric standards of reliability and validity. Developing and implementing a balanced approach to ass essment is not an easy or inexpensive undertaking. But because we value our children and respect those charged with their care, it is an investment worth making. 9 References Airasian, P. (2 002). A ssessment in the classroom. New Y ork Mc Graw-H ill. American Educational Research Association, American Psychological Association, & National Council of Measu rement in E ducation. (1 999). S tanda rds for edu cationa l and psy cholog ical testing.W ashington, DC American Psychological Association. Arter, J. A. , & Spande l, V. (199 2). exploitation p ortfolios of stud ent work in instru ction and a ssessment. E ducational Measurement Issues and Practice, 3644. Brede kamp, S. , & Rosegra nt, T. (Ed s. ) (1992 ). R eaching Potentials Appropriate Curriculum and Assessment for Young Children . Washington, DC National Association for the Education of Young Children. Graves, D . H. , & Sun stein, B. S. (19 92). P ortfolio p ortraits . New Hampshire Heinemann. Herma n, J. L. , & W inters, L. (199 4). Portfo lio research A slim collection . E duca tional Lea dership , 5 2 (2), 4855.McLa ughlin, M. , & Vogt, M . (1997) . P ortfolios in teacher education . Newark, Delaware International Reading Association. Meisels, S. (2003, 19 March). Can Head Start pass the test? E ducation Week , 2 2 (27), 44 & 29. National A ssociation for the Educa tion of Yo ung Childre n and Na tional Assoc iation of Ear ly Childhoo d Specia lists in State Dep artments of E ducation (2 003, N ovemb er. E arly Childhood Curriculum, Assessment, and Program Evaluation Building an Effective, Accountable System in Programs for Children Birth Through Age 8 . ) Washin gton, DC Authors. Av ailable online at www. naeyc.org/resources/position_statements/pscape. asp. N ational E ducation G oals Pane l. (1998). P rinciples and rec ommen dations for early childh ood assessm ents. Washington, DC Author. National R eading P anel. (200 0). T eaching children to read An evidence-based assessment of the scientific research literature on reading and its implications for reading instruction. Washin gton, DC National In stitute of Child Health and Human Developm ent, National Institutes of Health. National R esearch C ouncil. (20 00a). E ager to learn Educating our preschoo lers. W ashington, DC National Academy P ress. National R esearch C ouncil. (20 00b).N eurons to neighborhoods The science of early childhood development. Washington, D C National Acad emy Press. Paris, S. G . , & Ayers, L. R . (1994) . B ecom ing reflective s tudents a nd teach ers with po rtfolios and authen tic assessment. Washington DC American Psychological Association. Paulson, F. L. , Paulson, P. R. , & Meyer, C. A. (1991). What makes a portfolio a portfolio? E duca tional Lea dership , 48 (5), 6063. Schweinha rt, L. J. , Barne s, H. V. , & Weika rt, D. P. (19 93). S ignificant benefits The High/Scope Perry Preschool study through age 27 .Ypsilanti, MI High/Sco pe Press. Shaklee, B . D. , Barb our, N. E ., Ambros e, R. , & H ansford, S. J . (1997) . D esigning and using portfolios. Boston Allyn & Bacon. Valencia , S. W. (1 990). A portfolio ap proach to classroom reading asse ssment Th e whys, whats an d hows. T he Reading Teacher , 4 3 (4), 338340. Wiggins, G . (1992) . Creating tests wo rth taking. E duca tional Lea dership , 4 9 (8), 2633. Wolf, D. , Bixby, J. , Glenn, J. , & Gardner, H. (1991).To use their minds well Investigating new forms of student assessment. In G. Gran t (Ed. ), R eview of research in education, V ol 17 ( pp. 3174). Washington D. C. American Educational Research Association. Wolf, K . , & Siu-Run yan, Y.(19 96). Po rtfolio purpo ses and po ssibilities. J ournal of Adolescent and Adult Literacy, 40 (1), 3037. Zill, N. , Conn ell, D. , Mc Key, R. H . , OBrien, R . et al. (2001 , January). H ead Start FACESlongitudinal Findings on Pro gram P erforma nce, Third Progres s Report. W ashington, DC Administration on Children, Youth and Families, U. S. Depa rtment of Health a nd H uman Services. 10 High/Scope Assessment Resources High/Scope has developed and validated three preschool assessment instruments. Two are for children, one focusing specifically on literacy and the other more broadly on multiple domains of development.The third measure is used to assess and improve the quality of all aspects of early childhood programs. These alternative assessments are described below. Early Literacy Assessment In the Fall of 2004, High/Scope will release the Early Literacy Assessment (ELA), which will evaluate the four key principles of early literacy documented in the Early Reading First Grants and the No Child Left Behind legislation phonological awareness, alphabetic principle, comprehension, and concepts about print. Evaluation will take place in a meaningful context that is familiar to children.

No comments:

Post a Comment

Note: Only a member of this blog may post a comment.