Reliability and Comparability of TOEFL iBT Scores
Transcription
Reliability and Comparability of TOEFL iBT Scores
Insight TOEFL iBT Research TM Series I, Volume 3 Reliability and Comparability of TOEFL iBT™ Scores ETS — Listening. Learning. Leading.® Insight TOEFL iBT Research • Series 1, Volume 3 TM Foreword Preface We are very excited to announce the TOEFL iBT™ Research Insight Series, a bimonthly publication to make important research on the TOEFL iBT available to all test score users in a user-friendly format. Since the 1970’s, the TOEFL test has had a rigorous, productive and far-ranging research program. But why should test score users care about the research base for a test? In short, because it is only through a rigorous program of research that a testing company can demonstrate its forward-looking vision and substantiate claims about what test takers know or can do based on their test scores. This is why ETS has made the establishment of a strong research base a consistent feature of the evolution of the TOEFL test. The TOEFL iBT test is the most widely accepted English language assessment, used for admissions purposes in more than 130 countries including the United Kingdom, Canada, Australia, New Zealand and the United States. Since its initial launch in 1964, the TOEFL test has undergone several major revisions motivated by advances in theories of language ability and changes in English teaching practices. The most recent revision, the TOEFL iBT test, was launched in 2005. It contains a number of innovative design features, including the use of integrated tasks that engage multiple language skills to simulate language use in academic settings, and the use of test materials that reflect the reading and listening demands of real-world academic environments. At ETS we understand that you use TOEFL iBT test scores to help make important decisions about your students, and we would like to keep you up-to-date about the research results that assure the quality of these scores. Through the TOEFL iBT Research Insight Series we wish to both communicate to the institutions and English teachers who use the TOEFL iBT test scores the strong research and development base that underlies the TOEFL iBT test, and demonstrate our strong, continued commitment to research. We hope you will find this series relevant, informative and useful. We welcome your comments and suggestions about how to make it a better resource for you. Ida Lawrence Senior Vice President Research & Development Division Educational Testing Service The TOEFL test is developed and supported by a worldclass team of test developers, educational measurement specialists, statisticians and researchers. Our test developers have advanced degrees in such fields as English, language education and linguistics. They also possess extensive international experience, having taught English in Africa, Asia, Europe, North America and South America. Our research, measurement and statistics team includes some of the world’s most distinguished scientists and internationally recognized leaders in diverse areas such as test validity, language learning and testing, and educational measurement and statistics. To date, more than 150 peer-reviewed TOEFL research reports, technical reports and monographs have been published by ETS, many of which have also appeared in academic journals and book volumes. In addition to the 20-30 TOEFL-related research projects conducted by ETS Research & Development staff each year, the TOEFL Committee of Examiners (COE), comprised of language learning and testing experts from the academic community, funds an annual program of TOEFL research by external researchers from all over the world, including preeminent researchers from Australia, the UK, the US, Canada and Japan. In Series One of the TOEFL iBT Research Insight Series, we provide a comprehensive account of the essential concepts, procedures and research results that assure the quality of scores on the TOEFL iBT test. The six issues in this Series will cover the following topics: oooooooo 1 Reliability and Comparability of TOEFL iBT™ Scores Issue 1: TOEFL iBT Test Framework and Development The TOEFL iBT test is described along with the processes used to develop test questions and forms. These processes include rigorous review of test materials, with special attention to fairness concerns. Item pretesting, try outs and scoring procedures are also detailed. Issue 2: TOEFL Research The TOEFL Program has supported rigorous research to maintain and improve test quality. Over 150 reports and monographs are catalogued on the TOEFL website. A brief overview of some recent research on fairness and automated scoring is presented here. Issue 3: Reliability and Comparability of Test Scores Given that hundreds of thousands of test takers take the TOEFL iBT test each year, many different test forms are developed and administered. Procedures to achieve score comparability on different forms are described in this section. Issue 4: Validity Evidence Supporting Test Score Interpretation and Use The many types of evidence supporting the proposed interpretation and use of test scores as a measure of English-language proficiency in academic contexts are discussed. Issue 5: Information for Score Users, Teachers and Learners Materials and guidelines are available to aid in the interpretation and appropriate use of test scores, as well as resources for teachers and learners that support English-language instruction and test preparation. Insight TOEFL iBT Research • Series 1, Volume 3 TM Issue 6: TOEFL Program History A brief overview of the history and governance of the TOEFL Program is presented. The evolution of the TOEFL test constructs and contents from 1964 to the present is summarized. Future series will feature summaries of recent studies on topics of interest to our score users, such as “what TOEFL iBT test scores tell us about how examinees perform in academic settings,” and “how score users perceive and use TOEFL iBT test scores.” The close collaboration with TOEFL iBT score users, English language learning and teaching experts and university professors in the redesign of the TOEFL iBT test has contributed to its great success. Therefore, through this publication, we hope to foster an ever stronger connection with our score users by sharing the rigorous measurement and research base and solid test development that continues to ensure the quality of TOEFL iBT scores to meet the needs of score users. Xiaoming Xi Senior Research Scientist Research & Development Division Educational Testing Service Contributors The primary authors of this section are Mary Enright and Eileen Tyson. The following individuals also contributed to this section by providing their careful review as well as editorial suggestions (in alphabetical order). Cris Breining Rosalie Szabo Xiaofei Tang Xiaoming Xi Reliability and Comparability of TOEFL iBT™ ‑ Scores ETS has always been committed to the quality of its test scores. As an ETS assessment program, TOEFL® strives to ensure score reliability and comparability through strict adherence to guidelines and practices established for the development and operational implementation of its products and services. Evidence of score reliability and comparability is important because it suggests that test scores will have the same meaning across test forms. ETS strives to ensure that the test scores of the TOEFL Internet based test (iBT) are reliable and comparable by: • implementing and adhering to standardized administration and test security procedures • using detailed test specifications to guide test development • monitoring score reliability and generalizability • employing an appropriate scale for reporting scores • using equating and other means to maintain comparable scores across test forms Standardized Administration and Security Procedures TOEFL program also provides extensive material to test administrators and test takers so that violations of such procedures can be reported to ETS for investigation. The major procedures are: • certifying all test centers facilities and equipment for administering TOEFL iBT tests, including hardware, software and Internet connections • training test center staff on how to handle a test administration session, including test taker identity verification, test launch and incident and irregularity management • providing online practice tests and other supporting information for test takers to become familiar with the test and test-taking conditions, including section sequence, test duration, use of headphones and microphones and navigating within and across test sections (information for test takers is available at www.toeflgoanywhere.org) • using technology to control test delivery and transmission of test-related data to ensure the security of test contents and test results • informing test takers about how to report fraudulent behavior in a test session Test Specifications Another important way to help ensure the comparability of scores across test forms is to create detailed specifications to guide the test development process. In large-scale tests such as the TOEFL iBT test, a critical Test specifications are a detailed operational definition component to ensuring score validity and fairness of test characteristics that form an exact content sampling is accomplished by implementing and adhering to plan. For example, test specifications may define the standardized procedures for test administration and kind of content covered by the test, the number related test security. of test questions, the format of test questions and Standardized test administrations and responses, and the response options. The Standards test security measures ensure the As a result, the test for Educational and Psychological Testing (AERA, APA, TOEFL iBT test is given under comparable scores reflect the NCME, 1999, p. 43) provides general guidelines for conditions to all test takers, no matter test takers’ abilities developing and evaluating test specifications. When where or when they take the test. and are not unduly multiple forms of a test are developed according to influenced by other, well-defined test specifications, the test characteristics can unrelated factors. TOEFL iBT operational procedures for be expected to remain very similar across the test forms maintaining standardized conditions for test administration and across test administrations. and security follow the requirements laid out in the ETS Standards for Quality and Fairness (ETS, 2002). The 2 oooooooo oooooooo 3 Reliability and Comparability of TOEFL iBT™ Scores The TOEFL iBT test offers multiple test administrations each year, so it is critical to ensure that test forms used for these administrations are similar in content. This is accomplished by following the test specifications for the TOEFL iBT test in the test development process. Details of how these specifications were developed using Evidence Centered Design can be found in Pearlman (2008). Score Reliability and Generalizability TM equivalent. The person would receive two test scores, but neither of the two scores would be the person’s exact true ability score. However, both scores should be around his or her true ability score. SEM is a measure that defines a score range in which one’s true ability score lies with a certain level of probability. Obviously, the smaller the SEM, the better quality (more precise) the test scores will be. Reliability is expressed in a statistical index whose value ranges from 0 (not at all reliable) to 1 (perfectly reliable). Such an index, called a coefficient, can be estimated in different ways depending on the intended use of the coefficient and the underlying theoretical frameworks. In the TOEFL iBT test, the reliability estimation for the Reading and Listening sections that contain selected response questions is carried out using a method based on item response theory (IRT) (Lord, 1980). For the Speaking and Writing sections that contain constructed response tasks, generalizability theory (G-theory) is used (Brennan, 1983). An important measure of the quality of a test is how reliable the test scores are. Reliability is important because it indicates the replicability of the test scores when either a test could be given twice or more to the same group of people, or two tests constructed in the same manner could be given to the same group of people. Testing, like other measurement events, is subject to the influence of many factors that are not relevant to The more reliable the scores are, the Generally speaking, there are two steps taken in the ability being more confidence score users have in producing IRT-based reliability estimates for Reading measured. Such using the scores for making important and Listening scores. The first step estimates the SEM irrelevant factors decisions about test takers. for each scaled score point after a new test form is contribute to what is equated, and such a SEM specific to each score point called “measurement error,” which in turn determines how is called a conditional SEM or a CSEM. Item parameters and reliable test scores are. personal ability estimates from an IRT model are used in calculating the CSEM values. In the second step, the CSEM In educational measurement, score reliability is a statistical values across all the scaled score points are averaged, and index to quantify and evaluate the consistency of test scores. this averaged CSEM value is used in the calculation of the reliability estimate for the overall test of Reading or Listening In essence, “the concern of reliability is to quantify the by substituting this overall SEM (averaged CSEM) for the SEM precision of test scores and other measurements” (Haertel, in the formula below: 2006, p. 65). A test score is a measurement outcome of a person’s performance on a test, and as such, just like any other measurement, the score contains some degree of x xx measurement error due to many factors. In fact, a person’s real or true ability can never be observed in a test. A welldeveloped test is expected to yield a test score that will where σx is the standard deviation of the scaled score, rxx reflect a person’s real ability as much as possible and to keep is the reliability estimate, and solving for rxx will then obtain measurement error to a minimum. This is what “precision the estimated reliability. The SEM is on the same metric of test scores” really means. Precision of test scores is also as the scaled scores that are reported. expressed as a measurement index called the standard Generalizability theory has seen extensive application in error of measurement (SEM). A person’s true ability score is measuring the score reliability for tests with constructed never obtainable and is therefore estimated using statistical response tasks in which the test taker generates answers methods. Imagine that a person could take two tests that are to test questions instead of selecting from a list of possible constructed to the same specifications and are essentially responses. Using an analysis of variance, generalizability SEM = σ 4 oooooooo Insight TOEFL iBT Research • Series 1, Volume 3 1− r , theory separates different sources of variance (test taker ability, rater effects, and task effects) so that the effect of each source of variance (called a facet) can be evaluated. The index for score reliability in this framework is a generalizability coefficient (G-coefficient), which is also on a scale of 0 to 1 with a value closer to 1 being desired. A ‘person by task’ generalizability model is used for the Speaking scores, whereas a nested model is used for Writing (rating nested within task, and this is crossed with person) (see Lee & Kantor, 2005 for details). The above-mentioned reliability and generalizability analyses are conducted for every test form. Table 1 presents the average section and total score reliability estimates and standard errors of measurement based on operational data from 2007. Table 1. Reliabilities and Standard Errors of Measurement Score Scale Reliability Estimate SEM Reading 0-30 0.85 3.35 Listening 0-30 0.85 3.20 Speaking 0-30 0.88 1.62 Writing 0-30 0.74 2.76 Total 0-120 0.94 5.64 The reliability estimates for the Reading, Listening, Speaking, and Total scores are relatively high, while the reliability of the Writing score is somewhat lower. This is a typical result for writing measures composed of only two tasks (Breland, Bridgeman, & Fowles, 1999) and reflects one well-documented limitation of performance testing—reliability estimates for measures composed of a small number of time-consuming tasks are often lower than estimates for measures composed of many shorter, less time-consuming tasks. However, the construct of academic writing as defined for the TOEFL iBT test required the production of extended writing samples (Cumming, Kantor, Powers, Santos, & Taylor, 2000). One implication of these results is that, for making high-stakes decisions such as admissions to college or graduate school, the Total score provides the best information, both because it reflects all four language skills and because it is the most reliable. Nevertheless, there are circumstances under which decision makers may want to examine the profile of scores for test takers, such as the demands of the curriculum or a need for additional language training. Also note that ETS encourages score users to consider a number of other factors, when making admissions decisions, including grade point average, scores on other admissions exams, teacher recommendations, and interviews with individuals. The reliability estimates in Table 1 are what are used for the TOEFL iBT operational test scores. Other types of reliability estimates also exist that take into account other sources of variability such as differences in test forms or changes in examinees’ performances from day to day. Alternate form reliability, for example, is calculated based on examinees’ scores on two different forms of a test. This requires examinees to take two different test forms, something only a few examinees would volunteer to do. But some examinees do take the test twice during a period of time too short for much learning to occur, for reasons of their own. An analysis of the scores of these repeat test takers on the two test forms provides an approximation of alternate form reliability. Zhang (February 2008) compared the test scores of more than 12,000 examinees who were identified as having taken two TOEFL iBT tests within a period of one month. The correlations of their scores on the two test forms were 0.77 for the listening and writing sections, 0.78 for reading, 0.84 for speaking, and 0.91 for the total test score. Because these measures of reliability take into account additional sources of variability, they are typically lower than internal consistency measures. Nevertheless, they indicate a high degree of consistency in the rank ordering of the scores of these test repeaters. Scaling TOEFL iBT Scores Reported test scores are derived from performance on a test through a statistical process called “scaling.” In a simple example, a student who answered 55 questions correctly out of 60 questions on a test would receive a score of 55 if each correct answer was worth one score point. This score is the number correct score, which is also called a raw score. For standardized tests, raw score scales are almost never used directly as reported scores. This is because the raw score oooooooo 5 Reliability and Comparability of TOEFL iBT™ Scores TOEFL iBT Research • Series 1, Volume 3 TM is directly dependent on the specific items on a particular Such an equating process, however, is not practical or test form; this particular form may not have exactly the feasible for performance-based tests that contain only a same difficulty level few tasks. For example, a writing test may have only A carefully developed score scale, as other forms of one or two writing tasks that are scored by human together with an equating plan (see the same test. As raters using a scoring rubric (rules or guidelines for following section), is important in a result, the same assigning scores to constructed-response questions raw scores from two maintaining score comparability and or performance tasks). Many equating procedures meaningful interpretation of scores different test forms require repeating previously administered items in across test forms and over time. may not represent the current administration. This may not be feasible if the same level of a writing test has only one or two tasks that are easily performance or ability. Instead, the raw scale is transformed remembered and shared with other examinees. to a reporting score scale, which produces a scaled score. Threats to score comparability on such performance tests The scales for the measures on the TOEFL® iBT test were result from both differences in test form difficulty, and established such that the same scale range (0-30) for the inconsistency in human raters’ scoring activities. In the four sections was chosen to indicate that all sections should absence of any feasible equating procedures, various other be viewed as being equally important in measuring the statistical analyses can be used to evaluate and control the construct of academic language ability. The total score quality of scores (Baldwin, Fowles, & Livingston, 2005). would be the sum of the four section scores. The decision to use a 0-to-30 scale was based primarily on the need to provide reasonable raw to scale score mappings for each of ™ the sections, which differed in their maximum raw scores. The maximum number of raw score points on the four sections of the form used in the filed study ranged from A non-equivalent anchor test design (also called a common 20 for writing to 44 for reading. item non-equivalent group design) is used as the equating data collection design for TOEFL iBT Reading and Listening sections. This means that each new form of the TOEFL iBT test contains an anchor block, which is a set of items that have been pretested in previously administered forms and have already been analyzed. This design enables an adjustment for possible ability differences between the group in which the items were pretested in an old test For testing programs that have multiple administrations form and the group in which the items are used in the with different test forms, it is necessary to maintain score current new test form for equating. This is possible comparability across test forms for meaningful comparison because the same items are given to two groups of of scores. Score comparability across test forms is typically candidates, and the differences in item statistics for the maintained using certain statistical processes called equating. two groups reflect the ability differences between the two groups. Such differences need to be adjusted during Equating is a statistical process that is used to adjust the equating process. scores on test forms so that scores on the forms can be used interchangeably. “Equating adjusts for differences A statistical model within the item response theory (IRT) in difficulty among forms that are built to be similar in framework is used to analyze the items and candidates’ difficulty and content” (Kolen & Brennan, 1995, p. 2). abilities. From the IRT-based analysis, item parameters and candidates’ abilities are derived. The derived item parameters For tests containing selected response items, equating is and ability values are on the TOEFL iBT IRT scales that were routinely carried out to produce reporting scores for a already established. This way, items and candidates from new test form, as is the case for TOEFL iBT Reading and different test administrations can be directly compared. Listening sections. Equating TOEFL iBT Reading and Listening Sections Maintaining Score Comparability across Test Forms 6 oooooooo Insight The IRT true score equating method (see detailed the Writing or Speaking sections with the Reading and descriptions about this method in Kolen & Brennan, 1995) Listening sections. The performance of raters is evaluated is used to establish by statistics on rater agreement rates, which include The scaled scores for any test forms the relationship both exact agreement (no score difference between are directly comparable as they between scores on two raters) and adjacent agreement (1 point now indicate the same levels of the current new difference). The average of all the scores a rater performance and ability. form and scores on assigns in a scoring session is compared with the the base form. After average score of all the raters participating in the equating, a raw score on the new form is adjusted to be same session. A large difference between these two equivalent to a raw score on the base form. Because each average scores may alert a scoring leader to a possible raw score on the base form already corresponds to a problem in a rater’s scaled score, each raw score on the new form is now Several statistical methods are also performance. implemented to monitor and evaluate related to a scaled score between 0 and 30. the performance of the tasks in each test Whenever possible, monitor papers are form, and the performance of raters. also used to evaluate ™ cross-administration scoring consistency. Monitor papers are selected responses on a task from a prior test administration that were scored previously. If the task is occasionally included in a new test administration, these monitor papers are intermixed with the responses to the task on Speaking and Writing scores are not equated statistically the new test for scoring. Because these monitor papers are due to technical and test security constraints. The two indistinguishable from the responses to the task on the new test sections have six and two constructed response tasks, test, raters will score them in the same way as they score the respectively. Because these few tasks are prone to new responses. Then, the old and new scores on the monitor memorization, test security concerns preclude the possibility papers are compared, and the agreement rates between the of repeating tasks on every new test form, as is required two sets of scores indicate cross-administration rater by equating. To consistency in scoring. minimize differences Careful test development effort and in test form difficulty rigorous scoring standards are used Another type of statistical evidence of score and potential to maintain score quality. consistency across forms comes from the analysis inconsistency due to of repeat test takers. As noted in the section on human scoring, a number of non-statistical procedures are reliability, Zhang (2008) analyzed the test scores of put in place in test development and scoring. examinees who chose to take the test twice within a short period of time. The correlational analyses established that Detailed task specifications guide the development of the examinees were rank-ordered consistently on the two parallel tasks. Then, small-scale tryouts are used to screen test forms. Zhang also reported that the differences in out poorly performing tasks. Training is given to raters using scores on the two test forms for all four sections and for the well-defined and articulated scoring rubrics. In addition, total score were negligible for most examinees. raters are certified before they can begin scoring work and, prior to each scoring session, they must pass a calibration test. Furthermore, during all scoring sessions, raters are monitored and supervised by chief raters. Comparability of TOEFL iBT Speaking and Writing Sections across Forms All the constructed response tasks in Writing and Speaking sections are analyzed after a test is given. The analysis examines such statistics as average scores on a task, distributions of scores on a task, and correlations between oooooooo 7 Reliability and Comparability of TOEFL iBT™ Scores Conclusion Evidence of score reliability and comparability allows Because different forms of the TOEFL iBT test are decision makers to evaluate the trustworthiness of the test administered to test takers at different times and in scores when the scores are used to indicate candidates’ different locations, score reliability and comparability abilities or performances. are important criteria to evaluate the quality of the test. Therefore, ETS has implemented a Evidence of score reliability and comparability for variety of procedures TOEFL iBT scores comes both from statistical to enhance test analyses and from the application of accepted test score reliability and development, administration and scoring practices. comparability. References American Educational Research Association, American Psychological Association & National Council on Measurement in Education. (1999). The standards for educational and psychological testing. Washington, DC: American Educational Research Association. www.apa.org/science/standards.html Breland, H., Bridgeman, B. & Fowles, M. E. (1999). Writing assessment in admission to higher education: Review and framework (ETS Research Rep. No. 99-03). Princeton, NJ: ETS. Available on the Web at: www.ets.org/Media/Research/pdf/RR-99-03Breland.pdf Brennan, R. L. (1983). Elements of generalizability theory. Iowa City, IA: American College Testing Program. Cronbach, L. J. (1951). Coefficient alpha and the internal structure of tests. Psychometrika, (16), 297-334. Cumming, A., Kantor, R., Powers, D., Santos, T. & Taylor, C. (2000). TOEFL 2000 writing framework: A working paper. (TOEFL Monograph No. 18). Princeton, NJ: ETS. Available on the Web at: www.ets.org/Media/Research/pdf/RM-00-05.pdf Educational Testing Service. (2002). ETS standards for quality and fairness. Princeton, NJ: Author. Available on the Web at: www.ets.org/Media/About_ETS/pdf/standards.pdf Haertel, E. H. (2006). Reliability. In R. L. Brennan (Ed.), Educational measurement (4th ed., 65 – 110). Westport, CT: American Council on Education and Praeger. Kolen, M J. & Brennan, R.L. (1995). Test equating: Methods and practices. New York, NY: Springer-Verlag. Lee, Y.-W. & Kantor, R. (2005). Dependability of new ESL writing test scores: Evaluating prototype tasks and alternative rating schemes. ETS Research Report (RR-05-14). Princeton, NJ: ETS. http://infoport/sites/rrpts/publications/RR-05-14.pdf Lord, F. M. (1980). Applications of item response theory to practical testing problems. Hillsdale, NJ: Lawrence Erlbaum Associates. Pearlman, M. (2008). Finalizing the test blueprint. In C. Chapelle, M. K. Enright & J. M. Jamieson (Eds.), Building a validity argument for the Test of English as a Foreign Language™. New York: Routledge: Taylor & Francis Group. Zhang, Y. (2008). Repeater analyses for TOEFL iBT. ETS Research Report (RM-08-05). Princeton, NJ: ETS. 8 oooooooo Contact Us [email protected] Insight TOEFL iBT™ Research • Series 1, Volume 3 Copyright © 2011 by Educational Testing Service. All rights reserved. ETS, the ETS logo, LISTENING. LEARNING. LEADING. and TOEFL are registered trademarks of Educational Testing Service (ETS) in the United States and other countries. TOEFL iBT is a trademark of ETS. EDU00025 Listening. Learning. Leading.® www.ets.org 760650