Skip to main content Accessibility help
×
Hostname: page-component-78c5997874-g7gxr Total loading time: 0 Render date: 2024-11-10T10:29:26.115Z Has data issue: false hasContentIssue false

Part III - Assessments Used in Higher Education Admissions

Published online by Cambridge University Press:  09 January 2020

María Elena Oliveri
Affiliation:
Educational Testing Service, Princeton, New Jersey
Cathy Wendler
Affiliation:
Educational Testing Service, Princeton, New Jersey
Get access

Summary

The chapters in this part provide an overview and critical discussion of the types of assessments used to help inform admissions decisions. Examples include assessments that measure academic preparation (either general or subject-specific skills), tests of language proficiency, and assessments of noncognitive and personality traits. These assessments may or may not be required of all applicants, depending on the type of institution students apply to, educational model higher education institutions use, and rigor of the admitting institution’s selection criteria, among other factors.

Type
Chapter
Information
Higher Education Admissions Practices
An International Perspective
, pp. 233 - 302
Publisher: Cambridge University Press
Print publication year: 2020

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

References

References

ACT. (2017a). ACT technical manual. Iowa City, IA: ACT.Google Scholar
ACT. (2017b). Fairness report for the ACT tests: 2015/2016. Iowa City, IA: ACT.Google Scholar
ACT. (2018). The condition of college and career readiness, national 2018. Iowa City, IA: ACT.Google Scholar
Ağazade, A. S., Caner, H., Hasipoğlu, H. N., & Civelek, A. H. (2014). Turkish university entrance test and academic achievement in undergraduate programs: A criterion-related validity study. Procedia: Social and Behavioral Sciences 116, 45824590. https://doi.org/10.1016/j.sbspro.2014.01.990.Google Scholar
Akbas, I. (2015). Comparison between undergraduate placement examination (LYS) in Turkey and EdEXCEL international advanced levels examination (IAL) in the world. Green University Review of Social Sciences, 2(2), 7193. Retrieved from http://green.edu.bd/wp-content/uploads/PDFs/Journals/GURSS/v-2-i-2/Comparison_Between_Undergraduate_Placement_Examination_LYS_in_Turkey_and_Edexcel_International_Advanced_Levels_Examination_IAL_in_the_World.pdf.Google Scholar
Akkök, F., & Watts, A. G. (2003). Public policies and career development: Country report on Turkey. Washington, DC: World Bank. Retrieved from http://siteresources.worldbank.org/INTLL/Resources/Public-Policies-and-Career-Development-Policy/Turkey_report.pdf.Google Scholar
Alavi, T. (2012). The predictive validity of final English exams as a measure of success in Iranian National University Entrance English Exam. Journal of Language Teaching and Research, 3, 224228. https://doi.org/10.4304/jltr.3.1.224-228.CrossRefGoogle Scholar
American Educational Research Association, American Psychological Association, and National Council on Measurement in Education. (2014). Standards for educational and psychological testing. Washington, DC: American Educational Research Association.Google Scholar
Atalmis, E. H. (2016). Do the guideline violations influence test difficulty of high-stake test? An investigation on University Entrance Examination in Turkey. Journal of Education and Training Studies, 4(10), 17. https://doi.org/10.11114/jets.v4i10.1738.Google Scholar
Beard, J., & Marini, J. (2018). Validity of the SAT for predicting first-year grades: 2013 SAT validity sample. New York, NY: The College Board.Google Scholar
Benton, T., & Bramley, T. (2015). The use of evidence in setting and maintaining standards in GCSEs and A levels: Discussion paper. Cambridge: Cambridge Assessment.Google Scholar
Cascallar, A. S., & Dorans, N. J. (2003). Linking scores from tests of similar content given in different languages: Spanish language PAA and English language SAT I. New York, NY: College Entrance Examination Board.Google Scholar
Choi, H. J., & Park, J.-H. (2013) Historical analysis of the policy on the college entrance system in South Korea. International Education Studies, 6(11). https://doi.org/10.5539/ies.v6n11p106.CrossRefGoogle Scholar
Clinedinst, M., & Patel, P. (2018). 2018 state of college admission. Arlington, VA: National Association for College Admission Counseling.Google Scholar
The College Board. (2008). The Puerto Rico and Latin America Office. New York, NY: Author.Google Scholar
The College Board. (2017a). SAT suite of assessments technical manual: Characteristics of the SAT. New York, NY: Author.Google Scholar
The College Board. (2017b). PAA revisada: Preguntas y respuestas [Revised PAA: Questions and answers]. Retrieved from https://latam.collegeboard.org/wp-content/uploads/2017/10/PAA-Preguntas-y-Respuestas-Puerto-Rico-2017.pdfGoogle Scholar
Council of Chief State School Officers. (2014). Criteria for procuring and evaluating high quality assessments. Retrieved from https://ccsso.org/sites/default/files/2017-10/CCSSO%20Criteria%20for%20High%20Quality%20Assessments%2003242014.pdf.Google Scholar
Council of Chief State School Officers & Association of Test Publishers. (2013). Operational best practices for statewide large-scale assessment programs (2013 ed.). Washington, DC: Council of Chief State School Officers & Association of Test Publishers.Google Scholar
Farrokhi-Khajeh-Pasha, Y., Nedjat, S., Mohammadi, A., Rad, E. M., Majdzadeh, R., Monajemi, F., & Yasdani, S. (2012). The validity of Iran’s national university entrance examination (Konkoor) for predicting medical students’ academic performance. BMC Medical Education, 12(60), 18. https://doi.org/10.1186/1472-6920-12-60.CrossRefGoogle ScholarPubMed
Gronlund, N. E., & Linn, R. L. (1990). Measurement and evaluation in teaching. New York, NY: Macmillan.Google Scholar
Hatipoğlu, C. (2016). The impact of the University Entrance Exam on EFL education in Turkey: Pre-service English language teacher’s perspective. Procedia: Social and Behavioral Sciences, 232, 136144. https://doi.org/10.1016/j.sbspro.2016.10.038.Google Scholar
Helms, R. M. (2008). University admission worldwide. Washington, DC: International Bank for Reconstruction and Development/World Bank.Google Scholar
Hudson, T. (2005). Trends in assessment scales and criterion-referenced language assessment. Annual Review of Applied Linguistics, 25, 205227. https://doi.org/10.1017/S0267190505000115.Google Scholar
Huh, N. R., & Huang, C.-Y. (2016). Examining the validity of ACT composite score and high school grade point average for predicting first-year college GPA of special-tested students. Iowa City, IA: ACT, Inc.Google Scholar
International Test Commission (2001). International guidelines for test use. International Journal of Testing, 1, 93114. https://doi.org/10.1207/S15327574IJT0102_1.Google Scholar
International Test Commission (2006). International guidelines on computer-based and internet delivered testing. International Journal of Testing, 6, 143171. https://doi.org/10.1207/s15327574ijt0602_4.Google Scholar
International Test Commission (2012). ITC guidelines on quality control in scoring, test analysis, and reporting of test scores. Retrieved from www.intestcom.org.Google Scholar
International Test Commission (2014). International guidelines on the security of tests, examinations, and other assessments. Retrieved from www.intestcom.org/files/guideline_test_security.pdf.Google Scholar
International Test Commission (2018). ITC guidelines for the large-scale assessment of linguistically and culturally diverse populations. Retrieved from www.researchgate.net/publication/315761630_ITC_Guidelines_for_the_Large-Scale_Assessment_of_Linguistically_Diverse_Populations.Google Scholar
Kamyab, S. (2008). The University Entrance Exam crisis in Iran. International Higher Education, 51,2223. https://doi.org/10.6017/ihe.2008.51.8010.Google Scholar
Kane, M. T. (2013). Validating the interpretations and uses of test scores. Journal of Educational Measurement, 50, 173. https://doi.org/10.1111/jedm.12000.Google Scholar
Kuramoto, N., & Koizumi, R. (2018). Current issues in large-scale educational assessment in Japan: Focus on national assessment of academic ability and university entrance examinations. Assessment in Education: Principles, Policy, & Practice, 25, 415433. https://doi.org/10.1080/0969594X.2016.1225667.Google Scholar
Kwon, S. K., Lee, M., & Shin, D. (2017). Educational assessment in the Republic of Korea: Lights and shadows of high-stake exam-based education system. Assessment in Education: Principles, Policy, & Practice, 24, 6077. https://doi.org/10.1080/0969594X.2015.1074540.Google Scholar
McGrath, C. H., Henham, M. L., Corbett, A., Durazzi, N., Frearson, M., Janta, B., & Schweppenstedde, D. (2014.) Higher education entrance qualifications and exams in Europe: A comparison. Directorate-General for Internal Polices, Policy Department Structural and Cohesion Policies, European Parliament. Retrieved from www.europarl.europa.eu/thinktank/en/document.html?reference=IPOL-CULT_ET(2014)529057.Google Scholar
Marini, J. P., Beard, J., & Shaw, E. (2018). Student ranking differences within institutions using old and new SAT scores. New York, NY: The College Board.Google Scholar
Marini, J. P., Shaw, E. J., & Young, L. (2016). Using old and new SAT Scores for admission: A closer look at concordant scores in predictive models (College Board Research Report No. 2016-17). New York, NY: The College Board.Google Scholar
Messick, S. (1989). Validity. In Linn, R. L. (Ed.). Educational measurement (3rd ed.) (pp. 13103). New York, NY: Collier Macmillan.Google Scholar
Mundar, D., Kečik, D., & Matotek, D. (2015). Relationship between enrolment criteria and first-year student’ study-success. Proceedings of the 3rd Human and Social Sciences at the Common Conference, Slovakia, 3(1), 111115. https://doi.org/10.18638/hassacc.2015.3.1.179.Google Scholar
National Center for University Entrance Examinations. (2015). National Center for University Entrance Examinations Annual Report. Tokyo: National Center for University Entrance Examinations. Retrieved from www.dnc.ac.jp/albums/abm.php?f=abm00006725.pdf&n=2015%E5%A4.Google Scholar
Neather, E. J. (1993). The Abitur examination. Language Learning Journal, 7, 1921. https://doi.org/10.1080/09571739385200071.CrossRefGoogle Scholar
Office of Qualifications and Examinations Regulation. (2012). International comparisons in senior secondary assessment: Full report. Coventry: Ofqual.Google Scholar
Razmjoo, S. A., & Madani, H. (2013). A content analysis of the English section of university entrance exams based on Bloom’s revised taxonomy. International Journal of Language Learning and Applied Linguistics World, 4(3), 105129.Google Scholar
Salmi, J., & Fèvre, C. (2008). Tertiary education and lifelong learning in Brazil. Retrieved from www.anped.org.br/sites/default/files/resources/SALMI_Jamil_e_F_VRE_Chlo_._Tertiary_Education_and_Lifelong_Learning_in_Brazil..pdf.Google Scholar
Sazegar, Z., & Motallebzadeh, K. (2017). Iranian National University Entrance Examination (Konkoor) of B.A.: An analysis of its reliability and validity. Modern Journal of Language Teaching Methods, 3(7), 358365.Google Scholar
Shaw, E. (2015). An SAT validity primer. New York, NY: The College Board.Google Scholar
Shaw, S., & Crisp, V. (2010). How valid are A levels? Findings from a multi-method validation study of an international A level in geography. Paper presented at the Association for Educational Assessment in Europe, Oslo, Norway.Google Scholar
Stringer, N. (2014). The achieved weightings of assessment objective as a source of validity evidence (Ofqual Report No. 14/5375). Coventry: Office of Qualifications and Examinations Regulation. Retrieved from https://assets.publishing.service.gov.uk/government/uploads/system/uploads/attachment_data/file/605434/2014-02-11-the-achieved-weightings-of-assessment-objectives-as-a-source-of-validity-evidence.pdf.Google Scholar
Sutch, T., Zanini, N., & Benton, T. (2015). A level reform: Implications for subject uptake. (Cambridge Assessment Research Report). Cambridge: Cambridge Assessment.Google Scholar
Wang, X. B. (2006). An Introduction to the system and culture of the college entrance examination of China (Office of Research and Analysis Report No. RN-28). New York, NY: The College Board.Google Scholar
Watanabe, Y. (2013). The National Center Test for university admissions. Language Testing, 30, 565573. https://doi.org/10.1177/0265532213483095.Google Scholar
Wedman, J. (2017). Theory and validity evidence for a large-scale test for selection to higher education. Umea: Umea University.Google Scholar
World Bank. (2000). Brazil: Higher education sector study (Report No. 19392-BR). Washington, DC: Human Development Department, Latin America and the Caribbean Region.Google Scholar
Zanini, N., & Williamson, J. (2017). Learning aims: A preliminary exploration to monitor A/AS level reform (Cambridge Assessment Research Report). Cambridge: Cambridge Assessment.Google Scholar

References

Alderson, J. C. (2005). Diagnosing foreign language proficiency: The interface between learning and assessment. London: Continuum.Google Scholar
Althaus, H.-J. (2018). Warum C1 keine Lösung ist: Der Nachweis von Deutschkenntnissen für den Hochschulzugang, der GER und warum sie nicht zusammenpassen [Why C1 is not the solution: Proving knowledge of German for entrance to higher education, the CEFR, and why they do not fit together]. In Brandt, A., Buschmann-Göbels, A., & Harsch, C. (Eds.). Der Gemeinsame Europäische Referenzrahmen für Sprachen und seine Adaption im Hochschulkontext: Erträge des 6. Bremer Symposions, 2017 [The common European framework of reference for languages and its adaptation to the university context: Results from the 6th Bremen symposium, 2017] (pp. 78–98). Bochum: AKS-Verlag.Google Scholar
American Educational Research Association, American Psychological Association, & National Council on Measurement in Education. (2014). Standards for educational and psychological testing. Washington, DC: American Educational Research Association.Google Scholar
Bachman, L. F., & Palmer, A. S. (2010). Language assessment in practice: Developing language assessments and justifying their use in the real world. Oxford: Oxford University Press.Google Scholar
Biber, D., Reppen, R., & Staples, S. (2017). Exploring the relationship between TOEFL iBT scores and disciplinary writing performance. TESOL Quarterly, 51, 948960. https://doi.org/10.1002/tesq.359.CrossRefGoogle Scholar
Bridgeman, B., Cho, Y., & DiPietro, S. (2016). Predicting grades from an English language assessment: The importance of peeling the onion. Language Testing, 33, 307318. https://doi.org/10.1177/0265532215583066.Google Scholar
Byram, M., & Parmenter, L. (Eds.). (2012). The common European framework of reference: The globalisation of language education policy. Bristol: Multilingual Matters. https://doi.org/10.21832/9781847697318.Google Scholar
Camara, W. (2013). Defining and measuring college and career readiness: A validation framework. Educational Measurement: Issues and Practice, 32(4), 1627. https://doi.org/10.1111/emip.12016.Google Scholar
Carlsen, C. H. (2018). The adequacy of the B2 level as university entrance requirement. Language Assessment Quarterly, 15, 7589. https://doi.org/10.1080/15434303.2017.1405962.Google Scholar
Chalhoub-Deville, M. (2009). Content validity considerations in language testing contexts. In Lissitz, R. W. (Ed.). The concept of validity: Revisions, new directions, and applications (pp. 241263). Charlotte, NC: Information Age.Google Scholar
Cizek, G. J., & Bunch, M. B. (2007). Standard setting: A guide to establishing and evaluating performance standards on tests. Thousand Oaks, CA: Sage. https://doi.org/10.4135/9781412985918.Google Scholar
Council of Europe. (2001). Common European framework of reference for languages: Learning, teaching, assessment. Cambridge: Cambridge University Press.Google Scholar
Deygers, B., Van den Branden, K., & Peters, E. (2017). Checking assumed proficiency: Comparing L1 and L2 performance on a university entrance test. Assessing Writing, 32, 4356. https://doi.org/10.1016/j.asw.2016.12.005.Google Scholar
Deygers, B., Van den Branden, K., & Van Gorp, K. (2018). University entrance language tests: A matter of justice. Language Testing, 35, 449476. https://doi.org/10.1177/0265532217706196.Google Scholar
Deygers, B., Van Gorp, K., & Demeester, T. (2018). The B2 level and the dream of a common standard. Language Assessment Quarterly, 15, 4458. https://doi.org/10.1080/15434303.2017.1421955.CrossRefGoogle Scholar
Deygers, B., Zeidler, B., Vilcu, D., & Carlsen, C. H. (2018). One framework to unite them all? Use of the CEFR in European university entrance policies. Language Assessment Quarterly, 15, 315. https://doi.org/10.1080/15434303.2016.1261350.Google Scholar
Eckes, T. (2015). Introduction to many-facet Rasch measurement: Analyzing and evaluating rater-mediated assessments (2nd ed.). Frankfurt am Main: Peter Lang.Google Scholar
Eckes, T. (2017). Setting cut scores on an EFL placement test using the prototype group method: A receiver operating characteristic (ROC) analysis. Language Testing, 34, 383411. https://doi.org/10.1177/0265532216672703.Google Scholar
Elder, C. (2017). Language assessment in higher education. In Shohamy, E., Or, I. G., & May, S. (Eds.). Language testing and assessment (3rd ed., pp. 271286). Cham: Springer. https://doi.org/10.1007/978-3-319-02261-1_35.Google Scholar
Elder, C. (2018). Test review. Certifying French competency: The DELF tout public (B2). Language Testing, 35, 615623. https://doi.org/10.1177/0265532218781627.Google Scholar
Fawcett, T. (2006). An introduction to ROC analysis. Pattern Recognition Letters, 27, 861874. https://doi.org/10.1016/j.patrec.2005.10.010.Google Scholar
Floyd, C. B. (2015). Closing the gap: International student pathways, academic performance and academic acculturation. Journal of Academic Language and Learning, 9(2), A1A18.Google Scholar
Ginther, A., & Yan, X. (2018). Interpreting the relationships between TOEFL iBT scores and GPA: Language proficiency, policy, and profiles. Language Testing, 35, 271295. https://doi.org/10.1177/0265532217704010.Google Scholar
Graham, J. G. (1987). English language proficiency and the prediction of academic success. TESOL Quarterly, 21, 505521. https://doi.org/10.2307/3586500.Google Scholar
Green, A. (2012). Placement testing. In Coombe, C., Davidson, P., O’Sullivan, B., & Stoynoff, S. (Eds.). The Cambridge guide to second language assessment (pp. 164170). Cambridge: Cambridge University Press.Google Scholar
Green, A. (2018). Linking tests of English for academic purposes to the CEFR: The score user’s perspective. Language Assessment Quarterly, 15, 5974. https://doi.org/10.1080/15434303.2017.1350685.Google Scholar
Harsch, C. (2018). How suitable is the CEFR for setting university entrance standards? Language Assessment Quarterly, 15, 102108. https://doi.org/10.1080/15434303.2017.1420793.Google Scholar
Harsch, C., Ushioda, E., & Ladroue, C. (2017). Investigating the predictive validity of TOEFL iBT® test scores and their use in informing policy in a United Kingdom university setting (Research Report No. RR-17-41). Princeton, NJ: Educational Testing Service.Google Scholar
Institute of International Education (2017). A world on the move: Trends in global student mobility. New York, NY: Author.Google Scholar
International Test Commission. (2018). ITC Guidelines for the large-scale assessment of linguistically and culturally diverse populations. Retrieved from www.intestcom.org/files/guideline_diverse_populations.pdf.Google Scholar
Jin, Y., Wu, Z., Alderson, C., & Song, W. (2017). Developing the China Standards of English: Challenges at macropolitical and micropolitical levels. Language Testing in Asia, 7(1). https://languagetestingasia.springeropen.com/articles/10.1186/s40468-017-0032-5.Google Scholar
Kane, M. T. (2010). Validity and fairness. Language Testing, 27, 177182. https://doi.org/10.1177/0265532209349467.Google Scholar
Kane, M. T. (2013). Validating the interpretations and uses of test scores. Journal of Educational Measurement, 50, 173. https://doi.org/10.1111/jedm.12000.CrossRefGoogle Scholar
Kappe, R., & van der Flier, H. (2012). Predicting academic success in higher education: What’s more important than being smart? European Journal of Psychology of Education, 27, 605619. https://doi.org/10.1007/s10212–011-0099-9.CrossRefGoogle Scholar
Kecker, G., & Eckes, T. (2010). Putting the manual to the test: The TestDaF–CEFR linking project. In Martyniuk, W. (Ed.). Aligning tests with the CEFR: Reflections on using the Council of Europe’s draft manual (pp. 5079). Cambridge: Cambridge University Press.Google Scholar
Lee, Y.-W. (2015). Diagnosing diagnostic language assessment. Language Testing, 32, 299316. https://doi.org/10.1177/0265532214565387.CrossRefGoogle Scholar
Lim, G. S., Geranpayeh, A., Khalifa, H., & Buckendahl, C. W. (2013). Standard setting to an international reference framework: Implications for theory and practice. International Journal of Testing, 13, 3249. https://doi.org/10.1080/15305058.2012.678526.Google Scholar
Marks, D. (2015). Prüfen sprachlicher Kompetenzen internationaler Studienanfänger an deutschen Hochschulen: Was leistet der TestDaF? [Assessing international students’ language proficiency at German institutions of higher education: How well does the TestDaF perform?] Zeitschrift für Interkulturellen Fremdsprachenunterricht, 20, 2139. Retrieved from http://tujournals.ulb.tu-darmstadt.de/index.php/zif/article/view/189.Google Scholar
McNamara, T., & Ryan, K. (2011). Fairness versus justice in language testing: The place of English literacy in the Australian citizenship test. Language Assessment Quarterly, 8, 161178. https://doi.org/10.1080/15434303.2011.565438.CrossRefGoogle Scholar
Myford, C. M., & Wolfe, E. W. (2003). Detecting and measuring rater effects using many-facet Rasch measurement: Part I. Journal of Applied Measurement, 4, 386422.Google Scholar
Negishi, M., Takada, T., & Tono, Y. (2013). A progress report on the development of the CEFR-J. In Galaczi, E. D. & Weir, C. J. (Eds.). Exploring language frameworks: Proceedings of the ALTE Kraków Conference, July 2011 (pp. 135163). Cambridge: Cambridge University Press.Google Scholar
Norris, J., & Drackert, A. (2018). Test review: TestDaF. Language Testing, 35, 149157. https://doi.org/10.1177/0265532217715848.Google Scholar
North, B. (2014). The CEFR in practice. Cambridge: Cambridge University Press.Google Scholar
North, B., Martyniuk, W., & Panthier, J. (2010). The manual for relating language examinations to the common European framework of reference for languages in the context of the Council of Europe’s work on language education. In Martyniuk, W. (Ed.). Aligning tests with the CEFR: Reflections on using the Council of Europe’s draft manual (pp. 117). Cambridge: Cambridge University Press.Google Scholar
O’Loughlin, K. (2011). The interpretation and use of proficiency test scores in university selection: How valid and ethical are they? Language Assessment Quarterly, 8, 146160. https://doi.org/10.1080/15434303.2011.564698.Google Scholar
Organisation for Economic Co-operation and Development. (2017). Education at a glance 2017: OECD indicators. Paris: OECD Publishing.Google Scholar
Pitoniak, M. J., & Morgan, D. L. (2012). Setting and validating cut scores for tests. In Secolsky, C. & Denison, D. B. (Eds.). Handbook on measurement, assessment, and evaluation in higher education (pp. 343366). New York, NY: Routledge.Google Scholar
Project Atlas. (2017). A quick look at global mobility trends. New York, NY: Institute of International Education.Google Scholar
Read, J. (2015). Assessing English proficiency for university study. New York, NY: Palgrave Macmillan. https://doi.org/10.1057/9781137315694.CrossRefGoogle Scholar
Sawyer, R. (2013). Beyond correlations: Usefulness of high school GPA and test scores in making college admissions decisions. Applied Measurement in Education, 26, 89112. https://doi.org/10.1080/08957347.2013.765433.Google Scholar
Schmidt, F. L., & Hunter, J. E. (2015). Methods of meta-analysis: Correcting error and bias in research findings (3rd ed.). Thousand Oaks, CA: Sage. https://doi.org/10.4135/9781483398105.Google Scholar
Stemler, S. E. (2012). What should university admissions tests predict? Educational Psychologist, 47, 517. https://doi.org/10.1080/00461520.2011.611444.Google Scholar
Swets, J. A. (1988). Measuring the accuracy of diagnostic systems. Science, 240, 12851293. https://doi.org/10.1126/science.3287615.Google Scholar
Tannenbaum, R. J., & Baron, P. A. (2015). Mapping TOEIC® scores to the Vietnamese national standard: A study to recommend English language requirements for admissions into and graduation from Vietnamese universities (ETS RM-15-08). Princeton, NJ: Educational Testing Service.Google Scholar
Taylor, L. (2013). Communicating the theory, practice and principles of language testing to test stakeholders: Some reflections. Language Testing, 30, 403412. https://doi.org/10.1177/0265532213480338.Google Scholar
Wright, B. D. (1996). Reliability and separation. Rasch Measurement Transactions, 9, 472.Google Scholar
Xi, X. (2010). How do we go about investigating test fairness? Language Testing, 27, 147170. https://doi.org/10.1177/0265532209349465.Google Scholar
Xi, X., Bridgeman, B., & Wendler, C. (2014). Tests of English for academic purposes in university admissions. In Kunnan, A. J. (Ed.). The companion to language assessment: Evaluation, methodology, and interdisciplinary themes (Vol. 1, pp. 318337). Chichester: Wiley.Google Scholar
York, T. T., Gibson, C., & Rankin, S. (2015). Defining and measuring academic success. Practical Assessment, Research and Evaluation, 20(5), 120.Google Scholar
Zheng, Y., & Cheng, L. (2008). Test review: College English Test (CET) in China. Language Testing, 25, 408417. https://doi.org/10.1177/0265532208092433.Google Scholar

References

Aamodt, M. G., Bryan, D. A., & Whitcomb, A. J. (1993). Predicting performance with letters of recommendation. Public Personnel Management, 22, 8190. https://doi.org/10.1177/009102609302200106.Google Scholar
Al Ansari, A., Al Khalifa, K., Al Azzawi, M., Al Amer, R., Al Sharqi, D., Al-Mansoor, A., & Munshi, F. M. (2015). Cross-cultural challenges for assessing medical professionalism among clerkship physicians in a Middle Eastern country (Bahrain): Feasibility and psychometric properties of multisource feedback. Advances in Medical Education and Practice, 6, 509. https://doi.org/10.2147/AMEP.S86068.Google Scholar
American Educational Research Association, American Psychological Association, & National Council on Measurement in Education. (1985). Standards for educational and psychological testing. Washington, DC: American Psychological Association.Google Scholar
Beach, L. R. (2010). The psychology of narrative thought. Bloomington, IN: Xlibris.Google Scholar
Biernat, M., & Eidelman, S. (2007). Translating subjective language in letters of recommendation: The case of the sexist professor. European Journal of Social Psychology, 37, 11491175. https://doi.org/10.1002/ejsp.432.Google Scholar
Bouton, L. F. (1995). A cross-cultural analysis of the structure and content of letters of reference. Studies in Second Language Acquisition, 17, 211244. https://doi.org/10.1017/S0272263100014169.Google Scholar
Brown, A., & Maydeu-Olivares, A. (2013). How IRT can solve problems of ipsative data in forced-choice questionnaires. Psychological Methods, 18, 3652. https://doi.org/10.1037/a0030641.Google Scholar
Campion, M. A., Palmer, D. K., & Campion, J. E. (1997). A review of structure in the selection interview. Personnel Psychology, 50, 655702. https://doi.org/10.1111/j.1744-6570.1997.tb00709.x.Google Scholar
Cirillo, M., Smith, T., & Kiran, J. (2008). The development of biographical inventory for use in the prediction of first-year college success. TCNJ Journal of Student Scholarship, 10, 111.Google Scholar
Connelly, B. S., & Ones, D. S. (2010). Another perspective on personality: Meta-analytic integration of observers’ accuracy and predictive validity. Psychological Bulletin, 136, 1092. https://doi.org/10.1037/a0021212.Google Scholar
Costa, P. T. Jr., Terracciano, A., & McCrae, R. R. (2001). Gender differences in personality traits across cultures: Robust and surprising findings. Journal of Personality and Social Psychology, 81, 322331. https://doi.org/10.1037/0022-3514.81.2.322.Google Scholar
Credé, M., & Kuncel, N. R. (2008). Study habits, study skills, and study attitudes: A meta-analysis of their relationship to academic performance among college students. Perspectives on Psychological Science, 3, 425453. https://doi.org/10.1111/j.1745-6924.2008.00089.x.Google Scholar
Credé, M., Tynan, M. C., & Harms, P. D. (2017). Much ado about grit: A meta-analytic synthesis of the grit literature. Journal of Personality and Social Psychology, 113, 492511. https://doi.org/10.1037/pspp0000102.Google Scholar
Educational Testing Service. (1974). SIGI: A computer-based system of interactive guidance and information. Princeton, NJ: Educational Testing Service.Google Scholar
Fajardo, D. M. (1985). Author race, essay quality, and reverse discrimination. Journal of Applied Social Psychology, 15, 255268. https://doi.org/10.1111/j.1559-1816.1985.tb00900.x.CrossRefGoogle Scholar
Finegold, A. (1994). Gender differences in personality: A meta-analysis. Psychological Bulletin, 116, 429456. https://doi.org/10.1037/0033-2909.116.3.429.Google Scholar
Flanagan, J. C. (1954). The critical incident technique. Psychological Bulletin, 51, 327358. http://dx.doi.org/10.1037/h0061470.Google Scholar
Foldes, H. J., Duehr, E. E., & Ones, D. S. (2008). Group differences in personality: Meta-analyses comparing five U.S. racial groups. Personnel Psychology, 61, 579616. https://doi.org/10.1111/j.1744-6570.2008.00123.x.Google Scholar
Friedman, R., Fang, C. H., Hasbun, J., Han, H., Mady, L. J., Eloy, J. A., & Kalyoussef, E. (2017). Use of standardized letters of recommendation for otolaryngology head and neck surgery residency and the impact of gender. Laryngoscope, 127(12), 27382745. https://doi.org/10.1002/lary.26619.Google Scholar
Gale, J., Ooms, A., Grant, R., Paget, K., & Marks-Maran, D. (2016). Student nurse selection and predictability of academic success: The multiple mini interview project. Nurse Education Today, 40, 123127. https://doi.org/10.1016/j.nedt.2016.01.031.Google Scholar
Girzadas, D. V. Jr., Harwood, R. C., Dearie, J., & Garrett, S. (1998). A comparison of standardized and narrative letters of recommendation. Academic Emergency Medicine, 5(11), 11011104. https://doi.org/10.1111/j.1553-2712.1998.tb02670.x.Google Scholar
Grove, W. M., & Meehl, P. E. (1996). Comparative efficiency of informal (subjective, impressionistic) and formal (mechanical, algorithmic) prediction procedures: The clinical-statistical controversy. Psychology, Public Policy, and Law, 2, 293323. https://doi.org/10.1037/1076-8971.2.2.293.Google Scholar
Hawkins, D. A., & Lautz, J. (2005). State of college admission. Alexandria, VA: National Association for College Admission Counseling.Google Scholar
Henderson, M. C., Kelly, C. J., Griffin, E., Hall, T. R., Jerant, A., Peterson, E. M., Rainwater, J. A., Sousa, F. J., Wofsy, D. & Franks, P. (2018). Medical school applicant characteristics associated with performance in multiple mini-interviews: A multi-institutional study. Academic Medicine: Journal of the Association of American Medical Colleges , 93(7), 10291034.Google Scholar
Holland, J. L., & Nichols, R. C. (1964). Prediction of academic and extra-curricular achievement in college. Journal of Educational Psychology, 55, 55. https://doi.org/10.1037/h0047977.Google Scholar
Houser, C., & Lemmons, K. (2017). Implicit bias in letters of recommendation for an undergraduate research internship. Journal of Further and Higher Education, 38, 111. https://doi.org/10.1080/0309877X.2017.1301410.Google Scholar
Huber, C. (2017). Faking and the validity of personality tests: Using new faking-resistant measures to study some old questions (Doctoral dissertation). University of Minnesota.Google Scholar
Huffcutt, A. I., & Arthur, W. (1994). Hunter and Hunter (1984) revisited: Interview validity for entry-level jobs. Journal of Applied Psychology, 79(2), 184190. http://dx.doi.org/10.1037/0021-9010.79.2.184.Google Scholar
Huffcutt, A. I., Conway, J. M., Roth, P. L., & Stone, N. J. (2001). Identification and meta-analytic assessment of psychological constructs measured in employment interviews. Journal of Applied Psychology, 86, 897913. https://doi.org/10.1037/0021-9010.86.5.897.Google Scholar
Huffcutt, A. I., & Roth, P. L. (1998). Racial group differences in employment interview evaluations. Journal of Applied Psychology, 83, 179189. https://doi.org/10.1037/0021-9010.83.2.179.Google Scholar
Hurtz, G. M., & Donovan, J. J. (2000). Personality and job performance: The big five revisited. Journal of Applied Psychology, 85, 869879. https://doi.org/10.1037/0021-9010.85.6.869.Google Scholar
Keiser, H. N., Sackett, P. R., Kuncel, N. R., & Brothen, T. (2016). Why women perform better in college than admission scores would predict: Exploring the roles of conscientiousness and course-taking patterns. Journal of Applied Psychology, 101, 569581. https://doi.org/10.1037/apl0000069.CrossRefGoogle ScholarPubMed
Kryger, B. R., & Shikiar, R. (1978). Sexual discrimination in the use of letters of recommendation: A case of reverse discrimination. Journal of Applied Psychology, 63(3), 309314. http://dx.doi.org/10.1037/0021-9010.63.3.309.Google Scholar
Kuncel, N. R., Borneman, M., & Kiger, T. (2011). Innovative item response process and Bayesian faking detection methods: More questions than answers. In Ziegler, M., Maccann, C., & Roberts, R. D. (Eds.). New perspectives on faking in personality assessment. Oxford: Oxford University Press. https://doi.org/10.1093/acprof:oso/9780195387476.003.0036.Google Scholar
Kuncel, N. R., Brenneman, M., Petway, K., & Liu, J. (2018). Validation of the character skills snapshot (CSS). Skillman, NJ: Enrollment Management Association.Google Scholar
Kuncel, N. R., & Hezlett, S. A. (2010). Fact and fiction in standardized admissions testing. Current Directions in Psychological Science, 19, 339345. https://doi.org/10.1177/0963721410389459.Google Scholar
Kuncel, N. R., Klieger, D. M., Connelly, B. S., & Ones, D. S. (2013). Mechanical versus clinical data combination in selection and admissions decisions: A meta-analysis. Journal of Applied Psychology, 98, 10601072. https://doi.org/10.1037/a0034156.Google Scholar
Kuncel, N. R., Kochevar, R. J., & Ones, D. S. (2014). A meta-analysis of letters of recommendation in college and graduate admissions. Reasons for hope. International Journal of Selection and Assessment, 22, 101107. https://doi.org/10.1111/ijsa.12060.Google Scholar
Kuncel, N. R., & Tellegen, A. (2009). The measurement of the social desirability of items: A conceptual and empirical reexamination. Personnel Psychology, 62, 201228. https://doi.org/10.1111/j.1744-6570.2009.01136.x.Google Scholar
Kuncel, N. R., & Sackett, P. R. (2014). Resolving the assessment center construct validity problem (as we know it)Journal of Applied Psychology99(1), 38. http://dx.doi.org/10.1037/a0034147.Google Scholar
Kuncel, N. R., & Sackett, P. R. (2018, March 10). The gatekeeper tests. The Wall Street Journal, pp. C1–C2.Google Scholar
Lerner, J. S., & Tetlock, P. E. (2003). Bridging individual, interpersonal, and institutional approaches to judgment and decision making: The impact of accountability on cognitive bias. In Schneider, S. L. & Shanteau, J. (Eds.). Emerging Perspectives on Judgment and Decision Research (pp. 431–457). https://doi.org/10.1017/CBO9780511609978.015.Google Scholar
Lievens, F., Buyse, T., & Sackett, P. R. (2005). The operational validity of a video-based situational judgment test for medical college admissions: Illustrating the importance of matching predictor and criterion construct domains. Journal of Applied Psychology, 90, 442452. http://dx.doi.org/10.1037/0021-9010.90.3.442.Google Scholar
Lievens, F., Buyse, T., Sackett, P. R., & Connelly, B. S. (2012). The effects of coaching on situational judgment tests in high-stakes selection. International Journal of Selection and Assessment, 20, 272282. https://doi.org/10.1111/j.1468-2389.2012.00599.x.Google Scholar
Lievens, F., & Sackett, P. R. (2006). Video-based versus written situational judgment tests: A comparison in terms of predictive validity. Journal of Applied Psychology, 91(5), 1181. http://dx.doi.org/10.1037/0021-9010.91.5.1181.Google Scholar
Lievens, F., Sackett, P. R., & Buyse, T. (2009). The effects of response instructions on situational judgment test performance and validity in a high-stakes context. Journal of Applied Psychology, 94, 10951101. http://dx.doi.org/10.1037/a0014628.Google Scholar
Lumb, A. B., Homer, M., & Miller, A. (2010). Equity in interviews: Do personal characteristics impact on admission interview scores? Medical Education, 44, 10771083. https://doi.org/10.1111/j.1365-2923.2010.03771.x.Google Scholar
McDaniel, M. A., Morgeson, F. P., Finnegan, E. B., Campion, M. A., & Braverman, E. P. (2001). Predicting job performance using situational judgement tests: A clarification of the literature. Journal of Applied Psychology, 86, 730740.Google Scholar
Mackenzie, R. S. (1967). Predictive validity of a biographical inventory in higher education. Journal of Applied Psychology, 51, 544546. https://doi.org/10.1037/h0025108.Google Scholar
Max, B. A., Gelfand, B., Brooks, M. R., Beckerly, R., & Segal, S. (2010). Have personal statements become impersonal? An evaluation of personal statements in anesthesiology residency applications. Journal of Clinical Anesthesia, 22, 346351. https://doi.org/10.1016/j.jclinane.2009.10.007.Google Scholar
Morgan, W. B., Elder, K. B., & King, E. B. (2013). The emergence and reduction of bias in letters of recommendation. Journal of Applied Social Psychology, 43(11), 22972306. https://doi.org/10.1111/jasp.12179.Google Scholar
Murphy, S. R., Klieger, D. M., Borneman, M., & Kuncel, N. R. (2009). The predictive power of personal statements in admissions: A meta-analysis and cautionary tale. College and University, 84(4), 8388.Google Scholar
Nicklin, J. M., & Roch, S. G. (2008). Biases influencing recommendation letter contents: physical attractiveness and gender. Journal of Applied Social Psychology, 38, 30533074. http://dx.doi.org/10.1111/j.1559-1816.2008.00425.x.Google Scholar
Nye, C. D., Su, R., Rounds, J., & Drasgow, F. (2012). Vocational interests and performance: A quantitative summary of over 60 years of research. Perspectives on Psychological Science, 7, 384403. https://doi.org/10.1177/1745691612449021.Google Scholar
Oliveri, M. E., McCaffrey, D., Ezzo, C., & Holtzman, S. (2017). A multilevel factor analysis of third-party evaluations of noncognitive constructs used in admissions decision making. Applied Measurement in Education, 30, 297313. https://doi.org/10.1080/08957347.2017.1353989.Google Scholar
Osman, N. Y., Schonhardt‐Bailey, C., Walling, J. L., Katz, J. T., & Alexander, E. K. (2015). Textual analysis of internal medicine residency personal statements: Themes and gender differences. Medical Education, 49, 93102. https://doi.org/10.1111/medu.12487.Google Scholar
Oswald, F. L., Schmitt, N., Kim, B. H., Ramsey, L. J., & Gillespie, M. A. (2004). Developing a biodata measure and situational judgment inventory as predictors of college student performance. Journal of Applied Psychology, 89, 187207. http://dx.doi.org/10.1037/0021-9010.89.2.187.Google Scholar
The papers of Benjamin Franklin. n.d. Retrieved from http://franklinpapers.org/.Google Scholar
Pau, A., Jeevaratnam, K., Chen, Y. S., Fall, A. A., Khoo, C., & Nadarajah, V. D. (2013). The multiple mini-interview (MMI) for student selection in health professions training: A systematic review. Medical Teacher, 35, 10271041. https://doi.org/10.3109/0142159X.2013.829912.Google Scholar
Pavlov, G., Maydeu-Olivares, A., & Fairchild, A. J. (2018). Effects of applicant faking on forced-choice and Likert scores. Organizational Research Methods, 1–30. https://doi.org/10.1177/1094428117753683.Google Scholar
Peeters, H., & Lievens, P. (2005). Situational judgment tests and their predictiveness of college students’ success: The influence of faking. Educational and Psychological Measurement, 65, 7089. https://doi.org/10.1177/0013164404268672.Google Scholar
Poropat, A. E. (2009). A meta-analysis of the five-factor model of personality and academic performance. Psychological Bulletin, 135, 322338. https://doi.org/10.1037/a0014996.Google Scholar
Precht, K. (1998). A cross-cultural comparison of letters of recommendation. English for Specific Purposes, 17, 241265. https://doi.org/10.1016/S0889–4906(97)00012-4.Google Scholar
Roberts, B. W., Kuncel, N. R., Shiner, R., Caspi, A., & Goldberg, L. R. (2007). The comparative predictive validity of personality traits, SES, and cognitive ability. Perspectives on Psychological Science, 2, 331345. https://doi.org/10.1111/j.1745-6916.2007.00047.x.Google Scholar
Salgado, J. F. (1997). The five factor model of personality and job performance in the European community. Journal of Applied Psychology, 82, 3043. https://doi.org/10.1037/0021-9010.82.1.30.Google Scholar
Shahani, C., & Dipboye, R. L. (1993). Attractiveness bias in the interview: Exploring the boundaries of an effect. Basic and Applied Social Psychology, 14, 317328. https://doi.org/10.1207/s15324834basp1403_5.Google Scholar
Sisk, V. F., Burgoyne, A. P., Sun, J., Butler, J. L., & Macnamara, B. N. (2018). To what extent and under what circumstance are growth mind-sets important to academic achievement? Two meta-analyses. Psychological Science, 29, 549571. https://doi.org/10.1177/0956797617739704.Google Scholar
Stark, S., Chernyshenko, O. S., & Drasgow, F. (2005). An IRT approach to constructing and scoring pairwise preference items involving stimuli on different dimensions. Applied Psychological Measurement, 29, 184203. https://doi.org/10.1177/0146621604273988.Google Scholar
Terregino, C. A., McConnell, M., & Reiter, H. I. (2015). The effect of differential weighting of academics, experiences, and competencies measured by multiple mini interview (MMI) on race and ethnicity of cohorts accepted to one medical school. Academic Medicine: Journal of the Association of American Medical Colleges, 90, 16511657. https://doi.org/10.1097/ACM.0000000000000960.Google Scholar
Watson, C. (1987). Sex-linked differences in letters of recommendation. Women and Language, 10(2), 26. Retrieved from http://erepo.usiu.ac.ke/handle/11732/1283?show=full.Google Scholar
Whetzel, D. L., McDaniel, M. A., & Nguyen, N. T. (2008). Subgroup differences in situational judgment test performance: A meta-analysisHuman Performance21(3), 291-309.Google Scholar
Wright, S. R., & Bradley, P. M. (2010). Has the UK clinical aptitude test improved medical student selection? Medical Education, 44, 10691076. https://doi.org/10.1111/j.1365-2923.2010.03792.x.Google Scholar
Zhang, C. S. H., & Kuncel, N. R. (2018). Moving beyond the brag sheet: Predicting student outcomes with biodata measures. Unpublished manuscript. University of Minnesota.Google Scholar

Save book to Kindle

To save this book to your Kindle, first ensure coreplatform@cambridge.org is added to your Approved Personal Document E-mail List under your Personal Document Settings on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part of your Kindle email address below. Find out more about saving to your Kindle.

Note you can select to save to either the @free.kindle.com or @kindle.com variations. ‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi. ‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.

Find out more about the Kindle Personal Document Service.

Available formats
×

Save book to Dropbox

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Dropbox.

Available formats
×

Save book to Google Drive

To save content items to your account, please confirm that you agree to abide by our usage policies. If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account. Find out more about saving content to Google Drive.

Available formats
×