Article published In: Interpreting
Vol. 21:2 (2019) ► pp.245–269
Holistic assessment of consecutive interpretation
How interpreter trainers rate student performances
Published online: 11 November 2019
https://doi.org/10.1075/intp.00029.lee
https://doi.org/10.1075/intp.00029.lee
Abstract
The article aims to show how interpreter trainers holistically grade student performances. For this purpose,
experimental rating sessions were held for four undergraduate interpreter trainers. The raters were asked to think aloud their
quality judgments while holistically assessing six recordings of consecutive interpretation. Their concurrent verbal reports,
along with reflective reports, interview transcripts, and video recordings of computer screen activity, were collected and
analysed in detail. Findings revealed various facets of interpreting performance assessment, including what procedures the raters
followed, what aspects of the performance they focused on, what criteria they depended on for their judgment decisions, and why
two ratings of the same performance were divergent. This article also presents a tentative model for holistic rating of
consecutive interpretation.
Keywords: raters, quality judgment, rating process, think-aloud protocol (TAP), coding
Article outline
- 1.Introduction
- 2.Literature review
- 2.1Procedural dimension of interpreting performance assessment
- 2.2Thinking aloud as a data elicitation method
- 3.Method
- 3.1Participants
- 3.2Task materials
- 3.3Procedures
- 3.4Coding schemes
- 3.5Data analysis
- 4.Results
- 4.1Lengths of the verbal protocols
- 4.2Grades
- 4.3Rating styles and methods
- 4.3.1R1
- 4.3.2R2
- 4.3.3R3
- 4.3.4R4
- 4.4Distribution of positive and negative comments by protocol code
- 4.4.1R1
- 4.4.2R2
- 4.4.3R3
- 4.4.4R4
- 4.5A tentative model of the holistic rating process
- 5.Concluding remarks
- Acknowledgements
- Notes
References
References (57)
Bartłomiejczyk, M. (2007). Interpreting quality as perceived by trainee interpreters: Self-evaluation. The Interpreter and Translator Trainer 1 (2), 247–267.
Bejar, I. I. (2012). Rater cognition: Implications for validity. Educational Measurement: Issues and Practice 31 (3), 2–9.
Bernardini, S. (2001). Think-aloud protocols in translation research: Achievements, limits, future prospects. Target 13 (2), 241–263.
Bontempo, K. & Napier, J. (2011). Evaluating emotional stability as a predictor of interpreter competence and aptitude for interpreting. Interpreting 13 (1), 85–105.
Bühler, H. (1986). Linguistic (semantic) and extralinguistic (pragmatic) criteria for the evaluation of conference interpretation and interpreters. Multilingua 5 (4), 231–235.
Chabasse, C. & Kader, S. (2014). Putting interpreting admissions exams to the test. Interpreting 16 (1), 19–33.
Chiaro, D. & Nocella, G. (2004). Interpreters’ perception of linguistic and non-linguistic factors affecting quality: A survey through the World Wide Web. Meta 49 (2), 279–293.
Choi, J. Y. (2013). Assessing the impact of text length on consecutive interpreting. In D. Tsagari & R. van Deemter (Eds.), Assessment issues in language translation and interpreting. Frankfurt am Main: Peter Lang, 85–96.
Cifuentes-Férez, P. & Rojo, A. (2015). Thinking for translating: A think-aloud protocol on the translation of manner-of-motion verbs. Target 27 (2), 273–300.
Clifford, A. (2005). Putting the exam to the test: Psychometric validation and interpreter certification. Interpreting 7 (1), 97–131.
Coffey, A. & Atkinson, P. (1996). Making sense of qualitative data: Complementary research strategies. Thousand Oaks, CA: Sage.
Cohen, A. D. (2000). Exploring strategies in test-taking: Fine-tuning verbal reports from respondents. In G. Ekbatani & H. Pierson (Eds.), Learner-directed assessment in ESL. Mahwah, NJ: Lawrence Erlbaum, 127–150.
De Gregoris, G. (2014). The limits of expectations vs. assessment questionnaire-based surveys on simultaneous interpreting quality: The need for a gestaltic model of perception. Rivista internazionale di tecnica della traduzione 161, 57–87.
DeRember, M. L. (1998). Writing assessment: Raters’ elaboration of the rating task. Assessing Writing 5 (1), 7–29.
Englund Dimitrova, B. & Tiselius, E. (2014). Retrospection in interpreting and translation: Explaining the process? MonTI Special Issue 11, 177–200.
Ericsson, K. A. (2000/2001). Expertise in interpreting: An expert-performance perspective. Interpreting 5 (2), 187–220.
Ericsson, K. A. & Simon, H. A. (1993). Protocol analysis: Verbal reports as data (Rev. ed.). Cambridge, MA: MIT Press.
(1998). How to study thinking in everyday life: Contrasting think-aloud protocols with descriptions and explanations of thinking. Mind, Culture and Activity 5 (3), 178–186.
Eyckmans, J., Anckaert, P. & Segers, W. (2016). Translation and interpretation skills. In D. Tsagari & J. Banerjee (Eds.), Handbook of second language assessment. Berlin: Walter de Gruyter, 219–235.
Garzone, G. (2003). Reliability of quality criteria evaluation in survey research. In A. Collados Aís, M. M. Fernández Sánchez & D. Gile (Eds.), La evaluación de la calidad en interpretación: Investigación. Granada: Comares.
Green, A. (1998). Verbal protocol analysis in language testing research: A handbook. Cambridge: Cambridge University Press.
Han, C. (2015). Investigating rater severity/leniency in interpreter performance testing: A multifaceted Rasch measurement approach. Interpreting 17 (2), 255–283.
Iglesias Fernández, E. (2013). Unpacking delivery criteria in interpreting quality assessment. In D. Tsagari & R. van Deemter (Eds.), Assessment issues in language translation and interpreting. Frankfurt am Main: Peter Lang, 51–66.
Ivanova, A. (2000). The use of retrospection in research on simultaneous interpreting. In S. Tirkkonen-Condit & R. Jääskeläinen (Eds.), Tapping and mapping the processes of translation and interpreting: Outlooks on empirical research. Amsterdam: John Benjamins, 27–52.
Jakobsen, A. L. (2003). Effects of think aloud on translation speed, revision, and segmentation. In F. Alves (Ed.), Triangulating translation: Perspectives in process-oriented research. Amsterdam: John Benjamins, 69–95.
Jarvella, R. J., Jensen, A., Jensen, E. H. & Anderson, M. S. (2002). Towards characterizing translator expertise, knowledge and know-how: Some findings using TAPs and experimental methods. In A. Riccardi (Ed.), Translation studies: Perspectives on an emerging discipline. Cambridge: Cambridge University Press, 172–197.
Jourdenais, R. (2001). Cognition, instruction and protocol analysis. In P. Robinson (Ed.), Cognition and second language instruction. Cambridge: Cambridge University Press, 354–375.
Künzli, A. (2007). Translation revision: A study of the performance of ten professional translators revising a legal text. In Y. Gambier, M. Shlesinger & R. Stolze (Eds.), Doubts and directions in translation studies. Amsterdam: John Benjamins, 115–126.
Lee, J. (2009). Toward more reliable assessment of interpreting performance. In S. Hale, U. Ozolins & L. Stern (Eds.), The Critical Link 5: Quality in interpreting – a shared responsibility. Amsterdam: John Benjamins, 171–185.
Lee, S.-B. (2014). An interpreting self-efficacy (ISE) scale for undergraduate students majoring in consecutive interpreting: Construction and preliminary validation. The Interpreter and Translator Trainer 8 (2), 183–203.
Li, D. (2004). Trustworthiness of think-aloud protocols in the study of translation processes. International Journal of Applied Linguistics 14 (3), 301–313.
Liu, M. (2013). Design and analysis of Taiwan’s interpretation certification examination. In D. Tsagari & R. van Deemter (Eds.), Assessment issues in language translation and interpreting. Frankfurt am Main: Peter Lang, 163–178.
(2015). Assessment. In F. Pöchhacker (Ed.), Routledge encyclopedia of interpreting studies. London: Routledge, 20–22.
Lörscher, W. (2005). The translation process: Methods and problems of its investigation. Meta 50 (2), 597–608.
Miles, M. B., Huberman, A. M. & Saldaña, J. (2014). Qualitative data analysis: A methods sourcebook (3rd ed.). Thousand Oaks, CA: Sage.
Naumenko, O. (2015). Improving performance assessment score validation practices: An instructional module on generalizability theory. Working Papers on Language and Diversity in Education I (1), 1–17.
O’Hagan, S. (2014). Variability in assessor responses to undergraduate essays: An issue for assessment quality in higher education. Bern: Peter Lang.
Pöchhacker, F. (2001). Quality assessment in conference and community interpreting. Meta 46 (2), 410–425.
(2011). Assessing aptitude for interpreting: The SynCloze test. Interpreting 13 (1), 106–120.
Pradas Macías, M. (2006). Probing quality criteria in simultaneous interpreting: The role of silent pauses in fluency. Interpreting 8 (1), 25–43.
Pressley, M. & Afflerbach, P. (1995). Verbal protocols of reading: The nature of constructively responsive reading. Hillsdale, NJ: Lawrence Erlbaum.
Rallis, S. F. & Rossman, G. B. (2003). Mixed methods in evaluation contexts: A pragmatic framework. In A. Tashakkori & C. Teddlie (Eds.), Handbook of mixed methods in social & behavioral research. Thousand Oaks, CA: Sage, 491–512.
Russell, D. & Winston, B. (2014). Tapping into the interpreting process: Using participant reports to inform the interpreting process in educational settings. Translation & Interpreting 6 (1), 102–127.
Saldaña, J. (2016). The coding manual for qualitative researchers (3rd ed.). Thousand Oaks, CA: Sage.
Saldanha, G. & O’Brien, S. (2013). Research methodologies in translation studies. London: Routledge.
Sawyer, D. B. (2004). Fundamental aspects of interpreter education: Curriculum and assessment. Amsterdam: John Benjamins.
Sun, S. (2011). Think-aloud-based translation process research: Some methodological considerations. Meta 56 (4), 928–951.
Tiselius, E. (2009). Revisiting Carroll’s scales. In C. Angelelli & H. E. Jacobson (Eds.), Testing and assessment in translation and interpreting studies: A call for dialogue between research and practice. Amsterdam: John Benjamins.
Van Someren, M. W., Barnard, Y. F. & Sandberg, J. A. C. (1994). The think-aloud method: A practical guide to modeling cognitive processes. London: Academic Press.
Vik-Tuovinen, G. V. (2002). Retrospection as a method of studying the process of simultaneous interpreting. In G. Garzone & M. Viezzi (Eds.), Interpreting in the 21st century: Challenges and opportunities. Amsterdam: John Benjamins, 63–72.
Wallace, M. (2013). Rethinking bifurcated testing models in the court interpreter certification process. In D. Tsagari & R. van Deemter (Eds.), Assessment issues in language translation and interpreting. Frankfurt am Main: Peter Lang, 67–83.
Wang, J., Napier, J., Goswell, D. & Carmichael, A. (2015). The design and application of rubrics to assess signed language interpreting performance. The Interpreter and Translator Trainer 9 (1), 83–103.
Wigglesworth, G. (2005). Current approaches to researching second language learner processes. Annual Review of Applied Linguistics 251, 90–111.
Cited by (13)
Cited by 13 other publications
Chen, Shirong & Chao Han
Jiang, Zhaokun & Ziyin Zhang
Achkasov, Andrei Valentinovich
Han, Chao, Binghan Zheng, Mingqing Xie & Shirong Chen
Moratto, Riccardo & Zhimiao Yang
2024. Probing the cognitive load of consecutive interpreters. Translation and Interpreting Studies 19:2 ► pp. 234 ff.
Lu, Rong, Muhammad Alif Redzuan Abdullah & Lay Hoon Ang
Lu, Xiaolei & Chao Han
2023. Automatic assessment of spoken-language interpreting based on machine-translation evaluation metrics. Interpreting. International Journal of Research and Practice in Interpreting 25:1 ► pp. 109 ff.
Chen, Jing, Huabo Yang & Chao Han
Yang, Yuan & Xiangdong Li
Han, Chao, Rui Xiao & Wei Su
2021. Assessing the fidelity of consecutive interpreting. Interpreting. International Journal of Research and Practice in Interpreting 23:2 ► pp. 245 ff.
Wang, Weiwei
Abdel Latif, Muhammad M. M.
This list is based on CrossRef data as of 12 december 2025. Please note that it may not be complete. Sources presented here have been supplied by the respective publishers. Any errors therein should be reported to them.
