Shiva Kaivanpanah; Mohammad Alavi; Rose Meschi
Abstract
Writing is thought as the most complicated skill in second language acquisition; therefore, L2 researchers have always been in pursuit of discovering an effective approach to improve it. One of the most debated ways is feedback which has a key role in improving the quality of writing. Much of the previous ...
Read More
Writing is thought as the most complicated skill in second language acquisition; therefore, L2 researchers have always been in pursuit of discovering an effective approach to improve it. One of the most debated ways is feedback which has a key role in improving the quality of writing. Much of the previous research on feedback has focused on analyzing different types of feedback and their effect on the learners’ writings and few studies have examined the effectiveness of computer feedback. Therefore, the present study was conducted to 1) determine what aspects of students’ writings receive computer feedback, 2) examine the difference in the effect of computer-generatedfeedback (CBF) and Teacher-based feedback (TBF) on improving the students’ writing quality and 3) compare the differences in Depth of Processing (DOP) in processing computer and teacher feedback. The results indicated that content, style and organization of their essays received feedback from the teacher and the computer. Teacher feedback was more effective in terms of its impact on improving the quality of the writing of the students than computer-generatedfeedback and it resulted in deeper processing of lexical items, whereas computer-generated feedback invoked medium processing on grammar.
Mohammad Alavi; Shiva Kaivanpanah; Fatemeh Danesh
Volume 11, Issue 23 , June 2019, , Pages 1-25
Abstract
Assessment of writing skill is generally believed to be judged by a rater subjectively and qualitatively or by using analytic scoring rubrics which can potentially result in somehow not very reliable assessment. It seems that an evaluation of writing based on a model can result in a valid and reliable ...
Read More
Assessment of writing skill is generally believed to be judged by a rater subjectively and qualitatively or by using analytic scoring rubrics which can potentially result in somehow not very reliable assessment. It seems that an evaluation of writing based on a model can result in a valid and reliable writing assessment. To achieve such an objective, this study firstly aimed to develop an assessment model based on Activity Theory (AT), i. e., Activity Theory-Based Assessment Model (ATBAM), and then to employ it in the assessment of writing performances of Iranian language learners in a private language college. And finally, to achieve the concurrent validity of ATBAM, its results were compared with those of a traditional approach. Three groups of participants took part in this study: a group of upper intermediate English learners (N=29) who submitted one writing sample per week in four successive weeks, teachers (N=6) who provided learners with feedback and assigned holistic scores and course supervisors (N=2) who reassessed the writing samples on the basis of an analytic rubric of writing assessment. The results showed that using ATBAM in writing assessment results in an exploration of not only learners’ but also teachers’ engagement in the development of learners’ writing ability. The role of teachers’ feedback and teachers’ and learners’ social interaction in the development of learners’ writing ability could potentially provide comprehensive, fair, reliable, and valid scores in this model.
Mohammad Alavi; Fatemeh Ranjbaran
Volume 10, Issue 21 , June 2018, , Pages 1-31
Abstract
Of paramount importance in the study of cognitive diagnostic assessment (CDA) is the absence of tests developed for small-scale diagnostic purposes. Currently, much of the research carried out has been mainly on large-scale tests, e.g., TOEFL, MELAB, IELTS, etc. Even so, formative language assessment ...
Read More
Of paramount importance in the study of cognitive diagnostic assessment (CDA) is the absence of tests developed for small-scale diagnostic purposes. Currently, much of the research carried out has been mainly on large-scale tests, e.g., TOEFL, MELAB, IELTS, etc. Even so, formative language assessment with a focus on informing instruction and engaging in identification of student’s strengths and weaknesses to guide instruction has not been conducted in the Iranian English language learning context. In an attempt to respond to the call for developing diagnostic tests, this study explored developing a cognitive diagnostic reading comprehension test for CDA purposes. To achieve this, initially, a list of reading attributes was prepared based on the literature and then the attributes were used to construct 20 reading comprehension items. Then seven content raters were asked to identify the attributes of each item of the test. To obtain quantitative data for Q-matrix construction, the test battery was administered to 1986 students of a General English Language Course at the University of Tehran, Iran. In addition, 13 students were recruited to participate in think-aloud verbal protocols. On the basis of the overall agreement of the content raters’ judgements concerning the choices of attributes and results of think-aloud verbal protocol analysis, a Q-matrix that specified the relationships between test items and target attributes was developed. Finally, to examine the CDA of the test, the Fusion Model, a type of cognitive diagnostic model (CDM), was used for diagnosing the participants' strengths and weaknesses. Results suggest that nine major reading attributes are involved in these reading comprehension test items. The results obtained from such cognitive diagnostic analyses could be beneficial for both teachers and curriculum developers to prepare instructional materials that target specific weaknesses and inform them of the more problematic areas to focus on in class in order to plan for better instruction.
Seyyed Mohammad Alavi; Abbas Ali Rezaee; Seyyed Mohammad Reza Amirian
Volume 3, Issue 7 , December 2011, , Pages 39-65
Abstract
The purpose of this study was to detect differentially functioning items in the University of Tehran English Proficiency Test (UTEPT) which is a high stake test of English developed and administered by the Language Testing Centre of the University of Tehran. This paper is based on the answers of 400 ...
Read More
The purpose of this study was to detect differentially functioning items in the University of Tehran English Proficiency Test (UTEPT) which is a high stake test of English developed and administered by the Language Testing Centre of the University of Tehran. This paper is based on the answers of 400 test takers to the test. All participants earned a master degree either in humanities or science and engineering. To achieve the purpose of this study, the participants were divided into two equal groups. The results of generalized Mantel-Haenszel indicated that out of 100 items 12 items were displaying DIF. Logistic regression procedure also flagged 14 items as exhibiting DIF. Nevertheless, the associated test of effect size for logistic regression showed that none of these effect sizes were large according to the guidelines proposed by both Zumbo and Thomas (1997) and Jodoin and Gierl (2001). Therefore, it was concluded that UTEPT doesn't show significant academic discipline DIF and is equally fair to both humanities and science and engineering groups.