Home Page   About DIRECTIONS   Editorial Board   Contact Us   For Authors    

Current Issue

 
 

 

 

Bessie Mitsikopoulou
EVALUATION CRITERIA FOR THE KPG WRITING TEST IN ENGLISH

This article is concerned with the evaluation criteria for the writing test paper of the English exam, i.e., Module 2. The philosophy of this test paper is based on a functional view of language and a genre approach to writing assessment. Genre-based approaches emphasize the social constructedness of language and acknowledge that, while language is produced by individuals, its shape and structure are to a great extent socially determined. Taking into account that in everyday life we write as members of specific communities, producing texts which conform to different social rules (influenced by a variety of contextual factors, such as who is writing what to whom and for what purpose), similarly in assessment conditions, we ask our candidates to produce scripts which take into consideration specific contextual factors. In fact, the rubrics (instructions) of the writing activities require candidates to assume a role in order to produce a text of a particular text type, addressing a specific audience and meeting a predefined communicative purpose; that is, they determine to a great extent the content, text organization and the language to be used in a candidate’s script. 

Consequently, when evaluating scripts, KPG raters are trained to assess candidates’ ability to produce language which is appropriate for the situational context rather than simply correct in terms of form.

A second rule of the KPG writing evaluation system is to have achievement as the starting point rather than failure. In other words, raters are trained to consider what candidates have managed to accomplish rather than what they have failed to do. In fact the KPG rater guide and the rating grid serve as tools to help raters focus on candidates’ communicative performance rather than the grammatical, lexical, spelling or punctuation errors they might have made when writing. In assessing performance, they are guided to consider what candidates have done, how well they have responded to the writing activity and the degree to which they have used structures and forms appropriately, given the context which is always specified.

Evaluation criteria

There are three evaluation criteria which aim at helping raters focus not merely on sentence grammar and lexis, but on discourse and text as well as sentence grammar. 

Evaluation criterion 1

The first evaluation criterion has to do with task completion and is directly related to contextual features, i.e. the communicative purpose of the produced script, its appropriateness in terms of genre, register and style.

The notion of genre is used here to refer to both text type and generic process. A genre as text type (e.g. new report, email, recipe, film review) is characterised by relatively stable structural forms (e.g. particular beginnings, middle and ends), particular ways of organizing information (e.g. in paragraphs or in bullet forms) and lexicogrammatical features and patterns used according to the social purpose of the text. However, texts are not only determined by an overall social purpose but they are also formed out of the dynamics of social processes, such as to instruct, to argue, to narrate, to describe or to explain. Each one of these processes is associated with different language features; for example, a newspaper article whose purpose is to report a racist event employs different language features from an article whose purpose is to argue against racism. Raters are trained to assess the degree to which a script has addressed the task set and has developed it in terms of theme/topic. They also consider whether candidates have produced an appropriate text type and responded to the required generic processes.

For example, let us say that the writing activity requires candidates to produce a report (text type) which informs readers (generic processes) about the work of a volunteer programme (theme) and that instead of producing a report some candidates produced a letter. This would be partially ok, if these candidates managed at least to inform readers about the volunteer programme’s work. However, if instead of informing about the work that the programme involves candidates provide information about the advantages of joining the programme, they may have failed to meet Criterion 1, especially because it is likely that they have not used appropriate register and style, which are additional requirements of this criterion –depending of course of the level of language proficiency being tested.

Evaluation criterion 2  

Criterion 2 is related to text grammar (text organization, coherence and cohesion). The notion of text grammar as understood here addresses issues above the sentence level. Raters are trained to assess the degree to which candidates have managed to produce a coherent and cohesive script. Coherence refers to the presentation of ideas in a logical and understandable way. Candidates are expected to produce coherent texts by drawing on knowledge of how to organize and present their ideas from their previous experience as text producers and from their experience as readers. For instance, they know that events in a story are presented in chronological order, while arguments in an essay are often presented in terms of their importance (starting from the less important and moving to the most important arguments, or the opposite). 

Candidates are also expected to produce cohesive texts. Cohesion refers to the ways a part of a text is linked to another part of the text and it can be achieved in a script through a variety of ways: through the use of connectives (linking words and expressions), pronoun reference, repetition of key words, etc:

Connectives indicate how an idea presented in one sentence relates to the next one (e.g. in an antithetical way through the use of connectives such as but, on the other hand, however, etc). The use of connectives is an indicator of writing development: the more advanced a candidate’s level of language competence, the more complex and logical connectives that s/he uses for the construction of complex sentences. However, some candidates make repeated use of some formal connectives (e.g. in addition, furthermore, to conclude) which are to be found in a formal essay, but which may be inappropriately used in various other genres, as well. 

Reference is another way through which cohesion may be achieved. The use of personal pronouns is the most common way of maintaining reference which avoids the repetition of names. Raters are trained to take into account the fact that control of reference is an indicator of how well the flow of information from one sentence to the next or from one part of a script to another is maintained.

Tense consistency refers to appropriate use and control of tenses in a script. The use of tense changes from genre to genre. For example, factual descriptions are generally written in the present tense, while narratives in the past tense. Raters are trained to assess a script’s use and control of tense and possible changes in the same script.     

Overall, the criterion of text grammar considers how all parts of a script are structured, organized and coded, in order to make it effective for the purposes of a particular communicative context.

Evaluation Criterion 3

Criterion 3 is related to sentence grammar and lexical features. Sentence grammar refers to the use of language or lexicogrammar according to formal rules of grammar, syntax and morphology. Raters are trained to assess candidates’ writing performance considering the features of their sentences; e.g., correctness of clause pattern, subject-verb agreement, verb form, preposition, articles, plurals, etc. Errors are expected in varying degrees, depending on the level. However, raters are trained not to seriously penalize errors as long as they do not interfere with intelligibility. The importance of errors which violate rules of formal grammar are assessed on the basis of whether a script manages to convey a socially meaningful message, despite formal errors.

Concerning lexical features, different text types require use of different types of vocabulary as well as a different range of vocabulary (or repetition), depending on determining categories such as topic, purpose and audience. For example, an academic report will use a range of technical vocabulary including nominalizations and technical noun groups; a literary description, on the other hand, will use descriptive verbs, adjectives and adverbs, and affective language in order to create an emotive effect on the reader. Raters should assess vocabulary appropriacy in terms of a specific text type and lexical range (for more advanced levels). Spelling errors are considered differently, depending on whether or not they interfere with intelligibility and depending on the level.

Punctuation and writing conventions are also assessed when considering Criterion 3.

Applying the evaluation criteria

The rating process is done on the basis of a rating grid which guides raters to follow a procedure which moves from a holistic view (i.e. overall impression) to finer points of assessment. Though the grid for each level is different, all grids have been designed on the basis of the same philosophy and therefore, they can be applied by using the same methodology. The idea behind it is to have a zone based assessment rather than to subtract points on the basis of how many errors a script has. The decision the rater has to make, by moving from right to left on the grid and from left to right, is to decide whether the script is fully satisfactory (responding to all three evaluation criteria), moderately satisfactory (satisfying some of the criteria or satisfying them partially) or unsatisfactory (partly responding to a limited number of criteria or points of the criteria).

The application of the rating grid is generally a demanding process. In order to ensure reliable assessment and marking, the system needs to make sure that during the script rating process, raters use the rating grid systematically and correctly. In order to achieve this goal the system takes a variety of measures. The two most important ones are:

1.   To produce material that provides very concrete guidelines as to how to assess and mark KPG writing scripts. It is for this reason that for every exam period a very detailed Rater Guide is produced and given (free of charge) to KPG raters. This includes analyses of the evaluation criteria, the rating grids and guidelines regarding how to mark the scripts of the particular exam period with articulated expected outcomes and real sample scripts which have already been marked. To help further, a Handbook for this purpose is being prepared. It is an RCeL publication, to appear within the RCeL publication series, of which the editors are Bessie Dendrinos & Kia Karavas. The Handbook (edited by myself) is entitled The KPG Writing Test in English.

2.  To systematically work with raters. In the last issue of the ELT News, the “KPG Corner” presented the KPG script rating programme which runs every exam period before script assessment begins. The overall aim of this programme is to train raters to be able to reliably mark the scripts which have resulted from the writing activities of that period’s exams. In order to enhance assessment reliability, KPG scripts are evaluated and marked by two raters, and the final mark of each script is the average of their marks. If there is high discrepancy between them, scripts are re-evaluated, and the incident a cause for concern. The raters involved are informed. Of course, to avoid such incidents, raters are consistently supervised by highly qualified coordinators who are also assigned with the responsibility of evaluating the raters for intra- and inter-rater reliability.

 [Back]

   
 
 


Contents

Forthcoming Issue
Current Issue
Back Issues

Call for Papers

 

©2009-2011  ÔìÞìá ÁããëéêÞò Ãëþóóáò êáé Öéëïëïãßáò - ÅÊÐÁ  /  Faculty of English Language and Literature - UOA

Developed By A.Sarafantoni
Designed By
C.Frantzeskaki