Coming To Grips With
Coming To Grips With
n g l i s h
e a c h i n g
o r u m
Number
If we assume that a well-planned course should measure the extent to which students have fulfilled course objectives, then progress tests are a central part of the learning
process. Other reasons for testing can be identified:
Theoretical Considerations
37
The second important issue is whether there is a discrepancy between teaching and testing. Weir (1990:14)
has pointed out that the only difference between teaching and testing within the communicative paradigm relates to the amount of help that is available to the student
from the teacher or his/her peers. Still there are some constraints that the process of testing imposes, such as time,
anxiety, grading, and competition. But, on the whole, we
agree with Davies (1968:5) when he says that a good test
is an obedient servant since it follows and apes teaching.
Our tests should be based on the classroom experience
in terms of syllabus, activities and criteria of assessment.
Their final aim is to measure the language that students
have learned or acquired in the classroom both receptively and productively. We could conclude by saying that
the more our tests resemble the classroom, the more valid
they will be.
The theoretical requisites that a test must achieve are
validity, reliability, and practicality.
A test is valid if it measures what you want it to
measure.
Construct validity refers to the concomitance between
the test and the underlying teaching principles. It follows
from this that tests should reflect the objectives of the
course and underlie its teaching principles. As regards communicative testing, it is crucial that tests be as direct and
authentic as possible; they should relate to real life and real
communicative tasks.
A progress test has content validity if it measures
the contents of the syllabus and the skills specified in the
coursebook. Hence, we should take into consideration the
learners needs and their particular domain of use to ensure content validity. Success with regard to this aspect is
quite easy to achieve since the coursebook designer has decided on the course content. The task of the test writerthe
teacheris to sample this domain, measure it, score it, set
up pass/fail cutoffs, and give grades.
If a test is appealing to laymenstudents, administrators, etc.it has face validity. In other words, tests should
be based on the contents of the textbook and the methodological teaching approaches, as well as measuring what it
is supposed to measure.
Tests are reliable if their results are consistent, i.e., if administered to the same students on another occasion, they
would obtain the same results. There are two main sources
of reliability: the consistency of performance from candidates and scoring.
Finally, a test has practicality if it does not involve
much time or money in their construction, implementation, and scoring.
38
2012
Planning Stage
Development Stage
In this stage we start the process of test design. I propose the following guidelines for their construction:
Number
n g l i s h
e a c h i n g
o r u m
n g l i s h
e a c h i n g
o r u m
Number
cedure in which the answers are either correct or incorrect, mainly used for testing the receptive skills. The latter is subjective and used for the productive skills. One
way of making subjective, impressionist judgements more
objective is to devise a marking scheme through bands
and scales in which the judging criteria is described as
precisely as possible. These bands should be made as
simple and intelligible as possible (e.g., fluency, range of
vocabulary, accuracy, appropriateness, etc.) so that scorers
will not have to take into account too many aspects at the
same time.
8. Analyse the test statistically. Basic statistics are more
straightforward than we imagine. Calculate the reliability coefficientKuder-Richardsonand the difficulty
and discrimination coefficients. The first mathematical
operation tells you how reliable a test is; the other two
measures show if the items are at the right level of difficulty and how well they discriminate. These mathematical operations are simple enough to be carried out in a
manual calculator, and they can indicate the validity of
the test and the performance of the examinee.
9. Consider the pedagogical effects that the test may have
on teaching. Morrow (1986) stated that the most important validity of a test was that which would measure how far the intended washback effect was actually
realized.
If we want our test to influence teaching and learning,
we should ask our students and ourselves the following
questions:
What do students think about the fairness of the
test?
What poor results are due to poor item construction? How could the items be improved?
What poor results are due to poor or insufficient
teaching?
What poor results are due to the coursebook or
other materials?
What areas of weakness in student performance
have we detected for remedial work?
Can we make any assumptions on the relation between teaching and learning?
What changes should be implemented in our
classroom as a result of the test feedback?
10. Present the test and feedback results to the students
with the aim of reviewing and revising the teaching
of content or skills in which the test has shown students to be weak. Teachers should listen to what students have to say about the test and profit from their
comments.
2012
39
Conclusion
References
Alderson, J. C. 1981. Report of the discussion on communicative language testing. In Issues in Language Testing.
ELT Docs, III. ed. J.C. Alderson, and A. Hughes London: The British Council.
. 1990. Bands and scores. In Language testing in the
Nineties: The communicative legacy, ed. J. C. Alderson and
B. North. Oxford: Modern English Publications.
Anastasi, A. 1982. Psychological testing. London: Macmillan.
Carroll, B. and P. J. Hall. 1985. Make your own tests: A
practical guide to writing language performance tests.
New York: Pergamon.
Davies, A. 1968. Language testing symposium: A psycholinguistic perspective. Oxford: Oxford University Press.
Morrow, K. 1986. The evaluation of tests of communicative
performance. In Innovations in Language Testing, ed. M.
Portal. London: Nfer Nelson.
40
2012
Picket, D. 1984, cited by P. Dore. 1991. Authenticity in foreign language testing. In Current Developments in Language Testing, ed. S. Anivan. Singapore: SEAMO Anthology Series.
Pollit, A. 1990. Giving students a sporting chance: Assessment by counting and by judging. In Language Testing
in the Nineties, ed. J. C. Alderson and B. North. Oxford:
Modern English Publications.
Porter, D. 1990. Affective factors in language testing. In
Language Testing in the Nineties, ed. J. C. Alderson and B.
North. Oxford: Modern English Publications.
Spolsky, B. 1968. Language testing: The problem of validation. TESOL Quarterly, 2, 2.
Stevenson, D. K. and U. Riew. 1981. Teachers attitudes towards language tests and testing. In Occasional Papers,
29: Practice and problems in language testing, ed. by T.
Culhane, C. Klein-Braley, and D. K. Stevenson. Department of Language and Linguistics, University of Essex.
Walter, C. and I. McGrath. 1979. Testing: What you need to
know. In Teacher Training, ed. S. Holden. Oxford: Modern English Publications.
Weir, C. 1988. Communicative language testing, with special
reference to English as a foreign language. Exeter University: Exeter Linguistic Series, 1.
. 1993. Understanding and developing language tests.
Hemel Hemstead: Prentice-Hall International.
Number
n g l i s h
e a c h i n g
o r u m