0 ratings 0% found this document useful (0 votes) 214 views 14 pages Harmer Testing
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, 
claim it here .
Available Formats
Download as PDF or read online on Scribd
Go to previous items Go to next items 
Save Harmer+Testing For Later students
A The characteristics of tests
A1 Different types of test
‘There are four main reasons for testing which give rise to four categories of tests
© Placement tests: placing new students in the right lass in a school is
facilitated with the use of placement tests, Usually based on syllabuses and
materials the stuckents will follow and use once thet level has been decided om,
these test grammar and vocabulary knowledge and assess students’ productive
and receptive skills
Some schools ask students to assess themselves as part of the placement
process, adding this self-analysis into the final placing decision.
© Diagnostic tests: while placement tests are designed to show how good a
student's English is in relation to a previously agreed system of level, diagnostic
tests can be used to expose learner difficulties, gaps in their knowledge, and skill
deficiencies during a course. Thus, when we know what the problems are, we
can do something about them.
© Progress or achievement tests: these tests are designed to measure learners’
language and skill progress in relation to the syllabus they have been following,
‘Achievement tests only work if they contain item types which the students
are familiar with. This does not mean that in a reading test, for example, we give
them texts they have seen before, but it does mean providing them with similar
texts and familia task types. If students are faced with completely new material,
the test will not measure the learning that has been taking place, even though it
can still measure general language proficiency.
Achievement tests at the end of aterm (like progress tests at the end of unit,
a fortnight, et.) should reflect progress, not fafure. They should reinforce the
‘earning that has taken place, not go out of their way to expose weaknesses. They
cant also help us to decide on changes to futute teaching programmes where
students do significantly worse in (paris of) the test than we might have expected,
© Proficiency tests: proficiency tests give a general picture ofa student's
knowledge and ability (rather than measure progress). They ae frequently used
8 stages people have to reach if they want to be admitted to a foreign
university, geta job, ar obtain some kind of certificate.
32a2
Bt
cHaPren 25
Proficiency tests have a profound backwash effect since, where they are
‘external exams, students obviously want to pass them, and teachers’ reputations
sometimes depend (probably unfairly} upon how many of them sucveed,
Characteristics of a good test
In order to judge the effectiveness of any test it is sensible to lay down criteria against
‘which the test can be measured, as follows:
* Validity: a testis valid if it tests what itis supposed to test. Thus itis not valid,
for example, to test writing ability with an essay question that requircs specialist
knowledge of history or biology — unless it is known that all students share this
rrowledge before they do the test.
A particular kind of validity‘ that concerns most test designers i face
validity. This means that the test should look, on the ‘face’ of it, asf it is valid.
A test which consisted of only three multiple choice items would not
convince students ofits face validity however reliable or practical teachers
thought it tobe,
‘© Reliability: 2 good test should give consistent results. For example, if the same
‘group of students took the same test twice within two days — without reflecting
‘on the first test before they sat it again they should get the same results on
each occasion. If they took another similar test, the resuts should be consistent.
If two groups who were demonstrably alike took the test, the marking range
‘would be the same.
In practice, ‘reliability is enhanced by making the test instructions
absolutely clear, restricting the scope for variety in the answers,and making
‘sure that test conditions remain constant.
Reliability also depends on the people who mark the tests the scorers
(Cleary a testis unreliable if the result depends to any large extent on who is
marking it, Much thought has gone into making the scoring of tests as reliable
as possible (see c2 below).
 
   
Types of test item
‘Whatever purpose atest or exam has, major factor in its success or failure as a
good measuring instrument willbe determined by the item types tht it contains.
Direct and indirect test items
‘Atest item is direct if i asks candidates to perform the communicative skill which
is being tested. Indirect test items, on the other hand, try to measure a student's
knowledge and ability by getting a what lies beneath ther receptive and productive
Skis, Whereas direct test items try to be as much like real-life language use as posible,
indirect tems try to find out about a student’ language knowledge through more
controlied items, such as multiple choice questions or grammar transformation
items. These ace often quicker to design and, crucially, easier to mark, and produce
greater scorer reliability
aB2
Another:
 
listinction needs to be made between discrete-point testing and
 
integrative testing, Whereas discrete-point testing only tests one thing ata time
(such as asking students to choose the correct tense of a verb), integrative test items
expect students to use a variety of language at any one given time —as they will have
to do when writing a composition or dois
In many proficiency tests where students sit a numberof different papers, there isa
a conversational oral test.
  
riinture of direct and indirect, discrete point and integrative testing, Test designers find
that this combination gives a good overall picture of student ability Placement tests often
‘use discrete-point testing to measure students against an existing language syllabus, but
-may then compare this with more direct and integrative tasks to get a fuller pictue.
Indirect test item types
Although there is a wide range of indirect test possibilities certain types are in
 
 
common use:
¢ Multiple choice questions (MCQs): traditional vocabulary multiple choice
question looks like this:
The journalist was by enemy fire as he tried to send a
story by radio.
dayprsed —b wounded —¢ injured damaged a
For many years MOQs were considered to be ideal test instruments for
‘measuring students’ knowledge of grammar and vocabulary. Above allthis was
because they wete easy to mark, and since the advent of computers the answet
books for these tests can be read by machines, not people, thereby cutting out
the possibility of scorer error.
‘However there area number of problems with multiple choice questions. In
the frst place, they are extremely difficult to write well, especially in the design
of the incorrect choices, These distractors’ may actually put ideas into students!
hheads that they did not have before they read them. Second, while i is possible
to train students so that their MCQ abilities are enhanced, this may not actually
improve their English, ‘The difference berween two student scores may be
between the person who has been trained in the technique and a person who
has not, rather than being a difference of language knowledge and ability.
-MCQs are sill widely used, but though they score highly in terms of
practicality and scorer reliability their‘validity’ and overall ‘reliability’ are suspect.
Cloze procedures: cloze procedures seem to offer us the ideal indirect but
integrative testing item. They can be prepared quickly and ifthe claims made
for them are true, they are an extremely enst-effective way of finding out about
a testee’s overall knowledge.
‘Cloze, in its purest form, is the deletion of every nth word in a text
(somewhere between every fifth or tenth word), Because the procedure is
random, it avoids test designer failings. It produces test items lke this:
323‘cuaeren 25
‘They sat on a bench attached 1___a picnic table. Below them
they 2 see the river gurgling between overgrown
3 . The sky was diamond blue, with 4 white clouds
dancing in the freshening 5 . They could hear the call of
6 and the buzzing of countless insects. 7 were
completely atone,
cms
 
 
 
 
 
 
 
 
Cloze testing seems, on the face of it ike a perfect test instrument, since, because
of the randomness ofthe deleted words, anything miay be tested (e.g. grammar,
collocation, fixed phrases, reading comprehension), and therefore it becomes
‘more integrative in its reach. However it turns out that the actual score a student
gets depends on the particular words that are deleted, rather than on any general
English knowledge. Some are more difficult to supply than others, and in some
cases there ae several possible answers, Even in the short sample text above itis
ear that whilst there is no doubt about items such as | and 8, for example item
4 is less predictable, Different passages produce different rests.
Despite such problems of ‘reliability, cloze is too useful a technique to
abandon altogether, however, because itis clear that supplying the correct word
for a blank does imply an understanding of context and a knowledge of that
‘word and how it operates, Perhaps it would be better, therefore, to use rational”
or modified’ cloze procedures (Alderson 1996: 222) where the test designer can.
‘be sure that the deleted words are recoverable from the context. This means
abandoning the completely random nature of traditional cloze procedure.
Instead, every eighth or tenth word is deleted, but the teacher has the option to
delete a word to the left or right if the context makes this more sensible,
“Modified cloze is useful for placement tests since students can be given texts
that they would be expected to cope with at certain levels — thus allowing us 0
Judge their suitability for those levels. They are useful, too, as part of atest
battery in either achievement ot proficiency tests.
‘* Transformation and paraphrase: « common test item asks candidates to
rewrite sentences in a slightly different form, retaining the exact meaning of the
original. For example, the following item tests the candidates’ knowledge of
verb and clause patterns that ave triggered by the use of Fish:
‘Tm sony that I didn’t get her an anniversary present. 2
Twish
In order to complete the item successfully the student has to understand the
first sentence, and then know how to construct an equivalent which i
grammatically possible, As such they do tell us something about the candidates
“Knowledge ofthe language system,
‘+ Sentence re-ordering: getting students to put words in the right order to
‘make appropriate sentences tells us quite @ lot about their underlying
omBS
knowledge of syntax and lexico- grammatical elements. The following example
is typical:
Put the words in order to make correct sentences.
called’/ 1/1 / in / sony / wase't / when / you
Re-ordering exercises are fairly easy to write, though it is not always possible to
censure only one correct order,
‘There are many other indirect techniques too, induding sentence fill-ins (lan__to
the gym every Tuesday morning), choosing the correct tense of verbs in sentences and
passages (J have arvived/arrived yesterday), finding ercors in sentences (She noticed
about his new jacket), and choosing the cottect form of a word (He didn't enjoy being
com the (lose)___ side). Al of these offer items which are quick and effcient to score
and which aim to tell us something about a student's underlying knowledge.
Direct test item types
For direct test items to achieve ‘validity’ and to be ‘reliable’ test designers need to do
the following:
© Create a ‘level playing field’: in the case of @ written test, teachers and
candidates would almost certainly complain about the following essay question:
Why was the discovery of DNA so important for the science ofthe
swontieth century?
since it unfairly favours candidates who have sound scientific knowledge and
presupposes a knowledge of twentieth-century scientific history.
However, the following topic comes close to ensuring that all candidates
have the same chance of success:
‘Some businesses now say that no one can smoke cigarettes in or
‘even near any of their offices, Some governments have banned 2
-smoking in all public places - whether outside or inside.
This is a good idea but it also takes away some of our freedom. ~
Do you agree or disagree?
Give reasons for your answer. rs
‘General writing question from The IELTS Hancdook (ee notes atthe end of
this chapter)
Receptive sill testing also needs to avoid making excessive demands on the
student's general or specialist knowledge. Receptive ability testing can also be
undermined if the means of testing requires students to perform well in writing
‘or speaking (when it sa test of reading or listening). In such a situation we can
tno longer be sue that itis the receptive skill we are measuring.
 
225© Replicate real-life interaction: in real life when people speak or write they
generally do so with some real purpose. Yet traditional writing tests have often
been based exclusively on general essay questions, and speaking tests often
included hypothetical questions about what candidates might say ifthey
happened to be ina certain situation. More modern test writers now include
tasks which attempt to replicate features of real life (Weir 1993: 167). They will
‘often look similar to the kind of speaking activities described in Chapter 19.
Tests of reading and listening should ako, as far as possible, reflect real life.
‘This means that texts should be as realistic as possible, even where they are not
authentic (see Chapter 14, m). Although there are ways of assessing student
‘understanding (using matching tasks or multiple choice questions) which do
not necessarily satisfy these criteria, test items should be as much like real
reading and listening as possible.
The following direct test item types are afew ofthe many which attempt to meet the
criteria we have mentioned above:
SPEAKING
‘an interviewer questioning a candidate about themselves
© ‘information gap’ activities where @ candidate has to find out information
«either from an interlocutor or a fellow candidate. (The role-play on page
279, Example 5, would not need much modification to serve asa test iter.)
‘¢ ‘decision-making? activities, such as showing paired candidates ten
photos of people and asking them to put them in order of the best and
worst dressed
‘ using pictures for candidates to compare and contrast, whether they can
both see them or whether (as in many communication games) they have
found similarities and differences without being able to look at each
cother’s material
+ role-play activities where candidates perform tasks such as introducing
themselves, or ringing a theatre to book tickets
wriTg
+ vevting compositions and stories
+ ‘transactional letters’ where candidates reply toa job advertisement, or
pen a complaint to a hotel based on information given in the exam paper
information leaflets about their school or a place in theic town
© aset of instructions for some common task
© newspaper articles about a recent event
READING
‘» ukiple choice questions to test comprehension of a text
‘© matching written descriptions with pictures of the items, or procedure,
they describe
336ct
 
‘resting STUDENTS
«transferring writen information to charts, graphs, maps, etc (Uhough
special care has to be taken not to disadvantage non-mathematicaly
minded candidates)
‘+ choosing the best summary of a paragraph or a whole text
‘© matching jumbled headings with paragraphs
* inserting sentences provided by the examiner in the correct place fn the
text
LISTENING
«© completing charts with facts and figures from a listening text
© identifying which of a number of objects (pictures on the test paper) is
being described
« identifying which (out of two or three speakers) says what
« identifying whether speakers are enthusiastic, encouraging in
isogreement, or amused
« following directions on a map and identifying the correct house or place
In the interests of ‘reliability, stening tests are most often supplied on tape to ensure
that all candidates have the same opportunities, irrespective of the speakers voices,
speeds, or expressions, Sometimes, asin the computerised TOEFL test (see the notes
atthe end of this chapter) candidates work with headphones from an individual
computer, Where a group of students listen to the same tape or disk, however, we
need to be sure thatthe material i clearly and easly audible (see Chapter 26, a2).
 
 
Writing and marking tests
_Atvarious times during our teaching careers we may have to write tests forthe students
‘we ae taching, and mark the tess they have completed for us. These may range from
lesson test at the end of the week to an achievement test at the end of a term or a year,
Writing tests:
‘Before designing a test and then giving it to a group of students, there are a number
of things we need to do:
‘+ Assess the test situation: before we start to write the test we need to remind
ourselves of the context in which the test takes place. We have to decide how
‘mach time should be given to the test-taking, when and where it willtake place,
and how much time there is for marking.
* Decide what to test: ve have to list what we want to include in our test. This
_means taking a conscious decision to include or exclude skills such as reading
comprehension or speaking (if speaking tests are impractical). t means
knowing what spllabus items can be legitimately included (in an achievement
test), and what kinds of topics and situations are appropriate for our students
Just because we havea list ofall the vocabulary items or grammar points the
students have studied over the term, this does not mean we have to test every
a7‘cuaPTeR 23
single item, Ifwe include a representative sample from across the whole lst, the
students’ success or failure with those items will bea good indicator of how well
they have learnt all of the language they have studied,
© Balance the elements: if we are to include direct and indirect test items we
hhave to make a decision about how many of each we should put in out test.
‘A 200-item multiple choice test with a short real-life writing task tacked onto
the end suggests that we think that MCQs are a better way of finding out about
students than more integrative writing tasks would be.
Balancing elements involves estimating how long we want each section of
the test to take and then writing test items within those time constraints. The
amount of space and time we give to the various clements should also reflect
their importance in our teaching.
© Weight the scores: however well we have balanced the elements in our test,
our perception of our students’ success or failure will depend upon how many
marks are given to each section or sections of the test. If we give two marks for
each of our ten MCQs, but only one mark for each of our ten transformation
items, it means that its more important for students to do well in the former
than in the latter.
© Making the test work: itis absolutely vital that we try out individual items
andJor whole tests on colleagues and students alike before administering them
to real candidates
‘When we write test items the fist thing to do isto get fellow teachers to try
them out. Frequently they spot problems which we are not aware of and/or
come up with possible answers and alternatives that we had not anticipated.
Later, having made changes based on our colleagues’ reactions, we will want
to try out the test on students. We will not do this with the students who are
going to take the test, of course, but if we can find a class that is roughly similar
ora class one level above the proposed test —then we will soon find out what
items cause unnecessary problems, We can also discover how long the test takes.
Such trialling is designed to avoid disaster, and to yicld a whole range of
possible answers/responses to the various test ters. This means that when
other people finally mark the test we can give them a list of possible alternatives
and thus ensure reliable scoring.
C2 Marking tests
‘When Cyril Weir gave copies of the same eight exam scripts to his postgraduate
students they marked them first on the basis of ‘impressionistic’ marking out of a
possible total of 20 marks. The results were alarming, Some scorers gave higher
‘marks overall than others. But for some of the EFL student scripts, the range of
marks was excessive. For one script the lowest mark awarded was 5, whereas another
scorer gave it 20, For another the range was 1-15. As Cyril Weir writes ‘the worst
scripts... if they had been marked by certain markers, might have been given higher
marks than the best scripts!” (993: 157).
ame‘There are a number of solutions to this kind of scorer subjectivity:
© Training: if scorers have seen examples of scripts at various different levels and,
discussed what marks they should be given, then their marking is likely to be
less erratic than ifthey come to the task fresh. If scoters are allowed to watch,
and discuss videoed orl tests, they can be trained to rate the samples of spoken
English accurately and consistently in terms of the pre-defined descriptions of
performance’ (Savile and Hargreaves 1999: 49).
'* More than one scorer: reliability can be greatly enhanced by having more than
‘one scorer. The more people who look ata scrip, the greater the chance that its
{ruc worth will be located somewhere between the various scores itis given, Two
‘examiners watching an oral test are likely to agree a more reliable score than one.
“Many public examination boards use moderators whose job it isto check
samples of individual scorers! work to sce that it conforms with the genctal
standards laid down for the exam,
© Global assessment scales: a way of specifying scores that cat be given to
productive skill work isto create‘pre-defined descriptions of performance’
‘Such descriptions say what students need to be capable of in order to gain the
fequited marks, asin the following assessment (or rating) scale for oral ability:
 
Score Description
0 The candidate is almost unital uses words wrongly. and
& shows no sign of any grammatical understanding,
1 The candidate fs able to transmit only very basi ideas using
Individual words rather than phrases or fuller patterns of
discourse, Speech fs very hesizint and che pronunciation makes
inteligbly dticue.
“The candidate eransmits basic Ideas ina fairly slted way
Pronunciation is sometimes problematic and there are examples
of grammatical and lexical misuse and gaps which impede
communication on occasions.
‘The candidate transmits ideas moderately clearly Speech is.
somewhat hesitant and there are frequent lapses in grammar and
vocabulary use. Nevertheless, the candidate makes hinvherself
understood.
‘The candidate speaks fairly fiently, showing an abiticy «0
‘communicate ideas with nat too much trouble-There are some
problems of grammatical accuracy and some words are a
Inappropriataly used
‘The candidate speaks fluendy wit few obvious mistakes anda wide
variety of lexis and expression. Pronunciation is almost always,
ncellgble, and there is ltue difficulty in communicating ideas,
 
acnarren 33
Global assessment scales are not without problems however: perhaps the
description does not exactly match the student whe is speaking asin a case
(for the scale above) where he or she had very poor pronunciation but was
nevertheless grammatically accurate. There is also the danger that different
teachers will not agree on the meaning of scale descriptors'(Upshur and.
‘Turner 1995: 5). Global assessment, on its own, still falls short of the kind of
reliability we wish to achieve.
© Analytic profiles: macking gets more reliable when a student's performance is
analysed in much greater detail, Instead of just a general assessment, marks are
awarded for different elements,
For oral assessment we can judge a students speaking in a number of
different ways such as pronunciation, fluency, use of lexis and grammar, and
intelligibility. We may want to rate their ability to get themselves out of trouble
(repair skills) and how well they successfully completed the task which we set
‘hem.
‘The resulting analytic profile might end up looking lke this:
   
Score (see analytic scales)
    
 
Pronunciation
  
 
 
 
 
Fluency
Use of vocabulary
Use of grammar
Inteligibitty
Repair skils
Task completion oe ay
For each separate criterion, we can now provide a separate ‘analytic scale as in
the following example for lucncy:
= 2 AE ATT
Score Description
 
0 The candidace cannot get words or phrases out at all
1 ‘The candidate speaks hesiatingy in short interrupted bursts.
2 ‘The candidate speaks slowly with frequent pauses. a
3 Tie candidate speaks ata comfortable speed with quite a lot
of pauses and hesitations.
4 ‘The candidate speaks at a comfortable speed with only an 4
‘oceasional pause or upset.
5 The candace speaks quickly wih ow hesiations. my
a0‘A combination of global and analytic scoring gives us the best chance of reliable
‘marking. However, a profusion of criteria may make the marking of atest
extremely lengthy and cumbersome; test designers and administrators will have
to decide how to accommodate the competing claims of reliability and
practicality.
Scoring and interacting during oral tests: scorer reliability in oral tests is
haclped not only by global assessment scores and analytic profiles but also by
separating the role of scorer (or examiner) from the role of interlocutor (the
examiner who guides and provokes conversation). This may cause practical
ptoblems, but it will allow the scorer to observe and assess, fee from the
responsibility of keeping the interaction with the candidate or candidates
going.
In many tests of speaking, students are now put in pairs or groups for
certain tasks since it is fet that this will ensure genuine interaction, and will
help to relax students in a way that interlocutor-candidate interaction might
fail ( do on its own, However, at least one commentator worries that pairing
students in this way leads candidates to perform below thei level of proficiency,
and that when students with the same mother tongue are paired together their
intelligibility to the examiner may suffer (Foot 1999: 52).
  
D Teaching the test
‘When students are preparing for a public exam or school test it is the teacher's
responsibility not only to help them get their English to the level required, but also
familiarise them with the kinds of exam items they are likely to encounter, and give
them training in how to succeed.
Students can be prepared for furure tests and exams in 2 variety of ways:
‘Training for test types: we can show the various test types and ask them what
the item is testing so that they are clear about what is required. We can help
them to understand what the test or exam designer is aiming for; by showing
‘them the kind of marking scales that ae used, we can make them aware of what
constitutes success, We can then give them training to help them approach such
items more effectively. After they have completed atest item type we can tll
‘them what score an examiner might give and why. We will also equip students
‘with appropriate negotiating language (see Chapter 1 a) to help them get
‘over awkward moments in sich tasks.
‘When training students to handle reading test items we will discuss with
‘them the best way to approach a first reading of the text, and how that can be
‘modified on second reading to allow them to answer the questions provided.
If the test or exam is likely to contain multiple choice questions, we can
help students to appreciate the advantages of finding the obvious distractor(s)
fist, They can then work out what similarities and differences the other
dlistractors have so that they can identify the area of meaning or grammnar that
isbeing targeted.
  
31enarren 23
332
 
Discussing general exam skills: most students benefit from being reminded
about general test and exam skills, without which much of the work they do
‘will be wasted. Such general skills for written tests include studying questions
properly and then reading them again so that they are absolutely sure what
they should do. Students need to check their work over thoroughly before
considering that they have finished. They need to pace themselves so that they
do not spend a disproportionate amount of time on only one part of an exam.
Doing practice tests: students need a chance to practise taking the test or
‘exam so that they get a feel for the experience, especially with regard to issues
stich as pacing, At various points in a course students can sit practice papers, or
Whole practice tests, but this should not be done too often since not only will it
ive teachers horrific marking, schedutes, but it will also be less productive than
‘other test and exam preparation procedures ate,
Promoting autonomy: although we can doa lot of exam preparation in class,
‘we need to impress on students that their chances of success are far greater if
they study and revise om their own. We will want them to read more Hsten
more, work on self-study exercises, and use dictionaries and other means to
build up their language store. All ofthese activities are desirable for any student
(see Chapter 24, 2) but are especially appropriate when an exam is approaching.
‘Having fun: just because students need to practise certain test types does not
‘mean this has to be done in a boring or tense manner. There are a nurnber of
‘ways of having fan with tests and exams.
If typical test item asks candidates to put words in order to make
sentences (see #2 above), the teacher might prepare the class for this kind of
item by giving students a number of words on cards which they have to
physically assemble into sentences. They can hold them above their heads (so
that chey cannot see the words on them} and their classmates have to tell them
‘where to stand to make a human sentence’ Students can play ‘transformation
tennis Where one student ‘serves’ a sentence, €.g India isthe country I would like
10 visit more than any other and the receiving student has to reply with a
transformation starting with The country,e.g, The country E would most like to
visit is India (Prodromou 1995: 22-23). They can change the sex of all the people
in direct and indirect test items to sec ifthe items still work and if not, why not.
Students can be encouraged to write their own test items, based on
language they have been working on and the examples they have seen so far.
“The new test items can now be given to other students to see hove well they have
‘been written and how dificult they are. This helps students to get into the
minds of their test and exam writers.
Ignoring the test: if students who are studying for an exam only ever look at
test types, discuss exam technique, and take practice tests, lessons may become
monotonous. There is also the possibility that general English improvement
will be compromised atthe expense of exam preparation,‘When we are preparing students for an exam, we need to ignore the exam
from time to time so that we have opportunities to work on general language
{ssues, and so that students can take part inthe kind of motivating activities
that are appropriate forall English lessons.
 
Chapter notes and further reading
© Public exams
Among the international exams which students can take are the following:
~ Cambridge exams are offered by the University of Cambridge Local
Examinations Syndicate (UICLES), Cambridge, UK (http://www cambridge-
fhorg.uk), These are exams in general English. There are five main levels:
ey English Test (KET) for elementary candidates
Preliminary English Test (PET) for lower intermediate candidates
First Certificate in Fnglish (FCE) for upper intermediate candidates
Certificate of Advanced English (CAE) for upper intermediate/advanced
candidates,
Certificate of Proficiency in English (CPE) for very advanced candidates
= City and Guilds Pitman qualifications offered by City and Guilds, London
(hitp://www.city-and-guilds.co.uk/Pitman/default htm). Exams are offered in:
‘Communication in Technical English
English for Business Communication
English for Office Skills
ESOL (English for Speakers of Other Languages)
ESOL for Young Learners
Spoken ESOL,
Spoken ESOL for Young Learners
~ IELTS (International English Language Testing System), administered jointly by
UCLES (see above), the British Couneil,and IDP Education, Australia
(hitp://wwwielts.org)
IELTS scores (on a 0-9 band) are used especially by British and Australian
“universities to gauge the level of would-be students or trainers/teachers.
‘There are papers in listening and speaking, Candidates then choose general or
academic writing, and general or academic reading.
— Diploma in English for International Comununication is offered by the Institute of
Linguists, London (hetp://ww.iolorg.uk) for anyone who has reached a degree-
‘equivalent evel of attainment and does not have English as thei ist language,
~ SESOL (Spoken English for Speakers of Other Languages) offered by Trinity
College London (wwew.rinitycollege.co.uk). These are one-to-one interviews
with an examiner at a level to suit the candidate,
333charren 25
TOEFL (Test of English as a Foreign Language) is offered by Educational
“Testing Services, New Jersey, USA (http://wwwtoefl.org)
‘TOEFL scores are used by colleges and universities in North America and
‘elsewhere to measure English proficiency for would-be students. The tests are
now computer-administered in parts ofthe world where thisis possible.
— TOEIC (Test of English for International Communication) is offered by TOEIC
‘Service International, Princeton, New Jersey, USA (http:(/www.toeic.comn),
‘TOEIC scores are used by a number of companies in the USA and elsewhere to
judge the level of English of potential employees.
~ Exams for Business are offered by the University of Cambridge Local
Examinations Syndicate (see above} and by the London Chambet of
‘Commerce and Industry (LOCI) ¢httpe/www.lecieb org uk/qualifications htm),
Testing in general
‘The best and most accessible books on testing are still A Hughes (1989) and
C Weir (1993). See also JC AMerson et al. (1993).
© Cloze procedures
‘A Hughes (1989: 67-70) repeats his earlier claims that cloze passages taken from
real conversation transeripts can be good predictors of oral abilt
© Oral testing
B Knight (1992) describes a fascinating workshop in which teachers investigate
how to measure students’ speaking skills
© Using rating scales
J Upshur and C Turner (1995) suggest replacing descriptive rating scales with
scales where test markers have to answer a series of binary (yes/no) questions
about student performance.
 
36
You might also like М. Н. Аверина, Современный урок иностранного языка, Ярославль, 2004 H. Douglas Brown, Teaching by Principles: An Interactive Approach to Language Pedagogy, San Francisco State University, 1994 PDF 
М. Н. Аверина, Современный урок иностранного языка, Ярославль, 2004 H. Douglas Brown, Teaching by Principles: An Interactive Approach to Language Pedagogy, San Francisco State University, 1994
14 pages