Writing
Writing
Writing
Introduction to Direct Writing Testing
possible
These tasks must be specified clearly in
content the test framework to ensure validity.
1. Operations: The types of actions or functions the
writing should perform (e.g., expressing, directing).
2. Types of Text: The format or genre of the writing
(e.g., letters, reports).
Key Elements 3. Addressees: The intended recipients of the text (e.g.,
in Specifying friends, business partners).
• Conclusion:
The CCSE Certificate framework offers a broad range of writing tasks that are relevant to general
communicative language courses.
• Key Specifications:
Operations: Tasks require students to describe, explain, compare
Example 2 – and contrast, and argue for/against a position.
Types of Text: Short examination answers (up to two paragraphs).
English for
Addressees: University lecturers (both native and non-native
Academic English speakers).
Purposes Topics: Broad, non-specialized academic topics relevant to first-
year undergraduate students.
(EAP) Writing Dialect and Style: Any standard variety of English (e.g., British,
Test American), formal style.
Length: About one page.
• Conclusion:
The test focuses on academic writing tasks that are highly relevant
to university-level English.
(ii) Include • Key Concept:
a Content Specifications: In order to create a
representa valid test, it’s essential to identify all possible
tasks students should be able to perform.
tive The test specifications help ensure that tasks
sample of align with the intended abilities.
• Challenges:
the It's difficult to include every possible task, but
specified test designers should strive for a
representative sample of tasks.
content
The ideal test would cover all relevant tasks to provide a complete
measure of writing ability.
Representative Tasks: If a test includes only a small set of tasks, it may
not fully capture a candidate’s writing ability. However, a broader sample
can improve the validity of the test.
A test with a broad and representative set of tasks is likely to have a
more beneficial backwash effect, improving student learning.
Example of
CCSE Test
(May/June 2000)
Test Overview:
The exam includes
tasks related to a
Summer Camp in
America.
Tasks Include:
1. Task 1: Write a letter to inquire about a
summer camp.
2. Task 2: Fill in an application form.
3. Task 3: Write a postcard to a friend about the
camp.
4. Task 4: Write a note to explain a change of
plans to other camp helpers.
• Conclusion:
The tasks are designed to be a representative
sample of real-world writing situations.
However, they cannot cover all potential
writing tasks, limiting content validity.
Task 1: Write a letter to inquire
about a summer camp.
Task 2: Fill in an
application form.
Task 3: Write a postcard to a friend about the camp.
Task 4: Write a note to explain a change of plans to other camp
helpers.
Balancing Wide Sampling and Practicality
Challenges:
• Wide Sampling: Ideally, a test would cover all possible tasks, but practical
constraints limit this.
• Practicality vs. Accuracy: We must balance the need for a broad sample with the
practicality of constructing the test.
Considerations:
• High Stakes vs. Low Stakes: The importance of the test result influences how
many tasks are needed. High-stakes tests (e.g., university admissions) require a
more representative and accurate sample of writing ability.
1. Identify Relevant Tasks: Use frameworks to specify
key tasks students should be able to perform.
2. Content Validity: Strive for a representative set of
The Complexity tasks, but balance practicality.
of Representative 3. Test Design: High-stakes tests require more extensive
Tasks task sampling to ensure fairness and accuracy.
• Strategy:
Use illustrations (e.g., pictures, charts,
diagrams) to reduce the reliance on reading
comprehension.
• Example:
A diagram of three types of bees could lead
to a task asking students to compare and
contrast them, reducing reading difficulty.
Tasks that Elicit Narrative
Writing
Restrict candidates
Providing Notes and
Information
• Important Consideration:
When using notes or information sheets,
avoid giving candidates too much of the
answer. Full sentences should not be
provided unless necessary.
• Example:
Comparison task: Notes should give key
points but not fully formed sentences to
ensure candidates still need to write their
responses.
• Core Principle:
Tasks should be as authentic as possible to
ensure they reflect real-world writing situations.
• Authenticity Consideration:
Authenticity of A task may be authentic for some candidates
Tasks (e.g., writing to a supervisor), but inauthentic
for others (e.g., language teachers writing to their
supervisor).
• Important: Consider the context and audience
for each task to maintain relevance and
authenticity.
Ensure valid and
reiable scoring
• SET TASKS WHICH CAN BE RELIABLY SCORED
• The tasks selected should allow for consistent, reliable
scoring across different candidates.
Key Characteristics:
Fast: Scoring is completed quickly, often in just minutes.
Used for overall judgment, rather than evaluating specific components.
1. Rapid Scoring: Experienced scorers
can evaluate a one-page composition in
under two minutes.
Advantages of 2. Multiple Scores Per Work: Each piece
Holistic Scoring can be scored multiple times for greater
reliability.
3. Efficiency in High-Volume Tests:
Useful for large-scale assessments like
TOEFL.
Scoring system used in the English-medium
university writing test
Example of • NS: Native speaker standard
Holistic • MA: More than adequate
Scoring • A: Adequate for study
• D: Doubtful
Scale • NA: Not adequate
• FBA: Far below adequate
Purpose: This scale was designed to
determine if a student’s writing ability is
adequate for university study in English.
• Considerations:
Scoring systems must be tailored to the level
of candidates and the purpose of the test.
A scale suitable for university entry may not
Appropriateness be appropriate for other contexts.
of Scoring • Example: The scale used for university tests in
Systems English was designed based on:
1. An examination of undergraduate students’
work.
2. Teacher judgments of acceptable English.
TOEFL Scoring System
Key Features:
Involves assessing multiple components (e.g., grammar,
content, structure, etc.).
Designed to offer a more nuanced understanding of writing
ability.
1. Uneven Development of Subskills: Analytic
scoring addresses the variation in skill
development across different aspects of
writing.
2. Comprehensive Evaluation: Forces scorers
to evaluate aspects they might overlook in
holistic scoring.
3. Greater Reliability: Multiple scores across
components increase the consistency and
reliability of the assessment.
Analytic
Scoring Common Aspects Scored:
1. Content
2. Organization
3. Grammar
4. Vocabulary
5. Mechanics (spelling, punctuation, etc.)
Challenges:
1. Time Consumption: Analytic
scoring takes more time than holistic
scoring due to the need for multiple
evaluations.
Disadvantages of
Analytic Scoring 2. Focus on Parts vs. Whole: Scoring
individual aspects may distract from
the overall impact of the writing,
potentially reducing validity.
o Solution: Impressionistic score
added to balance the overall
assessment.
Issue of Error Gravity
• Problem:
A small number of grammatical errors can
significantly affect communication, while
many errors might not. The frequency of
errors and their impact on communication
may not always correlate.
• Relevance:
This issue is present in both analytic and
holistic scoring systems.
Example: Jacobs et al. (1981) Scoring
• Description:
Five components: Content, Organization, Vocabulary,
Grammar, Mechanics.
Content has the highest weight, Mechanics the lowest.
Score Range: Each component allows a range of scores,
giving scorers flexibility in evaluating performance.
• Considerations:
Diagnostic Information: Analytic scoring is
Choosing essential when detailed insights into specific
areas (like grammar or organization) are
Between needed.
Holistic and Scoring Context:
Analytic o Small, cohesive group: Holistic scoring
might be more time-efficient.
Scoring o Large or less experienced group:
Analytic scoring is preferred for
consistency.
Scales inform candidates about the criteria by
which they will be judged, potentially creating a
backwash effect.
If candidates are aware of the scales, they can focus
on improving the specific criteria emphasized in
the scoring.
Goal: Understand the scoring criteria and see how different types
of work can fit within the same level.
1. Clarify Queries:
o Address any questions or confusion from Stage 1.
2. Calibrated Samples:
o Trainees receive calibrated samples of writing, covering
all levels of the scale.
o They independently assign a score or level to each sample.
Training Stage 2 -
3. Discussion:
Practical
o Review and compare ratings with the trainer’s agreed-
Application upon scores.
o Discuss any discrepancies and ensure consistency in
interpretation.
Feedback
Non-Writing-Specific Feedback Elements
• Key Points:
Incomplete task performance:
1. Topic: Not all parts addressed or treated superficially.
2. Operations: Failure to execute required operations (e.g.,
compare and contrast).
Pointless repetition: When the writing includes unnecessary
repetition, it detracts from the overall quality.
1. Misuse of quotation marks: Incorrect or
inconsistent use of quotations.
2. Inappropriate underlining: Overuse or
misuse of underlining in the text.
3. Capitalization issues: Errors in capitalizing
Writing- words unnecessarily or inconsistently.
Specific 4. Style conventions: Failure to adhere to
Feedback appropriate writing style standards.
Elements 5. Failure to split overlong sentences: Writing
that includes long, run-on sentences that
should be broken up.
6. Inappropriate use of sentence fragments:
Use of incomplete sentences or fragments
inappropriately.
• Hughes, A. (2003). Testing for language
teachers (2nd ed.). Cambridge
References University Press.
Thank You!