The multiple choice format is perhaps the most familiar test format to many of us. It is also commonly referred to as an objective test as there is seen to be “objectivity” in grading the test. In this blog, we will examine the multiple choice format with respect to its structure, use, and construction.
The structure of the multiple choice format test
Multiple choice items consist of a stem and a set of options; the stem, which is the beginning part of the item that presents the item as a problem to be solved, a question asked of the respondent, or an incomplete statement to be completed, as well as other relevant information and the options, which are the possible answers that the examiner can choose from, with the correct answer called the key and the incorrect answers known as distractors. These components are illustrated in the following diagram:
Figure 1: The structure of the multiple choice questions
In the diagram above, the major components of the multiple choice format are clearly laid out. It should be noted that the number of alternatives in a multiple choice item usually ranges between three to five.
Use of the multiple choice format
There are a number of situations in which a multiple choice format test may be useful and appropriate:-
- When there is a large number of students taking the test
- When you wish to reuse the questions in the test
- When you have to provide the grades quickly
- When highly reliable test scores must be obtained as efficiently as possible
- When impartiality of evaluation, fairness and freedom from possible test scoring influences such as fatigue are essential
- When you are more confident of your ability to construct valid objective test items clearly than of your ability to judge essay test answers fairly
- When you want specific information especially for diagnostic feedback
Advantages
- The ability to create a test item bank
- Quick grading
- High reliability
- Objective grading
- Wide coverage of content
- Precision in providing information regarding specific skills and abilities
Constructing the multiple choice item
There are many issues that need to be considered in constructing MCQ test including how we select distracters in the item as well as how we can construct good test items:-
i) To select good distractors, there are a number of sources from which we can obtain reasonable distractors which are:
o Students previous incorrect responses
o Incorrect form, meaning, style or register
o Translations from the native language
o Variations of the key
o Contextual relevance
Constructing good test items
There are two issues in constructing good test items:-
a) Unintentional cues in the item are as follows:
i) Inconsistent length of alternatives
The way this item is written provides undue attention to the last alternative. As we are interested in measuring the student’s true language ability comprehension, in this case – it is unfair to draw so much of the students’ attention to a single alternative as in this example. If indeed any of the students select this alternative, it will be difficult to determine whether this was because he was unduly attracted to its obviously different length or whether he or she genuinely believes it to be the answer. It is bad enough that option D does this but it is even worse if it is the correct answer. If this is the case, then we will not be able to tell whether the selection of this alternative was because the student actually knew the answer or because he was at a total loss and chose the alternative simply because it stood out.
ii) Inconsistent categories of alternatives
Look at the following list of distractors and consider what cue is provided to the test taker.
In the list of distractors above, all the options except D. talkative are verbs. This draws undue attention to option D and increases the likelihood that it ill be selected regardless of whether or not it is the key.
iii) Convergence cues
The cue that is provided in the following item is a little less obvious and may occur more often without detection. The diagram below consists of the alternatives however, excluding the stem to better illustrate how it may work.
Without the stem, many are predicted to select A. tea because the two options tea and tee sound alike. Between two options, tea would probably be selected instead of tee because all the options are about a gathering of some sort where food is served. Therefore, these cues seem to converge on the option A. tea as the most likely answer.
iv) Nonsense distractors
Unless the students are actually experiencing problems in the language and the teacher intentionally inserts nonsense words into the item as a distrator in order to determine whether or not the student has such problems, the nonsense distractor cue should be avoided. The diagram below shows the possible alternatives of nonsense distractors.
One of the options above which is option B, does not exist in the English language and should be immediately deleted from the list of alternatives.
v) Common knowledge response
Another problem with select type test formats is that the student may select an answer based on common knowledge rather than the skill that we intend to assess. For example, let us consider the following test item:
While it is true that the question may have come from the reading passage, the answer is obvious even to those who have not read and understood the passage.
b) Reaction to the item
The test context can be quite taxing and anxiety causing. Test writers should avoid adding to this anxiety. Instead, they should consciously attempt to help students by reducing anxiety, time wastage and all other unnecessary reactions in the test. Some of these problems in writing test items related to this issue are discussed in this section. They include problems such as redundancy, trick questions, and improper sequencing of alternatives.
i) Redundancy
Some items are written in such way that the students need to read unnecessary information. In a test where time is often a premium, we should avoid such items. The following is an example of a test which includes redundant language.
Each alternatives begins with the phrase because he. It would have been much more helpful to the students if the phrase had been included in the stem so they would not have to spend time reading it over and over again.
ii) Trick questions
In the following example, the item is written in such a way that the student needs to read it over in order to understand what it actually requires.
The use of multiple negatives in the stem – not appropriate, not to be absent – make the stem somewhat confusing. Since our intention in a test is not to confuse the student but rather to measure his ability, this sort of item should be avoided.
iii) Improper sequencing
Disorganised sequencing of alternatives also contribute to unnecessary additional cognitive load for students. if the alternatives are days in a week or months in a year, they should be arranged in chronological order. Similarly, for testing reading comprehension, the alternatives that are taken from the passage should be arranged according to the order that they appear in the passage.
Criticism of the MCQ format
Hughes (1989) lists several other problem with the MCQ test format which are listed as follows:
- The technique test only recognition knowledge and recall of facts
- Guessing may have a considerable but unknowable effect on test scores
- The technique severely restricts what can be tested to only lower order skills.
- It is very difficult to write successful items, due especially to the difficulty of finding good distractors.
- . Backwash can be harmful – for both teaching and learning as preparing for a multiple choice format test is not reflective of good language teaching and learning practice.
- . Cheating may be facilitated
- . It places a high degree of dependence on the student’s reading ability and instructor’s writing ability.
- . It is time consuming to construct.
Hughes (1989:3) goes on to say that “good multiple choice items are notoriously difficult to write. A great deal of time and effort has to go into their construction. Too many multiple choice tests are written where such care and attention is not given (and indeed may not be possible). The result is a set of poor items that cannot possibly provide accurate measurements”
No comments:
Post a Comment